855 results
Search Results
2. Presenting the Neurocomputing best paper award, volume 5 (1993)
- Author
-
V.A. David Sánchez
- Subjects
Artificial Intelligence ,Cognitive Neuroscience ,Computer graphics (images) ,Computer Science Applications ,Mathematics ,Volume (compression) - Published
- 1994
3. Towards simulations of long-term behavior of neural networks: Modeling synaptic plasticity of connections within and between human brain regions
- Author
-
Marcus Kaiser, Emmanouil Giannakakis, Bernd Weber, Cheol E. Han, and Frances Hutchings
- Subjects
Brain simulation ,Optimization ,0209 industrial biotechnology ,Brain development ,Traumatic brain injury ,Computer science ,Cognitive Neuroscience ,02 engineering and technology ,Plasticity ,020901 industrial engineering & automation ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,medicine ,Quantitative Biology::Neurons and Cognition ,Artificial neural network ,Human brain ,medicine.disease ,Computer Science Applications ,Brief Papers ,medicine.anatomical_structure ,Biological neural network modeling ,Synaptic plasticity ,Neural mass model ,020201 artificial intelligence & image processing ,Neuroscience - Abstract
Highlights • Development of a biological neural network model that allows long term simulation of brain activity. • Optimization of the model using multiple techniques that led to a speed-up of X200. • Presentation of alternative simulation frameworks for long term simulations., Simulations of neural networks can be used to study the direct effect of internal or external changes on brain dynamics. However, some changes are not immediate but occur on the timescale of weeks, months, or years. Examples include effects of strokes, surgical tissue removal, or traumatic brain injury but also gradual changes during brain development. Simulating network activity over a long time, even for a small number of nodes, is a computational challenge. Here, we model a coupled network of human brain regions with a modified Wilson-Cowan model representing dynamics for each region and with synaptic plasticity adjusting connection weights within and between regions. Using strategies ranging from different models for plasticity, vectorization and a different differential equation solver setup, we achieved one second runtime for one second biological time.
- Published
- 2020
4. FAIR: Fair adversarial instance re-weighting
- Abstract
With growing awareness of societal impact of artificial intelligence, fairness has become an important aspect of machine learning algorithms. The issue is that human biases towards certain groups of popu-lation, defined by sensitive features like race and gender, are introduced to the training data through data collection and labeling. Two important directions of fairness ensuring research have focused on (i) instance weighting in order to decrease the impact of more biased instances and (ii) adversarial training in order to construct data representations informative of the target variable, but uninformative of the sensitive attributes. In this paper we propose a Fair Adversarial Instance Re-weighting (FAIR) method, which uses adversarial training to learn instance weighting function that ensures fair predictions. Merging the two paradigms, it inherits desirable properties from both interpretability of reweighting and end-to-end trainability of adversarial training. We propose four different variants of the method and, among other things, demonstrate how the method can be cast in a fully probabilistic framework. Additionally, theoretical analysis of FAIR models' properties is provided. We compare FAIR models to ten other related and state-of-the-art models and demonstrate that FAIR is able to achieve a better trade-off between accuracy and unfairness. To the best of our knowledge, this is the first model that merges reweighting and adversarial approaches by means of a weighting function that can provide inter-pretable information about fairness of individual instances.
- Published
- 2022
5. Distributional reinforcement learning with unconstrained monotonic neural networks
- Author
-
Théate, Thibaut, Wehenkel, Antoine, Bolland, Adrien, Louppe, Gilles, and Ernst, Damien
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,Artificial Intelligence (cs.AI) ,Computer Science - Artificial Intelligence ,Artificial Intelligence ,Cognitive Neuroscience ,Machine Learning (cs.LG) ,Computer Science Applications - Abstract
The distributional reinforcement learning (RL) approach advocates for representing the complete probability distribution of the random return instead of only modelling its expectation. A distributional RL algorithm may be characterised by two main components, namely the representation of the distribution together with its parameterisation and the probability metric defining the loss. The present research work considers the unconstrained monotonic neural network (UMNN) architecture, a universal approximator of continuous monotonic functions which is particularly well suited for modelling different representations of a distribution. This property enables the efficient decoupling of the effect of the function approximator class from that of the probability metric. The research paper firstly introduces a methodology for learning different representations of the random return distribution (PDF, CDF and QF). Secondly, a novel distributional RL algorithm named unconstrained monotonic deep Q-network (UMDQN) is presented. To the authors' knowledge, it is the first distributional RL method supporting the learning of three, valid and continuous representations of the random return distribution. Lastly, in light of this new algorithm, an empirical comparison is performed between three probability quasi-metrics, namely the Kullback-Leibler divergence, Cramer distance, and Wasserstein distance. The results highlight the main strengths and weaknesses associated with each probability metric together with an important limitation of the Wasserstein distance., Comment: Research paper accepted for publication in the peer-reviewed Neurocomputing journal edited by Elsevier
- Published
- 2023
6. Bid optimization using maximum entropy reinforcement learning
- Author
-
Mengjuan Liu, Jinyu Liu, Zhengning Hu, Yuchen Ge, and Xuyun Nie
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,Artificial Intelligence (cs.AI) ,Computer Science - Artificial Intelligence ,Computer Science - Computer Science and Game Theory ,Artificial Intelligence ,Cognitive Neuroscience ,TheoryofComputation_GENERAL ,Machine Learning (cs.LG) ,Computer Science and Game Theory (cs.GT) ,Computer Science Applications - Abstract
Real-time bidding (RTB) has become a critical way of online advertising. In RTB, an advertiser can participate in bidding ad impressions to display its advertisements. The advertiser determines every impression's bidding price according to its bidding strategy. Therefore, a good bidding strategy can help advertisers improve cost efficiency. This paper focuses on optimizing a single advertiser's bidding strategy using reinforcement learning (RL) in RTB. Unfortunately, it is challenging to optimize the bidding strategy through RL at the granularity of impression due to the highly dynamic nature of the RTB environment. In this paper, we first utilize a widely accepted linear bidding function to compute every impression's base price and optimize it by a mutable adjustment factor derived from the RTB auction environment, to avoid optimizing every impression's bidding price directly. Specifically, we use the maximum entropy RL algorithm (Soft Actor-Critic) to optimize the adjustment factor generation policy at the impression-grained level. Finally, the empirical study on a public dataset demonstrates that the proposed bidding strategy has superior performance compared with the baselines.
- Published
- 2022
7. Deep transfer learning-based gaze tracking for behavioral activity recognition
- Author
-
De Lope Asiaín, Javier, Graña Romay, Manuel María, and European Commission
- Subjects
human activity recognition ,Artificial Intelligence ,Cognitive Neuroscience ,deep transfer learning ,gaze ethogram ,gaze tracking ,Computer Science Applications - Abstract
Computational Ethology studies focused on human beings is usually referred as Human Activity Recognition (HAR). Specifically, this paper belongs to a line of work on the identification of broad cognitive activities that users carry out with computers. The keystone of this kind of systems is the noninvasive detection of the subject's gaze fixations in selected display areas. Noninvasiveness is ensured by using the conventional laptop cameras without additional illumination or tracking devices. The gaze ethograms, composed as sequences of gaze fixations, are the basis to identify the user activities. To determine the gaze fixation display areas with the highest accuracy, this paper explores the use of a transfer learning approach applied to several well-known deep learning network (DLN) architectures whose input is the eye area extracted from the face image,and output is the identification of the gaze fixation area in the computer screen. Two different datasets are created and used in the validation experiments. We report encouraging results that may allow the general use of the system. This work has been supported by FEDER funds through MINECO project TIN2017-85827-P. This project has received funding from the European Union's Horizon 2020 research and innovation programme under the Marie Sklodowska-Curie grant agreement No. 777720. XinZhe Jin contributed some early computational experiences.
- Published
- 2022
8. Learning graph normalization for graph neural networks
- Author
-
Chen, Yihao, Tang, Xin, Qi, Xianbiao, Li, Chun-Guang, and Xiao, Rong
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,Artificial Intelligence ,Computer Vision and Pattern Recognition (cs.CV) ,Cognitive Neuroscience ,Computer Science - Computer Vision and Pattern Recognition ,Machine Learning (cs.LG) ,Computer Science Applications - Abstract
Graph Neural Networks (GNNs) have attracted considerable attention and have emerged as a new promising paradigm to process graph-structured data. GNNs are usually stacked to multiple layers and the node representations in each layer are computed through propagating and aggregating the neighboring node features with respect to the graph. By stacking to multiple layers, GNNs are able to capture the long-range dependencies among the data on the graph and thus bring performance improvements. To train a GNN with multiple layers effectively, some normalization techniques (e.g., node-wise normalization, batch-wise normalization) are necessary. However, the normalization techniques for GNNs are highly task-relevant and different application tasks prefer to different normalization techniques, which is hard to know in advance. To tackle this deficiency, in this paper, we propose to learn graph normalization by optimizing a weighted combination of normalization techniques at four different levels, including node-wise normalization, adjacency-wise normalization, graph-wise normalization, and batch-wise normalization, in which the adjacency-wise normalization and the graph-wise normalization are newly proposed in this paper to take into account the local structure and the global structure on the graph, respectively. By learning the optimal weights, we are able to automatically select a single best or a best combination of multiple normalizations for a specific task. We conduct extensive experiments on benchmark datasets for different tasks, including node classification, link prediction, graph classification and graph regression, and confirm that the learned graph normalization leads to competitive results and that the learned weights suggest the appropriate normalization techniques for the specific task. Source code is released here https://github.com/cyh1112/GraphNormalization., Comment: 15 pages, 3 figures, 6 tables
- Published
- 2022
9. A survey of deep learning approaches to image restoration
- Author
-
Jingwen Su, Boyan Xu, and Hujun Yin
- Subjects
Image restoration ,Artificial Intelligence ,Deblurring ,Super-resolution ,Cognitive Neuroscience ,Deep learning ,Computer Science Applications - Abstract
In this paper, we present an extensive review on deep learning methods for image restoration tasks. Deep learning techniques, led by convolutional neural networks, have received a great deal of attention in almost all areas of image processing, especially in image classification. However, image restoration is a fundamental and challenging topic and plays significant roles in image processing, understanding and representation. It typically addresses image deblurring, denoising, dehazing and super-resolution. There are substantial differences in the approaches and mechanisms in deep learning methods for image restoration. Discriminative learning based methods are able to deal with issues of learning a restoration mapping function effectively, while optimisation models based methods can further enhance the performance with certain learning constraints. In this paper, we offer a comparative study of deep learning techniques in image denoising, deblurring, dehazing, and super-resolution, and summarise the principles involved in these tasks from various supervised deep network architectures, residual or skip connection and receptive field to unsupervised autoencoder mechanisms. Image quality criteria are also reviewed and their roles in image restoration are assessed. Based on our analysis, we further present an efficient network for deblurring and a couple of multi-objective training functions for super-resolution restoration tasks. The proposed methods are compared extensively with the state-of-the-art methods with both quantitative and qualitative analyses. Finally, we point out potential challenges and directions for future research.
- Published
- 2022
10. The KIV model - nonlinear spatio-temporal dynamics of the primordial vertebrate forebrain
- Author
-
Kozma, R, Freeman, Walter J, III, and Erdi, P
- Subjects
neurodynamics ,chaos ,cortex ,hippocampus ,spatio-temporal EEG - Abstract
EEG measurements indicate the presence of common-mode, coherent oscillations in various cortical areas. In previous studies the KIII model has been introduced, which interprets the experimental observation as nonlinear, spatially distributed dynamical oscillations of coupled neural populations. In this paper we combine multiple KIII sets into the KIV model, which approximates the operation of the basic vertebrate forebrain together with the basal ganglia and motor systems. This paper outlines a summary description of the essential components of the KIV model, as the basis for future modeling of their cooperative dynamics guided by analysis of multichannel EEG in animals and humans. (C) 2002 Elsevier Science B.V. All rights reserved.
- Published
- 2003
11. Smart surgical control under RCM constraint using bio-inspired network
- Author
-
Ameer Tamoor Khan and Shuai Li
- Subjects
Test bench ,ComputingMethodologies_SIMULATIONANDMODELING ,Computer science ,Cognitive Neuroscience ,Control (management) ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Computer Science Applications ,Task (project management) ,Constraint (information theory) ,Artificial Intelligence ,Nonlinear model ,Point (geometry) ,MATLAB ,computer ,Surgical robot ,Simulation ,ComputingMethodologies_COMPUTERGRAPHICS ,computer.programming_language - Abstract
In this paper, we propose a control framework for intelligent surgical robots under the Remote Center of Motion (RCM). The goal of a surgical robot is to assist surgeons in performing complex surgeries. RCM constraint implies that the surgical tip attached to the end-effector of the surgical robot does not slide away from the point of the incision while performing surgery. Implementation of a control algorithm to comply with RCM constraints is a complicated task because of the nonlinear model of the surgical robots and stringent conditions of accuracy imposed by the patient’s safety. This paper proposes an optimization-driven approach to perform the surgical maneuver under RCM constraints. We then applied a bio-inspired optimization algorithm to solve the problem efficiently. For testing the performance of ZNNBAS, we used MATLAB to simulate a surgical procedure. A 7-DOF surgical robot (KUKA LBR IIWA 7) was used as a test bench for running the simulations. The simulation results show that the ZNNBAS is comparable with BAS, PSO, and GA and efficiently and robustly performed the task commanded maneuvers while enforcing the RCM constraints.
- Published
- 2022
12. Convex formulation for multi-task L1-, L2-, and LS-SVMs
- Author
-
Ruiz Pastor, Carlos, Alaiz Gudín, Carlos María, Dorronsoro Ibero, José Ramón, and UAM. Departamento de Ingeniería Informática
- Subjects
Informática ,LS-SVM ,L1-SVM ,Computer science ,Convex model combination ,Cognitive Neuroscience ,Regular polygon ,Computer Science Applications ,Dual (category theory) ,Support vector machine ,L2-SVM ,Task (computing) ,Multi-task learning ,Artificial Intelligence ,Optimal combination ,Leverage (statistics) ,Convex combination ,Algorithm - Abstract
Quite often a machine learning problem lends itself to be split in several well-defined subproblems, or tasks. The goal of Multi-Task Learning (MTL) is to leverage the joint learning of the problem from two different perspectives: on the one hand, a single, overall model, and on the other hand task-specific models. In this way, the found solution by MTL may be better than those of either the common or the task-specific models. Starting with the work of Evgeniou et al., support vector machines (SVMs) have lent themselves naturally to this approach. This paper proposes a convex formulation of MTL for the L1-, L2- and LS-SVM models that results in dual problems quite similar to the single-task ones, but with multi-task kernels; in turn, this makes possible to train the convex MTL models using standard solvers. As an alternative approach, the direct optimal combination of the already trained common and task-specific models can also be considered. In this paper, a procedure to compute the optimal combining parameter with respect to four different error functions is derived. As shown experimentally, the proposed convex MTL approach performs generally better than the alternative optimal convex combination, and both of them are better than the straight use of either common or task-specific models, With partial support from Spain’s grant TIN2016-76406-P. Work supported also by the UAM–ADIC Chair for Data Science and Machine Learning.
- Published
- 2021
13. Robot learning system based on dynamic movement primitives and neural network
- Author
-
Miao Li, Ying Zhang, and Chenguang Yang
- Subjects
0209 industrial biotechnology ,Artificial neural network ,business.industry ,Computer science ,Generalization ,Cognitive Neuroscience ,Process (computing) ,Motion controller ,02 engineering and technology ,Robot learning ,Computer Science Applications ,Computer Science::Robotics ,020901 industrial engineering & automation ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,Trajectory ,Robot ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,Robotic arm - Abstract
In the process of Human-robot skill transfer, we require the robot to reproduce the trajectory of teacher and expect that the robot can generalize the learned trajectory. For the trajectory after generalization, we expect that the robot arm can accurately track. However, because the model of the robot can not be accurately obtained, some researchers have proposed using a neural network to approximate the unknown term. The parameters of the traditional RBF neural network are usually selected through the empirical and trial-and-error method, which maybe biased and inefficient. In addition, due to the end-effector of the mechanical arm trajectory will be constantly changing according to the needs of the task, when the neural network of compact set cannot contain the whole input vector, the neural network cannot achieve the ideal approximation effect. In this paper, the broad neural network is used to approximate the unknown terms of the robot. This method can reuse the motion controller that has been learned and complete other motions in the robot operating space without relearning its weight parameters. In this paper, the effectiveness of the proposed method is proved by the ultrasound scanning task.
- Published
- 2021
14. An evolving neuro-fuzzy system based on uni-nullneurons with advanced interpretability capabilities
- Author
-
Edwin Lughofer and Paulo Vitor de Campos Souza
- Subjects
Data stream ,0209 industrial biotechnology ,Neuro-fuzzy ,Artificial neural network ,business.industry ,Active learning (machine learning) ,Computer science ,Cognitive Neuroscience ,Context (language use) ,02 engineering and technology ,Fuzzy control system ,Computer Science Applications ,020901 industrial engineering & automation ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,Feature (machine learning) ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,Interpretability - Abstract
This paper proposes a hybrid architecture based on neural networks, fuzzy systems, and n-uninorms for solving pattern classification problems, termed as ENFS-Uni0 (short for evolving neuro-fuzzy system based on uni-nullneurons). The model can produce knowledge in an on-line (single-pass) and evolving learning context in a particular form of neuro-fuzzy rules representing the dependencies among input features through IF-THEN type relations. The rules antecedents are thereby realized through uni-nullneurons, which are constructed from n-uninorms, leading to the possibility to express both, AND- and OR-connections (and a mixture of these) among the single antecedent parts of a rule (and thus achieving an advanced interpretability aspect of the rules). The neurons’ evolution is done through an extended version of an autonomous data partition method (ADPA). On-line interpretation of the timely evolution of rules is addressed by (i) a concept for tracking the degree of changes of the rules over data stream samples, which may indicate experts/operators how much dynamics is in the process and may be used as a structural active learning component to request operator’s feedback in the case of significant changes and (ii) a concept for updating feature weights incrementally. These weights express the (possibly changing) impact degrees of features on the classification problem: features with low weights can be seen as unimportant and masked out when showing rules to an expert ( → rule length reduction). The rules’ consequents are represented by certainty vectors and are recursively updated by an indicator-based recursive weighted least squares (I-RWLS) approach (one RWLS estimator per class) where the weights are given through the neuron activation levels in order to gain stable local learning. The model proposed in this paper was successfully compared to related hybrid and evolving approaches in the literature for classifying binary and multi-class patterns. The results obtained by the proposed model show an outperformance of the related works in terms of higher accuracy trend lines over time, while offering a high degree of interpretability through coherent neuro-fuzzy rules to solve the classification problems.
- Published
- 2021
15. A survey: Deep learning for hyperspectral image classification with few labeled samples
- Author
-
Sen Jia, Nanying Li, Shuguo Jiang, Meng Xu, Zhijie Lin, and Shiqi Yu
- Subjects
FOS: Computer and information sciences ,0209 industrial biotechnology ,Source code ,Computer Science - Artificial Intelligence ,Computer science ,Active learning (machine learning) ,Computer Vision and Pattern Recognition (cs.CV) ,Cognitive Neuroscience ,media_common.quotation_subject ,Computer Science - Computer Vision and Pattern Recognition ,02 engineering and technology ,Machine learning ,computer.software_genre ,Field (computer science) ,020901 industrial engineering & automation ,Artificial Intelligence ,FOS: Electrical engineering, electronic engineering, information engineering ,0202 electrical engineering, electronic engineering, information engineering ,Hyperspectral image classification ,media_common ,business.industry ,Deep learning ,Image and Video Processing (eess.IV) ,Hyperspectral imaging ,Electrical Engineering and Systems Science - Image and Video Processing ,Computer Science Applications ,Artificial Intelligence (cs.AI) ,020201 artificial intelligence & image processing ,Artificial intelligence ,Transfer of learning ,Focus (optics) ,business ,computer - Abstract
With the rapid development of deep learning technology and improvement in computing capability, deep learning has been widely used in the field of hyperspectral image (HSI) classification. In general, deep learning models often contain many trainable parameters and require a massive number of labeled samples to achieve optimal performance. However, in regard to HSI classification, a large number of labeled samples is generally difficult to acquire due to the difficulty and time-consuming nature of manual labeling. Therefore, many research works focus on building a deep learning model for HSI classification with few labeled samples. In this article, we concentrate on this topic and provide a systematic review of the relevant literature. Specifically, the contributions of this paper are twofold. First, the research progress of related methods is categorized according to the learning paradigm, including transfer learning, active learning and few-shot learning. Second, a number of experiments with various state-of-the-art approaches has been carried out, and the results are summarized to reveal the potential research directions. More importantly, it is notable that although there is a vast gap between deep learning models (that usually need sufficient labeled samples) and the HSI scenario with few labeled samples, the issues of small-sample sets can be well characterized by fusion of deep learning methods and related techniques, such as transfer learning and a lightweight model. For reproducibility, the source codes of the methods assessed in the paper can be found at https://github.com/ShuGuoJ/HSI-Classification.git.
- Published
- 2021
16. Combining pretrained CNN feature extractors to enhance clustering of complex natural images
- Author
-
Stéphane Thiery, Joris Guérin, Olivier Gibaru, Eric Nyiri, and Byron Boots
- Subjects
Computer Science - Machine Learning ,0209 industrial biotechnology ,Point (typography) ,Contextual image classification ,business.industry ,Computer science ,Cognitive Neuroscience ,Feature extraction ,Computer Science - Computer Vision and Pattern Recognition ,Pattern recognition ,02 engineering and technology ,Convolutional neural network ,Computer Science Applications ,Image (mathematics) ,Task (computing) ,020901 industrial engineering & automation ,Artificial Intelligence ,Feature (computer vision) ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,Cluster analysis ,business - Abstract
Recently, a common starting point for solving complex unsupervised image classification tasks is to use generic features, extracted with deep Convolutional Neural Networks (CNN) pretrained on a large and versatile dataset (ImageNet). However, in most research, the CNN architecture for feature extraction is chosen arbitrarily, without justification. This paper aims at providing insight on the use of pretrained CNN features for image clustering (IC). First, extensive experiments are conducted and show that, for a given dataset, the choice of the CNN architecture for feature extraction has a huge impact on the final clustering. These experiments also demonstrate that proper extractor selection for a given IC task is difficult. To solve this issue, we propose to rephrase the IC problem as a multi-view clustering (MVC) problem that considers features extracted from different architectures as different "views" of the same data. This approach is based on the assumption that information contained in the different CNN may be complementary, even when pretrained on the same data. We then propose a multi-input neural network architecture that is trained end-to-end to solve the MVC problem effectively. This approach is tested on nine natural image datasets, and produces state-of-the-art results for IC., Comment: 21 pages, 16 figures, 10 tables, preprint of our paper published in Neurocomputing
- Published
- 2021
17. Revisiting paraphrase question generator using pairwise discriminator
- Author
-
Vinod K Kurmi, Vinay P. Namboodiri, Badri N. Patro, and Dev Chauhan
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,0209 industrial biotechnology ,Discriminator ,Computer science ,Cognitive Neuroscience ,Context (language use) ,02 engineering and technology ,computer.software_genre ,Paraphrase ,Machine Learning (cs.LG) ,020901 industrial engineering & automation ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,Computer Science - Computation and Language ,business.industry ,Sentiment analysis ,Function (mathematics) ,Computer Science Applications ,020201 artificial intelligence & image processing ,Pairwise comparison ,Artificial intelligence ,business ,Computation and Language (cs.CL) ,computer ,Natural language processing ,Sentence ,Generator (mathematics) - Abstract
In this paper, we propose a method for obtaining sentence-level embeddings. While the problem of securing word-level embeddings is very well studied, we propose a novel method for obtaining sentence-level embeddings. This is obtained by a simple method in the context of solving the paraphrase generation task. If we use a sequential encoder-decoder model for generating paraphrase, we would like the generated paraphrase to be semantically close to the original sentence. One way to ensure this is by adding constraints for true paraphrase embeddings to be close and unrelated paraphrase candidate sentence embeddings to be far. This is ensured by using a sequential pair-wise discriminator that shares weights with the encoder that is trained with a suitable loss function. Our loss function penalizes paraphrase sentence embedding distances from being too large. This loss is used in combination with a sequential encoder-decoder network. We also validated our method by evaluating the obtained embeddings for a sentiment analysis task. The proposed method results in semantic embeddings and outperforms the state-of-the-art on the paraphrase generation and sentiment analysis task on standard datasets. These results are also shown to be statistically significant., Comment: This work is an extension of our COLING-2018 paper arXiv:1806.00807
- Published
- 2021
18. Density-adaptive kernel based efficient reranking approaches for person reidentification
- Author
-
Ruopei Guo, Jiaru Lin, Jun Guo, Yonghua Li, and Chunguang Li
- Subjects
FOS: Computer and information sciences ,0209 industrial biotechnology ,Computer science ,business.industry ,Computer Vision and Pattern Recognition (cs.CV) ,Cognitive Neuroscience ,Adaptive kernel ,Computer Science - Computer Vision and Pattern Recognition ,02 engineering and technology ,Machine learning ,computer.software_genre ,Computer Science Applications ,Kernel (linear algebra) ,020901 industrial engineering & automation ,Ranking ,Artificial Intelligence ,Kernel (statistics) ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer - Abstract
Person reidentification (ReID) refers to the task of verifying the identity of a pedestrian observed from nonoverlapping views in a surveillance camera network. It has recently been validated that reranking can achieve remarkable performance improvements in person ReID systems. However, current reranking approaches either require feedback from users or suffer from burdensome computational costs. In this paper, we propose to exploit a density-adaptive smooth kernel technique to achieve efficient and effective reranking. Specifically, we adopt a smooth kernel function to formulate the neighbor relationships among data samples with a density-adaptive parameter. Based on this new formulation, we present two simple yet effective reranking methods, termed \emph{inverse} density-adaptive kernel based reranking (inv-DAKR) and \emph{bidirectional} density-adaptive kernel based reranking (bi-DAKR), in which the local density information in the vicinity of each gallery sample is elegantly exploited. Moreover, we extend the proposed inv-DAKR and bi-DAKR methods to incorporate the available extra probe samples and demonstrate that when and why these extra probe samples are able to improve the local neighborhood and thus further refine the ranking results. Extensive experiments are conducted on six benchmark datasets, including: PRID450s, VIPeR, CUHK03, GRID, Market-1501 and Mars. The experimental results demonstrate that our proposals are effective and efficient., 39 pages, 18 figures and 12 tables. This paper is an extended version of our preliminary work on ICPR 2018
- Published
- 2020
19. Reducing human efforts in video segmentation annotation with reinforcement learning
- Author
-
András Lőrincz and Viktor Varga
- Subjects
0209 industrial biotechnology ,Computer science ,business.industry ,Cognitive Neuroscience ,Deep learning ,02 engineering and technology ,Machine learning ,computer.software_genre ,Computer Science Applications ,Task (project management) ,Reduction (complexity) ,Annotation ,020901 industrial engineering & automation ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,Selection (linguistics) ,Reinforcement learning ,020201 artificial intelligence & image processing ,Segmentation ,Artificial intelligence ,business ,computer - Abstract
Manual annotation of video segmentation datasets requires an immense amount of human effort, thus, reduction of human annotation costs is an active topic of research. While many papers deal with the propagation of masks through frames of a video, only a few results attempt to optimize annotation task selection. In this paper we present a deep learning based solution to the latter problem and train it using Reinforcement Learning. Our approach utilizes a modified version of the Dueling Deep Q-Network sharing weight parameters across the temporal axis of the video. This technique enables the trained agent to select annotation tasks from the whole video. We evaluate our annotation task selection method by means of a hierarchical supervoxel segmentation based mask propagation algorithm.
- Published
- 2020
20. Weighted sum synchronization of memristive coupled neural networks
- Author
-
Chunhua Wang, Yichuang Sun, Chao Zhou, and Wei Yao
- Subjects
Lyapunov function ,0209 industrial biotechnology ,Artificial neural network ,Computer science ,Cognitive Neuroscience ,Node (networking) ,Intermittent control ,02 engineering and technology ,Function (mathematics) ,Topology ,Synchronization ,Computer Science Applications ,symbols.namesake ,020901 industrial engineering & automation ,Artificial Intelligence ,Control theory ,Control system ,Synchronization (computer science) ,0202 electrical engineering, electronic engineering, information engineering ,symbols ,020201 artificial intelligence & image processing ,Differential inequalities - Abstract
It is well known that weighted sum of node states plays an essential role in function implementation of neural networks. Therefore, this paper proposes a new weighted sum synchronization model for memristive neural networks. Unlike the existing synchronization models of memristive neural networks which control each network node to reach synchronization, the proposed model treats the networks as dynamic entireties by weighted sum of node states and makes the entireties instead of each node reach expected synchronization. In this paper, weighted sum complete synchronization and quasi-synchronization are both investigated by designing feedback controller and aperiodically intermittent controller, respectively. Meanwhile, a flexible control scheme is designed for the proposed model by utilizing some switching parameters and can improve anti-interference ability of control system. By applying Lyapunov method and some differential inequalities, some effective criteria are derived to ensure the synchronizations of memristive neural networks. Moreover, the error level of the quasi-synchronization is given. Finally, numerical simulation examples are used to certify the effectiveness of the derived results.
- Published
- 2020
21. Deep multi-center learning for face alignment
- Author
-
Lizhuang Ma, Yangyang Hao, Xin Tan, Zhiwen Shao, and Hengliang Zhu
- Subjects
FOS: Computer and information sciences ,0209 industrial biotechnology ,Computer science ,business.industry ,Computer Vision and Pattern Recognition (cs.CV) ,Cognitive Neuroscience ,Deep learning ,Computer Science - Computer Vision and Pattern Recognition ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Pattern recognition ,02 engineering and technology ,Computer Science Applications ,020901 industrial engineering & automation ,Artificial Intelligence ,Face (geometry) ,0202 electrical engineering, electronic engineering, information engineering ,Code (cryptography) ,020201 artificial intelligence & image processing ,Center (algebra and category theory) ,Artificial intelligence ,Layer (object-oriented design) ,business - Abstract
Facial landmarks are highly correlated with each other since a certain landmark can be estimated by its neighboring landmarks. Most of the existing deep learning methods only use one fully-connected layer called shape prediction layer to estimate the locations of facial landmarks. In this paper, we propose a novel deep learning framework named Multi-Center Learning with multiple shape prediction layers for face alignment. In particular, each shape prediction layer emphasizes on the detection of a certain cluster of semantically relevant landmarks respectively. Challenging landmarks are focused firstly, and each cluster of landmarks is further optimized respectively. Moreover, to reduce the model complexity, we propose a model assembling method to integrate multiple shape prediction layers into one shape prediction layer. Extensive experiments demonstrate that our method is effective for handling complex occlusions and appearance variations with real-time performance. The code for our method is available at https://github.com/ZhiwenShao/MCNet-Extension., This paper has been accepted by Neurocomputing
- Published
- 2020
22. Non-unique decision differential entropy-based feature selection
- Author
-
Ansheng Deng, Qiang Shen, Yanpeng Qu, Changjing Shang, and Rong Li
- Subjects
0209 industrial biotechnology ,Computer science ,Entropy (statistical thermodynamics) ,business.industry ,Cognitive Neuroscience ,Feature selection ,Pattern recognition ,02 engineering and technology ,Computer Science Applications ,Differential entropy ,020901 industrial engineering & automation ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,Entropy (information theory) ,020201 artificial intelligence & image processing ,Rough set ,Artificial intelligence ,Entropy (energy dispersal) ,business - Abstract
Feature selection plays an important role in reducing irrelevant and redundant features, while retaining the underlying semantics of selected ones. An effective feature selection method is expected to result in a significantly reduced subset of the original features without sacrificing the quality of problem-solving (e.g., classification). In this paper, a non-unique decision measure is proposed that captures the degree of a given feature subset being relevant to different categories. This helps to represent the uncertainty information in the boundary region of a granular model, such as rough sets or fuzzy-rough sets in an efficient manner. Based on this measure, the paper further introduce a differentiation entropy as an evaluator of feature subsets to implement a novel feature selection algorithm. The resulting feature selection method is capable of dealing with either nominal or real-valued data. Experimental results on both benchmark data sets and a real application problem demonstrate that the features selected by the proposed approach outperform those attained by state-of-the-art feature selection techniques, in terms of both the size of feature reduction and the classification accuracy.
- Published
- 2020
23. Deep convolution network based emotion analysis towards mental health care
- Author
-
David Day-Uei Li, Zixiang Fei, Huiyu Zhou, Stephen H. Butler, Winifred Ijomah, Erfu Yang, and Xia Li
- Subjects
0209 industrial biotechnology ,Facial expression ,Computer science ,business.industry ,Process (engineering) ,Cognitive Neuroscience ,Deep learning ,02 engineering and technology ,Machine learning ,computer.software_genre ,RS ,Computer Science Applications ,Convolution ,020901 industrial engineering & automation ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,Evolution of emotion ,business ,computer - Abstract
Facial expressions play an important role during communications, allowing information regarding the emotional state of an individual to be conveyed and inferred. Research suggests that automatic facial expression recognition is a promising avenue of enquiry in mental healthcare, as facial expressions can also reflect an individual's mental state. In order to develop user-friendly, low-cost and effective facial expression analysis systems for mental health care, this paper presents a novel deep convolution network based emotion analysis framework to support mental state detection and diagnosis. The proposed system is able to process facial images and interpret the temporal evolution of emotions through a new solution in which deep features are extracted from the Fully Connected Layer 6 of the AlexNet, with a standard Linear Discriminant Analysis Classifier exploited to obtain the final classification outcome. It is tested against 5 benchmarking databases, including JAFFE, KDEF,CK+, and databases with the images obtained ‘in the wild’ such as FER2013 and AffectNet. Compared with the other state-of-the-art methods, we observe that our method has overall higher accuracy of facial expression recognition. Additionally, when compared to the state-of-the-art deep learning algorithms such as Vgg16, GoogleNet, ResNet and AlexNet, the proposed method demonstrated better efficiency and has less device requirements. The experiments presented in this paper demonstrate that the proposed method outperforms the other methods in terms of accuracy and efficiency which suggests it could act as a smart, low-cost, user-friendly cognitive aid to detect, monitor, and diagnose the mental health of a patient through automatic facial expression analysis.
- Published
- 2020
24. Robust real-time hand detection and localization for space human–robot interaction based on deep learning
- Author
-
Qing Gao, Jinguo Liu, and Zhaojie Ju
- Subjects
0209 industrial biotechnology ,business.industry ,Computer science ,Cognitive Neuroscience ,Deep learning ,hand detection and localization ,deep learning ,02 engineering and technology ,Space (commercial competition) ,Object detection ,Human–robot interaction ,Computer Science Applications ,020901 industrial engineering & automation ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,Robot ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,business ,astronaut assistant robot ,SSD ,Gesture - Abstract
Hand gestures are quite suitable for space human-robot interaction (SHRI) because of their natural and convenient features. While the detection and localization of hands are the premise and foundation for SHRI based on hand gestures. But hand gestures are very complicated and hand sizes are very small in some images. These problems make the robust real-time hand detection and local- ization very difficult. In this paper, a feature-map-fused single shot multibox detector (FF-SSD) which is a deep learning network is designed to deal with the problems of hand detection and localization in SHRI. First, the background of the method is introduced in this paper, including an astronaut assistant robot platform, the difficulties of hand detection and localization, and introduction of the state-of-the-art deep learning networks for object detection and localiza- tion. Then, the FF-SSD is proposed for detecting and localizing hands especially pony-size hands. This network magentatakes into consideration both accuracy and speed with balanced performance. And in the experiment part, the FF-SSD is trained and tested on hand databases which include a homemade database and two public databases. At last, the superiority of the proposed method is demonstrated compared with the state-of-the-art methods.
- Published
- 2020
25. DEVDAN: Deep evolving denoising autoencoder
- Author
-
Andri Ashfahani, Edwin Lughofer, Mahardhika Pratama, Yew-Soon Ong, Ashfahani, Andri, Pratama, Mahardhika, Lughofer, Edwin, Ong, Yew-Soon, and School of Computer Science and Engineering
- Subjects
FOS: Computer and information sciences ,Data stream ,Computer Science - Machine Learning ,0209 industrial biotechnology ,Computer science ,Cognitive Neuroscience ,data streams ,Machine Learning (stat.ML) ,02 engineering and technology ,Machine Learning (cs.LG) ,020901 industrial engineering & automation ,Discriminative model ,Statistics - Machine Learning ,Artificial Intelligence ,denoising autoencoder ,0202 electrical engineering, electronic engineering, information engineering ,Protocol (object-oriented programming) ,incremental learning ,Flexibility (engineering) ,Denoising autoencoder ,business.industry ,Pattern recognition ,Computer Science Applications ,Computer science and engineering [Engineering] ,020201 artificial intelligence & image processing ,Artificial intelligence ,business - Abstract
The Denoising Autoencoder (DAE) enhances the flexibility of the data stream method in exploiting unlabeled samples. Nonetheless, the feasibility of DAE for data stream analytic deserves an in-depth study because it characterizes a fixed network capacity that cannot adapt to rapidly changing environments. Deep evolving denoising autoencoder (DEVDAN), is proposed in this paper. It features an open structure in the generative phase and the discriminative phase where the hidden units can be automatically added and discarded on the fly. The generative phase refines the predictive performance of the discriminative model exploiting unlabeled data. Furthermore, DEVDAN is free of the problem-specific threshold and works fully in the single-pass learning fashion. We show that DEVDAN can find competitive network architecture compared with state-of-the-art methods on the classification task using ten prominent datasets simulated under the prequential test-then-train protocol., This paper has been accepted for publication in Neurocomputing 2019. arXiv admin note: substantial text overlap with arXiv:1809.09081
- Published
- 2020
26. A robust approach to reading recognition of pointer meters based on improved mask-RCNN
- Author
-
Lin Zuo, Zhehan Zhang, He Peilin, and Changhua Zhang
- Subjects
0209 industrial biotechnology ,Computer science ,business.industry ,Cognitive Neuroscience ,02 engineering and technology ,Computer Science Applications ,020901 industrial engineering & automation ,Artificial Intelligence ,Pointer (computer programming) ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,business - Abstract
In this paper, we address a challenging task in real-word applications, i.e., automatic reading recognition for pointer meters, called PRM. 1 This application is valuable in the fields of military, industry, and aerospace. However, the accuracy of recognizing the readings of pointer meters by machine vision is, oftentimes, affected by several factors, such as uneven illumination in each image, large range variation of illumination in different images, complex backgrounds, tilting of pointer meters, image blur, and scale change, resulting in the recognized readings with unacceptable accuracy. In this paper, a new robust approach to reading recognition of pointer meters is proposed. The proposed method consists of three main contributions: (1) constructing a novel deep learning algorithm in which the PrRoIPooling is used in lieu of the RoiAlign in the existing Mask-RCNN, (2) classifying the type of pointer meters while fitting the pointer binary mask, and (3) calculating the readings of pointer meters by the proposed angle method. In addition, we also report and release a new dataset for the community. Experiments show that the new algorithm can significantly improve the accuracy of the recognized readings of pointer meters, meanwhile, the proposed approach is also robust to the natural environments and computationally efficient.
- Published
- 2020
27. SA-Net: A deep spectral analysis network for image clustering
- Author
-
Jinghua Wang and Jianmin Jiang
- Subjects
FOS: Computer and information sciences ,0209 industrial biotechnology ,business.industry ,Computer science ,Computer Vision and Pattern Recognition (cs.CV) ,Cognitive Neuroscience ,Deep learning ,Computer Science - Computer Vision and Pattern Recognition ,Pattern recognition ,02 engineering and technology ,Computer Science Applications ,Interactive Learning ,ComputingMethodologies_PATTERNRECOGNITION ,020901 industrial engineering & automation ,Artificial Intelligence ,Robustness (computer science) ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Spectral analysis ,Artificial intelligence ,Laplacian matrix ,business ,Cluster analysis ,Feature learning - Abstract
Although supervised deep representation learning has attracted enormous attentions across areas of pattern recognition and computer vision, little progress has been made towards unsupervised deep representation learning for image clustering. In this paper, we propose a deep spectral analysis network for unsupervised representation learning and image clustering. While spectral analysis is established with solid theoretical foundations and has been widely applied to unsupervised data mining, its essential weakness lies in the fact that it is difficult to construct a proper affinity matrix and determine the involving Laplacian matrix for a given dataset. In this paper, we propose a SA-Net to overcome these weaknesses and achieve improved image clustering by extending the spectral analysis procedure into a deep learning framework with multiple layers. The SA-Net has the capability to learn deep representations and reveal deep correlations among data samples. Compared with the existing spectral analysis, the SA-Net achieves two advantages: (i) Given the fact that one spectral analysis procedure can only deal with one subset of the given dataset, our proposed SA-Net elegantly integrates multiple parallel and consecutive spectral analysis procedures together to enable interactive learning across different units towards a coordinated clustering model; (ii) Our SA-Net can identify the local similarities among different images at patch level and hence achieves a higher level of robustness against occlusions. Extensive experiments on a number of popular datasets support that our proposed SA-Net outperforms 11 benchmarks across a number of image clustering applications., Comment: arXiv admin note: text overlap with arXiv:2009.05235
- Published
- 2020
28. Impact of fully connected layers on performance of convolutional neural networks for image classification
- Author
-
Snehasis Mukherjee, Viswanath Pulabaigari, Shiv Ram Dubey, and S. H. Shabbeer Basha
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,0209 industrial biotechnology ,Computer science ,Computer Vision and Pattern Recognition (cs.CV) ,Cognitive Neuroscience ,Computer Science - Computer Vision and Pattern Recognition ,Context (language use) ,02 engineering and technology ,Machine learning ,computer.software_genre ,Convolutional neural network ,Machine Learning (cs.LG) ,020901 industrial engineering & automation ,Artificial Intelligence ,FOS: Electrical engineering, electronic engineering, information engineering ,0202 electrical engineering, electronic engineering, information engineering ,Selection (linguistics) ,Neural and Evolutionary Computing (cs.NE) ,Thesaurus (information retrieval) ,Contextual image classification ,business.industry ,Image and Video Processing (eess.IV) ,Process (computing) ,Computer Science - Neural and Evolutionary Computing ,Electrical Engineering and Systems Science - Image and Video Processing ,Computer Science Applications ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer - Abstract
The Convolutional Neural Networks (CNNs), in domains like computer vision, mostly reduced the need for handcrafted features due to its ability to learn the problem-specific features from the raw input data. However, the selection of dataset-specific CNN architecture, which mostly performed by either experience or expertise is a time-consuming and error-prone process. To automate the process of learning a CNN architecture, this paper attempts at finding the relationship between Fully Connected (FC) layers with some of the characteristics of the datasets. The CNN architectures, and recently datasets also, are categorized as deep, shallow, wide, etc. This paper tries to formalize these terms along with answering the following questions. (i) What is the impact of deeper/shallow architectures on the performance of the CNN w.r.t. FC layers?, (ii) How the deeper/wider datasets influence the performance of CNN w.r.t. FC layers?, and (iii) Which kind of architecture (deeper/ shallower) is better suitable for which kind of (deeper/ wider) datasets. To address these findings, we have performed experiments with three CNN architectures having different depths. The experiments are conducted by varying the number of FC layers. We used four widely used datasets including CIFAR-10, CIFAR-100, Tiny ImageNet, and CRCHistoPhenotypes to justify our findings in the context of the image classification problem. The source code of this research is available at https://github.com/shabbeersh/Impact-of-FC-layers., This paper is accepted for publication in Neurocomputing Journal
- Published
- 2020
29. Hybrid multisynchronization of coupled multistable memristive neural networks with time delays
- Author
-
Wei Yao, Yichuang Sun, Chao Zhou, Jinde Cao, and Chunhua Wang
- Subjects
Equilibrium point ,0209 industrial biotechnology ,Artificial neural network ,Computer science ,Cognitive Neuroscience ,Activation function ,02 engineering and technology ,Fixed point ,Topology ,Synchronization ,Computer Science Applications ,Piecewise linear function ,020901 industrial engineering & automation ,Differential inclusion ,Exponential stability ,Artificial Intelligence ,Control theory ,0202 electrical engineering, electronic engineering, information engineering ,Periodic orbits ,020201 artificial intelligence & image processing ,Multistability - Abstract
In this paper, we focus on synchronization issue of coupled multistable memristive neural networks (CMMNNs) with time delay under multiple stable equilibrium states. First, we build delayed CMMNNs consisting of one master subnetwork without controller and N−1 identical slave subnetworks with controllers, and every subnetwork has n nodes. Moreover, this paper investigates multistability of delayed CMMNNs with continuous nonmonotonic piecewise linear activation function (PLAF) owning 2 r + 2 corner points. By using the theorems of differential inclusion and fixed point, sufficient conditions are derived such that master subnetwork of CMMNNs can acquire ( r + 2 ) n exponentially stable equilibrium points, stable periodic orbits or hybrid stable equilibrium states. Then, this paper proposes hybrid multisynchronization of delayed CMMNNs related with various external inputs under multiple stable equilibrium states for the first time. There exist ( r + 2 ) n hybrid multisynchronization manifolds in CMMNNs with different initial conditions and external inputs. Finally, two numerical simulations are given to illustrate the effectiveness of the obtained results.
- Published
- 2019
30. Multi-scale dictionary learning for ocular artifact reduction from single-channel electroencephalograms
- Author
-
Hideki Asoh, Atsunori Kanemura, and Suguru Kanoga
- Subjects
0209 industrial biotechnology ,Artifact (error) ,Channel (digital image) ,medicine.diagnostic_test ,business.industry ,Computer science ,Cognitive Neuroscience ,Eye movement ,02 engineering and technology ,Electroencephalography ,Computer Science Applications ,020901 industrial engineering & automation ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,medicine ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,business - Abstract
This paper addresses two issues toward practical use of wearable electroencephalogram (EEG) measurement devices. Ocular (eye movement and blink) artifacts often contaminate EEGs and deteriorate the performance of EEG-based brain–computer interfaces (BCIs). Although wearable consumer-grade EEG devices with single electrode allow users to operate BCIs conveniently in daily lives, it remains a challenging issue to attenuate ocular artifacts from single-channel measurements without spatial information. Existing ocular artifact reduction methods are, however, not simple enough for single-channel EEG data in the sense that they require an additional reference channel and/or pre-processing for artifact segment detection. Another issue is how to assess the performance of artifact reduction; the existing studies have used their own datasets that are not accessible from other researchers. Then, this paper makes two major contributions. (1) This paper proposes a novel ocular artifact reduction method, multi-scale dictionary learning (MSDL), which operates under single-channel measurements and without artifact segment detection. (2) We also develop a semi-simulation setting for quantitative evaluation with a publicly available EEG dataset. In particular, we employed BCI Competition IV Dataset 2a, on which the proposed method was compared with state-of-art methods. The proposed technique showed the best performance for recovering artifact-reduced waveforms from single-channel data compared to the other artifact reduction methods. The Matlab scripts for semi-simulation data generation and single-channel artifact reduction are available on GitHub .
- Published
- 2019
31. Adaptive online extreme learning machine by regulating forgetting factor by concept drift map
- Author
-
Hualong Yu and Geoffrey I. Webb
- Subjects
0209 industrial biotechnology ,Class (computer programming) ,Concept drift ,Computer science ,business.industry ,Data stream mining ,Generalization ,Cognitive Neuroscience ,Novelty ,02 engineering and technology ,Construct (python library) ,Space (commercial competition) ,Machine learning ,computer.software_genre ,Computer Science Applications ,020901 industrial engineering & automation ,Artificial Intelligence ,Metric (mathematics) ,0202 electrical engineering, electronic engineering, information engineering ,Benchmark (computing) ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer ,Extreme learning machine - Abstract
In online-learning, the data is incrementally received and the distributions from which it is drawn may keep changing over time. This phenomenon is widely known as concept drift. Such changes may affect the generalization of a learned model to future data. This problem may be exacerbated by the form of the drift itself changing over time. Quantitative measures to describe and analyze the concept drift have been proposed in previous work. A description composed from these measures is called a concept drift map. We believe that these maps could be useful for guiding how much knowledge in the old model should be forgotten. Therefore, this paper presents an adaptive online learning model that uses a concept drift map to regulate the forgetting factor of an extreme learning machine. Specifically, when a batch of new instances are labeled, the distribution of each class on each attribute is firstly estimated, and then it is compared with the distribution estimated in the previous batch to calculate the magnitude of concept drift, which is further used to regulate the forgetting factor and to update the learning model. Therefore, the novelty of this paper lies in that a quantitative distance metric between two distributions constructed on continuous attribute space is presented to construct concept drift map which can be further associated with the forgetting factor to make the learning model adapt the concept drift. Experimental results on several benchmark stream data sets show the proposed model is generally superior to several previous algorithms when classifying a variety of data streams subject to drift, indicating its effectiveness and feasibility.
- Published
- 2019
32. Monotonic classification: An overview on algorithms, performance measures and data sets
- Author
-
Pedro Antonio Gutiérrez, Bartosz Krawczyk, Michał Woźniak, Salvador García, and José Ramón Cano
- Subjects
FOS: Computer and information sciences ,Computer Science - Machine Learning ,0209 industrial biotechnology ,Computer Science - Artificial Intelligence ,Computer science ,Cognitive Neuroscience ,media_common.quotation_subject ,Monotonic function ,02 engineering and technology ,Machine learning ,computer.software_genre ,Field (computer science) ,Machine Learning (cs.LG) ,020901 industrial engineering & automation ,Knowledge extraction ,Artificial Intelligence ,Taxonomy (general) ,0202 electrical engineering, electronic engineering, information engineering ,Quality (business) ,Medical diagnosis ,media_common ,business.industry ,Class (biology) ,Computer Science Applications ,Data set ,Artificial Intelligence (cs.AI) ,Valuation of options ,Bankruptcy prediction ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer - Abstract
Currently, knowledge discovery in databases is an essential first step when identifying valid, novel and useful patterns for decision making. There are many real-world scenarios, such as bankruptcy prediction, option pricing or medical diagnosis, where the classification models to be learned need to fulfill restrictions of monotonicity (i.e. the target class label should not decrease when input attributes values increase). For instance, it is rational to assume that a higher debt ratio of a company should never result in a lower level of bankruptcy risk. Consequently, there is a growing interest from the data mining research community concerning monotonic predictive models. This paper aims to present an overview of the literature in the field, analyzing existing techniques and proposing a taxonomy of the algorithms based on the type of model generated. For each method, we review the quality metrics considered in the evaluation and the different data sets and monotonic problems used in the analysis. In this way, this paper serves as an overview of monotonic classification research in specialized literature and can be used as a functional guide for the field.
- Published
- 2019
33. Twin Neural Networks for the classification of large unbalanced datasets
- Author
-
Mayank Sharma, Sumit Soman, Himanshu Pant, and Jayadeva
- Subjects
0209 industrial biotechnology ,Artificial neural network ,Computer science ,business.industry ,Cognitive Neuroscience ,Pattern recognition ,02 engineering and technology ,Machine learning ,computer.software_genre ,Matrix multiplication ,Computer Science Applications ,Support vector machine ,ComputingMethodologies_PATTERNRECOGNITION ,020901 industrial engineering & automation ,Hyperplane ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,Feature (machine learning) ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer - Abstract
Twin Support Vector Machines (TWSVMs) have emerged as an efficient alternative to Support Vector Machines (SVM) for learning from imbalanced datasets. The TWSVM learns two non-parallel classifying hyperplanes by solving a couple of smaller sized problems. However, it is unsuitable for large datasets, as it involves matrix operations. In this paper, we discuss a Twin Neural Network (Twin NN) architecture for learning from large unbalanced datasets. The objective functions of the networks in the Twin NN are designed to realize the idea of the Twin SVM with non-parallel decision boudaries for the respective classes, while also being able to reduce model complexity. The Twin NN optimizes the feature map, allowing for better discrimination between classes. The paper also discusses an extension of the Twin NN for multiclass datasets. This architecture trains as many neural networks as the number of classes, and has the additional advantage that it does not have any hyper-parameter which requires tuning. Results presented in the paper demonstrate that the Twin NN generalizes well and scales well on large unbalanced datasets.
- Published
- 2019
34. A new multi-objective wrapper method for feature selection – Accuracy and stability analysis for BCI
- Author
-
Julio Ortega, P. Martin-Smith, Miguel Damas, John Q. Gan, and Jesús González
- Subjects
0209 industrial biotechnology ,Computer science ,Generalization ,Cognitive Neuroscience ,Evolutionary algorithm ,Stability (learning theory) ,Feature selection ,02 engineering and technology ,Motor imagery ,020901 industrial engineering & automation ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,EEG ,BCI ,Representation (mathematics) ,Brain–computer interface ,business.industry ,Pattern recognition ,Classification ,Computer Science Applications ,Multi-objective problem ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,Stability ,Ensemble - Abstract
Feature selection is an important step in building classifiers for high-dimensional data problems, such as EEG classification for BCI applications. This paper proposes a new wrapper method for feature selection, based on a multi-objective evolutionary algorithm, where the representation of the individuals or potential solutions, along with the breeding operators and objective functions, have been carefully designed to select a small subset of features that has good generalization capability, trying to avoid the over-fitting problems that wrapper methods usually suffer. A novel feature ranking procedure is also proposed in order to analyze the stability of the proposed wrapper method. Four different classification schemes have been applied within the proposed wrapper method in order to evaluate its accuracy and stability for feature selection on a real motor imagery dataset. Experimental results show that the wrapper method presented in this paper is able to obtain very small subsets of features, which are quite stable and also achieve high classification accuracy, regardless of the classifiers used., Project TIN2015-67020-P (Spanish “Ministerio de Economía y Competitividad”), European Regional Development Funds (ERDF)
- Published
- 2019
35. Matrix recovery with implicitly low-rank data
- Author
-
Jianlong Wu, Guangcan Liu, Xingyu Xie, and Jun Wang
- Subjects
FOS: Computer and information sciences ,0209 industrial biotechnology ,Rank (linear algebra) ,Computer science ,Computer Vision and Pattern Recognition (cs.CV) ,Cognitive Neuroscience ,Feature vector ,Computer Science - Computer Vision and Pattern Recognition ,Structure (category theory) ,02 engineering and technology ,Computer Science Applications ,Matrix (mathematics) ,020901 industrial engineering & automation ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Algorithm ,Robust principal component analysis - Abstract
In this paper, we study the problem of matrix recovery, which aims to restore a target matrix of authentic samples from grossly corrupted observations. Most of the existing methods, such as the well-known Robust Principal Component Analysis (RPCA), assume that the target matrix we wish to recover is low-rank. However, the underlying data structure is often non-linear in practice, therefore the low-rankness assumption could be violated. To tackle this issue, we propose a novel method for matrix recovery in this paper, which could well handle the case where the target matrix is low-rank in an implicit feature space but high-rank or even full-rank in its original form. Namely, our method pursues the low-rank structure of the target matrix in an implicit feature space. By making use of the specifics of an accelerated proximal gradient based optimization algorithm, the proposed method could recover the target matrix with non-linear structures from its corrupted version. Comprehensive experiments on both synthetic and real datasets demonstrate the superiority of our method.
- Published
- 2019
36. 3G structure for image caption generation
- Author
-
Xiaoqiang Lu, Aihong Yuan, and Xuelong Li
- Subjects
FOS: Computer and information sciences ,Structure (mathematical logic) ,0209 industrial biotechnology ,Computer science ,business.industry ,Computer Vision and Pattern Recognition (cs.CV) ,Cognitive Neuroscience ,Computer Science - Computer Vision and Pattern Recognition ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,02 engineering and technology ,Computer Science Applications ,Image (mathematics) ,Task (project management) ,020901 industrial engineering & automation ,Recurrent neural network ,Artificial Intelligence ,Feature (computer vision) ,0202 electrical engineering, electronic engineering, information engineering ,Embedding ,020201 artificial intelligence & image processing ,Computer vision ,Artificial intelligence ,business ,Sentence ,Generator (mathematics) - Abstract
It is a big challenge of computer vision to make machine automatically describe the content of an image with a natural language sentence. Previous works have made great progress on this task, but they only use the global or local image feature, which may lose some important subtle or global information of an image. In this paper, we propose a model with 3-gated model which fuses the global and local image features together for the task of image caption generation. The model mainly has three gated structures. 1) Gate for the global image feature, which can adaptively decide when and how much the global image feature should be imported into the sentence generator. 2) The gated recurrent neural network (RNN) is used as the sentence generator. 3) The gated feedback method for stacking RNN is employed to increase the capability of nonlinearity fitting. More specially, the global and local image features are combined together in this paper, which makes full use of the image information. The global image feature is controlled by the first gate and the local image feature is selected by the attention mechanism. With the latter two gates, the relationship between image and text can be well explored, which improves the performance of the language part as well as the multi-modal embedding part. Experimental results show that our proposed method outperforms the state-of-the-art for image caption generation., Comment: 35 pages, 7 figures, magazine
- Published
- 2019
37. Recent advances in convolutional neural network acceleration
- Author
-
Bei Yu, Meng Zhang, Zhifei Sun, Yuzhe Ma, Qianru Zhang, and Tinghuan Chen
- Subjects
0209 industrial biotechnology ,Contextual image classification ,Computer science ,business.industry ,Cognitive Neuroscience ,Inference ,02 engineering and technology ,Machine learning ,computer.software_genre ,Convolutional neural network ,Computer Science Applications ,Acceleration ,Range (mathematics) ,020901 industrial engineering & automation ,Dimension (vector space) ,Artificial Intelligence ,Pattern recognition (psychology) ,0202 electrical engineering, electronic engineering, information engineering ,Feature (machine learning) ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer - Abstract
In recent years, convolutional neural networks (CNNs) have shown great performance in various fields such as image classification, pattern recognition, and multi-media compression. Two of the feature properties, local connectivity and weight sharing, can reduce the number of parameters and increase processing speed during training and inference. However, as the dimension of data becomes higher and the CNN architecture becomes more complicated, the end-to-end approach or the combined manner of CNN is computationally intensive, which becomes limitation to CNN’s further implementation. Therefore, it is necessary and urgent to implement CNN in a faster way. In this paper, we first summarize the acceleration methods that contribute to but not limited to CNN by reviewing a broad variety of research papers. We propose a taxonomy in terms of three levels, i.e. structure level, algorithm level, and implementation level, for acceleration methods. We also analyze the acceleration methods in terms of CNN architecture compression, algorithm optimization, and hardware-based improvement. At last, we give a discussion on different perspectives of these acceleration and optimization methods within each level. The discussion shows that the methods in each level still have large exploration space. By incorporating such a wide range of disciplines, we expect to provide a comprehensive reference for researchers who are interested in CNN acceleration.
- Published
- 2019
38. Consensus of multi-agent systems with faults and mismatches under switched topologies using a delta operator method
- Author
-
Hongbin Zhang, Gang Wang, Dianhao Zheng, and J. Andrew Zhang
- Subjects
0209 industrial biotechnology ,Spanning tree ,Matching (graph theory) ,Computer science ,Cognitive Neuroscience ,Multi-agent system ,Stability (learning theory) ,02 engineering and technology ,Delta operator ,Fault (power engineering) ,Topology ,Network topology ,Computer Science Applications ,Computer Science::Multiagent Systems ,020901 industrial engineering & automation ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing - Abstract
This paper studies the consensus of multi-agent systems with faults and mismatches under switched topologies using a delta operator method. Since faults and mismatches can result in failure of the consensus even for a fixed topology with a spanning tree, how to reach a consensus is a complicated and challenging problem under such circumstances especially when part topologies have no spanning tree. Although some works studied the influence of faults and mismatches on the consensus, there is little work on reaching a consensus for the multi-agent systems with faults and mismatches. In this paper, we introduce the delta operator to unify the consensus analysis for continuous, discrete, or sampled systems under one framework. We develop the theories on the delta operator systems first and then apply theories of the delta operator systems to the consensus problems. By converting the consensus problems into stability problems, we investigate and prove consensus and the associated conditions for systems 1) without any fault, 2) with a known fault, and 3) with unknown faults, under switching topologies with matching or mismatching coefficients. Numerical examples are provided and validate the effectiveness of the theoretical results.
- Published
- 2018
39. Self-adaptive decision-making mechanisms to balance the execution of multiple tasks for a multi-robots team
- Author
-
Floriano De Rango, Amilcare Francesco Santamaria, Xin-She Yang, and Nunzia Palmieri
- Subjects
FOS: Computer and information sciences ,0209 industrial biotechnology ,Optimization problem ,Operations research ,Computer Science - Artificial Intelligence ,Computer science ,Cognitive Neuroscience ,Terrain ,02 engineering and technology ,65D19, 68T40, 78M32, 90C26 ,Computer Science - Robotics ,020901 industrial engineering & automation ,Artificial Intelligence ,Hazardous waste ,Order (exchange) ,0202 electrical engineering, electronic engineering, information engineering ,Neural and Evolutionary Computing (cs.NE) ,Coordination game ,Work (physics) ,Computer Science - Neural and Evolutionary Computing ,Mobile robot ,Computer Science Applications ,Artificial Intelligence (cs.AI) ,Robot ,020201 artificial intelligence & image processing ,Robotics (cs.RO) - Abstract
This work addresses the coordination problem of multiple robots with the goal of finding specific hazardous targets in an unknown area and dealing with them cooperatively. The desired behaviour for the robotic system entails multiple requirements, which may also be conflicting. The paper presents the problem as a constrained bi-objective optimization problem in which mobile robots must perform two specific tasks of exploration and at same time cooperation and coordination for disarming the hazardous targets. These objectives are opposed goals, in which one may be favored, but only at the expense of the other. Therefore, a good trade-off must be found. For this purpose, a nature-inspired approach and an analytical mathematical model to solve this problem considering a single equivalent weighted objective function are presented. The results of proposed coordination model, simulated in a two dimensional terrain, are showed in order to assess the behaviour of the proposed solution to tackle this problem. We have analyzed the performance of the approach and the influence of the weights of the objective function under different conditions: static and dynamic. In this latter situation, the robots may fail under the stringent limited budget of energy or for hazardous events. The paper concludes with a critical discussion of the experimental results., Comment: 40 pages
- Published
- 2018
40. Improving deep neural network with Multiple Parametric Exponential Linear Units
- Author
-
Chunxiao Fan, Yue Ming, Li Yang, Yong Li, and Wu Qiong
- Subjects
FOS: Computer and information sciences ,0209 industrial biotechnology ,Artificial neural network ,Computer science ,Generalization ,business.industry ,Computer Vision and Pattern Recognition (cs.CV) ,Cognitive Neuroscience ,Deep learning ,Activation function ,Computer Science - Computer Vision and Pattern Recognition ,Initialization ,02 engineering and technology ,Rectifier (neural networks) ,Residual ,Computer Science Applications ,Range (mathematics) ,020901 industrial engineering & automation ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,Algorithm ,Parametric statistics - Abstract
Activation function is crucial to the recent successes of deep neural networks. In this paper, we first propose a new activation function, Multiple Parametric Exponential Linear Units (MPELU), aiming to generalize and unify the rectified and exponential linear units. As the generalized form, MPELU shares the advantages of Parametric Rectified Linear Unit (PReLU) and Exponential Linear Unit (ELU), leading to better classification performance and convergence property. In addition, weight initialization is very important to train very deep networks. The existing methods laid a solid foundation for networks using rectified linear units but not for exponential linear units. This paper complements the current theory and extends it to the wider range. Specifically, we put forward a way of initialization, enabling training of very deep networks using exponential linear units. Experiments demonstrate that the proposed initialization not only helps the training process but leads to better generalization performance. Finally, utilizing the proposed activation function and initialization, we present a deep MPELU residual architecture that achieves state-of-the-art performance on the CIFAR-10/100 datasets. The code is available at https://github.com/Coldmooon/Code-for-MPELU .
- Published
- 2018
41. Computer vision and deep learning techniques for pedestrian detection and tracking: A survey
- Author
-
Vitoantonio Bevilacqua, Antonio Brunetti, Gianpaolo Francesco Trotta, and Domenico Buongiorno
- Subjects
Artificial neural network ,0209 industrial biotechnology ,Computer science ,Human tracking ,Cognitive Neuroscience ,Pedestrian detection ,Features extraction ,Convolutional neural network ,02 engineering and technology ,Field (computer science) ,020901 industrial engineering & automation ,Artificial Intelligence ,Robustness (computer science) ,Machine learning ,Deep learning ,0202 electrical engineering, electronic engineering, information engineering ,Computer vision ,business.industry ,Robotics ,Computer Science Applications ,Benchmark (computing) ,020201 artificial intelligence & image processing ,Artificial intelligence ,business - Abstract
Pedestrian detection and tracking have become an important field in the computer vision research area. This growing interest, started in the last decades, might be explained by the multitude of potential applications that could use the results of this research field, e.g. robotics, entertainment, surveillance, care for the elderly and disabled, and content-based indexing. In this survey paper, vision-based pedestrian detection systems are analysed based on their field of application, acquisition technology, computer vision techniques and classification strategies. Three main application fields have been individuated and discussed: video surveillance, human-machine interaction and analysis. Due to the large variety of acquisition technologies, this paper discusses both the differences between 2D and 3D vision systems, and indoor and outdoor systems. The authors reserved a dedicated section for the analysis of the Deep Learning methodologies, including the Convolutional Neural Networks in pedestrian detection and tracking, considering their recent exploding adoption for such a kind systems. Finally, focusing on the classification point of view, different Machine Learning techniques have been analysed, basing the discussion on the classification performances on different benchmark datasets. The reported results highlight the importance of testing pedestrian detection systems on different datasets to evaluate the robustness of the computed groups of features used as input to classifiers.
- Published
- 2018
42. Multistability of delayed neural networks with hard-limiter saturation nonlinearities
- Author
-
M. Grazzini, Mauro Di Marco, Luca Pancioni, and Mauro Forti
- Subjects
Equilibrium point ,Delayed neural networks ,Artificial neural network ,Generalization ,Computer science ,Cognitive Neuroscience ,020208 electrical & electronic engineering ,Linear system ,Cooperative systems ,Hard-limiter saturation nonlinearities ,Multistability ,Nonsmooth dynamical systems ,Computer Science Applications1707 Computer Vision and Pattern Recognition ,Artificial Intelligence ,02 engineering and technology ,Topology ,Computer Science Applications ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Hypercube - Abstract
The paper considers a class of nonsmooth neural networks where hard-limiter saturation nonlinearities are used to constrain solutions of a linear system with concentrated and distributed delays to evolve within a closed hypercube of R n . Such networks are termed delayed linear systems in saturated mode (D-LSSMs) and they are a generalization to the delayed case of a relevant class of neural networks previously introduced in the literature. The paper gives a rigorous foundation to the D-LSSM model and then it provides a fundamental result on convergence of solutions toward equilibrium points in the case where there are nonsymmetric cooperative (nonnegative) interconnections between neurons. The result ensures convergence for any finite value of the maximum delay and is physically robust with respect to perturbations of the interconnections. More importantly, it encompasses situations where there exist multiple stable equilibria, thus guaranteeing multistability of cooperative D-LSSMs. From an application viewpoint the delays in combination with the property of multistability make D-LSSMs potentially useful in the fields of associative memories, motion detection and processing of temporal patterns.
- Published
- 2018
43. Total stability of kernel methods
- Author
-
Dao-Hong Xiang, Andreas Christmann, and Ding-Xuan Zhou
- Subjects
Cognitive Neuroscience ,Hilbert space ,Stability (learning theory) ,010103 numerical & computational mathematics ,Lipschitz continuity ,01 natural sciences ,Computer Science Applications ,010104 statistics & probability ,symbols.namesake ,Kernel method ,Artificial Intelligence ,Kernel (statistics) ,Hyperparameter optimization ,symbols ,Applied mathematics ,Empirical risk minimization ,0101 mathematics ,Probability measure ,Mathematics - Abstract
Regularized empirical risk minimization using kernels and their corresponding reproducing kernel Hilbert spaces (RKHSs) plays an important role in machine learning. However, the actually used kernel often depends on one or on a few hyperparameters or the kernel is even data dependent in a much more complicated manner. Examples are Gaussian RBF kernels, kernel learning, and hierarchical Gaussian kernels which were recently proposed for deep learning. Therefore, the actually used kernel is often computed by a grid search or in an iterative manner and can often only be considered as an approximation to the “ideal” or “optimal” kernel. The paper gives conditions under which classical kernel based methods based on a convex Lipschitz loss function and on a bounded and smooth kernel are stable, if the probability measure P, the regularization parameter λ, and the kernel K may slightly change in a simultaneous manner. Similar results are also given for pairwise learning. Therefore, the topic of this paper is somewhat more general than in classical robust statistics, where usually only the influence of small perturbations of the probability measure P on the estimated function is considered.
- Published
- 2018
44. GVM based intuitive simulation web application for collision detection
- Author
-
Jun Shen, Zebang Shen, Huaming Chen, Xin Wang, Binbin Yong, and Qingguo Zhou
- Subjects
020203 distributed computing ,Theoretical computer science ,Artificial neural network ,Computer science ,business.industry ,Cognitive Neuroscience ,Distributed computing ,020207 software engineering ,02 engineering and technology ,Computer Science Applications ,Scheduling (computing) ,Support vector machine ,Software ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,Web application ,Collision detection ,business - Abstract
Computer simulation, which has been proved to be an effective approach to problem solving, is nowadays widely used in modern science. However, it requires a lot of computing resources, which are difficult for general users to acquire. In this paper, we design a Web based system to implement on-line simulation system for ordinary users. As a useful example, the simulation of one type of collision detection model is presented in this paper. Moreover, the software application of simulation is offered as a service on Web. Meanwhile, the incorporation of general vector machine (GVM, a type of neural network) to intelligently predict the relationship between simulation parameters and computation resources is presented, which could further provide more information for system monitoring and scheduling. The system has demonstrated efficiency and intuitiveness for users of this type of applications.
- Published
- 2018
45. Fuzzy mixed-prototype clustering algorithm for microarray data analysis
- Author
-
Jin Liu, Zhizheng Liang, Hong Yan, and Tuan D. Pham
- Subjects
0209 industrial biotechnology ,Fuzzy clustering ,FMP ,Computer Sciences ,Microarray analysis techniques ,Cognitive Neuroscience ,Pattern analysis ,02 engineering and technology ,computer.software_genre ,Fuzzy logic ,Computer Science Applications ,Data modeling ,Datavetenskap (datalogi) ,ComputingMethodologies_PATTERNRECOGNITION ,020901 industrial engineering & automation ,Microarray data analysis ,Hyperplane ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Data mining ,Cluster analysis ,computer ,Mathematics - Abstract
Being motivated by combining the advantages of hyperplane-based pattern analysis and fuzzy clustering techniques, we present in this paper a fuzzy mix-prototype (FMP) clustering for microarray data analysis. By integrating spherical and hyper-planar cluster prototypes, the FMP is capable of capturing latent data models with both spherical and non-spherical geometric structures. Our contributions of the paper can be summarized into three folds: first, the objective function of the FMP is formulated. Second, an iterative solution which minimizes the objective function under given constraints is derived. Third, the effectiveness of the proposed FMP is demonstrated through experiments on yeast and leukemia data sets.
- Published
- 2018
46. Age-related differences in SSVEP-based BCI performance
- Author
-
Piotr Stawicki, Ivan Volosyak, and Felix Gembler
- Subjects
medicine.medical_specialty ,Computer science ,Cognitive Neuroscience ,Speech recognition ,0206 medical engineering ,02 engineering and technology ,Audiology ,020601 biomedical engineering ,Computer Science Applications ,03 medical and health sciences ,0302 clinical medicine ,Artificial Intelligence ,Age related ,medicine ,Evoked potential ,030217 neurology & neurosurgery ,Brain–computer interface - Abstract
BrainComputer Interface (BCI) systems analyze brain signals to generate control commands for computer applications or external devices. Utilized as alternative communication channel, BCIs have the potential to assist people with severe motor disabilities to interact with their environment and to participate in daily life activities. Handicapped people from all age groups could benefit from such BCI technologies. Although some papers have previously reported slightly worse BCI performance by older subjects, in many studies BCI systems were tested with young subjects only.In the presented paper age-associated differences in BCI performance were investigated. We compared accuracy and speed of a steady-state visual evoked potential (SSVEP)-based BCI spelling application controlled by participants of two different equally sized age groups. Twenty subjects (eleven female and nine male) participated in this study; each age group consisted of ten subjects, ranging from 19 to 27 years and from 64 to 76 years. Our results confirm that elderly people may have a deteriorated information transfer rate (ITR). The mean (SD) ITR of the young age group was 27.36 (6.50) bit/min while the elderly people achieved a significantly lower ITR of 16.10 (5.90) bit/min. The average time window length associated with the signal classification was usually larger for the participants of advanced age. These findings show that the subject age must be taken into account during the development of SSVEP-based applications.
- Published
- 2017
47. Application of self-organizing map to failure modes and effects analysis methodology
- Author
-
Wui Lee Chang, Kai Meng Tay, and Lie Meng Pang
- Subjects
Self-organizing map ,021103 operations research ,business.product_category ,Artificial neural network ,Process (engineering) ,Computer science ,Cognitive Neuroscience ,0211 other engineering and technologies ,02 engineering and technology ,computer.software_genre ,Computer Science Applications ,Visualization ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Data mining ,business ,Cluster analysis ,Failure mode and effects analysis ,computer ,Implementation ,Worksheet - Abstract
In this paper, a self-organizing map (SOM) neural network is used to visualize corrective actions of failure modes and effects analysis (FMEA). SOM is a popular unsupervised neural network model that aims to produce a low-dimensional map (typically a two-dimensional map) for visualizing high-dimensional data. With regards to FMEA, it is a popular methodology to identify potential failure modes for a product or a process, to assess the risk associated with those failure modes, also, to identify and carry out corrective actions to address the most serious concerns. Despite the popularity of FMEA in a wide range of industries, two well-known shortcomings are the complexity of the FMEA worksheet and its intricacy of use. To the best of our knowledge, the use of computation techniques for solving the aforementioned shortcomings is limited. The use of SOM in FMEA is new. In this paper, corrective actions in FMEA are described in their severity, occurrence and detect scores. SOM is then used as a visualization aid for FMEA users to see the relationship among corrective actions via a map. Color information from the SOM map is then included to the FMEA worksheet for better visualization. In addition, a Risk Priority Number Interval is used to allow corrective actions to be evaluated and ordered in groups. Such approach provides a quick and easily understandable framework to elucidate important information from a complex FMEA worksheet; therefore facilitating the decision-making tasks by FMEA users. The significance of this study is two-fold, viz., the use of SOM as an effective neural network learning paradigm to facilitate FMEA implementations, and the use of a computational visualization approach to tackle the two well-known shortcomings of FMEA.
- Published
- 2017
48. Non-fragile H∞ state estimation for nonlinear networked system with probabilistic diverging disturbance and multiple missing measurements
- Author
-
Yongqing Yang, Yan Wang, Li Li, and Linghua Xie
- Subjects
Lyapunov function ,0209 industrial biotechnology ,Stochastic process ,Cognitive Neuroscience ,Probabilistic logic ,Estimator ,02 engineering and technology ,Interval (mathematics) ,Computer Science Applications ,Matrix (mathematics) ,Nonlinear system ,symbols.namesake ,020901 industrial engineering & automation ,Artificial Intelligence ,Control theory ,0202 electrical engineering, electronic engineering, information engineering ,symbols ,020201 artificial intelligence & image processing ,Random variable ,Mathematics - Abstract
This paper is concerned with the non-fragile H ∞ state estimation problem for a class of discrete-time networked system with probabilistic diverging disturbance and multiple missing measurements. The measurement missing phenomenon is assumed to occur randomly and the missing probability for each sensor is governed by an individual random variable satisfying a certain probabilistic distribution over the interval 0 , 1 . The aim of this paper is to estimate the networked system by designing a non-fragile H ∞ estimator such that the augmented estimation error system is asymptotically mean square stable with a prescribed H ∞ disturbance attention level γ. By using the Lyapunov method and stochastic analysis, we derive a sufficient condition for the existence of the desired estimator. By solving the linear matrix inequalities (LMIs), the estimator gain matrix is given. Two numerical examples are employed to demonstrate the effectiveness and applicability of the proposed design technique.
- Published
- 2017
49. Influencing over people with a social emotional model
- Author
-
Damiano Zanardini, Jaime Andres Rincon Arango, Carlos Carrascosa, Fernando De la Prieta, Juan Manuel Corchado, and Vicente Julian
- Subjects
Computer science ,Cognitive Neuroscience ,Emotional contagion ,02 engineering and technology ,Space (commercial competition) ,Emotional competence ,Social group ,03 medical and health sciences ,0302 clinical medicine ,Artificial Intelligence ,ComputerApplications_MISCELLANEOUS ,030225 pediatrics ,0202 electrical engineering, electronic engineering, information engineering ,Social emotional learning ,Emotional expression ,Emotional software agents ,GeneralLiterature_REFERENCE(e.g.,dictionaries,encyclopedias,glossaries) ,business.industry ,Multi-agent system ,Multi-agent systems ,Representation (systemics) ,PAD emotional state model ,Computer Science Applications ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,LENGUAJES Y SISTEMAS INFORMATICOS ,Neural networks ,Cognitive psychology - Abstract
[EN] This paper presents an approach of a social emotional model, which allows to extract the social emotion of a group of intelligent entities. The emotional model PAD allows to represent the emotion of an intelligent entity in 3-D space, allowing the representation of different emotional states. The social emotional model presented in this paper uses individual emotions of each one of the entities, which are represented in the emotional space PAD. Using a social emotional model within intelligent entities allows the creation of more real simulations, in which emotional states can influence decision-making. The result of this social emotional mode is represented by a series of examples, which are intended to represent a number of situations in which the emotions of each individual modify the emotion of the group. Moreover, the paper introduces an example which employs the proposed model in order to learn and predict future actions trying to influence in the social emotion of a group of people., This work is partially supported by the MINECO/FEDER TIN2015-65515-C4-1-R and the FPI grant AP2013-01276 awarded to Jaime-Andres Rincon.
- Published
- 2017
50. Discriminative analysis-synthesis dictionary learning for image classification
- Author
-
Meng Yang, Heyou Chang, and Weixin Luo
- Subjects
K-SVD ,Contextual image classification ,Computer science ,business.industry ,Cognitive Neuroscience ,020206 networking & telecommunications ,Pattern recognition ,Linear classifier ,02 engineering and technology ,Sparse approximation ,Facial recognition system ,Computer Science Applications ,Discriminative model ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,Classifier (UML) - Abstract
Dictionary learning has played an important role in the success of sparse representation. Although discriminative synthesis dictionary learning for sparse representation with a high-computational-complexity l0 or l1 norm constraint has been well studied for image classification, jointly and discriminatively learning an analysis dictionary and a synthesis dictionary is still in its infant stage. As a dual of synthesis dictionary, the recently developed analysis dictionary can provide a complementary view of data representation, which can have a much lower time complexity than sparse synthesis representation. Although several class-specific analysis-synthesis dictionary, which may have a big correlation between different classes' dictionaries, have been developed, how to learn a more compact and discriminative universal analysis-synthesis dictionary is still open. In this paper, to provide a more complete view of discriminative data representation, we propose a novel model of discriminative analysis-synthesis dictionary learning (DASDL), in which a linear classifier based on the coding coefficient is jointly learned with the dictionary pair, thus the performance of the classifier and the representational power of the dictionary pair being considered at the same time by the same optimization procedure. The size of the learned dictionaries can be very small since the analysis-synthesis dictionary is shared by all class data. An iterative algorithm to efficiently solve the proposed DASDL is presented in this paper. The experiments on face recognition, gender classification, action recognition and image classification clearly show the superiority of the proposed DASDL.
- Published
- 2017
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.