40 results on '"Recurrent networks"'
Search Results
2. Complex Recurrent Spectral Network.
- Author
-
Chicchi, Lorenzo, Giambagli, Lorenzo, Buffoni, Lorenzo, Marino, Raffaele, and Fanelli, Duccio
- Subjects
- *
ARTIFICIAL neural networks , *ARTIFICIAL intelligence , *INFORMATION processing , *EIGENVALUES - Abstract
This paper presents a novel approach to advancing artificial intelligence (AI) through the development of the Complex Recurrent Spectral Network (ℂ -RSN), an innovative variant of the Recurrent Spectral Network (RSN) model. The ℂ -RSN model introduces localized non-linearity, complex fixed eigenvalues, and a distinct separation of memory and input processing functionalities. These features enable the ℂ -RSN to evolve towards a dynamic, oscillating final state that bear some degree of similarity with biological cognition. The model's ability to classify data through a time-dependent function, and the localization of information processing, is demonstrated by using the MNIST dataset. Remarkably, distinct items supplied as a sequential input yield patterns in time which bear the indirect imprint of the insertion order (and of the separation in time between contiguous insertions). • The C-RSN overcomes some limitations in existing neural network models. • C-RSN introduces localized non-linearity, complex eigenvalues and a separated memory. • The network evolves towards a dynamic, oscillating final state. • The model's efficacy is demonstrated through empirical evaluation. • C-RSN is able to process sequential inputs keeping track of the insertion order. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
3. An Efficient Recurrent Adversarial Framework for Unsupervised Real-Time Video Enhancement
- Author
-
Dario Fuoli, Zhiwu Huang, Danda Pani Paudel, Luc Van Gool, and Radu Timofte
- Subjects
FOS: Computer and information sciences ,Generative adversarial networks ,Computer Vision and Pattern Recognition (cs.CV) ,Image and Video Processing (eess.IV) ,Computer Science - Computer Vision and Pattern Recognition ,Video enhancement ,Recurrent networks ,Video quality mapping ,Electrical Engineering and Systems Science - Image and Video Processing ,Real-time ,Joint distribution learning ,Artificial Intelligence ,FOS: Electrical engineering, electronic engineering, information engineering ,Computer Vision and Pattern Recognition ,Software - Abstract
Video enhancement is a challenging problem, more than that of stills, mainly due to high computational cost, larger data volumes and the difficulty of achieving consistency in the spatio-temporal domain. In practice, these challenges are often coupled with the lack of example pairs, which inhibits the application of supervised learning strategies. To address these challenges, we propose an efficient adversarial video enhancement framework that learns directly from unpaired video examples. In particular, our framework introduces new recurrent cells that consist of interleaved local and global modules for implicit integration of spatial and temporal information. The proposed design allows our recurrent cells to efficiently propagate spatio-temporal information across frames and reduces the need for high complexity networks. Our setting enables learning from unpaired videos in a cyclic adversarial manner, where the proposed recurrent units are employed in all architectures. Efficient training is accomplished by introducing one single discriminator that learns the joint distribution of source and target domain simultaneously. The enhancement results demonstrate clear superiority of the proposed video enhancer over the state-of-the-art methods, in all terms of visual quality, quantitative metrics, and inference speed. Notably, our video enhancer is capable of enhancing over 35 frames per second of FullHD video (1080x1920)., International Journal of Computer Vision, 131 (4), ISSN:0920-5691, ISSN:1573-1405
- Published
- 2023
- Full Text
- View/download PDF
4. Characteristics of sequential activity in networks with temporally asymmetric Hebbian learning
- Author
-
Maxwell Gillett, Nicolas Brunel, and Ulises Pereira
- Subjects
0301 basic medicine ,Computer science ,Nerve net ,Models, Neurological ,Hippocampus ,Posterior parietal cortex ,Inhibitory postsynaptic potential ,Mice ,03 medical and health sciences ,Synaptic weight ,0302 clinical medicine ,Parietal Lobe ,Learning rule ,medicine ,Animals ,Learning ,Premovement neuronal activity ,Computer Simulation ,030304 developmental biology ,Neurons ,0303 health sciences ,Neuronal Plasticity ,Multidisciplinary ,Quantitative Biology::Neurons and Cognition ,business.industry ,Pattern recognition ,Biological Sciences ,Network dynamics ,Hebbian plasticity ,030104 developmental biology ,medicine.anatomical_structure ,Hebbian theory ,recurrent networks ,Synaptic plasticity ,Excitatory postsynaptic potential ,sequences ,Unsupervised learning ,Neural Networks, Computer ,Artificial intelligence ,Nerve Net ,business ,030217 neurology & neurosurgery ,Neuroscience - Abstract
Significance Sequential activity is a prominent feature of many neural systems, in multiple behavioral contexts. Here, we investigate how Hebbian rules lead to storage and recall of random sequences of inputs in both rate and spiking recurrent networks. In the case of the simplest (bilinear) rule, we characterize extensively the regions in parameter space that allow sequence retrieval and compute analytically the storage capacity of the network. We show that nonlinearities in the learning rule can lead to sparse sequences and find that sequences maintain robust decoding but display highly labile dynamics to continuous changes in the connectivity matrix, similar to recent observations in hippocampus and parietal cortex., Sequential activity has been observed in multiple neuronal circuits across species, neural structures, and behaviors. It has been hypothesized that sequences could arise from learning processes. However, it is still unclear whether biologically plausible synaptic plasticity rules can organize neuronal activity to form sequences whose statistics match experimental observations. Here, we investigate temporally asymmetric Hebbian rules in sparsely connected recurrent rate networks and develop a theory of the transient sequential activity observed after learning. These rules transform a sequence of random input patterns into synaptic weight updates. After learning, recalled sequential activity is reflected in the transient correlation of network activity with each of the stored input patterns. Using mean-field theory, we derive a low-dimensional description of the network dynamics and compute the storage capacity of these networks. Multiple temporal characteristics of the recalled sequential activity are consistent with experimental observations. We find that the degree of sparseness of the recalled sequences can be controlled by nonlinearities in the learning rule. Furthermore, sequences maintain robust decoding, but display highly labile dynamics, when synaptic connectivity is continuously modified due to noise or storage of other patterns, similar to recent observations in hippocampus and parietal cortex. Finally, we demonstrate that our results also hold in recurrent networks of spiking neurons with separate excitatory and inhibitory populations.
- Published
- 2020
- Full Text
- View/download PDF
5. Input Addition and Deletion in Reinforcement: Towards Protean Learning
- Author
-
Abdelkader Gouaich, Fabien Michel, Iago Bonnici, Système Multi-agent, Interaction, Langage, Evolution (SMILE), Laboratoire d'Informatique de Robotique et de Microélectronique de Montpellier (LIRMM), and Centre National de la Recherche Scientifique (CNRS)-Université de Montpellier (UM)-Centre National de la Recherche Scientifique (CNRS)-Université de Montpellier (UM)
- Subjects
Theoretical computer science ,Computer science ,Process (engineering) ,Data stream mining ,Formalism (philosophy) ,Recurrent networks ,02 engineering and technology ,Construct (python library) ,Arity ,Signature (logic) ,Reinforcement ,Transfer learning ,[INFO.INFO-AI]Computer Science [cs]/Artificial Intelligence [cs.AI] ,03 medical and health sciences ,0302 clinical medicine ,Artificial Intelligence ,Online learning ,0202 electrical engineering, electronic engineering, information engineering ,Reinforcement learning ,020201 artificial intelligence & image processing ,Adaptation (computer science) ,030217 neurology & neurosurgery - Abstract
International audience; Reinforcement Learning (RL) agents are commonly thought of as adaptive decision procedures. They work on input/output data streams called "states", "actions" and "rewards". Most current research about RL adaptiveness to changes works under the assumption that the streams signatures (i.e. arity and types of inputs and outputs) remain the same throughout the agent lifetime. As a consequence, natural situations where the signatures vary (e.g. when new data streams become available, or when others become obsolete) are not studied. In this paper, we relax this assumption and consider that signature changes define a new learning situation called Protean Learning (PL). When they occur, traditional RL agents become undefined, so they need to restart learning. Can better methods be developed under the PL view? To investigate this, we first construct a stream-oriented formalism to properly define PL and signature changes. Then, we run experiments in an idealized PL situation where input addition and deletion occur during the learning process. Results show that a simple PL-oriented method enables graceful adaptation of these arity changes, and is more efficient than restarting the process.
- Published
- 2022
- Full Text
- View/download PDF
6. Recurrent dynamics in the cerebral cortex: Integration of sensory evidence with stored knowledge
- Author
-
Wolf Singer
- Subjects
Sensory processing ,Computer science ,medicine.medical_treatment ,media_common.quotation_subject ,Sensory system ,temporal codes ,neuronal dynamics ,rate codes ,Memory ,Perception ,medicine ,Feature (machine learning) ,Humans ,predictive coding ,media_common ,Cerebral Cortex ,Structure (mathematical logic) ,Multidisciplinary ,Hierarchy (mathematics) ,business.industry ,Deep learning ,Pattern recognition ,Biological Sciences ,Complex dynamics ,recurrent networks ,Artificial intelligence ,Nerve Net ,business ,Neuroscience ,Signal Transduction - Abstract
Significance This review attempts to unite three hitherto rather unconnected concepts of basic functions of the cerebral cortex, taking the visual system as an example: 1) feed-forward processing in multilayer hierarchies (labeled line coding), 2) dynamic association of features (assembly coding), and 3) matching of sensory evidence with stored priors (predictive coding). The latter two functions are supposed to rely on the high-dimensional dynamics of delay-coupled recurrent networks. Discharge rates of neurons (rate code) and temporal relations among discharges (temporal code) are identified as conveying complementary information. Thus, the new concept accounts for the coexistence of feed-forward and recurrent processing, accommodates both rate and temporal codes, and assigns crucial functions to the complex dynamics emerging from recurrent interactions., Current concepts of sensory processing in the cerebral cortex emphasize serial extraction and recombination of features in hierarchically structured feed-forward networks in order to capture the relations among the components of perceptual objects. These concepts are implemented in convolutional deep learning networks and have been validated by the astounding similarities between the functional properties of artificial systems and their natural counterparts. However, cortical architectures also display an abundance of recurrent coupling within and between the layers of the processing hierarchy. This massive recurrence gives rise to highly complex dynamics whose putative function is poorly understood. Here a concept is proposed that assigns specific functions to the dynamics of cortical networks and combines, in a unifying approach, the respective advantages of recurrent and feed-forward processing. It is proposed that the priors about regularities of the world are stored in the weight distributions of feed-forward and recurrent connections and that the high-dimensional, dynamic space provided by recurrent interactions is exploited for computations. These comprise the ultrafast matching of sensory evidence with the priors covertly represented in the correlation structure of spontaneous activity and the context-dependent grouping of feature constellations characterizing natural objects. The concept posits that information is encoded not only in the discharge frequency of neurons but also in the precise timing relations among the discharges. Results of experiments designed to test the predictions derived from this concept support the hypothesis that cerebral cortex exploits the high-dimensional recurrent dynamics for computations serving predictive coding.
- Published
- 2021
7. The Research on Distributed Fusion Estimation Based on Machine Learning
- Author
-
Zhengxiao Peng, Yun Li, and Gang Hao
- Subjects
Fusion ,Radar tracker ,Training set ,General Computer Science ,Artificial neural network ,Computer science ,business.industry ,Scalar (mathematics) ,General Engineering ,Kalman filter ,Machine learning ,computer.software_genre ,Distributed fusion ,BP network ,Matrix (mathematics) ,machine learning ,Minimum-variance unbiased estimator ,recurrent networks ,Diagonal matrix ,General Materials Science ,lcsh:Electrical engineering. Electronics. Nuclear engineering ,Artificial intelligence ,business ,lcsh:TK1-9971 ,computer - Abstract
Multi-sensor distributed fusion estimation algorithms based on machine learning are proposed in this paper. Firstly, using local estimations as inputs and estimations of three classic distributed fusion (weighted by matrices, by diagonal matrices and by scalars) as the training sets, three distributed fusion algorithms based on BP network (BP net-based fusion weighted by matrices, by diagonal matrices and by scalar) are proposed and the selection basis of the number of nodes in hidden layer is given. Furthermore, by using local estimations as inputs and centralized fusion estimation as training set, another recurrent net-based distributed fusion algorithm is proposed, in the case that neither true states nor cross-covariance matrices is available. This method is not limited to the linear minimum variance (LMV) criterion, so its accuracy is higher than the classical three distributed fusion algorithms. A radar tracking simulation verifies the effectiveness of the proposed fusion networks.
- Published
- 2020
- Full Text
- View/download PDF
8. GLOBAL OPTIMIZATION FOR THE FORWARD NEURAL NETWORKS AND THEIR APPLICATIONS.
- Author
-
REDDY, K. SUNIL MANOHAR, BABU, G. RAVINDRA, and RAO, S. KRISHNA MOHAN
- Subjects
NEURAL computers ,ARTIFICIAL intelligence ,ARTIFICIAL neural networks ,COMPUTER science ,ALGORITHMS - Abstract
This paper describes and evaluates several global optimization issues of Artificial Neural Networks (ANN) and their applications. In this paper, the authors examine the properties of the feed-forward neural networks and the process of determining the appropriate network inputs and architecture, and built up a short-term gas load forecast system - the Tell Future system. This system performs very well for short-term gas load forecasting, which is built based on various Back-Propagation (BP) algorithms. The standard Back-Propagation (BP) algorithm for training feed-forward neural networks have proven robust even for difficult problems. In order to forecast the future load from the trained networks, the history loads, temperature, wind velocity, and calendar information should be used in addition to the predicted future temperature and wind velocity. Compared to other regression methods, the neural networks allow more flexible relationships between temperature, wind, calendar information and load pattern. Feed-forward neural networks can be used in many kinds of forecasting in different industrial areas. Similar models can be built to make electric load forecasting, daily water consumption forecasting, stock and markets forecasting, traffic flow and product sales forecasting. [ABSTRACT FROM AUTHOR]
- Published
- 2015
9. Attention Based Vehicle Trajectory Prediction
- Author
-
Fawzi Nashashibi, Kaouther Messaoud, Anne Verroust-Blondet, Itheri Yahiaoui, Robotics & Intelligent Transportation Systems (RITS), Inria de Paris, Institut National de Recherche en Informatique et en Automatique (Inria)-Institut National de Recherche en Informatique et en Automatique (Inria), Centre de Recherche en Sciences et Technologies de l'Information et de la Communication - EA 3804 (CRESTIC), Université de Reims Champagne-Ardenne (URCA), and Projet PIA CAMPUS
- Subjects
Trajectory prediction ,050210 logistics & transportation ,Control and Optimization ,Exploit ,Multi-head attention ,Computer science ,05 social sciences ,Real-time computing ,[INFO.INFO-CV]Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV] ,Recurrent networks ,010501 environmental sciences ,01 natural sciences ,Motion (physics) ,Vehicle dynamics ,Artificial Intelligence ,0502 economics and business ,Automotive Engineering ,Trajectory ,Task analysis ,Pairwise comparison ,Multi-modality ,Hidden Markov model ,Futures contract ,Vehicles interactions ,0105 earth and related environmental sciences - Abstract
International audience; Self-driving vehicles need to continuously analyse the driving scene, understand the behavior of other road users and predict their future trajectories in order to plan a safe motion and reduce their reaction time. Motivated by this idea, this paper addresses the problem of vehicle trajectory prediction over an extended horizon. On highways, human drivers continuously adapt their speed and paths according to the behavior of their neighboring vehicles. Therefore, vehicles' trajectories are very correlated and considering vehicle interactions makes motion prediction possible even before the start of a clear maneuver pattern. To this end, we introduce and analyze trajectory prediction methods based on how they model the vehicles interactions. Inspired by human reasoning, we use an attention mechanism that explicitly highlights the importance of neighboring vehicles with respect to their future states. We go beyond pairwise vehicle interactions and model higher order interactions. Moreover, the existence of different goals and driving behaviors induces multiple potential futures. We exploit a combination of global and partial attention paid to surrounding vehicles to generate different possible trajectory. Experiments on highway datasets show that the proposed model outperforms the state-of-the-art performances.
- Published
- 2021
- Full Text
- View/download PDF
10. Forecasting and trading the EUR/USD exchange rate with Gene Expression and Psi Sigma Neural Networks
- Author
-
Sermpinis, Georgios, Laws, Jason, Karathanasopoulos, Andreas, and Dunis, Christian L.
- Subjects
- *
PREDICTION models , *GENE expression , *ARTIFICIAL neural networks , *RECURSIVE sequences (Mathematics) , *GENETIC programming , *PERFORMANCE evaluation , *ARTIFICIAL intelligence , *PERCEPTRONS , *ALGORITHMS - Abstract
Abstract: The motivation for this paper is to investigate the use of two promising classes of artificial intelligence models, the Psi Sigma Neural Network (PSI) and the Gene Expression algorithm (GEP), when applied to the task of forecasting and trading the EUR/USD exchange rate. This is done by benchmarking their results with a Multi-Layer Perceptron (MLP), a Recurrent Neural Network (RNN), a genetic programming algorithm (GP), an autoregressive moving average model (ARMA) plus a naïve strategy. We also examine if the introduction of a time-varying leverage strategy can improve the trading performance of our models. [Copyright &y& Elsevier]
- Published
- 2012
- Full Text
- View/download PDF
11. RECURRENT NEURO FUZZY AND FUZZY NEURAL HYBRID NETWORKS: A REVIEW.
- Author
-
Subathra, B. and Radhakrishnan, T.K.
- Subjects
- *
NEURAL circuitry , *ARTIFICIAL intelligence , *NEUROBIOLOGY , *FUZZY logic , *FUZZY systems , *COGNITIVE neuroscience - Abstract
An attempt is made to provide a comprehensive survey of the current trends in hybrid recurrent structures of fuzzy logic (FL) and neural networks (NN) for solving temporal problems. Their applications extending to universal approximations are also discussed with available reported literature on recurrent neuro fuzzy networks (RNFN) and recurrent fuzzy neural networks (RFNN). [ABSTRACT FROM PUBLISHER]
- Published
- 2012
- Full Text
- View/download PDF
12. Image and Video Captioning with Augmented Neural Architectures
- Author
-
Jorma Laaksonen, Rakshith Shetty, and Hamed R. Tavakoli
- Subjects
Closed captioning ,image captioning ,Computer science ,Feature extraction ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Context (language use) ,02 engineering and technology ,computer vision ,computing ,0202 electrical engineering, electronic engineering, information engineering ,Media Technology ,Computer vision ,ta113 ,Artificial neural network ,business.industry ,Deep learning ,Cognitive neuroscience of visual object recognition ,deep learning ,020207 software engineering ,artificial intelligence ,Computer Science Applications ,Visualization ,Hardware and Architecture ,recurrent networks ,Signal Processing ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,Software ,Generator (mathematics) ,applications and expert knowledge-intensive systems - Abstract
Neural-network-based image and video captioning can be substantially improved by utilizing architectures that make use of special features from the scene context, objects, and locations. A novel discriminatively trained evaluator network for choosing the best caption among those generated by an ensemble of caption generator networks further improves accuracy.
- Published
- 2018
13. Approximation of state-space trajectories by locally recurrent globally feed-forward neural networks
- Author
-
Patan, Krzysztof
- Subjects
- *
APPROXIMATION theory , *ARTIFICIAL neural networks , *ARTIFICIAL intelligence , *SPACE flight - Abstract
Abstract: The paper deals with investigating approximation abilities of a special class of discrete-time dynamic neural networks. The networks considered are called locally recurrent globally feed-forward, because they are designed with dynamic neuron models which contain inner feedbacks, but interconnections between neurons are strict feed-forward ones like in the well-known multi-layer perceptron. The paper presents analytical results showing that a locally recurrent network with two hidden layers is able to approximate a state-space trajectory produced by any Lipschitz continuous function with arbitrary accuracy. Moreover, based on these results, the network can be simplified and transformed into a more practical structure needed in real world applications. [Copyright &y& Elsevier]
- Published
- 2008
- Full Text
- View/download PDF
14. RECURRENT NEURAL NETWORKS FOR MUSICAL PITCH MEMORY AND CLASSIFICATION.
- Author
-
FRANKLIN, JUDY A. and LOCKE, KRYSTAL K.
- Subjects
- *
ARTIFICIAL neural networks , *ARTIFICIAL intelligence , *MUSICAL pitch , *JAZZ , *MUSIC - Abstract
We present results from experiments in using several pitch representations for jazz-oriented musical tasks performed by a recurrent neural network. We have run experiments with several kinds of recurrent networks for this purpose, and have found that Long Short-term Memory networks provide the best results. We show that a new pitch representation called Circles of Thirds works as well as two other published representations for these tasks, yet it is more succinct and enables faster learning. We then discuss limited results using other types of networks on the same tasks. [ABSTRACT FROM AUTHOR]
- Published
- 2005
- Full Text
- View/download PDF
15. Independent Random Recurrent Neural Networks for Infrared Spatial Point Targets Classification
- Author
-
Moufa Hu, Huanzhang Lu, Bendong Zhao, and Dongya Wu
- Subjects
0209 industrial biotechnology ,Computer science ,02 engineering and technology ,lcsh:Technology ,01 natural sciences ,lcsh:Chemistry ,010309 optics ,020901 industrial engineering & automation ,Infrared signature ,0103 physical sciences ,General Materials Science ,Point (geometry) ,United States Space Surveillance Network ,Point target ,lcsh:QH301-705.5 ,Instrumentation ,Fluid Flow and Transfer Processes ,lcsh:T ,business.industry ,Process Chemistry and Technology ,General Engineering ,Process (computing) ,Pattern recognition ,lcsh:QC1-999 ,Computer Science Applications ,Recurrent neural network ,lcsh:Biology (General) ,lcsh:QD1-999 ,lcsh:TA1-2040 ,Face (geometry) ,recurrent networks ,infrared signature ,Artificial intelligence ,Noise (video) ,time series ,lcsh:Engineering (General). Civil engineering (General) ,business ,lcsh:Physics ,spatial point target - Abstract
Exo-atmospheric infrared (IR) point target discrimination is an important research topic of space surveillance systems. It is difficult to describe the characteristic information of the shape and micro-motion states of the targets and to discriminate different targets effectively by the characteristic information. This paper has constructed the infrared signature model of spatial point targets and obtained the infrared radiation intensity sequences dataset of different types of targets. This paper aims to design an algorithm for the classification problem of infrared radiation intensity sequences of spatial point targets. Recurrent neural networks (RNNs) are widely used in time series classification tasks, but face several problems such as gradient vanishing and explosion, etc. In view of shortcomings of RNNs, this paper proposes an independent random recurrent neural network (IRRNN) model, which combines independent structure RNNs with randomly weighted RNNs. Without increasing the training complexity of network learning, our model solves the problem of gradient vanishing and explosion, improves the ability to process long sequences, and enhances the comprehensive classification performance of the algorithm effectively. Experiments show that the IRRNN algorithm performs well in classification tasks and is robust to noise.
- Published
- 2019
- Full Text
- View/download PDF
16. Performing Deep Recurrent Double Q-Learning for Atari Games
- Author
-
Felipe Moreno-Vera, Universidad Nacional de Ingeniería (UNI), Universidad Católica San Pablo (UCSP), LatinX in AI Workshop, and Moreno-Vera, Felipe
- Subjects
[INFO.INFO-AI] Computer Science [cs]/Artificial Intelligence [cs.AI] ,FOS: Computer and information sciences ,Computer Science - Machine Learning ,Double Q-Learning ,Deep Reinforcement Learning ,Computer Science - Artificial Intelligence ,Computer science ,Atari ,DRQN ,Q-learning ,Machine Learning (stat.ML) ,[INFO] Computer Science [cs] ,Machine Learning (cs.LG) ,[INFO.INFO-AI]Computer Science [cs]/Artificial Intelligence [cs.AI] ,[INFO.INFO-CV] Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV] ,Recurrent Networks ,Statistics - Machine Learning ,[SHS.STAT] Humanities and Social Sciences/Methods and statistics ,DDQN ,Reinforcement learning ,[INFO]Computer Science [cs] ,Neural and Evolutionary Computing (cs.NE) ,DQN ,Atari Games ,[SHS.STAT]Humanities and Social Sciences/Methods and statistics ,business.industry ,Recurrent Q-Learning ,Convolutional Networks ,Computer Science - Neural and Evolutionary Computing ,[INFO.INFO-CV]Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV] ,Reinforcement Learning ,Artificial Intelligence (cs.AI) ,Action (philosophy) ,Video Games ,Artificial intelligence ,purl.org/pe-repo/ocde/ford#2.02.04 [http] ,business - Abstract
Currently, many applications in Machine Learning are based on define new models to extract more information about data, In this case Deep Reinforcement Learning with the most common application in video games like Atari, Mario, and others causes an impact in how to computers can learning by himself with only information called rewards obtained from any action. There is a lot of algorithms modeled and implemented based on Deep Recurrent Q-Learning proposed by DeepMind used in AlphaZero and Go. In this document, We proposed Deep Recurrent Double Q-Learning that is an implementation of Deep Reinforcement Learning using Double Q-Learning algorithms and Recurrent Networks like LSTM and DRQN., Comment: Accepted paper on LatinXinAI Workshop co-located with the International Conference on Machine Learning (ICML) 2019
- Published
- 2019
17. Comparison between Recurrent Networks and Temporal Convolutional Networks Approaches for Skeleton-Based Action Recognition
- Author
-
Adina Magda Florea, Cezar Cătălin Iacob, Mihai Trascau, and Mihai Nan
- Subjects
Computer science ,Inference ,02 engineering and technology ,Skeleton (category theory) ,lcsh:Chemical technology ,Machine learning ,computer.software_genre ,Biochemistry ,Article ,Analytical Chemistry ,sequence-to-sequence ,0202 electrical engineering, electronic engineering, information engineering ,temporal convolutional networks ,lcsh:TP1-1185 ,Electrical and Electronic Engineering ,Instrumentation ,Sequence ,action recognition ,business.industry ,Search engine indexing ,020207 software engineering ,Atomic and Molecular Physics, and Optics ,Recurrent neural network ,Action (philosophy) ,recurrent networks ,Benchmark (computing) ,Action recognition ,020201 artificial intelligence & image processing ,Artificial intelligence ,business ,computer - Abstract
Action recognition plays an important role in various applications such as video monitoring, automatic video indexing, crowd analysis, human-machine interaction, smart homes and personal assistive robotics. In this paper, we propose improvements to some methods for human action recognition from videos that work with data represented in the form of skeleton poses. These methods are based on the most widely used techniques for this problem—Graph Convolutional Networks (GCNs), Temporal Convolutional Networks (TCNs) and Recurrent Neural Networks (RNNs). Initially, the paper explores and compares different ways to extract the most relevant spatial and temporal characteristics for a sequence of frames describing an action. Based on this comparative analysis, we show how a TCN type unit can be extended to work even on the characteristics extracted from the spatial domain. To validate our approach, we test it against a benchmark often used for human action recognition problems and we show that our solution obtains comparable results to the state-of-the-art, but with a significant increase in the inference speed.
- Published
- 2021
- Full Text
- View/download PDF
18. Accelerating Deep Action Recognition Networks for Real-Time Applications
- Author
-
Pablo Martinez-Gonzalez, David Ivorra-Piqueres, John Alejandro Castro Vargas, Universidad de Alicante. Departamento de Tecnología Informática y Computación, and Universidad de Alicante. Instituto Universitario de Investigación Informática
- Subjects
GPU Acceleration ,business.industry ,Computer science ,Deep learning ,Video decoding ,Real-time computing ,Optical flow ,ComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISION ,Ciencia de la Computación e Inteligencia Artificial ,02 engineering and technology ,Real-Time ,Video Decoding ,Machine Learning ,Deep Learning ,Optical Flow ,Recurrent Networks ,Action Recognition ,0202 electrical engineering, electronic engineering, information engineering ,Action recognition ,020201 artificial intelligence & image processing ,Action Understanding ,Artificial intelligence ,business - Abstract
In this work, the authors propose several techniques for accelerating a modern action recognition pipeline. This article reviewed several recent and popular action recognition works and selected two of them as part of the tools used for improving the aforementioned acceleration. Specifically, temporal segment networks (TSN), a convolutional neural network (CNN) framework that makes use of a small number of video frames for obtaining robust predictions which have allowed to win the first place in the 2016 ActivityNet challenge, and MotionNet, a convolutional-transposed CNN that is capable of inferring optical flow RGB frames. Together with the last proposal, this article integrated a new software for decoding videos that takes advantage of NVIDIA GPUs. This article shows a proof of concept for this approach by training the RGB stream of the TSN network in videos loaded with NVIDIA Video Loader (NVVL) of a subset of daily actions from the University of Central Florida 101 dataset.
- Published
- 2019
19. Low dimensional dynamics for working memory and time encoding
- Author
-
Stefano Fusi, C. Daniel Salzman, Christopher J. Cueva, Mehrdad Jazayeri, Ranulfo Romo, Encarni Marcos, Aldo Genovesio, Michael N. Shadlen, Alex Saez, National Science Foundation (US), Gatsby Charitable Foundation, Simons Foundation, Universidad Nacional Autónoma de México, National Institutes of Health (US), Consejo Nacional de Ciencia y Tecnología (México), Fondazione Regionale per la Ricerca Biomedica, and National Institute of Mental Health (US)
- Subjects
Primates ,Computer science ,Chaotic ,Sensory system ,03 medical and health sciences ,0302 clinical medicine ,Encoding (memory) ,Attractor ,Animals ,Backpropagation through time ,030304 developmental biology ,Neurons ,Brain Mapping ,0303 health sciences ,Multidisciplinary ,Artificial neural network ,Quantitative Biology::Neurons and Cognition ,Working memory ,business.industry ,Reservoir computing ,Brain ,Pattern recognition ,Biological Sciences ,Backpropagation ,3. Good health ,Memory, Short-Term ,Recurrent neural network ,neural dynamics ,recurrent networks ,reservoir computing ,time decoding ,working memory ,animals ,brain ,brain mapping ,memory, short-term ,nerve net ,neural networks, computer ,neurons ,primates ,Trajectory ,Neural Networks, Computer ,Artificial intelligence ,Nerve Net ,business ,030217 neurology & neurosurgery ,Decoding methods ,Curse of dimensionality - Abstract
Our decisions often depend on multiple sensory experiences separated by time delays. The brain can remember these experiences and, simultaneously, estimate the timing between events. To understand the mechanisms underlying working memory and time encoding, we analyze neural activity recorded during delays in four experiments on nonhuman primates. To disambiguate potential mechanisms, we propose two analyses, namely, decoding the passage of time from neural data and computing the cumulative dimensionality of the neural trajectory over time. Time can be decoded with high precision in tasks where timing information is relevant and with lower precision when irrelevant for performing the task. Neural trajectories are always observed to be low-dimensional. In addition, our results further constrain the mechanisms underlying time encoding as we find that the linear “ramping” component of each neuron’s firing rate strongly contributes to the slow timescale variations that make decoding time possible. These constraints rule out working memory models that rely on constant, sustained activity and neural networks with high-dimensional trajectories, like reservoir networks. Instead, recurrent networks trained with backpropagation capture the time-encoding properties and the dimensionality observed in the data., Research was supported by NSF Next Generation Network for Neuroscience Award DBI-1707398, the Gatsby Charitable Foundation, the Simons Foundation, the Swartz Foundation (C.J.C. and S.F.), NIH Training Grant 5T32NS064929 (to C.J.C.), and the Kavli Foundation (S.F.). M.N.S. was supported by National Institute of Neurological Disorders and Stroke Brain Initiative Grant R01NS113113. R.R. was supported by the Dirección General de Asuntos del Personal Académico de la Universidad Nacional Autónoma de México (PAPIIT-IN210819) and Consejo Nacional de Ciencia y Tecnología (CONACYT-240892). A.G. was supported by National Institute of Mental Health Division of Intramural Research Grant Z01MH-01092 and by Italian Fondo per gli investimenti della ricerca di base 2010 Grant RBFR10G5W9_001.
- Published
- 2018
- Full Text
- View/download PDF
20. Recurrent Multiresolution Convolutional Networks for VHR Image Classification
- Author
-
Claudio Persello, John Ray Bergado, Alfred Stein, Department of Earth Observation Science, UT-I-ITC-ACQUAL, and Faculty of Geo-Information Science and Earth Observation
- Subjects
FOS: Computer and information sciences ,010504 meteorology & atmospheric sciences ,Computer science ,Computer Vision and Pattern Recognition (cs.CV) ,Feature extraction ,Multispectral image ,0211 other engineering and technologies ,Computer Science - Computer Vision and Pattern Recognition ,02 engineering and technology ,01 natural sciences ,Upsampling ,land cover classification ,Labeling ,Image fusion ,Electrical and Electronic Engineering ,Image resolution ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences ,Spatial resolution ,Contextual image classification ,business.industry ,Deep learning ,deep learning ,Pattern recognition ,Panchromatic film ,Kernel ,Kernel (image processing) ,recurrent networks ,ITC-ISI-JOURNAL-ARTICLE ,Task analysis ,General Earth and Planetary Sciences ,Embedding ,very high-resolution (VHR) image ,Artificial intelligence ,business ,Interpolation ,Convolutional networks - Abstract
Classification of very high-resolution (VHR) satellite images has three major challenges: 1) inherent low intraclass and high interclass spectral similarities; 2) mismatching resolution of available bands; and 3) the need to regularize noisy classification maps. Conventional methods have addressed these challenges by adopting separate stages of image fusion, feature extraction, and postclassification map regularization. These processing stages, however, are not jointly optimizing the classification task at hand. In this paper, we propose a single-stage framework embedding the processing stages in a recurrent multiresolution convolutional network trained in an end-to-end manner. The feedforward version of the network, called FuseNet , aims to match the resolution of the panchromatic and multispectral bands in a VHR image using convolutional layers with corresponding downsampling and upsampling operations. Contextual label information is incorporated into FuseNet by means of a recurrent version called ReuseNet . We compared FuseNet and ReuseNet against the use of separate processing steps for both image fusions, e.g., pansharpening and resampling through interpolation and map regularization such as conditional random fields . We carried out our experiments on a land-cover classification task using a Worldview-03 image of Quezon City, Philippines, and the International Society for Photogrammetry and Remote Sensing 2-D semantic labeling benchmark data set of Vaihingen, Germany. FuseNet and ReuseNet surpass the baseline approaches in both the quantitative and qualitative results.
- Published
- 2018
21. Multi-Temporal Land Cover Classification with Sequential Recurrent Encoders
- Author
-
Marco Körner and Marc Rußwurm
- Subjects
FOS: Computer and information sciences ,Earth observation ,010504 meteorology & atmospheric sciences ,Machine translation ,Computer science ,Computer Vision and Pattern Recognition (cs.CV) ,Geography, Planning and Development ,Computer Science - Computer Vision and Pattern Recognition ,0211 other engineering and technologies ,sequence encoder ,lcsh:G1-922 ,deep learning ,multi-temporal classification ,land use and land cover classification ,recurrent networks ,crop classification ,sequence-to-sequence ,Sentinel 2 ,02 engineering and technology ,Land cover ,computer.software_genre ,01 natural sciences ,Earth and Planetary Sciences (miscellaneous) ,Preprocessor ,Computers in Earth Sciences ,021101 geological & geomatics engineering ,0105 earth and related environmental sciences ,Sequence ,business.industry ,Deep learning ,Pattern recognition ,ddc ,Temporal database ,Artificial intelligence ,business ,computer ,Encoder ,lcsh:Geography (General) - Abstract
Earth observation (EO) sensors deliver data at daily or weekly intervals. Mostland use and land cover classification(LULC) approaches, however, are designed for cloud-free and mono-temporal observations. The increasing temporal capabilities of today’s sensors enable the use of temporal, along with spectral and spatial features. Domains such as speech recognition or neural machine translation, work with inherently temporal data and, today, achieve impressive results by using sequential encoder-decoder structures. Inspired by these sequence-to-sequence models, we adapt an encoder structure with convolutional recurrent layers in order to approximate a phenological model for vegetation classes based on a temporal sequence ofSentinel 2(S2) images. In our experiments, we visualize internal activations over a sequence of cloudy and non-cloudy images and find several recurrent cells that reduce the input activity for cloudy observations. Hence, we assume that our network has learned cloud-filtering schemes solely from input data, which could alleviate the need for tedious cloud-filtering as a preprocessing step for many EO approaches. Moreover, using unfiltered temporal series oftop-of-atmosphere(TOA) reflectance data, our experiments achieved state-of-the-art classification accuracies on a large number of crop classes with minimal preprocessing, compared to other classification approaches.
- Published
- 2018
- Full Text
- View/download PDF
22. Foreign Currency Exchange Rates Prediction Using CGP and Recurrent Neural Network
- Author
-
Sahibzada Ali Mahmud, Mehreen Rehman, and Gul Muhammad Khan
- Subjects
Neural Networks ,Artificial neural network ,business.industry ,Computer science ,Feature selection ,Cartesian Genetic Programming ,Neuro-evolution ,Machine learning ,computer.software_genre ,Foreign exchange rate forecasting ,Recurrent neural network ,Recurrent Networks ,Currency ,Time Series Prediction ,Liberian dollar ,Feature (machine learning) ,Artificial intelligence ,business ,computer ,Foreign exchange market ,Predictive modelling - Abstract
Feedback in Neuro-Evolution is explored and evaluated for its application in devising prediction models for foreign currency exchange rates. A novel approach to foreign currency exchange rates forecasting based on Recurrent Neuro-Evolution is introduced. Cartesian Genetic Programming (CGP) is the algorithm deployed for the forecasting model. Recurrent Cartesian Genetic Programming evolved Artificial Neural Network (RCGPANN) is demonstrated to produce computationally efficient and accurate model for forex prediction with an accuracy of as high as 98.872% for a period of 1000 days. The approach utilizes the trends that are being followed in historical data to predict five currency rates against Australian dollar. The model is evaluated using statistical metrics and compared. The computational method outperforms the other methods particularly due to its capability to select the best possible feature in real time and the flexibility that the system provides in feature selection, connectivity pattern and network.
- Published
- 2014
- Full Text
- View/download PDF
23. Does the Cerebral Cortex Exploit High-Dimensional, Non-linear Dynamics for Information Processing?
- Author
-
Wolf Singer and Andreea Lazar
- Subjects
0301 basic medicine ,Elementary cognitive task ,Computer science ,media_common.quotation_subject ,Neuroscience (miscellaneous) ,Stimulus (physiology) ,lcsh:RC321-571 ,03 medical and health sciences ,Cellular and Molecular Neuroscience ,non-linear dynamics ,0302 clinical medicine ,Hypothesis and Theory ,Perception ,medicine ,visual cortex ,lcsh:Neurosciences. Biological psychiatry. Neuropsychiatry ,media_common ,Signal processing ,Quantitative Biology::Neurons and Cognition ,business.industry ,Information processing ,Pattern recognition ,Cognition ,plasticity and learning ,030104 developmental biology ,medicine.anatomical_structure ,Visual cortex ,Cerebral cortex ,recurrent networks ,Artificial intelligence ,synchrony and oscillations ,business ,Neuroscience ,030217 neurology & neurosurgery - Abstract
The discovery of stimulus induced synchronization in the visual cortex suggested the possibility that the relations among low-level stimulus features are encoded by the temporal relationship between neuronal discharges. In this framework, temporal coherence is considered a signature of perceptual grouping. This insight triggered a large number of experimental studies which sought to investigate the relationship between temporal coordination and cognitive functions. While some core predictions derived from the initial hypothesis were confirmed, these studies, also revealed a rich dynamical landscape beyond simple coherence whose role in signal processing is still poorly understood. In this paper, a framework is presented which establishes links between the various manifestations of cortical dynamics by assigning specific coding functions to low-dimensional dynamic features such as synchronized oscillations and phase shifts on the one hand and high-dimensional non-linear, non-stationary dynamics on the other. The data serving as basis for this synthetic approach have been obtained with chronic multisite recordings from the visual cortex of anesthetized cats and from monkeys trained to solve cognitive tasks. It is proposed that the low-dimensional dynamics characterized by synchronized oscillations and large-scale correlations are substates that represent the results of computations performed in the high-dimensional state-space provided by recurrently coupled networks.
- Published
- 2016
- Full Text
- View/download PDF
24. Neural Network Evidence for the Coupling of Presaccadic Visual Remapping to Predictive Eye Position Updating
- Author
-
Hrishikesh M Rao, Juan eSan Juan, Fred Y Shen, Jennifer E Villa, Kimia S Rafie, and Marc A Sommer
- Subjects
0301 basic medicine ,genetic structures ,media_common.quotation_subject ,Neuroscience (miscellaneous) ,Topographica ,Video camera ,lcsh:RC321-571 ,law.invention ,03 medical and health sciences ,Cellular and Molecular Neuroscience ,0302 clinical medicine ,Corollary ,law ,Perception ,visual stability ,Computer vision ,lcsh:Neurosciences. Biological psychiatry. Neuropsychiatry ,Original Research ,media_common ,Artificial neural network ,business.industry ,Process (computing) ,Eye movement ,saccades ,Saccadic masking ,eye movements ,030104 developmental biology ,recurrent networks ,Saccade ,Artificial intelligence ,Psychology ,business ,030217 neurology & neurosurgery ,Neuroscience - Abstract
As we look around a scene, we perceive it as continuous and stable even though each saccadic eye movement changes the visual input to the retinas. How the brain achieves this perceptual stabilization is unknown, but a major hypothesis is that it relies on presaccadic remapping, a process in which neurons shift their visual sensitivity to a new location in the scene just before each saccade. This hypothesis is difficult to test in vivo because complete, selective inactivation of remapping is currently intractable. We tested it in silico with a hierarchical, sheet-based neural network model of the visual and oculomotor system. The model generated saccadic commands to move a video camera abruptly. Visual input from the camera and internal copies of the saccadic movement commands, or corollary discharge, converged at a map-level simulation of the frontal eye field (FEF), a primate brain area known to receive such inputs. FEF output was combined with eye position signals to yield a suitable coordinate frame for guiding arm movements of a robot. Our operational definition of perceptual stability was “useful stability,” quantified as continuously accurate pointing to a visual object despite camera saccades. During training, the emergence of useful stability was correlated tightly with the emergence of presaccadic remapping in the FEF. Remapping depended on corollary discharge but its timing was synchronized to the updating of eye position. When coupled to predictive eye position signals, remapping served to stabilize the target representation for continuously accurate pointing. Graded inactivations of pathways in the model replicated, and helped to interpret, previous in vivo experiments. The results support the hypothesis that visual stability requires presaccadic remapping, provide explanations for the function and timing of remapping, and offer testable hypotheses for in vivo studies. We conclude that remapping allows for seamless coordinate frame transformations and quick actions despite visual afferent lags. With visual remapping in place for behavior, it may be exploited for perceptual continuity.
- Published
- 2016
- Full Text
- View/download PDF
25. Coherence and recurrency: maintenance, control and integration in working memory
- Author
-
Antonino Raffone and Gezinus Wolters
- Subjects
Maintenance ,Cognitive Neuroscience ,Integration ,control of attention ,integration ,maintenance ,prefrontal cortex ,recurrent networks ,synchronization ,working memory ,Prefrontal Cortex ,Experimental and Cognitive Psychology ,Review ,Synchronization ,Generalization, Psychological ,Artificial Intelligence ,Interneurons ,Chunking (psychology) ,Humans ,Attention ,Cortical Synchronization ,Prefrontal cortex ,Set (psychology) ,Cognitive science ,Neurons ,Artificial neural network ,Working memory ,Control of attention ,Attentional control ,Association Learning ,Brain ,Retention, Psychology ,Cognition ,Recurrent networks ,Neural Inhibition ,General Medicine ,Memory, Short-Term ,Pattern Recognition, Visual ,Neural Networks, Computer ,Nerve Net ,Psychology ,Information integration - Abstract
Working memory (WM), including a ‘central executive’, is used to guide behavior by internal goals or intentions. We suggest that WM is best described as a set of three interdependent functions which are implemented in the prefrontal cortex (PFC). These functions are maintenance, control of attention and integration. A model for the maintenance function is presented, and we will argue that this model can be extended to incorporate the other functions as well. Maintenance is the capacity to briefly maintain information in the absence of corresponding input, and even in the face of distracting information. We will argue that maintenance is based on recurrent loops between PFC and posterior parts of the brain, and probably within PFC as well. In these loops information can be held temporarily in an active form. We show that a model based on these structural ideas is capable of maintaining a limited number of neural patterns. Not the size, but the coherence of patterns (i.e., a chunking principle based on synchronous firing of interconnected cell assemblies) determines the maintenance capacity. A mechanism that optimizes coherent pattern segregation, also poses a limit to the number of assemblies (about four) that can concurrently reverberate. Top-down attentional control (in perception, action and memory retrieval) can be modelled by the modulation and re-entry of top-down information to posterior parts of the brain. Hierarchically organized modules in PFC create the possibility for information integration. We argue that large-scale multimodal integration of information creates an ‘episodic buffer’, and may even suffice for implementing a central executive.
- Published
- 2007
26. Modelling adaptation aftereffects in associative memory
- Author
-
Alessandro Treves, Nicola van Rijsbergen, and Federica Menghini
- Subjects
associative memory ,Artificial neural network ,visual aftereffects ,Computer science ,business.industry ,Cognitive Neuroscience ,Speech recognition ,recurrent networks ,multiple attractor states ,firing rate adaptation ,Content-addressable memory ,Stimulus (physiology) ,Computer Science Applications ,Artificial Intelligence ,Bidirectional associative memory ,Artificial intelligence ,business ,Priming (psychology) - Abstract
We present a simple autoassociative neural network model of cortical adaptation that replicates a high-level category aftereffect. We study in our model the effect of a priming stimulus on retrieval from a briefly presented partial cue. We also explore, in our general framework, which learning conditions permit correct pattern retrieval.
- Published
- 2007
- Full Text
- View/download PDF
27. Distributed Recurrent Neural Forward Models with Synaptic Adaptation and CPG-based control for Complex Behaviors of Walking Robots
- Author
-
Sakyasingha eDasgupta, Dennis eGoldschmidt, Florentin eWörgötter, and Poramate eManoonpong
- Subjects
Computer science ,Forward internal model ,Biomedical Engineering ,neural control ,forward models ,recurrent networks ,locomotion ,adaptive behavior ,walking robots ,synaptic adaptation ,Forward models ,Neural control ,Machine learning ,computer.software_genre ,lcsh:RC321-571 ,Artificial Intelligence ,Learning ,Adaptation (computer science) ,lcsh:Neurosciences. Biological psychiatry. Neuropsychiatry ,Adaptive Behavior ,Original Research ,Adaptive behavior ,Flexibility (engineering) ,Biorobotics ,business.industry ,Central pattern generator ,Embodied learning ,Legged robots ,Recurrent neural network ,Climbing ,Robot ,Artificial intelligence ,business ,computer ,Locomotion ,Neuroscience - Abstract
Walking animals, like stick insects, cockroaches or ants, demonstrate a fascinating range of locomotive abilities and complex behaviors. The locomotive behaviors can consist of a variety of walking patterns along with adaptation that allow the animals to deal with changes in environmental conditions, like uneven terrains, gaps, obstacles etc. Biological study has revealed that such complex behaviors are a result of a combination of biomechanics and neural mechanism thus representing the true nature of embodied interactions. While the biomechanics helps maintain flexibility and sustain a variety of movements, the neural mechanisms generate movements while making appropriate predictions crucial for achieving adaptation. Such predictions or planning ahead can be achieved by way of internal models that are grounded in the overall behavior of the animal. Inspired by these findings, we present here, an artificial bio-inspired walking system which effectively combines biomechanics (in terms of the body and leg structures) with the underlying neural mechanisms. The neural mechanisms consist of (1) central pattern generator based control for generating basic rhythmic patterns and coordinated movements, (2) distributed (at each leg) recurrent neural network based adaptive forward models with efference copies as internal models for sensory predictions and instantaneous state estimations, and (3) searching and elevation control for adapting the movement of an individual leg to deal with different environmental conditions. Using simulations we show that this bio-inspired approach with adaptive internal models allows the walking robot to perform complex locomotive behaviors as observed in insects, including walking on undulated terrains, crossing large gaps, leg damage adaptations, as well as climbing over high obstacles. Furthermore, we demonstrate that the newly developed recurrent network based approach to online forward models outperforms the adaptive neuron forward models, which have hitherto been the state of the art, to model a subset of similar walking behaviors in walking robots. peerReviewed
- Published
- 2015
- Full Text
- View/download PDF
28. Exploring multiple feature combination strategies with a recurrent neural network architecture for off-line handwriting recognition
- Author
-
Luc Mioulet, Clément Chatelain, Stephan Brunessaux, Gautier Bideault, Thierry Paquet, Laboratoire d'Informatique, de Traitement de l'Information et des Systèmes (LITIS), Institut national des sciences appliquées Rouen Normandie (INSA Rouen Normandie), Institut National des Sciences Appliquées (INSA)-Normandie Université (NU)-Institut National des Sciences Appliquées (INSA)-Normandie Université (NU)-Université de Rouen Normandie (UNIROUEN), Normandie Université (NU)-Université Le Havre Normandie (ULH), Normandie Université (NU), Equipe Apprentissage (DocApp - LITIS), Normandie Université (NU)-Institut national des sciences appliquées Rouen Normandie (INSA Rouen Normandie), Airbus Defence and Space, and Chatelain, Clément
- Subjects
Handwriting recognition ,Feature combination ,Artificial neural network ,Computer science ,Time delay neural network ,Intelligent character recognition ,business.industry ,Feature vector ,Speech recognition ,[INFO.INFO-TT] Computer Science [cs]/Document and Text Processing ,020207 software engineering ,Pattern recognition ,Recurrent networks ,02 engineering and technology ,Data modeling ,[INFO.INFO-TT]Computer Science [cs]/Document and Text Processing ,Recurrent neural network ,0202 electrical engineering, electronic engineering, information engineering ,020201 artificial intelligence & image processing ,Artificial intelligence ,System Combination ,business ,Word (computer architecture) ,Neural networks - Abstract
International audience; The BLSTM-CTC is a novel recurrent neural network architecture that has outperformed previous state of the art algorithms in tasks such as speech recognition or handwriting recognition. It has the ability to process long term dependencies in temporal signals in order to label unsegmented data. This paper describes different ways of combining features using a BLSTM-CTC architecture. Not only do we explore the low level combination (feature space combination) but we also explore high level combination (decoding combination) and mid-level (internal system representation combination). The results are compared on the RIMES word database. Our results show that the low level combination works best, thanks to the powerful data modeling of the LSTM neurons.
- Published
- 2015
29. A novel analytical characterization for short-term plasticity parameters in spiking neural networks
- Author
-
Corey M. Thibeault, Michael J. O'Brien, and Narayan Srinivasa
- Subjects
STP ,Computer science ,Neuroscience (miscellaneous) ,Parameter space ,lcsh:RC321-571 ,Physics::Geophysics ,Cellular and Molecular Neuroscience ,Random search ,self-sustaining ,Original Research Article ,lcsh:Neurosciences. Biological psychiatry. Neuropsychiatry ,Spiking neural network ,Quantitative Biology::Neurons and Cognition ,Artificial neural network ,business.industry ,Replicate ,neural networks ,Term (time) ,Power (physics) ,Asynchronous communication ,recurrent networks ,short term plasticity ,Artificial intelligence ,business ,Biological system ,Neuroscience - Abstract
Short-term plasticity (STP) is a phenomenon that widely occurs in the neocortex with implications for learning and memory. Based on a widely used STP model, we develop an analytical characterization of the STP parameter space to determine the nature of each synapse (facilitating, depressing or both) in a spiking neural network based on presynaptic firing rate and the corresponding STP parameters. We demonstrate consistency with previous work by leveraging the power of our characterization to replicate the functional volumes that are integral for the previous network stabilization results. We then use our characterization to predict the precise transitional point from the facilitating regime to the depressing regime in a simulated synapse, suggesting 'in vitro' experiments to verify the underlying STP model. We conclude the work by integrating our characterization into a framework for finding suitable STP parameters for self-sustaining random, asynchronous activity in a prescribed recurrent spiking neural network. The systematic process resulting from our analytical characterization improves the success rate of finding the requisite parameters for such networks by three orders of magnitude over a random search.
- Published
- 2014
- Full Text
- View/download PDF
30. Connectionism coming of age: legacy and future challenges
- Author
-
Pablo Gomez, Franklin Chang, Gary Lupyan, and Julien Mayor
- Subjects
computational modeling ,Computer science ,lcsh:BF1-990 ,interactive processing ,Computational linguistics ,050105 experimental psychology ,computational linguistics ,03 medical and health sciences ,0302 clinical medicine ,Connectionism ,ddc:150 ,Psychology ,0501 psychology and cognitive sciences ,Language processing ,connectionism ,word learning ,Word learning ,General Psychology ,probabilistic cognition ,Cognitive science ,Interactive processing ,Artificial neural network ,business.industry ,Deep learning ,05 social sciences ,Speech perception ,language processing ,Editorial Article ,Information processing ,Recurrent networks ,Computational modeling ,Language acquisition ,Catastrophic interference ,Perceptron ,language acquisition ,lcsh:Psychology ,Probabilistic cognition ,recurrent networks ,Speech Perception ,Artificial intelligence ,business ,030217 neurology & neurosurgery ,Nervous system network models - Abstract
In 1986, Rumelhart and McClelland took the cognitive science community by storm with the Parallel Distributed Processing (PDP) framework. Rather than abstracting from the biological substrate as was sought by the “information processing” paradigms of the 1970s, connectionism, as it has come to be called, embraced it. An immediate appeal of the connectionist agenda was its aim: to construct at the algorithmic level models of cognition that were compatible with their implementation in the biological substrate. The PDP group argued that this could be achieved by turning to networks of artificial neurons, originally introduced by McCulloch and Pitts (1943) which the group showed were able to provide insights into a wide range of psychological domains, from categorization, to perception, to memory, to language. This work built on an earlier formulation by Rosenblatt (1958) who introduced a simple type of feed-forward neural network called the perceptron. Perceptrons were limited to solving simple linearly-separable problems and although networks composed of perceptrons were known to be able to compute any Boolean function (including XOR, Minsky and Papert, 1969), there was no effective way of training such networks. In 1986, Rumelhart, Hinton and Williams introduced the back-propagation algorithm, providing an effective way of training multi-layered neural networks, which could easily learn non linearly-separable functions. In addition to providing the field with an effective learning algorithm, the PDP group published a series of demonstrations of how long standing questions in cognitive psychology could be elegantly solved using simple learning rules, distributed representations, and interactive processing. To take a classic example, consider the word-superiority effect, in which people can detect letters within a word faster than individual letters or letters within a non-word (Reicher, 1969). This result is difficult to square with serial “information-processing” theories of cognition that were dominant at the time (how could someone recognize “R” before “FRIEND” if recognizing the word required recognizing the letters?). Accounting for such findings demanded a framework which could naturally accommodate interactive processes within a bidirectional flow of information. The so-called “Interactive-activation model” (McClelland and Rumelhart, 1981) provided just such a framework. The connectionist paradigm was not without its critics. The principal critiques can be divided into three classes. First, some neuroscientists (Crick, 1989) questioned the biological plausibility of backpropagation, when they failed to observe experimentally complex and differentiated back-propagating signals that are required to learn in multi-layered neural networks. A second critique concerned stability-plasticity of the learned representations in these models. Some phenomena require the ability to rapidly learn new information, but sometimes newly learned knowledge overwrites previously learned information (catastrophic interference; McCloskey and Cohen, 1989). Third, representing spatial and temporal invariance—something that apparently came easily to people—was difficult for models, e.g., recognizing that the letter “T” in “TOM” was the “same” as the “T” in “POT.” This invariance problem was typically solved by multiplying a large number of hard-wired units that were space- or time-locked (see e.g., McClelland and Elman, 1986). Finally, critics pointed out that the networks were incapable of learning true rules on which a number of human behavioral, namely language-learning was thought to depend (e.g., Marcus, 2003; cf. Fodor and Pylyshyn, 1988; Seidenberg, 1999). The connectionist approach has embraced these challenges: Although some connectionist models continue to rely on backpropagation, others have moved to more biologically realistic learning rules (Giese and Poggio, 2003; Masquelier and Thorpe, 2007). Far from being a critical flaw of connectionism, the phenomenon of catastrophic interference (Mermillod et al., 2013) proved to be a feature that led to the development of complementary learning systems (McClelland et al., 1995). Progress has also been made on the invariance problem. For example, within the speech domain representing the similarity between similar speech sounds regardless of their location within a word has been addressed in the past by Grossberg and Myers (2000) and Norris (1994) and this issue presents a new more streamlined and computationally efficient model (Hannagan et al., 2013). An especially powerful approach to solving the location invariance problem in the visual domain is presented by Di Bono and Zorzi (2013), also in this issue. A key challenge for connectionism is to explain the learning of abstract structural representations. The use of recurrent networks (Elman, 1990; Dominey, 2013) and self-organizing maps, has captured important aspects of language learning (e.g., Mayor and Plunkett, 2010; Li and Zhao, 2013), while work on deep learning (Hinton and Salakhutdinov, 2006) has made it possible to model the emergence of structured and abstract representations within multi-layered hierarchical networks (Zorzi et al., 2013). The work on verbal analogies by Kollias and McClelland (2013) continues to address the challenges of modeling more abstract representations, but truly understanding how neural architectures give rise to symbolic cognition is a gap that remains. Although learning and representing formal language rules may not be completely outside of the abilities of neural networks (e.g., Chang, 2009), it seems clear that understanding human cognition requires understanding how we solve these symbolic problems (Clark and Karmiloff-Smith, 1993; Lupyan, 2013). Future generations of connectionist modelers may wish to fill this gap and in so doing provide a fuller picture of how neural networks give rise to intelligence of the sort enables us to ponder the very workings of our cognition.
- Published
- 2014
- Full Text
- View/download PDF
31. Modeling Behavior in Different Delay Match to Sample Tasksin One Simple Network
- Author
-
Volodya Yakovlev, Yali Amit, and Shaul Hochstein
- Subjects
Computer science ,Sample (statistics) ,Machine learning ,computer.software_genre ,working memory ,lcsh:RC321-571 ,Behavioral Neuroscience ,Memory ,Original Research Article ,lcsh:Neurosciences. Biological psychiatry. Neuropsychiatry ,Biological Psychiatry ,Network model ,Random graph ,forgetting ,Forgetting ,Quantitative Biology::Neurons and Cognition ,Hebbian Learning ,Working memory ,business.industry ,Pattern recognition ,Familiarity ,Psychiatry and Mental health ,Neuropsychology and Physiological Psychology ,Hebbian theory ,Neurology ,readout mechanism ,reset mechanism ,recurrent networks ,Noise (video) ,Artificial intelligence ,recognition ,business ,Reset (computing) ,computer ,Neuroscience - Abstract
Delay match to sample (DMS) experiments provide an important link between the theory of recurrent network models and behavior and neural recordings. We define a simple recurrent network of binary neurons with stochastic neural dynamics and Hebbian synaptic learning. Most DMS experiments involve heavily learned images, and in this setting we propose a readout mechanism for match occurrence based on a smaller increment in overall network activity when the matched pattern is already in working memory, and a reset mechanism to clear memory from stimuli of previous trials using random network activity. Simulations show that this model accounts for a wide range of variations on the original DMS tasks, including ABBA tasks with distractors, and more general repetition detection tasks with both learned and novel images. The differences in network settings required for different tasks derive from easily defined changes in the levels of noise and inhibition. The same models can also explain experiments involving repetition detection with novel images, although in this case the readout mechanism for match is based on higher overall network activity. The models give rise to interesting predictions that may be tested in neural recordings.
- Published
- 2013
- Full Text
- View/download PDF
32. Multiscale analysis of slow-fast neuronal learning models with noise
- Author
-
Mathieu Galtier, Gilles Wainrib, Mathematical and Computational Neuroscience (NEUROMATHCOMP), Inria Sophia Antipolis - Méditerranée (CRISAM), Institut National de Recherche en Informatique et en Automatique (Inria)-Institut National de Recherche en Informatique et en Automatique (Inria)-Laboratoire Jean Alexandre Dieudonné (JAD), Université Côte d'Azur (UCA)-Université Nice Sophia Antipolis (... - 2019) (UNS), COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-Centre National de la Recherche Scientifique (CNRS)-Université Côte d'Azur (UCA)-Université Nice Sophia Antipolis (... - 2019) (UNS), COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-Centre National de la Recherche Scientifique (CNRS), School of Engineering and Science [Bremen] (JU-SES), Jacobs University [Bremen], Laboratoire Analyse, Géométrie et Applications (LAGA), Université Paris 8 Vincennes-Saint-Denis (UP8)-Centre National de la Recherche Scientifique (CNRS)-Institut Galilée-Université Paris 13 (UP13), Institut National de Recherche en Informatique et en Automatique (Inria)-Institut National de Recherche en Informatique et en Automatique (Inria)-Laboratoire Jean Alexandre Dieudonné (LJAD), Université Nice Sophia Antipolis (1965 - 2019) (UNS), COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-Centre National de la Recherche Scientifique (CNRS)-Université Côte d'Azur (UCA)-Université Nice Sophia Antipolis (1965 - 2019) (UNS), COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-COMUE Université Côte d'Azur (2015-2019) (COMUE UCA)-Centre National de la Recherche Scientifique (CNRS)-Université Côte d'Azur (UCA), and Université Paris 8 Vincennes-Saint-Denis (UP8)-Université Paris 13 (UP13)-Institut Galilée-Centre National de la Recherche Scientifique (CNRS)
- Subjects
Trace (linear algebra) ,Computer science ,Neuroscience (miscellaneous) ,inhomogeneous Markov process ,unsupervised learning ,STDP ,03 medical and health sciences ,Matrix (mathematics) ,slow-fast systems ,0302 clinical medicine ,model reduction ,Statistical physics ,030304 developmental biology ,averaging ,0303 health sciences ,Cross-correlation ,business.industry ,Research ,stochastic differential equations ,Mathematical theory ,Noise ,Hebbian theory ,recurrent networks ,Time derivative ,[SDV.NEU]Life Sciences [q-bio]/Neurons and Cognition [q-bio.NC] ,Artificial intelligence ,Focus (optics) ,business ,Hebbian learning ,030217 neurology & neurosurgery - Abstract
International audience; This paper deals with the application of temporal averaging methods to recurrent networks of noisy neurons undergoing a slow and unsupervised modification of their connectivity matrix called learning. Three time-scales arise for these models: (i) the fast neuronal dynamics, (ii) the intermediate external input to the system, and (iii) the slow learning mechanisms. Based on this time-scale separation, we apply an extension of the mathematical theory of stochastic averaging with periodic forcing in order to derive a reduced deterministic model for the connectivity dynamics. We focus on a class of models where the activity is linear to understand the specificity of several learning rules (Hebbian, trace or anti-symmetric learning). In a weakly connected regime, we study the equilibrium connectivity which gathers the entire 'knowledge' of the network about the inputs. We develop an asymptotic method to approximate this equilibrium. We show that the symmetric part of the connectivity post-learning encodes the correlation structure of the inputs, whereas the anti-symmetric part corresponds to the cross correlation between the inputs and their time derivative. Moreover, the time-scales ratio appears as an important parameter revealing temporal correlations.
- Published
- 2012
- Full Text
- View/download PDF
33. Learning of temporal motor patterns: An analysis of continuous vs. reset timing
- Author
-
Rodrigo eLaje, Karen eCheng, and Dean V Buonomano
- Subjects
computational modeling ,neural dynamics ,Computer science ,Cognitive Neuroscience ,Speech recognition ,Population ,Musical instrument ,Context (language use) ,lcsh:RC346-429 ,lcsh:RC321-571 ,Cellular and Molecular Neuroscience ,timing ,education ,Video game ,lcsh:Neurosciences. Biological psychiatry. Neuropsychiatry ,lcsh:Neurology. Diseases of the nervous system ,Original Research ,education.field_of_study ,time estimation and production ,business.industry ,Recurrent networks ,Variance (accounting) ,human psychophysics ,temporal processing ,Sensory Systems ,Term (time) ,Recurrent neural network ,Artificial intelligence ,Timer ,business ,Neuroscience - Abstract
Our ability to generate well-timed sequences of movements is critical to an array of behaviors, including the ability to play a musical instrument or a video game. Here we address two questions relating to timing with the goal of better understanding the neural mechanisms underlying temporal processing. First, how does accuracy and variance change over the course of learning of complex spatiotemporal patterns? Second, is the timing of sequential responses most consistent with starting and stopping an internal timer at each interval or with continuous timing? To address these questions we used a psychophysical task in which subjects learned to reproduce a sequence of finger taps in the correct order and at the correct times – much like playing a melody at the piano. This task allowed us to calculate the variance of the responses at different time points using data from the same trials. Our results show that while “standard” Weber’s law is clearly violated, variance does increase as a function of time squared, as expected according to the generalized form of Weber’s law – which separates the source of variance into time-dependent and time-independent components. Over the course of learning, both the time-independent variance and the coefficient of the time-dependent term decrease. Our analyses also suggest that timing of sequential events does not rely on the resetting of an internal timer at each event. We describe and interpret our results in the context of computer simulations that capture some of our psychophysical findings. Specifically, we show that continuous timing, as opposed to “reset” timing, is consistent with “population clock” models in which timing emerges from the internal dynamics of recurrent neural networks.
- Published
- 2011
- Full Text
- View/download PDF
34. Improved neural network for SVM learning
- Author
-
A. Boni and Davide Anguita
- Subjects
Wake-sleep algorithm ,Computer Networks and Communications ,Computer science ,Active learning (machine learning) ,Competitive learning ,Quadratic programming ,Machine learning ,computer.software_genre ,Relevance vector machine ,Artificial Intelligence ,Least squares support vector machine ,support vector machine ,recurrent networks ,Structured support vector machine ,Artificial neural network ,Learning automata ,Time delay neural network ,business.industry ,Deep learning ,Online machine learning ,Recurrent neural nets ,General Medicine ,Computer Science Applications ,Support vector machine ,Recurrent neural network ,Computational learning theory ,Feedforward neural network ,Artificial intelligence ,Types of artificial neural networks ,business ,computer ,Software - Abstract
The recurrent network of Xia et al. (1996) was proposed for solving quadratic programming problems and was recently adapted to support vector machine (SVM) learning by Tan et al. (2000). We show that this formulation contains some unnecessary circuits which, furthermore, can fail to provide the correct value of one of the SVM parameters and suggest how to avoid these drawbacks.
- Published
- 2008
35. General-purpose computation with neural networks: a survey of complexity theoretic results
- Author
-
Jiří Šíma, Pekka Orponen, Aalto-yliopisto, and Aalto University
- Subjects
Time Factors ,Computational complexity theory ,perceptrons ,Cognitive Neuroscience ,education ,Models, Neurological ,Arts and Humanities (miscellaneous) ,probabilistic computation ,Mathematical Computing ,Mathematics ,computational power ,Spiking neural network ,Neurons ,Network architecture ,computational complexity ,Models, Statistical ,Artificial neural network ,spiking neurons ,business.industry ,Probabilistic logic ,radial basis functions ,Perceptron ,Classification ,feedforward networks ,Discrete time and continuous time ,analog computation ,recurrent networks ,State (computer science) ,Artificial intelligence ,Neural Networks, Computer ,business ,Algorithms - Abstract
We survey and summarize the literature on the computational aspects of neural network models by presenting a detailed taxonomy of the various models according to their complexity theoretic characteristics. The criteria of classification include the architecture of the network (feedforward versus recurrent), time model (discrete versus continuous), state type (binary versus analog), weight constraints (symmetric versus asymmetric), network size (finite nets versus infinite families), and computation type (deterministic versus probabilistic), among others. The underlying results concerning the computational power and complexity issues of perceptron, radial basis function, winner-take-all, and spiking neural networks are briefly surveyed, with pointers to the relevant literature. In our survey, we focus mainly on the digital computation whose inputs and outputs are binary in nature, although their values are quite often encoded as analog neuron states. We omit the important learning issues.
- Published
- 2003
36. Neural network learning for analog VLSI implementations of support vector machines: a survey
- Author
-
A. Boni and Davide Anguita
- Subjects
Hyperparameter ,Computer science ,business.industry ,Cognitive Neuroscience ,Model selection ,SVM learning ,Neural network learning ,Recurrent networks ,Quadratic programming ,Machine learning ,computer.software_genre ,Computer Science Applications ,Nonlinear programming ,Support vector machine ,Recurrent neural network ,Artificial Intelligence ,Analog VLSI ,Vlsi implementations ,Artificial intelligence ,SVM learning, Recurrent networks, Analog VLSI, Quadratic programming ,business ,computer - Abstract
In the last few years several kinds of recurrent neural networks (RNNs) have been proposed for solving linear and nonlinear optimization problems. In this paper, we provide a survey of RNNs that can be used to solve both the constrained quadratic optimization problem related to support vector machine (SVM) learning, and the SVM model selection by automatic hyperparameter tuning. The appeal of this approach is the possibility of implementing such networks on analog VLSI systems with relative easiness. We review several proposals appeared so far in the literature and test their behavior when applied to solve a telecommunication application, where a special purpose adaptive hardware is of great interest.
- Published
- 2003
37. The evolution of cognition - a hypothesis
- Author
-
Holk Cruse
- Subjects
Cognitive model ,Cognitive science ,Artificial neural network ,Cognitive Neuroscience ,media_common.quotation_subject ,Control (management) ,Perspective (graphical) ,subjective experience ,Experimental and Cognitive Psychology ,Cognition ,consciousness ,Evolution reactive system ,deliberation ,internal world model ,Connectionism ,cognitive system ,Artificial Intelligence ,recurrent networks ,internal perspective ,Consciousness ,Psychology ,Reactive system ,media_common - Abstract
Behavior may be controlled by reactive systems. In a reactive system the motor output is exclusively driven by actual sensory input. An alternative solution to control behavior is given by "cognitive" systems capable of planning ahead. To this end the system has to be equipped with some kind of internal world model. A sensible basis of an internal world model might be a model of the system's own body. I show that a reactive system with the ability to control a body of complex geometry requires only a slight reorganization to form a cognitive system. This implies that the assumption that the evolution of cognitive properties requires the introduction of new, additional modules, namely internal world models, is not justified. Rather, these modules may already have existed before the system obtained cognitive properties. Furthermore, I discuss whether the occurrence of such world models may lead to systems having internal perspective. (C) 2002 Cognitive Science Society, Inc. All rights reserved.
- Published
- 2003
- Full Text
- View/download PDF
38. Adaptive Networks for Physical Modeling
- Author
-
Claude Cadoz, Nicolas Szilas, ACROE - Ingénierie de la Création Artistique (ACROE-ICA), Ministère de la Culture et de la Communication (MCC)-Institut National Polytechnique de Grenoble (INPG), Laboratoire Leibniz (Leibniz - IMAG), Université Joseph Fourier - Grenoble 1 (UJF)-Institut National Polytechnique de Grenoble (INPG)-Centre National de la Recherche Scientifique (CNRS), and French Ministère de la Culture
- Subjects
Active learning ,Wake-sleep algorithm ,Computer science ,Active learning (machine learning) ,Cognitive Neuroscience ,Competitive learning ,Stability (learning theory) ,02 engineering and technology ,03 medical and health sciences ,0302 clinical medicine ,Artificial Intelligence ,Leabra ,0202 electrical engineering, electronic engineering, information engineering ,Instance-based learning ,Physical modeling ,Learning classifier system ,Artificial neural network ,business.industry ,Deep learning ,Algorithmic learning theory ,Online machine learning ,Recurrent networks ,Generalization error ,[INFO.INFO-GR]Computer Science [cs]/Graphics [cs.GR] ,Computer Science Applications ,Recurrent neural network ,Computational learning theory ,[INFO.INFO-SD]Computer Science [cs]/Sound [cs.SD] ,Unsupervised learning ,020201 artificial intelligence & image processing ,Artificial intelligence ,Types of artificial neural networks ,business ,030217 neurology & neurosurgery - Abstract
International audience; This paper presents an original link between neural networks theory and mechanical modeling networks. The problem is to find the parameters characterizing mechanical structures in order to reproduce given mechanical behaviors. Replacing "neural" units with mechanically based units and applying classical learning algorithms dedicated to supervised dynamic networks to these mechanical networks allows us to find the parameters for a physical model. Some new variants of real-time recurrent learning (RTRL) are also introduced, based on mechanical principles. The notion of interaction during learning is discussed at length and the results of tests are presented. Instead of the classical {machine learning system, environment} pair, we propose to study the {machine learning system, human operator, environment} triplet. Experiments have been carried out in simulated scenarios and some original experiments with a force-feedback device are also described.
- Published
- 1998
- Full Text
- View/download PDF
39. On the Computational Power of Recurrent Neural Networks for Structures
- Author
-
Alessandro Sperduti
- Subjects
neural networks for structures ,recurrent networks ,tree automata ,Finite-state machine ,Artificial neural network ,business.industry ,Computer science ,Cognitive Neuroscience ,Computer Science::Neural and Evolutionary Computation ,Tree (graph theory) ,Turing machine ,symbols.namesake ,Recurrent neural network ,Artificial Intelligence ,Cellular neural network ,Theory of computation ,symbols ,Artificial intelligence ,Types of artificial neural networks ,business - Abstract
Recurrent neural networks can simulate any finite state automata as well as any multi-stack Turing machine. When constraining the network architecture, however, this computational power may no longer hold. For example, recurrent cascade-correlation cannot simulate any finite state automata. Thus, it is important to assess the computational power of a given network architecture, since this characterizes the class of functions which, in principle, can be computed by it. We discuss the computational power of neural networks for structures. Elman-style networks, cascade-correlation networks and neural trees for structures are introduced. We show that Elman-style networks can simulate any frontier-to-root tree automation, while neither cascade-correlation networks nor neural trees can. As a special case of the latter result, we obtain that neural trees for sequences cannot simulate any finite state machine. © 1997 Elsevier Science Ltd. All Rights Reserved.
- Published
- 1997
40. Occlusion-Based Explanations in Deep Recurrent Models for Biomedical Signals
- Author
-
Anna Monreale, Michele Resta, and Davide Bacciu
- Subjects
Correctness ,Computer science ,Science ,QC1-999 ,biomedical signals ,media_common.quotation_subject ,General Physics and Astronomy ,occlusion ,02 engineering and technology ,Astrophysics ,Machine learning ,computer.software_genre ,Article ,Field (computer science) ,020204 information systems ,0202 electrical engineering, electronic engineering, information engineering ,Production (economics) ,Quality (business) ,media_common ,Interpretability ,business.industry ,Physics ,020206 networking & telecommunications ,Regression ,Biomedical signals ,Occlusion ,Recurrent networks ,QB460-466 ,Recurrent neural network ,recurrent networks ,Clinical diagnosis ,Artificial intelligence ,interpretability ,business ,computer - Abstract
The biomedical field is characterized by an ever-increasing production of sequential data, which often come in the form of biosignals capturing the time-evolution of physiological processes, such as blood pressure and brain activity. This has motivated a large body of research dealing with the development of machine learning techniques for the predictive analysis of such biosignals. Unfortunately, in high-stakes decision making, such as clinical diagnosis, the opacity of machine learning models becomes a crucial aspect to be addressed in order to increase the trust and adoption of AI technology. In this paper, we propose a model agnostic explanation method, based on occlusion, that enables the learning of the input’s influence on the model predictions. We specifically target problems involving the predictive analysis of time-series data and the models that are typically used to deal with data of such nature, i.e., recurrent neural networks. Our approach is able to provide two different kinds of explanations: one suitable for technical experts, who need to verify the quality and correctness of machine learning models, and one suited to physicians, who need to understand the rationale underlying the prediction to make aware decisions. A wide experimentation on different physiological data demonstrates the effectiveness of our approach both in classification and regression tasks.
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.