19 results on '"invariance"'
Search Results
2. Unsupervised invariance learning of transformation sequences in a model of object recognition yields selectivity for non-accidental properties
- Author
-
Thomas eSerre and Sarah eParker
- Subjects
Learning ,object recognition ,invariance ,Inferotemporal cortex ,HMAX ,ventral stream ,Neurosciences. Biological psychiatry. Neuropsychiatry ,RC321-571 - Abstract
Non-accidental properties (NAPs) correspond to image properties that are invariant to changes in viewpoint (e.g., straight vs. curved contours) and are distinguished from metric properties (MPs) that can change continuously with in-depth object rotation (e.g., aspect ratio, degree of curvature, etc). Behavioral and electrophysiological studies of shape processing have demonstrated greater sensitivity to differences in NAPs than in MPs. However, previous work has shown that such sensitivity is lacking in multiple-views models of object recognition such as textsc{Hmax}. These models typically assume that object processing is based on populations of view-tuned neurons with distributed symmetrical bell-shaped tuning that are modulated at least as much by differences in MPs as in NAPs.Here, we test the hypothesis that unsupervised learning of invariances to object transformations may increase the sensitivity to differences in NAPs vs. MPs in textsc{Hmax}. We collected a database of video sequences with objects slowly rotating in-depth in an attempt to mimic sequences viewed during object manipulation by young children during early developmental stages. We show that unsupervised learning yields shape-tuning in higher stages with greater sensitivity to differences in NAPs vs. MPs in agreement with monkey IT data. Together, these results suggest that greater NAP sensitivity may arise from experiencing different in-depth rotations of objects.
- Published
- 2015
- Full Text
- View/download PDF
3. Finding and recognising objects in natural scenes: complementary computations in the dorsal and ventral visual systems
- Author
-
Edmund eRolls and Tristan James Webb
- Subjects
object recognition ,invariance ,inferior temporal visual cortex ,VisNet ,saliency ,trace learning rule ,Neurosciences. Biological psychiatry. Neuropsychiatry ,RC321-571 - Abstract
Searching for and recognising objects in complex natural scenes is implemented by multiple saccades until the eyes reach within the reduced receptive field sizes of inferior temporal cortex (IT) neurons. We analyse and model how the dorsal and ventral visual streams both contribute to this. Saliency detection in the dorsal visual system including area LIP is modelled by graph-based visual saliency, and allows the eyes to fixate potential objects within several degrees. Visual information at the fixated location subtending approximately 9 degrees corresponding to the receptive fields of IT neurons is then passed through a four layer hierarchical model of the ventral cortical visual system, VisNet. We show that VisNet can be trained using a synaptic modification rule with a short-term memory trace of recent neuronal activity to capture both the required view and translation invariances to allow in the model approximately 90% correct object recognition for 4 objects shown in any view across a range of 135 degrees anywhere in a scene.The model was able to generalize correctly within the four trained views and the 25 trained translations.This approach analyses the principles by which complementary computations in the dorsal and ventral visual cortical streams enable objects to be located and recognised in complex natural scenes.
- Published
- 2014
- Full Text
- View/download PDF
4. Unsupervised invariance learning of transformation sequences in a model of object recognition yields selectivity for non-accidental properties.
- Author
-
Parker, Sarah M., Serre, Thomas, and Tarr, Michael J.
- Subjects
OBJECT recognition (Computer vision) ,NEURONS ,CHILD development ,OBJECT constancy (Psychoanalysis) ,INFORMATION technology - Abstract
Non-accidental properties (NAPs) correspond to image properties that are invariant to changes in viewpoint (e.g., straight vs. curved contours) and are distinguished from metric properties (MPs) that can change continuously with in-depth object rotation (e.g., aspect ratio, degree of curvature, etc.). Behavioral and electrophysiological studies of shape processing have demonstrated greater sensitivity to differences in NAPs than in MPs. However, previous work has shown that such sensitivity is lacking in multiple-views models of object recognition such as HMAX. These models typically assume that object processing is based on populations of view-tuned neurons with distributed symmetrical bell-shaped tuning that are modulated at least as much by differences in MPs as in NAPs. Here, we test the hypothesis that unsupervised learning of invariances to object transformations may increase the sensitivity to differences in NAPs vs. MPs in HMAX. We collected a database of video sequences with objects slowly rotating in-depth in an attempt to mimic sequences viewed during object manipulation by young children during early developmental stages. We show that unsupervised learning yields shape-tuning in higher stages with greater sensitivity to differences in NAPs vs. MPs in agreement with monkey IT data. Together, these results suggest that greater NAP sensitivity may arise from experiencing different in-depth rotations of objects. [ABSTRACT FROM AUTHOR]
- Published
- 2015
- Full Text
- View/download PDF
5. Finding and recognizing objects in natural scenes: complementary computations in the dorsal and ventral visual systems.
- Author
-
Rolls, Edmund T. and Webb, Tristan J.
- Subjects
OPTICAL pattern recognition ,PATTERN recognition systems ,IMAGE recognition (Computer vision) ,OBJECT recognition (Computer vision) ,COMPUTER vision ,IMAGE processing - Abstract
Searching for and recognizing objects in complex natural scenes is implemented by multiple saccades until the eyes reach within the reduced receptive field sizes of inferior temporal cortex (IT) neurons. We analyze and model how the dorsal and ventral visual streams both contribute to this. Saliency detection in the dorsal visual system including area LIP is modeled by graph-based visual saliency, and allows the eyes to fixate potential objects within several degrees. Visual information at the fixated location subtending approximately 9° corresponding to the receptive fields of IT neurons is then passed through a four layer hierarchical model of the ventral cortical visual system, VisNet. We show that VisNet can be trained using a synaptic modification rule with a short-term memory trace of recent neuronal activity to capture both the required view and translation invariances to allow in the model approximately 90% correct object recognition for 4 objects shown in any view across a range of 135° anywhere in a scene. The model was able to generalize correctly within the four trained views and the 25 trained translations. This approach analyses the principles by which complementary computations in the dorsal and ventral visual cortical streams enable objects to be located and recognized in complex natural scenes. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
6. Deformation-specific and deformation-invariant visual object recognition: pose vs. identity recognition of people and deforming objects.
- Author
-
Webb, Tristan J. and Rolls, Edmund T.
- Subjects
ISOMONODROMIC deformation method ,MEMORY ,LEARNING ,RECOGNITION (Psychology) ,TEMPOROMANDIBULAR disorders - Abstract
When we see a human sitting down, standing up, or walking, we can recognize one of these poses independently of the individual, or we can recognize the individual person, independently of the pose. The same issues arise for deforming objects. For example, if we see a flag deformed by the wind, either blowing out or hanging languidly, we can usually recognize the flag, independently of its deformation; or we can recognize the deformation independently of the identity of the flag. We hypothesize that these types of recognition can be implemented by the primate visual system using temporo-spatial continuity as objects transform as a learning principle. In particular, we hypothesize that pose or deformation can be learned under conditions in which large numbers of different people are successively seen in the same pose, or objects in the same deformation. We also hypothesize that person-specific representations that are independent of pose, and object-specific representations that are independent of deformation and view, could be built, when individual people or objects are observed successively transforming from one pose or deformation and view to another. These hypotheses were tested in a simulation of the ventral visual system, VisNet, that uses temporal continuity, implemented in a synaptic learning rule with a short-term memory trace of previous neuronal activity, to learn invariant representations. It was found that depending on the statistics of the visual input, either pose-specific or deformation-specific representations could be built that were invariant with respect to individual and view; or that identity-specific representations could be built that were invariant with respect to pose or deformation and view. We propose that this is how pose-specific and pose-invariant, and deformation-specific and deformation-invariant, perceptual representations are built in the brain. [ABSTRACT FROM AUTHOR]
- Published
- 2014
- Full Text
- View/download PDF
7. Low-level contrast statistics are diagnostic of invariance of natural textures.
- Author
-
Iris I. A. Groen, Sennay eGhebreab, Victor A. F. Lamme, and H. Steven Scholte
- Subjects
EEG ,illumination ,textures ,image statistics ,contrast ,invariance ,Neurosciences. Biological psychiatry. Neuropsychiatry ,RC321-571 - Abstract
Texture may provide important clues for real world object and scene perception. For these clues to be reliable, they should ideally be invariant to common viewing variations such as changes in illumination and orientation. In a large image database of natural materials, we found textures with low-level contrast statistics that varied substantially under viewing variations, as well as textures that remained relatively constant. This led us to ask whether textures with constant contrast statistics give rise to more invariant representations compared to other textures. To test this, we selected natural texture images with either high (HV) or low (LV) variance in statistics and presented these to human observers. In two distinct behavioral categorization paradigms, participants more often judged HV textures as 'different' compared to LV textures, showing that textures with constant contrast statistics are perceived as being more invariant. In a separate EEG experiment, evoked responses to single texture images (single-image ERPs) were collected. The results show that differences in contrast statistics predicted both early and late differences in ERP amplitude between individual images. Importantly, ERP differences between images of HV textures were mainly driven by illumination angle, which was not the case for LV images: there, differences were completely driven by texture membership, as predicted by contrast statistics. These converging neural and behavioral results imply that some natural textures are surprisingly invariant and that low-level contrast statistics predict the extent of this invariance.
- Published
- 2012
- Full Text
- View/download PDF
8. Invariant visual object and face recognition: neural and computational bases, and a model, VisNet
- Author
-
Edmund T eRolls
- Subjects
Visual Cortex ,Visual Perception ,object recognition ,invariance ,inferior temporal visual cortex ,face neurons ,Neurosciences. Biological psychiatry. Neuropsychiatry ,RC321-571 - Abstract
Neurophysiological evidence for invariant representations of objects and faces in the primate inferior temporal visual cortex is described. Then a computational approach to how invariant representations are formed in the brain is described that builds on the neurophysiology. A feature hierarchy modelin which invariant representations can be built by self-organizing learning based on the temporal and spatialstatistics of the visual input produced by objects as they transform in the world is described. VisNet can use temporal continuity in an associativesynaptic learning rule with a short term memory trace, and/or it can use spatialcontinuity in Continuous Spatial Transformation learning which does not require a temporal trace. The model of visual processing in theventral cortical stream can build representations of objects that are invariant withrespect to translation, view, size, and also lighting. The modelhas been extended to provide an account of invariant representations in the dorsal visualsystem of the global motion produced by objects such as looming, rotation, and objectbased movement. The model has been extended to incorporate top-down feedback connectionsto model the control of attention by biased competition in for example spatial and objectsearch tasks. The model has also been extended to account for how the visual system canselect single objects in complex visual scenes, and how multiple objects can berepresented in a scene. The model has also been extended to provide, with an additional layer, for the development of representations of spatial scenes of the type found in the hippocampus.
- Published
- 2012
- Full Text
- View/download PDF
9. Learning and disrupting invariance in visual recognition with a temporal association rule
- Author
-
Leyla eIsik, Joel Z Leibo, and Tomaso ePoggio
- Subjects
Vision ,object recognition ,invariance ,Trace rule ,Cortical models ,Inferotemporal cortex ,Neurosciences. Biological psychiatry. Neuropsychiatry ,RC321-571 - Abstract
Learning by temporal association rules such as Foldiak’s trace rule is an attractive hypothesis that explains the development of invariance in visual recognition. Consistent with these rules, several recent experiments have shown that invariance can be broken at both the psychophysical and single cell levels. We show a) that temporal association learning provides appropriate invariance in models of object recognition inspired by the visual cortex, b) that we can replicate the invariance disruption experiments using these models with a temporal association learning rule to develop and maintain invariance, and c) that despite dramatic single cell effects, a population of cells is very robust to these disruptions. We argue that these models account for the stability of perceptual invariance despite the underlying plasticity of the system, the variability of the visual world and expected noise in the biological mechanisms.
- Published
- 2012
- Full Text
- View/download PDF
10. A biologically plausible transform for visual recognition that is invariant to translation, scale and rotation
- Author
-
Pavel eSountsov, David M Santucci, and John E Lisman
- Subjects
Visual System ,object recognition ,invariance ,Interval Detector ,Spatial Frequency Analysis ,Neurosciences. Biological psychiatry. Neuropsychiatry ,RC321-571 - Abstract
Visual object recognition occurs easily despite differences in position, size, and rotation of the object, but the neural mechanisms responsible for this invariance are not known. We have found a set of transforms that achieve invariance in a neurally plausible way. We find that a transform based on local spatial frequency analysis of oriented segments and on logarithmic mapping, when applied twice in an iterative fashion, produces an output image that is unique to the object and that remains constant as the input image is shifted, scaled or rotated.
- Published
- 2011
- Full Text
- View/download PDF
11. Learning and disrupting invariance in visual recognition a temporal association rule.
- Author
-
Isik, Leyla, Leibo, Joel Z., and Poggio, Tomaso
- Subjects
OBJECT recognition (Computer vision) ,PRINCIPLE of relativity (Physics) ,CEREBRAL cortex ,WEBER-Fechner law ,SINGLE cell proteins - Abstract
Learning by temporal association rules such as Foldiak's trace rule is an attractive hypothesis that explains the development of invariance in visual recognition. Consistent with these rules, several recent experiments have shown that invariance can be broken at both the psychophysical and single cell levels. We show (1) that temporal association learning provides appropriate invariance in models of object recognition inspired by the visual cortex, (2) that we can replicate the "invariance disruption" experiments using these models with a temporal association learning rule to develop and maintain invariance, and (3) that despite dramatic single cell effects, a population of cells is very robust to these disruptions. We argue that these models account for the stability of perceptual invariance despite the underlying plasticity of the system, the variability of the visual world and expected noise in the biological mechanisms. [ABSTRACT FROM AUTHOR]
- Published
- 2012
- Full Text
- View/download PDF
12. Invariant visual object and face recognition: neural and computational bases, and a model, VisNet.
- Author
-
Rolls, Edmund T.
- Subjects
PRINCIPLE of relativity (Physics) ,HUMAN facial recognition software ,OBJECT recognition (Computer vision) ,VISUAL cortex ,NEUROPHYSIOLOGY ,MAGNETIC resonance imaging - Abstract
Neurophysiological evidence for invariant representations of objects and faces in the primate inferior temporal visual cortex is described. Then a computational approach to how invariant representations are formed in the brain is described that builds on the neuro-physiology. A feature hierarchy model in which invariant representations can be built by self-organizing learning based on the temporal and spatial statistics of the visual input produced by objects as they transform in the world is described. VisNet can use temporal continuity in an associative synaptic learning rule with a short-term memory trace, and/or it can use spatial continuity in continuous spatial transformation learning which does not require a temporal trace. The model of visual processing in the ventral cortical stream can build representations of objects that are invariant with respect to translation, view, size, and also lighting. The model has been extended to provide an account of invariant representations in the dorsal visual system of the global motion produced by objects such as looming, rotation, and object-based movement. The model has been extended to incorporate top-down feedback connections to model the control of attention by biased competition in, for example, spatial and object search tasks.The approach has also been extended to account for how the visual system can select single objects in complex visual scenes, and how multiple objects can be represented in a scene. The approach has also been extended to provide, with an additional layer, for the development of representations of spatial scenes of the type found in the hippocampus. [ABSTRACT FROM AUTHOR]
- Published
- 2012
- Full Text
- View/download PDF
13. Low-level contrast statistics are diagnostic of invariance of natural textures.
- Author
-
Groen, Iris I. A., Ghebreab, Sennay, Lamme, Victor A. F., and Steven Scholte, H.
- Subjects
TEXTURE analysis (Image processing) ,ELECTROENCEPHALOGRAPHY ,CONTRAST media ,PRINCIPLE of relativity (Physics) ,LIGHTING - Abstract
Texture may provide important clues for real world object and scene perception. To be reliable, these clues should ideally be invariant to common viewing variations such as changes in illumination and orientation. In a large image database of natural materials, we found textures with low-level contrast statistics that varied substantially under viewing variations, as well as textures that remained relatively constant. This led us to ask whether textures with constant contrast statistics give rise to more invariant representations compared to other textures. To test this, we selected natural texture images with either high (HV) or low (LV) variance in contrast statistics and presented these to human observers. In two distinct behavioral categorization paradigms, participants more often judged HV textures as "different" compared to LV textures, showing that textures with constant contrast statistics are perceived as being more invariant. In a separate electroencephalogram (EEG) experiment, evoked responses to single texture images (single-image ERPs) were collected. The results show that differences in contrast statistics correlated with both early and late differences in occipital ERP amplitude between individual images. Importantly, ERP differences between images of HV textures were mainly driven by illumination angle, which was not the case for LV images: there, differences were completely driven by texture membership. These converging neural and behavioral results imply that some natural textures are surprisingly invariant to illumination changes and that low-level contrast statistics are diagnostic of the extent of this invariance. [ABSTRACT FROM AUTHOR]
- Published
- 2012
- Full Text
- View/download PDF
14. Editorial: Integrating Computational and Neural Findings in Visual Object Perception.
- Author
-
Peters, Judith C., Op de Beeck, Hans P., and Goebel, Rainer
- Subjects
NEURAL codes ,COMPUTER vision ,OBJECT recognition (Computer vision) - Abstract
An introduction is presented in which the editor discusses various reports within the issue on topics including visual object perception, object recognition and neural coding.
- Published
- 2016
- Full Text
- View/download PDF
15. Integrating Computational and Neural Findings in Visual Object Perception
- Author
-
Hans Op de Beeck, Rainer Goebel, Judith C. Peters, Vision, RS: FPN CN 1, and Netherlands Institute for Neuroscience (NIN)
- Subjects
0301 basic medicine ,Computer science ,Computer Vision ,Neuroscience (miscellaneous) ,Machine learning ,computer.software_genre ,Field (computer science) ,object recognition ,Task (project management) ,lcsh:RC321-571 ,Visual processing ,03 medical and health sciences ,Cellular and Molecular Neuroscience ,0302 clinical medicine ,invariance ,Set (psychology) ,lcsh:Neurosciences. Biological psychiatry. Neuropsychiatry ,business.industry ,fMRI ,Cognitive neuroscience of visual object recognition ,Object (computer science) ,Feature representation ,Editorial ,030104 developmental biology ,ventral visual pathway ,Artificial intelligence ,business ,Neural coding ,Heuristics ,computer ,030217 neurology & neurosurgery ,Neuroscience - Abstract
Recognizing objects despite infinite variations in their appearance is a highly challenging computational task the visual system performs in a remarkably fast, accurate, and robust fashion. The complexity of the underlying mechanisms is reflected in the large proportion of cortical real-estate dedicated to visual processing, as well as in the difficulties encountered when trying to build models whose performance matches human proficiency. The articles in this Research Topic provide an overview of recent advances in our understanding of the neural mechanisms underlying visual object perception, focusing on integrative approaches which encompass both computational and empirical work. Given the vast expanse of topics covered in the discipline of computational visual neuroscience, it is impossible to provide a comprehensive overview of the field's status-quo. Instead, the presented papers highlight interesting extensions to existing models and novel insights into computational principles and their neural underpinnings. Contributions could be coarsely subdivided into three different sections: Two papers focused on implementing biologically-valid learning rules and heuristics in well-established neural models of the visual pathway (i.e., “VisNet” and “HMAX”) to improve flexible object recognition. Three other studies investigated the role of sparseness, selectivity, and correlation in optimizing neural coding of object features. Finally, another set of contributions focused on integrating computational vision models and human brain responses to gain more insights in the computational mechanisms underlying neural object representations.
- Published
- 2016
- Full Text
- View/download PDF
16. Unsupervised invariance learning of transformation sequences in a model of object recognition yields selectivity for non-accidental properties
- Author
-
Sarah Parker and Thomas Serre
- Subjects
Computer science ,Neuroscience (miscellaneous) ,Object processing ,Machine learning ,computer.software_genre ,Image properties ,Inferotemporal cortex ,object recognition ,lcsh:RC321-571 ,03 medical and health sciences ,Cellular and Molecular Neuroscience ,0302 clinical medicine ,ventral stream ,Learning ,invariance ,lcsh:Neurosciences. Biological psychiatry. Neuropsychiatry ,030304 developmental biology ,Original Research ,object constancy ,0303 health sciences ,business.industry ,Cognitive neuroscience of visual object recognition ,Pattern recognition ,Video sequence ,Invariant (physics) ,HMAX ,Unsupervised learning ,Artificial intelligence ,business ,computer ,030217 neurology & neurosurgery ,Object constancy ,Neuroscience - Abstract
Non-accidental properties (NAPs) correspond to image properties that are invariant to changes in viewpoint (e.g., straight vs. curved contours) and are distinguished from metric properties (MPs) that can change continuously with in-depth object rotation (e.g., aspect ratio, degree of curvature, etc). Behavioral and electrophysiological studies of shape processing have demonstrated greater sensitivity to differences in NAPs than in MPs. However, previous work has shown that such sensitivity is lacking in multiple-views models of object recognition such as textsc{Hmax}. These models typically assume that object processing is based on populations of view-tuned neurons with distributed symmetrical bell-shaped tuning that are modulated at least as much by differences in MPs as in NAPs.Here, we test the hypothesis that unsupervised learning of invariances to object transformations may increase the sensitivity to differences in NAPs vs. MPs in textsc{Hmax}. We collected a database of video sequences with objects slowly rotating in-depth in an attempt to mimic sequences viewed during object manipulation by young children during early developmental stages. We show that unsupervised learning yields shape-tuning in higher stages with greater sensitivity to differences in NAPs vs. MPs in agreement with monkey IT data. Together, these results suggest that greater NAP sensitivity may arise from experiencing different in-depth rotations of objects.
- Published
- 2015
- Full Text
- View/download PDF
17. Deformation-specific and deformation-invariant visual object recognition: pose vs. identity recognition of people and deforming objects
- Author
-
Edmund T. Rolls and Tristan J. Webb
- Subjects
Temporal continuity ,media_common.quotation_subject ,Neuroscience (miscellaneous) ,BF ,Engram ,object recognition ,QA76 ,lcsh:RC321-571 ,Cellular and Molecular Neuroscience ,VisNet ,Perception ,Learning rule ,invariance ,Computer vision ,Original Research Article ,object vision ,pose ,lcsh:Neurosciences. Biological psychiatry. Neuropsychiatry ,media_common ,pose recognition ,business.industry ,Cognitive neuroscience of visual object recognition ,deformation ,inferior temporal visual cortex ,Identity recognition ,Invariant (physics) ,Individual person ,Artificial intelligence ,trace learning rule ,Psychology ,business ,Neuroscience - Abstract
When we see a human sitting down, standing up, or walking, we can recognize one of these poses independently of the individual, or we can recognize the individual person, independently of the pose. The same issues arise for deforming objects. For example, if we see a flag deformed by the wind, either blowing out or hanging languidly, we can usually recognize the flag, independently of its deformation; or we can recognize the deformation independently of the identity of the flag. We hypothesize that these types of recognition can be implemented by the primate visual system using temporo-spatial continuity as objects transform as a learning principle. In particular, we hypothesize that pose or deformation can be learned under conditions in which large numbers of different people are successively seen in the same pose, or objects in the same deformation. We also hypothesize that person-specific representations that are independent of pose, and object-specific representations that are independent of deformation and view, could be built, when individual people or objects are observed successively transforming from one pose or deformation and view to another. These hypotheses were tested in a simulation of the ventral visual system, VisNet, that uses temporal continuity, implemented in a synaptic learning rule with a short-term memory trace of previous neuronal activity, to learn invariant representations. It was found that depending on the statistics of the visual input, either pose-specific or deformation-specific representations could be built that were invariant with respect to individual and view; or that identity-specific representations could be built that were invariant with respect to pose or deformation and view. We propose that this is how pose-specific and pose-invariant, and deformation-specific and deformation-invariant, perceptual representations are built in the brain.
- Published
- 2013
18. A Biologically Plausible Transform for Visual Recognition that is Invariant to Translation, Scale, and Rotation
- Author
-
John E. Lisman, David M. Santucci, and Pavel Sountsov
- Subjects
Logarithm ,Computer science ,Visual System ,Neuroscience (miscellaneous) ,hybrid model ,computer.software_genre ,050105 experimental psychology ,object recognition ,lcsh:RC321-571 ,cortico-striatal ,03 medical and health sciences ,Cellular and Molecular Neuroscience ,0302 clinical medicine ,invariance ,unsupervised ,0501 psychology and cognitive sciences ,Spatial frequency analysis ,lcsh:Neurosciences. Biological psychiatry. Neuropsychiatry ,Original Research ,reinforcement ,Interval Detector ,hierarchical ,business.industry ,05 social sciences ,Cognitive neuroscience of visual object recognition ,Pattern recognition ,Invariant (physics) ,Visual recognition ,biological classifier ,Data mining ,Artificial intelligence ,Spatial Frequency Analysis ,business ,computer ,Hybrid model ,030217 neurology & neurosurgery ,Neuroscience - Abstract
Visual object recognition occurs easily despite differences in position, size, and rotation of the object, but the neural mechanisms responsible for this invariance are not known. We have found a set of transforms that achieve invariance in a neurally plausible way. We find that a transform based on local spatial frequency analysis of oriented segments and on logarithmic mapping, when applied twice in an iterative fashion, produces an output image that is unique to the object and that remains constant as the input image is shifted, scaled, or rotated.
- Published
- 2011
- Full Text
- View/download PDF
19. Learning and disrupting invariance in visual recognition with a temporal association rule.
- Author
-
Isik L, Leibo JZ, and Poggio T
- Abstract
Learning by temporal association rules such as Foldiak's trace rule is an attractive hypothesis that explains the development of invariance in visual recognition. Consistent with these rules, several recent experiments have shown that invariance can be broken at both the psychophysical and single cell levels. We show (1) that temporal association learning provides appropriate invariance in models of object recognition inspired by the visual cortex, (2) that we can replicate the "invariance disruption" experiments using these models with a temporal association learning rule to develop and maintain invariance, and (3) that despite dramatic single cell effects, a population of cells is very robust to these disruptions. We argue that these models account for the stability of perceptual invariance despite the underlying plasticity of the system, the variability of the visual world and expected noise in the biological mechanisms.
- Published
- 2012
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.