1
|
Harris IM. Interpreting the orientation of objects: A cross-disciplinary review. Psychon Bull Rev 2024; 31:1503-1515. [PMID: 38302790 PMCID: PMC11358284 DOI: 10.3758/s13423-024-02458-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 01/09/2024] [Indexed: 02/03/2024]
Abstract
Is object orientation an inherent aspect of the shape of the object or is it represented separately and bound to the object shape in a similar way to other features, such as colour? This review brings together findings from neuropsychological studies of patients with agnosia for object orientation and experimental studies of object perception in healthy individuals that provide converging evidence of separate processing of object identity and orientation. Individuals with agnosia for object orientation, which typically results from damage to the right parietal lobe, can recognize objects presented in a range of orientations yet are unable to interpret or discriminate the objects' orientation. Healthy individuals tested with briefly presented objects demonstrate a similar dissociation: object identity is extracted rapidly in an orientation-invariant way, whereas processing the object's orientation is slower, requires attention and is influenced by the degree of departure from the canonical orientation. This asymmetry in processing can sometimes lead to incorrect bindings between the identity and orientation of objects presented in close temporal proximity. Overall, the available evidence indicates that object recognition is achieved in a largely orientation-invariant manner and that interpreting the object's orientation requires an additional step of mapping this orientation-invariant representation to a spatial reference frame.
Collapse
Affiliation(s)
- Irina M Harris
- School of Psychology, University of Sydney, Brennan MacCallum Building A18, Sydney, NSW, 2006, Australia.
| |
Collapse
|
2
|
Izard V, Pica P, Spelke ES. Visual foundations of Euclidean geometry. Cogn Psychol 2022; 136:101494. [PMID: 35751917 DOI: 10.1016/j.cogpsych.2022.101494] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/25/2019] [Revised: 05/10/2022] [Accepted: 06/06/2022] [Indexed: 01/29/2023]
Abstract
Geometry defines entities that can be physically realized in space, and our knowledge of abstract geometry may therefore stem from our representations of the physical world. Here, we focus on Euclidean geometry, the geometry historically regarded as "natural". We examine whether humans possess representations describing visual forms in the same way as Euclidean geometry - i.e., in terms of their shape and size. One hundred and twelve participants from the U.S. (age 3-34 years), and 25 participants from the Amazon (age 5-67 years) were asked to locate geometric deviants in panels of 6 forms of variable orientation. Participants of all ages and from both cultures detected deviant forms defined in terms of shape or size, while only U.S. adults drew distinctions between mirror images (i.e. forms differing in "sense"). Moreover, irrelevant variations of sense did not disrupt the detection of a shape or size deviant, while irrelevant variations of shape or size did. At all ages and in both cultures, participants thus retained the same properties as Euclidean geometry in their analysis of visual forms, even in the absence of formal instruction in geometry. These findings show that representations of planar visual forms provide core intuitions on which humans' knowledge in Euclidean geometry could possibly be grounded.
Collapse
Affiliation(s)
- Véronique Izard
- Université Paris Cité, CNRS, Integrative Neuroscience and Cognition Center, F-75006 Paris, France
- Department of Psychology, Harvard University, 33 Kirkland St, Cambridge, MA 02138, USA.
| | - Pierre Pica
- Instituto do Cérebro, Universidade Federal do Rio grande do Norte, R. do Horto, Lagoa Nova, Natal, RN 59076-550, Brazil
- UMR 7023, Structures Formelles du Langage, Université Paris 8, 2 rue de la Liberté, 93200 Saint-Denis, France
| | - Elizabeth S Spelke
- Department of Psychology, Harvard University, 33 Kirkland St, Cambridge, MA 02138, USA; NSF-STC Center for Brains, Minds and Machines, 43 Vassar St, Cambridge, MA 02139, USA
| |
Collapse
|
3
|
Seyll L, Content A. Letter-Like Shape Recognition in Preschool Children: Does Graphomotor Knowledge Contribute? Front Psychol 2022; 12:726454. [PMID: 35250685 PMCID: PMC8888515 DOI: 10.3389/fpsyg.2021.726454] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2021] [Accepted: 12/06/2021] [Indexed: 11/20/2022] Open
Abstract
Based on evidence that learning new characters through handwriting leads to better recognition than learning through typing, some authors proposed that the graphic motor plans acquired through handwriting contribute to recognition. More recently two alternative explanations have been put forward. First, the advantage of handwriting could be due to the perceptual variability that it provides during learning. Second, a recent study suggests that detailed visual analysis might be the source of the advantage of handwriting over typing. Indeed, in that study, handwriting and composition –a method requiring a detailed visual analysis but no specific graphomotor activity– led to equivalent recognition accuracy, both higher than typing. The aim of the present study was to assess whether the contribution of detailed visual analysis is observed in preschool children and to test the variability hypothesis. To that purpose, three groups of preschool children learned new symbols either by handwriting, typing, or composition. After learning, children performed first a four-alternative recognition task and then a categorization task. The same pattern of results as the one observed in adults emerged in the four-alternative recognition task, confirming the importance of the detailed visual analysis in letter-like shape learning. In addition, results failed to reveal any difference across learning methods in the categorization task. The latter results provide no evidence for the variability hypothesis which would predict better categorization after handwriting than after typing or composition.
Collapse
Affiliation(s)
- Lola Seyll
- Laboratoire Cognition Langage et Développement, Centre de Recherche Cognition et Neurosciences, Université Libre de Bruxelles (ULB), Brussels, Belgium
| | - Alain Content
- Laboratoire Cognition Langage et Développement, Centre de Recherche Cognition et Neurosciences, Université Libre de Bruxelles (ULB), Brussels, Belgium
| |
Collapse
|
4
|
Lonardo L, Versace E, Huber L. Recognition of rotated objects and cognitive offloading in dogs. iScience 2022; 25:103820. [PMID: 35198883 PMCID: PMC8841888 DOI: 10.1016/j.isci.2022.103820] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2021] [Revised: 12/30/2021] [Accepted: 01/21/2022] [Indexed: 11/19/2022] Open
Abstract
Recognition of rotated images can challenge visual systems. Humans often diminish the load of cognitive tasks employing bodily actions (cognitive offloading). To investigate these phenomena from a comparative perspective, we trained eight dogs (Canis familiaris) to discriminate between bidimensional shapes. We then tested the dogs with rotated versions of the same shapes, while measuring their accuracy and head tilts. Although generalization to rotated stimuli challenged dogs (overall accuracy: 55%), three dogs performed differently from chance level with rotated stimuli. The amplitude of stimulus rotation did not influence dogs' performance. Interestingly, dogs tilted their head following the direction and amplitude of rotated stimuli. These small head movements did not influence their performance. Hence, we show that dogs might be capable of recognizing rotated 2D objects, but they do not use a cognitive offloading strategy in this task. This work paves the way to further investigation of cognitive offloading in non-human species.
Collapse
Affiliation(s)
- Lucrezia Lonardo
- Comparative Cognition, Messerli Research Institute, University of Veterinary Medicine Vienna, Medical University of Vienna and University of Vienna, 1210, Vienna, Austria
| | - Elisabetta Versace
- School of Biological and Behavioural Sciences, Department of Biological and Experimental Psychology, Queen Mary University of London, E1 4NS, London, UK
| | - Ludwig Huber
- Comparative Cognition, Messerli Research Institute, University of Veterinary Medicine Vienna, Medical University of Vienna and University of Vienna, 1210, Vienna, Austria
| |
Collapse
|
5
|
Hollis J, Humphreys GW, Allen PM. Intermediate, Wholistic Shape Representation in Object Recognition: A Pre-Attentive Stage of Processing? Front Hum Neurosci 2021; 15:761174. [PMID: 35002652 PMCID: PMC8735852 DOI: 10.3389/fnhum.2021.761174] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/19/2021] [Accepted: 11/18/2021] [Indexed: 11/13/2022] Open
Abstract
Evidence is presented for intermediate, wholistic visual representations of objects and non-objects that are computed online and independent of visual attention. Short-term visual priming was examined between visually similar shapes, with targets either falling at the (valid) location cued by primes or at another (invalid) location. Object decision latencies were facilitated when the overall shapes of the stimuli were similar irrespective of whether the location of the prime was valid or invalid, with the effects being equally large for object and non-object targets. In addition, the effects were based on the overall outlines of the stimuli and low spatial frequency components, not on local parts. In conclusion, wholistic shape representations based on outline form, are rapidly computed online during object recognition. Moreover, activation of common wholistic shape representations prime the processing of subsequent objects and non-objects irrespective of whether they appear at attended or unattended locations. Rapid derivation of wholistic form provides a key intermediate stage of object recognition.
Collapse
Affiliation(s)
- Jarrod Hollis
- Vision and Hearing Sciences Research Centre, Anglia Ruskin University, Cambridge, United Kingdom
| | - Glyn W. Humphreys
- Department of Experimental Psychology, University of Oxford, Oxford, United Kingdom
| | - Peter M. Allen
- Vision and Hearing Sciences Research Centre, Anglia Ruskin University, Cambridge, United Kingdom
| |
Collapse
|
6
|
Baker N, Kellman PJ. Constant curvature modeling of abstract shape representation. PLoS One 2021; 16:e0254719. [PMID: 34339436 PMCID: PMC8328290 DOI: 10.1371/journal.pone.0254719] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2020] [Accepted: 07/01/2021] [Indexed: 11/19/2022] Open
Abstract
How abstract shape is perceived and represented poses crucial unsolved problems in human perception and cognition. Recent findings suggest that the visual system may encode contours as sets of connected constant curvature segments. Here we describe a model for how the visual system might recode a set of boundary points into a constant curvature representation. The model includes two free parameters that relate to the degree to which the visual system encodes shapes with high fidelity vs. the importance of simplicity in shape representations. We conducted two experiments to estimate these parameters empirically. Experiment 1 tested the limits of observers’ ability to discriminate a contour made up of two constant curvature segments from one made up of a single constant curvature segment. Experiment 2 tested observers’ ability to discriminate contours generated from cubic splines (which, mathematically, have no constant curvature segments) from constant curvature approximations of the contours, generated at various levels of precision. Results indicated a clear transition point at which discrimination becomes possible. The results were used to fix the two parameters in our model. In Experiment 3, we tested whether outputs from our parameterized model were predictive of perceptual performance in a shape recognition task. We generated shape pairs that had matched physical similarity but differed in representational similarity (i.e., the number of segments needed to describe the shapes) as assessed by our model. We found that pairs of shapes that were more representationally dissimilar were also easier to discriminate in a forced choice, same/different task. The results of these studies provide evidence for constant curvature shape representation in human visual perception and provide a testable model for how abstract shape descriptions might be encoded.
Collapse
Affiliation(s)
- Nicholas Baker
- Department of Psychology, University of California Los Angeles, Los Angeles, California, United States of America
- * E-mail:
| | - Philip J. Kellman
- Department of Psychology, University of California Los Angeles, Los Angeles, California, United States of America
| |
Collapse
|
7
|
Popovkina DV, Palmer J, Moore CM, Boynton GM. Is there a serial bottleneck in visual object recognition? J Vis 2021; 21:15. [PMID: 33704373 PMCID: PMC7961120 DOI: 10.1167/jov.21.3.15] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/04/2022] Open
Abstract
Divided attention has little effect for simple tasks, such as luminance detection, but it has large effects for complex tasks, such as semantic categorization of masked words. Here, we asked whether the semantic categorization of visual objects shows divided attention effects as large as those observed for words, or as small as those observed for simple feature judgments. Using a dual-task paradigm with nameable object stimuli, performance was compared with the predictions of serial and parallel models. At the extreme, parallel processes with unlimited capacity predict no effect of divided attention; alternatively, an all-or-none serial process makes two predictions: a large divided attention effect (lower accuracy for dual-task trials, compared to single-task trials) and a negative response correlation in dual-task trials (a given response is more likely to be incorrect when the response about the other stimulus is correct). These predictions were tested in two experiments examining object judgments. In both experiments, there was a large divided attention effect and a small negative correlation in responses. The magnitude of these effects was larger than for simple features, but smaller than for words. These effects were consistent with serial models, and rule out some but not all parallel models. More broadly, the results help establish one of the first examples of likely serial processing in perception.
Collapse
Affiliation(s)
- Dina V Popovkina
- Department of Psychology, University of Washington, Seattle, WA, USA.,
| | - John Palmer
- Department of Psychology, University of Washington, Seattle, WA, USA.,
| | - Cathleen M Moore
- Department of Psychological and Brain Sciences, University of Iowa, Iowa City, IA, USA.,
| | | |
Collapse
|
8
|
Blything R, Biscione V, Vankov II, Ludwig CJH, Bowers JS. The human visual system and CNNs can both support robust online translation tolerance following extreme displacements. J Vis 2021; 21:9. [PMID: 33620380 PMCID: PMC7910631 DOI: 10.1167/jov.21.2.9] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/03/2022] Open
Abstract
Visual translation tolerance refers to our capacity to recognize objects over a wide range of different retinal locations. Although translation is perhaps the simplest spatial transform that the visual system needs to cope with, the extent to which the human visual system can identify objects at previously unseen locations is unclear, with some studies reporting near complete invariance over 10 degrees and other reporting zero invariance at 4 degrees of visual angle. Similarly, there is confusion regarding the extent of translation tolerance in computational models of vision, as well as the degree of match between human and model performance. Here, we report a series of eye-tracking studies (total N = 70) demonstrating that novel objects trained at one retinal location can be recognized at high accuracy rates following translations up to 18 degrees. We also show that standard deep convolutional neural networks (DCNNs) support our findings when pretrained to classify another set of stimuli across a range of locations, or when a global average pooling (GAP) layer is added to produce larger receptive fields. Our findings provide a strong constraint for theories of human vision and help explain inconsistent findings previously reported with convolutional neural networks (CNNs).
Collapse
Affiliation(s)
- Ryan Blything
- School of Psychological Science, University of Bristol, Bristol, UK.,
| | - Valerio Biscione
- School of Psychological Science, University of Bristol, Bristol, UK.,
| | - Ivan I Vankov
- Department of Cognitive Science and Psychology, Sofia, New Bulgarian University, Bulgaria.,
| | | | - Jeffrey S Bowers
- School of Psychological Science, University of Bristol, Bristol, UK.,
| |
Collapse
|
9
|
The impact of graphic motor programs and detailed visual analysis on letter-like shape recognition. Cognition 2020; 205:104443. [DOI: 10.1016/j.cognition.2020.104443] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2019] [Revised: 08/14/2020] [Accepted: 08/18/2020] [Indexed: 11/21/2022]
|
10
|
Dujmović M, Malhotra G, Bowers JS. What do adversarial images tell us about human vision? eLife 2020; 9:e55978. [PMID: 32876562 PMCID: PMC7467732 DOI: 10.7554/elife.55978] [Citation(s) in RCA: 17] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2020] [Accepted: 08/14/2020] [Indexed: 12/24/2022] Open
Abstract
Deep convolutional neural networks (DCNNs) are frequently described as the best current models of human and primate vision. An obvious challenge to this claim is the existence of adversarial images that fool DCNNs but are uninterpretable to humans. However, recent research has suggested that there may be similarities in how humans and DCNNs interpret these seemingly nonsense images. We reanalysed data from a high-profile paper and conducted five experiments controlling for different ways in which these images can be generated and selected. We show human-DCNN agreement is much weaker and more variable than previously reported, and that the weak agreement is contingent on the choice of adversarial images and the design of the experiment. Indeed, we find there are well-known methods of generating images for which humans show no agreement with DCNNs. We conclude that adversarial images still pose a challenge to theorists using DCNNs as models of human vision.
Collapse
Affiliation(s)
- Marin Dujmović
- School of Psychological Science, University of BristolBristolUnited Kingdom
| | - Gaurav Malhotra
- School of Psychological Science, University of BristolBristolUnited Kingdom
| | - Jeffrey S Bowers
- School of Psychological Science, University of BristolBristolUnited Kingdom
| |
Collapse
|
11
|
Wood JN, Wood SMW. One-shot learning of view-invariant object representations in newborn chicks. Cognition 2020; 199:104192. [PMID: 32199170 DOI: 10.1016/j.cognition.2020.104192] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/08/2018] [Revised: 01/13/2020] [Accepted: 01/15/2020] [Indexed: 11/19/2022]
Abstract
Can newborn brains perform one-shot learning? To address this question, we reared newborn chicks in strictly controlled environments containing a single view of a single object, then tested their object recognition performance across 24 uniformly-spaced viewpoints. We found that chicks can build view-invariant object representations from a single view of an object: a case of one-shot learning in newborn brains. Chicks can also build the same view-invariant object representation from different views of an object, showing that newborn brains converge on common object representations from different sets of sensory inputs. Finally, by rearing chicks with larger numbers of object views, we found that chicks develop enhanced recognition for familiar views. These results illuminate the earliest stages of object recognition, revealing (1) powerful one-shot learning that builds invariant object representations from the first views of an object and (2) view-based learning that enriches object representations, producing enhanced recognition for familiar views.
Collapse
Affiliation(s)
- Justin N Wood
- Indiana University, Department of Informatics, 700 N Woodlawn Ave., Bloomington, IN 47408, United States of America.
| | - Samantha M W Wood
- Indiana University, Department of Informatics, 700 N Woodlawn Ave., Bloomington, IN 47408, United States of America.
| |
Collapse
|
12
|
Holler DE, Fabbri S, Snow JC. Object responses are highly malleable, rather than invariant, with changes in object appearance. Sci Rep 2020; 10:4654. [PMID: 32170123 PMCID: PMC7070005 DOI: 10.1038/s41598-020-61447-8] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/06/2020] [Accepted: 02/17/2020] [Indexed: 11/09/2022] Open
Abstract
Theoretical frameworks of human vision argue that object responses remain stable, or 'invariant', despite changes in viewing conditions that can alter object appearance but not identity. Here, in a major departure from previous approaches that have relied on two-dimensional (2-D) images to study object processing, we demonstrate that changes in an object's appearance, but not its identity, can lead to striking shifts in behavioral responses to objects. We used inverse multidimensional scaling (MDS) to measure the extent to which arrangements of objects in a sorting task were similar or different when the stimuli were displayed as scaled 2-D images, three-dimensional (3-D) augmented reality (AR) projections, or real-world solids. We were especially interested in whether sorting behavior in each display format was based on conceptual (e.g., typical location) versus physical object characteristics. We found that 2-D images of objects were arranged according to conceptual (typical location), but not physical, properties. AR projections, conversely, were arranged primarily according to physical properties such as real-world size, elongation and weight, but not conceptual properties. Real-world solid objects, unlike both 2-D and 3-D images, were arranged using multidimensional criteria that incorporated both conceptual and physical object characteristics. Our results suggest that object responses can be strikingly malleable, rather than invariant, with changes in the visual characteristics of the stimulus. The findings raise important questions about limits of invariance in object processing, and underscore the importance of studying responses to richer stimuli that more closely resemble those we encounter in real-world environments.
Collapse
Affiliation(s)
| | - Sara Fabbri
- Department of Psychology, University of Nevada, Reno, USA.,Department of Experimental Psychology, University of Groningen, Groningen, the Netherlands
| | | |
Collapse
|
13
|
Humphries A, Chen Z, Wiltshire J. Repetition priming with no antipriming in picture identification. Vision Res 2020; 168:9-17. [PMID: 32044587 DOI: 10.1016/j.visres.2019.09.011] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2018] [Revised: 09/14/2019] [Accepted: 09/17/2019] [Indexed: 11/29/2022]
Abstract
Previous studies have shown that the processing of a stimulus is facilitated when that stimulus is repeated compared to when it appears the first time, and this phenomenon is called repetition priming (RP). One explanation for RP is that initial processing of a stimulus strengthens connections within the visual representation, enabling subsequent processing of the same stimulus to be more efficient. More recently, it has been reported that presenting an object with features that overlap with those in a subsequent stimulus impairs the latter's processing, and this cost is termed antipriming (AP). AP is said to be the natural antithesis of RP, and it manifests when two objects share component features, thereby having overlapping representations. In two experiments, we investigated RP and AP in a picture naming task. Following previous research, we used a 4-phase paradigm, in which RP and AP were measured, respectively, by an increase or a decrease in performance for repeated or novel stimuli in Phase 4 compared with the baseline performance in Phase 2. We used a fully randomized design in Experiment 1, and a pseudo-randomized design in stimulus selection but a randomized design in presentation location in Experiment 2. We found robust RP in both experiments, but neither experiment showed any evidence of AP. Our results indicate that RP and AP do not always manifest within the same experiment, and that the relationship between these two effects may be more complex than previously understood.
Collapse
Affiliation(s)
- Ailsa Humphries
- Department of Psychology, University of Canterbury, Christchurch, New Zealand.
| | - Zhe Chen
- Department of Psychology, University of Canterbury, Christchurch, New Zealand
| | - Jonathan Wiltshire
- Department of Psychology, University of Canterbury, Christchurch, New Zealand
| |
Collapse
|
14
|
Emotional learning promotes perceptual predictions by remodeling stimulus representation in visual cortex. Sci Rep 2019; 9:16867. [PMID: 31727912 PMCID: PMC6856165 DOI: 10.1038/s41598-019-52615-6] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2018] [Accepted: 06/13/2019] [Indexed: 12/13/2022] Open
Abstract
Emotions exert powerful effects on perception and memory, notably by modulating activity in sensory cortices so as to capture attention. Here, we examine whether emotional significance acquired by a visual stimulus can also change its cortical representation by linking neuronal populations coding for different memorized versions of the same stimulus, a mechanism that would facilitate recognition across different appearances. Using fMRI, we show that after pairing a given face with threat through conditioning, viewing this face activates the representation of another viewpoint of the same person, which itself was never conditioned, leading to robust repetition-priming across viewpoints in the ventral visual stream (including medial fusiform, lateral occipital, and anterior temporal cortex). We also observed a functional-anatomical segregation for coding view-invariant and view-specific identity information. These results indicate emotional signals may induce plasticity of stimulus representations in visual cortex, serving to generate new sensory predictions about different appearances of threat-associated stimuli.
Collapse
|
15
|
Surface diagnosticity predicts the high-level representation of regular and irregular object shape in human vision. Atten Percept Psychophys 2019; 81:1589-1608. [PMID: 30864108 PMCID: PMC6647524 DOI: 10.3758/s13414-019-01698-4] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
The human visual system has an extraordinary capacity to compute three-dimensional (3D) shape structure for both geometrically regular and irregular objects. The goal of this study was to shed new light on the underlying representational structures that support this ability. Observers (N = 85) completed two complementary perceptual tasks. Experiment 1 involved whole–part matching of image parts to whole geometrically regular and irregular novel object shapes. Image parts comprised either regions of edge contour, volumetric parts, or surfaces. Performance was better for irregular than for regular objects and interacted with part type: volumes yielded better matching performance than surfaces for regular but not for irregular objects. The basis for this effect was further explored in Experiment 2, which used implicit part–whole repetition priming. Here, we orthogonally manipulated shape regularity and a new factor of surface diagnosticity (how predictive a single surface is of object identity). The results showed that surface diagnosticity, not object shape regularity, determined the differential processing of volumes and surfaces. Regardless of shape regularity, objects with low surface diagnosticity were better primed by volumes than by surfaces. In contrast, objects with high surface diagnosticity showed the opposite pattern. These findings are the first to show that surface diagnosticity plays a fundamental role in object recognition. We propose that surface-based shape primitives—rather than volumetric parts—underlie the derivation of 3D object shape in human vision.
Collapse
|
16
|
Abstract
Two decades of research in semantic priming has provided substantial evidence for a distinction between short- and long-term semantic priming effects. Early models of cognition suggested a single mechanism to explain priming at short and long lags. Later models refuted this explanation and proposed that different mechanisms are necessary to account for different durations of priming effects. Two alternative explanations of long-term semantic priming effects have been proposed in the extant literature. The first explanation is that long-term semantic priming effects rely upon the incremental strengthening of abstract semantic memory representations. The second explanation is that long-term priming is the result of memory for prior cognitive operations. In two experiments, we used different semantic content - word meaning versus category membership - to investigate the mechanisms accountable for long-term semantic priming. Evidence from the two experiments suggests that long-term semantic priming effects are due to different memory processes for different semantic content. Long-term semantic priming of word meanings was dependent on strengthening abstract semantic memory representations and persistent priming of category membership was dependent on memory for prior cognitive operations.
Collapse
|
17
|
Effects of canonical color, luminance, and orientation on sustained inattentional blindness for scenes. Atten Percept Psychophys 2018; 80:1833-1846. [PMID: 29987532 DOI: 10.3758/s13414-018-1558-z] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Whether scene gist perception occurs automatically and unconsciously has been the subject of much debate. In addition to demonstrating a new method that adapts the Mack and Rock (1998) inattentional blindness cross procedure to allow for sustained inattentional blindness over a large number of trials, we report evidence from a series of experiments that shows that canonical scene features reduce inattentional blindness to scenes by facilitating the extraction of scene gist. When attentional demands are high, the combination of canonical color, canonical luminance, and canonical orientation reduces rates of inattentional blindness. However, when attentional demands are reduced, canonical features are independently sufficient to facilitate gist extraction and to capture attention. These results demonstrate that canonical color, canonical luminance, and canonical orientation all contribute to scene gist perception, and that when attentional demands are high, only highly canonical stimuli are sufficient to capture attention.
Collapse
|
18
|
Quinlan PT, Allen K. The nature of shape constancy mechanisms as revealed by shape priming. J Vis 2018; 18:14. [PMID: 30029222 DOI: 10.1167/18.6.14] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Five shape priming experiments are reported in which the target was either a five- or six-sided line-drawn figure and participants made a speeded two-alternative forced-choice judgment about the target's number of sides. On priming trials, the target was preceded by a briefly presented smaller line figure (the prime) and performance on these trials was gauged relative to a no-prime condition. In the first two experiments, primes were rendered invisible by the presentation of a backwards visual noise mask, respectively for a short (∼40 ms) or long duration (∼93 ms). No reliable priming effects arose under masked conditions. When these experiments were repeated without the mask, participants were speeded when the prime and target were related by a rigid through-the-plane rotation but not when the prime was a nonrigid, stretched version of the target. The same pattern of priming effects arose when, in a final experiment, novel irregular shapes were used. Collectively, the data reveal the operation of shape constancy mechanisms that are particularly sensitive to shape rigidity. The findings suggest that the visual system attempts to secure a correspondence between the rapid and successive presentations of the prime and the target by matching shapes according to a rigidity constraint.
Collapse
Affiliation(s)
- Philip T Quinlan
- Department of Psychology, The University of York, Heslington, York, UK
| | - Keith Allen
- Department of Philosophy, The University of York, Heslington, York, UK
| |
Collapse
|
19
|
Ramírez FM. Orientation Encoding and Viewpoint Invariance in Face Recognition: Inferring Neural Properties from Large-Scale Signals. Neuroscientist 2018; 24:582-608. [PMID: 29855217 DOI: 10.1177/1073858418769554] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Viewpoint-invariant face recognition is thought to be subserved by a distributed network of occipitotemporal face-selective areas that, except for the human anterior temporal lobe, have been shown to also contain face-orientation information. This review begins by highlighting the importance of bilateral symmetry for viewpoint-invariant recognition and face-orientation perception. Then, monkey electrophysiological evidence is surveyed describing key tuning properties of face-selective neurons-including neurons bimodally tuned to mirror-symmetric face-views-followed by studies combining functional magnetic resonance imaging (fMRI) and multivariate pattern analyses to probe the representation of face-orientation and identity information in humans. Altogether, neuroimaging studies suggest that face-identity is gradually disentangled from face-orientation information along the ventral visual processing stream. The evidence seems to diverge, however, regarding the prevalent form of tuning of neural populations in human face-selective areas. In this context, caveats possibly leading to erroneous inferences regarding mirror-symmetric coding are exposed, including the need to distinguish angular from Euclidean distances when interpreting multivariate pattern analyses. On this basis, this review argues that evidence from the fusiform face area is best explained by a view-sensitive code reflecting head angular disparity, consistent with a role of this area in face-orientation perception. Finally, the importance is stressed of explicit models relating neural properties to large-scale signals.
Collapse
Affiliation(s)
- Fernando M Ramírez
- 1 Bernstein Center for Computational Neuroscience, Charité Universitätsmedizin Berlin, Berlin, Germany
| |
Collapse
|
20
|
Elgendi M, Kumar P, Barbic S, Howard N, Abbott D, Cichocki A. Subliminal Priming-State of the Art and Future Perspectives. Behav Sci (Basel) 2018; 8:bs8060054. [PMID: 29849006 PMCID: PMC6027235 DOI: 10.3390/bs8060054] [Citation(s) in RCA: 23] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/26/2018] [Revised: 05/18/2018] [Accepted: 05/23/2018] [Indexed: 11/26/2022] Open
Abstract
The influence of subliminal priming (behavior outside of awareness) in humans is an interesting phenomenon and its understanding is crucial as it can impact behavior, choices, and actions. Given this, research about the impact of priming continues to be an area of investigative interest, and this paper provides a technical overview of research design strengths and issues in subliminal priming research. Efficient experiments and protocols, as well as associated electroencephalographic and eye movement data analyses, are discussed in detail. We highlight the strengths and weaknesses of different priming experiments that have measured affective (emotional) and cognitive responses. Finally, very recent approaches and findings are described to summarize and emphasize state-of-the-art methods and potential future directions in research marketing and other commercial applications.
Collapse
Affiliation(s)
- Mohamed Elgendi
- Faculty of Medicine, University of British Columbia, Vancouver, BC V6T 1Z3, Canada.
- BC Children's & Women's Hospital, Vancouver, BC V6H 3N1, Canada.
- School of Electrical and Computer Engineering, University of British Columbia, Vancouver, BC V6T 1Z4, Canada.
| | - Parmod Kumar
- College of Engineering, Madda Walabu University, Bale Robe 247, Ethiopia.
| | - Skye Barbic
- Department of Occupational Science and Occupational Therapy, University of British Columbia, Vancouver, BC V6T 2B5, Canada.
| | - Newton Howard
- Nuffield Department of Surgical Sciences, University of Oxford, Oxford OX1 2JD, UK.
| | - Derek Abbott
- School of Electrical and Electronic Engineering, The University of Adelaide, Adelaide SA 5005, Australia.
- Centre for Biomedical Engineering, The University of Adelaide, Adelaide SA 5005, Australia.
| | - Andrzej Cichocki
- Skoltech Center for Computational and Data-Intensive Science and Engineering, Skolkowo Institute of Science and Technology, Moscow 143026, Russia.
- College of Computer Science, Hangzhou Dianzi University, Hangzhou 310000, China.
- Department of Informatics, Nicolaus Copernicus University, 87-100 Torun, Poland.
| |
Collapse
|
21
|
Yoon EY, Humphreys GW. Dissociative effects of viewpoint and semantic priming on action and semantic decisions: Evidence for dual routes to action from vision. Q J Exp Psychol (Hove) 2018; 60:601-23. [PMID: 17455069 DOI: 10.1080/17470210600701007] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/24/2022]
Abstract
This article reports evidence that two variables—object viewpoint and semantic priming—differentially affect action and semantic decisions to objects. Changing the viewing angle of an object, so that its graspable parts are oriented away from observers, disrupted action decisions but had little effect on semantic categorization. In contrast, semantic priming influenced semantic categorization and object naming, but not action decisions. Action priming, however, did not affect performance. These selective effects are consistent with the existence of separate semantic and direct (nonsemantic) routes to action from vision.
Collapse
Affiliation(s)
- Eun Young Yoon
- Behavioural Brain Science Centre, School of Psychology, University of Birmingham, Birmingham, UK
| | | |
Collapse
|
22
|
Incremental learning of perceptual and conceptual representations and the puzzle of neural repetition suppression. Psychon Bull Rev 2017; 23:1055-71. [PMID: 27294423 DOI: 10.3758/s13423-015-0855-y] [Citation(s) in RCA: 20] [Impact Index Per Article: 2.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Incremental learning models of long-term perceptual and conceptual knowledge hold that neural representations are gradually acquired over many individual experiences via Hebbian-like activity-dependent synaptic plasticity across cortical connections of the brain. In such models, variation in task relevance of information, anatomic constraints, and the statistics of sensory inputs and motor outputs lead to qualitative alterations in the nature of representations that are acquired. Here, the proposal that behavioral repetition priming and neural repetition suppression effects are empirical markers of incremental learning in the cortex is discussed, and research results that both support and challenge this position are reviewed. Discussion is focused on a recent fMRI-adaptation study from our laboratory that shows decoupling of experience-dependent changes in neural tuning, priming, and repetition suppression, with representational changes that appear to work counter to the explicit task demands. Finally, critical experiments that may help to clarify and resolve current challenges are outlined.
Collapse
|
23
|
McCarthy JD, Song JH. Global attention facilitates the planning, but not execution of goal-directed reaches. J Vis 2017; 16:7. [PMID: 27467450 PMCID: PMC4979628 DOI: 10.1167/16.9.7] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
In daily life, humans interact with multiple objects in complex environments. A large body of literature demonstrates that target selection is biased toward recently attended features, such that reaches are faster and trajectory curvature is reduced when target features (i.e., color) are repeated (priming of pop-out). In the real world, however, objects are comprised of several features—some of which may be more suitable for action than others. When fetching a mug from the cupboard, for example, attention not only has to be allocated to the object, but also the handle. To date, no study has investigated the impact of hierarchical feature organization on target selection for action. Here, we employed a color-oddity search task in which targets were Pac-men (i.e., a circle with a triangle cut out) oriented to be either consistent or inconsistent with the percept of a global Kanizsa triangle. We found that reaches were initiated faster when a task-irrelevant illusory figure was present independent of color repetition. Additionally, consistent with priming of pop-out, both reach planning and execution were facilitated when local target colors were repeated, regardless of whether a global figure was present. We also demonstrated that figures defined by illusory, but not real contours, afforded an early target selection benefit. In sum, these findings suggest that when local targets are perceptually grouped to form an illusory surface, attention quickly spreads across the global figure and facilitates the early stage of reach planning, but not execution. In contrast, local color priming is evident throughout goal-directed reaching.
Collapse
|
24
|
Bülthoff I, Newell FN. Crossmodal priming of unfamiliar faces supports early interactions between voices and faces in person perception. VISUAL COGNITION 2017. [DOI: 10.1080/13506285.2017.1290729] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
Affiliation(s)
| | - Fiona N. Newell
- School of Psychology and Institute of Neuroscience, Trinity College Dublin, Dublin 2, Ireland
| |
Collapse
|
25
|
Five shades of grey: Generalization in distractor-based retrieval of S-R episodes. Atten Percept Psychophys 2016; 78:2307-2312. [DOI: 10.3758/s13414-016-1210-8] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
26
|
Carbon CC, Leder H. When Feature Information Comes First! Early Processing of Inverted Faces. Perception 2016; 34:1117-34. [PMID: 16245489 DOI: 10.1068/p5192] [Citation(s) in RCA: 45] [Impact Index Per Article: 5.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
Abstract
We investigated the early stages of face recognition and the role of featural and holistic face information. We exploited the fact that, on inversion, the alienating disorientation of the eyes and mouth in thatcherised faces is hardly detectable. This effect allows featural and holistic information to be dissociated and was used to test specific face-processing hypotheses. In inverted thatcherised faces, the cardinal features are already correctly oriented, whereas in undistorted faces, the whole Gestalt is coherent but all information is disoriented. Experiment 1 and experiment 3 revealed that, for inverted faces, featural information processing precedes holistic information. Moreover, the processing of contextual information is necessary to process local featural information within a short presentation time (26 ms). Furthermore, for upright faces, holistic information seems to be available faster than for inverted faces (experiment 2). These differences in processing inverted and upright faces presumably cause the differential importance of featural and holistic information for inverted and upright faces.
Collapse
Affiliation(s)
- Claus-Christian Carbon
- Institute of Cognitive Psychology, Freie Universität Berlin, Habelschwerdter Allee 45, D 14169 Berlin, Germany.
| | | |
Collapse
|
27
|
Srinivas K, Greene AJ, Easton RD. Implicit and Explicit Memory for Haptically Experienced Two-Dimensional Patterns. Psychol Sci 2016. [DOI: 10.1111/j.1467-9280.1997.tb00419.x] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022] Open
Abstract
Recent research with visual objects has delineated important representational differences between memory measures that tap identification (implicit tests) and measures that require episodic recognition (explicit tests) We investigated whether these differences reflect a fundamental architecture for the representation of object information in memory In the present experiment, we contrasted identification and episodic recognition for haptically presented two-dimensional patterns Haptic identification was not affected by elaborative processing at study, whereas haptic episodic recognition was enhanced by elaborative processing This finding suggests important similarities in the organization of object information in the visual and haptic modalities
Collapse
|
28
|
Abstract
Many results implicate perceptual processing in repetition priming, but little is known of potential mechanisms for priming. A new method was used to help determine the processing stage at which priming occurs. Priming pictures were presented under dominance or suppression generated by binocular rivalry. Although low-level, sensory attributes can be processed under rivalry suppression, there is no evidence that repetition priming can be supported by such low-level processing. Priming was found only for stimuli that were processed sufficiently to be identified in the priming stage. The results demonstrate that repetition priming requires processing of stimulus attributes into relatively high-level representations.
Collapse
|
29
|
Tarr MJ, Bülthoff HH, Zabinski M, Blanz V. To What Extent Do Unique Parts Influence Recognition Across Changes in Viewpoint? Psychol Sci 2016. [DOI: 10.1111/j.1467-9280.1997.tb00439.x] [Citation(s) in RCA: 112] [Impact Index Per Article: 14.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022] Open
Abstract
We investigated how varying the number of unique parts within an object influences recognition across changes in viewpoint The stimuli were shaded objects composed of five three-dimensional volumes linked end to end with varying connection angles Of the five volumes, zero, one, three, or five were qualitatively distinct (e g, brick vs cone), the rest being tubes Sequential-matching and naming tasks were used to assess the recognition of these stimuli over rotations in depth Three major results stand out First, regardless of the number of distinct parts, there was increasingly poorer recognition performance with increasing change in viewpoint Second, the impact of viewpoint change for objects with one unique part was less than that for the other objects Third, additional parts beyond a single unique part produced strong viewpoint dependency comparable to that obtained for objects with no distinct parts Thus, visual recognition may be explained by a view-based theory in which viewpoint-specific representations encode both quantitative and qualitative features.
Collapse
|
30
|
Abstract
Masked pictures of objects were flashed so briefly that only 13.5% of them could be named. Forced-choice accuracy for the unidentified objects was at chance. When the pictures were shown again, about 15 min and 20 intervening trials later, without any indication of possible repetitions, naming accuracy increased to 34.5%. The priming was completely visual, rather than semantic or verbal, as there was no priming of same-name, different-shape images. This is the first demonstration of facilitatory visual recognition priming by unidentified pictures when the subject could not anticipate if, when, or where the previously unidentified picture was to be shown again. A change in the position of the object reduced but did not eliminate the priming, allowing a speculation that the locus of subliminal visual priming is at an intermediate stage in the ventral cortical pathway for shape recognition.
Collapse
|
31
|
Abstract
Research on animals, infants, children, and adults provides evidence that distinct cognitive systems underlie navigation and object recognition. Here we examine whether and how these systems interact when children interpret 2D edge-based perspectival line drawings of scenes and objects. Such drawings serve as symbols early in development, and they preserve scene and object geometry from canonical points of view. Young children show limits when using geometry both in non-symbolic tasks and in symbolic map tasks that present 3D contexts from unusual, unfamiliar points of view. When presented with the familiar viewpoints in perspectival line drawings, however, do children engage more integrated geometric representations? In three experiments, children successfully interpreted line drawings with respect to their depicted scene or object. Nevertheless, children recruited distinct processes when navigating based on the information in these drawings, and these processes depended on the context in which the drawings were presented. These results suggest that children are flexible but limited in using geometric information to form integrated representations of scenes and objects, even when interpreting spatial symbols that are highly familiar and faithful renditions of the visual world.
Collapse
Affiliation(s)
- Moira R. Dillon
- Psychology Department, Harvard University, Cambridge, MA 02138, USA
| | | |
Collapse
|
32
|
The visual system supports online translation invariance for object identification. Psychon Bull Rev 2015; 23:432-8. [DOI: 10.3758/s13423-015-0916-2] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
|
33
|
Repetition priming in picture naming: sustained learning through the speeding of multiple processes. Psychon Bull Rev 2015; 21:1301-8. [PMID: 24590468 DOI: 10.3758/s13423-014-0610-9] [Citation(s) in RCA: 18] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Picture naming has been used by vision researchers to study object identification, by language researchers to study word production, and by memory researchers to study implicit memory. Response times for naming repeated pictures decrease with successive repetitions. Repetition priming in picture naming involves an implicit, nonhippocampal form of memory. In this review, the processes speeded with repetition are decomposed, the time course of the effect is characterized, the factors affecting the magnitude of priming are enumerated, and possible mechanisms of priming are evaluated. Both behavioral response time and neuroimaging studies are considered. The processes that are speeded with repetition include high-level object identification and word production processes, but not low-level visual processes or articulation. Repetition priming lasts for at least several weeks and follows a typical forgetting function. The mechanism of priming is concluded to be speeded completion of the component processes of picture naming.
Collapse
|
34
|
Abstract
Performance on visual short-term memory for features has been known to depend on stimulus complexity, spatial layout, and feature context. However, with few exceptions, memory capacity has been measured for abruptly appearing, single-instance displays. In everyday life, objects often have a spatiotemporal history as they or the observer move around. In three experiments, we investigated the effect of spatiotemporal history on explicit memory for color. Observers saw a memory display emerge from behind a wall, after which it disappeared again. The test display then emerged from either the same side as the memory display or the opposite side. In the first two experiments, memory improved for intermediate set sizes when the test display emerged in the same way as the memory display. A third experiment then showed that the benefit was tied to the original motion trajectory and not to the display object per se. The results indicate that memory for color is embedded in a richer episodic context that includes the spatiotemporal history of the display.
Collapse
|
35
|
The resilience of object predictions: early recognition across viewpoints and exemplars. Psychon Bull Rev 2015; 21:682-8. [PMID: 24234168 DOI: 10.3758/s13423-013-0546-5] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Recognition of everyday objects can be facilitated by top-down predictions. We have proposed that these predictions are derived from rudimentary image information, or gist, extracted rapidly from the low spatial frequencies (LSFs) (Bar Journal of Cognitive Neuroscience 15: 600–609, 2003). Because of the coarse nature of LSF representations, we hypothesized here that such predictions can accommodate changes in viewpoint as well as facilitate the recognition of visually similar objects. In a repetition-priming task, we indeed observed significant facilitation of target recognition that was primed by LSF objects across moderate viewpoint changes, as well as across visually similar exemplars. These results suggest that the LSF representations are specific enough to activate accurate predictions, yet flexible enough to overcome small changes in visual appearance. Such gist representations facilitate object recognition by accommodating changes in visual appearance due to viewing conditions, and help generalize from familiar to novel exemplars.
Collapse
|
36
|
Ratan Murty NA, Arun SP. Dynamics of 3D view invariance in monkey inferotemporal cortex. J Neurophysiol 2015; 113:2180-94. [PMID: 25609108 PMCID: PMC4416554 DOI: 10.1152/jn.00810.2014] [Citation(s) in RCA: 38] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/14/2014] [Accepted: 01/20/2015] [Indexed: 11/22/2022] Open
Abstract
Rotations in depth are challenging for object vision because features can appear, disappear, be stretched or compressed. Yet we easily recognize objects across views. Are the underlying representations view invariant or dependent? This question has been intensely debated in human vision, but the neuronal representations remain poorly understood. Here, we show that for naturalistic objects, neurons in the monkey inferotemporal (IT) cortex undergo a dynamic transition in time, whereby they are initially sensitive to viewpoint and later encode view-invariant object identity. This transition depended on two aspects of object structure: it was strongest when objects foreshortened strongly across views and were similar to each other. View invariance in IT neurons was present even when objects were reduced to silhouettes, suggesting that it can arise through similarity between external contours of objects across views. Our results elucidate the viewpoint debate by showing that view invariance arises dynamically in IT neurons out of a representation that is initially view dependent.
Collapse
Affiliation(s)
| | - Sripati P Arun
- Centre for Neuroscience, Indian Institute of Science, Bangalore, India
| |
Collapse
|
37
|
Guggenmos M, Thoma V, Cichy RM, Haynes JD, Sterzer P, Richardson-Klavehn A. Non-holistic coding of objects in lateral occipital complex with and without attention. Neuroimage 2014; 107:356-363. [PMID: 25512039 DOI: 10.1016/j.neuroimage.2014.12.013] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/04/2014] [Revised: 10/30/2014] [Accepted: 12/05/2014] [Indexed: 10/24/2022] Open
Abstract
A fundamental issue in visual cognition is whether high-level visual areas code objects in a part-based or a view-based (holistic) format. Previous behavioral and neuroimaging studies that examined the viewpoint invariance of object recognition have yielded ambiguous results, providing evidence for either type of representational format. A critical factor distinguishing the two formats could be the availability of attentional resources, as a number of priming studies have found greater viewpoint invariance for attended compared to unattended objects. It has therefore been suggested that the activation of part-based representations requires attention, whereas the activation of holistic representations occurs automatically irrespective of attention. Using functional magnetic resonance imaging in combination with a novel multivariate pattern analysis approach, the present study probed the format of object representations in human lateral occipital complex and its dependence on attention. We presented human participants with intact and half-split versions of objects that were either attended or unattended. Cross-classifying between intact and split objects, we found that the object-related information coded in activation patterns of intact objects is fully preserved in the patterns of split objects and vice versa. Importantly, the generalization between intact and split objects did not depend on attention. We conclude that lateral occipital complex codes objects in a non-holistic format, both in the presence and absence of attention.
Collapse
Affiliation(s)
- Matthias Guggenmos
- Bernstein Center for Computational Neuroscience, Berlin, Germany; Visual Perception Laboratory, Charité Universitätsmedizin, Berlin, Germany.
| | - Volker Thoma
- School of Psychology, University of East London, London, UK
| | - Radoslaw Martin Cichy
- Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Cambridge, MA, USA
| | | | - Philipp Sterzer
- Bernstein Center for Computational Neuroscience, Berlin, Germany; Visual Perception Laboratory, Charité Universitätsmedizin, Berlin, Germany
| | - Alan Richardson-Klavehn
- Department of Neurology, Otto von Guericke University, Magdeburg, Germany; Leibniz Institute for Neurobiology, Magdeburg, Germany
| |
Collapse
|
38
|
Kolinsky R, Fernandes T. A cultural side effect: learning to read interferes with identity processing of familiar objects. Front Psychol 2014; 5:1224. [PMID: 25400605 PMCID: PMC4215613 DOI: 10.3389/fpsyg.2014.01224] [Citation(s) in RCA: 18] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2014] [Accepted: 10/09/2014] [Indexed: 11/23/2022] Open
Abstract
Based on the neuronal recycling hypothesis (Dehaene and Cohen, 2007), we examined whether reading acquisition has a cost for the recognition of non-linguistic visual materials. More specifically, we checked whether the ability to discriminate between mirror images, which develops through literacy acquisition, interferes with object identity judgments, and whether interference strength varies as a function of the nature of the non-linguistic material. To these aims we presented illiterate, late literate (who learned to read at adult age), and early literate adults with an orientation-independent, identity-based same-different comparison task in which they had to respond “same” to both physically identical and mirrored or plane-rotated images of pictures of familiar objects (Experiment 1) or of geometric shapes (Experiment 2). Interference from irrelevant orientation variations was stronger with plane rotations than with mirror images, and stronger with geometric shapes than with objects. Illiterates were the only participants almost immune to mirror variations, but only for familiar objects. Thus, the process of unlearning mirror-image generalization, necessary to acquire literacy in the Latin alphabet, has a cost for a basic function of the visual ventral object recognition stream, i.e., identification of familiar objects. This demonstrates that neural recycling is not just an adaptation to multi-use but a process of at least partial exaptation.
Collapse
Affiliation(s)
- Régine Kolinsky
- Fonds de la Recherche Scientifique-FNRS Brussels, Belgium ; Unité de Recherche en Neurosciences Cognitives, Center for Research in Cognition and Neurosciences, Université Libre de Bruxelles Brussels, Belgium
| | - Tânia Fernandes
- Faculty of Psychology, Center for Research in Psychology, Universidade de Lisboa Lisboa, Portugal
| |
Collapse
|
39
|
Stevenson RJ. Object concepts in the chemical senses. Cogn Sci 2014; 38:1360-83. [PMID: 24641582 DOI: 10.1111/cogs.12111] [Citation(s) in RCA: 17] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2012] [Revised: 03/27/2013] [Accepted: 06/26/2013] [Indexed: 11/27/2022]
Abstract
This paper examines the applicability of the object concept to the chemical senses, by evaluating them against a set of criteria for object-hood. Taste and chemesthesis do not generate objects. Their parts, perceptible from birth, never combine. Orthonasal olfaction (sniffing) presents a strong case for generating objects. Odorants have many parts yet they are perceived as wholes, this process is based on learning, and there is figure-ground segregation. While flavors are multimodal representations bound together by learning, there is no functional need for flavor objects in the mouth. Rather, food identification occurs prior to ingestion using the eye and nose, with the latter retrieving multimodal flavor objects via sniffing (e.g., sweet smelling caramel). While there are differences in object perception between vision, audition, and orthonasal olfaction, the commonalities suggest that the brain has adopted the same basic solution when faced with extracting meaning from complex stimulus arrays.
Collapse
|
40
|
Abstract
While emotion is a central component of human health and well-being, traditional approaches to understanding its biological function have been wanting. A dynamic systems model, however, broadly redefines and recasts emotion as a primary sensory system-perhaps the first sensory system to have emerged, serving the ancient autopoietic function of "self-regulation." Drawing upon molecular biology and revelations from the field of epigenetics, the model suggests that human emotional perceptions provide an ongoing stream of "self-relevant" sensory information concerning optimally adaptive states between the organism and its immediate environment, along with coupled behavioral corrections that honor a universal self-regulatory logic, one still encoded within cellular signaling and immune functions. Exemplified by the fundamental molecular circuitry of sensorimotor control in the E coli bacterium, the model suggests that the hedonic (affective) categories emerge directly from positive and negative feedback processes, their good/bad binary appraisals relating to dual self-regulatory behavioral regimes-evolutionary purposes, through which organisms actively participate in natural selection, and through which humans can interpret optimal or deficit states of balanced being and becoming. The self-regulatory sensory paradigm transcends anthropomorphism, unites divergent theoretical perspectives and isolated bodies of literature, while challenging time-honored assumptions. While suppressive regulatory strategies abound, it suggests that emotions are better understood as regulating us, providing a service crucial to all semantic language, learning systems, evaluative decision-making, and fundamental to optimal physical, mental, and social health.
Collapse
Affiliation(s)
- Katherine T Peil
- College of Professional Studies, Northeastern University, Boston, Massachusetts; Harvard Divinity School, Cambridge, Massachusetts, United States
| |
Collapse
|
41
|
Collins JA, Curby KM. Conceptual knowledge attenuates viewpoint dependency in visual object recognition. VISUAL COGNITION 2013. [DOI: 10.1080/13506285.2013.836138] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/26/2022]
|
42
|
Rommers J, Meyer AS, Huettig F. Object Shape and Orientation Do Not Routinely Influence Performance During Language Processing. Psychol Sci 2013; 24:2218-25. [DOI: 10.1177/0956797613490746] [Citation(s) in RCA: 21] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
The role of visual representations during language processing remains unclear: They could be activated as a necessary part of the comprehension process, or they could be less crucial and influence performance in a task-dependent manner. In the present experiments, participants read sentences about an object. The sentences implied that the object had a specific shape or orientation. They then either named a picture of that object (Experiments 1 and 3) or decided whether the object had been mentioned in the sentence (Experiment 2). Orientation information did not reliably influence performance in any of the experiments. Shape representations influenced performance most strongly when participants were asked to compare a sentence with a picture or when they were explicitly asked to use mental imagery while reading the sentences. Thus, in contrast to previous claims, implied visual information often does not contribute substantially to the comprehension process during normal reading.
Collapse
Affiliation(s)
- Joost Rommers
- Psychology of Language Department, Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
- International Max Planck Research School for Language Sciences, Nijmegen, The Netherlands
| | - Antje S. Meyer
- Psychology of Language Department, Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
- Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen
| | - Falk Huettig
- Psychology of Language Department, Max Planck Institute for Psycholinguistics, Nijmegen, The Netherlands
- Donders Institute for Brain, Cognition and Behaviour, Radboud University Nijmegen
| |
Collapse
|
43
|
Abstract
Human adults from diverse cultures share intuitions about the points, lines, and figures of Euclidean geometry. Do children develop these intuitions by drawing on phylogenetically ancient and developmentally precocious geometric representations that guide their navigation and their analysis of object shape? In what way might these early-arising representations support later-developing Euclidean intuitions? To approach these questions, we investigated the relations among young children's use of geometry in tasks assessing: navigation; visual form analysis; and the interpretation of symbolic, purely geometric maps. Children's navigation depended on the distance and directional relations of the surface layout and predicted their use of a symbolic map with targets designated by surface distances. In contrast, children's analysis of visual forms depended on the size-invariant shape relations of objects and predicted their use of the same map but with targets designated by corner angles. Even though the two map tasks used identical instructions and map displays, children's performance on these tasks showed no evidence of integrated representations of distance and angle. Instead, young children flexibly recruited geometric representations of either navigable layouts or objects to interpret the same spatial symbols. These findings reveal a link between the early-arising geometric representations that humans share with diverse animals and the flexible geometric intuitions that give rise to human knowledge at its highest reaches. Although young children do not appear to integrate core geometric representations, children's use of the abstract geometry in spatial symbols such as maps may provide the earliest clues to the later construction of Euclidean geometry.
Collapse
Affiliation(s)
- Moira R. Dillon
- Psychology Department, Harvard University, Cambridge, MA 02138; and
| | - Yi Huang
- State Key Laboratory of Cognitive Neuroscience and Learning, Beijing Normal University, Beijing 100875, China
| | | |
Collapse
|
44
|
Baldassi C, Alemi-Neissi A, Pagan M, DiCarlo JJ, Zecchina R, Zoccolan D. Shape similarity, better than semantic membership, accounts for the structure of visual object representations in a population of monkey inferotemporal neurons. PLoS Comput Biol 2013; 9:e1003167. [PMID: 23950700 PMCID: PMC3738466 DOI: 10.1371/journal.pcbi.1003167] [Citation(s) in RCA: 72] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2013] [Accepted: 06/19/2013] [Indexed: 12/02/2022] Open
Abstract
The anterior inferotemporal cortex (IT) is the highest stage along the hierarchy of visual areas that, in primates, processes visual objects. Although several lines of evidence suggest that IT primarily represents visual shape information, some recent studies have argued that neuronal ensembles in IT code the semantic membership of visual objects (i.e., represent conceptual classes such as animate and inanimate objects). In this study, we investigated to what extent semantic, rather than purely visual information, is represented in IT by performing a multivariate analysis of IT responses to a set of visual objects. By relying on a variety of machine-learning approaches (including a cutting-edge clustering algorithm that has been recently developed in the domain of statistical physics), we found that, in most instances, IT representation of visual objects is accounted for by their similarity at the level of shape or, more surprisingly, low-level visual properties. Only in a few cases we observed IT representations of semantic classes that were not explainable by the visual similarity of their members. Overall, these findings reassert the primary function of IT as a conveyor of explicit visual shape information, and reveal that low-level visual properties are represented in IT to a greater extent than previously appreciated. In addition, our work demonstrates how combining a variety of state-of-the-art multivariate approaches, and carefully estimating the contribution of shape similarity to the representation of object categories, can substantially advance our understanding of neuronal coding of visual objects in cortex.
Collapse
Affiliation(s)
- Carlo Baldassi
- Department of Applied Science and Technology & Center for Computational Sciences, Politecnico di Torino, Torino, Italy
- Human Genetics Foundation (HuGeF), Torino, Torino, Italy
| | - Alireza Alemi-Neissi
- Human Genetics Foundation (HuGeF), Torino, Torino, Italy
- International School for Advanced Studies (SISSA), Trieste, Italy
| | - Marino Pagan
- Department of Brain and Cognitive Sciences and McGovern Institute for Brain Research, Massachusetts Institute of Technology (MIT), Cambridge, Massachusetts, United States of America
- Department of Psychology, University of Pennsylvania, Philadelphia, Pennsylvania, United States of America
| | - James J. DiCarlo
- Department of Brain and Cognitive Sciences and McGovern Institute for Brain Research, Massachusetts Institute of Technology (MIT), Cambridge, Massachusetts, United States of America
| | - Riccardo Zecchina
- Department of Applied Science and Technology & Center for Computational Sciences, Politecnico di Torino, Torino, Italy
- Human Genetics Foundation (HuGeF), Torino, Torino, Italy
| | - Davide Zoccolan
- International School for Advanced Studies (SISSA), Trieste, Italy
- Department of Brain and Cognitive Sciences and McGovern Institute for Brain Research, Massachusetts Institute of Technology (MIT), Cambridge, Massachusetts, United States of America
| |
Collapse
|
45
|
Isolating shape from semantics in haptic-visual priming. Exp Brain Res 2013; 227:311-22. [DOI: 10.1007/s00221-013-3489-1] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2012] [Accepted: 03/14/2013] [Indexed: 11/26/2022]
|
46
|
Wakui E, Jüttner M, Petters D, Kaur S, Hummel JE, Davidoff J. Earlier development of analytical than holistic object recognition in adolescence. PLoS One 2013; 8:e61041. [PMID: 23577188 PMCID: PMC3618112 DOI: 10.1371/journal.pone.0061041] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2012] [Accepted: 03/05/2013] [Indexed: 11/19/2022] Open
Abstract
BACKGROUND Previous research has shown that object recognition may develop well into late childhood and adolescence. The present study extends that research and reveals novel differences in holistic and analytic recognition performance in 7-12 year olds compared to that seen in adults. We interpret our data within a hybrid model of object recognition that proposes two parallel routes for recognition (analytic vs. holistic) modulated by attention. METHODOLOGY/PRINCIPAL FINDINGS Using a repetition-priming paradigm, we found in Experiment 1 that children showed no holistic priming, but only analytic priming. Given that holistic priming might be thought to be more 'primitive', we confirmed in Experiment 2 that our surprising finding was not because children's analytic recognition was merely a result of name repetition. CONCLUSIONS/SIGNIFICANCE Our results suggest a developmental primacy of analytic object recognition. By contrast, holistic object recognition skills appear to emerge with a much more protracted trajectory extending into late adolescence.
Collapse
Affiliation(s)
- Elley Wakui
- Department of Psychology, Goldsmiths, University of London, London, United Kingdom
- School of Psychology, University of East London, London, United Kingdom
| | - Martin Jüttner
- Psychology, School of Life and Health Sciences, Aston University, Birmingham, United Kingdom
- * E-mail:
| | - Dean Petters
- Psychology, School of Life and Health Sciences, Aston University, Birmingham, United Kingdom
| | - Surinder Kaur
- Psychology, School of Life and Health Sciences, Aston University, Birmingham, United Kingdom
| | - John E. Hummel
- Department of Psychology, University of Illinois, Urbana, United States of America
| | - Jules Davidoff
- Department of Psychology, Goldsmiths, University of London, London, United Kingdom
| |
Collapse
|
47
|
MacEvoy SP, Yang Z. Joint neuronal tuning for object form and position in the human lateral occipital complex. Neuroimage 2012; 63:1901-8. [DOI: 10.1016/j.neuroimage.2012.08.043] [Citation(s) in RCA: 10] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022] Open
|
48
|
Wi NTN, Loo CK, Chockalingam L. Biologically inspired face recognition: toward pose-invariance. Int J Neural Syst 2012. [PMID: 23186278 DOI: 10.1142/s0129065712500293] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
Abstract
A small change in image will cause a dramatic change in signals. Visual system is required to be able to ignore these changes, yet specific enough to perform recognition. This work intends to provide biological-backed insights into 2D translation and scaling invariance and 3D pose-invariance without imposing strain on memory and with biological justification. The model can be divided into lower and higher visual stages. Lower visual stage models the visual pathway from retina to the striate cortex (V1), whereas the modeling of higher visual stage is mainly based on current psychophysical evidences.
Collapse
Affiliation(s)
- Noel Tay Nuo Wi
- Centre of Diploma Programmes, Multimedia University, JalanAyerKeroh Lama, Melaka, Malaysia.
| | | | | |
Collapse
|
49
|
The time course of activation of object shape and shape+colour representations during memory retrieval. PLoS One 2012; 7:e48550. [PMID: 23155393 PMCID: PMC3498244 DOI: 10.1371/journal.pone.0048550] [Citation(s) in RCA: 14] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2012] [Accepted: 09/26/2012] [Indexed: 11/19/2022] Open
Abstract
Little is known about the timing of activating memory for objects and their associated perceptual properties, such as colour, and yet this is important for theories of human cognition. We investigated the time course associated with early cognitive processes related to the activation of object shape and object shape+colour representations respectively, during memory retrieval as assessed by repetition priming in an event-related potential (ERP) study. The main findings were as follows: (1) we identified a unique early modulation of mean ERP amplitude during the N1 that was associated with the activation of object shape independently of colour; (2) we also found a subsequent early P2 modulation of mean amplitude over the same electrode clusters associated with the activation of object shape+colour representations; (3) these findings were apparent across both familiar (i.e., correctly coloured – yellow banana) and novel (i.e., incorrectly coloured - blue strawberry) objects; and (4) neither of the modulations of mean ERP amplitude were evident during the P3. Together the findings delineate the timing of object shape and colour memory systems and support the notion that perceptual representations of object shape mediate the retrieval of temporary shape+colour representations for familiar and novel objects.
Collapse
|
50
|
Baeck A, Windey I, Op de Beeck HP. The transfer of object learning across exemplars and their orientation is related to perceptual similarity. Vision Res 2012; 68:40-7. [PMID: 22819729 DOI: 10.1016/j.visres.2012.06.023] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/27/2012] [Revised: 06/06/2012] [Accepted: 06/29/2012] [Indexed: 11/19/2022]
Abstract
Recognition of objects improves after training. The exact characteristics of this visual learning process remain unclear. We examined to which extent object learning depends on the exact exemplar and orientation used during training. Participants were trained to name object pictures at as short a picture presentation time as possible. The required presentation time diminished over training. After training participants were tested with a completely new set of objects as well as with two variants of the trained object set, namely an orientation change and a change of the exact exemplar shown. Both manipulations led to a decrease in performance compared to the original picture set. Nevertheless, performance with the manipulated versions of the trained stimuli was better than performance with the completely new set, at least when only one manipulation was performed. Amount of transfer to new images of an object was related to perceptual similarity, but not to pixel overlap or to measurements of similarity in the different layers of a popular hierarchical object recognition model (HMAX). Thus, object learning generalizes only partially over changes in exemplars and orientation, which is consistent with the tuning properties of neurons in object-selective cortical regions and the role of perceptual similarity in these representations.
Collapse
Affiliation(s)
- Annelies Baeck
- Laboratory of Biological Psychology, University of Leuven (KU Leuven), Tiensestraat 102, 3000 Leuven, Belgium.
| | | | | |
Collapse
|