1
|
Ioannucci S, Vetter P. Semantic audio-visual congruence modulates visual sensitivity to biological motion across awareness levels. Cognition 2025; 262:106181. [PMID: 40378502 DOI: 10.1016/j.cognition.2025.106181] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/06/2025] [Revised: 05/06/2025] [Accepted: 05/07/2025] [Indexed: 05/19/2025]
Abstract
Whether cross-modal interaction requires conscious awareness of multisensory information or whether it can occur in the absence of awareness, is still an open question. Here, we investigated if sounds can enhance detection sensitivity of semantically matching visual stimuli at varying levels of visual awareness. We presented biological motion stimuli of human actions (walking, rowing, sawing) during dynamic continuous flash suppression (CFS) to 80 participants and measured the effect of co-occurring, semantically matching or non-matching action sounds on visual sensitivity (d'). By individually thresholding stimulus contrast, we distinguished participants who detected motion either above or at chance level. Participants who reliably detected visual motion above chance showed higher sensitivity to upright versus inverted biological motion across all experimental conditions. In contrast, participants detecting visual motion at chance level, i.e. during successful suppression, demonstrated this upright advantage exclusively during trials with semantically congruent sounds. Across the whole sample, the impact of sounds on visual sensitivity increased as participants' visual detection performance decreased, revealing a systematic trade-off between auditory and visual processing. Our findings suggest that semantic congruence between auditory and visual information can selectively modulate biological motion perception when visual awareness is minimal or absent, while more robust visual signals enable perception of biological motion independent of auditory input. Thus, semantically congruent sounds may impact visual representations as a function of the level of visual awareness.
Collapse
Affiliation(s)
- Stefano Ioannucci
- Visual and Cognitive Neuroscience Lab, Dept. of Psychology, University of Fribourg, Switzerland.
| | - Petra Vetter
- Visual and Cognitive Neuroscience Lab, Dept. of Psychology, University of Fribourg, Switzerland
| |
Collapse
|
2
|
Yang H, Huang W, Wen W, Long S, Zhang Y, Chi X, Luo D. The impact of frequency-specific music stimulation on consciousness in patients with disorders of consciousness. Front Neurol 2025; 16:1506261. [PMID: 40070671 PMCID: PMC11893416 DOI: 10.3389/fneur.2025.1506261] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/04/2024] [Accepted: 02/05/2025] [Indexed: 03/14/2025] Open
Abstract
Objective This study aimed to evaluate the effects of frequency-specific music stimulation on the awareness and brain connectivity in patients with disorders of consciousness (DOC). Methods A total of 25 DOC patients were exposed to auditory stimulation through music at varying frequencies (low: <0.5 Hz, middle: 0.5 Hz-3.5 kHz, high: >3.5 kHz). Brain responses were assessed using Functional Near-Infrared Spectroscopy (fNIRS) to monitor objective markers of brain activity. The analysis focused on effective connectivity and coupling strength (CS) values in response to different frequency stimulations, targeting regions such as the motor and somatosensory cortices. Results The mean age of the patients was 49.4 years, with an average coma duration of 1.96 months. While no significant differences were observed in general brain arousal across different frequency stimuli, notable differences in effective connectivity were identified. High-frequency stimulation resulted in significantly higher CS values in the right primary motor cortex (p < 0.05), while middle-frequency stimulation showed significant effects in the right primary somatosensory cortex (p = 0.016). Conclusion The findings suggest that middle- and high-frequency music stimulation may enhance effective connectivity in specific brain regions, potentially contributing to the rehabilitation of DOC patients. These results indicate that frequency-specific music could stimulate motor networks and areas associated with autobiographical memory, highlighting its therapeutic potential in promoting awareness in this patient population.
Collapse
Affiliation(s)
- Haitao Yang
- The First Affiliated Hospital of Guangzhou University of Chinese Medicine, Guangzhou, China
| | - Wenhao Huang
- The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Wei Wen
- Guangzhou University of Chinese Medicine, Guangzhou, China
| | - Shoufeng Long
- The Second Clinical College of Guangzhou University of Chinese Medicine, Guangzhou, China
| | - Yujie Zhang
- Guangzhou University of Chinese Medicine, Guangzhou, China
| | - Xiangfeng Chi
- Guangdong No. 2 Hospital of Traditional Chinese Medicine, Guangzhou, China
| | - Daihong Luo
- Beijing University of Chinese Medicine Shenzhen Hospital (Longgang), Shenzhen, China
| |
Collapse
|
3
|
Lunn J, Berggren N, Ward J, Forster S. Irrelevant sights and sounds require spatial suppression: ERP evidence. Psychophysiology 2023; 60:e14181. [PMID: 36114739 PMCID: PMC10078104 DOI: 10.1111/psyp.14181] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2021] [Revised: 06/28/2022] [Accepted: 08/19/2022] [Indexed: 01/27/2023]
Abstract
Both real-world experience and behavioral laboratory research suggest that entirely irrelevant stimuli (distractors) can interfere with a primary task. However, it is as yet unknown whether such interference reflects competition for spatial attention - indeed, prominent theories of attention predict that this should not be the case. Whilst electrophysiological indices of spatial capture and spatial suppression have been well-investigated, experiments have primarily utilized distractors which share a degree of task-relevance with targets, and are limited to the visual domain. The present research measured behavioral and ERP responses to test the ability of salient yet entirely task-irrelevant visual and auditory distractors to compete for spatial attention during a visual task, while also testing for potentially enhanced competition from multisensory distractors. Participants completed a central letter search task, while ignoring lateralized visual (e.g., image of a dog), auditory (e.g., barking), or multisensory (e.g., image + barking) distractors. Results showed that visual and multisensory distractors elicited a PD component indicative of active lateralized suppression. We also establish for the first time an auditory analog of the PD component, the PAD , elicited by auditory and multisensory distractors. Interestingly, there was no evidence to suggest enhanced ability of multisensory distractors to compete for attentional selection, despite previous proposals of a "special" saliency status for such items. Our findings hence suggest that irrelevant multisensory and unisensory distractors are similarly capable of eliciting a spatial "attend-to-me" signal - a precursor of spatial attentional capture - but at least in the present data set did not elicit full spatial attentional capture.
Collapse
Affiliation(s)
- Jessica Lunn
- School of Psychology and Sussex Neuroscience, University of Sussex, Brighton, UK
| | - Nick Berggren
- Department of Psychological Sciences, Birkbeck, University of London, London, UK
| | - Jamie Ward
- School of Psychology and Sussex Neuroscience, University of Sussex, Brighton, UK
| | - Sophie Forster
- School of Psychology and Sussex Neuroscience, University of Sussex, Brighton, UK
| |
Collapse
|
4
|
Magnetoencephalography recordings reveal the neural mechanisms of auditory contributions to improved visual detection. Commun Biol 2023; 6:12. [PMID: 36604455 PMCID: PMC9816120 DOI: 10.1038/s42003-022-04335-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/02/2022] [Accepted: 12/01/2022] [Indexed: 01/07/2023] Open
Abstract
Sounds enhance the detection of visual stimuli while concurrently biasing an observer's decisions. To investigate the neural mechanisms that underlie such multisensory interactions, we decoded time-resolved Signal Detection Theory sensitivity and criterion parameters from magneto-encephalographic recordings of participants that performed a visual detection task. We found that sounds improved visual detection sensitivity by enhancing the accumulation and maintenance of perceptual evidence over time. Meanwhile, criterion decoding analyses revealed that sounds induced brain activity patterns that resembled the patterns evoked by an actual visual stimulus. These two complementary mechanisms of audiovisual interplay differed in terms of their automaticity: Whereas the sound-induced enhancement in visual sensitivity depended on participants being actively engaged in a detection task, we found that sounds activated the visual cortex irrespective of task demands, potentially inducing visual illusory percepts. These results challenge the classical assumption that sound-induced increases in false alarms exclusively correspond to decision-level biases.
Collapse
|
5
|
Ono M, Hirose N, Mori S. Tactile information affects alternating visual percepts during binocular rivalry using naturalistic objects. Cogn Res Princ Implic 2022; 7:40. [PMID: 35543826 PMCID: PMC9095789 DOI: 10.1186/s41235-022-00390-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/14/2020] [Accepted: 04/17/2022] [Indexed: 12/14/2022] Open
Abstract
INTRODUCTION Past studies have provided evidence that the effects of tactile stimulation on binocular rivalry are mediated by primitive features (orientation and spatial frequency) common in vision and touch. In this study, we examined whether such effects on binocular rivalry can be obtained through the roughness of naturalistic objects. In three experiments, the total dominant time of visual percepts of two objects was measured under binocular rivalry when participants touched one of the objects. RESULT In Experiment 1, the total dominant time for the image of artificial turf and bathmat was prolonged by congruent tactile stimulation and shortened by incongruent tactile stimulation. In Experiment 2, we used the same stimuli but rotated their visual images in opposite directions. The dominant time for either image was prolonged by congruent tactile stimulation. In Experiment 3, we used different types of stimuli, smooth marble and rough fabric, and noted significant effects of the congruent and incongruent tactile stimulation on the dominant time of visual percepts. CONCLUSION These three experiments demonstrated that visuo-tactile interaction on binocular rivalry can be mediated by roughness.
Collapse
Affiliation(s)
- Mikoto Ono
- grid.177174.30000 0001 2242 4849Department of Informatics, Graduate school of Information Science and Electrical Engineering, Kyushu University, 744 Motooka, Nishi-ku, Fukuoka City, Fukuoka 819-0395 Japan
| | - Nobuyuki Hirose
- grid.177174.30000 0001 2242 4849Department of Informatics, Graduate school of Information Science and Electrical Engineering, Kyushu University, 744 Motooka, Nishi-ku, Fukuoka City, Fukuoka 819-0395 Japan
| | - Shuji Mori
- grid.177174.30000 0001 2242 4849Department of Informatics, Graduate school of Information Science and Electrical Engineering, Kyushu University, 744 Motooka, Nishi-ku, Fukuoka City, Fukuoka 819-0395 Japan
| |
Collapse
|
6
|
Filimonov D, Railo H, Revonsuo A, Koivisto M. Modality-specific and modality-general electrophysiological correlates of visual and auditory awareness: Evidence from a bimodal ERP experiment. Neuropsychologia 2022; 166:108154. [PMID: 35016890 DOI: 10.1016/j.neuropsychologia.2022.108154] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2021] [Revised: 11/03/2021] [Accepted: 01/07/2022] [Indexed: 10/19/2022]
Abstract
To date, most studies on the event-related potential (ERP) correlates of conscious perception have examined a single perceptual modality. We compared electrophysiological correlates of visual and auditory awareness in the same experiment to test whether there are modality-specific and modality-general correlates of conscious perception. We used near threshold stimulation and analyzed event-related potentials in response to aware and unaware trials in visual, auditory and bimodal conditions. The results showed modality-specific negative amplitude correlates of conscious perception between 200 and 300 ms after stimulus onset. A combination of these auditory and visual awareness negativities was observed in the bimodal condition. A later positive amplitude difference, whose early part was modality-specific, possibly reflecting access to global workspace, and later part shared modality-general features, possibly indicating higher level cognitive processing involving the decision making, was also observed.
Collapse
Affiliation(s)
| | - Henry Railo
- Turku Brain and Mind Centre, University of Turku, Finland; Department of Clinical Neurophysiology, University of Turku, Turku, Finland
| | - Antti Revonsuo
- Department of Psychology, University of Turku, Finland; Turku Brain and Mind Centre, University of Turku, Finland; Division of Cognitive Neuroscience and Philosophy, University of Skövde, Sweden
| | - Mika Koivisto
- Department of Psychology, University of Turku, Finland; Turku Brain and Mind Centre, University of Turku, Finland
| |
Collapse
|
7
|
Liang J, Li Y, Zhang Z, Luo W. Sound gaps boost emotional audiovisual integration independent of attention: Evidence from an ERP study. Biol Psychol 2021; 168:108246. [PMID: 34968556 DOI: 10.1016/j.biopsycho.2021.108246] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/16/2021] [Revised: 12/18/2021] [Accepted: 12/23/2021] [Indexed: 11/02/2022]
Abstract
The emotion discrimination paradigm was adopted to study the effect of interrupted sound on visual emotional processing under different attentional states. There were two experiments: Experiment 1: judging facial expressions (explicit task), Experiment 2: judging the position of a bar (implicit task). In Experiment 1, ERP results showed that there was a sound gap accelerating the effect of P1 present only under neutral faces. In Experiment 2, the accelerating effect (P1) existed regardless of the emotional condition. Combining two experiments, P1 findings suggest that sound gap enhances bottom-up attention. The N170 and late positive component (LPC) were found to be regulated by emotion face in both experiments, with fear over the neutral. Comparing the two experiments, the explicit task induced a larger LPC than the implicit task. Overall, sound gaps boosted the audiovisual integration by bottom-up attention in early integration, while cognitive expectations led to top-down attention in late stages.
Collapse
Affiliation(s)
- Junyu Liang
- Research Center of Brain and Cognitive Neuroscience, Liaoning Normal University, Dalian 116029, China; Key Laboratory of Brain and Cognitive Neuroscience, Dalian 116029, Liaoning Province, China
| | - Yuchen Li
- Research Center of Brain and Cognitive Neuroscience, Liaoning Normal University, Dalian 116029, China; Key Laboratory of Brain and Cognitive Neuroscience, Dalian 116029, Liaoning Province, China
| | - Zhao Zhang
- Research Center of Brain and Cognitive Neuroscience, Liaoning Normal University, Dalian 116029, China; Institute of Psychology, Weifang Medical University, Weifang 216053, China; Key Laboratory of Brain and Cognitive Neuroscience, Dalian 116029, Liaoning Province, China
| | - Wenbo Luo
- Research Center of Brain and Cognitive Neuroscience, Liaoning Normal University, Dalian 116029, China; Key Laboratory of Brain and Cognitive Neuroscience, Dalian 116029, Liaoning Province, China.
| |
Collapse
|
8
|
Abstract
Pieron’s and Chocholle’s seminal psychophysical work predicts that human response time to information relative to visual contrast and/or sound frequency decreases when contrast intensity or sound frequency increases. The goal of this study is to bring to the forefront the ability of individuals to use visual contrast intensity and sound frequency in combination for faster perceptual decisions of relative depth (“nearer”) in planar (2D) object configurations based on physical variations in luminance contrast. Computer controlled images with two abstract patterns of varying contrast intensity, one on the left and one on the right, preceded or not by a pure tone of varying frequency, were shown to healthy young humans in controlled experimental sequences. Their task (two-alternative, forced-choice) was to decide as quickly as possible which of two patterns, the left or the right one, in a given image appeared to “stand out as if it were nearer” in terms of apparent (subjective) visual depth. The results showed that the combinations of varying relative visual contrast with sounds of varying frequency exploited here produced an additive effect on choice response times in terms of facilitation, where a stronger visual contrast combined with a higher sound frequency produced shorter forced-choice response times. This new effect is predicted by audio-visual probability summation.
Collapse
|
9
|
Kim S, Kim J. Effects of Multimodal Association on Ambiguous Perception in Binocular Rivalry. Perception 2019; 48:796-819. [DOI: 10.1177/0301006619867023] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
When two eyes view dissimilar images, an observer typically reports ambiguous perception called binocular rivalry where the subjective perception fluctuates between the two inputs. This perceptual instability is often comprised of exclusive dominance of each image and a transition state called piecemeal state where the two images are intermingled in patchwork manner. Herein, we investigated the effects of multimodal association of sensory congruent pair, arbitrary pair, and reverse pair on piecemeal state in order to see how each level of association affects the ambiguous perception during binocular rivalry. To induce the multisensory associations, we designed a matching task with audiovisual feedback where subjects were required to respond according to given pairing rules. We found that explicit audiovisual associations can substantially affect the piecemeal state during binocular rivalry and that this congruency effect that reduces the amount of visual ambiguity originates primarily from explicit audiovisual association training rather than common sensory features. Furthermore, when one information is associated with multiple information, recent and preexisting associations work collectively to influence the perceptual ambiguity during rivalry. Our findings show that learned multimodal association directly affects the temporal dynamics of ambiguous perception during binocular rivalry by modulating not only the exclusive dominance but also the piecemeal state in a systematic manner.
Collapse
Affiliation(s)
- Sungyong Kim
- Graduate School of Culture Technology, Korea Advanced Institute of Science and Technology, Daejeon, Republic of Korea
| | - Jeounghoon Kim
- Graduate School of Culture Technology, Korea Advanced Institute of Science and Technology, Daejeon, Republic of Korea; School of Humanities and Social Sciences, Korea Advanced Institute of Science and Technology, Daejeon, Republic of Korea
| |
Collapse
|
10
|
Lunn J, Sjoblom A, Ward J, Soto-Faraco S, Forster S. Multisensory enhancement of attention depends on whether you are already paying attention. Cognition 2019; 187:38-49. [PMID: 30825813 DOI: 10.1016/j.cognition.2019.02.008] [Citation(s) in RCA: 26] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2018] [Revised: 02/12/2019] [Accepted: 02/13/2019] [Indexed: 10/27/2022]
Abstract
Multisensory stimuli are argued to capture attention more effectively than unisensory stimuli due to their ability to elicit a super-additive neuronal response. However, behavioural evidence for enhanced multisensory attentional capture is mixed. Furthermore, the notion of multisensory enhancement of attention conflicts with findings suggesting that multisensory integration may itself be dependent upon top-down attention. The present research resolves this discrepancy by examining how both endogenous attentional settings and the availability of attentional capacity modulate capture by multisensory stimuli. Across a series of four studies, two measures of attentional capture were used which vary in their reliance on endogenous attention: facilitation and distraction. Perceptual load was additionally manipulated to determine whether multisensory stimuli are still able to capture attention when attention is occupied by a demanding primary task. Multisensory stimuli presented as search targets were consistently detected faster than unisensory stimuli regardless of perceptual load, although they are nevertheless subject to load modulation. In contrast, task irrelevant multisensory stimuli did not cause greater distraction than unisensory stimuli, suggesting that the enhanced attentional status of multisensory stimuli may be mediated by the availability of endogenous attention. Implications for multisensory alerts in practical settings such as driving and aviation are discussed, namely that these may be advantageous during demanding tasks, but may be less suitable to signaling unexpected events.
Collapse
|
11
|
Sánchez-García C, Kandel S, Savariaux C, Soto-Faraco S. The Time Course of Audio-Visual Phoneme Identification: a High Temporal Resolution Study. Multisens Res 2018; 31:57-78. [DOI: 10.1163/22134808-00002560] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/07/2016] [Accepted: 02/20/2017] [Indexed: 11/19/2022]
Abstract
Speech unfolds in time and, as a consequence, its perception requires temporal integration. Yet, studies addressing audio-visual speech processing have often overlooked this temporal aspect. Here, we address the temporal course of audio-visual speech processing in a phoneme identification task using a Gating paradigm. We created disyllabic Spanish word-like utterances (e.g., /pafa/, /paθa/, …) from high-speed camera recordings. The stimuli differed only in the middle consonant (/f/, /θ/, /s/, /r/, /g/), which varied in visual and auditory saliency. As in classical Gating tasks, the utterances were presented in fragments of increasing length (gates), here in 10 ms steps, for identification and confidence ratings. We measured correct identification as a function of time (at each gate) for each critical consonant in audio, visual and audio-visual conditions, and computed the Identification Point and Recognition Point scores. The results revealed that audio-visual identification is a time-varying process that depends on the relative strength of each modality (i.e., saliency). In some cases, audio-visual identification followed the pattern of one dominant modality (either A or V), when that modality was very salient. In other cases, both modalities contributed to identification, hence resulting in audio-visual advantage or interference with respect to unimodal conditions. Both unimodal dominance and audio-visual interaction patterns may arise within the course of identification of the same utterance, at different times. The outcome of this study suggests that audio-visual speech integration models should take into account the time-varying nature of visual and auditory saliency.
Collapse
Affiliation(s)
- Carolina Sánchez-García
- Departament de Tecnologies de la Informació i les Comunicacions, Universitat Pompeu Fabra, Barcelona, Spain
| | - Sonia Kandel
- Université Grenoble Alpes, GIPSA-lab (CNRS UMR 5216), Grenoble, France
| | | | - Salvador Soto-Faraco
- Departament de Tecnologies de la Informació i les Comunicacions, Universitat Pompeu Fabra, Barcelona, Spain
- Institució Catalana de Recerca i Estudis Avançats (ICREA), Barcelona, Spain
| |
Collapse
|