1
|
An NM, Roh H, Kim S, Kim JH, Im M. Machine Learning Techniques for Simulating Human Psychophysical Testing of Low-Resolution Phosphene Face Images in Artificial Vision. ADVANCED SCIENCE (WEINHEIM, BADEN-WURTTEMBERG, GERMANY) 2025; 12:e2405789. [PMID: 39985243 PMCID: PMC12005743 DOI: 10.1002/advs.202405789] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/27/2024] [Revised: 01/18/2025] [Indexed: 02/24/2025]
Abstract
To evaluate the quality of artificial visual percepts generated by emerging methodologies, researchers often rely on labor-intensive and tedious human psychophysical experiments. These experiments necessitate repeated iterations upon any major/minor modifications in the hardware/software configurations. Here, the capacity of standard machine learning (ML) models is investigated to accurately replicate quaternary match-to-sample tasks using low-resolution facial images represented by arrays of phosphenes as input stimuli. Initially, the performance of the ML models trained to approximate innate human facial recognition abilities across a dataset comprising 3600 phosphene images of human faces is analyzed. Subsequently, due to the time constraints and the potential for subject fatigue, the psychophysical test is limited to presenting only 720 low-resolution phosphene images to 36 human subjects. Notably, the superior model adeptly mirrors the behavioral trend of human subjects, offering precise predictions for 8 out of 9 phosphene quality levels on the overlapping test queries. Subsequently, human recognition performances for untested phosphene images are predicted, streamlining the process and minimizing the need for additional psychophysical tests. The findings underscore the transformative potential of ML in reshaping the research paradigm of visual prosthetics, facilitating the expedited advancement of prostheses.
Collapse
Affiliation(s)
- Na Min An
- Brain Science InstituteKorea Institute of Science and Technology (KIST)Seoul02792Republic of Korea
- Present address:
Kim Jaechul Graduate School of AIKAISTSeoul02455Republic of Korea
| | - Hyeonhee Roh
- Brain Science InstituteKorea Institute of Science and Technology (KIST)Seoul02792Republic of Korea
| | - Sein Kim
- Brain Science InstituteKorea Institute of Science and Technology (KIST)Seoul02792Republic of Korea
| | - Jae Hun Kim
- Brain Science InstituteKorea Institute of Science and Technology (KIST)Seoul02792Republic of Korea
- Sensor System Research CenterAdvanced Materials and Systems Research DivisionKISTSeoul02792Republic of Korea
| | - Maesoon Im
- Brain Science InstituteKorea Institute of Science and Technology (KIST)Seoul02792Republic of Korea
- Division of Bio‐Medical Science and TechnologyUniversity of Science and Technology (UST)Seoul02792Republic of Korea
- KHU‐KIST Department of Converging Science and TechnologyKyung Hee UniversitySeoul02447Republic of Korea
| |
Collapse
|
2
|
Nejad A, Küçükoǧlu B, de Ruyter van Steveninck J, Bedrossian S, Heutink J, de Haan GA, Cornelissen FW, van Gerven M. Point-SPV: end-to-end enhancement of object recognition in simulated prosthetic vision using synthetic viewing points. Front Hum Neurosci 2025; 19:1549698. [PMID: 40196449 PMCID: PMC11973266 DOI: 10.3389/fnhum.2025.1549698] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2024] [Accepted: 02/24/2025] [Indexed: 04/09/2025] Open
Abstract
Prosthetic vision systems aim to restore functional sight for visually impaired individuals by replicating visual perception by inducing phosphenes through electrical stimulation in the visual cortex, yet there remain challenges in visual representation strategies such as including gaze information and task-dependent optimization. In this paper, we introduce Point-SPV, an end-to-end deep learning model designed to enhance object recognition in simulated prosthetic vision. Point-SPV takes an initial step toward gaze-based optimization by simulating viewing points, representing potential gaze locations, and training the model on patches surrounding these points. Our approach prioritizes task-oriented representation, aligning visual outputs with object recognition needs. A behavioral gaze-contingent object discrimination experiment demonstrated that Point-SPV outperformed a conventional edge detection method, by facilitating observers to gain a higher recognition accuracy, faster reaction times, and a more efficient visual exploration. Our work highlights how task-specific optimization may enhance representations in prosthetic vision, offering a foundation for future exploration and application.
Collapse
Affiliation(s)
- Ashkan Nejad
- Department of Research and Improvement of Care, Royal Dutch Visio, Huizen, Netherlands
- Department of Machine Learning and Neural Computing, Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, Netherlands
- Laboratory for Experimental Ophthalmology, University Medical Center Groningen, University of Groningen, Groningen, Netherlands
| | - Burcu Küçükoǧlu
- Department of Machine Learning and Neural Computing, Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, Netherlands
| | - Jaap de Ruyter van Steveninck
- Department of Machine Learning and Neural Computing, Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, Netherlands
| | - Sandra Bedrossian
- Faculty of Science and Engineering, University of Groningen, Groningen, Netherlands
| | - Joost Heutink
- Department of Research and Improvement of Care, Royal Dutch Visio, Huizen, Netherlands
- Department of Clinical and Developmental Neuropsychology, University of Groningen, Groningen, Netherlands
| | - Gera A. de Haan
- Department of Research and Improvement of Care, Royal Dutch Visio, Huizen, Netherlands
- Department of Clinical and Developmental Neuropsychology, University of Groningen, Groningen, Netherlands
| | - Frans W. Cornelissen
- Laboratory for Experimental Ophthalmology, University Medical Center Groningen, University of Groningen, Groningen, Netherlands
| | - Marcel van Gerven
- Department of Machine Learning and Neural Computing, Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, Netherlands
| |
Collapse
|
3
|
van Hoof R, Lozano A, Wang F, Klink PC, Roelfsema PR, Goebel R. Optimal placement of high-channel visual prostheses in human retinotopic visual cortex. J Neural Eng 2025; 22:026016. [PMID: 39870040 DOI: 10.1088/1741-2552/adaeef] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/28/2024] [Accepted: 01/27/2025] [Indexed: 01/29/2025]
Abstract
Objective.Recent strides in neurotechnology show potential to restore vision in individuals with visual impairments due to early visual pathway damage. As neuroprostheses mature and become available to a larger population, manual placement and evaluation of electrode designs become costly and impractical. An automatic method to simulate and optimize the implantation process of electrode arrays at large-scale is currently lacking.Approach.Here, we present a comprehensive method to automatically optimize electrode placement for visual prostheses, with the objective of matching predefined phosphene distributions. Our approach makes use of retinotopic predictions combined with individual anatomy data to minimize discrepancies between simulated and target phosphene patterns. While demonstrated with a 1000-channel 3D electrode array in V1, our simulation pipeline is versatile, potentially accommodating any electrode design and allowing for design evaluation.Main results.Notably, our results show that individually optimized placements in 362 brain hemispheres outperform average brain solutions, underscoring the significance of anatomical specificity. We further show how virtual implantation of multiple individual brains highlights the challenges of achieving full visual field coverage owing to single electrode constraints, which may be overcome by introducing multiple arrays of electrodes. Including additional surgical considerations, such as intracranial vasculature, in future iterations could refine the optimization process.Significance.Our open-source software streamlines the refinement of surgical procedures and facilitates simulation studies, offering a realistic exploration of electrode configuration possibilities.
Collapse
Affiliation(s)
- Rick van Hoof
- Department of Cognitive Neuroscience, Maastricht University, Maastricht, The Netherlands
| | - Antonio Lozano
- Department of Vision and Cognition, Netherlands Institute for Neuroscience, Amsterdam, The Netherlands
| | - Feng Wang
- Department of Vision and Cognition, Netherlands Institute for Neuroscience, Amsterdam, The Netherlands
| | - P Christiaan Klink
- Department of Vision and Cognition, Netherlands Institute for Neuroscience, Amsterdam, The Netherlands
- Experimental Psychology, Helmholtz Institute, Utrecht University, Utrecht, The Netherlands
| | - Pieter R Roelfsema
- Department of Vision and Cognition, Netherlands Institute for Neuroscience, Amsterdam, The Netherlands
- Department of Integrative Neurophysiology, VU University, Amsterdam, The Netherlands
- Department of Psychiatry, Academic Medical Centre, Amsterdam, The Netherlands
- Laboratory of Visual Brain Therapy, Sorbonne Université, INSERM, CNRS, Paris, France
| | - Rainer Goebel
- Department of Cognitive Neuroscience, Maastricht University, Maastricht, The Netherlands
- Maastricht Brain Imaging Centre, Maastricht University, Maastricht, The Netherlands
| |
Collapse
|
4
|
de Ruyter van Steveninck J, Nipshagen M, van Gerven M, Güçlü U, Güçlüturk Y, van Wezel R. Gaze-contingent processing improves mobility, scene recognition and visual search in simulated head-steered prosthetic vision. J Neural Eng 2024; 21:026037. [PMID: 38502957 DOI: 10.1088/1741-2552/ad357d] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2023] [Accepted: 03/19/2024] [Indexed: 03/21/2024]
Abstract
Objective.The enabling technology of visual prosthetics for the blind is making rapid progress. However, there are still uncertainties regarding the functional outcomes, which can depend on many design choices in the development. In visual prostheses with a head-mounted camera, a particularly challenging question is how to deal with the gaze-locked visual percept associated with spatial updating conflicts in the brain. The current study investigates a recently proposed compensation strategy based on gaze-contingent image processing with eye-tracking. Gaze-contingent processing is expected to reinforce natural-like visual scanning and reestablished spatial updating based on eye movements. The beneficial effects remain to be investigated for daily life activities in complex visual environments.Approach.The current study evaluates the benefits of gaze-contingent processing versus gaze-locked and gaze-ignored simulations in the context of mobility, scene recognition and visual search, using a virtual reality simulated prosthetic vision paradigm with sighted subjects.Main results.Compared to gaze-locked vision, gaze-contingent processing was consistently found to improve the speed in all experimental tasks, as well as the subjective quality of vision. Similar or further improvements were found in a control condition that ignores gaze-dependent effects, a simulation that is unattainable in the clinical reality.Significance.Our results suggest that gaze-locked vision and spatial updating conflicts can be debilitating for complex visually-guided activities of daily living such as mobility and orientation. Therefore, for prospective users of head-steered prostheses with an unimpaired oculomotor system, the inclusion of a compensatory eye-tracking system is strongly endorsed.
Collapse
Affiliation(s)
| | - Mo Nipshagen
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, The Netherlands
| | - Marcel van Gerven
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, The Netherlands
| | - Umut Güçlü
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, The Netherlands
| | - Yağmur Güçlüturk
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, The Netherlands
| | - Richard van Wezel
- Donders Institute for Brain, Cognition and Behaviour, Radboud University, Nijmegen, The Netherlands
- Biomedical Signals and Systems Group, University of Twente, Enschede, The Netherlands
| |
Collapse
|
5
|
Granley J, Fauvel T, Chalk M, Beyeler M. Human-in-the-Loop Optimization for Deep Stimulus Encoding in Visual Prostheses. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 2023; 36:79376-79398. [PMID: 38984104 PMCID: PMC11232484] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Subscribe] [Scholar Register] [Indexed: 07/11/2024]
Abstract
Neuroprostheses show potential in restoring lost sensory function and enhancing human capabilities, but the sensations produced by current devices often seem unnatural or distorted. Exact placement of implants and differences in individual perception lead to significant variations in stimulus response, making personalized stimulus optimization a key challenge. Bayesian optimization could be used to optimize patient-specific stimulation parameters with limited noisy observations, but is not feasible for high-dimensional stimuli. Alternatively, deep learning models can optimize stimulus encoding strategies, but typically assume perfect knowledge of patient-specific variations. Here we propose a novel, practically feasible approach that overcomes both of these fundamental limitations. First, a deep encoder network is trained to produce optimal stimuli for any individual patient by inverting a forward model mapping electrical stimuli to visual percepts. Second, a preferential Bayesian optimization strategy utilizes this encoder to optimize patient-specific parameters for a new patient, using a minimal number of pairwise comparisons between candidate stimuli. We demonstrate the viability of this approach on a novel, state-of-the-art visual prosthesis model. We show that our approach quickly learns a personalized stimulus encoder, leads to dramatic improvements in the quality of restored vision, and is robust to noisy patient feedback and misspecifications in the underlying forward model. Overall, our results suggest that combining the strengths of deep learning and Bayesian optimization could significantly improve the perceptual experience of patients fitted with visual prostheses and may prove a viable solution for a range of neuroprosthetic technologies.
Collapse
Affiliation(s)
- Jacob Granley
- Department of Computer Science, University of California, Santa Barbara
| | - Tristan Fauvel
- Institut de la Vision, Sorbonne Université, 17 rue Moreau, F-75012 Paris, France, Now with Quinten Health
| | - Matthew Chalk
- Institut de la Vision, Sorbonne Université, 17 rue Moreau, F-75012 Paris, France
| | - Michael Beyeler
- Department of Computer Science, Department of Psychological & Brain Sciences, University of California, Santa Barbara
| |
Collapse
|