1
|
Newport RA, Russo C, Liu S, Suman AA, Di Ieva A. SoftMatch: Comparing Scanpaths Using Combinatorial Spatio-Temporal Sequences with Fractal Curves. SENSORS (BASEL, SWITZERLAND) 2022; 22:7438. [PMID: 36236535 PMCID: PMC9570610 DOI: 10.3390/s22197438] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/01/2022] [Revised: 09/24/2022] [Accepted: 09/26/2022] [Indexed: 06/16/2023]
Abstract
Recent studies matching eye gaze patterns with those of others contain research that is heavily reliant on string editing methods borrowed from early work in bioinformatics. Previous studies have shown string editing methods to be susceptible to false negative results when matching mutated genes or unordered regions of interest in scanpaths. Even as new methods have emerged for matching amino acids using novel combinatorial techniques, scanpath matching is still limited by a traditional collinear approach. This approach reduces the ability to discriminate between free viewing scanpaths of two people looking at the same stimulus due to the heavy weight placed on linearity. To overcome this limitation, we here introduce a new method called SoftMatch to compare pairs of scanpaths. SoftMatch diverges from traditional scanpath matching in two different ways: firstly, by preserving locality using fractal curves to reduce dimensionality from 2D Cartesian (x,y) coordinates into 1D (h) Hilbert distances, and secondly by taking a combinatorial approach to fixation matching using discrete Fréchet distance measurements between segments of scanpath fixation sequences. These matching "sequences of fixations over time" are a loose acronym for SoftMatch. Results indicate high degrees of statistical and substantive significance when scoring matches between scanpaths made during free-form viewing of unfamiliar stimuli. Applications of this method can be used to better understand bottom up perceptual processes extending to scanpath outlier detection, expertise analysis, pathological screening, and salience prediction.
Collapse
Affiliation(s)
- Robert Ahadizad Newport
- Faculty of Medicine, Health and Human Sciences, Macquarie Medical School, Macquarie University, Balaclava Road, Sydney, NSW 2109, Australia
- Computational NeuroSurgery (CNS) Lab, Macquarie Medical School, Macquarie University, Balaclava Road, Sydney, NSW 2109, Australia
| | - Carlo Russo
- Computational NeuroSurgery (CNS) Lab, Macquarie Medical School, Macquarie University, Balaclava Road, Sydney, NSW 2109, Australia
| | - Sidong Liu
- Faculty of Medicine, Health and Human Sciences, Macquarie Medical School, Macquarie University, Balaclava Road, Sydney, NSW 2109, Australia
- Computational NeuroSurgery (CNS) Lab, Macquarie Medical School, Macquarie University, Balaclava Road, Sydney, NSW 2109, Australia
| | - Abdulla Al Suman
- Faculty of Medicine, Health and Human Sciences, Macquarie Medical School, Macquarie University, Balaclava Road, Sydney, NSW 2109, Australia
- Computational NeuroSurgery (CNS) Lab, Macquarie Medical School, Macquarie University, Balaclava Road, Sydney, NSW 2109, Australia
| | - Antonio Di Ieva
- Faculty of Medicine, Health and Human Sciences, Macquarie Medical School, Macquarie University, Balaclava Road, Sydney, NSW 2109, Australia
- Computational NeuroSurgery (CNS) Lab, Macquarie Medical School, Macquarie University, Balaclava Road, Sydney, NSW 2109, Australia
| |
Collapse
|
2
|
Chen K, Hwu T, Kashyap HJ, Krichmar JL, Stewart K, Xing J, Zou X. Neurorobots as a Means Toward Neuroethology and Explainable AI. Front Neurorobot 2020; 14:570308. [PMID: 33192435 PMCID: PMC7604467 DOI: 10.3389/fnbot.2020.570308] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2020] [Accepted: 08/25/2020] [Indexed: 12/18/2022] Open
Abstract
Understanding why deep neural networks and machine learning algorithms act as they do is a difficult endeavor. Neuroscientists are faced with similar problems. One way biologists address this issue is by closely observing behavior while recording neurons or manipulating brain circuits. This has been called neuroethology. In a similar way, neurorobotics can be used to explain how neural network activity leads to behavior. In real world settings, neurorobots have been shown to perform behaviors analogous to animals. Moreover, a neuroroboticist has total control over the network, and by analyzing different neural groups or studying the effect of network perturbations (e.g., simulated lesions), they may be able to explain how the robot's behavior arises from artificial brain activity. In this paper, we review neurorobot experiments by focusing on how the robot's behavior leads to a qualitative and quantitative explanation of neural activity, and vice versa, that is, how neural activity leads to behavior. We suggest that using neurorobots as a form of computational neuroethology can be a powerful methodology for understanding neuroscience, as well as for artificial intelligence and machine learning.
Collapse
Affiliation(s)
- Kexin Chen
- Department of Cognitive Sciences, University of California, Irvine, Irvine, CA, United States
| | - Tiffany Hwu
- HRL Laboratories (formerly Hughes Research Laboratory), LLC, Malibu, CA, United States
| | - Hirak J Kashyap
- Department of Computer Science, University of California, Irvine, Irvine, CA, United States
| | - Jeffrey L Krichmar
- Department of Cognitive Sciences, University of California, Irvine, Irvine, CA, United States.,Department of Computer Science, University of California, Irvine, Irvine, CA, United States
| | - Kenneth Stewart
- Department of Computer Science, University of California, Irvine, Irvine, CA, United States
| | - Jinwei Xing
- Department of Cognitive Sciences, University of California, Irvine, Irvine, CA, United States
| | - Xinyun Zou
- Department of Computer Science, University of California, Irvine, Irvine, CA, United States
| |
Collapse
|
3
|
Abstract
In visual search tasks, observers look for targets among distractors. In the lab, this often takes the form of multiple searches for a simple shape that may or may not be present among other items scattered at random on a computer screen (e.g., Find a red T among other letters that are either black or red.). In the real world, observers may search for multiple classes of target in complex scenes that occur only once (e.g., As I emerge from the subway, can I find lunch, my friend, and a street sign in the scene before me?). This article reviews work on how search is guided intelligently. I ask how serial and parallel processes collaborate in visual search, describe the distinction between search templates in working memory and target templates in long-term memory, and consider how searches are terminated.
Collapse
Affiliation(s)
- Jeremy M. Wolfe
- Department of Ophthalmology, Harvard Medical School, Boston, Massachusetts 02115, USA
- Department of Radiology, Harvard Medical School, Boston, Massachusetts 02115, USA
- Visual Attention Lab, Brigham & Women's Hospital, Cambridge, Massachusetts 02139, USA
| |
Collapse
|
4
|
Neuromodulated attention and goal-driven perception in uncertain domains. Neural Netw 2020; 125:56-69. [PMID: 32070856 DOI: 10.1016/j.neunet.2020.01.031] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2019] [Revised: 12/13/2019] [Accepted: 01/27/2020] [Indexed: 11/23/2022]
Abstract
In uncertain domains, the goals are often unknown and need to be predicted by the organism or system. In this paper, contrastive Excitation Backprop (c-EB) was used in two goal-driven perception tasks - one with pairs of noisy MNIST digits and the other with a robot in an action-based attention scenario. The first task included attending to even, odd, low, and high digits, whereas the second task included action goals, such as "eat", "work-on-computer", "read", and "say-hi" that led to attention to objects associated with those actions. The system needed to increase attention to target items and decrease attention to distractor items and background noise. Because the valid goal was unknown, an online learning model based on the cholinergic and noradrenergic neuromodulatory systems was used to predict a noisy goal (expected uncertainty) and re-adapt when the goal changed (unexpected uncertainty). This neurobiologically plausible model demonstrates how neuromodulatory systems can predict goals in uncertain domains and how attentional mechanisms can enhance the perception for that goal.
Collapse
|
5
|
Tsotsos JK, Kotseruba I, Wloka C. Rapid visual categorization is not guided by early salience-based selection. PLoS One 2019; 14:e0224306. [PMID: 31648265 PMCID: PMC6812801 DOI: 10.1371/journal.pone.0224306] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/20/2019] [Accepted: 10/11/2019] [Indexed: 11/19/2022] Open
Abstract
The current dominant visual processing paradigm in both human and machine research is the feedforward, layered hierarchy of neural-like processing elements. Within this paradigm, visual saliency is seen by many to have a specific role, namely that of early selection. Early selection is thought to enable very fast visual performance by limiting processing to only the most salient candidate portions of an image. This strategy has led to a plethora of saliency algorithms that have indeed improved processing time efficiency in machine algorithms, which in turn have strengthened the suggestion that human vision also employs a similar early selection strategy. However, at least one set of critical tests of this idea has never been performed with respect to the role of early selection in human vision. How would the best of the current saliency models perform on the stimuli used by experimentalists who first provided evidence for this visual processing paradigm? Would the algorithms really provide correct candidate sub-images to enable fast categorization on those same images? Do humans really need this early selection for their impressive performance? Here, we report on a new series of tests of these questions whose results suggest that it is quite unlikely that such an early selection process has any role in human rapid visual categorization.
Collapse
Affiliation(s)
- John K. Tsotsos
- Department of Electrical Engineering and Computer Science, York University, Toronto, ON, Canada
| | - Iuliia Kotseruba
- Department of Electrical Engineering and Computer Science, York University, Toronto, ON, Canada
| | - Calden Wloka
- Department of Electrical Engineering and Computer Science, York University, Toronto, ON, Canada
| |
Collapse
|
6
|
Tsotsos JK. Complexity Level Analysis Revisited: What Can 30 Years of Hindsight Tell Us about How the Brain Might Represent Visual Information? Front Psychol 2017; 8:1216. [PMID: 28848458 PMCID: PMC5552749 DOI: 10.3389/fpsyg.2017.01216] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/15/2016] [Accepted: 07/03/2017] [Indexed: 11/13/2022] Open
Abstract
Much has been written about how the biological brain might represent and process visual information, and how this might inspire and inform machine vision systems. Indeed, tremendous progress has been made, and especially during the last decade in the latter area. However, a key question seems too often, if not mostly, be ignored. This question is simply: do proposed solutions scale with the reality of the brain's resources? This scaling question applies equally to brain and to machine solutions. A number of papers have examined the inherent computational difficulty of visual information processing using theoretical and empirical methods. The main goal of this activity had three components: to understand the deep nature of the computational problem of visual information processing; to discover how well the computational difficulty of vision matches to the fixed resources of biological seeing systems; and, to abstract from the matching exercise the key principles that lead to the observed characteristics of biological visual performance. This set of components was termed complexity level analysis in Tsotsos (1987) and was proposed as an important complement to Marr's three levels of analysis. This paper revisits that work with the advantage that decades of hindsight can provide.
Collapse
Affiliation(s)
- John K Tsotsos
- Department of Electrical Engineering and Computer Science, York UniversityToronto, ON, Canada
| |
Collapse
|