Elsevier

Journal of Physiology-Paris

Volume 98, Issues 1–3, January–June 2004, Pages 235-248
Journal of Physiology-Paris

Ocular tracking as a measure of auditory motion perception

https://doi.org/10.1016/j.jphysparis.2004.03.010Get rights and content

Abstract

Motion is a potent sub-modality of vision. Motion cues alone can be used to segment images into figure and ground and break camouflage. Specific patterns of motion support vivid percepts of form, guide locomotion by specifying directional heading and the passage of objects, and in case of an impending collision, the time to impact. Visual motion also drives smooth pursuit eye movements (SPEMs) that serve to stabilize the retinal image of objects in motion.

In contrast, the auditory system does not appear to be particularly sensitive to motion. We review the ambiguous status of auditory motion processing from the psychophysical and electrophysiological perspectives. We then report the results of two experiments that use ocular tracking performance as an objective measure of the perception of auditory motion in humans. We examine ocular tracking of auditory motion, visual motion, combined auditory + visual motion and imagined motion in both the frontal plane and in depth. The results demonstrate that ocular tracking of auditory motion is no better than ocular tracking of imagined motion. These results are consistent with the suggestion that, unlike the visual system, the human auditory system is not endowed with low-level motion sensitive elements. We hypothesize however, that auditory information may gain access to a recently described high-level motion processing system that is heavily dependent on `top-down' influences, including attention.

Introduction

The goal of perceptual processing is to create an internal (neural) representation of the external environment. A primary dimension of most sensory representations is concerned with the specification of spatial location. This is certainly the case for vision and audition, for it is indeed difficult to imagine a visual or auditory percept that does not include a localization component.

A great deal of neural circuitry is devoted to performing the computations needed to localize visual and auditory stimuli. In fact, the most prominent theory of the neural mechanisms of visual processing hypothesizes that an extensive network of cortical areas extending throughout the occipital and parietal lobes is concerned primarily with coding visuo-spatial information [67]. One reason why the spatial dimension of sensory processing is so important is that spatial representations are needed to guide movements. Thus, it is not surprising that much of the visuo-spatial processing that occurs in the `dorsal processing stream' is concerned with the transformation of visuo-spatial information into pre-motor commands relating to reaching and the control of eye movements (e.g., [19], [20], [37], [50]). Auditory spatial information also is distributed to portions of this dorsal processing stream where a similar transformation into pre-motor signals appears to occur (e.g., [13], [15], [39], [55], [69], [73]).

In humans, psychophysical evidence suggests that the neural representation of visual space is more elaborated and precise than the representation of the acoustic space. Consider for example vernier acuity, which is the finest measure of visual localization (e.g., [70], [71]). In a typical vernier acuity task, an observer is required to detect a spatial misalignment of two visual contours, as is illustrated in Fig. 1. The threshold for such judgments corresponds to retinal offsets that are substantially smaller than the diameter of individual foveal cones. Under optimal conditions, vernier acuity thresholds are approximately 5 s of visual arc (0.0014°). By this standard, auditory localization is fairly coarse. The minimum audible angle (MAA), which is the finest measure of auditory spatial offset, is slightly greater than 1° of azimuth [49]. Thus, the threshold for spatial offsets in the frontal plane is about 700-times greater for auditory stimuli than visual stimuli. Consistent with these psychophysical data are the observations that saccadic eye movements to auditory stimuli are not as accurate as visually-guided saccades (e.g., [22], [32], [35]).

The manner in which auditory position information is distributed to the oculomotor system appears to be designed specifically to permit poly-modal access to the saccadic control system [35]. Consistent with this idea is the observation that saccades to bimodal targets (auditory + visual) are faster than saccades to auditory stimuli alone or visual stimuli alone. The magnitude of facilitation cannot be accounted for by probability summation (the gain produced by two independent detection processes––one for vision and one for audition [32], [33]). As a result, inter-sensory facilitation of saccades has been interpreted as a psychophysical correlate of the convergence of auditory and visual inputs onto single neurons that is known to occur within the superior colliculus (e.g., [18], [36], [47]).

While the saccade data indicate that neural representations of static locations of auditory and visual stimuli are distributed to neural centers that generate saccadic eye movements (or reaches), it is unclear whether analogous circuits exist for auditory and visual motion. Moreover, if such circuits do exist, is there poly-modal facilitation comparable to that seen in saccadic eye movements?

To address these questions, we compared smooth pursuit eye movements (SPEM) for moving auditory and visual stimuli. SPEMs are classically associated with tracking of a moving visual stimulus. Consequently, these movements typically require a neural signal specifying target velocity (e.g., [10], [61]). We reasoned that, since SPEMs rely on motion signals, any evidence of SPEMs based on auditory motion would imply that the human auditory system extracts information specifying stimulus velocity and distributes that information to the slow pursuit control system. If SPEMs are elicited by auditory motion, we may conclude that information about auditory motion is coded somewhere within the central auditory pathways and transmitted to the smooth pursuit system.

If, however, ocular tracking of auditory motion is not possible, we are left with two possible explanations. It may be that the auditory system generates the relevant motion signals, but those signals do not gain access to the pursuit control system. Alternatively, a failure to observe SPEMs elicited by auditory motion might be attributed to an absence of motion detectors in the human auditory system. It is this latter possibility that is the primary focus of this article.

We evaluated the quality of ocular tracking in response to two types of motion: motion in the frontal plane and motion in depth. Motion in the frontal plane produces conjunctive SPEMs, whereas motion in depth produces disjunctive SPEMs, or vergence eye movements. In the case of visually-guided pursuit, it is thought that different mechanisms drive conjunctive and disjunctive SPEMs––stereoscopic cues are considered essential for disjunctive SPEMs, and motion cues are considered essential for conjunctive SPEMs. Based on the evidence that each type of movement relies on different visual cues and different motor systems (cf., [10]), we decided to study possible auditory control over each type of movement.

Section snippets

Ocular pursuit of auditory motion in the frontal plane

We begin with an examination of ocular tracking of auditory motion in the frontal plane for several reasons. First, the processes that support auditory localization in the frontal plane are reasonably well understood (cf., [6], [14], [21], [28], [48], [56], [72]). The eccentricity of a sound source is computed through comparisons of the auditory signals arriving at each ear. Since the classic analysis by Lord Rayleigh [56], it has been recognized that two distinct binaural cues support sound

Methods

The results from Experiment 1 indicate that ocular tracking of auditory motion is accomplished almost exclusively by generating a succession of saccades. These observations confirm and extend earlier evidence that auditory motion is incapable of supporting smooth pursuit eye movements. However, it is more than a matter of poor smooth pursuit of auditory motion: the quality of ocular tracking of auditory motion was statistically indistinguishable from attempts to track imagined motion.

General discussion

The subjective impression that we experience auditory motion in much the same manner as we experience visual motion can be quite compelling. Yet objective data lending support to this phenomenology has not been forthcoming. To our knowledge, neurons that are specifically responsive to moving sound sources have not been reported, although not because of a lack of effort. A variety of subtle effects have been described that could be a component operation in the computation of auditory motion [1],

Acknowledgements

This work was supported by in part by NIH NS17778. The authors thank the Fondation des Treilles for their generous support of the conference which formed the basis for this issue of the Journal of Physiology, Paris.

References (72)

  • G. Westheimer et al.

    Spatial configurations for visual hyperacuity

    Vision Res

    (1977)
  • M. Ahissar et al.

    Encoding of sound-source location and movement: Activity of single neurons and interactions between adjacent neurons in the monkey auditory cortex

    J. Neurophysiol

    (1992)
  • W. Ball et al.

    Infant responses to impending collision: Optical and real

    Science

    (1971)
  • H.B. Barlow

    Summation and inhibition in the frog's retina

    J. Physiol. (Lond.)

    (1953)
  • H.B. Barlow et al.

    Selective sensitivity to direction of movement in ganglion cells of the rabbit retina

    Science

    (1963)
  • F. Baumgart et al.

    A movement-sensitive area in auditory cortex

    Nature

    (1999)
  • J. Blauert

    An introduction to binaural technology

  • T.G.R. Bower et al.

    Infant responses to approaching objects. An indicator of response to distal variables

    Percept. Psychophys

    (1970)
  • V. Braitenberg et al.

    Landing reaction of Musca domestica induced by visual stimuli

    Naturwissenschaften

    (1966)
  • R.H.S. Carpenter

    Movements of the Eyes

    (1988)
  • P. Cavanagh et al.

    Motion: the long and short of it

    Spatial Vision

    (1989)
  • C. Chubb et al.

    Drift-balanced random stimuli: A general basis for studying non-Fourier motion perception

    JOSA

    (1988)
  • Y.E. Cohen et al.

    A common reference frame for movement plans in the posterior parietal cortex

    Nat. Rev. Neurosci

    (2002)
  • C.-J. Dong et al.

    A contingent aftereffect in the auditory sytem

    Nat. Neurosci

    (1999)
  • C.-J. Dong et al.

    The auditory motion aftereffect: its tuning and specificity in the spatial and frequency domains

    Percept. Psychophys

    (2000)
  • U.C. Drager et al.

    Physiology of visual cells in mouse superior colliculus and correlation with somatosensory and auditory input

    Nature

    (1975)
  • C.J. Duffy et al.

    Responses of monkey MST neurons to optic flow stimuli with shifted centers of motion

    J. Neurosci

    (1995)
  • J.-R. Duhamel et al.

    The updating of the representation of visual space in parietal cortex by intended eye movements

    Science

    (1992)
  • W.E. Feddersen et al.

    Localization of high-frequency tones

    J. Acoust. Soc. Am

    (1957)
  • G.M. Gauthier et al.

    Eye tracking of self-moved targets in the absence of vision

    Exp. Brain Res

    (1976)
  • G.M. Gauthier et al.

    Eye movements in response to real and apparent motions of acoustic targets

    Percept. Motor Skills

    (1976)
  • G. Glenny et al.

    Hans Gertz revisited: the different effects of invisibility and darkness of pursuit eye movements

    Perception

    (1979)
  • D.W. Grantham

    Motion aftereffects with horizontally moving sound sources in the free field

    Percept. Psychophys

    (1989)
  • D.W. Grantham et al.

    Auditory motion aftereffects

    Percept. Psychophys

    (1979)
  • D.M. Green

    An Introduction to Hearing

    (1976)
  • T.D. Griffiths et al.

    Human brain areas involved in the analysis of auditory movement

    Hum. Brain Map

    (2000)
  • Cited by (19)

    • Sensorimotor control of standing balance

      2018, Handbook of Clinical Neurology
      Citation Excerpt :

      Auditory cues can induce illusion of self-motion (Urbantschitsch, 1897) and audiokinetic nystagmus (Dodge, 1923), similar (but weaker) to visual cues (Marme-Karelse and Bles, 1977; Tanahashi et al., 2015; for review, see Väljamäe, 2009 and Campos et al., 2018). For example, psychophysical testing of visual and auditory localization demonstrates a 0.0014° threshold during visual (Vernier) acuity tests and 1° threshold during minimum audible angle tests (Mills, 1958; Westheimer and McKee, 1977; Boucher et al., 2004). Sound intensity, interaural time, and interaural level differences, as well as Doppler effects, potentially all contribute to our sense of auditory motion (for review on auditory motion perception, see Carlile and Leung, 2016).

    • Do we track what we see? Common versus independent processing for motion perception and smooth pursuit eye movements: A review

      2011, Vision Research
      Citation Excerpt :

      p. 1371). Interestingly, tracking of non-visual objects such as auditory, tactile or proprioceptive targets with pursuit eye movements has been reported as very poor, with particularly low velocity gains in response to moving auditory stimuli (Berryhill, Chiu, & Hughes, 2006; Boucher, Lee, Cohen, & Hughes, 2004; Gauthier & Hofferer, 1976). These findings exclude a comparison between perception and pursuit, even though motion perception of non-visual targets can be intact.

    • How the eyes respond to sounds

      2024, Annals of the New York Academy of Sciences
    View all citing articles on Scopus
    View full text