Quantcast

Foveated display eye-tracking system and method

Imported: 13 Feb '17 | Published: 18 Jan '11

Brian T. Mitchell

USPTO - Utility Patents

Abstract

A system for detecting and recording eye-tracking data by presenting to an observer a display image with an area in focus surrounded by blurred areas. The observer may shift the focus area within the image by head movements which are wirelessly transmitted to a receiver and used to modify the displayed image. This system solves that part of the eye-tracking problem that is most critical for measuring cognitive processes.

Description

REFERENCE TO RELATED APPLICATION

This application claims priority from U.S. Provisional Patent Application Ser. No. 60/470,940, filed May 15, 2003, the entire content of which is incorporated herein by reference.

FIELD OF THE INVENTION

This invention relates to a system and method for analyzing and recording the eye movement of a subject exposed to a display of an image, and more particularly, to such a system in which the display image has a foveated region and a non-foveated region and the system detects head movement of the subject to move the foveated region within the image.

BACKGROUND OF THE INVENTION

There is little doubt that eye-tracking provides valuable insight into the current mental state and problem solving strategy of a human subject. The human retina is anisotropic, and has a high-resolution fovea surrounded by a much lower resolution peripheral retina. In order for this architecture to work properly, the eyes must move rapidly about the scene. The net effect of this movement, is the illusion that the human vision system (HVS) has both a large field-of-view and a high acuity across the entire visual field. What is actually happening, however, is that the rapid movements of the eye are sequentially capturing the information needed to solve a particular problem. It is the fact that eyes acquire information in small amounts and in a sequential manner that makes eye-tracking so important to understanding the state of the human subject. It is the only tool that allows a researcher to explore the portions of the cognitive process that cannot be articulated.

The research literature is filled with eye-tracking studies that validate that statement, and illustrate that eye-tracking can be used to measure learning. For example, eye-tracking was used to show strategic game playing differences between novice and expert players of the game Tetris. As reported, expert players do not play the game faster than novice players. Quite the contrary. Experts stretch out the placement of the current piece for as long as possible so that they can concentrate on where to place the next piece. This is actually a clever strategy that maximizes decision time. Although the expert player could not articulate this strategy, it was clearly evident in the eye movement histories. The novice player simply focused on placing the current piece, and in fact, played the game at a much faster rate than the expert. This simple example illustrates the power that eye-tracking has on uncovering non-intuitive problem solving strategies and the potential that eye-tracking has for developing measures of performance and training methods.

Although current methods of eye-tracking have demonstrated tremendous potential for the use of this data, they have fallen short in their overall ability to facilitate widespread use of this technology for training. Many of these methods capture and process a video stream of the working eye, and require extensive image processing methods to obtain eye position information. Unfortunately, these contemporary eye-trackers fall short of perfection, and as a result, they have received much criticism: they are too expensive, they are too difficult or cumbersome to wear, they are too hard to use, a chin rest or bite bar may be required to restrict head movement, they do not work for individuals that wear glasses or contact lenses, they require frequent calibrations and recalibrations, and the data they collect can be corrupted by blinks or glances.

In short, current eye-tracking methods are far from perfect, and none of the current devices completely satisfy all the requirements identified for an ideal eye-tracker. Alternative approaches to eye-tracking based on video oculography (VOG), infrared oculography (IROG), and electro-oculography (EOG) methods have also been developed.

The VOG approach relies on determining the relative position of the subject's cornea and a glint of infrared light reflected off the pupil. An infrared emitter is used to produce the glint and a video camera is used to capture a sequence of images that contain the glint and subject's eye. Image processing techniques are used to determine the position of the cornea and the glint. The position of the glint relative to the cornea changes as the eye moves. Thus, a VOG-based eye-tracker calculates eye position based on the relative position of the cornea and the reflected pupil glint. In general, this method of eye-tracking requires an infrared emitter and a video camera mounted to maintain a fixed relationship to the subject's eye that often results in cumbersome and expensive eye-tracking devices. This method has good spatial and temporal resolution, but head movements and eye blinks can effect image quality and tracking performance.

The IROG approach relies on measuring the intensity of infrared light reflected back from the subject's eye. Infrared emitters and infrared detectors are located in fixed positions around the eye. The amount of light reflected back to a fixed detector varies with eye position. Thus, an IROG-based eye-tracker calculates eye position based on the amount of reflected infrared light. In general, this method of eye-tracking requires goggles with mounted infrared emitters and detectors, and such devices are often both intrusive and expensive. This method has good spatial resolution and high temporal resolution, but is better for measuring horizontal than vertical eye movements and has difficulty with eye blinks which alter the amount of reflected light.

The EOG approach relies on the fact that the eye has a standing electrical potential across it with the front of the eye positive and the back of the eye negative. This potential varies from one to several millivolts, depending on the individual and illumination levels. EOG is measured by placing electrodes above and below the eye and on the outside of each eye. Changes in the EOG signals are directly related to changes in eye position. Thus, an EOG-based eye-tracker calculates eye position based on these signals. In general, this method of eye-tracking requires extensive calibration and lacks the precision needed for many eye-tracking applications. Foveated or gaze-contingent variable-resolution displays have been well developed in previous works. Foveated display techniques select the foveated region by actively tracking the subject's eyes and presenting an area of high spatial resolution at the point of gaze. Widespread application of such displays has been slowed by several difficulties. The technique requires fast and continuous tracking of the gaze point, and previous eyetrackers have been too expensive for widespread application, too imprecise, or too invasive for routine use. At present, the widespread use of gaze-contingent applications depends primarily on the development of low-cost eye-tracking systems.

SUMMARY OF THE INVENTION

The present invention relates to a new approach to eye-tracking. This approach simplifies the eye-tracking problem by using simple head movements, other body cues, or a combination of head and other body cues to control a foveated display. In a preferred embodiment of the invention, which will subsequently be disclosed in detail, the foveated region of the display is presented in-focus. The remainder of the display is blurred to correspond to the peripheral field of view found in the HVS. Head movements are used to move the foveated region within the field of view. Movements of the subject's head are detected and operate through the display system to change the foveated region of the image so that the subject acquires visual information by moving the foveated region through a series of head movements. The foveated region movements correspond to eye movements and may be recorded for subsequent analysis.

The benefits of this approach are significant. Gaze-point tracking simply becomes a matter of following the foveated region on the display. Not only is this simple to calculate, but it is more precise. Since the exact location of the fovea is known, this solution offers more precise calculation of the gaze point than conventional methods. It eliminates many of the calibration and recalibration issues normally associated with eye-tracker equipment, it is also less expensive. Since the gaze point corresponds to the display position of the foveal region, no calculations are needed to translate gaze point to screen coordinates. No optics or video cameras are required. Instead, a simple control device based on body motion is used to move the foveated region on the display screen, in short, the proposed method is more precise, easier to use, and less expensive than current methods.

While the preferred embodiment of the present invention uses head motion to control the foveated region, it might be possible in other embodiments to detect other body movement cues or combined head and other body cues to control the foveated display.

The system of the present invention requires a close correlation between head movement and eye movement. It is possible to move the eyes without moving the head, but the subject must make an effort to move the head to reposition the eyes. The user will learn the eye-head relationship over time through perceptual feedback provided through the foveated region placement.

Once learned, the position of the foveated region will correspond to eye gaze point, and the entire system will be perceptually seamless to the user. While the user may be initially drawn toward eye movement without head movement, since this will not change the in-focus area on the displayed image, the user will quickly learn that head movements are required to acquire information from the image.

Rather than explicitly measuring the position or orientation of the eye, the foveated display eye-tracker invention (FD eye-tracker) relies on a control strategy that works in concert with the head-eye control mechanism to move a foveated region on a computer screen. The idea is that the foveated display provides the same mechanism for acquiring visual information as the Human Vision System (HVS); a region of high-resolution must be sequentially moved around in the field of view at a rapid rate so that the human can obtain a complete understanding of the scene. By combining this approach with the natural head-eye control system of the Human Vision System, a device can be produced that would allow simple recording of the foveal position. Such a device would mirror the mechanism that the HVS uses to collect information, and would be controlled by an augmented version of the HVS head-eye control system.

DETAILED DESCRIPTION OF THE INVENTION

FIG. 1 is a schematic drawing of a subject generally indicated at 10, using the system of the present invention by observing a foveated display on a screen 12 such as a computer display screen. The image 14 displayed on the screen 12 has an area 16 that is in-focus and other areas 18 that are blurred. In the preferred embodiment, the foveated region 16 is circular in shape to correspond to the human retina. The subject 10 is positioned and directed so that the movement of the eyes to acquire visual information from the foveated image 14 is achieved by motions of the head which move the area 16.

The subject's head has a motion transmitter 20 attached. The signals generated by the transmitter 20 as the head moves, are picked up by a wireless receiver 24. The output of the receiver is provided to a foveated display generator 26. The foveated display generator 26 produces the image 14 with an in-focus area 16 surrounded by areas 18 that are blurred and moves the in-focus area 16 with respect to the general image, as a function of the detected head movements. In-focus area 16 positions are provided to a recorder 28.

The head tracker, generally including a transmitter 20 and a receiver 24, may take the form of any one of a variety of commercially available systems. One group of these head trackers is used to give disabled users control of the mouse function on a personal computer. These systems are generally low in cost and are directly applicable to the present invention. For example, the HeadMouse sold by Origin Instruments Corporation of Grand Prairie, Tex., would be useful with the present invention. It incorporates a small, flexible, paper-thin tracking dot typically attached to the forehead of the user and a receiver which detects the position of the dot using near-infrared signals.

FIGS. 2A and 2B represent displays of an image, generally indicated at 30 in both views, in which different areas of the image are displayed in-focus: the area 32 in FIG. 2A and the area 34 in FIG. 2B. The balance of both images are blurred. By movements of the head of the subject 10, the subject can shift the in-focus area to acquire information of the image. Those movements are recorded on a recorder 28 and may be played back in super position to the image 30, shown in FIG. 3, in which the entire image 30 is shown in-focus and the scanpath 36 used by the subject 10 to gain information of the image is superimposed on the image 30.

The image generated may be of the type disclosed in an article by Jansen, A. R., Blackwell, A. F., and Marriott, K., “A Tool for Tracking Visual Attention: The Restricted Focus Viewer,” Behavioral Research Methods Instruments and Computers, 35(1): 57-69, February 2003. The system disclosed in this paper has a functional goal similar to that disclosed herein, but operates in an entirely different manner, with the user moving a mouse to change the region in focus of a foveated image. The foveal display technique found in the restricted focus viewer (described in that paper) may be used with the present invention for the display unit 26, with the head motion receiver 24 of the present invention and a control system that mirrors the control system of the human vision system substituting for the mouse. The referenced paper describes a method of generating a graded, blurring effect about the central, in-focus portion of the foveated image, which is useful with the present invention.

The images 30 which may be displayed and used for detecting eye tracking through head tracking with the present invention are not limited to static images and may include moving images such as streaming video, motion pictures, and computer-generated video streams dynamically produced from computer applications.

The inventive system and method disclosed herein may be used with other methodologies to improve performance. For example, ancillary data such as EOG or other data may be combined with head movements to control the foveated region. The FD data may be time-synchronized with FD position, to produce a form of FD position time-stamping, and a sequence of FD points, such as those shown in FIG. 3, may be processed or post-processed for analysis purposes.

The FD position may coincide with a comfortable head position upon initialization. For example, the initialization of the FD position may appear in the center of the screen, or precisely at eye level. During initialization a comfortable relationship between the foveal display region and the user's head alignment is established before commencing movement of the foveal region. This is accomplished by displaying the foveal region in the center of the screen, aligning the user's head in a comfortable position with respect to that position of the foveal region, and then commencing movement of the foveal region as determined by subsequent head movements.

The foveal region may also be parameterized to allow for varying viewing distances, and a calculation may be used with respect to the distance between the head and the display to change the foveated region display characteristics to mirror those associated with HVS perceptual capabilities. The method used to parameterize the foveal region for viewing distance may be of the type disclosed in an article by T. Kuyel, W. S. Geisler and J. Ghosh, “Retinally reconstructed images: Digital images having a resolution match with the human eye”, IEEE Trans. Systems, Man and Cybernetics—Part A, 29(2), March 1999, pp. 235-243. The viewing distance between computer display and user's eyes may be dynamically calculated from the signals generated by the transmitter 20 as the head moves and received by the wireless receiver 24 using well-known methods such as triangulation or time-of-flight. The resulting distance can be used to dynamically modify the viewing characteristics of the foveated region to correspond to HVS perceptual capabilities.

A foveated display eye-tracker with no perceptual side effects has many positive attributes. Such a device would be relatively inexpensive, easy to use, produce very precise data, and be minimally invasive. It has all the qualities that the current technology does not. The FD eye-tracker is not ideal. It does, however, provide a set of qualities that would enable widespread use of the device for training and other applications.

The method of this invention preserves the portion of eye-tracking that provides a glimpse into cognitive processes (sequential visual data acquisition), and, as such, it creates a tool for understanding cognitive performance levels. Thus, even though it only solves a part of the eye-tracking problem in general, it only solves a part that is most critical for measuring cognitive processes. Accordingly, the system and method find numerous applications, including the ability to determine cognitive workload control over human-computer interaction. Human performance may be analyzed in these reactions, or real-world tasks may be simulated and analyzed in conjunction with the system and method.

For the military and commercial markets, such a device has the potential to revolutionize the methods by which individuals are trained to operate high-performance systems. Modern military systems require advanced training methods that assure the cognitive readiness of our warfighters. Achieving cognitive readiness is important for the following reasons; 1) the warfighter is mentally prepared, 2) the warfighter is performing at an optimal level, 3) the tools and techniques for training the warfighter are the most effective and affordable, and 4) the tools and techniques that the warfighter uses are the most effective and affordable. Because the post-Cold War challenges include the potential of simultaneous, multiple, geographically separate, high- or low-intensity conflicts, as well as peacekeeping, counter-terrorism, and disaster support missions, training for such diverse threats has the potential to strain limited resources. This, coupled with the increasing complexity of modern warfare techniques, has placed an increasing importance on developing training techniques that maximize the use of resources. This invention may be used to obtain a quick snapshot of student capabilities so that optimal training techniques can be prescribed. Thus, the disclosed eye-tracker has the potential to revolutionize the method in which modern warfighters are trained.

Within the commercial markets, the eye-tracker could support the development of products in such areas as advanced training, telemedicine, remote vehicle operation, and video teleconferencing.

Claims

1. An eye-tracking method, comprising the steps of:
generating an image in which a selected foveated region is more focused than the balance of the image;
displaying the generated image to a user;
tracking movements of the user's head;
changing the area of the image which mirrors user eye movement and is in-focus relative to the balance of the image as a function of the user's head movements to obtain information about the image;
recording the movements of the position of the in-focus area on the display screen while acquiring information from the image; and
using combined head movement data and eyetracking or EOG derived data to control the foveated region.
generating an image in which a selected foveated region is more focused than the balance of the image;
displaying the generated image to a user;
tracking movements of the user's head;
changing the area of the image which mirrors user eye movement and is in-focus relative to the balance of the image as a function of the user's head movements to obtain information about the image;
recording the movements of the position of the in-focus area on the display screen while acquiring information from the image; and
using combined head movement data and eyetracking or EOG derived data to control the foveated region.
2. An eye-tracking method, comprising the steps of:
generating an image in which a selected foveated region is more focused than the balance of the image;
displaying the generated image to a user;
tracking movements of the user's head;
changing the area of the image which mirrors user eye movement and is in-focus relative to the balance of the image as a function of the user's head movements to obtain information about the image;
recording the movements of the position of the in-focus area on the display screen while acquiring information from the image; and
time-synchronizing the foveated region to minor user eye movement as a function of combined eye and head position data.
generating an image in which a selected foveated region is more focused than the balance of the image;
displaying the generated image to a user;
tracking movements of the user's head;
changing the area of the image which mirrors user eye movement and is in-focus relative to the balance of the image as a function of the user's head movements to obtain information about the image;
recording the movements of the position of the in-focus area on the display screen while acquiring information from the image; and
time-synchronizing the foveated region to minor user eye movement as a function of combined eye and head position data.