US6200139B1 - Operator training system - Google Patents

Operator training system Download PDF

Info

Publication number
US6200139B1
US6200139B1 US09/258,848 US25884899A US6200139B1 US 6200139 B1 US6200139 B1 US 6200139B1 US 25884899 A US25884899 A US 25884899A US 6200139 B1 US6200139 B1 US 6200139B1
Authority
US
United States
Prior art keywords
operator
vehicle
looks
information
course
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
US09/258,848
Inventor
Edward O. Clapper
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Intel Corp
Original Assignee
Intel Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Intel Corp filed Critical Intel Corp
Assigned to INTEL CORPORATION reassignment INTEL CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CLAPPER, EDWARD C.
Priority to US09/258,848 priority Critical patent/US6200139B1/en
Priority to TW088123109A priority patent/TW440807B/en
Priority to DE60009057T priority patent/DE60009057T2/en
Priority to PCT/US2000/000723 priority patent/WO2000051097A2/en
Priority to CNB008043183A priority patent/CN1248174C/en
Priority to JP2000601623A priority patent/JP3787493B2/en
Priority to AU25028/00A priority patent/AU2502800A/en
Priority to AT00903253T priority patent/ATE262207T1/en
Priority to KR10-2001-7010227A priority patent/KR100416152B1/en
Priority to EP00903253A priority patent/EP1177544B1/en
Publication of US6200139B1 publication Critical patent/US6200139B1/en
Application granted granted Critical
Priority to JP2005323273A priority patent/JP2006163375A/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B9/00Simulators for teaching or training purposes
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B19/00Teaching not covered by other main groups of this subclass
    • G09B19/16Control of vehicles or other craft
    • G09B19/167Control of land vehicles
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B9/00Simulators for teaching or training purposes
    • G09B9/02Simulators for teaching or training purposes for teaching control of vehicles or other craft
    • G09B9/04Simulators for teaching or training purposes for teaching control of vehicles or other craft for teaching control of land vehicles
    • G09B9/042Simulators for teaching or training purposes for teaching control of vehicles or other craft for teaching control of land vehicles providing simulation in a real vehicle
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B9/00Simulators for teaching or training purposes
    • G09B9/02Simulators for teaching or training purposes for teaching control of vehicles or other craft
    • G09B9/04Simulators for teaching or training purposes for teaching control of vehicles or other craft for teaching control of land vehicles
    • G09B9/05Simulators for teaching or training purposes for teaching control of vehicles or other craft for teaching control of land vehicles the view from a vehicle being simulated
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B9/00Simulators for teaching or training purposes
    • G09B9/02Simulators for teaching or training purposes for teaching control of vehicles or other craft
    • G09B9/04Simulators for teaching or training purposes for teaching control of vehicles or other craft for teaching control of land vehicles
    • G09B9/052Simulators for teaching or training purposes for teaching control of vehicles or other craft for teaching control of land vehicles characterised by provision for recording or measuring trainee's performance

Definitions

  • This invention related generally to systems for training and evaluating operators of vehicles such as automobiles, trucks, airplanes, trains, boats and the like.
  • vehicle operators may be trained using simulators.
  • aircraft training simulators are known which expose the operator to a variety of different conditions and evaluate the operator's response to those conditions.
  • automated driver testing the driver may be asked a series of questions and his or her answers are then checked against the correct answers in a computer database. All of these approaches respond to the need to obtain better and lower cost evaluation and training of vehicle operators. This hopefully improves the performance of those operators and reduces their training cost.
  • Some characteristics of a good operator may be difficult to assess.
  • a trainer may watch the operator for certain characteristics such as checking rear view mirrors, and checking gauges and the like.
  • an objective evaluation of whether the person has mastered the skills necessary to operate the vehicle correctly may be difficult to obtain. For example, it may be difficult to compare the abilities of one driver quantitatively to those of other drivers.
  • a vehicle operator training system may include a sensor adapted to sense what an operator looks at.
  • a controller is adapted to record information about what the operator looked at.
  • FIG. 1 is a front elevational view of the front cockpit area of a vehicle as seen by the operator;
  • FIG. 2 is a block diagram of an apparatus for detecting what an operator looks at in the course of operating a vehicle
  • FIG. 3 is a flow chart for determining what the operator is looking at while operating a vehicle
  • FIG. 4 shows a flow chart for evaluating an operator in accordance with one embodiment of the present invention
  • FIG. 5 is a diagram showing one embodiment of a processor based system for implementing the system shown in the preceding figures.
  • FIG. 6 is a screen display showing a simulated playback display of a driver's performance in accordance with one embodiment of the invention.
  • a vehicle operator may sit in a cockpit 10 which may be a simulator or an actual vehicle such as an automobile, truck, airplane, train or boat, as examples.
  • the cockpit 10 may include a windshield 12 , a dashboard 14 , a steering wheel 16 , a rear view mirror 18 , and a side view mirror 20 . It may also include a gaze camera 26 and an eye camera 28 .
  • the gaze camera 26 may be coupled by a pan/tilt unit 22 to infrared light emitting diodes (LEDs) 24 .
  • the cameras 26 and 28 provide the input information to determine what the operator looks at in the course of operating the vehicle. This can be used to evaluate whether the operator is appropriately using sources of information to effectively operate the vehicle.
  • the system may determine whether the operator looks out the front windshield 12 , whether the operator scans around the front windshield, whether the operator checks the side view mirror 20 and the rear view mirror 18 and whether the operator checks the various gauges provided on the dashboard 14 .
  • the eye camera 28 may have a short focal distance and may be used to measure the three dimensional eye position.
  • the gaze camera 26 may be used to determine a gaze function. It may have a long focal distance and may be panned and tilted to follow the eye when the head moves.
  • the infrared LED array 24 may be mounted a short distance from the optical axis of gaze camera to illuminate the eye and to cause a reflection on the eye that may be useful in tracking the movement of the eye.
  • a system is described herein for determining what the user is looking at, using eye and gaze tracking, principle component analysis and an infrared LED array.
  • the invention is in no way limited to this technology. Instead, this technology is merely described to illustrate one technique for determining what the operator is looking at while operating the vehicle.
  • Other known techniques for determining what a person looks at include using head mounted cameras. For example, a grid of LEDs may be pulsed in sequence. The camera image may be processed to detect flashes.
  • Another system uses a number of video cameras to capture simultaneous images that are coordinated to track an object.
  • the eye tracker 34 and gaze tracker 38 receive inputs from the cameras 26 and 28 . Namely, as illustrated by the image 32 , the eye tracker receives a digital image from the camera 26 which corresponds to a face shot of the operator. At the same time, the gaze tracker 38 receives an image of one or both of the operator's eyes. The eye tracker analyzes the video image of the eye camera 28 and detects and tracks three dimensional eye position. In addition, either or both pupils may be detected and tracked to allow calculation of the three dimensional
  • the eye tracker 34 may cause the eye camera 28 to be panned and tilted, using the pan/tilt unit 40 , to follow the position of the eye as the operator's head moves.
  • Motion compensation may be provided as indicated at block 42 to compensate for operator head movement.
  • a still head camera image may be received for analysis. Dark regions called valleys may be identified and eliminated to expedite the ensuing analysis.
  • Principal component analysis may be used to locate the operator's eyes.
  • characteristics of human eyes may be recorded to locate eye position using PCA. These characteristics may be stored as reference eye patterns called eigenfaces.
  • the camera 26 image may then be analyzed and compared with the stored eye patterns to locate the eyes. In one embodiment, these determined eye locations may be searched in another stereoscopic camera image (if an additional camera is used). By stereo matching, the three dimensional positions of both eyes can be determined. This information may then be used to control the pan and tilt unit 22 of the eye camera 26 .
  • a cornea-reflex method may be used to determine the gaze direction.
  • Low power infrared LEDs 24 may be mounted on the pan tilt unit 22 at one side of the camera 26 with the normal of the array surface parallel to the optical axis of the camera 26 .
  • the LEDs 24 may illuminate the eye and generate a highlight on the cornea.
  • a cornea reflex algorithm may identify the center of one or both of the pupils and the center of the cornea surface reflection. The distance between the two centers and their orientation may be used to determine gaze direction.
  • the change in gaze direction due to head movement may be compensated using information regarding the three dimensional head position.
  • PCA represents the general characteristics of human eyes using relatively few dimensions.
  • PCA transforms a luminance description of the eyes into a different coordinate system. This coordinate system is such that the mean square error arising from truncating basic vectors of the coordinate system is reduced.
  • the input from the camera 28 may be used to develop the head view shown at 32 in FIG. 2 and as indicated at block 46 and FIG. 3 .
  • the system may deduce some initial characteristics of facial features. For example, dark areas may be eliminated since they may not be as useful in locating facial features that indicate eye location.
  • facial feature detection may be implemented using PCA.
  • the three dimensional eye position may be determined using either a single camera or a pair of stereoscopic cameras. The location of the pupil may then be detected using pattern recognition and/or cornea reflection analysis as indicated in block 58 using the input from camera 26 as indicated at 36 in FIG. 2 .
  • the highlight caused by the LED reflection on the cornea may be detected as indicated in block 60 and a displacement factor determined as indicated in block 62 .
  • the displacement factor corrects for movement of the head, using information from the other camera 28 .
  • the PCA analysis is utilized to determine the gaze direction (block 64 ), as indicated at block 66 .
  • a plurality of cameras may be distributed around the cockpit to facilitate face location when the user turns around. For example, when the operator turns to look behind the vehicle, the cameras 26 , 28 may “lose” the operator's face. The cameras distributed about the cockpit can be used to detect the position of the operator's face. This information may be analyzed to determine whether the operator's actions were appropriate.
  • the gaze direction may be used to evaluate or train the operator.
  • software 66 may begin by receiving eye movement information as indicated in block 68 . In one embodiment, this may be information about what the eye is actually looking at, at any instance in time. This information may then be resolved into gaze coordinates as indicated in block 70 .
  • the gaze coordinates may correspond to targets such as the side view or rear view mirrors.
  • the gaze coordinates may include information about what the eye is actually looking at any instance in time.
  • the gaze coordinates information may be resolved into gazed upon targets as indicated in block 70 . Knowing the gaze coordinates of the operator and coordinates of gaze targets, the particular object being viewed can be determined.
  • the gazed upon targets correspond to targets such as the side view or rear view mirrors, the instrument panel and the front windshield in an automobile.
  • the coordinates of each of these objects may be compared to a given gaze direction and when the gaze direction detected by the system 30 correlates generally to a given target's coordinates, that gaze information can be resolved to determine that the operator's gaze is directed to one of the known gaze targets.
  • the target gaze frequency may be determined by determining how many times the operator looked at a given gaze target or object, as indicated in block 72 . In this way, the system can determine in a given predefined course of travel how many times the operator looked at given objects.
  • the operator may operate the vehicle, travelling a simulated course in which the number of times which the operator should check each target may be known.
  • the operator may operate the vehicle over a known course wherein it is known how many times the operator should reasonably be expected to check the various gaze targets.
  • the target gaze frequency information may be correlated and compared to standard ranges of gaze frequencies. For example, it may be determined that on average the operator should check the rear view mirror X number of times per hour. Knowing the predetermined frequency, an analysis may be undertaken to determine if the operator's performance falls within established norms.
  • a check determines whether the operator should be prompted to change his or her gaze frequency. Rather then simply evaluating the user, the system may train the operator to increase his or her frequency of checking particular sources of information. For example, if it is determined that the operator is not checking the rear view mirror frequently enough, the operator could be prompted as indicated in block 76 to check the rear view mirror more often.
  • the system may also receive other information including the vehicle's speed as indicated in block 78 .
  • the speed history may be recorded and may be compared to known speed limits or known speed recommendations for a given course.
  • the system may receive control information as indicated in block 82 .
  • the system may record when and how often the operator applies the brakes or other controls and how often the operator uses the turn indicator or other indications. As before, this information may be compared to what is expected over a predefined course or to norms in other cases.
  • the system may also receive vehicle position information, as indicated by block 83 .
  • Vehicle locating apparatus such as a global positioning system (GPS) receiver may be used to determine the vehicle's position and that information may be correlated with gaze target information. The system can then evaluate whether the operator was looking at what the operator should have been looking at, at a given position, for example, in a predefined training or evaluation course.
  • GPS global positioning system
  • the operator's gaze target frequency, speed and control/indicator performance may be evaluated. This may be done by comparing the operator's performance to a database of normal values over a wide range of different operators over different courses and conditions. Alternatively, it may be developed by comparing the operator's performance over a known course to expected performance. The operator may then be provided information about how his or her operating skills compare to what is expected. This may be used to evaluate the operator in the case of operator testing or to provide feedback to the operator about how to improve his or her performance.
  • a processor based system 100 for implementing one embodiment of the invention, shown in FIG. 5, includes a processor 102 .
  • the processor may be coupled to an accelerated graphics port (AGP) (see Accelerated Graphics Port Interface Specification, version 1.0, published on Jul. 31, 1996, by Intel Corporation, Santa Clara, Calif.) chipset 104 for implementing an accelerated graphics port embodiment.
  • AGP Accelerated Graphics Port Interface Specification
  • Chipset 104 may communicate with the AGP port 105 and the graphics accelerator 106 .
  • a display 110 may be coupled to the video output of the graphics accelerator 106 .
  • chipset 104 may also be coupled to the system memory 108 and to a bus 112 .
  • the bus 112 is also coupled to the motion compensator 42 . It may receive input signals from the cameras 26 and 28 and provide output signals to the pan/tilt unit 22 and the array 24 .
  • the bus 112 may also be coupled to a bridge 116 which couples a hard disk drive 118 .
  • the software 66 may be stored on the hard disk drive 118 together with eigenface information 116 and the calibration database 114 .
  • the bridge 116 may also coupled to another bus 132 .
  • the bus 132 may be coupled to a serial input/output (SIO) device 134 .
  • the device 134 may receive serial inputs from a control interface 136 , an indicator interface 138 and a speed interface 140 , each coupled to receive information from the vehicle and a position location system 144 .
  • a basic input/output system (BIOS) 142 is also coupled to the bus 132 .
  • a processor based playback of a vehicle operating session may be undertaken to provide either performance evaluation or feedback to a given operator.
  • the display screen 110 may include a display which includes a dashboard 86 , steering column 88 and driver 90 in a simulated format. The driver's line of sight is indicated by dashed line 92 .
  • the driver's line of sight over time and at particular locations can be recreated as a graphical user interface such as the dashed line 92 shown in FIG. 6 .
  • the trainee and the trainer can view what the operator was looking at, at given instances of time and in given situations.
  • the display portion 94 may provide information about the circumstances at each instance of time.
  • a global positioning system landmark may be indicated on the display, in one embodiment, to provide information about where the vehicle was when the operator was looking in a given gaze direction.
  • the information displayed on the screen, as indicated in 94 may include, for each instance of time, course conditions, weather indications, speed indications, a braking factor indicating the use of the brakes and a blind spot factor indicating how effective the driver was in monitoring blind spots.
  • the trainee may view his or her performance replayed at regular time intervals or continuously along a given course.
  • the replay may also indicate where the vehicle was when the operator reacted in a given fashion.
  • the system 30 may also provide exterior events which may be triggered by the system to either evaluate the operator's performance or to provide a training exercise. Exterior event generation may be synchronized with the vehicle's position as provided by a global positioning system receiver. For example, another vehicle may be caused to move into the vehicle's line of travel when the vehicle approaches a given position. The operator's response to the event, including how the operator's line of sight detects the event and how the operator responds, may be evaluated by the processor based system.
  • the exterior events may be based on set scripts triggered by global positioning coordinates of the vehicle to test the operator's performance under given conditions.
  • the operator's response time to a given event may also be determined and recorded.
  • the system may adaptively generate events which require viewing in that region. In this way, the operator can be conditioned to overcome bad habits, having seen the consequences of failing to maintain a wide range of view.
  • the operator's response to conditions keyed to land markers may be judged. For example, if GPS coordinates are known for the lane stripes, one can determine, based on global positioning system coordinates, whether the operator maintained a central position within his or her lane or wandered to the left or the right.
  • the system may also receive weather and time of day inputs which may be taken into account in evaluating the operator's performance.
  • Weather, traffic conditions and other information may be retrieved from the Internet or by sensors along the way. This information may be used to evaluate the operator's performance under the existing conditions.
  • the line of sight indicator may be overlaid onto the recorded image.
  • video playback can provide more detailed feedback about what was occurring relative to the operator's line of sight and other operator reactions.

Abstract

A system for operator training and evaluation may detect and record information about various actions taken by an operator. For example, the system may record what the operator looks at in the course of operating the vehicle to determine whether the operator is looking at the things which the operator should appropriately be checking. In addition the system may provide other information about other aspects of the operator's performance including speed and use of controls and indicators. This information may be correlated and provided in the form of an evaluation. Alternatively, the operator may be prompted in the course of operating the vehicle to correct the operator's performance to improve the operator's skill.

Description

BACKGROUND
This invention related generally to systems for training and evaluating operators of vehicles such as automobiles, trucks, airplanes, trains, boats and the like.
Currently, vehicle operators may be trained using simulators. For example, aircraft training simulators are known which expose the operator to a variety of different conditions and evaluate the operator's response to those conditions. In addition, in automated driver testing, the driver may be asked a series of questions and his or her answers are then checked against the correct answers in a computer database. All of these approaches respond to the need to obtain better and lower cost evaluation and training of vehicle operators. This hopefully improves the performance of those operators and reduces their training cost.
Some characteristics of a good operator may be difficult to assess. A trainer may watch the operator for certain characteristics such as checking rear view mirrors, and checking gauges and the like. However, an objective evaluation of whether the person has mastered the skills necessary to operate the vehicle correctly may be difficult to obtain. For example, it may be difficult to compare the abilities of one driver quantitatively to those of other drivers.
Thus, there is a continuing need for better techniques for training and evaluating vehicle operators.
SUMMARY
In accordance with one embodiment, a vehicle operator training system may include a sensor adapted to sense what an operator looks at. A controller is adapted to record information about what the operator looked at.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a front elevational view of the front cockpit area of a vehicle as seen by the operator;
FIG. 2 is a block diagram of an apparatus for detecting what an operator looks at in the course of operating a vehicle;
FIG. 3 is a flow chart for determining what the operator is looking at while operating a vehicle;
FIG. 4 shows a flow chart for evaluating an operator in accordance with one embodiment of the present invention;
FIG. 5 is a diagram showing one embodiment of a processor based system for implementing the system shown in the preceding figures; and
FIG. 6 is a screen display showing a simulated playback display of a driver's performance in accordance with one embodiment of the invention.
DETAILED DESCRIPTION
Referring to FIG. 1, a vehicle operator may sit in a cockpit 10 which may be a simulator or an actual vehicle such as an automobile, truck, airplane, train or boat, as examples. The cockpit 10 may include a windshield 12, a dashboard 14, a steering wheel 16, a rear view mirror 18, and a side view mirror 20. It may also include a gaze camera 26 and an eye camera 28.
The gaze camera 26 may be coupled by a pan/tilt unit 22 to infrared light emitting diodes (LEDs) 24. The cameras 26 and 28 provide the input information to determine what the operator looks at in the course of operating the vehicle. This can be used to evaluate whether the operator is appropriately using sources of information to effectively operate the vehicle.
For example, in an automobile application, the system may determine whether the operator looks out the front windshield 12, whether the operator scans around the front windshield, whether the operator checks the side view mirror 20 and the rear view mirror 18 and whether the operator checks the various gauges provided on the dashboard 14.
The eye camera 28 may have a short focal distance and may be used to measure the three dimensional eye position. The gaze camera 26 may be used to determine a gaze function. It may have a long focal distance and may be panned and tilted to follow the eye when the head moves. The infrared LED array 24 may be mounted a short distance from the optical axis of gaze camera to illuminate the eye and to cause a reflection on the eye that may be useful in tracking the movement of the eye.
A system is described herein for determining what the user is looking at, using eye and gaze tracking, principle component analysis and an infrared LED array. However, the invention is in no way limited to this technology. Instead, this technology is merely described to illustrate one technique for determining what the operator is looking at while operating the vehicle. Other known techniques for determining what a person looks at include using head mounted cameras. For example, a grid of LEDs may be pulsed in sequence. The camera image may be processed to detect flashes. Another system uses a number of video cameras to capture simultaneous images that are coordinated to track an object.
The illustrated system is further explained in “Determination of the Point of Fixation in a Head-fixed Coordinate System,” by Jin Liu presented at the 14th International Conference in Pattern Recognition (ICPR'98) held at Brisbane, Australia Aug. 17-20, 1998. Additional information about the system may also be found in an article by Kay Talmi and Jin Liu, titled “Eye and Gaze Tracking for Visually Controlled Interactive Stereoscopic Displays” which may currently be found on the Internet at http://www.hhi.de/˜blick/papers/eyegaze 97/eye-gaze.html.
Referring to FIG. 2, the eye tracker 34 and gaze tracker 38 receive inputs from the cameras 26 and 28. Namely, as illustrated by the image 32, the eye tracker receives a digital image from the camera 26 which corresponds to a face shot of the operator. At the same time, the gaze tracker 38 receives an image of one or both of the operator's eyes. The eye tracker analyzes the video image of the eye camera 28 and detects and tracks three dimensional eye position. In addition, either or both pupils may be detected and tracked to allow calculation of the three dimensional
The eye tracker 34 may cause the eye camera 28 to be panned and tilted, using the pan/tilt unit 40, to follow the position of the eye as the operator's head moves. Motion compensation may be provided as indicated at block 42 to compensate for operator head movement.
Initially a still head camera image may be received for analysis. Dark regions called valleys may be identified and eliminated to expedite the ensuing analysis.
Principal component analysis (PCA) may be used to locate the operator's eyes. In a training phase, characteristics of human eyes may be recorded to locate eye position using PCA. These characteristics may be stored as reference eye patterns called eigenfaces. The camera 26 image may then be analyzed and compared with the stored eye patterns to locate the eyes. In one embodiment, these determined eye locations may be searched in another stereoscopic camera image (if an additional camera is used). By stereo matching, the three dimensional positions of both eyes can be determined. This information may then be used to control the pan and tilt unit 22 of the eye camera 26.
A cornea-reflex method may be used to determine the gaze direction. Low power infrared LEDs 24 may be mounted on the pan tilt unit 22 at one side of the camera 26 with the normal of the array surface parallel to the optical axis of the camera 26. The LEDs 24 may illuminate the eye and generate a highlight on the cornea. A cornea reflex algorithm may identify the center of one or both of the pupils and the center of the cornea surface reflection. The distance between the two centers and their orientation may be used to determine gaze direction. The change in gaze direction due to head movement may be compensated using information regarding the three dimensional head position.
PCA represents the general characteristics of human eyes using relatively few dimensions. PCA transforms a luminance description of the eyes into a different coordinate system. This coordinate system is such that the mean square error arising from truncating basic vectors of the coordinate system is reduced.
Referring now to FIG. 3, the input from the camera 28 may be used to develop the head view shown at 32 in FIG. 2 and as indicated at block 46 and FIG. 3. During preprocessing, indicated in block 48, the system may deduce some initial characteristics of facial features. For example, dark areas may be eliminated since they may not be as useful in locating facial features that indicate eye location. At block 50, facial feature detection may be implemented using PCA. At block 54, the three dimensional eye position may be determined using either a single camera or a pair of stereoscopic cameras. The location of the pupil may then be detected using pattern recognition and/or cornea reflection analysis as indicated in block 58 using the input from camera 26 as indicated at 36 in FIG. 2.
The highlight caused by the LED reflection on the cornea may be detected as indicated in block 60 and a displacement factor determined as indicated in block 62. The displacement factor corrects for movement of the head, using information from the other camera 28. Finally the PCA analysis is utilized to determine the gaze direction (block 64), as indicated at block 66.
In one embodiment of the invention, a plurality of cameras may be distributed around the cockpit to facilitate face location when the user turns around. For example, when the operator turns to look behind the vehicle, the cameras 26, 28 may “lose” the operator's face. The cameras distributed about the cockpit can be used to detect the position of the operator's face. This information may be analyzed to determine whether the operator's actions were appropriate.
The gaze direction may be used to evaluate or train the operator. Referring to FIG. 4, software 66 may begin by receiving eye movement information as indicated in block 68. In one embodiment, this may be information about what the eye is actually looking at, at any instance in time. This information may then be resolved into gaze coordinates as indicated in block 70.
The gaze coordinates may correspond to targets such as the side view or rear view mirrors. In one embodiment, the gaze coordinates may include information about what the eye is actually looking at any instance in time. For example, the gaze coordinates information may be resolved into gazed upon targets as indicated in block 70. Knowing the gaze coordinates of the operator and coordinates of gaze targets, the particular object being viewed can be determined. The gazed upon targets correspond to targets such as the side view or rear view mirrors, the instrument panel and the front windshield in an automobile. The coordinates of each of these objects may be compared to a given gaze direction and when the gaze direction detected by the system 30 correlates generally to a given target's coordinates, that gaze information can be resolved to determine that the operator's gaze is directed to one of the known gaze targets.
Next, the target gaze frequency may be determined by determining how many times the operator looked at a given gaze target or object, as indicated in block 72. In this way, the system can determine in a given predefined course of travel how many times the operator looked at given objects.
Thus, in one example, the operator may operate the vehicle, travelling a simulated course in which the number of times which the operator should check each target may be known. Alternatively, the operator may operate the vehicle over a known course wherein it is known how many times the operator should reasonably be expected to check the various gaze targets. Alternatively, the target gaze frequency information may be correlated and compared to standard ranges of gaze frequencies. For example, it may be determined that on average the operator should check the rear view mirror X number of times per hour. Knowing the predetermined frequency, an analysis may be undertaken to determine if the operator's performance falls within established norms.
At diamond 74, a check determines whether the operator should be prompted to change his or her gaze frequency. Rather then simply evaluating the user, the system may train the operator to increase his or her frequency of checking particular sources of information. For example, if it is determined that the operator is not checking the rear view mirror frequently enough, the operator could be prompted as indicated in block 76 to check the rear view mirror more often.
The system may also receive other information including the vehicle's speed as indicated in block 78. The speed history may be recorded and may be compared to known speed limits or known speed recommendations for a given course. Next, the system may receive control information as indicated in block 82. The system may record when and how often the operator applies the brakes or other controls and how often the operator uses the turn indicator or other indications. As before, this information may be compared to what is expected over a predefined course or to norms in other cases.
The system may also receive vehicle position information, as indicated by block 83. Vehicle locating apparatus, such as a global positioning system (GPS) receiver may be used to determine the vehicle's position and that information may be correlated with gaze target information. The system can then evaluate whether the operator was looking at what the operator should have been looking at, at a given position, for example, in a predefined training or evaluation course.
Referring to block 84, the operator's gaze target frequency, speed and control/indicator performance may be evaluated. This may be done by comparing the operator's performance to a database of normal values over a wide range of different operators over different courses and conditions. Alternatively, it may be developed by comparing the operator's performance over a known course to expected performance. The operator may then be provided information about how his or her operating skills compare to what is expected. This may be used to evaluate the operator in the case of operator testing or to provide feedback to the operator about how to improve his or her performance.
A processor based system 100 for implementing one embodiment of the invention, shown in FIG. 5, includes a processor 102. In one embodiment, the processor may be coupled to an accelerated graphics port (AGP) (see Accelerated Graphics Port Interface Specification, version 1.0, published on Jul. 31, 1996, by Intel Corporation, Santa Clara, Calif.) chipset 104 for implementing an accelerated graphics port embodiment. Chipset 104 may communicate with the AGP port 105 and the graphics accelerator 106. A display 110 may be coupled to the video output of the graphics accelerator 106. Then chipset 104 may also be coupled to the system memory 108 and to a bus 112.
The bus 112 is also coupled to the motion compensator 42. It may receive input signals from the cameras 26 and 28 and provide output signals to the pan/tilt unit 22 and the array 24. The bus 112 may also be coupled to a bridge 116 which couples a hard disk drive 118. The software 66 may be stored on the hard disk drive 118 together with eigenface information 116 and the calibration database 114. The bridge 116 may also coupled to another bus 132. The bus 132 may be coupled to a serial input/output (SIO) device 134. The device 134 may receive serial inputs from a control interface 136, an indicator interface 138 and a speed interface 140, each coupled to receive information from the vehicle and a position location system 144. Also coupled to the bus 132 is a basic input/output system (BIOS) 142.
In accordance with one embodiment of the present invention, a processor based playback of a vehicle operating session may be undertaken to provide either performance evaluation or feedback to a given operator. Referring to FIG. 6, the display screen 110 may include a display which includes a dashboard 86, steering column 88 and driver 90 in a simulated format. The driver's line of sight is indicated by dashed line 92.
Based on the information received from the motion compensator 42, the driver's line of sight over time and at particular locations can be recreated as a graphical user interface such as the dashed line 92 shown in FIG. 6. Thus, the trainee and the trainer can view what the operator was looking at, at given instances of time and in given situations. For example, the display portion 94 may provide information about the circumstances at each instance of time. A global positioning system landmark may be indicated on the display, in one embodiment, to provide information about where the vehicle was when the operator was looking in a given gaze direction.
The information displayed on the screen, as indicated in 94, may include, for each instance of time, course conditions, weather indications, speed indications, a braking factor indicating the use of the brakes and a blind spot factor indicating how effective the driver was in monitoring blind spots. In this way, the trainee may view his or her performance replayed at regular time intervals or continuously along a given course. The replay may also indicate where the vehicle was when the operator reacted in a given fashion.
The system 30 may also provide exterior events which may be triggered by the system to either evaluate the operator's performance or to provide a training exercise. Exterior event generation may be synchronized with the vehicle's position as provided by a global positioning system receiver. For example, another vehicle may be caused to move into the vehicle's line of travel when the vehicle approaches a given position. The operator's response to the event, including how the operator's line of sight detects the event and how the operator responds, may be evaluated by the processor based system.
The exterior events may be based on set scripts triggered by global positioning coordinates of the vehicle to test the operator's performance under given conditions. The operator's response time to a given event may also be determined and recorded.
For example, if the operator continually avoids looking in a given direction or at a given item, the system may adaptively generate events which require viewing in that region. In this way, the operator can be conditioned to overcome bad habits, having seen the consequences of failing to maintain a wide range of view.
By determining global positioning coordinates for the vehicle, the operator's response to conditions keyed to land markers may be judged. For example, if GPS coordinates are known for the lane stripes, one can determine, based on global positioning system coordinates, whether the operator maintained a central position within his or her lane or wandered to the left or the right.
The system may also receive weather and time of day inputs which may be taken into account in evaluating the operator's performance. Weather, traffic conditions and other information may be retrieved from the Internet or by sensors along the way. This information may be used to evaluate the operator's performance under the existing conditions.
In some embodiments it may be desirable to provide a camera to record the view out the windshield. Then in a playback mode, the actual recorded view out the vehicle windshield may be played back to make the playback more realistic. The line of sight indicator may be overlaid onto the recorded image. In addition, video playback can provide more detailed feedback about what was occurring relative to the operator's line of sight and other operator reactions.
While the invention has been disclosed with respect to a limited number of embodiments, those skilled in the art will appreciate numerous modifications and variations therefrom. It is intended that the appended claims cover all such modifications and variations as fall within the true spirit and scope of the invention.

Claims (25)

What is claimed is:
1. A vehicle operator training system comprising:
a sensor to sense what an operator looks at;
a storage storing information about a plurality of gaze targets; and
a controller to develop a record of what the operator looked at in the course of an operator training session and to compare the record to said information about a plurality of gaze targets to determine whether the operator looked at said gaze targets in the course of an operator training session.
2. The system of claim 1 wherein said sensor includes a camera.
3. The system of claim 2 wherein said camera is a digital camera.
4. The system of claim 3 including a first camera to locate head position and a second camera to locate eye position.
5. The system of claim 1 including a position location device.
6. The system of claim 5 wherein said position location device is a global positioning system receiver.
7. The system of claim 1 including an interface adapted to receive information from vehicle controls.
8. The system of claim 1 including an interface adapted to receive vehicle speed information.
9. The system of claim 1 including an interface adapted to receive information about the use of vehicle indicators.
10. A method for evaluating a vehicle operator comprising:
automatically sensing what the operator looks at;
developing a record of what the operator looks at in the course of an operator evaluation session;
compiling a list of gaze targets; and
comparing the record to the gaze targets to determine whether the operator looked at said gaze targets in the course of said evaluation session.
11. The method of claim 10 including receiving information about the current position of the vehicle.
12. The method of claim 10 including receiving information about the operation of vehicle controls.
13. The method of claim 10 including comparing the operator's performance to predefined standards.
14. The method of claim 10 including determining the frequency with which the operator looks at a given object.
15. The method of claim 10 including recording information about vehicle speed.
16. The method of claim 10 including prompting the operator to increase the frequency with which the operator looks at one or more objects.
17. The method of claim 16 including correlating what the operator looks at with the position the vehicle is at when the operator looked at the object.
18. An article comprising a medium that stores instructions for causing a processor based system to:
receive information about what a vehicle operator looks at;
make a record of what the operator looks at in the course of operator evaluation;
compile a list of gaze targets including items a vehicle operator may look at in the course of said operator evaluation; and
compare the record to said gaze targets to determine whether the operator looked at each of said gaze targets during the course of said operator evaluation.
19. The article of claim 18 storing instructions for causing a processor based system to receive information about the current position of the vehicle.
20. The article of claim 18 storing instructions for causing a processor based system to determine the frequency with which the operator looks at a given object.
21. The article claim 18 storing instructions for causing a processor based system to prompt the operator to increase the frequency with which the operator looks at one or more objects.
22. The article of claim 18 storing instructions for causing a processor based system to correlate what the operator looks at with the position the vehicle is at when the operator looked at the object.
23. A method for evaluating a vehicle operator comprising:
automatically sensing what the operator looks at;
prompting the operator to increase the frequency with which the operator looks at one or more objects; and
recording what the operator looks at.
24. The method of claim 23 including correlating what the operator looks at with the position the vehicle is at when the operator looked at the object.
25. A vehicle operator training system comprising:
a first vehicle mounted camera to scan the operator's head;
a second vehicle mounted camera to scan the operator's eye; and
a controller to develop a record of what the operator looked at in the course of operator training using information from said first and second cameras.
US09/258,848 1999-02-26 1999-02-26 Operator training system Expired - Lifetime US6200139B1 (en)

Priority Applications (11)

Application Number Priority Date Filing Date Title
US09/258,848 US6200139B1 (en) 1999-02-26 1999-02-26 Operator training system
TW088123109A TW440807B (en) 1999-02-26 1999-12-28 Operator training system
AU25028/00A AU2502800A (en) 1999-02-26 2000-01-12 Operator training system
PCT/US2000/000723 WO2000051097A2 (en) 1999-02-26 2000-01-12 Operator training system
CNB008043183A CN1248174C (en) 1999-02-26 2000-01-12 Operator training system
JP2000601623A JP3787493B2 (en) 1999-02-26 2000-01-12 Vehicle driver training system, vehicle driver evaluation method, and apparatus
DE60009057T DE60009057T2 (en) 1999-02-26 2000-01-12 TRAINING SYSTEM FOR OPERATORS
AT00903253T ATE262207T1 (en) 1999-02-26 2000-01-12 TRAINING SYSTEM FOR OPERATORS
KR10-2001-7010227A KR100416152B1 (en) 1999-02-26 2000-01-12 Operator training system
EP00903253A EP1177544B1 (en) 1999-02-26 2000-01-12 Operator training system
JP2005323273A JP2006163375A (en) 1999-02-26 2005-11-08 Vehicle operator training system and vehicle operator evaluating method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US09/258,848 US6200139B1 (en) 1999-02-26 1999-02-26 Operator training system

Publications (1)

Publication Number Publication Date
US6200139B1 true US6200139B1 (en) 2001-03-13

Family

ID=22982383

Family Applications (1)

Application Number Title Priority Date Filing Date
US09/258,848 Expired - Lifetime US6200139B1 (en) 1999-02-26 1999-02-26 Operator training system

Country Status (10)

Country Link
US (1) US6200139B1 (en)
EP (1) EP1177544B1 (en)
JP (2) JP3787493B2 (en)
KR (1) KR100416152B1 (en)
CN (1) CN1248174C (en)
AT (1) ATE262207T1 (en)
AU (1) AU2502800A (en)
DE (1) DE60009057T2 (en)
TW (1) TW440807B (en)
WO (1) WO2000051097A2 (en)

Cited By (59)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6397137B1 (en) * 2001-03-02 2002-05-28 International Business Machines Corporation System and method for selection of vehicular sideview mirrors via eye gaze
JP2002538498A (en) * 1999-02-26 2002-11-12 インテル コーポレイション Vehicle driver training system, vehicle driver evaluation method, and device
US20030012408A1 (en) * 2001-05-09 2003-01-16 Jean-Yves Bouguet Method and system using a data-driven model for monocular face tracking
US6612841B2 (en) * 2000-12-08 2003-09-02 Daimler Chrysler Ag Information system for providing explanatory information on functional elements or controls in motor vehicles
WO2004015653A1 (en) * 2002-08-09 2004-02-19 Best Practice Deployment As A system for computer assisted driving lessons
US20040150514A1 (en) * 2003-02-05 2004-08-05 Newman Timothy J. Vehicle situation alert system with eye gaze controlled alert signal generation
US20040167761A1 (en) * 2002-04-05 2004-08-26 Konstantin Sizov Portable in-the-vehicle road simulator
US20060200277A1 (en) * 2005-03-02 2006-09-07 Denso Corporation Driving evaluation system and server
US20070047517A1 (en) * 2005-08-29 2007-03-01 Hua Xu Method and apparatus for altering a media activity
US7245741B1 (en) * 2000-11-14 2007-07-17 Siemens Aktiengesellschaft Method and device for determining whether the interior of a vehicle is occupied
US20070219685A1 (en) * 2006-03-16 2007-09-20 James Plante Vehicle event recorders with integrated web server
US20070257815A1 (en) * 2006-05-08 2007-11-08 Drivecam, Inc. System and method for taking risk out of driving
US20070257804A1 (en) * 2006-05-08 2007-11-08 Drivecam, Inc. System and Method for Reducing Driving Risk With Foresight
US20070271078A1 (en) * 2002-06-25 2007-11-22 New York Air Brake Corporation Remote Control Locomotive Simulator
US20070271105A1 (en) * 2006-05-09 2007-11-22 Drivecam, Inc. System and Method for Reducing Driving Risk With Hindsignt
US20070280505A1 (en) * 1995-06-07 2007-12-06 Automotive Technologies International, Inc. Eye Monitoring System and Method for Vehicular Occupants
US20080048930A1 (en) * 1999-12-15 2008-02-28 Automotive Technologies International, Inc. Vehicular Heads-Up Display System
US20080111666A1 (en) * 2006-11-09 2008-05-15 Smartdrive Systems Inc. Vehicle exception event management systems
US20080158096A1 (en) * 1999-12-15 2008-07-03 Automotive Technologies International, Inc. Eye-Location Dependent Vehicular Heads-Up Display System
US20080208646A1 (en) * 2007-02-28 2008-08-28 Thompson Ralph E Method for increasing productivity and safety in the mining and heavy construction industries
US20080276191A1 (en) * 1999-12-15 2008-11-06 Automotive Technologies International, Inc. Vehicular Heads-Up Display System
US20090040196A1 (en) * 2003-08-27 2009-02-12 Bernd Duckstein Method of controlling the display of various data in a vehicle and Opto-acoustic data unit
US20090063053A1 (en) * 2007-09-04 2009-03-05 International Business Machines Corporation Method and system for blind spot identification and warning utilizing visual indicators
US20090157255A1 (en) * 2005-12-08 2009-06-18 Smart Drive Systems, Inc. Vehicle Event Recorder Systems
US20090191531A1 (en) * 2007-12-21 2009-07-30 Joseph Saccocci Method and Apparatus for Integrating Audio and/or Video With a Book
US20100160014A1 (en) * 2008-11-25 2010-06-24 Mario Galasso Methods and apparatus for virtual competition
EP2296124A1 (en) * 2008-06-06 2011-03-16 Yamashiro Driving School System for automatic evaluation of driving behavior
US20120259246A1 (en) * 2010-09-15 2012-10-11 Erbicol Sa Reflex measuring apparatus determining the reactivity of a driver and its application
US20130004920A1 (en) * 2010-03-24 2013-01-03 Krauss-Maffei Wegmann Gmbh & Co. Kg Method for Training a Crew Member of a, in Particular, Military Vehicle
US20130083197A1 (en) * 2010-05-25 2013-04-04 Fujitsu Limited Storage managing method and storage management device
US20130280678A1 (en) * 2012-04-23 2013-10-24 The Boeing Company Aircrew training system
US20130302759A1 (en) * 2009-09-29 2013-11-14 Advanced Training System Llc System, Method and Apparatus for Driver Training System with Dynamic Mirrors
US8880279B2 (en) 2005-12-08 2014-11-04 Smartdrive Systems, Inc. Memory management in event recording systems
US8892310B1 (en) 2014-02-21 2014-11-18 Smartdrive Systems, Inc. System and method to detect execution of driving maneuvers
US8989959B2 (en) 2006-11-07 2015-03-24 Smartdrive Systems, Inc. Vehicle operator performance history recording, scoring and reporting systems
US9082315B2 (en) 2012-03-08 2015-07-14 Industrial Technology Research Institute Surrounding bird view monitoring image generation method and training method, automobile-side device, and training device thereof
US20150310758A1 (en) * 2014-04-26 2015-10-29 The Travelers Indemnity Company Systems, methods, and apparatus for generating customized virtual reality experiences
US9183679B2 (en) 2007-05-08 2015-11-10 Smartdrive Systems, Inc. Distributed vehicle event recorder systems having a portable memory data transfer system
US9201842B2 (en) 2006-03-16 2015-12-01 Smartdrive Systems, Inc. Vehicle event recorder systems and networks having integrated cellular wireless communications systems
US20160107642A1 (en) * 2014-10-15 2016-04-21 Continental Automotive Systems, Inc. Collision avoidance system using driver eye monitoring during gear change
US9501878B2 (en) 2013-10-16 2016-11-22 Smartdrive Systems, Inc. Vehicle event playback apparatus and methods
US9554080B2 (en) 2006-11-07 2017-01-24 Smartdrive Systems, Inc. Power management systems for automotive video event recorders
CN106530890A (en) * 2016-02-03 2017-03-22 北京易驾佳信息科技有限公司 Intelligent driving training system and method based on vehicle posture and trajectory perception
US9610955B2 (en) 2013-11-11 2017-04-04 Smartdrive Systems, Inc. Vehicle fuel consumption monitor and feedback systems
US9639804B1 (en) 2016-03-22 2017-05-02 Smartdrive Systems, Inc. System and method to determine responsiveness of a driver of a vehicle to feedback regarding driving behaviors
US9663127B2 (en) 2014-10-28 2017-05-30 Smartdrive Systems, Inc. Rail vehicle event detection and recording system
US9728228B2 (en) 2012-08-10 2017-08-08 Smartdrive Systems, Inc. Vehicle event playback apparatus and methods
US9792319B2 (en) 2006-05-08 2017-10-17 Lytx, Inc. System and method for identifying non-event profiles
US10102773B2 (en) 2012-04-23 2018-10-16 The Boeing Company Methods for evaluating human performance in aviation
CN109102194A (en) * 2018-08-20 2018-12-28 合肥优控科技有限公司 A kind of driving behavior methods of marking based on global positioning system and inertial sensor
US20190304325A1 (en) * 2018-03-30 2019-10-03 Cae Inc. Standard operating procedures feedback during an interactive computer simulation
WO2020031949A1 (en) * 2018-08-07 2020-02-13 国立大学法人名古屋大学 Information processing device, information processing system, information processing method, and computer program
US10625745B1 (en) * 2019-01-07 2020-04-21 Sean Tremblay Automated driver's exam system
US10692392B2 (en) * 2016-11-02 2020-06-23 Kyocera Document Solutions Inc. Display apparatus and method for displaying guidance video
US10699556B1 (en) 2019-03-01 2020-06-30 Honeywell International Inc. System and method for plant operation gap analysis and guidance solution
US10930093B2 (en) 2015-04-01 2021-02-23 Smartdrive Systems, Inc. Vehicle event recording system and method
US11037461B2 (en) 2009-09-29 2021-06-15 Advance Training Systems LLC System, method and apparatus for adaptive driver training
US11069257B2 (en) 2014-11-13 2021-07-20 Smartdrive Systems, Inc. System and method for detecting a vehicle event and generating review criteria
US11875707B2 (en) 2009-09-29 2024-01-16 Advanced Training Systems, Inc. System, method and apparatus for adaptive driver training

Families Citing this family (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7565230B2 (en) * 2000-10-14 2009-07-21 Temic Automotive Of North America, Inc. Method and apparatus for improving vehicle operator performance
CN1294510C (en) * 2003-01-28 2007-01-10 中国南方航空股份有限公司 Computer-aided teaching system and method for stimulated aviation training
JP2005208478A (en) * 2004-01-26 2005-08-04 Ota Hiroo Simulation driving apparatus
JP4565445B2 (en) * 2004-03-18 2010-10-20 国立大学法人 奈良先端科学技術大学院大学 Face information measurement system
JP4540507B2 (en) * 2005-03-07 2010-09-08 株式会社日立国際電気 Behavior training content display method
JP4682714B2 (en) * 2005-06-14 2011-05-11 トヨタ自動車株式会社 Dialog system
JP2007264490A (en) * 2006-03-29 2007-10-11 National Univ Corp Shizuoka Univ Information presentation evaluation system
CN100483478C (en) * 2006-09-06 2009-04-29 李宏 Training system for imitating operate of excavator
JP2008139553A (en) * 2006-12-01 2008-06-19 National Agency For Automotive Safety & Victim's Aid Driving aptitude diagnosing method, evaluation standard determining method for driving aptitude diagnosis, and driving aptitude diagnostic program
KR100898417B1 (en) * 2007-06-08 2009-06-19 샬롬엔지니어링 주식회사 A simulation system for operating train
WO2008151942A2 (en) * 2007-06-14 2008-12-18 Driving Performance Solutions As System for performing driver education
EP3333831A1 (en) * 2007-07-13 2018-06-13 Volvo Construction Equipment AB A work machine
JP2009107373A (en) * 2007-10-26 2009-05-21 Equos Research Co Ltd Alarm sound output device
AU2010295239B2 (en) 2009-09-16 2015-04-23 Sydac Pty Ltd Visual presentation system
KR101364080B1 (en) 2009-12-17 2014-02-20 한국전자통신연구원 Method and apparatus for evaluating safe driving
WO2011126379A1 (en) * 2010-04-04 2011-10-13 Driving Performance Solutions As A system and method for drivers training
EP3527121B1 (en) * 2011-02-09 2023-08-23 Apple Inc. Gesture detection in a 3d mapping environment
US9377863B2 (en) 2012-03-26 2016-06-28 Apple Inc. Gaze-enhanced virtual touchscreen
JP5946349B2 (en) * 2012-07-24 2016-07-06 日本無線株式会社 Driving simulator
DE102012016765A1 (en) * 2012-08-27 2014-02-27 Dr. Ing. H.C. F. Porsche Ag Motor car simulation system for simulating and optimizing outer mirror to guide motor car, has simulation device virtually and three-dimensionally simulating mirror and image visible in mirror based on parameters of situation and mirror
DE102012220321A1 (en) * 2012-11-08 2014-06-12 Bayerische Motoren Werke Aktiengesellschaft Method for demonstrating driver assistance system for avoiding accidents of motor vehicle, involves displaying animated virtual traffic situation on display and detecting that driver gives no attention to traffic situation
CN103854539A (en) * 2012-12-07 2014-06-11 永泰软件有限公司 Electronic road test monitoring system based on video analysis
US9619695B2 (en) * 2013-06-06 2017-04-11 Visteon Global Technologies, Inc. Gaze time indicator for a vehicle
CN105788401B (en) * 2016-04-20 2018-09-28 公安部交通管理科学研究所 Defensive drive simulation based on real vehicle body experiences training system
US10083620B2 (en) * 2016-07-29 2018-09-25 Nissan North America, Inc. Smart tutorial that learns and adapts
SE2030301A1 (en) * 2020-09-30 2022-03-31 Chadalavada Ravi Teja Method and system for driving skill feedback
FR3123491A1 (en) * 2021-05-31 2022-12-02 Psa Automobiles Sa Car driving simulator

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5138555A (en) 1990-06-28 1992-08-11 Albrecht Robert E Helmet mounted display adaptive predictive tracking
EP0596868A2 (en) 1988-07-14 1994-05-11 Atr Communication Systems Research Laboratories Eye tracking method using an image pickup apparatus
US5373857A (en) * 1993-06-18 1994-12-20 Forte Technologies, Inc. Head tracking apparatus
EP0660291A1 (en) 1993-12-23 1995-06-28 Thomson-Csf Visual simulating device for a vehicle with monoscopic parallax
US5471542A (en) 1993-09-27 1995-11-28 Ragland; Richard R. Point-of-gaze tracker
US5583795A (en) * 1995-03-17 1996-12-10 The United States Of America As Represented By The Secretary Of The Army Apparatus for measuring eye gaze and fixation duration, and method therefor
US5764291A (en) * 1994-09-30 1998-06-09 Apple Computer, Inc. Apparatus and method for orientation-dependent camera exposure and focus setting optimization
US5839000A (en) * 1997-11-10 1998-11-17 Sharp Laboratories Of America, Inc. Automatic zoom magnification control using detection of eyelid condition
EP0990416A1 (en) 1998-10-01 2000-04-05 Mitsubishi Denki Kabushiki Kaisha System for classifying an individual's gaze direction

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0431172U (en) * 1990-07-11 1992-03-12
JPH04129173U (en) * 1991-05-13 1992-11-25 三菱重工業株式会社 Visibility target acquisition and tracking evaluation device
JPH06189906A (en) * 1992-12-26 1994-07-12 Nissan Motor Co Ltd Visual axial direction measuring device
JPH07141095A (en) * 1993-11-17 1995-06-02 Matsushita Electric Ind Co Ltd Training device
JPH0736299U (en) * 1993-12-03 1995-07-04 昌久 時山 Drowsiness prevention device for vehicles
JP2993345B2 (en) * 1993-12-29 1999-12-20 日産自動車株式会社 Display device for vehicles
JP3134667B2 (en) * 1994-06-02 2001-02-13 日産自動車株式会社 Display device for vehicles
JP3119076B2 (en) * 1994-06-02 2000-12-18 日産自動車株式会社 Warning display device for vehicles
JP3257310B2 (en) * 1994-12-22 2002-02-18 トヨタ自動車株式会社 Inattentive driving detection device
JPH1097179A (en) * 1996-09-20 1998-04-14 Sanyo Electric Works Ltd Running information recording device, running information analyzer, and vehicle training system using both devices
US6200139B1 (en) * 1999-02-26 2001-03-13 Intel Corporation Operator training system

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0596868A2 (en) 1988-07-14 1994-05-11 Atr Communication Systems Research Laboratories Eye tracking method using an image pickup apparatus
US5138555A (en) 1990-06-28 1992-08-11 Albrecht Robert E Helmet mounted display adaptive predictive tracking
US5373857A (en) * 1993-06-18 1994-12-20 Forte Technologies, Inc. Head tracking apparatus
US5471542A (en) 1993-09-27 1995-11-28 Ragland; Richard R. Point-of-gaze tracker
EP0660291A1 (en) 1993-12-23 1995-06-28 Thomson-Csf Visual simulating device for a vehicle with monoscopic parallax
US5764291A (en) * 1994-09-30 1998-06-09 Apple Computer, Inc. Apparatus and method for orientation-dependent camera exposure and focus setting optimization
US5583795A (en) * 1995-03-17 1996-12-10 The United States Of America As Represented By The Secretary Of The Army Apparatus for measuring eye gaze and fixation duration, and method therefor
US5839000A (en) * 1997-11-10 1998-11-17 Sharp Laboratories Of America, Inc. Automatic zoom magnification control using detection of eyelid condition
EP0990416A1 (en) 1998-10-01 2000-04-05 Mitsubishi Denki Kabushiki Kaisha System for classifying an individual's gaze direction

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Determination of the Point of Fixation in a Head Fixed Coordinate System, by Liu, prototype released 1997 and published Nov. 28, 1998. *
Eye Gaze Tracking for Visually Controlled Interactive Stereoscopic Displays by Talmi et al., published Nov. 28, 1998. *

Cited By (122)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070280505A1 (en) * 1995-06-07 2007-12-06 Automotive Technologies International, Inc. Eye Monitoring System and Method for Vehicular Occupants
US7788008B2 (en) * 1995-06-07 2010-08-31 Automotive Technologies International, Inc. Eye monitoring system and method for vehicular occupants
JP2002538498A (en) * 1999-02-26 2002-11-12 インテル コーポレイション Vehicle driver training system, vehicle driver evaluation method, and device
US20080048930A1 (en) * 1999-12-15 2008-02-28 Automotive Technologies International, Inc. Vehicular Heads-Up Display System
US8686922B2 (en) 1999-12-15 2014-04-01 American Vehicular Sciences Llc Eye-location dependent vehicular heads-up display system
US20080276191A1 (en) * 1999-12-15 2008-11-06 Automotive Technologies International, Inc. Vehicular Heads-Up Display System
US20080158096A1 (en) * 1999-12-15 2008-07-03 Automotive Technologies International, Inc. Eye-Location Dependent Vehicular Heads-Up Display System
US8032264B2 (en) 1999-12-15 2011-10-04 Automotive Technologies International, Inc. Vehicular heads-up display system
US8818647B2 (en) 1999-12-15 2014-08-26 American Vehicular Sciences Llc Vehicular heads-up display system
US7245741B1 (en) * 2000-11-14 2007-07-17 Siemens Aktiengesellschaft Method and device for determining whether the interior of a vehicle is occupied
US6612841B2 (en) * 2000-12-08 2003-09-02 Daimler Chrysler Ag Information system for providing explanatory information on functional elements or controls in motor vehicles
US6397137B1 (en) * 2001-03-02 2002-05-28 International Business Machines Corporation System and method for selection of vehicular sideview mirrors via eye gaze
US20030012408A1 (en) * 2001-05-09 2003-01-16 Jean-Yves Bouguet Method and system using a data-driven model for monocular face tracking
US9400921B2 (en) * 2001-05-09 2016-07-26 Intel Corporation Method and system using a data-driven model for monocular face tracking
US20110131026A1 (en) * 2002-04-05 2011-06-02 Konstantin Sizov Portable in-the-vehicle road simulator
US10217374B2 (en) 2002-04-05 2019-02-26 Konstantin Sizov Portable in-the-vehicle road simulator
US8412499B2 (en) 2002-04-05 2013-04-02 Konstantin Sizov Portable in-the-vehicle road simulator
US20040167761A1 (en) * 2002-04-05 2004-08-26 Konstantin Sizov Portable in-the-vehicle road simulator
US20070271078A1 (en) * 2002-06-25 2007-11-22 New York Air Brake Corporation Remote Control Locomotive Simulator
WO2004015653A1 (en) * 2002-08-09 2004-02-19 Best Practice Deployment As A system for computer assisted driving lessons
US20060057543A1 (en) * 2002-08-09 2006-03-16 Best Practice Deployment As System for computer assisted driving lessons
US6859144B2 (en) 2003-02-05 2005-02-22 Delphi Technologies, Inc. Vehicle situation alert system with eye gaze controlled alert signal generation
US20040150514A1 (en) * 2003-02-05 2004-08-05 Newman Timothy J. Vehicle situation alert system with eye gaze controlled alert signal generation
US20090040196A1 (en) * 2003-08-27 2009-02-12 Bernd Duckstein Method of controlling the display of various data in a vehicle and Opto-acoustic data unit
US8010283B2 (en) 2005-03-02 2011-08-30 Denso Corporation Driving evaluation system and server
US20060200277A1 (en) * 2005-03-02 2006-09-07 Denso Corporation Driving evaluation system and server
US20070047517A1 (en) * 2005-08-29 2007-03-01 Hua Xu Method and apparatus for altering a media activity
US9633318B2 (en) 2005-12-08 2017-04-25 Smartdrive Systems, Inc. Vehicle event recorder systems
US8880279B2 (en) 2005-12-08 2014-11-04 Smartdrive Systems, Inc. Memory management in event recording systems
US20090157255A1 (en) * 2005-12-08 2009-06-18 Smart Drive Systems, Inc. Vehicle Event Recorder Systems
US9226004B1 (en) 2005-12-08 2015-12-29 Smartdrive Systems, Inc. Memory management in event recording systems
US9911253B2 (en) 2005-12-08 2018-03-06 Smartdrive Systems, Inc. Memory management in event recording systems
US10878646B2 (en) 2005-12-08 2020-12-29 Smartdrive Systems, Inc. Vehicle event recorder systems
US9402060B2 (en) 2006-03-16 2016-07-26 Smartdrive Systems, Inc. Vehicle event recorders with integrated web server
US9942526B2 (en) 2006-03-16 2018-04-10 Smartdrive Systems, Inc. Vehicle event recorders with integrated web server
US9566910B2 (en) 2006-03-16 2017-02-14 Smartdrive Systems, Inc. Vehicle event recorder systems and networks having integrated cellular wireless communications systems
US9472029B2 (en) 2006-03-16 2016-10-18 Smartdrive Systems, Inc. Vehicle event recorder systems and networks having integrated cellular wireless communications systems
US20070219685A1 (en) * 2006-03-16 2007-09-20 James Plante Vehicle event recorders with integrated web server
US9691195B2 (en) 2006-03-16 2017-06-27 Smartdrive Systems, Inc. Vehicle event recorder systems and networks having integrated cellular wireless communications systems
US9208129B2 (en) 2006-03-16 2015-12-08 Smartdrive Systems, Inc. Vehicle event recorder systems and networks having integrated cellular wireless communications systems
US10404951B2 (en) 2006-03-16 2019-09-03 Smartdrive Systems, Inc. Vehicle event recorders with integrated web server
US9201842B2 (en) 2006-03-16 2015-12-01 Smartdrive Systems, Inc. Vehicle event recorder systems and networks having integrated cellular wireless communications systems
US9545881B2 (en) 2006-03-16 2017-01-17 Smartdrive Systems, Inc. Vehicle event recorder systems and networks having integrated cellular wireless communications systems
US8996240B2 (en) 2006-03-16 2015-03-31 Smartdrive Systems, Inc. Vehicle event recorders with integrated web server
US20070257804A1 (en) * 2006-05-08 2007-11-08 Drivecam, Inc. System and Method for Reducing Driving Risk With Foresight
US9792319B2 (en) 2006-05-08 2017-10-17 Lytx, Inc. System and method for identifying non-event profiles
US8314708B2 (en) * 2006-05-08 2012-11-20 Drivecam, Inc. System and method for reducing driving risk with foresight
US7659827B2 (en) * 2006-05-08 2010-02-09 Drivecam, Inc. System and method for taking risk out of driving
US20070257815A1 (en) * 2006-05-08 2007-11-08 Drivecam, Inc. System and method for taking risk out of driving
US20070271105A1 (en) * 2006-05-09 2007-11-22 Drivecam, Inc. System and Method for Reducing Driving Risk With Hindsignt
US10235655B2 (en) 2006-05-09 2019-03-19 Lytx, Inc. System and method for reducing driving risk with hindsight
US9836716B2 (en) 2006-05-09 2017-12-05 Lytx, Inc. System and method for reducing driving risk with hindsight
US9761067B2 (en) 2006-11-07 2017-09-12 Smartdrive Systems, Inc. Vehicle operator performance history recording, scoring and reporting systems
US10682969B2 (en) 2006-11-07 2020-06-16 Smartdrive Systems, Inc. Power management systems for automotive video event recorders
US10339732B2 (en) 2006-11-07 2019-07-02 Smartdrive Systems, Inc. Vehicle operator performance history recording, scoring and reporting systems
US9554080B2 (en) 2006-11-07 2017-01-24 Smartdrive Systems, Inc. Power management systems for automotive video event recorders
US10053032B2 (en) 2006-11-07 2018-08-21 Smartdrive Systems, Inc. Power management systems for automotive video event recorders
US8989959B2 (en) 2006-11-07 2015-03-24 Smartdrive Systems, Inc. Vehicle operator performance history recording, scoring and reporting systems
US10471828B2 (en) 2006-11-09 2019-11-12 Smartdrive Systems, Inc. Vehicle exception event management systems
US20080111666A1 (en) * 2006-11-09 2008-05-15 Smartdrive Systems Inc. Vehicle exception event management systems
US9738156B2 (en) 2006-11-09 2017-08-22 Smartdrive Systems, Inc. Vehicle exception event management systems
US11623517B2 (en) 2006-11-09 2023-04-11 SmartDriven Systems, Inc. Vehicle exception event management systems
US8868288B2 (en) 2006-11-09 2014-10-21 Smartdrive Systems, Inc. Vehicle exception event management systems
US20080208646A1 (en) * 2007-02-28 2008-08-28 Thompson Ralph E Method for increasing productivity and safety in the mining and heavy construction industries
US9679424B2 (en) 2007-05-08 2017-06-13 Smartdrive Systems, Inc. Distributed vehicle event recorder systems having a portable memory data transfer system
US9183679B2 (en) 2007-05-08 2015-11-10 Smartdrive Systems, Inc. Distributed vehicle event recorder systems having a portable memory data transfer system
US20090063053A1 (en) * 2007-09-04 2009-03-05 International Business Machines Corporation Method and system for blind spot identification and warning utilizing visual indicators
US8645001B2 (en) * 2007-09-04 2014-02-04 International Business Machines Corporation Method and system for blind spot identification and warning utilizing visual indicators
US20090191531A1 (en) * 2007-12-21 2009-07-30 Joseph Saccocci Method and Apparatus for Integrating Audio and/or Video With a Book
EP2296124A1 (en) * 2008-06-06 2011-03-16 Yamashiro Driving School System for automatic evaluation of driving behavior
EP2296124A4 (en) * 2008-06-06 2011-08-10 Yamashiro Driving School System for automatic evaluation of driving behavior
US20100160014A1 (en) * 2008-11-25 2010-06-24 Mario Galasso Methods and apparatus for virtual competition
US9108098B2 (en) * 2008-11-25 2015-08-18 Fox Factory, Inc. Methods and apparatus for virtual competition
US10325512B2 (en) 2009-09-29 2019-06-18 Advanced Training System Llc System, method and apparatus for driver training system with dynamic mirrors
US20130302759A1 (en) * 2009-09-29 2013-11-14 Advanced Training System Llc System, Method and Apparatus for Driver Training System with Dynamic Mirrors
US9418568B2 (en) * 2009-09-29 2016-08-16 Advanced Training System Llc System, method and apparatus for driver training system with dynamic mirrors
US11037461B2 (en) 2009-09-29 2021-06-15 Advance Training Systems LLC System, method and apparatus for adaptive driver training
US11263916B2 (en) 2009-09-29 2022-03-01 Advanced Training System Llc System, method and apparatus for adaptive driver training
US11875707B2 (en) 2009-09-29 2024-01-16 Advanced Training Systems, Inc. System, method and apparatus for adaptive driver training
US20160321941A1 (en) * 2009-09-29 2016-11-03 Advanced Training System Llc System, Method and Apparatus for Driver Training System with Dynamic Mirrors
US20130004920A1 (en) * 2010-03-24 2013-01-03 Krauss-Maffei Wegmann Gmbh & Co. Kg Method for Training a Crew Member of a, in Particular, Military Vehicle
US9202385B2 (en) * 2010-03-24 2015-12-01 Krauss-Maffei Wegmann Gmbh & Co. Kg Method for training a crew member of a, in particular, military vehicle
US9787946B2 (en) * 2010-05-25 2017-10-10 Fujitsu Limited Picture processing device and method
US20130083197A1 (en) * 2010-05-25 2013-04-04 Fujitsu Limited Storage managing method and storage management device
US20120259246A1 (en) * 2010-09-15 2012-10-11 Erbicol Sa Reflex measuring apparatus determining the reactivity of a driver and its application
US9082315B2 (en) 2012-03-08 2015-07-14 Industrial Technology Research Institute Surrounding bird view monitoring image generation method and training method, automobile-side device, and training device thereof
US10102773B2 (en) 2012-04-23 2018-10-16 The Boeing Company Methods for evaluating human performance in aviation
US20130280678A1 (en) * 2012-04-23 2013-10-24 The Boeing Company Aircrew training system
US9728228B2 (en) 2012-08-10 2017-08-08 Smartdrive Systems, Inc. Vehicle event playback apparatus and methods
US10818112B2 (en) 2013-10-16 2020-10-27 Smartdrive Systems, Inc. Vehicle event playback apparatus and methods
US10019858B2 (en) 2013-10-16 2018-07-10 Smartdrive Systems, Inc. Vehicle event playback apparatus and methods
US9501878B2 (en) 2013-10-16 2016-11-22 Smartdrive Systems, Inc. Vehicle event playback apparatus and methods
US9610955B2 (en) 2013-11-11 2017-04-04 Smartdrive Systems, Inc. Vehicle fuel consumption monitor and feedback systems
US11260878B2 (en) 2013-11-11 2022-03-01 Smartdrive Systems, Inc. Vehicle fuel consumption monitor and feedback systems
US11884255B2 (en) 2013-11-11 2024-01-30 Smartdrive Systems, Inc. Vehicle fuel consumption monitor and feedback systems
US10249105B2 (en) 2014-02-21 2019-04-02 Smartdrive Systems, Inc. System and method to detect execution of driving maneuvers
US9594371B1 (en) 2014-02-21 2017-03-14 Smartdrive Systems, Inc. System and method to detect execution of driving maneuvers
US10497187B2 (en) 2014-02-21 2019-12-03 Smartdrive Systems, Inc. System and method to detect execution of driving maneuvers
US11734964B2 (en) 2014-02-21 2023-08-22 Smartdrive Systems, Inc. System and method to detect execution of driving maneuvers
US8892310B1 (en) 2014-02-21 2014-11-18 Smartdrive Systems, Inc. System and method to detect execution of driving maneuvers
US11250649B2 (en) 2014-02-21 2022-02-15 Smartdrive Systems, Inc. System and method to detect execution of driving maneuvers
US9953470B1 (en) 2014-02-21 2018-04-24 Smartdrive Systems, Inc. System and method to detect execution of driving maneuvers
US20150310758A1 (en) * 2014-04-26 2015-10-29 The Travelers Indemnity Company Systems, methods, and apparatus for generating customized virtual reality experiences
US20160107642A1 (en) * 2014-10-15 2016-04-21 Continental Automotive Systems, Inc. Collision avoidance system using driver eye monitoring during gear change
CN105523043A (en) * 2014-10-15 2016-04-27 大陆汽车系统公司 Collision avoidance system using driver eye monitoring during gear change
US9663127B2 (en) 2014-10-28 2017-05-30 Smartdrive Systems, Inc. Rail vehicle event detection and recording system
US11069257B2 (en) 2014-11-13 2021-07-20 Smartdrive Systems, Inc. System and method for detecting a vehicle event and generating review criteria
US10930093B2 (en) 2015-04-01 2021-02-23 Smartdrive Systems, Inc. Vehicle event recording system and method
CN106530890A (en) * 2016-02-03 2017-03-22 北京易驾佳信息科技有限公司 Intelligent driving training system and method based on vehicle posture and trajectory perception
CN106530890B (en) * 2016-02-03 2018-07-06 北京易驾佳信息科技有限公司 A kind of intelligent driving training system perceived based on vehicle pose and track and method
US10793160B2 (en) 2016-03-22 2020-10-06 Smartdrive Systems, Inc. System and method to determine responsiveness of a driver of a vehicle to feedback regarding driving behaviors
US11254317B2 (en) 2016-03-22 2022-02-22 Smartdrive Systems, Inc. System and method to determine responsiveness of a driver of a vehicle to feedback regarding driving behaviors
US9639804B1 (en) 2016-03-22 2017-05-02 Smartdrive Systems, Inc. System and method to determine responsiveness of a driver of a vehicle to feedback regarding driving behaviors
US11731636B2 (en) 2016-03-22 2023-08-22 Smartdrive Systems, Inc. System and method to determine responsiveness of a driver of a vehicle to feedback regarding driving behaviors
US10692392B2 (en) * 2016-11-02 2020-06-23 Kyocera Document Solutions Inc. Display apparatus and method for displaying guidance video
US20190304325A1 (en) * 2018-03-30 2019-10-03 Cae Inc. Standard operating procedures feedback during an interactive computer simulation
US11715387B2 (en) * 2018-03-30 2023-08-01 Cae Inc. Standard operating procedures feedback during an interactive computer simulation
JP2020024278A (en) * 2018-08-07 2020-02-13 国立大学法人名古屋大学 Information processing device, information processing system, information processing method, and computer program
WO2020031949A1 (en) * 2018-08-07 2020-02-13 国立大学法人名古屋大学 Information processing device, information processing system, information processing method, and computer program
CN109102194A (en) * 2018-08-20 2018-12-28 合肥优控科技有限公司 A kind of driving behavior methods of marking based on global positioning system and inertial sensor
US10625745B1 (en) * 2019-01-07 2020-04-21 Sean Tremblay Automated driver's exam system
US10699556B1 (en) 2019-03-01 2020-06-30 Honeywell International Inc. System and method for plant operation gap analysis and guidance solution

Also Published As

Publication number Publication date
KR20010102090A (en) 2001-11-15
TW440807B (en) 2001-06-16
CN1248174C (en) 2006-03-29
DE60009057D1 (en) 2004-04-22
WO2000051097A3 (en) 2001-11-29
AU2502800A (en) 2000-09-14
ATE262207T1 (en) 2004-04-15
JP2002538498A (en) 2002-11-12
DE60009057T2 (en) 2004-08-05
KR100416152B1 (en) 2004-01-24
JP3787493B2 (en) 2006-06-21
WO2000051097A2 (en) 2000-08-31
EP1177544B1 (en) 2004-03-17
CN1470044A (en) 2004-01-21
EP1177544A2 (en) 2002-02-06
JP2006163375A (en) 2006-06-22

Similar Documents

Publication Publication Date Title
US6200139B1 (en) Operator training system
Braunagel et al. Ready for take-over? A new driver assistance system for an automated classification of driver take-over readiness
US10705600B2 (en) Method and apparatus for detecting and following an eye and/or the gaze direction thereof
CN105741643B (en) Automatic implementation method, device and system for training driving technology
US7162054B2 (en) Augmented reality technology
US5781505A (en) System and method for locating a trajectory and a source of a projectile
CN102510480B (en) Automatic calibrating and tracking system of driver sight line
Jiménez et al. Gaze fixation system for the evaluation of driver distractions induced by IVIS
US20110260965A1 (en) Apparatus and method of user interface for manipulating multimedia contents in vehicle
CN109849788B (en) Information providing method, device and system
US10325512B2 (en) System, method and apparatus for driver training system with dynamic mirrors
US20170213474A1 (en) System, Method and Apparatus for Driver Training System with Stress Management
EP3654143B1 (en) Method, apparatus and system for determining line of sight, and wearable eye movement device
CN107380064B (en) Vehicle-mounted sight tracking device based on augmented reality technology
US20200005663A1 (en) System, Method and Apparatus for Driver Training System with Stress Management
JPH07505308A (en) Perspective motility test method and device
Guasconi et al. A low-cost implementation of an eye tracking system for driver's gaze analysis
US9798941B2 (en) Driver visual sensor behavior study device
Mackenzie et al. Characterizing visual attention during driving and non-driving hazard perception tasks in a simulated environment
CN113767026A (en) Method for operating a motor vehicle
US20190339786A1 (en) Data processing device, monitoring system, awakening system, data processing method, and data processing program
Hamada et al. Detecting method applicable to individual features for drivers' drowsiness
US20210295731A1 (en) Information processing apparatus, information processing system, information processing method, and computer program
JP2988235B2 (en) Gaze direction detection device for vehicles
RU2425331C2 (en) Method to process data to define visual figures on visual stage

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTEL CORPORATION, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:CLAPPER, EDWARD C.;REEL/FRAME:009794/0971

Effective date: 19990219

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FPAY Fee payment

Year of fee payment: 4

FEPP Fee payment procedure

Free format text: PAYER NUMBER DE-ASSIGNED (ORIGINAL EVENT CODE: RMPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FPAY Fee payment

Year of fee payment: 8

FPAY Fee payment

Year of fee payment: 12