US20160039285A1 - Scene awareness system for a vehicle - Google Patents

Scene awareness system for a vehicle Download PDF

Info

Publication number
US20160039285A1
US20160039285A1 US14/920,420 US201514920420A US2016039285A1 US 20160039285 A1 US20160039285 A1 US 20160039285A1 US 201514920420 A US201514920420 A US 201514920420A US 2016039285 A1 US2016039285 A1 US 2016039285A1
Authority
US
United States
Prior art keywords
user
vehicle
scene
data pertaining
gaze
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/920,420
Inventor
Roy J. Mathieu
Joseph F. Szczerba
Micah R. Jones
Thomas A. Seder
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
GM Global Technology Operations LLC
Original Assignee
GM Global Technology Operations LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by GM Global Technology Operations LLC filed Critical GM Global Technology Operations LLC
Priority to US14/920,420 priority Critical patent/US20160039285A1/en
Assigned to GM Global Technology Operations LLC reassignment GM Global Technology Operations LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SEDER, THOMAS A., JONES, MICAH R., MATHIEU, ROY J., SZCZERBA, JOSEPH F.
Publication of US20160039285A1 publication Critical patent/US20160039285A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60KARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
    • B60K35/00Arrangement of adaptations of instruments
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60KARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
    • B60K28/00Safety devices for propulsion-unit control, specially adapted for, or arranged in, vehicles, e.g. preventing fuel supply or ignition in the event of potentially dangerous conditions
    • B60K28/02Safety devices for propulsion-unit control, specially adapted for, or arranged in, vehicles, e.g. preventing fuel supply or ignition in the event of potentially dangerous conditions responsive to conditions relating to the driver
    • B60K28/06Safety devices for propulsion-unit control, specially adapted for, or arranged in, vehicles, e.g. preventing fuel supply or ignition in the event of potentially dangerous conditions responsive to conditions relating to the driver responsive to incapacity of driver
    • B60K28/066Safety devices for propulsion-unit control, specially adapted for, or arranged in, vehicles, e.g. preventing fuel supply or ignition in the event of potentially dangerous conditions responsive to conditions relating to the driver responsive to incapacity of driver actuating a signalling device
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/08Interaction between the driver and the control system
    • B60W50/14Means for informing the driver, warning the driver or prompting a driver intervention
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B3/00Apparatus for testing the eyes; Instruments for examining the eyes
    • A61B3/10Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions
    • A61B3/113Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for determining or recording eye movement
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/02Detecting, measuring or recording pulse, heart rate, blood pressure or blood flow; Combined pulse/heart-rate/blood pressure determination; Evaluating a cardiovascular condition not otherwise provided for, e.g. using combinations of techniques provided for in this group with electrocardiography or electroauscultation; Heart catheters for measuring blood pressure
    • A61B5/0205Simultaneously evaluating both cardiovascular conditions and different types of body conditions, e.g. heart and respiratory condition
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/68Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
    • A61B5/6887Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient mounted on external non-worn devices, e.g. non-medical devices
    • A61B5/6893Cars
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60CVEHICLE TYRES; TYRE INFLATION; TYRE CHANGING; CONNECTING VALVES TO INFLATABLE ELASTIC BODIES IN GENERAL; DEVICES OR ARRANGEMENTS RELATED TO TYRES
    • B60C9/00Reinforcements or ply arrangement of pneumatic tyres
    • B60K35/23
    • B60K35/28
    • B60K35/29
    • B60K35/60
    • B60K35/81
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q9/00Arrangement or adaptation of signal devices not provided for in one of main groups B60Q1/00 - B60Q7/00, e.g. haptic signalling
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/09Beam shaping, e.g. changing the cross-sectional area, not otherwise provided for
    • G02B27/0927Systems for changing the beam intensity distribution, e.g. Gaussian to top-hat
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/09Beam shaping, e.g. changing the cross-sectional area, not otherwise provided for
    • G02B27/0938Using specific optical elements
    • G02B27/095Refractive optical elements
    • G02B27/0955Lenses
    • G02B27/0961Lens arrays
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B6/00Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
    • G02B6/0001Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings specially adapted for lighting devices or systems
    • G02B6/0011Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings specially adapted for lighting devices or systems the light guides being planar or of plate-like form
    • G02B6/0033Means for improving the coupling-out of light from the light guide
    • G02B6/005Means for improving the coupling-out of light from the light guide provided by one optical element, or plurality thereof, placed on the light output side of the light guide
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D3/00Control of position or direction
    • G06K9/00845
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/59Context or environment of the image inside of a vehicle, e.g. relating to seat occupancy, driver state or inner lighting conditions
    • G06V20/597Recognising the driver's state or behaviour, e.g. attention or drowsiness
    • GPHYSICS
    • G07CHECKING-DEVICES
    • G07CTIME OR ATTENDANCE REGISTERS; REGISTERING OR INDICATING THE WORKING OF MACHINES; GENERATING RANDOM NUMBERS; VOTING OR LOTTERY APPARATUS; ARRANGEMENTS, SYSTEMS OR APPARATUS FOR CHECKING NOT PROVIDED FOR ELSEWHERE
    • G07C5/00Registering or indicating the working of vehicles
    • G07C5/02Registering or indicating driving, working, idle, or waiting time only
    • GPHYSICS
    • G07CHECKING-DEVICES
    • G07CTIME OR ATTENDANCE REGISTERS; REGISTERING OR INDICATING THE WORKING OF MACHINES; GENERATING RANDOM NUMBERS; VOTING OR LOTTERY APPARATUS; ARRANGEMENTS, SYSTEMS OR APPARATUS FOR CHECKING NOT PROVIDED FOR ELSEWHERE
    • G07C5/00Registering or indicating the working of vehicles
    • G07C5/08Registering or indicating performance data other than driving, working, idle, or waiting time, with or without registering driving, working, idle or waiting time
    • GPHYSICS
    • G08SIGNALLING
    • G08CTRANSMISSION SYSTEMS FOR MEASURED VALUES, CONTROL OR SIMILAR SIGNALS
    • G08C17/00Arrangements for transmitting signals characterised by the use of a wireless electrical link
    • G08C17/02Arrangements for transmitting signals characterised by the use of a wireless electrical link using a radio link
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/02Detecting, measuring or recording pulse, heart rate, blood pressure or blood flow; Combined pulse/heart-rate/blood pressure determination; Evaluating a cardiovascular condition not otherwise provided for, e.g. using combinations of techniques provided for in this group with electrocardiography or electroauscultation; Heart catheters for measuring blood pressure
    • A61B5/024Detecting, measuring or recording pulse rate or heart rate
    • A61B5/02405Determining heart rate variability
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/05Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves 
    • A61B5/053Measuring electrical impedance or conductance of a portion of the body
    • A61B5/0531Measuring skin impedance
    • A61B5/0533Measuring galvanic skin response
    • B60K2360/177
    • B60K2360/197
    • B60K2360/21
    • B60K2360/25
    • B60K2360/334
    • B60K2360/785
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R2300/00Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
    • B60R2300/80Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement
    • B60R2300/8006Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement for monitoring and displaying scenes of vehicle interior, e.g. for monitoring passengers or cargo
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R2300/00Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
    • B60R2300/80Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement
    • B60R2300/8093Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement for obstacle warning
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/08Interaction between the driver and the control system
    • B60W50/14Means for informing the driver, warning the driver or prompting a driver intervention
    • B60W2050/146Display means
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60YINDEXING SCHEME RELATING TO ASPECTS CROSS-CUTTING VEHICLE TECHNOLOGY
    • B60Y2302/00Responses or measures related to driver conditions
    • B60Y2302/03Actuating a signal or alarm device
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/0112Head-up displays characterised by optical features comprising device for genereting colour display
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/0118Head-up displays characterised by optical features comprising devices for improving the contrast of the display / brillance control visibility
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/0141Head-up displays characterised by optical features characterised by the informative content of the display
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B2027/0192Supplementary details
    • G02B2027/0194Supplementary details with combiner of laminated type, for optical or mechanical aspects

Definitions

  • the present disclosure is related to an adaptive external road scene transfer to improve driver awareness.
  • Vehicles such as cars, typically include displays or indicators to provide information to the vehicle user.
  • Such displays or indicators may, for example, provide information regarding mileage, fuel consumption, and vehicle speed.
  • the vehicle user usually has to shift his eye gaze away from the road scene and onto an in-vehicle display in order to visually process the information presented by these displays or indicators.
  • One possible aspect of the disclosure provides a method of alerting a user of a vehicle as to a scene, external to the vehicle, which includes capturing data pertaining to the scene, external to the vehicle.
  • the captured data pertaining to the scene external to the vehicle, is transmitted to a processor.
  • the determination is made, in the processor that a characteristic is in the scene that requires the user's visual attention.
  • Data pertaining to a gaze of the user is captured.
  • Captured data, pertaining to the gaze of the user is transmitted to the processor.
  • a determination is made, in the processor that the user is gazing toward a user device.
  • a signal is transmitted to the user device such that the user device alerts the user as to the characteristic in the scene requiring the user's visual attention.
  • a scene awareness system for a vehicle that includes at least one exterior camera, at least one interior camera, and a vehicle controller.
  • the exterior camera is configured to capture data pertaining to a scene, external to the vehicle.
  • the interior camera is configured to capture data pertaining to an orientation of a gaze of a user of the vehicle.
  • the vehicle controller is in communication with each of the at least one exterior camera and the at least one interior camera.
  • the vehicle controller is configured to determine whether at least one characteristic in the scene, exterior to the vehicle, requires the user's visual attention.
  • the vehicle controller is also configured to determine whether the user is gazing at a user device.
  • the vehicle controller is configured to transmit a signal to the user device such that the user device alerts the user as to the characteristic in the scene requiring the user's visual attention when the controller determines there is at least one characteristic in the scene requiring the user's attention, simultaneous with the controller determining the user is gazing at the user device.
  • a vehicle in yet another aspect of the disclosure, includes a body and a scene awareness system.
  • the scene awareness system includes at least one exterior camera, at least one interior camera, and a vehicle controller.
  • the cameras are operatively attached to the body.
  • the exterior camera is configured to capture data pertaining to a scene, external to the vehicle.
  • the interior camera is configured to capture data pertaining to an orientation of a gaze of a user of the vehicle.
  • the vehicle controller is in communication with each of the cameras.
  • the vehicle controller is operable for receiving data pertaining to the scene, external to the vehicle, from the at least one exterior camera.
  • the vehicle controller includes a process and a characteristic in the scene requiring the user's visual attention that is determined in the processor.
  • Data, pertaining to a gaze of the user is received from the interior camera.
  • a determination is made that the user is gazing toward a user device.
  • a signal is transmitted to the user device, from the controller that alerts the user as to the characteristic in the scene requiring the user's
  • FIG. 1 is a schematic illustrative side view of a vehicle.
  • FIG. 2 is a schematic diagrammatic view of a scene awareness system of the vehicle of FIG. 1 .
  • FIG. 3A is schematic fragmentary view of an interior of the vehicle having a scene awareness system, illustrating the user of the vehicle gazing at a scene, external to the vehicle.
  • FIG. 3B is a schematic fragmentary view of the interior of the vehicle having the scene awareness system, illustrating the user of the vehicle gazing at a user device.
  • FIG. 4 is a schematic flow chart diagram of a method of alerting the user of the vehicle as to a scene, external to the vehicle, requiring the user's attention.
  • FIG. 1 schematically illustrates a vehicle 10 including a body 12 .
  • the vehicle 10 may be a land vehicle, such as a car, or any other type of vehicle such as an airplane, farm equipment, construction equipment, a boat, etc.
  • the vehicle 10 may be configured to be operated by a user 18 , operated autonomously, and/or operated semi-autonomously.
  • the vehicle 10 includes a windshield 14 coupled to the body 12 .
  • the windshield 14 is wholly or partly made of a substantially transparent material.
  • the vehicle 10 includes an interior 16 .
  • a vehicle user 18 can be in position within the interior 16 and see through the windshield 14 , external to the vehicle 10 .
  • the body 12 extends between a forward end 21 and a rearward end 23 .
  • the windshield 14 faces the forward end 21 of the vehicle 10 , opposite the rearward end 23 .
  • a scene awareness system 25 having one or more exterior cameras 26 , one or more internal cameras 28 , and a vehicle controller (C) 20 .
  • the vehicle controller 20 which is in communication with the internal and the exterior camera(s) 28 , 26 , is programmed to use the exterior camera(s) 26 to automatically locate and/or identify one or more characteristics in a scene 32 , exterior to the vehicle 10 , and to use the interior camera(s) 28 to simultaneously locate and identify where the user's 18 eyes are gazing within the vehicle 10 via execution of instructions embodying a method 100 , an example of which is described below with reference to FIG. 4 .
  • the characteristics may include, but should not be limited to, identifying one or more objects 30 , colors, brightness, darkness, temperatures, temperature gradients, graphical patterns, motion, and other like characteristics.
  • the exterior camera(s) 26 is configured to monitor a scene 32 , external to the vehicle 10 .
  • the exterior cameras 26 may be operatively attached to the vehicle 10 to view the scenery 32 in front of, beside, above, and/or behind the vehicle 10 .
  • at least one exterior camera 26 may be resident within a user device 42 located within the interior 16 of the vehicle 10 through the windshield 14 or some other window within the vehicle 10 .
  • the user device 42 would be positioned within the vehicle such that the exterior camera 26 is directed to view the scenery 32 around the vehicle.
  • the scenery 32 may be of a landscape, a billboard, a landmark, and the like.
  • the exterior camera(s) 26 may include a camera, a sensor, and the like.
  • the exterior camera(s) 26 is configured to capture images, in static or in motion, external to the vehicle 10 , in real-time.
  • the exterior camera(s) 26 is in operative communication with a processor 22 configured to determine if the scenery 32 captured in real-time is sufficient to require the user's 18 visual attention, i.e., the scenery 32 is visually salient.
  • the processor 22 may be resident within the vehicle controller 20 .
  • the exterior camera(s) 26 may be configured to use machine vision (MV) to recognize characteristics, such as objects 30 , in the scenery 32 , in real-time.
  • MV machine vision
  • the captured images may be transmitted as data (arrow 38 ) from the exterior camera(s) 26 to the processor 22 within the controller to make an imaging-based determination of whether the scenery 32 is visually salient.
  • Machine vision may make an imaging-based determination based on edge detection of objects 30 , color analysis to identify objects 30 , pattern recognition, feature detection, motion analysis, and the like.
  • the processor 22 may be pre-programmed with criteria.
  • scenery 32 sufficient to require the user's 18 visual attention may include an upcoming curve in the road, stop sign, landmark, destination, and the like.
  • such scenery 32 may include objects 30 which are static and/or objects which are in motion.
  • the interior camera(s) 28 is also in operative communication with the processor 22 .
  • the interior camera(s) 28 may be used to capture visual information with respect to the user 18 of the vehicle 10 . More specifically, the interior camera(s) 28 may capture information about the users 18 head, including their eyes, in real-time.
  • the interior camera(s) 28 may be positioned within the vehicle 10 , e.g., attached to a rearview mirror, and the like. Alternatively, the interior camera(s) 28 may be positioned within the vehicle, where the interior camera 28 is resident within the user device 42 . In such an example, the user device 42 may be configured with both the interior camera 28 and the exterior camera 26 .
  • the captured visual information may be transmitted as data (arrow 40 ) to the processor 22 within the vehicle controller 20 .
  • the interior camera(s) 28 is configured to determine where the eyes of the user 18 are looking and/or to determine an orientation of the user's 18 head, relative to the windshield 14 and/or the forward end 21 of the vehicle 10 .
  • the interior camera(s) 28 and/or the user device 42 may be configured to determine whether the user 18 is gazing out the windshield 14 , gazing at the user device 42 , or gazing in any other direction, using MV for facial detection.
  • the interior camera(s) 28 may be a camera, an infrared (IR) sensor, and the like.
  • the vehicle controller 20 is also configured to be in operative communication with the user device 42 .
  • the user device 42 includes a device controller 44 and a display screen 46 .
  • the user device 42 is configured to display information, including text and graphics, on the display screen 46 .
  • the user device 42 may be a portable personal device, such as a cell phone, a tablet, a computer, and the like. Alternatively, the user device 42 may be integrated into the vehicle 10 , e.g., as an integrated infotainment device.
  • the user device 42 is configured to be in operative communication with the vehicle controller 20 .
  • the vehicle controller 20 may selectively transmit a signal (arrow 48 ) to the device controller 44 to clear the display screen 46 and/or to replace the displayed image of the display screen 46 with the captured scenery 32 .
  • the vehicle 10 may be configured to be in selective communication with an off-board system 49 .
  • the vehicle controller 20 may be configured to receive data relating to traffic conditions, traffic signals, weather conditions, and the like, from the off-board system 49 .
  • An example of an off-board system 49 may include a service provider, which may be configured as a server located off-board the vehicle 10 , e.g., at a location remote from the vehicle 10 .
  • the off-board system 49 may be a vehicle integrated service provider, such as the OnStar® service system, which may be selectively linked to the vehicle interface device and/or in communication with the portable device.
  • the service provider may be configured to provide data to the vehicle controller 20 via Wi-Fi, a mobile telecommunications network, satellite, Bluetooth, and the like.
  • the off-board system 49 may include environmental sensors 52 disposed in conjunction with landmarks, such as a tree, and the like. The environmental sensor 52 may be configured to transmit a signal (arrow 55 ) to the vehicle controller 20 to alert the vehicle controller 20 as to the proximity of such a landmark to the vehicle 10 .
  • the vehicle controller 20 is programmed with, or has access to, the algorithm or method 100 , the execution of which provides a method of improving awareness of a user 18 of the vehicle 10 as to the scene 32 , external to a vehicle 10 , with the algorithm 100 explained in detail below and as shown in FIG. 4 .
  • the vehicle controller 20 of FIG. 1 may be embodied as one or more computer devices having a processor (P) 22 and tangible, non-transitory memory (M) 24 on which is recorded instructions for executing the method 100 .
  • the memory 24 may include magnetic or optical memory, electrically-erasable programmable read only memory (EEPROM), and the like. Additional transitory memory may be included as needed, e.g., random access memory (RAM), memory for internal signal buffers, etc.
  • Other hardware of the vehicle controller 20 may include a high-speed clock, analog-to-digital (A/D) circuitry, digital-to-analog (D/A) circuitry, and any required input/output (I/O) circuitry and devices, as well as signal conditioning and buffer electronics. Individual control algorithms resident in the vehicle controller 20 or readily accessible by the vehicle controller 20 may be stored in memory 24 and/or other suitable memory, and automatically executed via the processor 22 to provide the required control functionality.
  • A/D analog-to-digital
  • D/A digital-to-
  • the exterior camera(s) 26 are operable for imaging the scene 32 exterior to the vehicle 10 , in real-time
  • the interior camera(s) 28 are operable for imaging the user's 18 head including their eyes, inside of the vehicle 10 , in real-time.
  • Other sensors may also be used to determine or supplement the interior camera's determination of whether the user 18 is gazing at the user device 42 . More specifically, such sensors may sense a physical interaction between the device and the user 18 of the vehicle 10 .
  • the collected data (arrow 38 ) of the exterior camera(s) 26 and the collected data (arrow 40 ) of the interior camera(s) 28 are transmitted to the vehicle controller 20 for processing according to the method 100 .
  • the exterior camera(s) 26 and/or the interior camera(s) 28 may be three-dimensional (3D) point cloud cameras.
  • a 3D point cloud is a set of data points in a 3D coordinate system, such as the X, Y, Z Cartesian coordinate system.
  • Such cameras are able to capture any number of data points describing the surface contour of a target object 30 , and to output the collected data (arrow 38 ) as a depth data file with synchronized color data.
  • the vehicle controller 20 may be preprogrammed with predetermined target data and/or to receive predetermined target information from the off-board system 49 (arrow 50 ) and/or the environmental sensors 52 (arrow 55 ) which define the known size, shape, color, movements, and/or other descriptive parameters of the particular objects 30 to be located by the external and/or interior cameras 28 .
  • the vehicle controller 20 may be made aware of the dimensions, color, and/or other identifying features of the objects 30 it is attempting to identify. While one exterior camera 26 and one interior camera 28 are shown in FIG. 2 for illustrative simplicity, the present method 100 may work with more than one exterior camera 26 and/or more than one interior camera 28 .
  • an example embodiment of the method 100 begins with step 102 , where the vehicle controller 20 receives data (arrow 56 ) pertaining to operation of the vehicle 10 .
  • the data may be received from a plurality of sensors 54 .
  • some of the sensors may be configured to determine revolutions per minute (RPM) of a transmission and/or engine of the vehicle 10 , i.e., via a transmission sensor 54 A and/or an engine sensor 54 B; determine, via a brake pedal sensor 54 C, whether a brake pedal is being actuated; determine, via a steering wheel sensor 54 D, whether the user's 18 hand is touching a steering wheel; determine a wheel speed of the vehicle 10 , i.e., via a wheel speed sensor 54 E; determine whether a clutch switch has been actuated, i.e., via a clutch sensor 54 F; determine, via an accelerometer 54 G, a longitudinal acceleration of the vehicle 10 ; and the like.
  • RPM revolutions per minute
  • determining whether the vehicle 10 is being operated may include receiving geographic data into the vehicle controller 20 regarding vehicle 10 position, e.g., via a global positioning system (GPS), and the like.
  • This geographic data may, in turn, be used as a prompt to determine upcoming traffic conditions in order to suggest a driving route and/or suggest a traffic lane to be used.
  • This geographic data may also be used to prevent a preview of an upcoming scene, e.g., a hidden intersection, a hidden driveway, a curve, and other geographic data.
  • the method proceeds to step 104 .
  • the vehicle controller 20 determines, based on the data, received by the vehicle controller 20 , whether the vehicle 10 is being operated by the user 18 . If the vehicle controller 20 determines the vehicle 10 is not being operated, the method returns to step 102 . However, if the vehicle controller 20 determines the vehicle 10 is being operated, the method proceeds to step 106 .
  • the exterior camera(s) 26 captures data pertaining to the scene 32 , external to the vehicle 10 .
  • the method then proceeds to step 108 .
  • step 108 the captured data, pertaining to the scene 32 exterior to the vehicle 10 , is transmitted to the processor 22 , in real-time. The method then proceeds to step 110 .
  • the processor 22 determines whether the scene 32 is visually salient. More specifically, the processor 22 determines whether there are important objects 30 or characteristic within the captured data of the scene 32 . If a determination is made that an object 30 considered to be important is detected in the scene 32 , the method proceeds to step 112 . However, if a determination is made that no visually salient objects 30 or characteristics are detected in the scene 32 , the method returns to step 102 .
  • the interior camera(s) 28 captures data, in real-time, data pertaining to a gaze 29 of the user 18 , including but not limited to, an orientation of the head and/or eyes of the user 18 .
  • the method then proceeds to step 114 .
  • step 114 the captured data is transmitted to the processor 22 in the vehicle controller 20 , in real-time. The method then proceeds to step 116 .
  • the processor 22 determines where the eyes of the user 18 are gazing.
  • the vehicle controller 20 receives the captured data and processes the captured data using a processor 22 . If the processor 22 determines at step 116 that the user 18 is not gazing at the user device 42 , the method returns to step 118 .
  • the vehicle controller 20 may transmit a signal to the user device 42 such that the user device 42 displays content on the display screen 46 , not pertaining to the scenery 32 , i.e., the display screen 46 returns to a nominal condition or to an originally displayed condition. It should be appreciated that, in one embodiment, such a signal may only be transmitted to the user device 42 after certain criteria are satisfied, e.g., the passage of a pre-defined period of time where the user 18 is not gazing at the user device 42 , etc. The method then returns to step 102 .
  • step 116 determines at step 116 that the user 18 is gazing at the user device 42 .
  • the method proceeds to step 118 .
  • the controller may, in turn, transmit a signal to the user device 42 .
  • a signal may instruct the controller of the user device 42 to clear the display screen 46 .
  • a signal may instruct the controller of the user device 42 to display imagery corresponding to the visually salient scenery 32 .
  • the user 18 is alerted to scenery 32 external to the vehicle 10 , requiring the user's 18 attention.
  • the signal transmitted by the vehicle controller 20 may not be limited to clearing the display screen 46 and/or displaying scenery 32 external to the vehicle 10 .
  • the signal may be also be transmitted to the user device 42 to instruct the user device 42 to output an audible sound and/or cause the user device 42 to vibrate.
  • the signal may be transmitted to change displays of the driving displays, e.g., instrument cluster, driver information center (DIC), navigation screen, heads-up display, inside rearview mirror, outside rearview mirror, and the like when a determination is made that the user 18 may be gazing at such driving displays, instead of at the scenery 32 external to the vehicle 10 .
  • the driving displays e.g., instrument cluster, driver information center (DIC), navigation screen, heads-up display, inside rearview mirror, outside rearview mirror, and the like when a determination is made that the user 18 may be gazing at such driving displays, instead of at the scenery 32 external to the vehicle 10 .
  • DIC driver information center

Abstract

A method of alerting a user of a vehicle as to a scene, external to the vehicle, includes capturing data pertaining to the scene, external to the vehicle. The captured data, pertaining to the scene external to the vehicle, is transmitted to a processor. The determination is made, in the processor that a characteristic is in the scene that requires the user's visual attention. Data pertaining to a gaze of the user is captured. Captured data, pertaining to the gaze of the user, is transmitted to the processor. A determination is made, in the processor that the drive is gazing toward a user device. A signal is transmitted to the user device such that the user device alerts the user as to the characteristic in the scene requiring the user's visual attention.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of International Patent Application No. PCT/US2014/035398, filed on Apr. 25, 2014, which claims the benefit of, U.S. Provisional Application No. 61/816,089, filed Apr. 25, 2013, each of which are hereby incorporated by reference in their entirety.
  • TECHNICAL FIELD
  • The present disclosure is related to an adaptive external road scene transfer to improve driver awareness.
  • BACKGROUND
  • Vehicles, such as cars, typically include displays or indicators to provide information to the vehicle user. Such displays or indicators may, for example, provide information regarding mileage, fuel consumption, and vehicle speed. The vehicle user usually has to shift his eye gaze away from the road scene and onto an in-vehicle display in order to visually process the information presented by these displays or indicators.
  • SUMMARY
  • One possible aspect of the disclosure provides a method of alerting a user of a vehicle as to a scene, external to the vehicle, which includes capturing data pertaining to the scene, external to the vehicle. The captured data, pertaining to the scene external to the vehicle, is transmitted to a processor. The determination is made, in the processor that a characteristic is in the scene that requires the user's visual attention. Data pertaining to a gaze of the user is captured. Captured data, pertaining to the gaze of the user, is transmitted to the processor. A determination is made, in the processor that the user is gazing toward a user device. A signal is transmitted to the user device such that the user device alerts the user as to the characteristic in the scene requiring the user's visual attention.
  • Another aspect of the disclosure provides a scene awareness system for a vehicle that includes at least one exterior camera, at least one interior camera, and a vehicle controller. The exterior camera is configured to capture data pertaining to a scene, external to the vehicle. The interior camera is configured to capture data pertaining to an orientation of a gaze of a user of the vehicle. The vehicle controller is in communication with each of the at least one exterior camera and the at least one interior camera. The vehicle controller is configured to determine whether at least one characteristic in the scene, exterior to the vehicle, requires the user's visual attention. The vehicle controller is also configured to determine whether the user is gazing at a user device. The vehicle controller is configured to transmit a signal to the user device such that the user device alerts the user as to the characteristic in the scene requiring the user's visual attention when the controller determines there is at least one characteristic in the scene requiring the user's attention, simultaneous with the controller determining the user is gazing at the user device.
  • In yet another aspect of the disclosure, a vehicle includes a body and a scene awareness system. The scene awareness system includes at least one exterior camera, at least one interior camera, and a vehicle controller. The cameras are operatively attached to the body. The exterior camera is configured to capture data pertaining to a scene, external to the vehicle. The interior camera is configured to capture data pertaining to an orientation of a gaze of a user of the vehicle. The vehicle controller is in communication with each of the cameras. The vehicle controller is operable for receiving data pertaining to the scene, external to the vehicle, from the at least one exterior camera. The vehicle controller includes a process and a characteristic in the scene requiring the user's visual attention that is determined in the processor. Data, pertaining to a gaze of the user, is received from the interior camera. A determination is made that the user is gazing toward a user device. A signal is transmitted to the user device, from the controller that alerts the user as to the characteristic in the scene requiring the user's visual attention.
  • The above features and advantages and other features and advantages of the present teachings are readily apparent from the following detailed description of the best modes for carrying out the present teachings when taken in connection with the accompanying drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a schematic illustrative side view of a vehicle.
  • FIG. 2 is a schematic diagrammatic view of a scene awareness system of the vehicle of FIG. 1.
  • FIG. 3A is schematic fragmentary view of an interior of the vehicle having a scene awareness system, illustrating the user of the vehicle gazing at a scene, external to the vehicle.
  • FIG. 3B is a schematic fragmentary view of the interior of the vehicle having the scene awareness system, illustrating the user of the vehicle gazing at a user device.
  • FIG. 4 is a schematic flow chart diagram of a method of alerting the user of the vehicle as to a scene, external to the vehicle, requiring the user's attention.
  • DETAILED DESCRIPTION
  • Those having ordinary skill in the art will recognize that terms such as “above,” “below,” “upward,” “downward,” “top,” “bottom,” etc., are used descriptively for the figures, and do not represent limitations on the scope of the invention, as defined by the appended claims. Furthermore, the invention may be described herein in terms of functional and/or logical block components and/or various processing steps. It should be realized that such block components may be comprised of any number of hardware, software, and/or firmware components configured to perform the specified functions.
  • Referring now to the drawings, wherein the like numerals indicate corresponding parts throughout the several views, FIG. 1 schematically illustrates a vehicle 10 including a body 12. The vehicle 10 may be a land vehicle, such as a car, or any other type of vehicle such as an airplane, farm equipment, construction equipment, a boat, etc. The vehicle 10 may be configured to be operated by a user 18, operated autonomously, and/or operated semi-autonomously. Regardless of the specific kind of vehicle, the vehicle 10 includes a windshield 14 coupled to the body 12. The windshield 14 is wholly or partly made of a substantially transparent material. Referring now to FIGS. 3A and 3B, the vehicle 10 includes an interior 16. Accordingly, a vehicle user 18 can be in position within the interior 16 and see through the windshield 14, external to the vehicle 10. Referring again to FIG. 1, the body 12 extends between a forward end 21 and a rearward end 23. The windshield 14 faces the forward end 21 of the vehicle 10, opposite the rearward end 23.
  • With reference to FIG. 2, a scene awareness system 25 is shown having one or more exterior cameras 26, one or more internal cameras 28, and a vehicle controller (C) 20. The vehicle controller 20, which is in communication with the internal and the exterior camera(s) 28, 26, is programmed to use the exterior camera(s) 26 to automatically locate and/or identify one or more characteristics in a scene 32, exterior to the vehicle 10, and to use the interior camera(s) 28 to simultaneously locate and identify where the user's 18 eyes are gazing within the vehicle 10 via execution of instructions embodying a method 100, an example of which is described below with reference to FIG. 4. The characteristics may include, but should not be limited to, identifying one or more objects 30, colors, brightness, darkness, temperatures, temperature gradients, graphical patterns, motion, and other like characteristics.
  • Referring to FIG. 2, the exterior camera(s) 26 is configured to monitor a scene 32, external to the vehicle 10. The exterior cameras 26 may be operatively attached to the vehicle 10 to view the scenery 32 in front of, beside, above, and/or behind the vehicle 10. Alternatively, at least one exterior camera 26 may be resident within a user device 42 located within the interior 16 of the vehicle 10 through the windshield 14 or some other window within the vehicle 10. In such an embodiment, the user device 42 would be positioned within the vehicle such that the exterior camera 26 is directed to view the scenery 32 around the vehicle. The scenery 32 may be of a landscape, a billboard, a landmark, and the like. It should be appreciated that the portions of the scenery 32 may be static, in motion, and/or a combination of in static or in motion. The exterior camera(s) 26 may include a camera, a sensor, and the like. The exterior camera(s) 26 is configured to capture images, in static or in motion, external to the vehicle 10, in real-time. The exterior camera(s) 26 is in operative communication with a processor 22 configured to determine if the scenery 32 captured in real-time is sufficient to require the user's 18 visual attention, i.e., the scenery 32 is visually salient. The processor 22 may be resident within the vehicle controller 20. The exterior camera(s) 26 may be configured to use machine vision (MV) to recognize characteristics, such as objects 30, in the scenery 32, in real-time. The captured images may be transmitted as data (arrow 38) from the exterior camera(s) 26 to the processor 22 within the controller to make an imaging-based determination of whether the scenery 32 is visually salient. Machine vision may make an imaging-based determination based on edge detection of objects 30, color analysis to identify objects 30, pattern recognition, feature detection, motion analysis, and the like. The processor 22 may be pre-programmed with criteria. By way of a non-limiting example, scenery 32 sufficient to require the user's 18 visual attention may include an upcoming curve in the road, stop sign, landmark, destination, and the like. As previously mentioned, such scenery 32 may include objects 30 which are static and/or objects which are in motion.
  • The interior camera(s) 28 is also in operative communication with the processor 22. The interior camera(s) 28 may be used to capture visual information with respect to the user 18 of the vehicle 10. More specifically, the interior camera(s) 28 may capture information about the users 18 head, including their eyes, in real-time. The interior camera(s) 28 may be positioned within the vehicle 10, e.g., attached to a rearview mirror, and the like. Alternatively, the interior camera(s) 28 may be positioned within the vehicle, where the interior camera 28 is resident within the user device 42. In such an example, the user device 42 may be configured with both the interior camera 28 and the exterior camera 26. The captured visual information may be transmitted as data (arrow 40) to the processor 22 within the vehicle controller 20. The interior camera(s) 28 is configured to determine where the eyes of the user 18 are looking and/or to determine an orientation of the user's 18 head, relative to the windshield 14 and/or the forward end 21 of the vehicle 10. The interior camera(s) 28 and/or the user device 42 may be configured to determine whether the user 18 is gazing out the windshield 14, gazing at the user device 42, or gazing in any other direction, using MV for facial detection. The interior camera(s) 28 may be a camera, an infrared (IR) sensor, and the like.
  • The vehicle controller 20 is also configured to be in operative communication with the user device 42. The user device 42 includes a device controller 44 and a display screen 46. The user device 42 is configured to display information, including text and graphics, on the display screen 46. The user device 42 may be a portable personal device, such as a cell phone, a tablet, a computer, and the like. Alternatively, the user device 42 may be integrated into the vehicle 10, e.g., as an integrated infotainment device.
  • As described in more detail below, the user device 42 is configured to be in operative communication with the vehicle controller 20. As such, the vehicle controller 20 may selectively transmit a signal (arrow 48) to the device controller 44 to clear the display screen 46 and/or to replace the displayed image of the display screen 46 with the captured scenery 32.
  • Referring again to FIG. 2, the vehicle 10 may be configured to be in selective communication with an off-board system 49. More specifically, the vehicle controller 20 may be configured to receive data relating to traffic conditions, traffic signals, weather conditions, and the like, from the off-board system 49. An example of an off-board system 49 may include a service provider, which may be configured as a server located off-board the vehicle 10, e.g., at a location remote from the vehicle 10. The off-board system 49 may be a vehicle integrated service provider, such as the OnStar® service system, which may be selectively linked to the vehicle interface device and/or in communication with the portable device. By way of another non-limiting example, the service provider may be configured to provide data to the vehicle controller 20 via Wi-Fi, a mobile telecommunications network, satellite, Bluetooth, and the like. In yet another non-limiting example, as illustrated in FIG. 2, the off-board system 49 may include environmental sensors 52 disposed in conjunction with landmarks, such as a tree, and the like. The environmental sensor 52 may be configured to transmit a signal (arrow 55) to the vehicle controller 20 to alert the vehicle controller 20 as to the proximity of such a landmark to the vehicle 10.
  • The vehicle controller 20 is programmed with, or has access to, the algorithm or method 100, the execution of which provides a method of improving awareness of a user 18 of the vehicle 10 as to the scene 32, external to a vehicle 10, with the algorithm 100 explained in detail below and as shown in FIG. 4.
  • The vehicle controller 20 of FIG. 1 may be embodied as one or more computer devices having a processor (P) 22 and tangible, non-transitory memory (M) 24 on which is recorded instructions for executing the method 100. The memory 24 may include magnetic or optical memory, electrically-erasable programmable read only memory (EEPROM), and the like. Additional transitory memory may be included as needed, e.g., random access memory (RAM), memory for internal signal buffers, etc. Other hardware of the vehicle controller 20 may include a high-speed clock, analog-to-digital (A/D) circuitry, digital-to-analog (D/A) circuitry, and any required input/output (I/O) circuitry and devices, as well as signal conditioning and buffer electronics. Individual control algorithms resident in the vehicle controller 20 or readily accessible by the vehicle controller 20 may be stored in memory 24 and/or other suitable memory, and automatically executed via the processor 22 to provide the required control functionality.
  • Still referring to FIG. 2, when the vehicle 10 is being operated, the exterior camera(s) 26 are operable for imaging the scene 32 exterior to the vehicle 10, in real-time, and the interior camera(s) 28 are operable for imaging the user's 18 head including their eyes, inside of the vehicle 10, in real-time. Other sensors may also be used to determine or supplement the interior camera's determination of whether the user 18 is gazing at the user device 42. More specifically, such sensors may sense a physical interaction between the device and the user 18 of the vehicle 10. The collected data (arrow 38) of the exterior camera(s) 26 and the collected data (arrow 40) of the interior camera(s) 28 are transmitted to the vehicle controller 20 for processing according to the method 100. The exterior camera(s) 26 and/or the interior camera(s) 28 may be three-dimensional (3D) point cloud cameras. As is known in the art, a 3D point cloud is a set of data points in a 3D coordinate system, such as the X, Y, Z Cartesian coordinate system. Such cameras are able to capture any number of data points describing the surface contour of a target object 30, and to output the collected data (arrow 38) as a depth data file with synchronized color data. The vehicle controller 20 may be preprogrammed with predetermined target data and/or to receive predetermined target information from the off-board system 49 (arrow 50) and/or the environmental sensors 52 (arrow 55) which define the known size, shape, color, movements, and/or other descriptive parameters of the particular objects 30 to be located by the external and/or interior cameras 28.
  • The vehicle controller 20, shown in FIG. 2, may be made aware of the dimensions, color, and/or other identifying features of the objects 30 it is attempting to identify. While one exterior camera 26 and one interior camera 28 are shown in FIG. 2 for illustrative simplicity, the present method 100 may work with more than one exterior camera 26 and/or more than one interior camera 28.
  • Referring to FIG. 4, an example embodiment of the method 100 begins with step 102, where the vehicle controller 20 receives data (arrow 56) pertaining to operation of the vehicle 10. The data may be received from a plurality of sensors 54. By way of a non-limiting example, some of the sensors may be configured to determine revolutions per minute (RPM) of a transmission and/or engine of the vehicle 10, i.e., via a transmission sensor 54A and/or an engine sensor 54B; determine, via a brake pedal sensor 54C, whether a brake pedal is being actuated; determine, via a steering wheel sensor 54D, whether the user's 18 hand is touching a steering wheel; determine a wheel speed of the vehicle 10, i.e., via a wheel speed sensor 54E; determine whether a clutch switch has been actuated, i.e., via a clutch sensor 54F; determine, via an accelerometer 54G, a longitudinal acceleration of the vehicle 10; and the like.
  • Additionally, determining whether the vehicle 10 is being operated may include receiving geographic data into the vehicle controller 20 regarding vehicle 10 position, e.g., via a global positioning system (GPS), and the like. This geographic data may, in turn, be used as a prompt to determine upcoming traffic conditions in order to suggest a driving route and/or suggest a traffic lane to be used. This geographic data may also be used to prevent a preview of an upcoming scene, e.g., a hidden intersection, a hidden driveway, a curve, and other geographic data. Next, the method proceeds to step 104.
  • At step 104, the vehicle controller 20 determines, based on the data, received by the vehicle controller 20, whether the vehicle 10 is being operated by the user 18. If the vehicle controller 20 determines the vehicle 10 is not being operated, the method returns to step 102. However, if the vehicle controller 20 determines the vehicle 10 is being operated, the method proceeds to step 106.
  • At step 106, the exterior camera(s) 26 captures data pertaining to the scene 32, external to the vehicle 10. The method then proceeds to step 108.
  • At step 108, the captured data, pertaining to the scene 32 exterior to the vehicle 10, is transmitted to the processor 22, in real-time. The method then proceeds to step 110.
  • At step 110, the processor 22 determines whether the scene 32 is visually salient. More specifically, the processor 22 determines whether there are important objects 30 or characteristic within the captured data of the scene 32. If a determination is made that an object 30 considered to be important is detected in the scene 32, the method proceeds to step 112. However, if a determination is made that no visually salient objects 30 or characteristics are detected in the scene 32, the method returns to step 102.
  • At step 112, with the user 18 seated within the interior 16 of the vehicle 10, the interior camera(s) 28 captures data, in real-time, data pertaining to a gaze 29 of the user 18, including but not limited to, an orientation of the head and/or eyes of the user 18. The method then proceeds to step 114.
  • At step 114, the captured data is transmitted to the processor 22 in the vehicle controller 20, in real-time. The method then proceeds to step 116.
  • At step 116, the processor 22 determines where the eyes of the user 18 are gazing. The vehicle controller 20 receives the captured data and processes the captured data using a processor 22. If the processor 22 determines at step 116 that the user 18 is not gazing at the user device 42, the method returns to step 118.
  • At step 118, the vehicle controller 20 may transmit a signal to the user device 42 such that the user device 42 displays content on the display screen 46, not pertaining to the scenery 32, i.e., the display screen 46 returns to a nominal condition or to an originally displayed condition. It should be appreciated that, in one embodiment, such a signal may only be transmitted to the user device 42 after certain criteria are satisfied, e.g., the passage of a pre-defined period of time where the user 18 is not gazing at the user device 42, etc. The method then returns to step 102.
  • However, if the processor 22 determines at step 116 that the user 18 is gazing at the user device 42, the method proceeds to step 118.
  • At step 118, the controller may, in turn, transmit a signal to the user device 42. Such a signal may instruct the controller of the user device 42 to clear the display screen 46. Additionally, such a signal may instruct the controller of the user device 42 to display imagery corresponding to the visually salient scenery 32. By changing the display screen 46 of the user device 42, the user 18 is alerted to scenery 32 external to the vehicle 10, requiring the user's 18 attention.
  • It should be appreciated that the signal transmitted by the vehicle controller 20 may not be limited to clearing the display screen 46 and/or displaying scenery 32 external to the vehicle 10. By way of a non-limiting example, the signal may be also be transmitted to the user device 42 to instruct the user device 42 to output an audible sound and/or cause the user device 42 to vibrate. In another non-limiting example, the signal may be transmitted to change displays of the driving displays, e.g., instrument cluster, driver information center (DIC), navigation screen, heads-up display, inside rearview mirror, outside rearview mirror, and the like when a determination is made that the user 18 may be gazing at such driving displays, instead of at the scenery 32 external to the vehicle 10.
  • While the best modes for carrying out the many aspects of the present teachings have been described in detail, those familiar with the art to which these teachings relate will recognize various alternative aspects for practicing the present teachings that are within the scope of the appended claims.

Claims (18)

1. A method of alerting a user of a vehicle as to a scene, external to the vehicle, the method comprising:
capturing data pertaining to the scene, external to the vehicle;
transmitting the captured data pertaining to the scene, external to the vehicle, to a processor;
determining, in the processor, a characteristic of the scene requiring the user's visual attention;
capturing data pertaining to a gaze of the user;
transmitting the captured data pertaining to the gaze of the user to the processor;
determining, in the processor, the driver is gazing toward a user device; and
transmitting a signal to the user device such that the user device alerts the user as to the characteristic of the scene requiring the user's visual attention.
2. A method, as set forth in claim 1, wherein transmitting is further defined as transmitting a signal to the user device such that the user device alerts the user as to the characteristic of the scene requiring the user's visual attention by clearing the screen of the user device.
3. A method, as set forth in claim 2, wherein transmitting is further defined as transmitting a signal to the user device such that the user device alerts the user as to the characteristic of the scene requiring the user's visual attention by displaying a visual depiction of the characteristic of the scene requiring the user's visual attention on the display screen of the user device.
4. A method, as set forth in claim 1, further comprising:
receiving data pertaining to operation of the vehicle; and
determining the vehicle is being operated by the user.
5. A method, as set forth in claim 4, wherein capturing data pertaining to a gaze of the user is further defined as capturing data pertaining to a gaze of the user when a determination is made that the vehicle is being operated by the user.
6. A method, as set forth in claim 4, wherein receiving data is further defined as receiving geographic data regarding a vehicle position.
7. A method, as set forth in claim 1, wherein capturing data pertaining to the scene, external to the vehicle, is further defined as capturing data pertaining to the scene, external to the vehicle, with at least one external camera.
8. A method, as set forth in claim 1, wherein capturing data pertaining to a gaze of the user is further defined as capturing data pertaining to a gaze of the user when a determination is made that a characteristic of the scene requiring the user's attention is detected in the scene.
9. A method, as set forth in claim 1, wherein capturing data pertaining to a gaze of the user is further defined as capturing data pertaining to a gaze of the user with at least one internal camera.
10. A method, as set forth in claim 9, wherein capturing data pertaining to a gaze of the user is further defined as capturing data pertaining to an orientation of at least one of the head and the eyes of the user.
11. A method, as set forth in claim 1, wherein capturing data pertaining to a gaze of the user is further defined as capturing data pertaining to a gaze of the user after determining, in the processor, a characteristic of the scene requiring the user's visual attention.
12. A scene awareness system for a vehicle, the scene awareness system comprising:
at least one exterior camera configured to capture data pertaining to a scene, external to the vehicle;
at least one interior camera configured to capture data pertaining to an orientation of a gaze of a user of the vehicle;
a vehicle controller in communication with each of the at least one exterior camera and the at least one interior camera;
wherein the vehicle controller is configured to determine whether at least one characteristic in the scene, exterior to the vehicle, requires the user's visual attention;
wherein the vehicle controller is configured to determine whether the user is gazing at a user device; and
wherein the vehicle controller is configured to transmit a signal to the user device such that the user device alerts the user as to the characteristic in the scene requiring the user's visual attention when the controller determines there is at least one characteristic in the scene requiring the user's attention simultaneous with the controller determining the user is gazing at the user device.
13. A scene awareness system, as set forth in claim 12, further comprising at least one sensor configured for transmitting data to the vehicle controller;
wherein the data pertains to operation of the vehicle.
14. A vehicle comprising:
a body;
a scene awareness system including:
at least one exterior camera operatively attached to the body;
wherein the at least one exterior camera is configured to capture data pertaining to a scene, external to the vehicle;
at least one interior camera operatively attached to the body;
wherein the at least one interior camera is configured to capture data pertaining to an orientation of a gaze of a user of the vehicle;
a vehicle controller in communication with each of the at least one exterior camera and the at least one interior camera, the vehicle controller operable for:
receiving data pertaining to the scene, external to the vehicle, from the at least one exterior camera;
determining, in the processor, a characteristic in the scene requiring the user's visual attention;
receiving data pertaining to a gaze of the user from the at least one interior camera;
determining the driver is gazing toward a user device; and
transmitting a signal to a user device such that the user device alerts the user as to the characteristic in the scene requiring the user's visual attention.
15. A vehicle, as set forth in claim 14, wherein the body extends between a forward end and a rearward end; and
wherein the at least one exterior camera is operatively attached to the body, proximate the forward end.
16. A vehicle, as set forth in claim 15, wherein the body defines an interior configured for receiving the user of the vehicle therein;
wherein the at least one interior camera is operatively disposed in the interior of the body.
17. A vehicle, as set forth in claim 16, further comprising at least one sensor configured to be in operative communication with the vehicle controller;
wherein the vehicle controller is further operable for receiving data, pertaining to operation of the vehicle, from the at least one sensor.
18. A vehicle, as set forth in claim 14, wherein the vehicle controller is further operable for:
receiving data pertaining to operation of the vehicle; and
determining the vehicle is being operated by the user.
US14/920,420 2013-04-25 2015-10-22 Scene awareness system for a vehicle Abandoned US20160039285A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/920,420 US20160039285A1 (en) 2013-04-25 2015-10-22 Scene awareness system for a vehicle

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US201361816089P 2013-04-25 2013-04-25
PCT/US2014/035398 WO2014176478A1 (en) 2013-04-25 2014-04-25 Scene awareness system for a vehicle
US14/920,420 US20160039285A1 (en) 2013-04-25 2015-10-22 Scene awareness system for a vehicle

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2014/035398 Continuation WO2014176478A1 (en) 2013-04-25 2014-04-25 Scene awareness system for a vehicle

Publications (1)

Publication Number Publication Date
US20160039285A1 true US20160039285A1 (en) 2016-02-11

Family

ID=51792397

Family Applications (4)

Application Number Title Priority Date Filing Date
US14/918,073 Active 2034-06-28 US9688287B2 (en) 2013-04-25 2015-10-20 Situation awareness system and method
US14/920,499 Active 2035-06-10 US10131364B2 (en) 2013-04-25 2015-10-22 Ambient display
US14/920,420 Abandoned US20160039285A1 (en) 2013-04-25 2015-10-22 Scene awareness system for a vehicle
US14/920,413 Abandoned US20160041562A1 (en) 2013-04-25 2015-10-22 Method of controlling a component of a vehicle with a user device

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US14/918,073 Active 2034-06-28 US9688287B2 (en) 2013-04-25 2015-10-20 Situation awareness system and method
US14/920,499 Active 2035-06-10 US10131364B2 (en) 2013-04-25 2015-10-22 Ambient display

Family Applications After (1)

Application Number Title Priority Date Filing Date
US14/920,413 Abandoned US20160041562A1 (en) 2013-04-25 2015-10-22 Method of controlling a component of a vehicle with a user device

Country Status (4)

Country Link
US (4) US9688287B2 (en)
CN (1) CN105324268A (en)
DE (1) DE112014001607B4 (en)
WO (4) WO2014176473A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170054837A1 (en) * 2014-05-09 2017-02-23 Samsung Electronics Co., Ltd. Terminal and method for displaying caller information
US20170294065A1 (en) * 2014-04-18 2017-10-12 Gentex Corporation Trainable transceiver and cloud computing system architecture systems and methods
US10166996B2 (en) * 2017-02-09 2019-01-01 Toyota Motor Engineering & Manufacturing North America, Inc. Systems and methods for adaptively communicating notices in a vehicle
US11312300B1 (en) 2021-01-29 2022-04-26 Toyota Motor Engineering & Manufacturing North America, Inc. Object notification systems for identifying and notifying vehicles of relevant objects

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150321604A1 (en) * 2014-05-07 2015-11-12 Ford Global Technologies, Llc In-vehicle micro-interactions
DE102015212676A1 (en) * 2015-07-07 2017-01-12 Bayerische Motoren Werke Aktiengesellschaft Determining the driving ability of the driver of a first motor vehicle
CN105631977A (en) * 2016-02-18 2016-06-01 广东百事泰电子商务股份有限公司 Intelligent monitoring and recording instrument
TW201836890A (en) * 2017-03-31 2018-10-16 育全 李 Method of showing the inside status of a vehicle via a plurality of first icons
JP6325154B1 (en) * 2017-06-07 2018-05-16 スマート ビート プロフィッツ リミテッド Information processing system
WO2019073983A1 (en) * 2017-10-10 2019-04-18 積水化学工業株式会社 Passenger vehicle window glass and warning display method
US10709386B2 (en) 2017-12-12 2020-07-14 Lear Corporation Electrocardiogram waveform identification and diagnostics via electrophysiological sensor system fusion
CN110027473A (en) * 2018-01-04 2019-07-19 哈曼国际工业有限公司 The situation skylight of media experience for the enhancing in compartment
JPWO2020071169A1 (en) * 2018-10-01 2021-09-02 富士フイルム株式会社 display
US11151810B2 (en) * 2018-10-12 2021-10-19 Aurora Flight Sciences Corporation Adaptable vehicle monitoring system
US20200376937A1 (en) * 2019-05-29 2020-12-03 Toyota Boshoku Kabushiki Kaisha Light-adjusting system and vehicle light-adjusting system
CN112918381B (en) * 2019-12-06 2023-10-27 广州汽车集团股份有限公司 Vehicle-mounted robot welcome method, device and system
US11467401B2 (en) * 2021-03-02 2022-10-11 GM Global Technology Operations LLC Display and light blocking screens
US11506892B1 (en) 2021-05-03 2022-11-22 GM Global Technology Operations LLC Holographic display system for a motor vehicle
US11762195B2 (en) 2021-05-06 2023-09-19 GM Global Technology Operations LLC Holographic display system with conjugate image removal for a motor vehicle
US20240054528A1 (en) * 2022-08-10 2024-02-15 Toyota Motor Engineering & Manufacturing North America, Inc. Systems and methods for measuring a reaction of a user to an advertisement

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6320497B1 (en) * 1999-07-27 2001-11-20 Kubota Corporation Display control system for a working vehicle
US20040017300A1 (en) * 2002-07-25 2004-01-29 Kotzin Michael D. Portable communication device and corresponding method of operation
US20040150514A1 (en) * 2003-02-05 2004-08-05 Newman Timothy J. Vehicle situation alert system with eye gaze controlled alert signal generation
US20060071766A1 (en) * 2000-03-02 2006-04-06 O'brien Frank Tire pressure alert system
US7209833B2 (en) * 2004-01-19 2007-04-24 Denso Corporation Collision possibility determination device
US7400233B2 (en) * 2005-05-30 2008-07-15 Honda Motor Co., Ltd. Travel safety apparatus for vehicle
US7643737B2 (en) * 2006-03-27 2010-01-05 Honda Motor Co., Ltd. Line of sight detection apparatus
US20100079270A1 (en) * 2008-10-01 2010-04-01 Wolfgang Krautter Method for displaying a visual warning signal
US8085140B2 (en) * 2006-09-04 2011-12-27 Panasonic Corporation Travel information providing device
US20120050068A1 (en) * 1997-08-25 2012-03-01 Donnelly Corporation Interior rearview mirror system
US8144002B2 (en) * 2008-03-31 2012-03-27 Hyundai Motor Company Alarm system for alerting driver to presence of objects
US8599027B2 (en) * 2010-10-19 2013-12-03 Deere & Company Apparatus and method for alerting machine operator responsive to the gaze zone
US20160023604A1 (en) * 2013-07-08 2016-01-28 LightSpeed Automotive Technology Head-Up Display Controller

Family Cites Families (55)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5883605A (en) * 1992-02-25 1999-03-16 Gentex Corporation Automatic electrochromic control of light level of vacuum fluorescent display
DE4211728A1 (en) 1992-04-08 1993-10-14 Zeiss Carl Fa Holographic display device e.g. for vehicle or aircraft head=up display - uses curved windscreen incorporating monomode waveguide for supplied light and holographic gratings
JPH08272321A (en) * 1995-03-31 1996-10-18 Toyoda Gosei Co Ltd External display device of vehicle
US7346374B2 (en) * 1999-05-26 2008-03-18 Johnson Controls Technology Company Wireless communications system and method
US7449081B2 (en) * 2000-06-21 2008-11-11 E. I. Du Pont De Nemours And Company Process for improving the emission of electron field emitters
US6580973B2 (en) * 2000-10-14 2003-06-17 Robert H. Leivian Method of response synthesis in a driver assistance system
US7565230B2 (en) * 2000-10-14 2009-07-21 Temic Automotive Of North America, Inc. Method and apparatus for improving vehicle operator performance
JP3846872B2 (en) 2002-06-27 2006-11-15 パイオニア株式会社 Driver mental state information provision system
DE102004004302A1 (en) * 2003-02-03 2004-08-12 Denso Corp., Kariya Vehicle remote control air conditioning system has a control unit that activates only a ventilation or climate control component of the air conditioning system in order to reduce battery power consumption
US20050084659A1 (en) * 2003-10-20 2005-04-21 General Atomics Vehicle windshield head-up display
US7801283B2 (en) * 2003-12-22 2010-09-21 Lear Corporation Method of operating vehicular, hands-free telephone system
DE102004005816B4 (en) 2004-02-06 2007-02-08 Audi Ag motor vehicle
US7413328B2 (en) * 2004-12-30 2008-08-19 Honeywell International Inc. Remotely coupled hybrid HUD backlight
DE102005059216A1 (en) 2005-07-16 2007-01-25 Ralf Michel Supervision system in particular for motorcycle, comprises units for determination and evaluation of driving performance
JP4617226B2 (en) * 2005-08-30 2011-01-19 本田技研工業株式会社 Vehicle display device
KR100828965B1 (en) * 2006-07-31 2008-05-13 삼성전자주식회사 Method and apparatus for setting environment of cars in portable terminal
EP2084629A1 (en) * 2006-11-14 2009-08-05 Johnson Controls Technology Company System and method of synchronizing an in-vehicle control system with a remote source
US20080158510A1 (en) * 2007-01-02 2008-07-03 Gm Global Technology Operations, Inc. Apparatus And Method For Displaying Information Within A Vehicle Interior
WO2008086216A2 (en) * 2007-01-05 2008-07-17 Johnson Controls Technology Company System and method for customized audio prompting
CN101950122B (en) * 2007-05-17 2012-01-04 Prysm公司 Multilayered screens with light-emitting stripes for scanning beam display systems
US7908060B2 (en) * 2007-07-31 2011-03-15 International Business Machines Corporation Method and system for blind spot identification and warning utilizing portable and wearable devices
JP2009156898A (en) * 2007-12-25 2009-07-16 Seiko Epson Corp Display device
US8311734B2 (en) * 2008-07-01 2012-11-13 Sony Corporation Automatic speed limit adjust for road conditions
DE102009010623A1 (en) 2009-02-26 2010-09-02 Hella Kgaa Hueck & Co. Device for issuing visual warning information to driver of vehicle, particularly motor vehicle, has vehicle window pane, through which vehicle driver visually captures area of environment of vehicle
US8564502B2 (en) * 2009-04-02 2013-10-22 GM Global Technology Operations LLC Distortion and perspective correction of vector projection display
FR2946336B1 (en) * 2009-06-03 2011-05-20 Saint Gobain LAMINATED GLAZING FOR HIGH HEAD VISUALIZATION SYSTEM
US20110012720A1 (en) * 2009-07-15 2011-01-20 Hirschfeld Robert A Integration of Vehicle On-Board Diagnostics and Smart Phone Sensors
US20110025584A1 (en) * 2009-07-29 2011-02-03 Gm Global Technology Operations, Inc. Light-emitting diode heads-up display for a vehicle
KR20110038563A (en) * 2009-10-08 2011-04-14 최운호 Method, vehicle terminal, biometrics card and system for controlling vehicle through authenticating driver, and method for providing passenger protecting/tracking function using biometrics card and terminal
US8498757B2 (en) * 2009-10-09 2013-07-30 Visteon Global Technologies, Inc. Portable and personal vehicle presets
CN201525262U (en) * 2009-11-25 2010-07-14 王辉 Automobile front windshield glass with transparent LCD (liquid crystal display) device
US8384534B2 (en) * 2010-01-14 2013-02-26 Toyota Motor Engineering & Manufacturing North America, Inc. Combining driver and environment sensing for vehicular safety systems
KR101177660B1 (en) * 2010-01-20 2012-08-27 최인주 Vehicle management system through radio relay of vehicle remote controller
KR20120075672A (en) * 2010-12-29 2012-07-09 성균관대학교산학협력단 System and method for safe driving induction using detection of gaze direction
DE102011009840A1 (en) * 2011-01-31 2012-08-02 Continental Automotive Gmbh operating device
WO2012135018A2 (en) * 2011-03-25 2012-10-04 Tk Holdings Inc. System and method for determining driver alertness
KR20120113579A (en) * 2011-04-05 2012-10-15 현대자동차주식회사 Apparatus and method for displaying road guide information on the windshield
US9348492B1 (en) * 2011-04-22 2016-05-24 Angel A. Penilla Methods and systems for providing access to specific vehicle controls, functions, environment and applications to guests/passengers via personal mobile devices
US20120303178A1 (en) * 2011-05-26 2012-11-29 Hendry Jeffrey C Method and system for establishing user settings of vehicle components
KR20120136721A (en) * 2011-06-09 2012-12-20 현대자동차주식회사 Apparatus and method for alarming use of mobile phone on driving
KR101337014B1 (en) * 2011-07-12 2013-12-05 주식회사 팬택 Mobile terminal, system and method for controlling ecu of vehicle using the same
EP2564776B1 (en) * 2011-09-02 2019-08-28 Volvo Car Corporation Method, system and computer readable medium embodying a computer program product for determining a vehicle operator's expectation of a state of an object
CN202357886U (en) * 2011-12-09 2012-08-01 常州永旭车辆配件厂 Dashboard of electric vehicle
FR2985042B1 (en) * 2011-12-22 2014-01-17 Saint Gobain DEVICE FOR VISUALIZING AN IMAGE ON A SHEET SUPPORT
US8989961B2 (en) * 2012-04-13 2015-03-24 Htc Corporation Method of controlling interaction between mobile electronic device and in-vehicle electronic system and devices using the same
CN202806308U (en) * 2012-08-21 2013-03-20 惠州市德赛西威汽车电子有限公司 Automotive windshield
US9426274B2 (en) * 2012-09-27 2016-08-23 Intel Corporation Device, method, and system for portable configuration of vehicle controls
JP2014069592A (en) * 2012-09-27 2014-04-21 Mitsubishi Motors Corp Remote control system for on-vehicle equipment
DE102012219924A1 (en) * 2012-10-31 2014-04-30 Bayerische Motoren Werke Aktiengesellschaft Vehicle assistance device
US20140142948A1 (en) * 2012-11-21 2014-05-22 Somya Rathi Systems and methods for in-vehicle context formation
US20140163771A1 (en) * 2012-12-10 2014-06-12 Ford Global Technologies, Llc Occupant interaction with vehicle system using brought-in devices
US8751065B1 (en) * 2012-12-14 2014-06-10 Denso Corporation Smartphone controller of vehicle settings
KR101555444B1 (en) * 2014-07-10 2015-10-06 현대모비스 주식회사 An apparatus mounted in vehicle for situational awareness and a method thereof
CN105313898B (en) * 2014-07-23 2018-03-20 现代摩比斯株式会社 Driver status induction installation and its method
US20160109701A1 (en) * 2014-10-15 2016-04-21 GM Global Technology Operations LLC Systems and methods for adjusting features within a head-up display

Patent Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8779910B2 (en) * 1997-08-25 2014-07-15 Donnelly Corporation Interior rearview mirror system
US20120050068A1 (en) * 1997-08-25 2012-03-01 Donnelly Corporation Interior rearview mirror system
US6320497B1 (en) * 1999-07-27 2001-11-20 Kubota Corporation Display control system for a working vehicle
US20060071766A1 (en) * 2000-03-02 2006-04-06 O'brien Frank Tire pressure alert system
US7423522B2 (en) * 2000-03-02 2008-09-09 Donnelly Corporation Tire pressure alert system
US20040017300A1 (en) * 2002-07-25 2004-01-29 Kotzin Michael D. Portable communication device and corresponding method of operation
US20040150514A1 (en) * 2003-02-05 2004-08-05 Newman Timothy J. Vehicle situation alert system with eye gaze controlled alert signal generation
US7209833B2 (en) * 2004-01-19 2007-04-24 Denso Corporation Collision possibility determination device
US7400233B2 (en) * 2005-05-30 2008-07-15 Honda Motor Co., Ltd. Travel safety apparatus for vehicle
US7643737B2 (en) * 2006-03-27 2010-01-05 Honda Motor Co., Ltd. Line of sight detection apparatus
US8085140B2 (en) * 2006-09-04 2011-12-27 Panasonic Corporation Travel information providing device
US8144002B2 (en) * 2008-03-31 2012-03-27 Hyundai Motor Company Alarm system for alerting driver to presence of objects
US20100079270A1 (en) * 2008-10-01 2010-04-01 Wolfgang Krautter Method for displaying a visual warning signal
US8390440B2 (en) * 2008-10-01 2013-03-05 Robert Bosch Gmbh Method for displaying a visual warning signal
US8599027B2 (en) * 2010-10-19 2013-12-03 Deere & Company Apparatus and method for alerting machine operator responsive to the gaze zone
US20160023604A1 (en) * 2013-07-08 2016-01-28 LightSpeed Automotive Technology Head-Up Display Controller

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170294065A1 (en) * 2014-04-18 2017-10-12 Gentex Corporation Trainable transceiver and cloud computing system architecture systems and methods
US10096186B2 (en) * 2014-04-18 2018-10-09 Gentex Corporation Trainable transceiver and cloud computing system architecture systems and methods
US10127804B2 (en) 2014-04-18 2018-11-13 Gentex Corporation Trainable transceiver and cloud computing system architecture systems and methods
US20170054837A1 (en) * 2014-05-09 2017-02-23 Samsung Electronics Co., Ltd. Terminal and method for displaying caller information
US10166996B2 (en) * 2017-02-09 2019-01-01 Toyota Motor Engineering & Manufacturing North America, Inc. Systems and methods for adaptively communicating notices in a vehicle
US11312300B1 (en) 2021-01-29 2022-04-26 Toyota Motor Engineering & Manufacturing North America, Inc. Object notification systems for identifying and notifying vehicles of relevant objects

Also Published As

Publication number Publication date
US9688287B2 (en) 2017-06-27
DE112014001607T5 (en) 2015-12-24
US10131364B2 (en) 2018-11-20
US20160085070A1 (en) 2016-03-24
WO2014176474A1 (en) 2014-10-30
WO2014176473A1 (en) 2014-10-30
US20160082979A1 (en) 2016-03-24
US20160041562A1 (en) 2016-02-11
WO2014176478A1 (en) 2014-10-30
CN105324268A (en) 2016-02-10
DE112014001607B4 (en) 2021-09-02
WO2014176476A1 (en) 2014-10-30

Similar Documents

Publication Publication Date Title
US20160039285A1 (en) Scene awareness system for a vehicle
CN110678371B (en) Vehicle control system, vehicle control method, and storage medium
US9738224B2 (en) Vehicle vision system
US9507345B2 (en) Vehicle control system and method
JP5811804B2 (en) Vehicle periphery monitoring device
US9639990B2 (en) Display control apparatus, computer-implemented method, storage medium, and projection apparatus
CN108621943B (en) System and method for dynamically displaying images on a vehicle electronic display
KR102531888B1 (en) How to operate a display device in a car
CN106394553A (en) Driver assistance apparatus and control method for the same
US20070124071A1 (en) System for providing 3-dimensional vehicle information with predetermined viewpoint, and method thereof
CN106564432A (en) Apparatus and method for controlling viewing angle for vehicle, and vehicle including the apparatus
CN108791062B (en) Dynamic information system and method of operation
CN110462702B (en) Travel route providing system, control method thereof, and medium
CN113165510B (en) Display control device, method, and computer program
JP6186905B2 (en) In-vehicle display device and program
US10902273B2 (en) Vehicle human machine interface in response to strained eye detection
CN109415018A (en) Method and control unit for digital rearview mirror
JP2018203014A (en) Imaging display unit
US20230373309A1 (en) Display control device
JP6589775B2 (en) Vehicle display control device and vehicle display system
US11828947B2 (en) Vehicle and control method thereof
CN114103982A (en) Display control device for vehicle, display control method for vehicle, computer-readable storage medium, and display system for vehicle
JP2008040667A (en) Image processor and program
CN116868251A (en) Display control device and display control program
WO2023213416A1 (en) Method and user device for detecting an environment of the user device

Legal Events

Date Code Title Description
AS Assignment

Owner name: GM GLOBAL TECHNOLOGY OPERATIONS LLC, MICHIGAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MATHIEU, ROY J.;SZCZERBA, JOSEPH F.;JONES, MICAH R.;AND OTHERS;SIGNING DATES FROM 20151021 TO 20151022;REEL/FRAME:036882/0669

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION