Motorist user interface sensor

Information

  • Patent Grant
  • 11429230
  • Patent Number
    11,429,230
  • Date Filed
    Monday, November 25, 2019
    5 years ago
  • Date Issued
    Tuesday, August 30, 2022
    2 years ago
Abstract
A steering wheel that includes optoelectronic components, each specific optoelectronic component including a light projector projecting light out of the steering wheel at two different angles, denoted a1 and a2, a light sensor detecting reflections of the light projected by neighboring optoelectronic components by an object above the steering wheel, a lens oriented relative to the light sensor such that the light sensor receives maximum intensity when light enters the lens at either of two particular angles, specifically, when light enters the lens at a particular angle b1, and at a particular angle b2 different than b1, wherein angle b1 views reflections of light projected at angle a1 by the optoelectronic component neighboring the specific optoelectronic component on one side, and angle b2 views reflections of light projected at angle a2 by the optoelectronic component neighboring the specific optoelectronic component on the side opposite the one side.
Description
FIELD OF THE INVENTION

The field of the present invention is user interfaces for motorists and passengers, via the steering wheel and overhead console in a vehicle.


BACKGROUND OF THE INVENTION

Prior art user interfaces on steering wheels associate a function with an absolute position on the steering wheel. This is conceptually analogous to a touch-sensitive screen displaying icons where the user touches the location on the screen at which the icon is located to activate the icon.


Reference is made to FIG. 1, which is a simplified illustration of a prior art steering wheel. FIG. 1 shows a steering wheel 400, including a circular grip 401, also known as a ring member, one or more connecting members 402-404 that connect grip 401 to steering column 407, and buttons 405 and 406 on connecting members 402 and 403; namely, button 405 is used to answer an incoming phone call on the vehicle's speaker phone, and button 406 hangs up the call.


In contrast to user interfaces based on absolute positioning, the computer mouse introduced a user interface for controlling a cursor based on relative positioning. Namely, the mouse cursor moves on the screen in a direction that the mouse moves from point A to point B, but this movement is not at all contingent on the actual coordinates—the absolute positions—of points A and B. This shift from absolute positioning to relative positioning frees the user from having to look at, or be aware of, the location of the mouse on the table. The user only has to control the direction in which the mouse moves on the table, which he can do without looking at the mouse. One of the objectives of the present invention is to provide a user interface that does not require a driver to take his eyes off the road.


Overhead consoles have multiple functions that are located in a confined space. The functions need to be easily identified by the user and intuitive in operation to keep the driver's attention on the road and must comply with federal regulations specifying design, construction, performance and durability requirements for motor vehicles and regulated components, systems and design features. In the United States these regulations are contained in the Federal Motor Vehicle Safety Standards (FMVSS), in Canada in the Canada Motor Vehicle Safety Standards (CMVSS), and other countries have adopted UN regulations developed by the World Forum for Harmonization of Vehicle Regulations. FMVSS No. 118 regulating window, partition and roof panel systems, aims at preventing accidental operation, e.g., by a child, leg or knee. Thus, for example, roof panel switches that can close a roof panel by momentary switch actuation must be protected from operation by a 20 mm radius sphere which represents a knee or leg.


Reference is made to FIG. 2, which is an image of a prior art vehicle overhead console. FIG. 2 shows an overhead console 600 having map light 601, roof panel control 602 and slide switch 603 for controlling the vehicle's internal cabin light.


SUMMARY

The present disclosure relates to user interfaces for on-board vehicle systems, and teaches a user interface that does not require the user to look at the steering wheel in order to activate a function. The present disclosure teaches user gestures that can be mapped to a variety of applications.


There is thus provided in accordance with an embodiment of the present invention a steering wheel including a series of optoelectronic components mounted in the steering wheel grip, each specific optoelectronic component including a PCB, a light projector, having at least one light pulse emitter, mounted on the PCB, the light projector projecting light out of the steering wheel grip at two different angles, denoted a1 and a2, a light sensor, having at least one light detector mounted on the PCB, the light sensor detecting reflections of the light projected by the light pulse emitters of the optoelectronic components that neighbor the specific optoelectronic component on two opposite sides thereof, the light being reflected by an object above the steering wheel grip, a lens oriented relative to the light sensor in such a manner that the light sensor receives maximum intensity when light enters the lens at either of two particular angles, specifically, (i) the light sensor receives maximum intensity when the light reflected by the object enters the lens at a particular angle b1, and (ii) the light sensor also receives maximum intensity when light reflected by the object enters the lens at a particular angle b2 different than b1, wherein angle b1 views reflections of light projected at angle a1 by the light projector of the optoelectronic component neighboring the specific optoelectronic component on one side, and angle b2 views reflections of light projected at angle a2 by the light projector of the optoelectronic component neighboring the specific optoelectronic component on the side opposite the one side.


There is additionally provided, in accordance with an embodiment of the present invention, a steering wheel including a series of optoelectronic components mounted in the steering wheel grip, each specific optoelectronic component including a PCB, at least two light pulse emitters mounted on the PCB for projecting light out of the steering wheel grip, at least two light detectors mounted on the PCB detecting reflections of the light projected by light pulse emitters of optoelectronic components on opposite sides of the specific optoelectronic component, the light being reflected by an object above the steering wheel grip, and a lens oriented relative to the light detectors in such a manner that each light detector receives maximum intensity when light enters the lens at a particular angle, specifically, (i) one or more of the light pulse emitters project light out of the steering wheel grip at an angle a1, (ii) one or more others of the light pulse emitters project light out of the steering wheel grip at an angle a2 different than a1, (iii) one or more of the light detectors receive maximum intensity when the light reflected by the object enters the lens at a particular angle b1, and (iv) one or more others of the light detectors receive maximum intensity when light reflected by the object enters the lens at a particular angle b2 different than b1.


There is further provided, in accordance with an embodiment of the present invention, a method for detecting driver input by providing a series of optoelectronic components in a steering wheel grip, each specific optoelectronic component projecting light beams in two emission directions, denoted by angles a1 and a2, out of the steering wheel grip, and orienting a lens within each specific optoelectronic component so as to provide two viewing angles, denoted b1 and b2, that detect maximum intensity of reflections of the light projected by optoelectronic components that neighbor the specific optoelectronic component, the light being reflected by a reflective object above the steering wheel grip, wherein viewing angle b1 receives maximum intensity when light projected by a first neighboring optoelectronic component at angle a1 is reflected by the object, and viewing angle b2 receives maximum intensity when light projected by a second neighboring optoelectronic component at angle a2 is reflected by the object.


The present disclosure also relates to controls mounted in overhead consoles in vehicles. The user interface described hereinbelow can be applied to additional areas in and around the vehicle, in addition to the overhead console.


There is thus provided in accordance with an embodiment of the present invention a vehicle overhead console including a touch sensitive surface, a gesture sensor operable to identify in-air wave gestures performed in an airspace opposite the surface, a processor, connected to the surface and the sensor, controlling a plurality of a vehicle systems in response to a plurality of respective, multi-gesture operations, each multi-gesture operation including a touch gesture on the surface indicating a corresponding one of the vehicle systems, and an in-air wave gesture in the airspace opposite the surface indicating a function to be performed by the indicated vehicle system.


There is additionally provided, in accordance with an embodiment of the present invention, a method for controlling vehicle systems, including providing a touchscreen display in the interior roof of a vehicle, rendering representations of a plurality of user controlled vehicle systems, providing a gesture sensor operable to detect in-air wave gestures in an airspace opposite the touchscreen display, and performing a vehicle system function in response to a multi-gesture operation including a touch gesture on one of the rendered representations followed by an in-air wave gesture in the airspace opposite the touchscreen display.





BRIEF DESCRIPTION OF THE DRAWINGS

The present invention will be more fully understood and appreciated from the following detailed description, taken in conjunction with the drawings in which:



FIG. 1 is a simplified illustration of a prior art steering wheel;



FIG. 2 is an image of a prior art vehicle overhead console;



FIG. 3 is a simplified illustration of a steering wheel, in accordance with an embodiment of the present invention;



FIGS. 4 and 5 are simplified illustrations of a thumb swipe gesture detected by a steering wheel, in accordance with an embodiment of the present invention;



FIG. 6 is a simplified illustration of the light beams and corridors of maximum light detection provided for a gesture-detecting steering wheel, in accordance with an embodiment of the present invention;



FIG. 7 is a simplified illustration of optoelectronic components mounted on a PCB in a steering wheel, in accordance with an embodiment of the present invention;



FIG. 8 is the illustration of FIG. 7 with the addition of light beams and corridors of maximum light detection provided for a gesture-detecting steering wheel, in accordance with an embodiment of the present invention;



FIGS. 9-11 are simplified illustrations of optoelectronic components mounted in a segment of a steering wheel, in accordance with an embodiment of the present invention;



FIG. 12 is a simplified illustration of two optoelectronic components in a steering wheel, in accordance with an embodiment of the present invention;



FIG. 13 is a simplified illustration of an optoelectronic component, in accordance with an embodiment of the present invention;



FIG. 14 is an exploded view of the optoelectronic component of FIG. 13, in accordance with an embodiment of the present invention;



FIG. 15 is a simplified illustration of the optoelectronic component of FIG. 13 viewed from a different perspective, in accordance with an embodiment of the present invention;



FIG. 16 is a simplified illustration of a touch gesture on a vehicle overhead console touchscreen, in accordance with an embodiment of the present invention;



FIG. 17 is a simplified illustration of a hand performing a wave gesture in the airspace opposite the vehicle overhead console touchscreen of FIG. 16, in accordance with an embodiment of the present invention;



FIGS. 18 and 19 are simplified illustrations of different stages of a hand wave gesture in the airspace opposite the vehicle overhead console touchscreen of FIG. 16, in accordance with an embodiment of the present invention; and



FIGS. 20 and 21 are simplified illustrations of a touchscreen user interface for a vehicle overhead console, in accordance with an embodiment of the present invention.





In the disclosure and figures, the following numbering scheme is used. Like numbered elements are similar but not necessarily identical.
















Type of element
Numbering range









steering wheel
100-104



PCB
105-109



light emitter
110-114



light detector
115-119



lens or optical part
120-129



optoelectronic component
150-159



outgoing light beam
160-169



viewing angle
170-179



hotspot
180-189



hand
190-199



prior art elements
400-410



prior art vehicle overhead console
600



map light
601



roof panel control
602



cabin light slide switch
603



touchscreen display
604



map light icon
605-607



open/close sunroof icon
608



sunroof pop-up vent icon
609



open/close moonroof icon
610



finger
611



fingers
612



in-air detection plane
613



movement direction arrows
614



vehicle interior roof
615



rearview mirror
616



front windshield
617










DETAILED DESCRIPTION

Aspects of the present disclosure relate to light-based touch and gesture controls that allow a driver to keep his hands on the steering wheel and eyes on the road while operating electronic devices and automated features in a vehicle. Detection of ergonomic gestures is enabled by the sensors described herein, and the invention includes methods of interacting with systems in the vehicle using the ergonomic gestures on the steering wheel.


Reference is made to FIG. 3, which is a simplified illustration of a steering wheel, in accordance with an embodiment of the present invention. FIG. 3 shows steering wheel 100. The upper-right segment of the steering wheel ring member has two concentric bands of proximity sensors, 180 and 181. In FIG. 3 each proximity sensor has a light emitter (not shown) that emits a directed light beam 160 outwards from the steering wheel, and a light detector (not shown) having a specific viewing angle 170. In the figures, emitter beams and viewing angles are drawn using dashed lines. The emitter beams are drawn using longer dashes than the viewing angles. The proximity sensor light detector detects an object when light beam 160 is reflected back at the detector's specific viewing angle 170. FIG. 3 shows the intersection between each emitter beam 160 and a respective detector viewing angle 170. The intersection is the location at which an object will be maximally detected by the proximity sensor. The steering wheel's concentric bands of proximity sensors identify swipe gestures across the width of the steering wheel grip. Viewing angle 170 is also referred to as a corridor of maximum detection.


Reference is made to FIGS. 4 and 5, which are simplified illustrations of a thumb swipe gesture detected by a steering wheel, in accordance with an embodiment of the present invention. FIGS. 4 and 5 illustrate a swipe gesture performed by the driver's thumb of hand 190 gripping steering wheel 100. In FIG. 4 the driver's radially extended thumb is detected by inner band 181 of proximity sensors, and in FIG. 5 the driver's thumb is detected by outer band 180 of proximity sensors. This gesture is efficient and comfortable for the driver gripping the wheel, as the driver radially extends her thumb from the palm (FIG. 4) and then adducts the thumb to sweep it upward across the width of the steering wheel (FIG. 5). The reverse gesture—abducting the thumb from the position in FIG. 5 to the position in FIG. 4—is also efficient and comfortable for the driver. Other gestures, such as moving a finger or hand along one of the bands 180 and 181, are also supported by embodiments of the present invention.


Reference is made to FIG. 6, which is a simplified illustration of the light beams and corridors of maximum light detection provided for a gesture-detecting steering wheel, in accordance with an embodiment of the present invention. FIG. 6 shows two complete concentric circles 180, 181 of proximity sensors along the grip of steering wheel 100.


Reference is made to FIG. 7, which is a simplified illustration of optoelectronic components mounted on a PCB in a steering wheel, in accordance with an embodiment of the present invention. FIG. 7 shows the steering wheel of FIG. 6 with the upper portion of the steering wheel grip removed to expose PCB 105 and a plurality of optoelectronic components 150 mounted upright on PCB 105; i.e., optoelectronic components 150 are oriented substantially perpendicular to PCB 105 and to the surface of the steering wheel grip facing the driver. Some steering wheel ring members include an inner ring made of a stiff material shaped into a bent or curved cross-section that strengthens the steering wheel ring, as discussed in U.S. Pat. No. 3,714,844, incorporated herein in its entirety by reference, and illustrated therein in FIGS. 13A-13N. In certain embodiments of the invention, optoelectronic components 150 are mounted between the walls formed by the bent or curved inner ring cross-section to protect optoelectronic components 150. Each optoelectronic component 150 includes a PCB, one or more emitters, one or more detectors and a lens that directs light beams 160 and creates viewing angles 170 for the detectors. The upper portion of the steering wheel grip, not shown in FIG. 7, is light-transmissive to enable beams 160 to travel out of steering wheel 100 and to enable reflections of beams 160 to re-enter steering wheel 100 and reach the detectors. When optoelectronic components 150 are mounted between the walls formed by the bent or curved inner ring cross-section, beams 160 pass through the opening in the bent or curved cross-section.


Reference is made to FIG. 8, which is the illustration of FIG. 7 with the addition of light beams and corridors of maximum light detection provided for a gesture-detecting steering wheel, in accordance with an embodiment of the present invention. FIG. 8 shows the steering wheel of FIG. 7, optoelectronic components 150 and the light beams and viewing angles that enable the proximity sensors to detect objects. FIG. 8 shows the two concentric bands 180 and 181 of proximity sensors by illustrating intersections between emitter beams and viewing angles for all of the optoelectronic components.


Reference is made to FIGS. 9-11, which are simplified illustrations of optoelectronic components mounted in a segment of a steering wheel, in accordance with an embodiment of the present invention. FIGS. 9-11 show a portion of steering wheel 100 without its upper surface, exposing PCB 105 and six or seven optoelectronic components 150. The emitter beams 160 and detector viewing angles 170 for each optoelectronic component 150 are shown. The intersection between each emitter beam 160 and the viewing angle 170 of a corresponding detector is the location at which an object, reflecting beam 160, is detected. Each optoelectronic component 150 in FIGS. 9-11 has two emitter beams and two detector viewing angles. The emitter beams are projected out of the component at divergent angles, denoted a1 and a2, and the viewing angles are denoted b1 and b2. Projection angles a1 and a2 and viewing angles b1 and b2 are shown in FIG. 13. FIGS. 9-11 also show that the two emitter beams of optoelectronic component n are intersected by the viewing angles of optoelectronic component n−1 and optoelectronic component n+1, respectively. This relationship between neighboring optoelectronic components is further illustrated in FIG. 12.


Reference is made to FIG. 12, which is a simplified illustration of two optoelectronic components in a steering wheel, in accordance with an embodiment of the present invention. FIG. 12 shows two, neighboring optoelectronic components 151 and 152 mounted on PCB 105.


Optoelectronic component 151 projects emitter beams 161 and 162 and detects light arriving at viewing angles 171 and 172. Optoelectronic component 152 projects emitter beams 163 and 164 and detects light arriving at viewing angles 173 and 174. A location along an emitter beam that is located along the viewing angle of a detector is referred to as a “hotspot”. Reflections of the light beam by an object at this location are maximally detected by the detector. In certain embodiments of the invention, the hotspots are located at, or near, the outer surface of the steering wheel grip facing the driver. In other embodiments of the invention, the hotspots are located in airspace above the outer surface of the steering wheel grip facing the driver. In still other embodiments of the invention, some of the hotspots are nearer to the steering wheel grip than others. Two hotspots 182 and 183 are indicated in FIG. 12. Hotspot 183 is at the intersection of emitter beam 163 and viewing angle 171. Hotspot 182 is at the intersection of emitter beam 162 and viewing angle 173. Hotspot 183 is located along outer band 180 of the proximity sensors in FIGS. 3-6, and hotspot 182 is located along inner band 181 of the proximity sensors in FIGS. 3-6. Thus the concentric bands or arcs of the proximity sensors in FIGS. 3-6 are actually hotspots created by the emitters and detectors of optoelectronic components 150. The two emitter beams emitted by each optoelectronic component define a first plane, and the two viewing angles of that same optoelectronic component define a second plane that intersects the first plane. This is illustrated in FIG. 13.


Reference is made to FIG. 13, which is a simplified illustration of an optoelectronic component, in accordance with an embodiment of the present invention. FIG. 13 illustrates a single optoelectronic component. The two emitter beams 161 and 162, that are projected at angles denoted a1 and a2, respectively, define a first plane and the two viewing angles 171 and 172, denoted b1 and b2, respectively, define a second plane. In some embodiments of the invention these planes are orthogonal; in other embodiments these planes intersect, but not at right angles.



FIG. 13 shows two optical parts 120 and 121 separated by air. A single emitter, e.g., a vertical cavity surface-emitting laser (VCSEL) or a light-emitting diode (LED), is mounted underneath optical part 120 and its light is split into beams 161 and 162. A single photodiode detector is mounted underneath optical part 121. The combination of lenses and air-to-plastic surfaces in optical parts 120 and 121 provides two viewing angles 171 and 172 to this single photodiode detector. In certain embodiments of the invention each light beam is generated by a separate light emitter. In certain other embodiments of the invention a separate detector is provided for each viewing angle. And in still other embodiments of the invention each light beam is generated by a separate light emitter and a separate detector is provided for each viewing angle.


Reference is made to FIG. 14, which is an exploded view of the optoelectronic component of FIG. 13, in accordance with an embodiment of the present invention. FIG. 14 shows emitter 110 and photodiode detector 115 mounted on PCB 106, covered by optical parts 120 and 121. In some embodiments, optical parts 120 and 121 are molded as a single part that is mounted on PCB 106.


Referring back to FIGS. 9-11, it can be seen that each viewing angle can be configured to intersect two emitter beams emitted by two respective optoelectronic components. These two hotspots are at different heights above the steering wheel grip, and enable detecting when the driver's hand approaches the steering wheel grip, namely, the further hotspot detects the hand first, and the nearer hotspot detects the hand second. The nearer hotspot is formed by the detector on optoelectronic component n detecting a light beam emitted by optoelectronic component n+1, and the further hotspot is formed by the detector on optoelectronic component n detecting a light beam emitted by optoelectronic component n+2.


Certain proximity sensors provide a large array of light emitters and light detectors mounted on a single PCB. This configuration requires that each emitter and each detector be placed precisely on the PCB so that they are positioned correctly in relation to their respective lenses. However, a long PCB may suffer from significant bending or warping when exposed to heat, causing a misalignment between the lenses and the components. Furthermore, when one lens part is used for more than one component, it may be difficult to position the lens part correctly. In order to overcome these problems, embodiments of the present invention use multiple PCBs, and mount a small number of emitters and detectors on each PCB, e.g., only one or two emitters and only one or two detectors.


Reference is made to FIG. 15, which is a simplified illustration of the optoelectronic component of FIG. 12 viewed from a different perspective, in accordance with an embodiment of the present invention. FIG. 15 is a perspective view from above of the optoelectronic component of FIG. 13. The outer surface (near the viewer) of the central portion of optical part 120 has two inclined surfaces that direct light in the two directions 161 and 162.


Several options are available for connecting the optoelectronic components to each other. One option is to provide one or more rigid, main PCBs 105 along the entire steering wheel grip, and soldering or otherwise connecting each optoelectronic component's PCB 106 to the underlying rigid PCB 105. A second option is to use a rigid-flex PCB, where individual optoelectronic component PCBs 106 are connected by flexible circuit substrates, and no PCB 105 is required. The output signals from the optoelectronic components 150 are typically weak signals and therefore the connector between optoelectronic components may need to be shielded.


Aspects of the present disclosure relate to touch and mid-air gestures used to replace mechanical switches in vehicle overhead consoles for improved appearance, reduced cost, thinner profiles, reduced weight, and easy reconfiguration for different vehicles. The present invention replaces the mechanical buttons in prior art overhead consoles with touch and gesture controls, reducing complexity and part cost. In this description the terms “mid-air” and “in-air” are used interchangeably.


According to the present invention, an improved user interface for vehicle overhead console functions is provided that increases driver interaction while reducing driver distraction. The user interface of the present invention requires less time and attention to operate than prior art overhead consoles, thereby allowing the driver to focus more attention on the road. At the same time, the user interface of the present invention enables more nuanced control of console-operated functions by the driver. Thus, whereas prior art cabin lighting systems offer only rudimentary light controls, the present invention enables more comprehensive cabin mood lighting features while reducing the complexity of using these features.


The present invention combines a reconfigurable graphic display with on-surface gestures and mid-air gestures to operate the overhead console functions. The present invention requires a multi-gesture operation, namely, a touch gesture followed by a mid-air gesture to activate an overhead console function, reducing the possibility of inadvertent activation of functions. Reducing the possibility of inadvertent activation of functions is a motivation for requirements in FMVSS No. 118.


Reference is made to FIG. 16, which is a simplified illustration of a touch gesture on a vehicle overhead console touchscreen, in accordance with an embodiment of the present invention. FIG. 16 shows a touchscreen 604 mounted on an interior roof 615 of a vehicle to function as an overhead console. FIG. 16 shows front windshield 617 and rearview mirror 616, to illustrate that display 604 is located where vehicle overhead consoles are typically mounted in a vehicle. Finger 611 is shown touching display 604. Display 604 presents a graphical user interface (GUI), e.g., icons for the different functions activated by the overhead console. When the user touches or taps an icon, as illustrated in FIG. 16, that function is selected on the display but not activated. In order to activate the function, the user performs a mid-air gesture in the airspace opposite display 604, e.g., a wave gesture. The direction and extent of the wave gesture determine how the function is activated. E.g., when a lighting function is selected, the forward/backward direction of the wave indicates whether to increase or decrease the lighting, and the length of the wave gesture indicates how much to increase or decrease the lighting. Furthermore, a left/right direction of the wave can be used to select another lighting variable, such as lighting mood or color. In another example, when a roof panel function is selected, the forward/backward direction of the wave indicates whether to open or close the panel, and the length of the wave gesture indicates how much to open or close the panel. Furthermore, a left/right direction of the wave can be used to select another roof panel function, such as pop-up the panel as an air vent.


In certain embodiments of the invention, the touch functionality of display 604 is enabled by an optical proximity sensor as described in U.S. Pat. No. 9,164,625, entitled OPTICAL PROXIMITY SENSORS, and in U.S. Pat. No. 9,921,661, entitled OPTICAL PROXIMITY SENSOR AND ASSOCIATED USER INTERFACE, both of which are incorporated herein by reference in their entirety. This optical proximity sensor creates a detection plane substantially parallel to the surface of display 604.


In certain embodiments of the invention, display 604 is an electronic paper display. In other embodiments of the invention, display 604 is mounted behind a Fiber Optic Faceplate (FOFP) which is a coherent fiber optic plate that precisely transmits an image from its input surface to its output surface. Thus, an image on the display is transmitted by the FOFP to the exposed surface of the plate and appears to the user as if the image is on the exposed surface of the faceplate. This protects the display and provides a rich user experience. In other embodiments of the invention, icons or other representations of functions are printed, engraved or embossed, on material such as leather, wood, glass or plastic mounted in the interior roof of the vehicle cabin, and the sensor detects touches on these icons. In other embodiments of the invention, icons or other representations of functions are printed, engraved or embossed on the interior roof of the vehicle cabin. In some embodiments of the invention, icons are etched into a translucent plastic plate mounted in the roof of the vehicle cabin and light emitters are arranged around the plate whereby when a light emitter is activated a corresponding etched icon reflects the emitter light and is clearly visible. Initially, all icons are thus illuminated, and when an icon is selected by the user only the selected icon is illuminated by a respective light emitter and the remaining emitters are turned off making visibility of the remaining icons less prominent.


Reference is made to FIG. 17, which is a simplified illustration of a hand performing a wave gesture in the airspace opposite the vehicle overhead console touchscreen of FIG. 16, in accordance with an embodiment of the present invention. FIG. 17 shows the second gesture required to activate an overhead console function. FIG. 17 shows fingers 612 performing an in-air wave gesture in the airspace opposite display 604. In certain embodiments of the invention, the sensor used to detect the in-air gesture is also an optical proximity sensor as described in U.S. Pat. Nos. 9,164,625, and 9,921,661, directed away from the surface of display 604. The detection plane of this sensor is indicated in FIG. 17 as detection plane 613. The use of this type of proximity sensor to detect in-air gestures is illustrated, inter alia, in FIG. 37 of U.S. Pat. No. 9,921,661. Furthermore, these proximity sensors detect the difference between a single finger 611 and multiple fingers 612, inter alia, based on the number and pattern of detected reflected light beams. Thus, a user interface according to the present invention further prevents inadvertent activation by requiring a single finger to perform the touch gesture and multiple fingers to perform the wave gesture in order to activate the function.


Reference is made to FIGS. 18 and 19, which are simplified illustrations of different stages of a hand wave gesture in the airspace opposite the vehicle overhead console touchscreen of FIG. 16, in accordance with an embodiment of the present invention. FIGS. 18 and 19 show an in-air front-to-back hand wave gesture. In FIG. 18 fingers 612 are opposite the top of display 604, near rearview mirror 616, and in FIG. 19 fingers 612 have moved to the bottom of display 604.


Reference is made to FIGS. 20 and 21, which are simplified illustrations of a touchscreen user interface for a vehicle overhead console, in accordance with an embodiment of the present invention. FIG. 20 shows a user interface for display 604, including icons 605 and 606 indicating driver and passenger map lights, respectively; icon 608 indicating functions for opening and closing the sunroof; icon 609 for popping the roof panel up or down as an air vent; icon 610 indicating functions for opening and closing the moonroof; additional icons marked “OFF,” “DOOR” and “ON” controlling behavior of the interior dome light, to never go on, go on when doors are open, and remain on, respectively.



FIG. 21 shows a user interface for display 604, after the user has selected map light icon 605 in FIG. 20. In FIG. 21 the selected icon is enlarged and the non-selected icons are removed. In this state, hand wave gestures will activate the selected map light, e.g., the driver-side map light.


Tables I-XII hereinbelow provide a detailed inventory of a user interface for overhead console features according to the present invention.


Overhead consoles according to the present invention have fewer components than prior art mechanical consoles. Overhead consoles according to the present invention are also thinner and lighter, and have fewer connections, and therefore fewer potential failure points, than prior art mechanical consoles. When optical touch and gesture sensors are used, visible-lighting optics can share components with the near-infrared sensor optics, e.g., optical touch and gesture sensors include controllers for activating near-infrared LEDs used in the sensors, and these controllers can additionally control visible-light LEDs used to illuminate the cabin. Thus, overall cost is reduced and conventional bulbs are replaced with LEDs.


The present invention enables thinner and lighter packages for the overhead console, and enables using a wide range of materials for the console. It also enables blending the design of the overhead console with the vehicle interior and the vehicle theme. The components—the display, touch sensor and mid-air gesture sensor—can be reused in different vehicle themes. The sensors are offered in packages that provide LED and RGB lighting which can replace the light bulbs used in prior art map lights and interior lights, and enable focusing projected cabin lighting with direct and indirect lighting and creating nuanced mood lighting in the cabin.


The system of a touch sensor detecting touch on a surface on which representations of functions are presented and a gesture sensor detecting gestures in the airspace opposite the surface, and a user interface that combines these two gestures to activate a function, is not limited to overhead consoles. The invention can also be realized inter alia in a rearview mirror and in a side door panel.


In the foregoing specification, the invention has been described with reference to specific exemplary embodiments thereof. It will, however, be evident that various modifications and changes may be made to the specific exemplary embodiments without departing from the broader spirit and scope of the invention. In particular, sensors other than optical sensors may be used to implement the user interface, inter alia capacitive sensors disposed along the circumference of the steering wheel and on the overhead console, or cameras that capture images of the steering wheel and the airspace adjacent the overhead console. Accordingly, the specification and drawings are to be regarded in an illustrative rather than a restrictive sense.









TABLE I







Map light ON/OFF








Feature
Map light ON/OFF





Feature feedback to user
Toggle visible light on/off.


Input action
One finger press > 20 ms < 150 ms.


Input action location
1. Dedicated position on surface (on map



light).



2. Off surface dedicated position (above



map light).


Impact to driver attention
Intuitive attention directed to light.


to the road
Confirmation from light on or off.


Search & glance time


Result unintended
If ON selected inadvertently then power


activation
management turns light off after set time.


Haptic effect
Confirmation pop.


Feedback sound
Click.


Feedback light
Not required.
















TABLE II







Map light brightness








Feature
Map light brightness





Feature feedback to user
Light brightness change.


Input action
One finger hold and then move in sensing



zone left/right up/down.


Input action location
1. Dedicated position on surface.



2. Off surface dedicated position.



3. Off surface adjustment after map



selected on surface.


Impact to driver attention
Intuitive attention directed to light.


to the road
Confirmation from light increase or


Search & glance time
decrease. Haptics could increase



feedback but secondary to light changing.


Result unintended
If dome ON then light dims, if not ON no


activation
action.


Haptic effect
Confirmation & detent/pulse (change



acceleration) with movement.


Feedback sound
Tick or changing click.


Feedback light
Not required.
















TABLE III







Dome light operation selection








Feature feedback to user
Dome light operation selection





Input action
ON & OFF positions the light changes but



DOOR position not readily apparent.


Input action location
A. One finger hold in one of



ON/DOOR/OFF positions for > 150 ms.



B. Toggle or slide to change selection -



One finger in middle position < 150 ms



then finger moves direction.


Impact to driver attention
A.1. Three dedicated positions on surface.


to the road
A.2. Three dedicated positions off surface.


Search & glance time
A.3. Off surface selection after dome



selected on surface.



B.1. Three dedicated positions on surface.



B.2. Three dedicated positions off surface.



B.3. Off surface selection after done



selected on surface.


Result unintended
Search reduced with intuitive location of 3


activation
position selection.



Glance reduced with surface geometry to



feedback finger position.



Haptic feedback reduces glance time.


Haptic effect
If selected OFF or DOOR inadvertently



then map light activated as desired.


Feedback sound
Confirmation and release sweep-peak.


Feedback light
Click.


Feature feedback to user
Not required.
















TABLE IV







Sunroof open AUTOMATIC








Feature
Sunroof open AUTOMATIC





Feature feedback to user
Roof moves to open position plus motor



sound.


Input action
One finger hold in sunroof position and



then move in opposite direction of close -



left/right or up/down.


Input action location
1. Dedicated protected position on



surface.



2. Off surface after roof selected on



surface.


Impact to driver attention
Search reduced with intuitive location and


to the road
surface geometry.


Search & glance time
Glance reduced for OPEN position



selection with geometry and haptic



feedback.


Result unintended
FMVSS 118 Safety function for


activation
Homologation not activated by 20 mm



radius sphere unless auto reverse system


Haptic effect
Confirmation & detent/pulse (change



acceleration) with movement


Feedback sound
Not required


Feedback light
Not required
















TABLE V







Sunroof close AUTOMATIC








Feature
Sunroof close AUTOMATIC





Feature feedback to user
Roof moves to close position plus motor



sound.


Input action
One finger hold in sunroof position and



then move in opposite direction of open -



left/right or up/down.


Input action location
1. Dedicated protected position on



surface.



2. Off surface after Roof selected on



surface.


Impact to driver attention
Search reduced with intuitive location and


to the road
surface geometry.


Search & glance time
Glance reduced for CLOSE position



selection with geometry and haptic



feedback.


Result unintended
FMVSS 118 Safety function for


activation
Homologation not activated by 20 mm



radius sphere unless auto reverse system.


Haptic effect
Confirmation & detent/pulse (change



acceleration) with movement.


Feedback sound
Not required.


Feedback light
Not required.
















TABLE VI







Sunroof vent open AUTOMATIC








Feature
Sunroof vent open AUTOMATIC





Feature feedback to user
Roof moves to vent position plus motor



sound.


Input action
One finger hold in sunroof vent position



and then move in sensing zone left/right



or up/down.


Input action location
1. Dedicated protected position on



surface.



2. Off surface after Roof selected on



surface.


Impact to driver attention
Search reduced with intuitive position


to the road
close to roof open/close.


Search & glance time
Glance reduced for Vent position selection



with haptic feedback.


Result unintended
FMVSS 118 Safety function for


activation
Homologation not activated by 20 mm



radius sphere unless auto reverse system.


Haptic effect
Confirmation & detent/pulse (change



acceleration) with movement.


Feedback sound
Not required.


Feedback light
Not required.
















TABLE VII







Sunroof vent close AUTOMATIC








Feature
Sunroof vent close AUTOMATIC





Feature feedback to user
Roof moves to close position plus motor



sound.


Input action
One finger hold in sunroof vent position



and then move in sensing zone left/right or



up/down.


Input action location
1. Dedicated protected position on surface.



2. Off surface after Roof selected on



surface.


Impact to driver attention
Search reduced with intuitive position


to the road
close to roof open/close.


Search & glance time
Glance reduced for Vent position selection



with haptic feedback.


Result unintended
FMVSS 118 Safety function for


activation
Homologation not activated by 20 mm



radius sphere unless auto reverse system.


Haptic effect
Confirmation & detent/pulse (change



acceleration) with movement.


Feedback sound
Not required.


Feedback light
Not required.
















TABLE VIII







Sunshade open AUTOMATIC








Feature
Sunshade open AUTOMATIC





Feature feedback to user
Shade moves to open position plus motor



sound.


Input action
One finger hold in sunshade position and



then move in sensing zone left/right or



up/down.


Input action location
1. Dedicated position on surface



2. Off surface after Shade selected on



surface.


Impact to driver attention


to the road


Search & glance time


Result unintended
Not FMVSS 118.


activation


Haptic effect
Confirmation & detent/pulse (change



acceleration) with movement.


Feedback sound
Not required.


Feedback light
Not required.
















TABLE IX







Sunshade close AUTOMATIC








Feature
Sunshade close AUTOMATIC





Feature feedback to user
Shade moves to close position plus motor



sound.


Input action
One finger hold in sunshade position and



then move in sensing zone left/right or



up/down.


Input action location
1. Dedicated position on surface.



2. Off surface after Shade selected on



surface


Impact to driver attention


to the road


Search & glance time


Result unintended
Not FMVSS 118.


activation


Haptic effect
Confirmation & detent/pulse (change



acceleration) with movement.


Feedback sound
Not required.


Feedback light
Not required.
















TABLE X







Home link 1, 2, 3








Feature
Home link 1, 2, 3





Feature feedback to user
No readily apparent action.


Input action
One finger hold in 1/2/3 positions for >



150 ms.


Input action location
Dedicated position on surface.


Impact to driver attention
Search reduced with intuitive location and


to the road
graphics.


Search & glance time
Glanced reduced with haptics feedback of



correct selection and function initiation.


Result unintended
Confirmation required since operation not


activation
readily apparent.



Each function should have unique



feedback haptic/light/sound to confirm



correct operation.


Haptic effect
Confirmation and release sweep-peak.


Feedback sound
Click.


Feedback light
Can confirm selection and initiation of



function.
















TABLE XI







SOS Safety








Feature
SOS safety





Feature feedback to user
No readily apparent action.


Input action
One finger hold in SOS position for > 150



ms.


Input action location
Dedicated protected position on surface.


Impact to driver attention
Search reduced with graphics and


to the road
location.


Search & glance time
Glanced reduced with haptic feedback of



selection and initiation.


Result unintended
Safety feature, so protected input with


activation
geometry should have unique feedback



haptic/light/sound.


Haptic effect
Confirmation and release sweep-peak.


Feedback sound


Feedback light
Illumination color change or flash.
















TABLE XII







Storage bin








Feature
Storage bin





Feature feedback to user
Open door.


Input action
One finger hold in bin open position for >



150 ms.


Input action location
Dedicated position on surface.


Impact to driver attention
Search reduced with location on door.


to the road
Glance reduced with haptic confirmation


Search & glance time
and door starts to open.


Result unintended
Operation starts bin open or not with


activation
readily apparent feedback.


Haptic effect
Confirmation and release sweep-peak.


Feedback sound
Click.


Feedback light
Not required.








Claims
  • 1. A steering wheel comprising a series of optoelectronic components mounted in the steering wheel grip, each specific optoelectronic component comprising: a PCB;a light projector, comprising: at least one light pulse emitter mounted on said PCB; andat least one first lens oriented relative to said at least one light pulse emitter in such a manner that said at least one light pulse emitter projects light through the at least one first lens out of said steering wheel grip at two different angles relative to the plane of the steering wheel grip, denoted a1 and a2;a light sensor, comprising at least one light detector mounted on said PCB, the light sensor detecting reflections of the light projected by the light projectors of the optoelectronic components that neighbor the specific optoelectronic component on two opposite sides thereof, the light being reflected by an object above the steering wheel grip;at least one second lens oriented relative to said light sensor in such a manner that said light sensor receives maximum intensity when light enters the at least one second lens at either of two particular angles, specifically, said light sensor receives maximum intensity when the light reflected by the object enters the at least one second lens at a particular angle b1 relative to the plane of the steering wheel grip, and said light sensor also receives maximum intensity when light reflected by the object enters the at least one second lens at a particular angle b2 relative to the plane of the steering wheel grip, b2 being different than b1, wherein reflections of light projected at angle a1 by the light projector of the optoelectronic component neighboring the specific optoelectronic component on one side enter the at least one second lens at angle b1, and reflections of light projected at angle a2 by the light projector of the optoelectronic component neighboring the specific optoelectronic component on the side opposite the one side enter the at least one second lens at angle b2.
  • 2. The steering wheel of claim 1, wherein reflections of light projected at angle a1 by two optoelectronic components positioned on one side of the specific optoelectronic component enter the lens at angle b1.
  • 3. The steering wheel of claim 1, wherein said at least one first lens comprises a beam splitter that splits light from the at least one light pulse emitter into light projected out of the steering wheel grip at angles a1 and a2.
  • 4. The steering wheel of claim 1, wherein said light projector comprises a first light pulse emitter that projects light out of the steering wheel grip at angle a1, and a second light pulse emitter that projects light out of the steering wheel grip at angle a2.
  • 5. The steering wheel of claim 1, wherein each of the at least one light pulse detector receives maximum intensity when the light reflected by the object enters said at least one second lens at angle b1, and also receives maximum intensity when light reflected by the object enters said at least one second lens at angle b2.
  • 6. The steering wheel of claim 1, wherein said light sensor comprises a first light detector that receives maximum intensity when the light reflected by the object enters said at least one second lens at angle b1, and a second light detector that receives maximum intensity when light reflected by the object enters said at least one second lens at angle b2.
  • 7. The steering wheel of claim 1, further comprising a processor, connected to said optoelectronic components, configured to identify a swipe gesture by the object in a first direction across the steering wheel grip in response to a specific optoelectronic component detecting maximum intensity of reflections of the object at one of angles b1 and b2, followed by the specific optoelectronic component detecting maximum intensity of reflections of the object at the other of angles b1 and b2.
  • 8. The steering wheel of claim 1, further comprising a processor, connected to said optoelectronic components, configured to identify a swipe gesture by the object in a first direction across the steering wheel grip in response to a first specific optoelectronic component detecting maximum intensity of reflections of the object at one of angles b1 and b2, followed by a second optoelectronic component, neighboring the first optoelectronic component, detecting maximum intensity of reflections of the object at the other of angles b1 and b2.
  • 9. The steering wheel of claim 8, wherein said processor is further configured to identify a swipe gesture by the object in a second direction across the steering wheel grip in response to the first specific optoelectronic component detecting maximum intensity of reflections of the object at one of angles b1 and b2, followed by the second optoelectronic component detecting maximum intensity of reflections of the object at the same one of angles b1 and b2.
  • 10. The steering wheel of claim 1, further comprising a processor, connected to said optoelectronic components, configured to identify an approach gesture by the object toward the steering wheel grip in response to a specific optoelectronic component detecting maximum intensity of reflections of light projected by the light projector of a first neighboring optoelectronic component at one of angles b1 and b2, followed by detecting maximum intensity of reflections of light projected by the light projector of a second neighboring optoelectronic component at the same one of angles b1 and b2, wherein the second neighboring optoelectronic component is nearer to the specific optoelectronic component than is the first neighboring optoelectronic component.
  • 11. The steering wheel of claim 1, wherein said PCB is oriented substantially perpendicular to the surface of the steering wheel grip facing the driver.
  • 12. The steering wheel of claim 1, wherein reflected light directed into said at least one second lens at angles b1 and b2 is respectively perpendicular to light directed out of said at least one first lens at angles a1 and a2.
  • 13. A method for detecting driver input, comprising: providing a series of optoelectronic components in a steering wheel grip, each specific optoelectronic component including a light emitter and a light detector;orienting a lens within each specific optoelectronic component such that the light emitter projects light beams through the lens in two emission directions out of the steering wheel grip, at angles a1 and a2 relative to the plane of the steering wheel grip, and the specific optoelectronic component light detector detects maximum intensity of reflections of the light projected by optoelectronic components that neighbor the specific optoelectronic component when the reflected light enters the lens at angles b1 and b2, the light being reflected by an object above the steering wheel grip; andmapping driver input to locations along the steering wheel grip that correspond to detections of the reflections.
  • 14. The method of claim 13, further comprising identifying an approach gesture by the object in response to detecting reflections of light projected by a distant neighboring optoelectronic component followed by detecting reflections of light projected by a near neighboring optoelectronic component.
  • 15. The method of claim 13, further comprising identifying a radial swipe gesture by the object in a first direction across the steering wheel grip in response to detecting reflections of the object entering the lens at one of angles b1 and b2, followed by detecting reflections of the object entering the lens at the other of angles b1 and b2.
  • 16. The method of claim 15, further comprising identifying a swipe gesture by the object in a second direction across the steering wheel grip in response to a first optoelectronic component detecting reflections of the object entering the lens at one of angles b1 and b2, followed by a second optoelectronic component, neighboring the first optoelectronic component, detecting reflections of the object entering the lens at the same one of angles b1 and b2.
PCT Information
Filing Document Filing Date Country Kind
PCT/US2019/062909 11/25/2019 WO
Publishing Document Publishing Date Country Kind
WO2020/112585 6/4/2020 WO A
US Referenced Citations (196)
Number Name Date Kind
4243879 Carroll et al. Jan 1981 A
4267443 Carroll et al. May 1981 A
4301447 Funk et al. Nov 1981 A
4518249 Murata et al. May 1985 A
4550250 Mueller et al. Oct 1985 A
4703316 Sherbeck Oct 1987 A
4710760 Kasday Dec 1987 A
4782328 Denlinger Nov 1988 A
4790028 Ramage Dec 1988 A
4847606 Beiswenger Jul 1989 A
4880969 Lawrie Nov 1989 A
4928094 Smith May 1990 A
5003505 McClelland Mar 1991 A
5016008 Gruaz May 1991 A
5036187 Yoshida et al. Jul 1991 A
5053758 Cornett et al. Oct 1991 A
5103085 Zimmerman Apr 1992 A
5119079 Hube et al. Jun 1992 A
5162783 Moreno Nov 1992 A
5179369 Person et al. Jan 1993 A
5194863 Barker et al. Mar 1993 A
5220409 Bures Jun 1993 A
5283558 Chan Feb 1994 A
5406307 Hirayama et al. Apr 1995 A
5414413 Tamaru et al. May 1995 A
5422494 West Jun 1995 A
5463725 Henckel et al. Oct 1995 A
5559727 Deley et al. Sep 1996 A
5577733 Downing Nov 1996 A
5579035 Beiswenger Nov 1996 A
5603053 Gough et al. Feb 1997 A
5612719 Beemink et al. Mar 1997 A
5618232 Martin Apr 1997 A
5729250 Bishop et al. Mar 1998 A
5748185 Stephan et al. May 1998 A
5785439 Bowen Jul 1998 A
5825352 Bisset et al. Oct 1998 A
5838308 Knapp et al. Nov 1998 A
5880743 Moran et al. Mar 1999 A
5886697 Naughton et al. Mar 1999 A
5889236 Gillespie et al. Mar 1999 A
5900875 Haitani et al. May 1999 A
5914709 Graham et al. Jun 1999 A
5936615 Waters Aug 1999 A
5943043 Furuhata et al. Aug 1999 A
5943044 Martinelli et al. Aug 1999 A
5956030 Conrad et al. Sep 1999 A
5988645 Downing Nov 1999 A
6010061 Howell Jan 2000 A
6023265 Lee Feb 2000 A
6031989 Cordell Feb 2000 A
6052279 Friend et al. Apr 2000 A
6073036 Heikkinen et al. Jun 2000 A
6085204 Chijiwa et al. Jul 2000 A
6091405 Lowe et al. Jul 2000 A
6114949 Schmitz et al. Sep 2000 A
6135494 Lotito et al. Oct 2000 A
6246395 Goyins et al. Jun 2001 B1
6259436 Moon et al. Jul 2001 B1
6292179 Lee Sep 2001 B1
6310609 Morgenthaler Oct 2001 B1
6323846 Westerman et al. Nov 2001 B1
6340979 Beaton et al. Jan 2002 B1
6346935 Nakajima et al. Feb 2002 B1
6356287 Ruberry et al. Mar 2002 B1
6359632 Eastty et al. Mar 2002 B1
6362468 Murakami et al. Mar 2002 B1
6411283 Murphy Jun 2002 B1
6421042 Omura et al. Jul 2002 B1
6429857 Masters et al. Aug 2002 B1
6456952 Nathan Sep 2002 B1
6529920 Arons et al. Mar 2003 B1
6542191 Yonezawa Apr 2003 B1
6549217 De Greef et al. Apr 2003 B1
6570557 Westerman et al. May 2003 B1
6597345 Hirshberg Jul 2003 B2
6628268 Harada et al. Sep 2003 B1
6639584 Li Oct 2003 B1
6646633 Nicolas Nov 2003 B1
6677932 Westerman Jan 2004 B1
6690365 Hinckley et al. Feb 2004 B2
6690387 Zimmerman et al. Feb 2004 B2
6703999 Iwanami et al. Mar 2004 B1
6707449 Hinckley et al. Mar 2004 B2
6727917 Chew et al. Apr 2004 B1
6734883 Wynn et al. May 2004 B1
6757002 Oross et al. Jun 2004 B1
6788292 Nako et al. Sep 2004 B1
6803906 Morrison et al. Oct 2004 B1
6833827 Lui et al. Dec 2004 B2
6836367 Seino et al. Dec 2004 B2
6857746 Dyner Feb 2005 B2
6864882 Newton Mar 2005 B2
6874683 Keronen et al. Apr 2005 B2
6888536 Westerman et al. May 2005 B2
6944557 Hama et al. Sep 2005 B2
6947032 Morrison et al. Sep 2005 B2
6954197 Morrison et al. Oct 2005 B2
6958749 Matsushita et al. Oct 2005 B1
6972401 Akitt et al. Dec 2005 B2
6988246 Kopitzke et al. Jan 2006 B2
6992660 Kawano et al. Jan 2006 B2
7006077 Uusimaki Feb 2006 B1
7007239 Hawkins et al. Feb 2006 B1
7030861 Westerman et al. Apr 2006 B1
7046232 Inagaki et al. May 2006 B2
7126583 Breed Oct 2006 B1
7133032 Cok Nov 2006 B2
7155683 Williams Dec 2006 B1
7159763 Yap et al. Jan 2007 B2
7176905 Baharav et al. Feb 2007 B2
7184030 McCharles et al. Feb 2007 B2
7221462 Cavallucci May 2007 B2
7225408 ORourke May 2007 B2
7232986 Worthington et al. Jun 2007 B2
7254775 Geaghan et al. Aug 2007 B2
7265748 Ryynanen Sep 2007 B2
7283845 De Bast Oct 2007 B2
7286063 Gauthey et al. Oct 2007 B2
7339580 Westerman et al. Mar 2008 B2
7352940 Charters et al. Apr 2008 B2
7355594 Barkan Apr 2008 B2
7369724 Deane May 2008 B2
7372456 McLintock May 2008 B2
7429706 Ho Sep 2008 B2
7435940 Eliasson et al. Oct 2008 B2
7441196 Gottfurcht et al. Oct 2008 B2
7441800 Weber et al. Oct 2008 B2
7442914 Eliasson et al. Oct 2008 B2
7464110 Pyhalammi et al. Dec 2008 B2
7465914 Eliasson et al. Dec 2008 B2
7469381 Ording Dec 2008 B2
7474772 Russo et al. Jan 2009 B2
7479949 Jobs et al. Jan 2009 B2
7518738 Cavallucci et al. Apr 2009 B2
7587072 Russo et al. Sep 2009 B2
7633300 Keroe et al. Dec 2009 B2
7663607 Hotelling et al. Feb 2010 B2
7705835 Eikman Apr 2010 B2
7742290 Kaya Jun 2010 B1
7782296 Kong et al. Aug 2010 B2
7812828 Westerman et al. Oct 2010 B2
7855716 McCreary et al. Dec 2010 B2
7880724 Nguyen et al. Feb 2011 B2
7880732 Goertz Feb 2011 B2
8022941 Smoot Sep 2011 B2
8026798 Makinen et al. Sep 2011 B2
8068101 Goertz Nov 2011 B2
8089299 Rahman et al. Jan 2012 B1
8095879 Goertz Jan 2012 B2
8120625 Hinckley Feb 2012 B2
8193498 Cavallucci et al. Jun 2012 B2
8289299 Newton Oct 2012 B2
8564424 Evarts et al. Oct 2013 B2
8775023 Frojdh et al. Jul 2014 B2
8918252 Frojdh et al. Dec 2014 B2
8933876 Galor et al. Jan 2015 B2
9092093 Jubner et al. Jul 2015 B2
9770986 Sannomiya et al. Sep 2017 B2
20010043189 Brisebois et al. Nov 2001 A1
20020152010 Colmenarez et al. Oct 2002 A1
20020158453 Levine Oct 2002 A1
20030086588 Shinada May 2003 A1
20040044293 Burton Mar 2004 A1
20040199309 Hayashi et al. Oct 2004 A1
20050021190 Worrell et al. Jan 2005 A1
20050052426 Hagermoser et al. Mar 2005 A1
20060047386 Kanevsky et al. Mar 2006 A1
20080211779 Pryor Sep 2008 A1
20090139778 Butler et al. Jun 2009 A1
20090166098 Sunder Jul 2009 A1
20090278915 Kramer et al. Nov 2009 A1
20090322673 Cherradi El Fadili Dec 2009 A1
20100185341 Wilson et al. Jul 2010 A1
20110030502 Lathrop Feb 2011 A1
20110032214 Gruhlke et al. Feb 2011 A1
20110050589 Yan et al. Mar 2011 A1
20110087963 Brisebois et al. Apr 2011 A1
20110241850 Bosch et al. Oct 2011 A1
20110310005 Chen et al. Dec 2011 A1
20120109455 Newman et al. May 2012 A1
20120179328 Goldman-Shenhar Jul 2012 A1
20120232751 Guspan Sep 2012 A1
20120283894 Naboulsi Nov 2012 A1
20120326735 Bennett et al. Dec 2012 A1
20130024071 Sivertsen Jan 2013 A1
20130063336 Sugimoto et al. Mar 2013 A1
20130204457 King et al. Aug 2013 A1
20140081521 Frojdh et al. Mar 2014 A1
20140292665 Lathrop et al. Oct 2014 A1
20150100204 Gondo Apr 2015 A1
20150123947 Jubner et al. May 2015 A1
20160154475 Eriksson Jun 2016 A1
20170262134 Eriksson et al. Sep 2017 A1
20180105185 Watanabe et al. Apr 2018 A1
20180296915 Holmgren et al. Oct 2018 A1
Foreign Referenced Citations (39)
Number Date Country
4423744 Apr 1995 DE
202014104143 Oct 2014 DE
0330767 Sep 1989 EP
0513694 Nov 1992 EP
0601651 Jun 1994 EP
0618528 Oct 1994 EP
0703525 Mar 1996 EP
1059603 Dec 2000 EP
2937263 Oct 2015 EP
1107666 Mar 1968 GB
2319997 Jun 1998 GB
2423808 Sep 2006 GB
03-216719 Sep 1991 JP
5-173699 Jul 1993 JP
6-39621 May 1994 JP
10-148640 Jun 1998 JP
10-269012 Oct 1998 JP
11-232024 Aug 1999 JP
2001-216069 Aug 2001 JP
3240941 Dec 2001 JP
2009-248629 Oct 2009 JP
2010-088487 Apr 2010 JP
2011-254957 Dec 2011 JP
2012-181639 Sep 2012 JP
2014-225145 Dec 2014 JP
8600446 Jan 1986 WO
8600447 Jan 1986 WO
9615464 May 1996 WO
0102949 Jan 2001 WO
0140922 Jun 2001 WO
02095668 Nov 2002 WO
03038592 May 2003 WO
03083767 Oct 2003 WO
2005026938 Mar 2005 WO
2008147266 Dec 2008 WO
2009008786 Jan 2009 WO
2010093570 Aug 2010 WO
2010121031 Oct 2010 WO
2011119483 Sep 2011 WO
Non-Patent Literature Citations (22)
Entry
Pfleging, B., Schneegass, S., Schmidt, A., Multimodal Interaction in the Car—Combining Speech and Gestures on the Steering Wheel, Proceedings of the 4th International Conference on Automotive User Interfaces and Interactive Vehicular Applications (AutomotiveUI '12), Oct. 17-19, 2012, Portsmouth, NH, USA.
Pfeiffer, M., Döring, T., Kern, D Krüger, A., Schöning, J., Schmidt, A., A Multi-Touch Enabled Steering Wheel—Exploring the Design Space, CHI 2010, Apr. 10-15, 2010, Atlanta, Georgia, USA.
Mahr, A., Endres, C., Schneeberger, T., Müller, C., Determining Human-Centered Parameters of Ergonomic Micro-Gesture Interaction for Drivers Using the Theater Approach, AutomotiveUI 2011, Nov. 30-Dec. 2, 2011, Salzburg, Austria.
Döring, T., Kern, D., Marshall, P., Pfeiffer, M., Schöning, J., Gruhn, V., Schmidt, A., Gestural Interaction on the Steering Wheel—Reducing the Visual Demand, CHI 2011, May 7-12, 2011, Vancouver, BC, Canada.
Navarro, J., Mars, F., Hoc, J.-M., Lateral Control Support for Car Drivers: a Human-Machine Cooperation Approach, Proceedings of the ECCE 2007 Conference, Aug. 28-31, 2007, London, UK.
Angelini, L., et al., Gesturing on the Steering Wheel: a User-elicited taxonomy, AutomotiveUI '14, Sep. 17-19, 2014, Seattle, WA, USA.
Werner, Steffen, The Steering Wheel as a Touch Interface: Using Thumb-Based Gesture Interfaces as Control Inputs While Driving, AutomotiveUI '14, Sep. 17-19, 2014, Seattle, WA, USA.
González, I. E., et al., Eyes on the Road, Hands on the Wheel: Thumb-based Interaction Techniques for Input on Steering Wheels, Graphics Interface Conference 2007, May 28-30, 2007, Montreal, Canada.
Murer, M., et al., Exploring the Back of the Steering Wheel: Text Input with Hands on the Wheel and Eyes on the Road, AutomotiveUI'12, Oct. 17-19, Portsmouth, NH, USA.
Koyama, S., et al., Multi-Touch Steering Wheelfor In-Car Tertiary Applications Using Infrared Sensors, AH '14, Mar. 7-9, 2014, Kobe, Japan.
Non-Final Office action received for U.S. Appl. No. 14/088,458 dated Feb. 7, 2014, 8 pages.
Notice of Allowance for U.S. Appl. No. 14/088,458 dated Mar. 6, 2014, 8 pages.
Search Report and Written Opinion for PCT application No. PCT/US13/71557 dated Apr. 25, 2014, 25 pages.
Examination Report No. 1 received for Australian patent application No. 2013352456 dated Dec. 23, 2014, 9 pages.
First Office Action received for Chinese patent application No. 201380021907.X dated Mar. 28, 2016, 12 pages.
Search Report for European patent application No. 13 859 391.8 dated Mar. 18, 2016, 8 pages.
Search Report for European patent application No. 17 184 782.5 dated Jul. 9, 2018, 10 pages.
Office Action from the Japanese Patent Office re: JP Patent Application: No. 2015-530174, dated Aug. 6, 2015, 7 pages.
First Office action for Korean patent application No. 10-2015-7001419 dated May 20, 2015, 3 pages.
Non-Final Office action received for U.S. Appl. No. 14/312,787 dated Jan. 8, 2015, 15 pages.
European Patent Application No. 19 890 109.2, European Search Report, dated Nov. 22, 2021,. 5 pages.
PCT/US2019/062909, International Search Report and Written Opinion, dated Mar. 24, 2020, 10 pp.
Related Publications (1)
Number Date Country
20220035477 A1 Feb 2022 US
Provisional Applications (2)
Number Date Country
62778390 Dec 2018 US
62772574 Nov 2018 US