Integrating a sensing system into headlight optics

Information

  • Patent Grant
  • 12228653
  • Patent Number
    12,228,653
  • Date Filed
    Friday, October 7, 2022
    2 years ago
  • Date Issued
    Tuesday, February 18, 2025
    2 months ago
  • Inventors
  • Original Assignees
    • Magna Electronics, LLC (Southfield, MI, US)
  • Examiners
    • Williams; Joseph L
    Agents
    • Thayne; Matthew D.
    • Thayne and Davis LLC
Abstract
An example system includes an infrared emitter to output infrared light towards a target, where the infrared light reflects from the target to produce reflected infrared light, and a detector to receive the reflected infrared light and to provide a signal based on the reflected infrared light. The system also includes a lighting system that includes a light emitter to output visible light, a mirror configured (i) to allow the visible light to pass through the mirror and to reflect the reflected infrared light onto the detector, or (ii) to allow the reflected infrared light to pass through the mirror and onto the detector and to reflect the visible light, and one or more optical elements configured to affect the visible light and the reflected infrared light.
Description
TECHNICAL FIELD

This specification describes examples of techniques for integrating a sensing system into headlight optics.


BACKGROUND

A vehicle, such as an automobile, includes an illumination system. A headlight, for example, is located on the front of the vehicle and illuminates a path in front the vehicle, at least partly. The vehicle may also include a sensor on a front of the vehicle to sense objects in the vehicle's path or otherwise in front of the vehicle.


SUMMARY

An example system includes an infrared emitter to output infrared light towards a target, where the infrared light reflects from the target to produce reflected infrared light, and a detector to receive the reflected infrared light and to provide a signal based on the reflected infrared light. The system also includes a lighting system that includes a light emitter to output visible light, a mirror configured (i) to allow the visible light to pass through the mirror and to reflect the reflected infrared light onto the detector, or (ii) to allow the reflected infrared light to pass through the mirror and onto the detector and to reflect the visible light, and one or more optical elements configured to affect the visible light and the reflected infrared light. The mirror is between the light emitter and the one or more optical elements or the mirror is between the detector and the one or more optical elements. The example system may include one or more of the following features, either alone or in combination.


The lighting system may include a headlight of a vehicle. The detector may include a silicon photomultiplier (SiPM) sensor that is responsive to infrared wavelengths of light. The infrared emitter may include a near infrared (NIR) laser emitter. The infrared light may include laser light. The mirror may include a dielectric mirror having a dichroic filter to allow the visible light to pass and to reflect the infrared light. The infrared emitter and the detector may be components of a light detection and ranging (LIDAR) system for a vehicle. The LIDAR system may be configured for use in at least one of: automatic emergency braking for the vehicle, forward sensing for the vehicle, or automated driving for the vehicle.


Along a first direction of an optical path containing the one or more optical elements, the one or more optical elements may direct the visible light away from the light emitter. Along a second direction of the optical path, the one or more optical elements may direct the reflected infrared light towards the mirror. Along the first direction of the optical path, the one or more optical elements may direct the visible light. Along the second direction of the optical path, the one or more optical elements may at least partially focus the reflected infrared light. In at least one dimension, at least one of the optical elements may have a dimension that is at least ten times greater in size than a dimension of optics in the infrared emitter.


The detector may include cells that are addressable for operation. The cells may be addressable in sequence by the reflected infrared light to produce a scan of a field of view in front of the infrared emitter. The one or more optical elements may include at least one convex lens. The one or more optical elements may include at least one concave-convex lens. The one or more optical elements may include at least one telecentric stop. The light emitter may include a light-emitting diode matrix array configured to project the visible light toward the mirror.


An example method of scanning a target includes the following operations: outputting infrared light toward the target; receiving reflected infrared light from the target via first optics that are also configured to output visible light toward the target, where the first optics are physically separate from second optics used to output the infrared light toward the target; directing, via the first optics, the reflected infrared light toward a dielectric mirror comprising a dichroic filter configured (i) to reflect infrared light and to allow the visible light to pass through the dielectric mirror, or (ii) to reflect the visible light and to allow the infrared light to pass through the dielectric mirror; and receiving, at a detector, the reflected infrared light via (e.g., by way of or from) the dielectric mirror, where the detector provides one or more signals based on the reflected infrared light. The example method may include one or more of the following features, either alone or in combination.


Along a first direction of an optical path containing the optics, the optics may direct the visible light away from a light emitter. Along a second direction of the optical path, the optics may direct the reflected infrared light towards the dielectric mirror. Along the first direction of the optical path, the optics may direct the visible light. Along the second direction of the optical path, the optics may direct the reflected infrared light. In at least one dimension, the first optics are at least ten times greater in size than the second optics. The method may performed by a light detection and ranging (LIDAR) system for a vehicle. The LIDAR system may be used in at least one of: automatic emergency braking for the vehicle, forward sensing for the vehicle, or automated driving for the vehicle.


Any two or more of the features described in this specification, including in this summary section, may be combined to form implementations not specifically described in this specification.


The systems, techniques, components, structures, and variations thereof described herein, or portions thereof, can be implemented using, or controlled by, a computer program product that includes instructions that are stored on one or more non-transitory machine-readable storage media, and that are executable on one or more processing devices to execute at least some of the operations described herein. The systems, techniques, components, structures, and variations thereof described herein, or portions thereof, can be implemented as an apparatus, method, or electronic system that can include one or more processing devices and computer memory to store executable instructions to implement various operations. The systems, techniques, components, structures, and variations thereof described herein may be configured, for example, through design, construction, size, shape, arrangement, placement, programming, operation, activation, deactivation, and/or control.


The details of one or more implementations are set forth in the accompanying drawings and the following description. Other features and advantages will be apparent from the description and drawings, and from the claims.





DESCRIPTION OF THE DRAWINGS


FIG. 1 is a diagram showing part of a front end of an automobile.



FIG. 2 is a top view of the automobile of FIG. 1.



FIG. 3 is a side view of a sensing system integrated into example headlight optics and a block diagram of a control system therefor.



FIG. 4 is a perspective view of the system of FIG. 3.



FIG. 5 is a side view of some components of the system of FIG. 3 showing an illumination path.



FIG. 6 is a side view of an alternative example of a sensing system integrated into example headlight optics.



FIG. 7 is a perspective view of the system of FIG. 6.



FIG. 8 is a side view of some components of the system of FIG. 6 showing an illumination path.



FIG. 9 is a side view of an alternative example of a sensing system integrated into example headlight optics.



FIG. 10 is a flowchart showing a process for using a sensing system that incorporates headlight optics.





Like reference numerals in different figures indicate like elements.


DETAILED DESCRIPTION

Described herein are example implementations of systems that use separate optical elements (“optics”) for transmitting and receiving sensor signals. A type of sensor that may be used includes, for example, a light detection and ranging (LIDAR) system to detect one or more objects exterior to the vehicle. LIDAR is a technique for determining ranges (e.g., variable distance) by targeting an object with a laser and measuring the time for the reflected light to return to the receiver.


In the example systems described herein, the LIDAR system is integrated into headlight optics, which are different and separate from the LIDAR emitter optics. More specifically, an infrared emitter outputs infrared light, such as an infrared laser beam, towards a target, such as an object in the vehicle's path. The incident infrared light is reflected from the object back to optics in the vehicle's headlight. The optics in the vehicle's headlight direct the reflected infrared light to a detector. The detector receives the reflected infrared light and is responsive to the reflected infrared light to determine the presence, location, size, and/or other features of the object.


Using optics in the vehicle's headlight to receive and to direct the reflected infrared light may be advantageous. For example, optics needed to receive reflected infrared light accurately are usually of a greater size than that required to emit or to transmit the infrared light. By using existing optics in the headlight, which is larger than the optics in the infrared emitter, the amount of optics on the vehicle can be reduced. That is, there is no need for large, separate sensor optics to receive the reflected infrared light, since pre-existing, large optics in the headlight performs that function. Moreover, the large optics in the vehicle's headlight may improve the quality of the signal that reaches the detector, thereby providing more accurate detection results.


Another potential advantage of the example systems described herein is that existing headlight cleaning mechanisms can be used to clean the larger headlight optics used in the LIDAR system. Because there is no large, separate sensor optics to receive the reflected infrared light, there is likewise no need for separate cleaning mechanisms or additional cleaning solution for such sensor optics.



FIG. 1 shows components of an example system 10 of the preceding type that integrates sensor optics into an existing headlight system. As shown in FIG. 1, system 10 is part of an automobile 12 and, in particular, the front-end 13 of the automobile.


The system's components include an infrared emitter 14 to output infrared light towards a target. The infrared light may be a laser beam having wavelengths in the range of about 700 nanometers (nm) to 1 millimeter (mm). In some implementations, the infrared light is near infrared light (NIR), although the system is not limited as such. NIR light is generally considered to be in the range of 700 nm to 2500 nm. In a particular example, the infrared light has a wavelength of about 940 nm. In some implementations, the infrared light is shortwave infrared light (SWIR). SWIR light is generally considered to be in a range of wavelengths from 0.9 to 1.7 microns.


Infrared emitter 14 may include an array of VCSELS (vertical-cavity surface-emitting laser) (for example, 1×8) that are individually addressable and may be fired in sequence to scan a field-of-view (FOV) in the front of the automobile. As shown in FIG. 2, “the front” of automobile 10 is not limited to directly in front, but rather includes any area 16 that is in front of a line 17 parallel to the automobile's front end. Likewise, “the back” of the automobile is not limited to directly in back, but rather includes any area 18 that is in back of a line 19 parallel to the vehicle's back end.


Infrared emitter 14 may be a relatively small device having a dimension of less than 0.5 mm or a larger device having a dimension of greater than 0.5 mm. The dimension may be, for example, a side of the device or a diameter of the device. During operation of the emitter, the infrared light may be moved side-to-side or up-and-down—in other words, scanned—in a plane by addressing the VCSELS in the array in linear sequence. As shown in FIG. 2, infrared light 23 from infrared emitter 14 may be scanned in a region, such as a plane 20, in front of automobile 10. The scanned infrared light may be pulsed, FMCW (frequency-modulated continuous wave) amplitude modulated, or pseudo-random pulse coded. In some implementations, an emitter may be used that emits laser light other than infrared light for LIDAR scanning.


The infrared light incident on an object 21 (in this example, a person) is reflected off of the object. The resulting reflected infrared light 22 propagates through the air back to automobile 10. All or at least some of that reflected infrared light is incident on headlight 25 (FIGS. 1 and 2). The optics—e.g., one or more optical elements, such as lenses and/or mirrors—in the headlight are configured to affect the direction and/or shape of the reflected infrared light such that the reflected infrared light is incident on a detector. The detector, which is described below, is part of the LIDAR system and is configured to detect the reflected infrared light and to generate an output, such as a signal, based on the reflected infrared light. The signal may be read by, or output to, an onboard control system, such as an onboard computing system, that interprets the signal along with other reflected infrared signals to determine the presence, location, size, or other features of the object that reflected the infrared light.



FIGS. 3 and 4 show components of a lighting system 28 included as part of a headlight 25 in automobile 10. Also, FIG. 1 shows where the components of lighting system 28 are located and FIG. 5 shows a subset of those components and the illumination path. The components may be held in, and mounted to, an enclosure 30 that contains the headlight and its various parts, as shown in FIG. 3. The components include a light emitter 29, such as a flat light emitting diode (LED) array, comprised of a matrix of collocated individual LEDs that are configured to emit white (visible) light. Use of an LED array may be advantageous because it eliminates or reduces the need for light guides in the headlight system.


The light emitted from the LED array may be visible light having wavelengths within in a range of about 400 nanometers (nm) to about 700 nm. Light having any wavelength in this range may be emitted. The light from the LED array is the light that is output from the headlight to illuminate a field in front of the automobile. The optics—e.g., lenses and/or mirrors—in the headlight are configured to affect the direction and/or shape of the visible light output to the environment.


The headlight may be controllable by the operator and/or a control system 32, which is described below with respect to FIG. 3, to operate in different modes such as low-beam operation and high-beam operations. During low-beam operation, the visible light from the headlight is typically on the order of 700 lumens. During high-beam operation, the visible light from the headlight is typically on the order of 1200 lumens. Regulations may specify headlight intensity in terms of candela (cd), where the range may be from 1200 cd for daytime running lights to 75,000 cd for high-beam operation. In any case, different vehicles may produce different light intensities for different modes of operation. In some implementations, during normal headlight operation, the direction of the output light is controllable; that is, the output light can be output in specified directions. For example, the LED array and/or optics may be controlled to move or to tilt in order direct the light output at specific angles based on signals received from the control system. In some implementations, during normal headlight operation, the direction of the output light is not controllable.


Components of system 28 also include a dichroic mirror (or simply “mirror”) 35. Mirror 35 is dichroic in the sense that it allows light of certain wavelengths to pass through the mirror unimpeded and reflects light of other wavelengths. In this example, mirror 35 is coated with a thin film that allows visible light to pass unimpeded while reflecting infrared light. A mirror of this type is known as a “hot mirror”.


In the example configuration of FIGS. 3 and 4, mirror 35 is angled at about 45° relative to detector 36 and LED array 29. In other implementations the angle may be different; for example, the mirror may be angled at 250, 350, 650, 750 and so forth, depending on the locations and angular position of detector 36 and LED array 29. Light from LED array 29 passes in the direction of arrow 37 and proceeds through mirror 35 unimpeded on to optics 38, which is described below, and, thereafter, out of the headlight. This light 41 is shown in FIG. 5. By contrast, as shown in FIGS. 3 and 4, reflected infrared light 22, which entered the headlight from the environment and passed through optics 39, reflects off mirror, 31 in this example at a 45° angle, and onto a surface of detector 36. The reflected infrared light used in the LIDAR system may enter headlight 25 when LED array 29 is activated (for example, while the headlight is in low-beam or high-beam mode) or when LED array 29 is not activated (for example, while the headlight is off). Mirror reflects the reflected infrared light in the same way regardless of the illumination of LED array 29. Accordingly, the LIDAR system described herein works the same way regardless of the illumination of LED array 29.


In some implementations, detector 36 is or includes a flat silicon photomultiplier (SiPM) detector having cells that are responsive to infrared wavelengths of light, such as 940 nm or whatever wavelength is output by the infrared emitter. The cells are addressable in sequence by reflected infrared light resulting from infrared light scanned across a FOV in front of the automobile. In some implementations, the cells of the detector react to incident infrared light 22, for example, by illuminating and/or generating signals based on the infrared light. In implementations that use laser light other than infrared light, detector 36 may be responsive to laser light other than infrared light.


In some implementations, control system 32 may synchronously sample the detector output using rolling shutter readout in order to identify where the reflected infrared impacted the detector. A relatively low duty factor may be used to maximize SNR (signal-to-noise) ratio. The control system uses the information obtained from the detector to determine the presence, location, size, and/or other features of the object from which the reflected infrared light was received. Multiple reflections—for example, from scanning of the incident infrared light—may be required for the controller to make a determination about the presence, location, size, and/or other features of the object.


Signals obtained via the LIDAR system may be used by the control system to control and/or to inform various automobile operations including, but not limited to, automatic emergency braking for the automobile, forward sensing for the automobile, or automated/self-driving for the automobile. For example, if an object is detected in the automobile's path of travel, the vehicles brakes may be activated, or the steering of the automobile may be controlled to avoid the object, as described below.


In this example, optics 38 also include two lenses 38a, 38b; however, more than two or fewer than two lenses may be used. For example, one lens may be used or three, four, five or more lenses may be used. In this example, the two lenses are aspherical and include a convex lens 38b and a concave-convex lens 38a. As shown in FIG. 5, in this optics configuration, white (visible) light 41 from LED array 29 passes through mirror 35 (not shown in FIG. 5) unimpeded and, thereafter, through lenses 38a, 38b in the manner shown to illuminate the environment 44 in front of the automobile. By contrast, referring to FIGS. 3 and 4, reflected infrared light 22 enters headlight 25 and optics 38 as shown. Lens 38a directs reflected infrared light 22 onto lens 38b, and lens 38b directs reflected infrared light 22 onto mirror 35. In an example, the lens system in its entirety focuses such that the incoming infrared light is focused when it reaches detector 36. Mirror 35 reflects reflected infrared light 22 onto detector 36, which reacts to the reflected infrared light as described above.


Stated otherwise, in a first direction 48 (FIG. 3) of the optical path containing lenses 38a, 38b, lenses 38a, 38b direct visible light 41 away from the LED array 29 and at least partially disperse the visible light into the environment 44 external to the automobile. In a second direction of that same optical path 49, lenses 38a, 38b direct the reflected infrared light 22 onto mirror 35, which reflects that light onto detector 36.


In some implementations, an example lens 38a has a 13 mm aperture entrance, which is more than ten times the dimension of an example infrared emitter having a side or a diameter of 0.5 mm. This dichotomy between the size of the infrared emitter 14 and the size of the headlight's lens 38a enables the LIDAR operation and advantages described above. The dimensions described herein, however, are only examples. In some implementations, the entrance to the headlight optics may be twice the dimension of the infrared emitter, five times the dimension of the infrared emitter, fifteen times the dimension of an infrared emitter, twenty times the dimension of the infrared emitter, and so forth, or any appropriate value in between these.


In some implementations, example optics 50 of FIGS. 6 to 8 may be substituted for the optics of FIG. 38 in the systems described herein. Optics 50 includes an aspherical lens 50a having one side that is flat and one side that is concave. Optics 50 also includes a concave-convex lens 50b and a telecentric stop having aperture of 14 mm in this example (although larger or smaller apertures may be used). Telecentric stop 50c (not shown in FIG. 8) partially collimates light output from, or input to, the remainder of optics 50. FIG. 8 shows how optics 50 directs white (visible) light 53 output from LED array 29 to illuminate in front 82 of the automobile. FIGS. 6 and 7 show how optics 50 direct, and at least partially focus, reflected infrared light 22 onto detector 36. In this implementation, detector 36, LED array 31, and mirror 35 have the same structures and functions as the implementation described with respect to FIGS. 3 to 5.



FIG. 9 shows an alternative system 60 of the type described herein. In system 60, mirror 61 is coated with a thin film that allows infrared light 63 (e.g., the same as reflected infrared light 22) to pass through mirror 61 unimpeded and that reflects visible light 64. This type of dichroic mirror is known as a “cold mirror”. Accordingly, in system 60, detector 36 may be positioned as shown to receive reflected infrared light 63 passing through mirror 61 and LED array 29 may be positioned as shown to provide the visible light 64 for the headlight that is reflected from mirror 61 and output to the environment. In the example configuration of FIG. 9, mirror 61 is angled at about 45° relative to detector 36 and LED array 29. In other implementations the angle may be different; for example, the mirror may be angled at 25°, 35°, 65°, 750 and so forth, depending on the locations and angular position of detector 36 and LED array 29, In system 60, optics 67 include a telecentric lens and can be implemented using two lens elements, with at least one having an aspherical surface. Any appropriate optical configuration may be used to receive and to direct both the visible light and the infrared light to the appropriate components, including those described with respect to FIGS. 3 to 8. A potential additional advantage of system 60 is that it may be compatible with a flat detector 36 without requiring lightguides.


The implementations of the LIDAR system described herein may have a FOV of approximately 50°×10° and may reliably detect 10% reflective objects at a 40 meter (me) distance in full sunlight. These numbers, however are examples only and are not limiting. For example, the range of the system can be increased by scaling the optics.


The example systems described herein, such as systems 10, 28, and 60, each may be controlled a control system, such as control system 32 of FIG. 3, to control all or part of the operation of the system components. The control system may be part of an onboard control system on the automobile. As shown in FIG. 3, in some implementations, an onboard portion 70 of control system 30 includes one or more processing devices 71 of the type described herein that are programmable to control operations of at least some of the components of the system. The onboard portion 70 of control system 30 may also include memory 74 for storing data and programs executable by the one or more processing devices 71 to implement all or part of the functionality described herein. The control system 32 may also include an external computing system 72 that communicates to the onboard control portion 70. For example, the external computing system 72 may communicate with the onboard control portion 70 using a cellular network or other appropriate wireless functionality.



FIG. 10 shows operations that may be part of an example process 80 that may be performed using the example systems described herein, such as systems 10, 28, and 60. Process 10 includes controlling an infrared emitter to output (80a) infrared light to scan a region external to a vehicle, such as automobile 12. The infrared light reaches an object in the infrared emitter's scanning field (e.g., 20 of FIG. 2) and reflects from the object. All or part of the reflected infrared light is received (80b) at a vehicle's headlight, such as headlight 25. Optics included in systems 10, 28, and 60 direct the infrared light to a detector. The detector is responsive to the reflected infrared light to detect the reflected infrared light (80c). The detector is read (80d) by the control system as described herein and/or outputs one or more signals to the control system based on the detected infrared light. The control system interprets (80e) the readings/signals obtained from om the detector and reacts (80f) accordingly. For example, the control system may output a warning on a display screen on the interior of the cabin, output an audible warning, or take control of the vehicle whether or not an operator is otherwise controlling the vehicle. For example, the control system may control the automobile's brakes to stop or to slow-down before reaching the object or the control system may control the automobile's steering to avoid the object preemptively. Other controls not described herein also may be implemented.


Although the preceding descriptions focus on using LIDAR on a vehicle's front-end, LIDAR may be incorporated on the back-end of a vehicle to scan region 18 (FIG. 2) using to the techniques described herein, where the taillight optics are used as receiving optics for reflected infrared (or other wavelength) signals. Furthermore, the systems and techniques are not limited to use with automobiles, but rather may be used with any type of vehicle, whether operator-drive or automated. Still further, a LIDAR system of the type described herein may be implemented using more than one headlight or taillight. For example, each vehicle headlight or taillight may include LIDAR functionality of the type described herein. In some cases, a single infrared emitter may be serviced by headlight LIDAR functionality in two front-end headlights of the type described herein or each headlight LIDAR functionality may operate with a separate, dedicated infrared emitter. The same is true for taillight LIDAR configurations. Still further, the systems and techniques are not limited to use with vehicle lights, but rather may integrate any type of light source into a LIDAN system for use as receiving optics as described herein.


Furthermore, although the preceding descriptions focus on using LIDAR, other types of detectors may be used. For example, a line scan sensor may be used in place of LIDAR detector. In an example, a line scan sensor may be a camera that uses a single row of light-sensitive pixels that image across the object, line-by-line, using high intensity lighting. A completed image is constructed by combining the lines.


In an example system of this type, headlight optics, which are different and separate from line scan detector optics, are integrated into a line scan detection system. Similar to above, an emitter outputs light, such as scanning laser beam, towards a target, such as an object in the vehicle's path. The incident light is reflected from the object back to optics in the vehicle's headlight. The optics in the vehicle's headlight direct the reflected light to a line scan detector. The line detector receives the reflected infrared light and is responsive to the reflected light to generate signals used by the control system to determine the presence, location, size, and/or other features of the object using line scan technology.


All or part of the systems and processes described in this specification and their various modifications may be configured or controlled at least in part by one or more computing systems, such as control system 32, using one or more computer programs tangibly embodied in one or more information carriers, such as in one or more non-transitory machine-readable storage media. A computer program can be written in any form of programming language, including compiled or interpreted languages, and it can be deployed in any form, including as a stand-alone program or as a module, part, subroutine, or other unit suitable for use in a computing environment. A computer program can be deployed to be executed on one computer or on multiple computers at one site or distributed across multiple sites and interconnected by a network.


Actions associated with configuring or controlling the systems and processes described herein can be performed by one or more programmable processors executing one or more computer programs to control or to perform all or some of the operations described herein. All or part of the systems and processes can be configured or controlled by special purpose logic circuitry, such as, an FPGA (field programmable gate array) and/or an ASIC (application-specific integrated circuit) or embedded microprocessor(s) localized to the instrument hardware.


Processors suitable for the execution of a computer program include, by way of example, both general and special purpose microprocessors, and any one or more processors of any kind of digital computer. Generally, a processor will receive instructions and data from a read-only storage area or a random access storage area or both. Elements of a computer include one or more processors for executing instructions and one or more storage area devices for storing instructions and data. Generally, a computer will also include, or be operatively coupled to receive data from, or transfer data to, or both, one or more machine-readable storage media, such as mass storage devices for storing data, such as magnetic, magneto-optical disks, or optical disks. Non-transitory machine-readable storage media suitable for embodying computer program instructions and data include all forms of non-volatile storage area, including by way of example, semiconductor storage area devices, such as EPROM (erasable programmable read-only memory), EEPROM (electrically erasable programmable read-only memory), and flash storage area devices; magnetic disks, such as internal hard disks or removable disks; magneto-optical disks; and CD-ROM (compact disc read-only memory) and DVD-ROM (digital versatile disc read-only memory).


Elements of different implementations described may be combined to form other implementations not specifically set forth previously. Elements may be left out of the systems described previously without adversely affecting their operation or the operation of the system in general. Furthermore, various separate elements may be combined into one or more individual elements to perform the functions described in this specification.


Other implementations not specifically described in this specification are also within the scope of the following claims.

Claims
  • 1. A system comprising: an infrared emitter to output infrared light towards a target, the infrared light reflecting from the target to produce reflected infrared light;a detector to receive the reflected infrared light and to provide a signal based on the reflected infrared light; anda lighting system comprising: a light emitter to output visible light;a mirror configured (i) to allow the visible light to pass through the mirror and to reflect the reflected infrared light onto the detector, or (ii) to allow the reflected infrared light to pass through the mirror and onto the detector and to reflect the visible light; andone or more optical elements configured to affect the visible light and the reflected infrared light, the mirror being between the light emitter and the one or more optical elements or the mirror being between the detector and the one or more optical elements.
  • 2. The system of claim 1, wherein the lighting system comprises a headlight of a vehicle.
  • 3. The system of claim 1, wherein the detector comprises a silicon photomultiplier (SiPM) sensor that is responsive to infrared wavelengths of light.
  • 4. The system of claim 1, wherein the infrared emitter comprises a near infrared (NIR) laser emitter and the infrared light comprises laser light.
  • 5. The system of claim 1, wherein the mirror comprises a dielectric mirror comprising a dichroic filter to allow the visible light to pass and to reflect the infrared light.
  • 6. The system of claim 1, wherein the infrared emitter and the detector comprise components of a light detection and ranging (LIDAR) system for a vehicle.
  • 7. The system of claim 6, wherein the system is part of a vehicle; and wherein the LIDAR system is configured for use in at least one of: automatic emergency braking for the vehicle, forward sensing for the vehicle, or automated driving for the vehicle.
  • 8. The system of claim 1, wherein, along a first direction of an optical path containing the one or more optical elements, the one or more optical elements direct the visible light away from the light emitter; and wherein, along a second direction of the optical path, the one or more optical elements direct the reflected infrared light towards the mirror.
  • 9. The system of claim 8, wherein, along the first direction of the optical path, the one or more optical elements direct the visible light; and wherein, along the second direction of the optical path, the one or more optical elements at least partially focus the reflected infrared light.
  • 10. The system of claim 1, wherein, in at least one dimension, at least one of the optical elements has a dimension that is at least ten times greater in size than a dimension of optics in the infrared emitter.
  • 11. The system of claim 1, wherein the detector comprises cells that are addressable for operation; and wherein the cells are addressable in sequence by the reflected infrared light to produce a scan of a field of view in front of the infrared emitter.
  • 12. The system of claim 1, wherein the one or more optical elements comprise at least one convex lens.
  • 13. The system of claim 1, wherein the one or more optical elements comprise at least one concave-convex lens.
  • 14. The system of claim 1, wherein the one or more optical elements comprise at least one telecentric stop.
  • 15. The system of claim 1, wherein the light emitter comprises a light-emitting diode matrix array configured to project the visible light toward the mirror.
  • 16. A method of scanning a target, the method comprising: outputting infrared light toward the target;receiving reflected infrared light from the target via first optics that are also configured to output visible light toward the target, the first optics being physically separate from second optics used to output the infrared light toward the target;directing, via the first optics, the reflected infrared light toward a dielectric mirror comprising a dichroic filter configured (i) to reflect infrared light and to allow the visible light to pass through the dielectric mirror, or (ii) to reflect the visible light and to allow the infrared light to pass through the dielectric mirror; andreceiving, at a detector, the reflected infrared light via the dielectric mirror, the detector providing one or more signals based on the reflected infrared light.
  • 17. The method of claim 16, wherein, along a first direction of an optical path containing the optics, the optics directs the visible light away from a light emitter; and wherein, along a second direction of the optical path, the optics directs the reflected infrared light towards the dielectric mirror.
  • 18. The method of claim 17, wherein, along the first direction of the optical path, the optics direct the visible light; and wherein, along the second direction of the optical path, the optics at least partially focus the reflected infrared light.
  • 19. The method of claim 16, wherein, in at least one dimension, the first optics are at least ten times greater in size than the second optics.
  • 20. The method of claim 16, which is performed by a light detection and ranging (LIDAR) system for a vehicle; wherein the LIDAR system is used in at least one of: automatic emergency braking for the vehicle, forward sensing for the vehicle, or automated driving for the vehicle.
US Referenced Citations (215)
Number Name Date Kind
1520245 Jules Dec 1924 A
3712985 Swarner et al. Jan 1973 A
3898656 Jensen Aug 1975 A
4125864 Aughton Nov 1978 A
4184154 Albanese et al. Jan 1980 A
4362361 Campbell et al. Dec 1982 A
4439766 Kobayashi et al. Mar 1984 A
4765715 Matsudaira et al. Aug 1988 A
4957362 Peterson Sep 1990 A
5200606 Krasutsky et al. Apr 1993 A
5210586 Grage et al. May 1993 A
5274379 Carbonneau et al. Dec 1993 A
5428215 Dubois et al. Jun 1995 A
5604695 Cantin et al. Feb 1997 A
5793491 Wangler et al. Aug 1998 A
5889490 Wachter et al. Mar 1999 A
5966226 Gerber Oct 1999 A
6078395 Jourdain et al. Jun 2000 A
6122222 Hossack et al. Sep 2000 A
6292285 Wang et al. Sep 2001 B1
6384770 De Gouy et al. May 2002 B1
6437854 Hahlweg Aug 2002 B2
6556282 Jamieson et al. Apr 2003 B2
6559932 Halmos May 2003 B1
7202941 Munro Apr 2007 B2
7227116 Gleckler Jun 2007 B2
7272271 Kaplan et al. Sep 2007 B2
7440084 Kane Oct 2008 B2
7483600 Achiam et al. Jan 2009 B2
7489865 Varshneya et al. Feb 2009 B2
7544945 Tan et al. Jun 2009 B2
7570347 Ruff et al. Aug 2009 B2
7675610 Redman et al. Mar 2010 B2
7832762 Breed Nov 2010 B2
8044999 Mullen et al. Oct 2011 B2
8050863 Trepagnier et al. Nov 2011 B2
8134637 Rossbach et al. Mar 2012 B2
8223215 Oggier et al. Jul 2012 B2
8363511 Frank et al. Jan 2013 B2
8508723 Chang et al. Aug 2013 B2
8629975 Dierking et al. Jan 2014 B1
8742325 Droz et al. Jun 2014 B1
8836761 Wang et al. Sep 2014 B2
8836922 Pennecot et al. Sep 2014 B1
8879050 Ko Nov 2014 B2
9007569 Amzajerdian et al. Apr 2015 B2
9063549 Pennecot et al. Jun 2015 B1
9086273 Gruver et al. Jul 2015 B1
9090213 Lawlor et al. Jul 2015 B2
9097646 Campbell et al. Aug 2015 B1
9140792 Zeng et al. Sep 2015 B2
9157790 Shpunt et al. Oct 2015 B2
9267787 Shpunt et al. Feb 2016 B2
9285477 Smith et al. Mar 2016 B1
9482412 Schwaiger et al. Nov 2016 B2
9575162 Owechko Feb 2017 B2
9618742 Droz et al. Apr 2017 B1
9651417 Shpunt et al. May 2017 B2
9658322 Lewis May 2017 B2
9696427 Wilson et al. Jul 2017 B2
9711493 Lin Jul 2017 B2
9753351 Eldada Sep 2017 B2
9823351 Haslim et al. Nov 2017 B2
9857472 Mheen et al. Jan 2018 B2
9869754 Campbell et al. Jan 2018 B1
10018725 Liu Jul 2018 B2
10018726 Hall et al. Jul 2018 B2
10024655 Raguin et al. Jul 2018 B2
10078133 Dussan Sep 2018 B2
10088557 Yeun Oct 2018 B2
10148060 Hong et al. Dec 2018 B2
10175360 Zweigle et al. Jan 2019 B2
10183541 Van Den Bossche et al. Jan 2019 B2
10369922 Nakashima et al. Aug 2019 B2
10408924 Mheen et al. Sep 2019 B2
10411524 Widmer et al. Sep 2019 B2
10416292 De Mersseman et al. Sep 2019 B2
10473767 Xiang et al. Nov 2019 B2
10473784 Puglia Nov 2019 B2
10473943 Hughes Nov 2019 B1
10551501 Lachapelle Feb 2020 B1
10557923 Watnik et al. Feb 2020 B2
10558044 Pan Feb 2020 B2
10564268 Turbide et al. Feb 2020 B2
10578724 Droz et al. Mar 2020 B2
10627493 Morikawa et al. Apr 2020 B2
10678117 Shin et al. Jun 2020 B2
10768346 Miner et al. Sep 2020 B2
10775508 Rezk et al. Sep 2020 B1
10937773 T'ng et al. Mar 2021 B2
11326758 de Mersseman May 2022 B1
20010052872 Hahlweg Dec 2001 A1
20030043363 Jamieson et al. Mar 2003 A1
20040028418 Kaplan et al. Feb 2004 A1
20040031906 Gleckler Feb 2004 A1
20040135992 Munro Jul 2004 A1
20040155249 Narui et al. Aug 2004 A1
20050219506 Okuda et al. Oct 2005 A1
20060072189 Dimarzio et al. Apr 2006 A1
20060221250 Rossbach et al. Oct 2006 A1
20060232052 Breed Oct 2006 A1
20060239312 Kewitsch et al. Oct 2006 A1
20070140613 Achiam et al. Jun 2007 A1
20070181810 Tan et al. Aug 2007 A1
20070211786 Shattil Sep 2007 A1
20070219720 Trepagnier et al. Sep 2007 A1
20080088499 Bonthron et al. Apr 2008 A1
20080095121 Shattil Apr 2008 A1
20080100510 Bonthron et al. May 2008 A1
20080219584 Mullen et al. Sep 2008 A1
20080246944 Redman et al. Oct 2008 A1
20090002680 Ruff et al. Jan 2009 A1
20090010644 Varshneya et al. Jan 2009 A1
20090190007 Oggier et al. Jul 2009 A1
20090251361 Beasley Oct 2009 A1
20100027602 Abshire et al. Feb 2010 A1
20100157280 Kusevic et al. Jun 2010 A1
20100182874 Frank et al. Jul 2010 A1
20120075422 Wang et al. Mar 2012 A1
20120182540 Suzuki et al. Jul 2012 A1
20120206712 Chang et al. Aug 2012 A1
20120236379 Da Silva et al. Sep 2012 A1
20120310516 Zeng Dec 2012 A1
20120310519 Lawlor et al. Dec 2012 A1
20130088726 Goyal et al. Apr 2013 A1
20130093584 Schumacher Apr 2013 A1
20130120760 Raguin et al. May 2013 A1
20130166113 Dakin et al. Jun 2013 A1
20130206967 Shpunt et al. Aug 2013 A1
20130207970 Shpunt et al. Aug 2013 A1
20130222786 Hanson et al. Aug 2013 A1
20130250276 Chang et al. Sep 2013 A1
20130265561 Takahira et al. Oct 2013 A1
20140009747 Suzuki et al. Jan 2014 A1
20140036252 Amzajerdian et al. Feb 2014 A1
20140049609 Wilson et al. Feb 2014 A1
20140152975 Ko Jun 2014 A1
20140168631 Haslim et al. Jun 2014 A1
20140233942 Kanter Aug 2014 A1
20140313519 Shpunt et al. Oct 2014 A1
20150009485 Mheen et al. Jan 2015 A1
20150055117 Pennecot et al. Feb 2015 A1
20150234308 Lim et al. Aug 2015 A1
20150260843 Lewis Sep 2015 A1
20150301162 Kim Oct 2015 A1
20150371074 Lin Dec 2015 A1
20150378011 Owechko Dec 2015 A1
20160047895 Dussan Feb 2016 A1
20160047896 Dussan Feb 2016 A1
20160047903 Dussan Feb 2016 A1
20160138944 Lee et al. May 2016 A1
20160146926 Jungwirth May 2016 A1
20160178749 Lin et al. Jun 2016 A1
20160200161 Van Den Bossche et al. Jul 2016 A1
20160245902 Watnik et al. Aug 2016 A1
20160280229 Kasahara Sep 2016 A1
20160291160 Zweigle et al. Oct 2016 A1
20160357187 Ansari Dec 2016 A1
20160363669 Liu Dec 2016 A1
20160380488 Widmer et al. Dec 2016 A1
20170023678 Pink et al. Jan 2017 A1
20170090013 Paradie et al. Mar 2017 A1
20170102457 Li et al. Apr 2017 A1
20170199273 Morikawa et al. Jul 2017 A1
20170219696 Hayakawa et al. Aug 2017 A1
20170269215 Hall et al. Sep 2017 A1
20170270381 Itoh et al. Sep 2017 A1
20170285346 Pan Oct 2017 A1
20170307736 Donovan Oct 2017 A1
20170307737 Ishikawa et al. Oct 2017 A1
20170310948 Pei et al. Oct 2017 A1
20170329010 Warke et al. Nov 2017 A1
20170329011 Warke et al. Nov 2017 A1
20180052378 Shin et al. Feb 2018 A1
20180113193 Huemer et al. Apr 2018 A1
20180128903 Chang May 2018 A1
20180136328 Moss May 2018 A1
20180143309 Pennecot et al. May 2018 A1
20180180718 Lin Jun 2018 A1
20180224529 Wolf et al. Aug 2018 A1
20180241477 Turbide et al. Aug 2018 A1
20180275249 Campbell et al. Sep 2018 A1
20180275275 Lundquist Sep 2018 A1
20180284237 Campbell et al. Oct 2018 A1
20180284282 Hong et al. Oct 2018 A1
20180284286 Eichenholz et al. Oct 2018 A1
20180286909 Eichenholz et al. Oct 2018 A1
20180306913 Bartels Oct 2018 A1
20180341009 Niclass et al. Nov 2018 A1
20180364334 Xiang et al. Dec 2018 A1
20180372870 Puglia Dec 2018 A1
20190018143 Thayer et al. Jan 2019 A1
20190101644 De Mersseman et al. Apr 2019 A1
20190113200 Murakami Apr 2019 A1
20190123508 Hong et al. Apr 2019 A1
20190129009 Eichenholz et al. May 2019 A1
20190139951 T'ng et al. May 2019 A1
20190146060 Qiu et al. May 2019 A1
20190195990 Shand Jun 2019 A1
20190221988 Villeneuve et al. Jul 2019 A1
20190235064 Droz et al. Aug 2019 A1
20190242978 Weed et al. Aug 2019 A1
20190265336 Zhang et al. Aug 2019 A1
20190310351 Hughes et al. Oct 2019 A1
20200081129 De Mersseman et al. Mar 2020 A1
20200088847 De Mersseman et al. Mar 2020 A1
20200249354 Yeruhami et al. Aug 2020 A1
20200284906 Eichenholz et al. Sep 2020 A1
20200341120 Ahn et al. Oct 2020 A1
20200341121 Ahn et al. Oct 2020 A1
20210018602 De Mersseman et al. Jan 2021 A1
20210190919 de Mersseman Jun 2021 A1
20220146817 Erdl et al. May 2022 A1
20220333757 Li et al. Oct 2022 A1
20220403998 de Mersseman et al. Dec 2022 A1
Foreign Referenced Citations (37)
Number Date Country
509180 Jun 2011 AT
6638286 Jun 1987 AU
102508258 Jun 2012 CN
19731754 Feb 1999 DE
19757840 Sep 1999 DE
102004033944 Feb 2006 DE
102006031114 Jan 2008 DE
102008045387 Mar 2010 DE
102014218957 Mar 2016 DE
102015217908 Mar 2017 DE
102015224692 Jun 2017 DE
102016201606 Aug 2017 DE
0112188 Jun 1984 EP
0578129 Jan 1994 EP
2124069 Nov 2009 EP
2696166 Feb 2014 EP
2824418 Jan 2015 EP
3147685 Mar 2017 EP
3203259 Aug 2017 EP
3457080 Mar 2019 EP
201800001765 Jul 2019 IT
2002148556 May 2002 JP
2018041723 Mar 2018 JP
20190105889 Sep 2019 KR
1994019705 Sep 1994 WO
03009048 Jan 2003 WO
2008008970 Jan 2008 WO
2015014556 Feb 2015 WO
2016072483 May 2016 WO
2016097409 Jun 2016 WO
2016204138 Dec 2016 WO
2018229131 Dec 2018 WO
2018229131 Dec 2018 WO
2019050643 Mar 2019 WO
2019099166 May 2019 WO
2020243038 Dec 2020 WO
2020243038 Dec 2020 WO
Non-Patent Literature Citations (44)
Entry
“A milestone for laswer sensors in self-driving cars,” OSRAM Opto Semiconductors, Trade Press, Jul. 2016, 3 pages.
“Advanced Scientific Concepts,” http://www.advancedscientificconcepts.com/products/overview.html, 2015, 4 pages.
“Cameras,” Continental Automotive, https://www.continental-automotive.com/en-gl/Passenger-Cars/Chassis-Safety/Advanced-Driver-Assistance-Systems/Cameras, 2017, 2 pages.
“Hi-Res 3D Flash LIDAR will supplement Continental's existing portfolio for automated driving,” Continental AG, Mar. 2016, 2 pages.
“Multi Function Camera with Lidar,” Continental Automotive, https://www.continental-automotive.com/en-gl/Passenger-Cars/Chassis-Safety/Advanced-Driver-Assistance-Systems/Cameras/Multi-Function-Camera-with-Lidar, 2017, 2 pages.
Campbell et al., “Advanced sine wave modulation of continuous wave laser system for atmospheric CO2 differential absorption measurements,” NASA Langley Research Center; 32 pages.
Church et al., “Evaluation of a steerable 3D laser scanner using a double Risley prism pair,” SPIE Paper, 9 pages.
Hewlett-Packard Application Note 77-4, “Swept-Frequency Group Delay Measurements,” Hewlett-Packard Co., Sep. 1968, 7 pages.
Journet & Bazin, “A Low-Cost Laser Range Finder Based on an FMCW-Like Method,” IEEE Transactions on Instrumentation and Measurement, vol. 49, No. 4, Aug. 2000, 4 pages.
Kahn, “Modulation and Detection Techniques for Optical Communication Systems,” OSA/COTA, 2006, 3 pages.
Kasturi et al., “UAV-Borne LiDAR with MEMS Mirror Based Scanning Capability,” SPIE Defense and Commercial Sensing Conference, Apr. 2016, 10 pages.
Kravitz et al., “High-Resolution Low-Sidelobe Laser Ranging Based on Incoherent Pulse Compression,” IEEE Photonics Technology Letters, vol. 24, No. 23, Dec. 2012, 3 pages.
Levanon et al., “Non-coherent pulse compression—aperiodic and periodic waveforms,” IET Radar Sonar Navig., Jun. 2015, 9 pages.
Li et al., “Investigation of beam steering performances in rotation Risley-prism scanner,” OSA, Jun. 2016, 11 pages.
Li, “Time-of-Flight Camera—An Introduction,” Technical White Paper, SLOA190B, May 2014, 10 pages.
Luhmann, “A historical review on panorama photogrammetry,” University of Applied Sciences, Jul. 2008, 9 pages.
Niclass et al., “Development of Automotive LIDAR,” Electronics and Communications in Japan, vol. 98, No. 5, 2015, pp. 1-6.
Peer & Levanon, “Compression Waveforms for Non-Coherent Radar,” Compression Waveforms for Non-Coherent Radar, Tel Aviv University; 6 pages.
Pierrottet et al., “Linear FMCW Laser Radar for Precision Range and Vector Velocity Measurements,” Coherent Applications Inc. & NASA Langley Research Center, 9 pages.
Simpson et al., “Intensity-modulated, stepped frequency cw lidar for distributed aerosol an dhard target measurements,” Applied Optics, vol. 44, No. 33, Nov. 2005, 8 pages.
Skolnik, “Introduction to Radar Systems,” McGraw-Hill Higher Education, 2001, 6 pages.
Su et al., “2-D FFT and Time-Frequency Analysis Techniques for Multi-Target Recognition of FMCW Radar Signal,” Proceedings of the Asia-Pacific Microwave Conference, 2011, 4 pages.
THORLABS Application Note, Risley Prism Scanner; 33 pages.
Wang et al., “Range-Doppler image processing in linear FMCW radar and FPGA based real-time implementation,” Journal of Communication and Computer, vol. 6, No. 4, Apr. 2009, 5 pages.
Wien, “The Geometry of Airborne Laser Scanning in a Kinematical Framework,” Vienna University of Technology, Oct. 2016, 69 pages.
Winkler, “Range Doppler Detection for automotive FMCW Radars,” Proceedings of the 4th European Radar Conference, Oct. 2007, 4 pages.
Wojtkiewicz et al., “Two-dimensional signal processing in FMCW radars,” Instytut Podstaw Elektroniki, 6 pages.
Invitation to Pay Additional Fees and, Where Applicable, Protest Fee issued in International Application No. PCT/US2018/052849 on Mar. 8, 2019.
International Search Report and Written Opinion issued in International Application No. PCT/US2018/052849 on May 6, 2019.
International Search Report and Written Opinion issued in International Application No. PCT/US2018/057676 on Jan. 23, 2019.
International Search Report and Written Opinion issued in International Application No. PCT/US2019/046800 on Nov. 25, 2019.
International Search Report and Written Opinion issued in International Application No. PCT/US2020/039760 on Sep. 18, 2020.
International Search Report and Written Opinion issued in International Application No. PCT/US2020/064474 on Apr. 1, 2021.
International Search Report and Written Opinion issued in International Application No. PCT/US2017/033263 on Aug. 29, 2017.
International Search Report and Written Opinion issued in International Application No. PCT/US2017/033265 on Sep. 1, 2017.
International Search Report and Written Opinion issued in International Application No. PCT/US2017/033271 on Sep. 1, 2017.
International Search Report and Written Opinion issued in International Application No. PCT/US2018/048869 on Nov. 8, 2018.
International Search Report and Written Opinion issued in International Application No. PCT/US2018/049038 on Dec. 12, 2018.
International Search Report and Written Opinion issued in International Application No. PCT/US2018/051281 on Nov. 22, 2018.
International Search Report and Written Opinion issued in International Application No. PCT/US2018/052837 on Jan. 24, 2019.
International Search Report and Written Opinion issued in International Application No. PCT/US2018/054992 on Dec. 11, 2018.
International Search Report and Written Opinion issued in International Application No. PCT/US2018/057727 on Jan. 28, 2019.
International Search Report and Written Opinion issued in International Application No. PCT/US2023/034129 on Jan. 3, 2024.
International Search Report and Written Opinion issued in International Application No. PCT/US2023/034131 on Dec. 21, 2023.
Related Publications (1)
Number Date Country
20240118425 A1 Apr 2024 US