The present invention relates generally to a system for a vehicle that utilizes one or more microphones at a vehicle to detect sound at or near the vehicle.
Use of imaging sensors in vehicle imaging systems is common and known. Examples of such known systems are described in U.S. Pat. Nos. 5,949,331; 5,670,935 and/or 5,550,677, which are hereby incorporated herein by reference in their entireties. Such imaging systems may be used to determine other vehicles and/or lights of vehicles, such as headlights and/or taillights and/or flashing lights of emergency vehicles.
The present invention provides a control system for a vehicle that utilizes a plurality of microphones (preferably arranged in one or more arrays of at least two microphones) disposed at the exterior of the vehicle for sensing sound around the vehicle. An audio processor receives and processes outputs of the microphones to determine or recognize sounds (such as emergency vehicle sirens or the like) detected or sensed at the microphones and to identify or classify the detected sounds and determine the direction from which the sounds are coming from. Responsive to identification of an emergency vehicle siren, the system may generate an alert to a driver of a non-autonomous vehicle or to an occupant of an autonomous vehicle and/or may control a steering system and/or braking system and/or accelerating system of the equipped vehicle to make way for the determined emergency vehicle and/or to pull over the equipped vehicle to the side of the road and stop if it is determined that the detected emergency vehicle is on the road on which the equipped vehicle is traveling and is approaching the equipped vehicle (either from ahead or behind the equipped vehicle).
The system of the present invention thus provides, such as for an autonomous vehicle that is operating autonomously, a means to detect the likes of a siren or other emergency vehicle audible warning being emitted by the likes of a siren of an ambulance or fire engine or police vehicle or the like, and, responsive to such determination, a control of the equipped vehicle generally determines the location of the detected emergency vehicle relative to the equipped vehicle, determines whether the detected emergency vehicle is approaching or receding from the equipped vehicle, and determines proximity of the detected emergency vehicle to the equipped vehicle.
These and other objects, advantages, purposes and features of the present invention will become apparent upon review of the following specification in conjunction with the drawings.
In-vehicle environments are currently highly insulated against external noise. As a result, it is difficult for the driver to detect the presence of vehicles on emergency travel (such as ambulances, police, fire-fighting trucks and/or the like), especially when all of the vehicle's windows are fully closed and the vehicle's audio system is playing. As inner city roads become ever busier, it becomes important to design and implement systems that alert the driver of the proximity of incoming emergency vehicles and, as importantly, which direction they are approaching from.
For autonomous vehicles suitable for deployment with the system of the present invention, an occupant of the vehicle may, under particular circumstances, be desired or required to take over operation/control of the vehicle and drive the vehicle so as to avoid potential hazard for as long as the autonomous system relinquishes such control or driving. Such occupant of the vehicle thus becomes the driver of the autonomous vehicle. As used herein, the term “driver” refers to such an occupant, even when that occupant is not actually driving the vehicle, but is situated in the vehicle so as to be able to take over control and function as the driver of the vehicle when the vehicle control system hands over control to the occupant or driver or when the vehicle control system is not operating in an autonomous or semi-autonomous mode.
Typically an autonomous vehicle would be equipped with a suite of sensors, including multiple machine vision cameras deployed at the front, sides and rear of the vehicle, multiple radar sensors deployed at the front, sides and rear of the vehicle, and/or multiple lidar sensors deployed at the front, sides and rear of the vehicle. Typically, such an autonomous vehicle will also have wireless two way communication with other vehicles or infrastructure, such as via a car2car (V2V) or car2x communication system.
An autonomous vehicle needs to detect emergency vehicles when an emergency vehicle is actively responding to an emergency (such as when the emergency vehicle's sirens and flashing lights are active). The autonomous vehicle will have to react in a specific way that is mandated by local legislation. The autonomous vehicle will have to understand the distance and heading of the emergency vehicle with respect to the emergency vehicle's own position. When the emergency vehicle approaches from behind the autonomous vehicle, the autonomous vehicle may have to drive to the right side of the road and come to a full stop until the emergency vehicle has passed, or on a multi-lane highway, the autonomous vehicle may have to move away from the emergency vehicle in order to let the emergency vehicle pass. If the emergency vehicle approaches from the front of the autonomous vehicle, the autonomous vehicle may have to come to a stop as well (single lane highway), or may be able to continue on without specific reaction (such as if the autonomous vehicle is traveling along an interstate with a barrier). If the emergency vehicle approaches from the left or the right of the autonomous vehicle, the autonomous vehicle needs to be aware that the emergency vehicle may ignore road rules (such as stop signs, traffic lights and the like). The autonomous vehicle should detect an emergency vehicle in conditions when it is not possible to optically ‘see’ the emergency vehicle (such as in an “urban canyon”).
The present invention provides sound based emergency vehicle detection that is operable to detect presence and direction of emergency vehicles using microphones and digital signal processing (DSP) algorithms. The system may communicate signals using a vehicle communication network or bus (such as a controller area network (CAN) bus or the like) and may send signals to sensor fusion or some other module of the vehicle. The system comprises three stages: Acoustic Detection, Direction Extraction and Acoustic Information Relay.
The system of the present invention will recognize an emergency vehicle when the siren is activated but will not recognize an emergency vehicle that drives with flashing lights only. Thus, it is preferred that the invention will be part of a system that will rely on other means to detect active emergency vehicles (such as, for example, V2X communication systems and/or vehicle vision or optical systems or the like).
The system of the present invention uses an array of microphones and a digital signal processor (DSP) to discriminate sirens from ambient noise. The microphone array will be used to triangulate the source and to determine the angle and distance respective to the system. The resulting information may be transmitted on a vehicle communication bus (LIN, CAN, FlexRay or the like).
Sound-based emergency vehicle detection (SBEVD) can provide awareness to the driver of the equipped vehicle that an emergency vehicle is in the area and can provide awareness as to the direction of the sound. Modern automobiles have extensive methods to reduce outside noise which make it more difficult for drivers to detect emergency vehicles (EVs). As such, the system provides situational awareness of the EV in the area. The system also provides direction information of the EV for the driver in relation to the driven or equipped vehicle. This provides further information to the driver such that appropriate driver actions can be taken. The system may also take actions to enhance the driver's ability to hear the emergency vehicle, such actions including, for example, controlling the audio environment within the vehicle, such as turning down the radio or other audio sources. Further enhancements may include approximation of the path of the emergency vehicle to determine actions either for assisted driving or autonomous vehicle control. Potentially, the audio system in the vehicle can be programmed or controlled to provide a filtered or synthesized directional sound to give the driver an enhanced directional feel for the EV location (such as by controlling in-vehicle microphones to generate an EV sound or other alert that, due to the control of multiple in-vehicle microphones, sounds as if it coming from the determined direction of the EV, such as by utilizing aspects of the systems described in U.S. Publication No. US-2016-0029111, which is hereby incorporated herein by reference in its entirety). Networking to adjacent vehicles, GPS and cloud based sensor fusion are all possible enhancements with the SBEVD when implemented in an autonomous vehicle (AV).
The EV's (Emergency Vehicles, including police, fire and paramedic) often use high amplitude screeching sirens to warn road users and pedestrians that the emergency vehicle is approaching. A visual alarm provided by the vehicle detection of the oncoming emergency vehicle sound can improve driver safety for people with hearing impairments.
The SBEVD provides decoding of EV signatures with microphones (i.e., pressure transducers) placed on or disposed at the vehicle (such as at an exterior portion of the vehicle, such as at the exterior rearview mirrors or at the front and/or rear bumpers and/or at a shark-fin antenna or the like). Multiple microphones placed at some physical distance around the vehicle enable triangulation and better discrimination of the sound emanating from the EV. Multiple microphones enable more precision and reduced clutter, multipath, and other impairments in the acoustic environment. Consideration and technical solutions can be made to account for locally generated noise, such as vehicle, road, engine, body vibration and other noise sources which will degrade the acoustical signals. The use of small clusters of microphones can be used to provide beam steering mitigation of these complicating factors.
In general, the further the microphones are apart, the better the triangulation. In a clean environment, a minimum of 3 microphones is required. However, in the real world with impairments and external noise sources, having more microphones is advantageous, safer and more robust. The system of the present invention may include a limited number of microphones dispersed around the automobile. For example, and such as shown in
The computational overhead for acoustics is substantial to provide a high level of confidence against real world impairments. However, the computational overhead is an order of magnitude less than that of vision systems for AVs and thus can be implemented in a reasonably compact and low power system. It is envisioned that the CAN bus can be used to carry information from the microphone clusters to a dedicated acoustic processor for high level filtering and processing. Both the microphone clusters and the Acoustics/Sound-ECU can be accommodated on the vehicle CAN bus (see
The physical location of the microphones has two conflicting factors. On the one hand, physical protection is important for automotive use given the various environmental challenges. On the other hand, the best location for microphones is open on the surface of the vehicle. Thus, the system may include exposed and hidden microphone placed around the vehicle.
With reference to
The siren frequency is fixed and programmed in the microcontroller. The microcontroller can use the Doppler shift of the signal (higher frequency means that S is approaching, lower frequency means S is receding) to determine the speed of the siren relative to the vehicle. The system will utilize the vehicle speed information (such as from the vehicle CAN) and the angle information from the microphone array to estimate the emergency vehicle speed by subtracting the vectors. By using map data (which determines the emergency vehicle path determined by the road), the estimate can be refined and the system may determine if the emergency vehicle is traveling along a path that may intersect with the path of travel of the equipped vehicle.
The sound signatures (such as sound signatures of known emergency vehicle sirens) can be pre-programmed in the SBEVD units (so the system may more quickly and accurately detect and identify an emergency vehicle siren). Also if need be, wireless or over the air (OTA) communications could be used to provide updates or localizations for differing jurisdictions of the siren types (so the system may be configured or adapted or programmed for the types of sirens used by emergency vehicles in that geographical location).
The system may process extracted siren signals to obtain the source directions using various techniques, such as sound intensity processing, time delay methods or triangulation. For example, using an intensity based approach, an array of directional microphones may be arranged at the vehicle (such as in a circle facing away from the center of the circle), whereby the system may determine that the sound is coming from the direction where the microphone with the loudest signal is facing. The resolution in direction is limited by the number of microphones, and more microphones allow for enhanced monitoring of more sectors.
Optionally, using a time lag based approach, an array of omnidirectional microphones may be implemented and trigonometry is used to calculate the direction from the time lags between the sensors' sensing of a given sound. Given enough spacing between the microphones, resolution in direction can be very high with just three microphones. Such a configuration is less susceptible to signal to noise issues compared to the intensity based approach.
The system may also determine a frequency shift in the sound to determine if the emergency vehicle is going away from the equipped vehicle or approaching the equipped vehicle (the Doppler Effect). If it is determined that the emergency vehicle is ahead of the equipped vehicle and approaching the equipped vehicle, the system may control the vehicle braking and/or steering to pull the vehicle toward the side of the road. Similarly, if it is determined that the emergency vehicle is approaching an intersection ahead of the equipped vehicle, the system may control the vehicle braking to slow or stop the equipped vehicle before reaching the intersection.
Thus, the system may process output signals of at least three microphones at the vehicle to determine or detect the presence of an emergency vehicle with its siren on. For example, the microphones may communicate outputs to a processor that processes the outputs to detect and/or identify the siren or sound source, and to determine the direction or location of the sound source (see
By processing the microphone outputs, the system determines the direction of the emergency vehicle relative to the equipped vehicle. For example, an algorithm can determine the time delays between the microphones by finding how much the signals must be shifted to align them, and then can calculate the direction angle from the determined time delays. There is a tradeoff between microphone spacing and sampling frequency, but a spacing of about 1 foot and sampling frequency of about 20 kHz works well. The system may include a bandpass filter and may include filtering and/or pattern recognition to enhance detection and identification of specific sound sources (such as police sirens, etc.).
Preferably, the array of microphones are provided in a unitary module installable at an exterior body location of the vehicle. The microphone or audio module is provided complete with the microphones and any necessary power supply and an audio processor and associated circuitry, and provides an output control signal to the vehicle (preferably via connection to a vehicle communication bus, such as a LIN or LAN or CAN bus). Multiple microphone or audio modules may be installed at the equipped vehicle, such as at a driver-side portion (preferably at or incorporated into a driver-side exterior rearview mirror assembly or at a driver-side camera assembly, whereupon economy can be achieved by sharing circuitry and processors otherwise present in the exterior rearview mirror assembly or camera module or assembly), and such as at a passenger-side portion (preferably at or incorporated into a passenger-side exterior rearview mirror assembly or at a passenger-side camera assembly, whereupon economy can be achieved by sharing circuitry and processors otherwise present in the exterior rearview mirror assembly or camera module or assembly), and such as at a front and/or rear portion of the vehicle (preferably at or incorporated into a front or rear camera assembly, whereupon economy can be achieved by sharing circuitry and processors otherwise present in the camera module or assembly). Alternatively, or additionally, an audio module may be installed at an exterior roof portion of the vehicle, preferably at or along a centerline of the roof, such as at or part of a shark-fin antenna of the vehicle.
Thus, by using an array of microphones and the likes of audio triangulation based on individual audio processing of sound signals received by the respective microphones of the array, the location of the determined EV can be achieved. The system operates independent of visibility constraints (fog, snow, urban canyon). The system is relatively simple (3 or 4 microphones or microphone arrays and one microcontroller) and less expensive than vision systems and the like. The system can be used in non-autonomous vehicles to alert drivers listening to loud music or being hard of hearing.
Upon initial detection and initial determination of the presence and proximity/location of a detected emergency vehicle, sensors, such as cameras, radar sensors and/or lidar sensors can supplement or augment identification and location of the EV as the EV enters the field of view of the camera and/or field of sensing of the radar/lidar sensors. Also, the system may be augmented by optical or V2X systems to detect active emergency vehicles with flashing lights only.
Optionally, for example, the system may operate in conjunction with one or more cameras at the vehicle and/or one or more other sensors. For example, the sound processor may detect the presence of a siren of an emergency vehicle, and a vision system of the equipped vehicle (including one or more cameras and an image processor that processes image data captured by the one or more cameras) may provide confirmation (via image processing of captured image data) that the determined siren/object making the noise is actually part of an emergency vehicle. The vision system may operate to focus its image processing on image data representative of the area of the cameras' fields of view where the microphone system determines the presence of the activated siren. For example, if the microphone system or emergency vehicle detection system determines that an emergency vehicle siren is activated ahead of the equipped vehicle and to the left of the equipped vehicle, the image processor may focus processing of image data captured by the forward viewing camera and may further focus processing of image data captured by a left region imager of the forward viewing camera so as to quickly confirm the presence of the emergency vehicle.
The microphones may be incorporated in camera modules that are disposed at selected locations at the vehicle (such as at the front, rear and sides of the vehicle). Each camera module includes a camera that has a respective field of view (e.g., forward, rearward, sideward) exterior the vehicle. Each camera module includes at least one microphone that senses sounds emanating from exterior the vehicle. The microphones and cameras are collocated at the respective modules and may share components and/or circuitry, and the module may communicate or transmit outputs representative of microphone outputs and captured image data, such as to a central processor or control of the vehicle (such as via a vehicle network or communication bus or the like).
Cohabitation/colocation of the plurality of microphones with the likes of a forward/sideward/rearward viewing vehicular cameras has several advantages. A common microprocessor and/or digital signal processor can be used both for digital sound processing of signals provided by the plurality of microphones, and digital image processing of image data captured by one or more cameras. Moreover, digital image processing of image data captured by the camera may confirm and/or determine presence of an/the emergency vehicle (such as by character recognition of signage or the like on the vehicle indicating it to be an emergency vehicle, or by detection of flashing or strobing of lights on the vehicle indicating it to be an emergency vehicle or the like). For example, a vehicle camera may include in its housing at least one and preferably a plurality of microphones. Furthermore, modern vehicles come equipped with multiple cameras located around the vehicle (such as a front camera, a rear camera, and at least two side cameras). Thus, each camera module may include a camera and a single microphone, whereby the plurality of cameras provide the array of microphones.
The system may process outputs of the microphones to detect a siren or sound indicative of an emergency vehicle, and may determine a direction from the vehicle toward where the sound is emanating from. Responsive to the detected siren and determined source direction, the system may process image data captured by one or more of the cameras so that image data captured by at least one camera that has its field of view at least partially encompassing the area/direction from where the siren or noise is emanating. The system may process the captured image data of at least that camera to identify or confirm that the detected siren is emanating from an emergency vehicle (such as via processing image data to detect characters or signage or icons on the vehicle or to detect flashing lights or the like on the vehicle).
The camera or sensor may comprise any suitable camera or sensor. Optionally, the camera may comprise a “smart camera” that includes the imaging sensor array and associated circuitry and image processing circuitry and electrical connectors and the like as part of a camera module, such as by utilizing aspects of the vision systems described in International Publication Nos. WO 2013/081984 and/or WO 2013/081985, which are hereby incorporated herein by reference in their entireties.
The system may include an image processor operable to process image data captured by the camera or cameras, such as for detecting objects or other vehicles or pedestrians or the like in the field of view of one or more of the cameras. For example, the image processor may comprise an image processing chip selected from the EyeQ family of image processing chips available from Mobileye Vision Technologies Ltd. of Jerusalem, Israel, and may include object detection software (such as the types described in U.S. Pat. Nos. 7,855,755; 7,720,580 and/or 7,038,577, which are hereby incorporated herein by reference in their entireties), and may analyze image data to detect vehicles and/or other objects. Responsive to such image processing, and when an object or other vehicle is detected, the system may generate an alert to the driver of the vehicle and/or may generate an overlay at the displayed image to highlight or enhance display of the detected object or vehicle, in order to enhance the driver's awareness of the detected object or vehicle or hazardous condition during a driving maneuver of the equipped vehicle.
For example, the vision system and/or processing and/or camera and/or circuitry may utilize aspects described in U.S. Pat. Nos. 9,233,641; 9,146,898; 9,174,574; 9,090,234; 9,077,098; 8,818,042; 8,886,401; 9,077,962; 9,068,390; 9,140,789; 9,092,986; 9,205,776; 8,917,169; 8,694,224; 7,005,974; 5,760,962; 5,877,897; 5,796,094; 5,949,331; 6,222,447; 6,302,545; 6,396,397; 6,498,620; 6,523,964; 6,611,202; 6,201,642; 6,690,268; 6,717,610; 6,757,109; 6,802,617; 6,806,452; 6,822,563; 6,891,563; 6,946,978; 7,859,565; 5,550,677; 5,670,935; 6,636,258; 7,145,519; 7,161,616; 7,230,640; 7,248,283; 7,295,229; 7,301,466; 7,592,928; 7,881,496; 7,720,580; 7,038,577; 6,882,287; 5,929,786 and/or 5,786,772, and/or U.S. Publication Nos. US-2014-0340510; US-2014-0313339; US-2014-0347486; US-2014-0320658; US-2014-0336876; US-2014-0307095; US-2014-0327774; US-2014-0327772; US-2014-0320636; US-2014-0293057; US-2014-0309884; US-2014-0226012; US-2014-0293042; US-2014-0218535; US-2014-0218535; US-2014-0247354; US-2014-0247355; US-2014-0247352; US-2014-0232869; US-2014-0211009; US-2014-0160276; US-2014-0168437; US-2014-0168415; US-2014-0160291; US-2014-0152825; US-2014-0139676; US-2014-0138140; US-2014-0104426; US-2014-0098229; US-2014-0085472; US-2014-0067206; US-2014-0049646; US-2014-0052340; US-2014-0025240; US-2014-0028852; US-2014-005907; US-2013-0314503; US-2013-0298866; US-2013-0222593; US-2013-0300869; US-2013-0278769; US-2013-0258077; US-2013-0258077; US-2013-0242099; US-2013-0215271; US-2013-0141578 and/or US-2013-0002873, which are all hereby incorporated herein by reference in their entireties.
The system may utilize sensors, such as radar or lidar sensors or the like. The sensing system may utilize aspects of the systems described in U.S. Pat. Nos. 9,753,121; 9,689,967; 9,599,702; 9,575,160; 9,146,898; 9,036,026; 8,027,029; 8,013,780; 6,825,455; 7,053,357; 7,408,627; 7,405,812; 7,379,163; 7,379,100; 7,375,803; 7,352,454; 7,340,077; 7,321,111; 7,310,431; 7,283,213; 7,212,663; 7,203,356; 7,176,438; 7,157,685; 6,919,549; 6,906,793; 6,876,775; 6,710,770; 6,690,354; 6,678,039; 6,674,895 and/or 6,587,186, and/or International Publication No. WO 2011/090484 and/or U.S. Publication Nos. US-2017-0276788; US-2017-0254873; US-2017-0222311 and/or US-2010-0245066, which are hereby incorporated herein by reference in their entireties.
The system may also communicate with other systems, such as via a vehicle-to-vehicle communication system or a vehicle-to-infrastructure communication system or the like. Such car2car or vehicle to vehicle (V2V) and vehicle-to-infrastructure (car2X or V2X or V2I or a 4G or 5G broadband cellular network) technology provides for communication between vehicles and/or infrastructure based on information provided by one or more vehicles and/or information provided by a remote server or the like. Such vehicle communication systems may utilize aspects of the systems described in U.S. Pat. Nos. 6,693,517 and/or 7,580,795, and/or U.S. Publication Nos. US-2014-0375476; US-2014-0218529; US-2013-0222592; US-2012-0218412 and/or US-2012-0062743, which are hereby incorporated herein by reference in their entireties.
Changes and modifications in the specifically described embodiments can be carried out without departing from the principles of the invention, which is intended to be limited only by the scope of the appended claims, as interpreted according to the principles of patent law including the doctrine of equivalents.
The present application claims the filing benefits of U.S. provisional application Ser. No. 62/450,680, filed Jan. 26, 2017, which is hereby incorporated herein by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
3997868 | Ribnick et al. | Dec 1976 | A |
4930742 | Schofield et al. | Jun 1990 | A |
4956866 | Bernstein et al. | Sep 1990 | A |
4959865 | Stettiner et al. | Sep 1990 | A |
4975966 | Sapiejewski | Dec 1990 | A |
5287411 | Hill et al. | Feb 1994 | A |
5329593 | Lazzeroni et al. | Jul 1994 | A |
5495242 | Kick et al. | Feb 1996 | A |
5550677 | Schofield et al. | Aug 1996 | A |
5670935 | Schofield et al. | Sep 1997 | A |
5671996 | Bos et al. | Sep 1997 | A |
5703957 | McAteer | Dec 1997 | A |
5820245 | Desmond et al. | Oct 1998 | A |
5828012 | Repolle et al. | Oct 1998 | A |
5850016 | Jung et al. | Dec 1998 | A |
5877897 | Schofield et al. | Mar 1999 | A |
5878147 | Killion et al. | Mar 1999 | A |
5894279 | Rose et al. | Apr 1999 | A |
5949331 | Schofield et al. | Sep 1999 | A |
5979586 | Farmer et al. | Nov 1999 | A |
6166625 | Teowee et al. | Dec 2000 | A |
6243003 | DeLine et al. | Jun 2001 | B1 |
6278377 | DeLine et al. | Aug 2001 | B1 |
6326613 | Heslin et al. | Dec 2001 | B1 |
6329925 | Skiver et al. | Dec 2001 | B1 |
6362749 | Brill | Mar 2002 | B1 |
6363156 | Roddy | Mar 2002 | B1 |
6366213 | DeLine et al. | Apr 2002 | B2 |
6420975 | DeLine et al. | Jul 2002 | B1 |
6428172 | Hutzel et al. | Aug 2002 | B1 |
6433676 | DeLine et al. | Aug 2002 | B2 |
6485081 | Bingle et al. | Nov 2002 | B1 |
6501387 | Skiver et al. | Dec 2002 | B2 |
6570992 | Folan et al. | May 2003 | B1 |
6614911 | Watson et al. | Sep 2003 | B1 |
6648477 | Hutzel et al. | Nov 2003 | B2 |
6690268 | Schofield et al. | Feb 2004 | B2 |
6693517 | McCarthy et al. | Feb 2004 | B2 |
6717524 | DeLine et al. | Apr 2004 | B2 |
6774356 | Heslin et al. | Aug 2004 | B2 |
6798890 | Killion et al. | Sep 2004 | B2 |
6882734 | Watson et al. | Apr 2005 | B2 |
6906632 | DeLine et al. | Jun 2005 | B2 |
6958707 | Siegel | Oct 2005 | B1 |
6980092 | Turnbull et al. | Dec 2005 | B2 |
6980663 | Linhard | Dec 2005 | B1 |
7038577 | Pawlicki et al. | May 2006 | B2 |
7061402 | Lawson | Jun 2006 | B1 |
7245232 | Caouette, Sr. | Jul 2007 | B1 |
7308341 | Schofield et al. | Dec 2007 | B2 |
7415116 | Fels | Aug 2008 | B1 |
7580795 | McCarthy et al. | Aug 2009 | B2 |
7657052 | Larson et al. | Feb 2010 | B2 |
7675431 | Caouette, Sr. | Mar 2010 | B1 |
7720580 | Higgins-Luthman | May 2010 | B2 |
7791499 | Mohan et al. | Sep 2010 | B2 |
7855755 | Weller et al. | Dec 2010 | B2 |
8094040 | Cornett et al. | Jan 2012 | B1 |
8258932 | Wahlstrom | Sep 2012 | B2 |
8275145 | Buck et al. | Sep 2012 | B2 |
8319620 | Usher et al. | Nov 2012 | B2 |
8355521 | Larson et al. | Jan 2013 | B2 |
8392064 | Thrun et al. | Mar 2013 | B2 |
8676427 | Ferguson et al. | Mar 2014 | B1 |
8824697 | Christoph | Sep 2014 | B2 |
8849557 | Levandowski | Sep 2014 | B1 |
9224294 | St. John | Dec 2015 | B1 |
9275136 | Sharifi et al. | Mar 2016 | B1 |
9278689 | Delp | Mar 2016 | B1 |
9397630 | Wang et al. | Jul 2016 | B2 |
9412273 | Ricci | Aug 2016 | B2 |
9417838 | Baalu et al. | Aug 2016 | B2 |
9576208 | Agnew et al. | Feb 2017 | B2 |
9800983 | Wacquant et al. | Oct 2017 | B2 |
10789840 | Boykin | Sep 2020 | B2 |
20020032510 | Turnbull et al. | Mar 2002 | A1 |
20020080021 | Skiver et al. | Jun 2002 | A1 |
20020110255 | Killion et al. | Aug 2002 | A1 |
20020110256 | Watson et al. | Aug 2002 | A1 |
20040155795 | Quintana | Aug 2004 | A1 |
20040170286 | Durach et al. | Sep 2004 | A1 |
20040258252 | Inoue et al. | Dec 2004 | A1 |
20050074131 | McCall et al. | Apr 2005 | A1 |
20060023892 | Schultz | Feb 2006 | A1 |
20060055630 | Cheang et al. | Mar 2006 | A1 |
20070008175 | Johnson et al. | Jan 2007 | A1 |
20070146127 | Stilp et al. | Jun 2007 | A1 |
20070216539 | D'Antoni et al. | Sep 2007 | A1 |
20080068187 | Bonefas et al. | Mar 2008 | A1 |
20080150786 | Breed | Jun 2008 | A1 |
20090028425 | Cavallaro et al. | Jan 2009 | A1 |
20090097674 | Watson et al. | Apr 2009 | A1 |
20090125311 | Haulick et al. | May 2009 | A1 |
20090179774 | Mohan et al. | Jul 2009 | A1 |
20090322559 | Yen et al. | Dec 2009 | A1 |
20100100284 | Kudo et al. | Apr 2010 | A1 |
20110199230 | Stahlin et al. | Aug 2011 | A1 |
20110296794 | Thomas et al. | Dec 2011 | A1 |
20120062743 | Lynam et al. | Mar 2012 | A1 |
20120072051 | Koon et al. | Mar 2012 | A1 |
20120121113 | Li | May 2012 | A1 |
20120136559 | Rothschild | May 2012 | A1 |
20120218412 | Dellantoni et al. | Aug 2012 | A1 |
20120230504 | Kuroda | Sep 2012 | A1 |
20130049985 | Eisenson et al. | Feb 2013 | A1 |
20130222592 | Gieseke | Aug 2013 | A1 |
20130223643 | Sato et al. | Aug 2013 | A1 |
20140078291 | Hammarstrom et al. | Mar 2014 | A1 |
20140218529 | Mahmoud et al. | Aug 2014 | A1 |
20140266853 | Orr et al. | Sep 2014 | A1 |
20140350927 | Yamabe et al. | Nov 2014 | A1 |
20140375476 | Johnson et al. | Dec 2014 | A1 |
20150025709 | Spaulding et al. | Jan 2015 | A1 |
20150061895 | Ricci | Mar 2015 | A1 |
20150065060 | Stahlin | Mar 2015 | A1 |
20150066284 | Yopp | Mar 2015 | A1 |
20150137998 | Marti et al. | May 2015 | A1 |
20150142244 | You et al. | May 2015 | A1 |
20150161458 | Agnew et al. | Jun 2015 | A1 |
20150181175 | Camilleri et al. | Jun 2015 | A1 |
20150294169 | Zhou et al. | Oct 2015 | A1 |
20150296135 | Wacquant et al. | Oct 2015 | A1 |
20150348580 | van Hoff | Dec 2015 | A1 |
20150382102 | Akino | Dec 2015 | A1 |
20160029111 | Wacquant | Jan 2016 | A1 |
20160094911 | Kropf | Mar 2016 | A1 |
20160191815 | Annau | Jun 2016 | A1 |
20160252905 | Tian et al. | Sep 2016 | A1 |
20160355124 | Cervantes Guarneros et al. | Dec 2016 | A1 |
20160355125 | Herbert | Dec 2016 | A1 |
20170323540 | Boykin | Nov 2017 | A1 |
20180050698 | Polisson | Feb 2018 | A1 |
20180077506 | Wacquant et al. | Mar 2018 | A1 |
20180137756 | Moosaei | May 2018 | A1 |
20180261237 | Moore et al. | Sep 2018 | A1 |
20180350391 | Moore et al. | Dec 2018 | A1 |
Number | Date | Country |
---|---|---|
102013004271 | Sep 2013 | DE |
1998017046 | Apr 1998 | WO |
199031637 | Jun 1999 | WO |
2001037519 | May 2001 | WO |
2003091066 | Nov 2003 | WO |
2014204794 | Dec 2014 | WO |
Entry |
---|
Real F.D., Berry F. (2009) Smart Cameras: Technologies and Applications. In: Belbachir A. (eds) Smart Cameras. Springer, Boston , MA. 2009 (Year: 2009). |
Bendler et al., “Emergency Vehicle Detector,” ECE4007 Senior Design Project, Sep. 15, 2008. |
Fazenda et al., “Acoustic based safety emergency vehicle detection for intelligent transport systems,” IEEE ICCAS-SICE Sep. 2009. |
Fragoulis et al., “A Siren Detection System Based on Mechanical Resonant Filters,” Sensors 2001, vol. 1, pp. 121-137, ISSN 1424-8220, MDPI, 2001. |
Park et al., “Automatic Detection of Emergency Vehicles for Hearing Impaired Drivers,” Texas A&M University-Kingsville, EE/CS Department, 2000. |
Schröder et al., “Automatic Acoustic Siren Detection in Traffic Noise By Part-Based Models,” University of Oldenburg, Germany, May 2013. |
Schwander et al., “Effect of two-microphone noise reduction on speech recognition by normal-hearing listeners” Veterans Administration, Journal of Rehabilitation Research and Development, vol. 24, No. 4, pp. 87-92, Fall 1987. |
Simmer et al., “Adaptive Microphone Arrays for Noise Suppression in the Frequency Domain” Second Cost 229 Workshop on Adaptive Algorithms in Communications, Bordeaux, 30.92.10.1992. |
Zhang et al., “High-speed Noise Cancellation with Microphone Array” Technology Reports, NTT DoCoMo Technical Journal vol. 9, No. 4. |
Number | Date | Country | |
---|---|---|---|
20180211528 A1 | Jul 2018 | US |
Number | Date | Country | |
---|---|---|---|
62450680 | Jan 2017 | US |