The present invention relates generally to a vehicle vision system for a vehicle and, more particularly, to a vehicle vision system that utilizes a forward facing camera, such as at a vehicle windshield.
Use of imaging sensors in vehicle imaging systems is common and known. Examples of such known systems are described in U.S. Pat. Nos. 5,949,331; 5,670,935 and/or 5,550,677, which are hereby incorporated herein by reference in their entireties.
Dedicated Short Range Communication (DSRC) radio technology has been developed in recent past to enable communications-based active safety systems. The communication link for such applications needs a reliable, high speed, low latency that is immune to extreme weather conditions, works reliably in high speed mobility conditions and multipath roadway environments.
The present invention provides a communication system and vision system or imaging system for a vehicle that utilizes one or more cameras (preferably one or more CMOS cameras) to capture image data representative of images exterior of the vehicle, and provides a windshield camera having an integrated DSRC radio with multi diversified array antenna. The present invention comprises a methods and apparatus to enhance the performance of DSRC communication in the vehicle utilizing antenna diversity, dynamic beam pattern control, and dynamic transmission power control methodology utilizing the forward viewing camera.
In accordance with an aspect of the present invention, a communication system for a vehicle includes an antenna array for transmitting data to at least one other vehicle or structure. A control is operable to adjust a beam transmission of the antenna array responsive to determination of a driving condition of the vehicle. The antenna array may be disposed at a camera of the vehicle, such as at a forward viewing windshield mounted camera of the vehicle. For example, the antenna array may be integrated in a camera mounting bracket at the windshield of the vehicle. Optionally, a second antenna array may be disposed at a rearward portion of the vehicle, such as at a “shark fin” antenna element or a rear camera of the vehicle. The control may be operable to adjust a beam transmission of one of the antenna arrays independent of the beam transmission of the other of the antenna arrays. For example, the control may adjust the beam transmission of one of the antenna arrays (such as the front array) to an omnidirectional beam and to adjust the beam transmission of the other of the antenna arrays (such as the rear array) to a directional beam so as to provide enhanced transmission range of the other antenna array.
The control may adjust the beam transmission from an omnidirectional beam to a directed beam to enhance the transmission range of the beam. For example, the control may adjust the beam transmission to the directed beam directed rearward of the vehicle responsive to a determination of a highway driving condition. The control may adjust the beam transmission responsive to a determination of at least one of (i) a highway driving condition, (ii) a high traffic driving condition, (iii) an intersection driving condition and (iv) an adverse weather condition.
The antenna array is associated with a DSRC radio of the vehicle. The control may adjust a power output of the DSRC radio responsive to a driving condition. For example, the control may adjust the power output of the DSRC radio responsive to a determination of an adverse weather condition.
These and other objects, advantages, purposes and features of the present invention will become apparent upon review of the following specification in conjunction with the drawings.
A vehicle vision system and/or driver assist system and/or object detection system and/or alert system operates to capture images exterior of the vehicle and may process the captured image data to detect objects at or near the vehicle and in the predicted path of the vehicle. The vision system includes an image processor or image processing system that is operable to receive image data from one or more cameras and to process the captured image data.
Referring now to the drawings and the illustrative embodiments depicted therein, a vehicle imaging system or vision system 10 includes at least one exterior facing imaging sensor or camera, such as a forward facing imaging sensor or camera 12 disposed at the windshield 14 of the vehicle 16 and viewing forwardly through the windshield of the vehicle, with the camera having a lens for focusing images at or onto an imaging array or imaging plane or imager of the camera (
DSRC radio performance is sensitive to the location of the antenna, and utilizing the antenna diversity techniques a high reliable signal reception in multipath environments can be achieved. Mounting the antenna on the roof or front of the vehicle detracts from the appearance of the vehicle. The present invention provides an elegant look while also providing a reliable signal reception.
The range of the DSRC communication link may be degraded as the snow or rain water density increases. For safety critical applications it is very important to have good range during such bad weather conditions. The range may be improved if the system is aware of such situations and the transmission power of the radio is dynamically controlled during such situations.
It is important to have better range toward the rear and rearward of the vehicle especially during highway driving, such that the following vehicles will be aware of the accident or environment condition in advance and limit or prevent multi vehicle pileup accidents. Thus, the present invention may dynamically control the antenna beam depending on the driving situation to provide enhanced range for the given transmission power.
The performance of the DSRC radio may be improved by utilizing multiple antennas installed at various selected diversified locations at the vehicle. A windshield camera 119 is located at the windshield, and one array antenna 115 may be an integral part of the camera mounting bracket 114 (such as shown in
As the range of the DSRC link may be improved by controlling the antenna beam, during the highway driving as the traffic flow in one direction, a concentrated beam 111 could be formed rearward of the vehicle, such as shown in
As the DSRC radio performance degrades such as due to bad weather conditions like snow, rain, fog and/or the like, the system of the present invention may sense the weather condition (such as by utilizing the windshield camera and processing image data captured by the windshield camera) and, responsive to a determination of bad weather conditions that may adversely affect or degrade the radio performance, the transmission power of the DSRC radio may be dynamically adjusted to improve the range during such bad weather conditions, such as shown in the flowchart of
Such vehicle communication systems may provide for communication between vehicles and/or between a vehicle and a remote server. Such car2car or vehicle to vehicle (V2V) and vehicle to infrastructure (car2X or V2X or V2I) technology provides for communication between vehicles and/or infrastructure based on information provided by one or more vehicles and/or information provided by a remote server or the like. The vehicle communication systems may utilize aspects of the systems described in U.S. Pat. Nos. 6,690,268; 6,693,517; 7,156,796 and/or 7,580,795, and/or U.S. Publication Nos. US-2012-0218412, US-2012-0062743, US-2015-0158499; US-2015-0124096 and/or US-2015-0352953, which are all hereby incorporated herein by reference in their entireties.
The main processor may also interface with other modules of the vehicle via a vehicle communication network or bus. The DSRC radio comprises two array antenna modules 103 and 102 for antenna diversity, with one array antenna located in the camera mounting bracket 114 (such as shown in
Therefore, the present invention provides a control or system that adjusts the beam of an antenna array responsive to detection of a driving condition, such as weather conditions at or surrounding the vehicle, highway driving conditions, intersection driving conditions, high or low traffic conditions, parking conditions and/or the like. The driving condition or conditions may be determined via any suitable means, such as responsive to a communication received from a GPS system of the vehicle or a vehicle to infrastructure (V2X) system or such as responsive to processing by an image processor of image data captured by one or more cameras of the vehicle, such as the windshield-mounted camera or such as one or more other exterior viewing cameras of the vehicle (such as a forward viewing front-mounted camera and/or sideward viewing side-mounted cameras and/or a rearward viewing rear-mounted camera of the vehicle that operate as part of a multi-camera surround view vision system of the vehicle).
The system may utilize an omnidirectional beam that covers a generally circular area around the vehicle, and when it is desired to extend the beam further in one direction (such as rearward of the vehicle during highway driving conditions), the system adjusts or controls the beam to provide a directed beam (such as a rearward directed beam) that has an enhanced range in the directed direction as compared to the omnidirectional beam. The system of the present invention thus provides dynamic control or shaping of the antenna beam and range and direction responsive to determined driving conditions at or around the vehicle. Optionally, the system or control may control or dynamically adjust the transmission power of the DSRC radio to improve the range during determined driving conditions, such bad weather conditions, when the transmission power may be degraded.
The camera or sensor may comprise any suitable camera or sensor. Optionally, the camera may comprise a “smart camera” that includes the imaging sensor array and associated circuitry and image processing circuitry and electrical connectors and the like as part of a camera module, such as by utilizing aspects of the vision systems described in International Publication Nos. WO 2013/081984 and/or WO 2013/081985, which are hereby incorporated herein by reference in their entireties.
The system includes an image processor operable to process image data captured by the camera or cameras, such as for detecting objects or other vehicles or pedestrians or the like in the field of view of one or more of the cameras. For example, the image processor may comprise an EYEQ2 or EYEQ3 image processing chip available from Mobileye Vision Technologies Ltd. of Jerusalem, Israel, and may include object detection software (such as the types described in U.S. Pat. Nos. 7,855,755; 7,720,580 and/or 7,038,577, which are hereby incorporated herein by reference in their entireties), and may analyze image data to detect vehicles and/or other objects. Responsive to such image processing, and when an object or other vehicle is detected, the system may generate an alert to the driver of the vehicle and/or may generate an overlay at the displayed image to highlight or enhance display of the detected object or vehicle, in order to enhance the driver's awareness of the detected object or vehicle or hazardous condition during a driving maneuver of the equipped vehicle.
The vehicle may include any type of sensor or sensors, such as imaging sensors or radar sensors or lidar sensors or ladar sensors or ultrasonic sensors or the like. The imaging sensor or camera may capture image data for image processing and may comprise any suitable camera or sensing device, such as, for example, a two dimensional array of a plurality of photosensor elements arranged in at least 640 columns and 480 rows (at least a 640×480 imaging array, such as a megapixel imaging array or the like), with a respective lens focusing images onto respective portions of the array. The photosensor array may comprise a plurality of photosensor elements arranged in a photosensor array having rows and columns. Preferably, the imaging array has at least 300,000 photosensor elements or pixels, more preferably at least 500,000 photosensor elements or pixels and more preferably at least 1 million photosensor elements or pixels. The imaging array may capture color image data, such as via spectral filtering at the array, such as via an RGB (red, green and blue) filter or via a red/red complement filter or such as via an RCC (red, clear, clear) filter or the like. The logic and control circuit of the imaging sensor may function in any known manner, and the image processing and algorithmic processing may comprise any suitable means for processing the images and/or image data.
The camera module and circuit chip or board and imaging sensor may be implemented and operated in connection with various vehicular vision-based systems, and/or may be operable utilizing the principles of such other vehicular systems, such as a vehicle headlamp control system, such as the type disclosed in U.S. Pat. Nos. 5,796,094; 6,097,023; 6,320,176; 6,559,435; 6,831,261; 7,004,606; 7,339,149 and/or 7,526,103, which are all hereby incorporated herein by reference in their entireties, a rain sensor, such as the types disclosed in commonly assigned U.S. Pat. Nos. 6,353,392; 6,313,454; 6,320,176 and/or 7,480,149, which are hereby incorporated herein by reference in their entireties, a vehicle vision system, such as a forwardly, sidewardly or rearwardly directed vehicle vision system utilizing principles disclosed in U.S. Pat. Nos. 5,550,677; 5,670,935; 5,760,962; 5,877,897; 5,949,331; 6,222,447; 6,302,545; 6,396,397; 6,498,620; 6,523,964; 6,611,202; 6,201,642; 6,690,268; 6,717,610; 6,757,109; 6,802,617; 6,806,452; 6,822,563; 6,891,563; 6,946,978 and/or 7,859,565, which are all hereby incorporated herein by reference in their entireties, a traffic sign recognition system, a system for determining a distance to a leading vehicle or object, such as a system utilizing the principles disclosed in U.S. Pat. Nos. 6,396,397 and/or 7,123,168, which are hereby incorporated herein by reference in their entireties, and/or the like.
Optionally, the vision system (utilizing the forward facing camera and a rearward facing camera and other cameras disposed at the vehicle with exterior fields of view) may be part of or may provide a display of a top-down view or birds-eye view system of the vehicle or a surround view at the vehicle, such as by utilizing aspects of the vision systems described in International Publication Nos. WO 2010/099416; WO 2011/028686; WO 2012/075250; WO 2013/019795; WO 2012/075250; WO 2012/145822; WO 2013/081985; WO 2013/086249 and/or WO 2013/109869, and/or U.S. Publication No. US-2012-0162427, which are hereby incorporated herein by reference in their entireties. Optionally, for example, the system may include multiple exterior facing imaging sensors or cameras, such as a forwardly facing camera at the front of the vehicle, and a sidewardly/rearwardly facing camera at respective sides of the vehicle), which capture images exterior of the vehicle.
Changes and modifications in the specifically described embodiments can be carried out without departing from the principles of the invention, which is intended to be limited only by the scope of the appended claims, as interpreted according to the principles of patent law including the doctrine of equivalents.
The present application is a continuation of U.S. patent application Ser. No. 17/247,079, filed Nov. 30, 2020, now U.S. Pat. No. 11,134,220, which is a continuation of U.S. patent application Ser. No. 16/571,644, filed Sep. 16, 2019, now U.S. Pat. No. 10,855,953, which is a continuation of U.S. patent application Ser. No. 15/189,326, filed Jun. 22, 2016, now U.S. Pat. No. 10,419,723, which claims the filing benefits of U.S. provisional application Ser. No. 62/184,546, filed Jun. 25, 2015, which is hereby incorporated herein by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
5550677 | Schofield et al. | Aug 1996 | A |
5670935 | Schofield et al. | Sep 1997 | A |
5854609 | Pyo et al. | Dec 1998 | A |
5949331 | Schofield et al. | Sep 1999 | A |
6434472 | Minowa et al. | Aug 2002 | B1 |
6587186 | Bamji et al. | Jul 2003 | B2 |
6674895 | Rafii et al. | Jan 2004 | B2 |
6678039 | Charbon | Jan 2004 | B2 |
6690268 | Schofield et al. | Feb 2004 | B2 |
6690354 | Sze | Feb 2004 | B2 |
6693517 | McCarthy et al. | Feb 2004 | B2 |
6710770 | Tomasi et al. | Mar 2004 | B2 |
6825455 | Schwarte | Nov 2004 | B1 |
6876775 | Torunoglu | Apr 2005 | B2 |
6906793 | Bamji et al. | Jun 2005 | B2 |
6919549 | Bamji et al. | Jul 2005 | B2 |
7053357 | Schwarte | May 2006 | B2 |
7156796 | Makley | Jan 2007 | B2 |
7157685 | Bamji et al. | Jan 2007 | B2 |
7176438 | Bamji et al. | Feb 2007 | B2 |
7203356 | Gokturk et al. | Apr 2007 | B2 |
7212663 | Tomasi | May 2007 | B2 |
7283213 | O'Connor et al. | Oct 2007 | B2 |
7310431 | Gokturk et al. | Dec 2007 | B2 |
7321111 | Bamji et al. | Jan 2008 | B2 |
7340077 | Gokturk et al. | Mar 2008 | B2 |
7352454 | Bamji et al. | Apr 2008 | B2 |
7375803 | Bamji | May 2008 | B1 |
7379100 | Gokturk et al. | May 2008 | B2 |
7379163 | Rafii et al. | May 2008 | B2 |
7405812 | Bamji | Jul 2008 | B1 |
7408627 | Bamji et al. | Aug 2008 | B2 |
7446650 | Scholfield et al. | Nov 2008 | B2 |
7579939 | Schofield et al. | Aug 2009 | B2 |
7579940 | Schofield et al. | Aug 2009 | B2 |
7580795 | McCarthy et al. | Aug 2009 | B2 |
7583184 | Schofield et al. | Sep 2009 | B2 |
7616781 | Schofield et al. | Nov 2009 | B2 |
7792329 | Schofield et al. | Sep 2010 | B2 |
7873187 | Schofield et al. | Jan 2011 | B2 |
7949152 | Schofield et al. | May 2011 | B2 |
8013780 | Lynam | Sep 2011 | B2 |
8027029 | Lu et al. | Sep 2011 | B2 |
8044776 | Schofield et al. | Oct 2011 | B2 |
8217830 | Lynam | Jul 2012 | B2 |
8427288 | Schofield et al. | Apr 2013 | B2 |
8643724 | Schofield et al. | Feb 2014 | B2 |
8676483 | Miura | Mar 2014 | B2 |
8681218 | Jensen et al. | Mar 2014 | B1 |
8830087 | Tijink et al. | Sep 2014 | B2 |
9008369 | Schofield et al. | Apr 2015 | B2 |
9036026 | Dellantoni et al. | May 2015 | B2 |
9070973 | Hanisch et al. | Jun 2015 | B2 |
9146898 | Ihlenburg et al. | Sep 2015 | B2 |
9214086 | Onishi | Dec 2015 | B1 |
9233645 | Schofield et al. | Jan 2016 | B2 |
9321449 | Johansson et al. | Apr 2016 | B2 |
9428192 | Schofield et al. | Aug 2016 | B2 |
9494093 | Crombez et al. | Nov 2016 | B2 |
9499139 | Koravadi | Nov 2016 | B2 |
9529360 | Melamed et al. | Dec 2016 | B1 |
9575160 | Davis et al. | Feb 2017 | B1 |
9599702 | Bordes et al. | Mar 2017 | B1 |
9609289 | Schofield et al. | Mar 2017 | B2 |
9637053 | Schofield et al. | May 2017 | B2 |
9674490 | Koravadi | Jun 2017 | B2 |
9688199 | Koravadi | Jun 2017 | B2 |
9689967 | Stark et al. | Jun 2017 | B1 |
9707896 | Boegel et al. | Jul 2017 | B2 |
9729636 | Koravadi et al. | Aug 2017 | B2 |
9753121 | Davis et al. | Sep 2017 | B1 |
9843777 | Schofield et al. | Dec 2017 | B2 |
9881220 | Koravadi | Jan 2018 | B2 |
9948904 | Schofield et al. | Apr 2018 | B2 |
10015452 | Schofield et al. | Jul 2018 | B1 |
10059265 | Schofield et al. | Aug 2018 | B2 |
10110860 | Schofield et al. | Oct 2018 | B1 |
10228460 | Jinkins | Mar 2019 | B1 |
10306190 | Schofield et al. | May 2019 | B1 |
10340996 | Wakid | Jul 2019 | B1 |
10389016 | Van Dan Elzen et al. | Aug 2019 | B2 |
10419723 | Koravadi et al. | Sep 2019 | B2 |
10427604 | Schofield et al. | Oct 2019 | B2 |
10462426 | Schofield et al. | Oct 2019 | B2 |
10493899 | Solar | Dec 2019 | B2 |
10855953 | Koravadi et al. | Dec 2020 | B2 |
11134220 | Koravadi et al. | Sep 2021 | B2 |
11187800 | Fersdahl | Nov 2021 | B1 |
11194043 | Meehan | Dec 2021 | B2 |
20020113743 | Judd et al. | Aug 2002 | A1 |
20030064752 | Adachi et al. | Apr 2003 | A1 |
20040145457 | Schofield et al. | Jul 2004 | A1 |
20070109103 | Jedrey et al. | May 2007 | A1 |
20070109146 | Tengler et al. | May 2007 | A1 |
20080082261 | Tengler et al. | Apr 2008 | A1 |
20080160932 | Smith et al. | Jul 2008 | A1 |
20080194204 | Duet et al. | Aug 2008 | A1 |
20080212215 | Schofield et al. | Sep 2008 | A1 |
20090290369 | Schofield et al. | Nov 2009 | A1 |
20100033984 | Sugimoto | Feb 2010 | A1 |
20100045797 | Schofield et al. | Feb 2010 | A1 |
20100182199 | Jeong | Jul 2010 | A1 |
20100245066 | Sarioglu et al. | Sep 2010 | A1 |
20100271274 | Gibson et al. | Oct 2010 | A1 |
20100312446 | Schofield et al. | Dec 2010 | A1 |
20110050489 | Maenpa et al. | Mar 2011 | A1 |
20110093179 | Schofield et al. | Apr 2011 | A1 |
20120062743 | Lynam et al. | Mar 2012 | A1 |
20120062744 | Schofield et al. | Mar 2012 | A1 |
20120218412 | Dellantoni et al. | Aug 2012 | A1 |
20120299373 | Yoshida | Nov 2012 | A1 |
20120302287 | Wright | Nov 2012 | A1 |
20130002470 | Kambe et al. | Jan 2013 | A1 |
20130027218 | Schwarz et al. | Jan 2013 | A1 |
20130222592 | Gieseke | Aug 2013 | A1 |
20130229522 | Schofield et al. | Sep 2013 | A1 |
20140105054 | Sægrov et al. | Apr 2014 | A1 |
20140218529 | Mahmoud et al. | Aug 2014 | A1 |
20140297171 | Minemura et al. | Oct 2014 | A1 |
20140362221 | Schofield et al. | Dec 2014 | A1 |
20140375476 | Johnson et al. | Dec 2014 | A1 |
20150124096 | Koravadi | May 2015 | A1 |
20150131086 | Morishita et al. | May 2015 | A1 |
20150158499 | Koravadi | Jun 2015 | A1 |
20150251599 | Koravadi | Sep 2015 | A1 |
20150348412 | Onishi | Dec 2015 | A1 |
20150352953 | Koravadi | Dec 2015 | A1 |
20160036917 | Koravadi et al. | Feb 2016 | A1 |
20160100084 | Schofield et al. | Apr 2016 | A1 |
20160210853 | Koravadi | Jul 2016 | A1 |
20160223651 | Kamo et al. | Aug 2016 | A1 |
20160267781 | Papay | Sep 2016 | A1 |
20160381326 | Schofield et al. | Dec 2016 | A1 |
20160381571 | Koravadi et al. | Dec 2016 | A1 |
20170043702 | Park et al. | Feb 2017 | A1 |
20170066440 | Koravadi | Mar 2017 | A1 |
20170158133 | Chundrlik, Jr. et al. | Jun 2017 | A1 |
20170161571 | Zhao et al. | Jun 2017 | A1 |
20170174471 | Salmikuukka et al. | Jun 2017 | A1 |
20170222311 | Hess et al. | Aug 2017 | A1 |
20170222852 | Eitan et al. | Aug 2017 | A1 |
20170237946 | Schofield et al. | Aug 2017 | A1 |
20170254873 | Koravadi | Sep 2017 | A1 |
20170276788 | Wodrich | Sep 2017 | A1 |
20170285754 | Holman et al. | Oct 2017 | A1 |
20170292315 | Koravadi | Oct 2017 | A1 |
20170302889 | Koravadi | Oct 2017 | A1 |
20170315231 | Wodrich | Nov 2017 | A1 |
20170320440 | Boegel et al. | Nov 2017 | A1 |
20170334484 | Koravadi | Nov 2017 | A1 |
20170356994 | Wodrich et al. | Dec 2017 | A1 |
20180009374 | Kim et al. | Jan 2018 | A1 |
20180015875 | May et al. | Jan 2018 | A1 |
20180020192 | Schofield et al. | Jan 2018 | A1 |
20180045812 | Hess | Feb 2018 | A1 |
20180082315 | Smid et al. | Mar 2018 | A1 |
20180109764 | Schofield et al. | Apr 2018 | A1 |
20180113461 | Potnis et al. | Apr 2018 | A1 |
20180141563 | Becker | May 2018 | A1 |
20180158337 | Koravadi | Jun 2018 | A1 |
20180167551 | Koravadi | Jun 2018 | A1 |
20180217254 | Hong et al. | Aug 2018 | A1 |
20180217255 | Kim | Aug 2018 | A1 |
20180309963 | Schofield et al. | Oct 2018 | A1 |
20190009721 | Schofield et al. | Jan 2019 | A1 |
20190080606 | Niwa | Mar 2019 | A1 |
20190158790 | Schofield et al. | May 2019 | A1 |
20190273896 | Schofield et al. | Sep 2019 | A1 |
20210149039 | Koike-Akino et al. | May 2021 | A1 |
20210191399 | Verghese | Jun 2021 | A1 |
Number | Date | Country |
---|---|---|
2011090484 | Jul 2011 | WO |
2018007995 | Jan 2018 | WO |
Number | Date | Country | |
---|---|---|---|
20220014714 A1 | Jan 2022 | US |
Number | Date | Country | |
---|---|---|---|
62184546 | Jun 2015 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 17247079 | Nov 2020 | US |
Child | 17448946 | US | |
Parent | 16571644 | Sep 2019 | US |
Child | 17247079 | US | |
Parent | 15189326 | Jun 2016 | US |
Child | 16571644 | US |