The subject disclosure relates to radar system control to perform cross-traffic management in a vehicle with a trailer.
Vehicles (e.g., automobiles, trucks, motor cycles, construction equipment, farm equipment, automated factory equipment) include an ever-increasing array of sensors. These sensors obtain information about the vehicle and its environment and facilitate autonomous implementation of some or all operations of the vehicles or alerts to the driver. Exemplary sensors include a radar system, a lidar system, and a camera. Exemplary semi-autonomous systems in vehicles may perform collision avoidance, adaptive cruise control, or automatic braking. Another exemplary system in a vehicle is a cross-traffic alert and/or automated braking system. When the vehicle is reversing, the cross-traffic alert and/or automated braking system warns the driver of other vehicles entering the reverse path (i.e., cross-traffic) and may additionally automatically apply the brake to avoid a collision with another vehicle. This type of system generally relies on a radar system to detect the cross-traffic. When the vehicle has a trailer behind it, the radar system function may be negatively affected. This is because the trailer, which is relatively close to the radar system and has a high radar signature, would produce high amplitude reflections that increase the detection floor and, thus, obscure other, farther objects. Accordingly, it is desirable to provide radar system control to perform cross-traffic management in a vehicle with a trailer.
In one exemplary embodiment, a method for controlling a radar system to perform cross-traffic management in a vehicle with a trailer includes determining a location of the trailer behind the vehicle, and adjusting an initial field of view of the radar system to a modified field of view that excludes the location of the trailer. The method also includes detecting one or more other vehicles with the radar system having the modified field of view, and implementing a cross-traffic alert based on the detecting the one or more other vehicles.
In addition to one or more of the features described herein, the method also includes obtaining dimensions of the trailer.
In addition to one or more of the features described herein, the method also includes obtaining an articulation angle, the articulation angle being an angle between a center axial line of the vehicle and a center axial line of the trailer.
In addition to one or more of the features described herein, the obtaining the articulation angle includes performing image processing on an image obtained by a camera with a field of view that includes the trailer.
In addition to one or more of the features described herein, the determining the location of the trailer behind the vehicle is based on the dimensions of the trailer and on the articulation angle.
In addition to one or more of the features described herein, the method also includes determining an overlap region between the initial field of view and the location of the trailer.
In addition to one or more of the features described herein, the adjusting the initial field of view includes excluding the overlap region in the modified field of view based on turning off a subset of antenna elements or on performing beamforming.
In addition to one or more of the features described herein, the method also includes filtering reflections received as a result of transmission by the radar system in the modified field of view to eliminate any of the reflections that result from the trailer.
In addition to one or more of the features described herein, the method also includes determining, using the controller, whether the vehicle is in reverse gear, wherein implementing the method is based on determining that the vehicle is in the reverse gear.
In addition to one or more of the features described herein, the method also includes computing a time to collision based on the detecting the one or more other vehicles in a path of the vehicle and providing an alert or implementing automatic braking based on the time to collision being less than a threshold value.
In another exemplary embodiment, a system to perform cross-traffic management in a vehicle with a trailer includes a radar system with a modified field of view. The radar system detects one or more other vehicles. The system also includes a controller to determine a location of the trailer behind the vehicle, to adjust an initial field of view of the radar system to the modified field of view that excludes the location of the trailer, and to implement a cross-traffic alert based on the radar system detecting the one or more other vehicles.
In addition to one or more of the features described herein, the controller obtains dimensions of the trailer.
In addition to one or more of the features described herein, the controller obtains an articulation angle. The articulation angle is an angle between a center axial line of the vehicle and a center axial line of the trailer.
In addition to one or more of the features described herein, the system also includes a camera. The controller obtains the articulation angle by performing image processing on an image obtained by the camera with a field of view that includes the trailer.
In addition to one or more of the features described herein, the controller determines the location of the trailer behind the vehicle based on the dimensions of the trailer and on the articulation angle.
In addition to one or more of the features described herein, the controller determines an overlap region between the initial field of view and the location of the trailer.
In addition to one or more of the features described herein, the controller excludes the overlap region to adjust the initial field of view to the modified field of view. The radar system implements the modified field of view based on turning off a subset of antenna elements or on performing beamforming.
In addition to one or more of the features described herein, the controller filters reflections received as a result of transmission by the radar system in the modified field of view to eliminate any of the reflections that result from the trailer.
In addition to one or more of the features described herein, the controller determines whether the vehicle is in reverse gear and implements the cross-traffic alert based on determining that the vehicle is in the reverse gear.
In addition to one or more of the features described herein, the controller computes a time to collision based on the one or more other vehicles being in a path of the vehicle and implements an alert or automatic braking based on the time to collision being less than a threshold value.
The above features and advantages, and other features and advantages of the disclosure are readily apparent from the following detailed description when taken in connection with the accompanying drawings.
Other features, advantages and details appear, by way of example only, in the following detailed description, the detailed description referring to the drawings in which:
The following description is merely exemplary in nature and is not intended to limit the present disclosure, its application or uses. It should be understood that throughout the drawings, corresponding reference numerals indicate like or corresponding parts and features.
As previously noted, many vehicle operations may be aided or performed autonomously based on sensor information. A cross-traffic alert and/or automated braking system uses radar information about cross-traffic behind the vehicle during a reversing maneuver. As previously noted, a trailer attached to the back of the vehicle can negatively affect the radar system results. This is because the trailer obscures the field of view (FOV) of the radar system. Thus, prior cross-traffic alert and/or automated braking systems are disabled when a trailer is attached to the vehicle.
Embodiments of the systems and methods detailed herein relate to radar system control to perform cross-traffic management in a vehicle with a trailer. The radar system control according to one or more embodiments is facilitated by information about the dimensions of the trailer and about the articulation angle of the trailer relative to the vehicle. This information is obtained using one or more cameras. Existing vehicles may use this information to augment a rear-view image displayed to the driver in order to make the trailer invisible in the image. According to one or more embodiments, the dimension and articulation angle information is used to adjust the FOV of one or more rear radar systems and to filter any unwanted reflections from the trailer.
In accordance with an exemplary embodiment,
The vehicle 100 also includes a controller 120, which may instead be a collection of two or more controllers 120 that communicate with each other to perform the functionality discussed herein. The controller 120 may interact with an infotainment system 125 that provides alerts and other output to an operator (e.g., driver) of the vehicle 100 and may also obtain inputs from the operator. The inputs include dimensions of the trailer 105. The processing of reflections received by the radar systems 110 and the control of the radar systems 110 may be performed by the radar systems 110 themselves, by the controller 120, or by a combination of the two. Information (e.g., articulation angle AA) from one or more cameras 130 and information from an operator (e.g., dimensions of trailer 105) may be obtained by the controller 120 to facilitate the control of the radar systems 110, as further discussed with reference to
The controller 120 includes processing circuitry that may include an application specific integrated circuit (ASIC), an electronic circuit, a processor (shared, dedicated, or group) and memory that executes one or more software or firmware programs, a combinational logic circuit, and/or other suitable components that provide the described functionality. For example, one or more memory devices of the controller 120 may store received information, such as the dimensions of the trailer 105, and one or more processors of the controller 120 may execute stored instructions to perform the functionality detailed herein.
The radar control operations are discussed with reference to blocks 340 and 350. As indicated, the processing at block 350 may be optional. At block 340, adjusting the field of view of one or more radar systems 110 includes determining an overlap between the default field of view of each radar system 110 and the location of the trailer 105. The field of view is then modified to ensure that the trailer 105 is not within the field of view. Modifying the field of view may be achieved in a number of ways based on the specific type of radar system 110 being used. According to an exemplary embodiment, modifying the field of view includes turning off (i.e., not transmitting from) antenna elements whose main beam includes the trailer 105. According to another exemplary embodiment, modifying the field of view includes limiting a beamforming angle to exclude an area with the trailer 105. The particular implementation used to modify the field of view is not intended to be limited by the examples. Any known mechanism to modify the field of view to exclude an area with the trailer 105 is within the contemplated scope.
At block 350, filtering reflections resulting from transmissions by one or more radar systems 110 is optional, as indicated. As previously noted, one or more cameras 130 may be used to obtain rear-facing images. Processing these images (e.g., by the controller 120) facilitates determining the articulation angle AA and, in conjunction with the dimensions of the trailer 105 (obtained at block 310), the location of the trailer 105 (at block 330). The processes 320, 330 may be performed periodically while the reverse gear is engaged. Adjustment of the field of view, at block 340, may also be performed periodically. However, there may be a delay (e.g., a few seconds) between determination of the location of the trailer 105 (at block 330) and adjustment of the field of view of one or more radar systems 110 according to the location of the trailer 105 (at block 340). Thus, a part of the trailer 105 may move into the modified field of view of one or more of the radar systems 110. By the time the location of the trailer 105 is updated (at block 330), one or more radar systems 110 may have emitted transmissions within the previously modified field of view (at block 340).
That is, there may be a chance of at least a portion of the trailer 105 being in the field of one or more radar systems 110 despite the modification of the field of view as part of the processing at block 340. This will result in a high-amplitude reflected energy from the portion of the trailer 105 that is encountered by the emitted transmission. Filtering the reflections to address this scenario, at block 350, may be implemented in a number of ways. According to an exemplary embodiment, the high amplitude of the reflected energy may be used to associate the reflection with the trailer 105. According to another exemplary embodiment, the range of the reflected energy additionally or alternately with the known (possibly inaccurate) location of the trailer 105 (from block 330) may be used to associate the reflected energy with the trailer 105. Once identified, the reflected energy associated with the trailer 105 may then be filtered (e.g., zeroed out) to avoid artificially increasing the detection threshold according to a constant false alarm rate (CFAR) scheme, for example. As previously noted, the processing of the reflections resulting from transmissions by one or more radar systems 110 may be performed by the radar systems 110, the controller 120, or both.
At block 360, implementing cross-traffic alert and/or braking is not modified from existing systems and, therefore, is only generally described herein. Once the process at block 340 and, optionally, at block 350 has been performed, the cross-traffic management may be handled in the same way as when the trailer 105 is not present. However, the position of the trailer 105 must be considered for purposes of determining the potential for a collision. Specifically, reflections received at one or more radar systems 110 may be processed by the radar systems 110, controller 120, or both to identify any cross-traffic (e.g., the truck 210 shown in
While the above disclosure has been described with reference to exemplary embodiments, it will be understood by those skilled in the art that various changes may be made and equivalents may be substituted for elements thereof without departing from its scope. In addition, many modifications may be made to adapt a particular situation or material to the teachings of the disclosure without departing from the essential scope thereof. Therefore, it is intended that the present disclosure not be limited to the particular embodiments disclosed, but will include all embodiments falling within the scope thereof
Number | Name | Date | Kind |
---|---|---|---|
4431995 | Barton | Feb 1984 | A |
5023617 | Deering | Jun 1991 | A |
5675343 | Champeau | Oct 1997 | A |
6670910 | Delcheccolo | Dec 2003 | B2 |
8432309 | MacDonald | Apr 2013 | B2 |
9156496 | Greenwood | Oct 2015 | B2 |
9566911 | Greenwood | Feb 2017 | B2 |
9854209 | Aich | Dec 2017 | B2 |
9971943 | Greenwood | May 2018 | B2 |
10046800 | Hu | Aug 2018 | B2 |
10276049 | Prasad | Apr 2019 | B2 |
10351146 | Prasad | Jul 2019 | B2 |
10481255 | Prasad et al. | Nov 2019 | B2 |
10609340 | Aich | Mar 2020 | B2 |
10638025 | Gali | Apr 2020 | B2 |
10807639 | Hu | Oct 2020 | B2 |
10836376 | Wodrich | Nov 2020 | B2 |
11086333 | Balogh | Aug 2021 | B2 |
11127301 | Lai | Sep 2021 | B1 |
20020147534 | Delcheccolo | Oct 2002 | A1 |
20030055596 | Studt | Mar 2003 | A1 |
20080231701 | Greenwood | Sep 2008 | A1 |
20120133547 | MacDonald | May 2012 | A1 |
20120200706 | Greenwood | Aug 2012 | A1 |
20130226390 | Luo | Aug 2013 | A1 |
20140267688 | Aich | Sep 2014 | A1 |
20160001705 | Greenwood | Jan 2016 | A1 |
20160041258 | Cashler | Feb 2016 | A1 |
20160170412 | Yamamoto | Jun 2016 | A1 |
20170083773 | Greenwood | Mar 2017 | A1 |
20170363727 | Prasad | Dec 2017 | A1 |
20170363728 | Prasad | Dec 2017 | A1 |
20180043933 | Hu | Feb 2018 | A1 |
20180061239 | Prasad | Mar 2018 | A1 |
20180109762 | Aich | Apr 2018 | A1 |
20180220034 | Katayama | Aug 2018 | A1 |
20180319437 | Hu | Nov 2018 | A1 |
20190170867 | Wang | Jun 2019 | A1 |
20190220034 | Balogh | Jul 2019 | A1 |
20190297233 | Gali | Sep 2019 | A1 |
20200259976 | Gali | Aug 2020 | A1 |
20200341118 | Chen | Oct 2020 | A1 |
20210055373 | Smith | Feb 2021 | A1 |
20210055375 | Smith | Feb 2021 | A1 |
20210061261 | Wodrich | Mar 2021 | A1 |
20210287548 | Lai | Sep 2021 | A1 |
20210325897 | Balogh | Oct 2021 | A1 |
Entry |
---|
Chinese Office Action for CN Application No. 202011201658.0 dated Jun. 2, 2022, pp. 1-9. |
Number | Date | Country | |
---|---|---|---|
20210149044 A1 | May 2021 | US |