The subject disclosure relates to a surround view camera system for object detection and tracking.
Cameras are increasingly used in vehicles (e.g., automobiles, construction equipment, farm equipment, automated manufacturing facilities) for automation and safety systems. Surround-view or rear cameras provide images that facilitate an enhanced view during parking, for example. Forward-looking cameras are used alone or in combination with other sensors (e.g., radar, lidar) to detect and track objects and enable semi-autonomous driving, for example. However, the field of view of the forward-looking camera is insufficient in many scenarios. For example, in a parking lot, in which other vehicles or pedestrians may be approaching from any direction, the forward-looking camera system cannot detect a potential threat of collision. As another example, when an adjacent vehicle changes lanes without allowing sufficient space, that vehicle may not be detected by a forward-looking camera system. Accordingly, it is desirable to provide a surround view camera system for object detection and tracking.
In one exemplary embodiment, a surround view camera system in a vehicle includes two or more cameras arranged respectively at two or more locations of the vehicle. The two or more cameras capture images within a field of view of the two or more cameras. The system also includes a processing system to obtain the images from the two or more cameras and perform image processing to detect and track objects in the field of view of the two or more cameras.
In addition to one or more of the features described herein, the processing system performing image processing includes the processing system pre-processing each of the images individually including de-warping each of the images.
In addition to one or more of the features described herein, the processing system being configured to perform image processing includes the processing system being configured to perform visual recognition techniques to detect the objects in each of the images in which the objects appear.
In addition to one or more of the features described herein, the processing system being configured to perform image processing includes the processing system being configured to perform inter-image detection to detect the objects based on overlapping areas in the images obtained by the two or more cameras.
In addition to one or more of the features described herein, the processing system being configured to perform image processing includes the processing system being configured to perform temporal detection on a frame-by-frame basis to track movement of the objects.
In addition to one or more of the features described herein, the processing system is further configured to obtain vehicle dynamics information about the vehicle.
In addition to one or more of the features described herein, the processing system is further configured to obtain information from other sensors of the vehicle, the other sensors including a radar system, a lidar system, or an ultrasonic sensor system.
In addition to one or more of the features described herein, the processing system is further configured to output information about the locations of the objects in the field of view of the two or more cameras in a vehicle coordinate system.
In addition to one or more of the features described herein, the processing system is further configured to present the objects in the field of view of the two or more cameras in a three-dimensional bounding box (BBOX).
In addition to one or more of the features described herein, the processing system is further configured to provide information about the objects in the field of view of the two or more cameras to a controller in the vehicle, the controller being configured to control safety and autonomous systems of the vehicle.
In another exemplary embodiment, a method of equipping a vehicle to perform object detection and tracking with a surround view camera system includes arranging two or more cameras at respective two or more locations of the vehicle. The two or more cameras capture images within a field of view of the two or more cameras. The method also includes a processing system obtaining the images from the two or more cameras and performing image processing to detect and track objects in the field of view of the two or more cameras.
In addition to one or more of the features described herein, the performing the image processing includes pre-processing each of the images individually, the pre-processing including de-warping each of the images.
In addition to one or more of the features described herein, the performing the image processing includes performing visual recognition techniques to detect the objects in each of the images in which the objects appear.
In addition to one or more of the features described herein, the performing the image processing includes performing inter-image detection to detect the objects based on overlapping areas in the images obtained by the two or more cameras.
In addition to one or more of the features described herein, the performing the image processing includes performing temporal detection on a frame-by-frame basis to track movement of the objects.
In addition to one or more of the features described herein, the performing the image processing includes obtaining vehicle dynamics information about the vehicle.
In addition to one or more of the features described herein, the performing the image processing includes obtaining information from other sensors of the vehicle, the other sensors including a radar system, a lidar system, or an ultrasonic sensor system.
In addition to one or more of the features described herein, the method includes the processing system outputting information about the locations of the objects in the field of view of the two or more cameras in a vehicle coordinate system.
In addition to one or more of the features described herein, the method includes the processing system presenting the objects in the field of view of the two or more cameras in a three-dimensional bounding box (BBOX).
In addition to one or more of the features described herein, the method includes the processing system providing information about the objects in the field of view of the two or more cameras to a controller in the vehicle and the controller controlling safety and autonomous systems of the vehicle.
The above features and advantages, and other features and advantages of the disclosure are readily apparent from the following detailed description when taken in connection with the accompanying drawings.
Other features, advantages and details appear, by way of example only, in the following detailed description, the detailed description referring to the drawings in which:
The following description is merely exemplary in nature and is not intended to limit the present disclosure, its application or uses. It should be understood that throughout the drawings, corresponding reference numerals indicate like or corresponding parts and features.
As previously noted, forward-looking camera systems have been used for object detection. The information obtained about objects in front of the vehicle may be used for adaptive cruise control (ACC), automatic emergency braking (AEB), or forward collision warning (FCW), for example. To address other scenarios and to enhance automated systems, information is desirable about objects in proximity to the vehicle that are not necessarily only in front of the vehicle. While surround view cameras provide images around the vehicle, these camera images have not been used for object detection and tracking. Embodiments of the systems and methods detailed herein relate to a surround view camera system for object detection and tracking. As detailed, the surround view camera system is not simply an extension of the processing used in the forward-looking camera system to multiple cameras disposed around the vehicle. Instead, the multiple views can provide enhanced information that cannot be obtained with a single camera image. For example, images from each of the different views are pre-processed, overlapping images are resolved, and images in the different views are used to filter false alarms or adjust detection thresholds.
The images from the different cameras 140 are sent to the processing system 110 of the surround view camera system 100 for processing. The communication between the cameras 140 and processing system 110 may be over wires that are routed around the vehicle 101 or may be wireless. The processing system 110 may include an application specific integrated circuit (ASIC), an electronic circuit, a processor (shared, dedicated, or group) and memory that executes one or more software or firmware programs, a combinational logic circuit, and/or other suitable components that provide the described functionality. A controller 120 of the vehicle 101 is shown in
The cameras 140 may include extreme wide angle lenses such that the images obtained by the cameras 140 are distorted (i.e., fisheye images). The extreme wide angle lenses have an ultra-wide field of view and, thus, provide images that facilitate 360 degree coverage around the vehicle 101 with the four cameras 140 shown in
In scenario 210b, an object 220, which is another vehicle, cuts into the lane of the vehicle 101. A forward-looking camera system may only see the object 220 when it is in the position shown in
Scenarios 210c and 210d show several objects 220 that are in the FOV 201 at various positions relative to the vehicle 101. The forward-looking camera system would only detect some of the objects 220 shown within the FOV 201 of the surround view camera system 100. As further discussed with reference to
Obtaining vehicle information, at block 330, includes obtaining motion information, for example, what can aid in tracking of objects 220. Exemplary vehicle information includes speed, angle of motion, acceleration, or information from the global positioning system (GPS) receiver. This information may be provided to the processing system 110 through the controller 120 or directly from other vehicle systems that obtain information about vehicle dynamics. According to alternate or additional embodiments, the vehicle information obtained at block 330 can also include data from other sensors 130 (e.g., radar, lidar) mounted on the vehicle 101.
At block 340, operations are performed to detect and track objects 220 based on the images obtained by the cameras 140. These operations include known image processing, computer vision, and machine learning operations and may be performed by a deep learning neural network, for example. Known algorithms and processes that may be used as part of block 340 include a deep learning method, for example, a deep convolution neural network (DCNN), or other computer vision methods, such as deformable part models (DPM), along with other visual recognition techniques. The processing at block 340 facilitates organizing and outputting detection and tracking information at block 350.
The processing at block 340 includes performing individual frame detection at block 343. This process may use the known DPM algorithm, for example, to perform detection of objects 220 within each of the individual frames obtained by each of the cameras 140. Performing inter-image detection, at block 345, is also part of the processing at block 340. The inter-image detection operation involves associating and matching objects 220 that are captured by more than one camera 140 of the surround view camera system 100. Essentially, the position of an object 220 can be triangulated based on the images from two or more cameras 140. The process facilitates resolving overlapping images by filtering false alarms or adjusting detection thresholds, for example.
According to the exemplary arrangement shown in
Performing temporal detection, at block 347, is also part of the processing at block 340. The position of an object 220 that is detected (according to block 343 or, additionally, 345) is tracked in time based on its location from one frame to the next. While the temporal tracking (at block 347) relies on detection at block 343 or 345, the temporal tracking (at block 347) may enhance the detection at block 343 or 345, as well. For example, an object 220 that may otherwise be dismissed as a false alarm may instead be determined to have moved out of an overlapping area of coverage of two cameras 140 based on the temporal detection at block 347. The temporal detection (at block 347) facilitates determining the movement of an object 220 relative to the vehicle 101. For example, a determination of whether an object 220 is moving toward or away from the vehicle 101 can affect information provided to other vehicle systems (e.g., ACC, AEB) through the controller 120. That is, an object 220 moving away from the vehicle 101 may not be used to trigger the AEB system while an object 220 moving toward the vehicle 101 may trigger the AEB system.
As the discussion indicates and as shown in
The output 510a shows a top-down view that shows the vehicle 101 and five different objects 220 around the vehicle 101. The angle of each object 220 relative to the vehicle 101 is shown and indicates the direction of travel of each object 220. The output 510b also shows a top-down view of the vehicle 101 and five objects 220 around the vehicle 101. The objects 220 may be color-coded or coded by pattern, as shown in
While the above disclosure has been described with reference to exemplary embodiments, it will be understood by those skilled in the art that various changes may be made and equivalents may be substituted for elements thereof without departing from its scope. In addition, many modifications may be made to adapt a particular situation or material to the teachings of the disclosure without departing from the essential scope thereof. Therefore, it is intended that the present disclosure not be limited to the particular embodiments disclosed, but will include all embodiments falling within the scope thereof.
This application claims the benefit of priority of U.S. Provisional Application No. 62/324,602 filed Apr. 19, 2016, the disclosure of which is incorporated herein by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
62324602 | Apr 2016 | US |