The present invention generally relates to trailer backup assist systems, and more particularly, to trailer backup assist systems employing hitch angle detection through image processing.
Reversing a vehicle while towing a trailer can be challenging for many drivers, particularly for drivers that drive with a trailer on an infrequent basis or with various types of trailers. Some systems used to assist a driver in backing a trailer rely on hitch angle measurements to determine the position of the trailer relative to the vehicle. Thus, the accuracy and reliability of the hitch angle measurements can be critical to the operation of the trailer backup assist system.
According to one aspect of the present invention, a trailer type identification system is provided. The system includes an imaging device for capturing images of a trailer connected to a vehicle, and a controller for analyzing the captured images. The controller identifies vehicle and trailer contours, predicts that the trailer is a fifth wheel trailer if unable to detect a connection between the identified vehicle and trailer contours, and validates the prediction if the identified trailer contour exhibits motion during a vehicle turn event.
According to another aspect of the present invention, a trailer type identification system is provided. The system includes an imaging device for capturing images of a trailer connected to a vehicle, and a controller for analyzing the captured. The controller identifies vehicle and trailer contours, predicts a trailer type based on detection of a connection between the identified vehicle and trailer contours, and validates the prediction if the identified trailer contour exhibits motion during a vehicle turn event.
According to yet another aspect of the present invention, a trailer type identification system is provided. The system includes an imaging device for capturing images of a trailer connected to a vehicle, and a controller for analyzing the captured images. The controller detects vehicle and trailer contours, predicts a trailer type based on detection of a connection between the identified vehicle and trailer contours, and validates the prediction if the detected trailer contour exhibits an expected motion during a vehicle turn event.
These and other features, advantages, and objects of the present invention will be further understood and appreciated by those skilled in the art by reference to the following specification, claims, and appended drawings.
In the drawings:
For purposes of description herein, it is to be understood that the disclosed trailer backup assist system and the related methods may assume various alternative embodiments and orientations, except where expressly specified to the contrary. It is also to be understood that the specific devices and processes illustrated in the attached drawings, and described in the following specification, are simply exemplary embodiments of the inventive concepts defined in the appended claims. While various aspects of the trailer backup assist system and the related methods are described with reference to a particular illustrative embodiment, the disclosed invention is not limited to such embodiments, and additional modifications, applications, and embodiments may be implemented without departing from the disclosed invention. Hence, specific dimensions and other physical characteristics relating to the embodiments disclosed herein are not to be considered as limiting, unless the claims expressly state otherwise.
As used herein, the term “and/or,” when used in a list of two or more items, means that any one of the listed items can be employed by itself, or any combination of two or more of the listed items can be employed. For example, if a composition is described as containing components A, B, and/or C, the composition can contain A alone; B alone; C alone; A and B in combination; A and C in combination; B and C in combination; or A, B, and C in combination.
Referring to
The trailer backup assist system 10 also includes an imaging device 34 located at the rear of the vehicle 14 and configured to image a rear-vehicle scene. The imaging device 34 may be centrally located at an upper region of a vehicle tailgate 35 such that the imaging device 34 is elevated relative to the tongue 26 of the trailer 12. The imaging device 34 has a field of view 36 located and oriented to capture one or more images that may include the tongue 26 of the trailer 12 and the hitch ball 30, among other things. Captured images are supplied to a controller 38 of the trailer backup assist system 10 and are processed by the controller 38 to determine the hitch angle between the vehicle 14 and the trailer 12, as will be described in greater detail herein. The controller 38 is configured with a microprocessor 40 and/or other analog and/or digital circuitry for processing one or more logic routines stored in a memory 42. The logic routines may include one or more hitch angle detection routines 44 and operating routines 46. Information from the imaging device 34 or other components of the trailer backup assist system 10 can be supplied to the controller 38 via a communication network of the vehicle 14, which can include a controller area network (CAN), a local interconnect network (LIN), or other conventional protocols used in the automotive industry. It should be appreciated that the controller 38 may be a stand-alone dedicated controller or may be a shared controller integrated with the imaging device 34 or other component of the trailer backup assist system 10 in addition to any other conceivable onboard or off-board vehicle control systems.
With respect to the present embodiment, the controller 38 of trailer backup assist system 10 may be configured to communicate with a variety of vehicle equipment. The trailer backup assist system 10 may include a vehicle sensor module 48 that monitors certain dynamics of the vehicle 14. The vehicle sensor module 48 may generate a plurality of signals that are communicated to the controller 38 and may include a vehicle speed signal generated by a speed sensor 50 and a vehicle yaw rate signal generated by a yaw rate sensor 52. A steering input device 54 may be provided to enable a driver to control or otherwise modify the desired curvature of the backing path of the trailer 12. The steering input device 54 may be communicatively coupled to the controller 38 in a wired or wireless manner and provides the controller 38 with information defining the desired curvature of the backing path of the trailer 12. In response, the controller 38 processes the information and generates corresponding steering commands that are supplied to a power assist steering system 56 of the vehicle 14. In one embodiment, the steering input device 54 includes a rotatable knob 58 operable between a number of rotated positions that each provide an incremental change to the desired curvature of the backing path of the trailer 12.
According to one embodiment, the controller 38 of the trailer backup assist system 10 may control the power assist steering system 56 of the vehicle 14 to operate the steered wheels 60 of the vehicle 14 for moving the vehicle 14 in such a manner that the trailer 12 reacts in accordance with the desired curvature of the backing path of the trailer 12. The power assist steering system 56 may be an electric power-assisted steering (EPAS) system that includes an electric steering motor 62 for turning the steered wheels 60 to a steering angle based on a steering command generated by the controller 38, whereby the steering angle may be sensed by a steering angle sensor 64 of the power assist steering system 56 and provided to the controller 38. The steering command may be provided for autonomously steering the vehicle 14 during a backup maneuver and may alternatively be provided manually via a rotational position (e.g., a steering wheel angle) of a steering wheel 66 or the rotatable knob 58. However, in some embodiments, the steering wheel 66 of the vehicle 14 may be mechanically coupled with the steered wheels 60 of the vehicle 14, such that the steering wheel 66 moves in concert with steered wheels 60 via an internal torque, thereby preventing manual intervention with the steering wheel 66 during autonomous steering of the vehicle 14. In such instances, the power assist steering system 56 may include a torque sensor 68 that senses torque (e.g., gripping and/or turning) on the steering wheel 66 that is not expected from autonomous control of the steering wheel 66 and therefore indicative of manual intervention by the driver. In some embodiments, external torque applied to the steering wheel 66 may serve as a signal to the controller 38 that the driver has taken manual control and for the trailer backup assist system 10 to discontinue autonomous steering functionality.
The controller 38 of the trailer backup assist system 10 may also communicate with a vehicle brake control system 70 of the vehicle 14 to receive vehicle speed information such as individual wheel speeds of the vehicle 14. Additionally or alternatively, vehicle speed information may be provided to the controller 38 by a powertrain control system 72 and/or the speed sensor 50, among other conceivable means. It is conceivable that individual wheel speeds may be used to determine a vehicle yaw rate, which can be provided to the controller 38 in the alternative, or in addition to, the vehicle yaw rate measured by yaw rate sensor 52 of the vehicle sensor module 48. In some embodiments, the controller 38 may provide braking commands to the vehicle brake control system 70, thereby allowing the trailer backup assist system 10 to regulate the speed of the vehicle 14 during a backup maneuver of the trailer 12. It should be appreciated that the controller 38 may additionally or alternatively regulate the speed of the vehicle 14 via interaction with the powertrain control system 72.
Through interaction with the power assist steering system 56, the vehicle brake control system 70, and/or the powertrain control system 72 of the vehicle 14, the potential for unacceptable trailer backup conditions can be reduced. Examples of unacceptable trailer backup conditions include, but are not limited to, a vehicle over-speed condition, a high hitch angle rate, hitch angle dynamic instability, a trailer jackknife condition, sensor failure, and the like. In such circumstances, the driver may be unaware of the failure until the unacceptable trailer backup condition is imminent or already happening. Therefore, it is disclosed herein that the controller 38 of the trailer backup assist system 10 can generate an alert signal corresponding to a notification of an actual, impending, and/or anticipated unacceptable trailer backup condition, and prior to driver intervention, generate a counter measure to prevent such an unacceptable trailer backup condition.
According to one embodiment, the controller 38 may communicate with one or more devices, including a vehicle alert system 74, which may prompt visual, auditory, and tactile warnings. For instance, vehicle brake lights 76 and vehicle emergency flashers may provide a visual alert and a vehicle horn 78 and/or speaker 80 may provide an audible alert. Additionally, the controller 38 and/or vehicle alert system 74 may communicate with a human machine interface (HMI) 82 of the vehicle 14. The HMI 82 may include a touchscreen vehicle display 84 such as a center-stack mounted navigation or entertainment display capable of displaying images indicating the alert. Such an embodiment may be desirable to notify the driver of the vehicle 14 that an unacceptable trailer backup condition is afoot. Further, it is contemplated that the controller 38 may communicate via wireless communication with one or more electronic portable devices such as portable electronic device 86, which is embodied as a smartphone. The portable electronic device 86 may include a display 88 for displaying one or more images and other information to a user. In response, the portable electronic device 86 may provide feedback information, such as visual, audible, and tactile alerts.
Referring to
For purposes of illustration, a captured image 90 is exemplarily shown in
The template matching method may begin at step 100, where the driver or other occupant initiates the trailer backup assist system 10. This may be achieved via user-input made through the display 84 of the vehicle 14 or other conceivable means. At step 110, the driver is instructed to pull the trailer 12 in a straight direction such that the hitch angle between the vehicle 14 and the trailer 12 is substantially zero. While the vehicle 14 and trailer 12 are engaged in the straight pull maneuver, the controller 38 derives an averaged image of all images captured by the imaging device 34 during a period of time at step 120. It has been discovered that 1-3 seconds typically suffices. Notably, the trailer 12 appears stationary within the images captured by the imaging device 34 whereas ground noise constantly changes from image to image. Thus, with respect to the averaged image, pixels associated with the trailer 12 will keep their contrast whereas pixels associated with ground noise will be blurred. To illustrate this effect, an averaged image 125 is exemplarily shown in
At step 130, the controller 38 derives an edge map of the averaged image by calculating the intensity gradient for each pixel of the averaged image 125. The intensity gradient, or edge value, of each pixel may range from 0 to 255. For purposes of illustration, an edge map 135 is exemplarily shown in
According to one embodiment, as shown in
According to one embodiment, the imaged hitch point 172 may be determined through process of elimination. For instance, as exemplarily shown in
While matching the template image 158 to the search image 171, the controller 38 may additionally determine the presence of an imminent jackknife scenario at step 180. With continued reference to
In the event the template image 158 crosses into the proximity zone 182 of the search image 171, the controller 38 determines that an imminent jackknife scenario is present and initiates a jackknife countermeasure at step 190. Otherwise, if it is determined that an imminent jackknife scenario is not present, the controller 38 may continue to determine the hitch angle between the vehicle 14 and the trailer 12, as discussed previously with respect to step 170. The jackknife countermeasure may include generating an auditory warning via the vehicle alert system 74, generating a visual warning via the display 84, generating a braking command to the vehicle brake control system 70, reducing the torque of the powertrain control system 72, modifying the steering angle of the vehicle 14, or a combination thereof in addition to any other conceivable countermeasures. Since the trailer 12 will likely be in motion upon the controller 38 determining that an imminent jackknife scenario is present, it is generally desirable to locate and dimension the proximity zone 182 in a manner that provides sufficient time for a jackknife scenario to be detected and a countermeasure to be implemented, thereby minimizing the potential of an actual jackknifing and/or collision between the trailer 12 and the vehicle 14. Doing so also overcomes any response latency that may be inherent in the trailer backup assist system 10. While steps 170 and 180 have been illustrated in a linear fashion, it should be appreciated that both steps may be performed simultaneously.
Referring to
The centerline method may begin at step 200, where the controller 38 processes successive images captured by the imaging device 34 to derive an averaged image. At step 210, the controller 38 derives an edge map by calculating an intensity gradient, or edge value, for each pixel in the averaged image. At step 220, the controller 38 identifies trailer pixels in the edge map by comparing the edge value of each pixel to a threshold value and selecting only those pixels meeting or exceeding the threshold value to correspond to trailer pixels. For purposes of illustration, an edge map 222 is exemplarily shown in
At step 240, the controller 38 selects the candidate line (e.g., candidate line 242) having approximately the same number of trailer pixels on each of its sides, or said differently, the candidate line, or centerline, about which the trailer pixels are substantially symmetric. Once the controller 38 has made a candidate line selection, the controller 38 may determine the hitch angle between the vehicle 14 and the trailer 12 based on the angular position of the selected candidate line 242 relative to the reference line 234 at step 250. More specifically, the angle θ between the selected candidate line 242 and the reference line 234 can be correlated to the hitch angle between the vehicle 14 and the trailer 12. As the vehicle 14 and trailer 12 continue along its course, steps 200-250 may be iterated with subsequent images captured by the imaging device 34 to continually provide hitch angle measurements.
Referring to
The drawbar scan method generally begins at step 300, where the controller 38 selects two images captured by the imaging device 34 that show the trailer 12 at distinct hitch angles. The two images may be successive or non-successive depending on the frame rate of the imaging device 34. In practice, a noticeable difference in hitch angles between the two images is generally preferred. At step 310, the controller 38 derives an edge map for both images by calculating the intensity gradient, or edge value, for each of their corresponding pixels. For purposes of illustration,
At step 330, the controller 38 determines an average pixel intensity of the pixels bounded by each detection window 322 when centered at the current pixel position, and at step 340, the controller 38 calculates an averaged pixel difference, which is defined herein as the absolute value of the difference between the average pixel intensities, as calculated at step 330. Once the average pixel difference has been calculated, at step 350, the controller 38 shifts each detection window 322 in an upward vertical direction (as specified by arrow 342) so that each detection window 322 is commonly centered at a new pixel position that is 1 or more pixel positions higher on the corresponding reference line 326 than the previous pixel position. Thus, by making multiple iterations of steps 330-350, the controller 38 may calculate averaged pixel differences for when each detection window 322 is commonly centered at a number of pixel positions along the reference line 326. Once this has been done, at step 360, the controller 38 extrapolates the location of the hitch point 328 based on variations in the calculated averaged pixel differences.
With respect to images 312 and 314, it is generally expected that little variation will occur between the calculated averaged pixel differences associated with pixel positions that coincide with the imaged drawbar 327 due in part to the imaged drawbar 327 appearing in a common fixed position in both images 312, 314. In contrast, it is generally expected that greater variation to occur between the calculated averaged pixel differences associated with pixel positions that are located on portions of the reference line 326 that extend beyond the imaged drawbar 327 due in part to the trailer 12 appearing in different positions in both images 312, 314. For purposes of illustration, a graph is shown in
As shown in
Referring to
δ: steering angle at steered wheels 60 of the vehicle 14;
α: yaw angle of the vehicle 14;
β: yaw angle of the trailer 12;
γ: hitch angle between the vehicle 14 and the trailer 12 (γ=β−α);
W: wheelbase length between a front axle 370 and a rear axle 372 of the vehicle 14;
L: drawbar length between the hitch point 32 and the rear axle 372 of the vehicle 14;
D: trailer length between the hitch point 32 and axle 20 of the trailer 12 or effective axle for a multiple axle trailer; and
ν: vehicle longitudinal speed.
From the kinematic model shown in
Furthermore, the yaw rate of the trailer 12 may be represented with the following equation:
Accordingly, when the yaw rate of the vehicle 14 and the trailer 12 become equal, the hitch angle γ and the steering angle δ will be constant. This condition, referred to herein as “steady state”, can occur when a steering command is steadily maintained during a backing maneuver such as when the trailer 12 is reversed in a straight line with the vehicle 14 or when the vehicle 14 and trailer 12 are turning at a constant curvature for at least a threshold period of time or over a threshold distance of motion. Under such steady state driving conditions, the resulting hitch angle γ can be described using the following equation:
c=a cos γ+b sin γ
This equation can be rewritten as follows:
c=a√{square root over (1−sin2 γ)}+b sin γ
The above equation can be rearranged into quadratic form and rewritten as follows:
c2−a2−2bc sin γ+(b2+a2)sin γ=0
Solving the quadratic equation for the hitch angle γ yields the following hitch angle equation:
Accordingly, for a particular vehicle and trailer combination, the trailer length D, the wheelbase length W, and the drawbar length L are constant and assumed known. Thus, when the steady state condition is satisfied, the hitch angle γ between the vehicle 14 and trailer 12 may be determined as a function of the trailer length D, the wheelbase length W, the drawbar length L, and the steering angle δ.
Referring to
Referring to
The hitch angle initialization method may begin at step 600, where the controller 38 determines whether a template image is available for the trailer 12 being towed. If so, the controller 38 proceeds to step 610 to determine the hitch angle via the template matching method. The template matching method may determine the hitch angle in approximately 1 second and is generally the most reliable when compared to the centerline method and the steady state method. So long as the template image remains available, the template matching method is selected as the best available hitch angle detection. In the event no template image is available or the template matching method is unable to be executed (e.g., system error), the controller 38 proceeds to step 620 to determine whether the vehicle 14 and trailer 12 are moving in a straight direction. According to one embodiment, the direction of the vehicle 14 and trailer 12 may be determined by obtaining a steering angle from the steering angle sensor 64 over a period of time. If it is determined that the vehicle 14 and trailer 12 are moving in a straight direction, the controller 38 proceeds to step 625 and processes images captured by the imaging device 34 to derive a template image of the trailer 12 before proceeding to step 610 to determine the hitch angle via the template matching method. Otherwise, the controller 38 proceeds to step 630 to determine the hitch angle via the centerline method. The centerline method may determine the hitch angle in less than 1 second but is generally less reliable when compared to the template matching method and the steady state method.
Once the centerline method is selected, the controller 38 will continue to determine the hitch angle via the centerline method until a steady state condition is satisfied at step 640. As described previously herein, the steady state condition may be satisfied when the vehicle 14 and trailer 12 are moving in a straight direction or moving along a path at constant curvature. Or in other words, the steady state condition is satisfied when the yaw rate of the vehicle 14 and the trailer 12 become equal, thereby resulting in the hitch angle and the steering angle becoming constant. If the steady state condition is satisfied, the controller proceeds to step 650, where it determines whether the hitch angle is substantially zero. In instances where the steady state condition is satisfied due to the vehicle 14 and trailer 12 moving in a straight direction at a constant zero hitch angle value (γ=0), the controller 38 proceeds to step 625 and processes images captured by the imaging device 34 to derive a template image of the trailer 12 before proceeding to step 610 to determine the hitch angle via the template matching method. Otherwise, in instances where the steady state condition is satisfied due to the vehicle 14 and trailer 12 moving along a path at a constant non-zero hitch angle value (λ≠0), the controller proceeds to step 650 to determine the hitch angle via the steady state method. The steady state method may determine the hitch angle in approximately 1-3 seconds and is generally less reliable than the template matching method but more reliable than the centerline method. So long as the steady state condition is satisfied, the controller 38 will select either the template matching method or the steady state method. If the steady state method is the currently selected hitch angle detection method and the steady state condition is no longer satisfied, the controller 38 returns to step 630 to determine the hitch angle via the centerline method.
Referring to
Referring to
The optical flow method generally begins at step 800 where the controller 38 receives captured images from the imaging device 34 for imaging processing. At step 810, the controller 38 determines an optical flow from the image data of the captured images. According to one embodiment, the controller 38 processes successive captured images taken by imager 34 to generate an optical flow field at step 820. For purposes of illustration, a captured image 900 is generally shown in
At step 830, the controller 38 identifies optical flow vectors of pixels associated with the operating environment 908 of the trailer 12 and optical flow vectors of pixels associated with the imaged trailer 906. Generally, the optical flow vector of pixels associated with the operating environment of the trailer 12, particularly the travelling surface of the trailer 12, is based on both the longitudinal motion and the yaw of the trailer 12 and can be readily predicted. Accordingly, the controller 38 derives a predictive optical flow vector for each pixel in the captured image 900 by assuming each pixel is associated with the operating environment 908 of the trailer 12 based on a vehicle speed and a steering angle of the vehicle 14, both of which may be supplied to the controller 38 via the speed sensor 50 and the steering angle sensor 64, respectively. The magnitudes of the optical flow vectors may be scaled based on a scaling factor determined by camera parameters such as field of view, resolution, and the like. Having derived the predicted optical flow vector for each pixel in the captured image 900, the controller 38 compares the actual optical flow vector of each pixel to the corresponding predicted optical flow vector. If they are close (e.g., within a predetermined threshold) in both direction and magnitude, the controller 38 determines that the actual optical flow vector is associated with the operating environment 908 of the trailer 12. Otherwise, the controller 38 determines that the actual optical flow vector is associated with the imaged trailer 906. For example, with respect to the optical flow field of the captured image 900 shown in
Having identified the optical flow vectors of pixels associated with the imaged trailer 906, the controller 38 determines an angular velocity about a reference point (e.g., imaged hitch point 910 in
where {dot over (γ)} is the average angular velocity, δ is the steering angle, L is the drawbar length between the hitch point and a rear axle of the vehicle 14, D is the trailer length between the hitch point and an axle of the trailer 12, W is the wheelbase length between a front axle and the rear axle of the vehicle 14, and ν is the longitudinal speed of the vehicle 14. It is to be noted that the function ν/W tan δ corresponds to the yaw rate of the vehicle 14 and can be otherwise supplied by yaw rate sensor 52 (
Alternatively, the instantaneous yaw rate of the vehicle 14 may be determined by analyzing the optical field vectors of pixels associated with the operating environment 908 of the trailer 12. As shown in
Referring to
With further reference to
Referring to
At step 1030, a straight pull maneuver is performed where the vehicle 1000 is driven straight forward for a predetermined distance and/or time period such that the vehicle is longitudinally aligned with a connected trailer (e.g., trailer 1002 or 1055). In alternative embodiments, the vehicle and trailer combination may be backed up to accomplish the same. However, driving the vehicle 1000 forward may be preferable since it's typically easier for the driver to align the trailer to the vehicle 1000 in the forward direction. At step 1040, the imaging device 1008 captures images while the vehicle 1000 and trailer are longitudinally aligned. The captured images are processed by the controller 1010 to identify vehicle and trailer contours at step 1050. The trailer contours may be detected pursuant to steps 120-140 of the template matching method described previously herein. As outlined in those steps, successive images (e.g., image frames) are averaged or combined into an averaged image, which generally results in the blurring of imaged objects (e.g., the ground) that move in relation to the vehicle 1000 and the trailer. Next, an edge map or gradient map is derived from the averaged image and the trailer contour(s) can be detected by analyzing the intensity gradient for each pixel of the averaged image. It is contemplated that the vehicle contour(s) may be similarly detected. Alternatively, the vehicle contour(s) may be predetermined by the vehicle OEM since the shape/location of the vehicle 1000 and related components is known and unchanged in images captured by the imaging device 1008. In some embodiments where the vehicle contour(s) is known, all other identified contours may be characterized as trailer contours by default.
For purposes of illustration,
As depicted, the connecting member 1057 may be defined by a tongue 1058 of the trailer 1055 and a drawbar 1059 of the vehicle 1000. The tongue 1058 and drawbar 1059 may be coupled together via a ball and socket coupling 1060, for example. The respective contours of the tongue 1058, drawbar 1059, and any related accessories and components may be identified in the same manner as the trailer contour 1056. By virtue of the imaging device 1008 being centrally located on the tailgate 1012, the connecting member 1057 typically appears in captured images to be longitudinally aligned with the vehicle 1000 and the trailer 1055 during a straight pull maneuver. Thus, in identifying the connecting member 1057, the controller 1010 may consider whether the connecting member 1057 provides a qualified connection between the vehicle 1000 and the trailer 1055. Additionally, the controller 1010 may consider the size and/or shape of the connecting member 1057 in order to distinguish the same from any electrical connections (not shown) between the vehicle 1000 and the trailer 1055. However, in embodiments where the imaging device 1008 is offset, it is to be understood that the connecting member 1057 may appear differently (e.g., orientation, location, etc.) in the captured images. In view of the above, it is disclosed herein that the detection of a connection (e.g., connecting member 1057) in captured images provides key information in determining a trailer type, namely whether a connected trailer is a fifth wheel trailer or a standard trailer.
Referring back to
Similarly, if the controller 1010 predicts the trailer to be a standard trailer, the controller 1010 can validate the prediction if the identified contour (e.g., trailer contour 1056) appears in captured images to move in an expected direction. That is, if the vehicle 1000 makes a leftward turn, the identified trailer contour 1056 would appear in captured images to move in a clockwise direction generally specified by arrow 1096. Alternatively, if the vehicle 1000 makes a rightward turn, the identified trailer contour 1056 would appear in captured images to move in a counterclockwise direction generally specified by arrow 1098. With respect to the validation at step 1090, if the identified trailer contour(s) fails to move, the method 1020 may be reiterated from step 1030 so that a new trailer contour may be identified, if available, and the previously identified trailer contour may be ignored in subsequent iterations. In instances where an identified trailer contour fails to exhibit motion during a vehicle turn event, it is likely that a false identification occurred and the identified trailer contour in question is actually part of the vehicle 1000 (e.g., a bike rack).
Once an identified trailer contour has been validated to correspond to a fifth wheel trailer or a standard trailer, the trailer contour may be monitored by the controller 1010 while the vehicle 1000 is operational. It is conceived that the controller 1010 may save one or more images captured by the imaging device 1008 throughout the operation of the vehicle and may designate the captured images as belonging to a fifth wheel trailer or a standard trailer. In one specific scenario, the controller 1010 saves one or more images captured by the imaging device 1008 when the vehicle 1000 is placed in park. In this manner, if the vehicle 1000 is turned OFF, captured images will be saved showing the trailer, designated as either a fifth wheel or standard trailer, in its last known position. Assuming the trailer is left hitched to the vehicle 1000 and the vehicle 1000 is later turned ON, the controller 1010 may readily identify the trailer as being a fifth wheel trailer or a standard trailer by simply matching the saved captured images to new ones taken by the imaging device 1008 at vehicle start-up. In such scenarios, where a match occurs, the need to perform method 1020 may be bypassed. According to one embodiment, the captured images may be saved in non-volatile memory and retrieved for a later key cycle. In such instances, the captured images may be saved after the vehicle 1000 is turned OFF and retrieved before the vehicle is turned ON. In other words, power may be provided to the controller 1010 or other module after the vehicle 1000 is turned OFF. For example, the controller 1010 may be maintained in a low power state while the vehicle 1000 is turned OFF. However, it will be appreciated that the captured images may alternatively be stored to the controller 1010 before the vehicle 1000 is turned OFF, if desired. Likewise, the saved captured images may be retrieved at or after vehicle start-up.
It is to be understood that variations and modifications can be made on the aforementioned structures and methods without departing from the concepts of the present invention, and further it is to be understood that such concepts are intended to be covered by the following claims unless these claims by their language expressly state otherwise.
This application is a continuation-in-part of U.S. patent application Ser. No. 15/220,750, which was filed on Jul. 27, 2016, entitled “HITCH ANGLE DETECTION FOR TRAILER BACKUP ASSIST SYSTEM,” now U.S. Pat. No. 9,796,228 which is a continuation-in-part of U.S. patent application Ser. No. 14/972,761, filed Dec. 17, 2015, and entitled “HITCH ANGLE DETECTION FOR TRAILER BACKUP ASSIST SYSTEM,” now U.S. Pat. No. 9,610,975 the entire disclosures of which are hereby incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
4897642 | DiLullo et al. | Jan 1990 | A |
5142278 | Moallemi et al. | Aug 1992 | A |
5282641 | McLaughlin | Feb 1994 | A |
5442810 | Jenquin | Aug 1995 | A |
5905433 | Wortham | May 1999 | A |
5999091 | Wortham | Dec 1999 | A |
6111524 | Lesesky et al. | Aug 2000 | A |
6142372 | Wright | Nov 2000 | A |
6501376 | Dieckmann et al. | Dec 2002 | B2 |
6687609 | Hsiao et al. | Feb 2004 | B2 |
6847916 | Ying | Jan 2005 | B1 |
6999856 | Lee et al. | Feb 2006 | B2 |
7039504 | Tanaka et al. | May 2006 | B2 |
7136754 | Hahn et al. | Nov 2006 | B2 |
7175194 | Ball | Feb 2007 | B2 |
7696862 | Herschell et al. | Apr 2010 | B2 |
7715953 | Shepard | May 2010 | B2 |
7731302 | Tandy, Jr. et al. | Jun 2010 | B2 |
7760077 | Day | Jul 2010 | B2 |
7825782 | Hermann | Nov 2010 | B2 |
7932815 | Martinez et al. | Apr 2011 | B2 |
7950751 | Offerle et al. | May 2011 | B2 |
8010252 | Getman et al. | Aug 2011 | B2 |
8068019 | Bennie et al. | Nov 2011 | B2 |
8165770 | Getman et al. | Apr 2012 | B2 |
8179238 | Roberts, Sr. et al. | May 2012 | B2 |
8223204 | Hahn | Jul 2012 | B2 |
8362888 | Roberts, Sr. et al. | Jan 2013 | B2 |
8768535 | Kossira et al. | Jul 2014 | B2 |
8851521 | Rhymer | Oct 2014 | B2 |
8930140 | Trombley et al. | Jan 2015 | B2 |
9008913 | Sears et al. | Apr 2015 | B1 |
9156496 | Greenwood et al. | Oct 2015 | B2 |
9164955 | Lavoie et al. | Oct 2015 | B2 |
9227474 | Liu | Jan 2016 | B2 |
9238483 | Hafner et al. | Jan 2016 | B2 |
9315212 | Kyrtsos | Apr 2016 | B1 |
9335162 | Kyrtsos et al. | May 2016 | B2 |
9352777 | Lavoie et al. | May 2016 | B2 |
9428188 | Schwindt et al. | Aug 2016 | B2 |
9616923 | Lavoie et al. | Apr 2017 | B2 |
9623904 | Lavoie et al. | Apr 2017 | B2 |
9676377 | Hafner et al. | Jun 2017 | B2 |
9827818 | Hu et al. | Nov 2017 | B2 |
9840278 | Lavoie et al. | Dec 2017 | B2 |
20040222881 | Deng et al. | Nov 2004 | A1 |
20050071373 | Long | Mar 2005 | A1 |
20070058273 | Ito et al. | Mar 2007 | A1 |
20070271267 | Lim et al. | Nov 2007 | A1 |
20080231701 | Greenwood et al. | Sep 2008 | A1 |
20090045924 | Roberts, Sr. et al. | Feb 2009 | A1 |
20090093928 | Getman et al. | Apr 2009 | A1 |
20090219147 | Bradley et al. | Sep 2009 | A1 |
20090271078 | Dickinson | Oct 2009 | A1 |
20090322510 | Berger | Dec 2009 | A1 |
20100156667 | Bennie et al. | Jun 2010 | A1 |
20110001825 | Hahn | Jan 2011 | A1 |
20110257860 | Getman et al. | Oct 2011 | A1 |
20110281522 | Suda | Nov 2011 | A1 |
20120200706 | Greenwood et al. | Aug 2012 | A1 |
20130027195 | Van Wiemeersch et al. | Jan 2013 | A1 |
20130041524 | Brey | Feb 2013 | A1 |
20130148748 | Suda | Jun 2013 | A1 |
20130253814 | Wirthlin | Sep 2013 | A1 |
20130261843 | Kossira et al. | Oct 2013 | A1 |
20140005918 | Qiang | Jan 2014 | A1 |
20140052337 | Lavoie et al. | Feb 2014 | A1 |
20140058614 | Trombley et al. | Feb 2014 | A1 |
20140058622 | Trombley et al. | Feb 2014 | A1 |
20140058655 | Trombley et al. | Feb 2014 | A1 |
20140058668 | Trombley et al. | Feb 2014 | A1 |
20140160276 | Pliefke | Jun 2014 | A1 |
20140172232 | Rupp et al. | Jun 2014 | A1 |
20140188344 | Lavoie | Jul 2014 | A1 |
20140188346 | Lavoie | Jul 2014 | A1 |
20140210456 | Crossman | Jul 2014 | A1 |
20140218506 | Trombley et al. | Aug 2014 | A1 |
20140218522 | Lavoie et al. | Aug 2014 | A1 |
20140236532 | Trombley et al. | Aug 2014 | A1 |
20140249691 | Hafner et al. | Sep 2014 | A1 |
20140267688 | Aich et al. | Sep 2014 | A1 |
20140267689 | Lavoie | Sep 2014 | A1 |
20140277942 | Kyrtsos et al. | Sep 2014 | A1 |
20140297128 | Lavoie et al. | Oct 2014 | A1 |
20140303847 | Lavoie | Oct 2014 | A1 |
20140309888 | Smit et al. | Oct 2014 | A1 |
20140324295 | Lavoie | Oct 2014 | A1 |
20140379217 | Rupp et al. | Dec 2014 | A1 |
20150002669 | Reed et al. | Jan 2015 | A1 |
20150002670 | Bajpai | Jan 2015 | A1 |
20150025732 | Min et al. | Jan 2015 | A1 |
20150057903 | Rhode et al. | Feb 2015 | A1 |
20150066296 | Trombley et al. | Mar 2015 | A1 |
20150066349 | Chan | Mar 2015 | A1 |
20150077557 | Han et al. | Mar 2015 | A1 |
20150120141 | Lavoie et al. | Apr 2015 | A1 |
20150134183 | Lavoie et al. | May 2015 | A1 |
20150138340 | Lavoie | May 2015 | A1 |
20150158527 | Hafner et al. | Jun 2015 | A1 |
20150203156 | Hafner et al. | Jul 2015 | A1 |
20150210317 | Hafner et al. | Jul 2015 | A1 |
20160039456 | Lavoie et al. | Feb 2016 | A1 |
20160059780 | Lavoie | Mar 2016 | A1 |
20160059888 | Bradley et al. | Mar 2016 | A1 |
20160129939 | Singh et al. | May 2016 | A1 |
20160152263 | Singh et al. | Jun 2016 | A1 |
20160153778 | Singh et al. | Jun 2016 | A1 |
20160280267 | Lavoie et al. | Sep 2016 | A1 |
20170073005 | Ghneim et al. | Mar 2017 | A1 |
20170101130 | Lavoie | Apr 2017 | A1 |
20170177949 | Hu | Jun 2017 | A1 |
20170297619 | Lavoie et al. | Oct 2017 | A1 |
20170297620 | Lavoie et al. | Oct 2017 | A1 |
20170313351 | Lavoie | Nov 2017 | A1 |
Number | Date | Country |
---|---|---|
202159367 | Mar 2012 | CN |
102008045436 | Mar 2010 | DE |
1312492 | May 2003 | EP |
2014123575 | Aug 2014 | WO |
2015187467 | Dec 2015 | WO |
Entry |
---|
Skybitz, website, 2012, pp. 1-3, http://www.skybitz.com/products-services/hardware/bat-xtndr/. |
Verma, V.S.; Guntur, R.R.; Womg, J.Y.; “Directional Behavior During Braking of a Tractor/Semitrailer”, TRID, International Journal of Vehicle Design, May 1980, pp. 195-220, vol. 1, No. 3, Inderscience Enterprises Limited, ISSN: 1477-5360. |
Number | Date | Country | |
---|---|---|---|
20170177949 A1 | Jun 2017 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15220750 | Jul 2016 | US |
Child | 15352091 | US | |
Parent | 14972761 | Dec 2015 | US |
Child | 15220750 | US |