Centerline method for trailer hitch angle detection

Information

  • Patent Grant
  • 10155478
  • Patent Number
    10,155,478
  • Date Filed
    Thursday, December 17, 2015
    8 years ago
  • Date Issued
    Tuesday, December 18, 2018
    5 years ago
Abstract
A system for detecting a hitch angle between a vehicle and a trailer is provided herein. An imaging device is configured to capture images of the trailer and a controller is configured to process the captured images. The controller derives an edge map and identifies trailer pixels therein. The controller defines a number of candidate lines in the edge map and determines the hitch angle based on an angle between a selected candidate line and a reference line.
Description
FIELD OF THE INVENTION

The present invention generally relates to trailer backup assist systems, and more particularly, to trailer backup assist systems employing hitch angle detection through image processing.


BACKGROUND OF THE INVENTION

Reversing a vehicle while towing a trailer can be challenging for many drivers, particularly for drivers that drive with a trailer on an infrequent basis or with various types of trailers. Some systems used to assist a driver in backing a trailer rely on hitch angle measurements to determine the position of the trailer relative to the vehicle. Thus, the accuracy and reliability of the hitch angle measurements can be critical to the operation of the trailer backup assist system.


SUMMARY OF THE INVENTION

According to one aspect of the present invention, a system for detecting a hitch angle between a vehicle and a trailer is provided. An imaging device is configured to capture images of the trailer and a controller is configured to process the captured images. The controller derives an edge map and identifies trailer pixels therein. The controller defines a number of candidate lines in the edge map and determines the hitch angle based on an angle between a selected candidate line and a reference line.


According to another aspect of the present invention, a method of detecting a hitch angle between a vehicle and a trailer is provided. The method includes the steps of capturing images of the trailer using an imaging device and providing a controller for processing the captured images. The controller derives an edge map and identifies trailer pixels therein. The controller generates a number of candidate lines in the edge map and determines the hitch angle based on an angle between a selected candidate line and a reference line.


These and other features, advantages, and objects of the present invention will be further understood and appreciated by those skilled in the art by reference to the following specification, claims, and appended drawings.





BRIEF DESCRIPTION OF THE DRAWINGS

In the drawings:



FIG. 1 is a top perspective view of a vehicle attached to a trailer with one embodiment of a hitch angle sensor for operating a trailer backup assist system;



FIG. 2 is a block diagram illustrating one embodiment of the trailer backup assist system having a steering input device, a curvature controller, and a trailer braking system;



FIG. 3 is a flow diagram of a method of detecting a hitch angle, according to one embodiment;



FIG. 4 is a captured image showing a trailer in straight alignment with a vehicle and the presence of ground noise;



FIG. 5 is an edge map of the captured image shown in FIG. 4;



FIG. 6 illustrates the blurring of ground noise in an averaged image;



FIG. 7 is an edge map of the averaged image shown in FIG. 6;



FIG. 8 illustrates a trailer contour of a template image;



FIG. 9 illustrates a template image being matched to a search image to determine a hitch angle;



FIG. 10 is a search image having a proximity zone for jackknife detection and a number of candidate hitch point locations about which a template image can be rotated to determine an actual imaged hitch point and a hitch angle;



FIG. 11 is a flow diagram of a method of detecting a hitch angle, according to another embodiment;



FIG. 12 is an edge map of a captured image showing a number of candidate lines, one of which, is selected to determine a hitch angle based on its angular position relative to a reference line;



FIG. 13 is a flow diagram of a method of locating an imaged hitch point, according to one embodiment;



FIGS. 14 and 15 are each captured images in which a trailer appears at a distinct hitch angle and a detection window is used to scan an imaged drawbar to locate an imaged hitch point;



FIG. 16 is a graph illustrating averaged pixel differences for a number of pixel positions common to the captured images shown in FIGS. 14 and 15;



FIG. 17 is a kinematic model of the vehicle and trailer shown in FIG. 1;



FIG. 18 is a flow diagram of a method of detecting a hitch angle, according to yet another embodiment; and



FIG. 19 is a flow diagram of a method of initializing hitch angle detection, according to one embodiment.





DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS

For purposes of description herein, it is to be understood that the disclosed trailer backup assist system and the related methods may assume various alternative embodiments and orientations, except where expressly specified to the contrary. It is also to be understood that the specific devices and processes illustrated in the attached drawings, and described in the following specification, are simply exemplary embodiments of the inventive concepts defined in the appended claims. While various aspects of the trailer backup assist system and the related methods are described with reference to a particular illustrative embodiment, the disclosed invention is not limited to such embodiments, and additional modifications, applications, and embodiments may be implemented without departing from the disclosed invention. Hence, specific dimensions and other physical characteristics relating to the embodiments disclosed herein are not to be considered as limiting, unless the claims expressly state otherwise.


As used herein, the term “and/or,” when used in a list of two or more items, means that any one of the listed items can be employed by itself, or any combination of two or more of the listed items can be employed. For example, if a composition is described as containing components A, B, and/or C, the composition can contain A alone; B alone; C alone; A and B in combination; A and C in combination; B and C in combination; or A, B, and C in combination.


Referring to FIGS. 1 and 2, reference numeral 10 generally designates a trailer backup assist system for controlling a backing path of a trailer 12 attached to a vehicle 14 by allowing a driver of the vehicle 14 to specify a desired curvature of the backing path of the trailer 12. The vehicle 14 is embodied as a pickup truck that is pivotally attached to one embodiment of the trailer 12 that has a box frame 16 with an enclosed cargo area 18, a single axle 20 operably coupled to wheels 22 and 24, and a tongue 26 longitudinally extending forward from the enclosed cargo area 18. The illustrated trailer 12 also has a trailer hitch connector in the form of a coupler assembly 28 that is connected to a vehicle hitch connector in the form of a hitch ball 30 and drawbar 31. The coupler assembly 28 latches onto the hitch ball 30 to provide a pivoting hitch point 32 that allows for articulation of a hitch angle between the vehicle 14 and the trailer 12. As defined herein, the hitch angle corresponds to the angle formed between the center longitudinal axis of the vehicle 14 and of the trailer 12 (see hitch angle γ; FIG. 17). It should be appreciated that additional embodiments of the trailer 12 may alternatively couple with the vehicle 14 to provide a pivoting connection, such as by connecting with a fifth wheel connector. It is also contemplated that additional embodiments of the trailer 12 may include more than one axle and may have various shapes and sizes configured for different loads and items, such as a boat trailer or a flatbed trailer.


The trailer backup assist system 10 also includes an imaging device 34 located at the rear of the vehicle 14 and configured to image a rear-vehicle scene. The imaging device 34 may be centrally located at an upper region of a vehicle tailgate 35 such that the imaging device 34 is elevated relative to the tongue 26 of the trailer 12. The imaging device 34 has a field of view 36 located and oriented to capture one or more images that may include the tongue 26 of the trailer 12 and the hitch ball 30, among other things. Captured images are supplied to a controller 38 of the trailer backup assist system 10 and are processed by the controller 38 to determine the hitch angle between the vehicle 14 and the trailer 12, as will be described in greater detail herein. The controller 38 is configured with a microprocessor 40 and/or other analog and/or digital circuitry for processing one or more logic routines stored in a memory 42. The logic routines may include one or more hitch angle detection routines 44 and an operating routines 46. Information from the imaging device 34 or other components of the trailer backup assist system 10 can be supplied to the controller 38 via a communication network of the vehicle 14, which can include a controller area network (CAN), a local interconnect network (LIN), or other conventional protocols used in the automotive industry. It should be appreciated that the controller 38 may be a stand-alone dedicated controller or may be a shared controller integrated with the imaging device 34 or other component of the trailer backup assist system 10 in addition to any other conceivable onboard or off-board vehicle control systems.


With respect to the present embodiment, the controller 38 of trailer backup assist system 10 may be configured to communicate with a variety of vehicle equipment. The trailer backup assist system 10 may include a vehicle sensor module 48 that monitors certain dynamics of the vehicle 14. The vehicle sensor module 48 may generate a plurality of signals that are communicated to the controller 38 and may include a vehicle speed signal generated by a speed sensor 50 and a vehicle yaw rate signal generated by a yaw rate sensor 52. A steering input device 54 may be provided to enable a driver to control or otherwise modify the desired curvature of the backing path of the trailer 12. The steering input device 54 may be communicatively coupled to the controller 38 in a wired or wireless manner and provides the controller 38 with information defining the desired curvature of the backing path of the trailer 12. In response, the controller 38 processes the information and generates corresponding steering commands that are supplied to a power assist steering system 56 of the vehicle 14. In one embodiment, the steering input device 54 includes a rotatable knob 58 operable between a number of rotated positions that each provide an incremental change to the desired curvature of the backing path of the trailer 12.


According to one embodiment, the controller 38 of the trailer backup assist system 10 may control the power assist steering system 56 of the vehicle 14 to operate the steered wheels 60 of the vehicle 14 for moving the vehicle 14 in such a manner that the trailer 12 reacts in accordance with the desired curvature of the backing path of the trailer 12. The power assist steering system 56 may be an electric power-assisted steering (EPAS) system that includes an electric steering motor 62 for turning the steered wheels 60 to a steering angle based on a steering command generated by the controller 38, whereby the steering angle may be sensed by a steering angle sensor 64 of the power assist steering system 56 and provided to the controller 38. The steering command may be provided for autonomously steering the vehicle 14 during a backup maneuver and may alternatively be provided manually via a rotational position (e.g., a steering wheel angle) of a steering wheel 66 or the rotatable knob 58. However, in some embodiments, the steering wheel 66 of the vehicle 14 may be mechanically coupled with the steered wheels 60 of the vehicle 14, such that the steering wheel 66 moves in concert with steered wheels 60 via an internal torque, thereby preventing manual intervention with the steering wheel 66 during autonomous steering of the vehicle 14. In such instances, the power assist steering system 56 may include a torque sensor 68 that senses torque (e.g., gripping and/or turning) on the steering wheel 66 that is not expected from autonomous control of the steering wheel 68 and therefore indicative of manual intervention by the driver. In some embodiments, external torque applied to the steering wheel 66 may serve as a signal to the controller 38 that the driver has taken manual control and for the trailer backup assist system 10 to discontinue autonomous steering functionality.


The controller 38 of the trailer backup assist system 10 may also communicate with a vehicle brake control system 70 of the vehicle 14 to receive vehicle speed information such as individual wheel speeds of the vehicle 14. Additionally or alternatively, vehicle speed information may be provided to the controller 38 by a powertrain control system 72 and/or the speed sensor 50, among other conceivable means. It is conceivable that individual wheel speeds may be used to determine a vehicle yaw rate, which can be provided to the controller 38 in the alternative, or in addition to, the vehicle yaw rate measured by yaw rate sensor 52 of the vehicle sensor module 48. In some embodiments, the controller 38 may provide braking commands to the vehicle brake control system 70, thereby allowing the trailer backup assist system 10 to regulate the speed of the vehicle 14 during a backup maneuver of the trailer 12. It should be appreciated that the controller 38 may additionally or alternatively regulate the speed of the vehicle 14 via interaction with the powertrain control system 72.


Through interaction with the power assist steering system 56, the vehicle brake control system 70, and/or the powertrain control system 72 of the vehicle 14, the potential for unacceptable trailer backup conditions can be reduced. Examples of unacceptable trailer backup conditions include, but are not limited to, a vehicle over-speed condition, a high hitch angle rate, hitch angle dynamic instability, a trailer jackknife condition, sensor failure, and the like. In such circumstances, the driver may be unaware of the failure until the unacceptable trailer backup condition is imminent or already happening. Therefore, it is disclosed herein that the controller 38 of the trailer backup assist system 10 can generate an alert signal corresponding to a notification of an actual, impending, and/or anticipated unacceptable trailer backup condition, and prior to driver intervention, generate a counter measure to prevent such an unacceptable trailer backup condition.


According to one embodiment, the controller 38 may communicate with one or more devices, including a vehicle alert system 74, which may prompt visual, auditory, and tactile warnings. For instance, vehicle brake lights 76 and vehicle emergency flashers may provide a visual alert and a vehicle horn 78 and/or speaker 80 may provide an audible alert. Additionally, the controller 38 and/or vehicle alert system 74 may communicate with a human machine interface (HMI) 82 of the vehicle 14. The HMI 82 may include a touchscreen vehicle display 84 such as a center-stack mounted navigation or entertainment display capable of displaying images indicating the alert. Such an embodiment may be desirable to notify the driver of the vehicle 14 that an unacceptable trailer backup condition is afoot. Further, it is contemplated that the controller 38 may communicate via wireless communication with one or more electronic portable devices such as portable electronic device 86, which is embodied as a smartphone. The portable electronic device 86 may include a display 88 for displaying one or more images and other information to a user. In response, the portable electronic device 86 may provide feedback information, such as visual, audible, and tactile alerts.


Referring to FIG. 3, a method of detecting hitch angle is illustrated. The method, also referred to herein as “the template matching method,” may be executed by the controller 38 of the trailer backup assist system 10 and is shown as one embodiment of the hitch angle detection routine 44. The template matching method generally includes processing image information to distinguish trailer contour from ground noise in images captured by the imaging device 34. The trailer contour then serves as a template and is matched to a search image to determine the hitch angle between the vehicle 14 and the trailer. 12


For purposes of illustration, a captured image 90 is exemplarily shown in FIG. 4 illustrating the trailer 12 in straight alignment with the vehicle 14 and the presence of ground noise. As defined herein, ground noise generally corresponds to any ground structure capable of interfering with image acquisition of the trailer 12. With respect to the captured image 90, potential ground noise candidates may include large stones (e.g., stone 92) and irregular ground surfaces (e.g., ground surface 94). As such, it may be difficult to accurately identify the trailer 12 when an image acquisition technique, namely edge detection, is applied to the captured image 90, as exemplarily shown in FIG. 5. With these things in mind, the template matching method described herein is able to blur out ground noise to enable identification of one or more trailer contours. Once identified, the trailer contour(s) may be stored as a template image that is matched to a search image to determine the hitch angle between the vehicle 14 and the trailer 12. In practice, the method has been found highly robust and benefits from relatively fast and straightforward computations.


The template matching method may begin at step 100, where the driver or other occupant initiates the trailer backup assist system 10. This may be achieved via user-input made through the display 84 of the vehicle 14 or other conceivable means. At step 110, the driver is instructed to pull the trailer 12 in a straight direction such that the hitch angle between the vehicle 14 and the trailer 12 is substantially zero. While the vehicle 14 and trailer 12 are engaged in the straight pull maneuver, the controller 38 derives an averaged image of all images captured by the imaging device 34 during a period of time at step 120. It has been discovered that 1-3 seconds typically suffices. Notably, the trailer 12 appears stationary within the images captured by the imaging device 34 whereas ground noise constantly changes from image to image. Thus, with respect to the averaged image, pixels associated with the trailer 12 will keep their contrast whereas pixels associated with ground noise will be blurred. To illustrate this effect, an averaged image 125 is exemplarily shown in FIG. 6.


At step 130, the controller 38 derives an edge map of the averaged image by calculating the intensity gradient for each pixel of the averaged image 125. The intensity gradient, or edge value, of each pixel may range from 0 to 255. For purposes of illustration, an edge map 135 is exemplarily shown in FIG. 7, in which the edge values of pixels associated with ground noise have been substantially weakened due to the blurring effect. At step 140, the controller 38 compares the edge value of each pixel of the edge map 135 to a threshold value (e.g., 30). Pixels having an edge value meeting or exceeding the threshold value are identified as trailer pixels whereas pixels having an edge value not meeting or exceeding the threshold value are identified as ground noise pixels. Once the trailer pixels have been identified, the controller 38 determines one or more trailer contours at step 150. The trailer contour(s) are saved to the memory 42 of the controller 38 as a template image at step 160 and may include a substantial entirety of the imaged trailer 12 or portions thereof. For purposes of illustration, a trailer contour 152 is shown in FIG. 8. As shown, the trailer contour 152 has a square shape, which is generally more computationally efficient. In its current position, the trailer contour 152 may serve as a zero hitch angle reference and enables the hitch angle between the vehicle 14 and the trailer 12 to be determined in subsequent images (i.e., search images) via template matching at step 170.


According to one embodiment, as shown in FIG. 9, the hitch angle between the vehicle 14 and the trailer 12 may be determined based on a positional relationship between a template image 158 and a search image 171. More specifically, the hitch angle may be determined by superimposing the template image 158 over the search image 171 such that the template image 158 is initially in a zero hitch angle position and subsequently rotating the template image 158 about a rotation point, preferably the imaged hitch point 172. The direction of rotation can be predicted based on information received from the steering angle sensor 64 or other sensors from which an initial assessment can be made concerning the angular position of the trailer 12 relative to the vehicle 14. Once the template image 158 has been matched to the search image 171, the angle θ at which the template image 158 is rotated relative to the zero hitch angle position can be correlated to the hitch angle between the vehicle 14 and the trailer 12.


According to one embodiment, the imaged hitch point 172 may be determined through process of elimination. For instance, as exemplarily shown in FIG. 10, the controller 38 may define a number of candidate hitch point locations 173a-173d that are positioned along a reference line 174 that extends vertically across the middle column of the search image 171. The reference line 174 is defined by the controller 38 and is assumed to coincide with the center longitudinal axis of an imaged drawbar 175 and intersect with the hitch point 172 of the imaged vehicle 14 and trailer 12. The candidate hitch point locations 173a-173d are shown evenly spaced along the reference line 174 but may vary in number and spacing in other embodiments. Once the template image 158 has been derived and the vehicle 14 and trailer 12 are moving along a curved path, the controller 38 may superimpose the template image 158 onto the search image 171 at the zero hitch angle position and rotate the template image 158 about each of the candidate hitch point locations 173a-173d in an attempt to match the template image 158 with the search image 171. Based on the match quality, a confidence score is given to each candidate hitch point location 173a-173d and the candidate hitch point location 173a-173d receiving the highest confidence score is selected as the hitch point. In the event the matching quality associated with each candidate hitch point location 173a-173d is below a predetermined threshold, the controller 38 may define additional candidate hitch point locations (not shown) along the reference line 174 in either or both directions of the candidate hitch point location 173a-173d that received the highest confidence score and execute template matching with respect to each of the additional candidate hitch point locations. This process may be iterated as many times as needed until the predetermined threshold has been met. In so doing, the location of the candidate hitch point location that is ultimately selected as the imaged hitch point will closely mirror the location of the actual hitch point 172.


While matching the template image 158 to the search image 171, the controller 38 may additionally determine the presence of an imminent jackknife scenario at step 180. With continued reference to FIG. 10, the displacement of the template image 158 may be monitored relative to a proximity zone 182 while the template image 158 is rotated about the rotation point. In the illustrated embodiment, the proximity zone 182 may be defined as the space between an imaged rear bumper 184 of the vehicle 14 and a boundary line 186 that is defined by the controller 38 and overlaid onto the search image 171. The boundary line 186 may be v-shaped and includes a pair of straight segments 188 extending outwardly at an angle from a point 189 that is located on the reference line 174 and is disposed between the imaged hitch point 172 and the imaged rear bumper 184. It should be appreciated that the boundary line 186 may assume other shapes in alternative embodiments. The location and shape of the boundary line 186 may be determined based on various considerations such as, but not limited to, vehicle speed, trailer length, drawbar length, imager characteristics, trailer contour, and vehicle contour. It is generally assumed that vehicle speed, trailer length, drawbar length, and image characteristics are known or may be otherwise measured and inputted to the trailer backup assist system 10. Vehicle contour, such as that of the imaged rear bumper 184, may be programmed at the factory.


In the event the template image 158 crosses into the proximity zone 182 of the search image 171, the controller 38 determines that an imminent jackknife scenario is present and initiates a jackknife countermeasure at step 190. Otherwise, if it is determined that an imminent jackknife scenario is not present, the controller 38 may continue to determine the hitch angle between the vehicle 14 and the trailer 12, as discussed previously with respect to step 170. The jackknife countermeasure may include generating an auditory warning via the vehicle alert system 74, generating a visual warning via the display 84, generating a braking command to the vehicle brake control system 70, reducing the torque of the powertrain control system 72, modifying the steering angle of the vehicle 14, or a combination thereof in addition to any other conceivable countermeasures. Since the trailer 12 will likely be in motion upon the controller 38 determining that an imminent jackknife scenario is present, it is generally desirable to locate and dimension the proximity zone 182 in a manner that provides sufficient time for a jackknife scenario to be detected and a countermeasure to be implemented, thereby minimizing the potential of an actual jackknifing and/or collision between the trailer 12 and the vehicle 14. Doing so also overcomes any response latency that may be inherent in the trailer backup assist system 10. While steps 170 and 180 have been illustrated in a linear fashion, it should be appreciated that both steps may be performed simultaneously.


Referring to FIG. 11, another method of detecting hitch angle is illustrated. The method, also referred to herein as “the centerline method,” may be executed by the controller 38 of the trailer backup assist system 10 and is exemplarily shown as one embodiment of the hitch angle detection routine 44. The centerline method also utilizes image information obtained by processing images captured by the imaging device 34 to determine the hitch angle between the vehicle 14 and the trailer 12. The centerline method differs from the template matching method in that the vehicle 14 and the trailer 12 need not be moving a straight direction prior to hitch angle detection, thus making the centerline method particularly useful in instances where no template image is available for the trailer 12 and the driver is prevented from pulling the trailer 12 in a straight line. When compared to the template matching method, the centerline method generally benefits from faster processing times, but is also generally less reliable. Therefore, it can be said that the centerline method provides a quick start to hitch angle detection and may be replaced at a later time by the template matching method or other suitable methods providing more reliable hitch angle measurements. Thus, for purposes of illustration, the centerline method will be described in greater detail below under the condition that no template image is available and that the vehicle 14 and trailer 12 are initially moving along a curved path.


The centerline method may begin at step 200, where the controller 38 processes successive images captured by the imaging device 34 to derive an averaged image. At step 210, the controller 38 derives an edge map by calculating an intensity gradient, or edge value, for each pixel in the averaged image. At step 220, the controller 38 identifies trailer pixels in the edge map by comparing the edge value of each pixel to a threshold value and selecting only those pixels meeting or exceeding the threshold value to correspond to trailer pixels. For purposes of illustration, an edge map 222 is exemplarily shown in FIG. 12, wherein one embodiment of the imaged trailer 12 appears slightly blurred as a result of angular displacement of the trailer 12 relative to the vehicle 14. It is contemplated that additional threshold values may be used to differentiate between trailer pixels and vehicle pixels since vehicle pixels will generally have a higher degree of contrast relative to trailer pixels due to the blurring of the trailer 12. At step 230, the controller 38 defines a number of candidate lines on the edge map 222, as exemplarily shown in FIG. 12. The candidate lines project outwardly from a common projection point, preferably the imaged hitch point 232. The direction at which the candidate lines project may vary based on a width of the rear bumper of the vehicle 14, a length of the drawbar of the vehicle 14, and a length of the tongue of the trailer 12. In the illustrated embodiment, the candidate lines may vary from −90 degrees to 90 degrees with respect to a reference line 234 that is indicative of a predetermined hitch angle (e.g., a zero hitch angle) and extends vertically across the middle column of the edge map 222 and intersects the imaged hitch point 232. While the exact location of the imaged hitch point 232 may be unknown initially, the controller 38 may assign a default projection point from which to project the candidate lines. Since it's assumed the hitch point 232 will typically be located along the reference line 234 and usually falls within a predictable range (e.g., 10-20 centimeters from the rear bumper of the vehicle 14), the selection of a default projection point meeting the foregoing criteria is generally sufficient for the purposes of initial hitch angle detection.


At step 240, the controller 38 selects the candidate line (e.g., candidate line 242) having approximately the same number of trailer pixels on each of its sides, or said differently, the candidate line, or centerline, about which the trailer pixels are substantially symmetric. Once the controller 38 has made a candidate line selection, the controller 38 may determine the hitch angle between the vehicle 14 and the trailer 12 based on the angular position of the selected candidate line 242 relative to the reference line 234 at step 250. More specifically, the angle θ between the selected candidate line 242 and the reference line 234 can be correlated to the hitch angle between the vehicle 14 and the trailer 12. As the vehicle 14 and trailer 12 continue along its course, steps 200-250 may be iterated with subsequent images captured by the imaging device 34 to continually provide hitch angle measurements.


Referring to FIG. 13, a method of locating an imaged hitch point in images captured by the imaging device 34 is illustrated. The method, also referred to herein as “the drawbar scan method,” may be executed by the controller 38 of the trailer backup assist system 10 and may be embodied as a subroutine of the hitch angle detection routine 44. The drawbar scan method generally requires the trailer 12 to be moving relative to the vehicle 14 at a non-zero hitch angle in order to identify an imaged hitch point. As such, the drawbar scan method may be executed to provide a suitable rotation point or projection point when executing the centerline method or the template matching method in instances where the vehicle 14 and trailer 12 are moving along a curved path. By identifying the imaged hitch point, more accurate hitch angle measurements can be achieved.


The drawbar scan method generally begins at step 300, where the controller 38 selects two images captured by the imaging device 34 that show the trailer 12 at distinct hitch angles. The two images may be successive or non-successive depending on the frame rate of the imaging device 34. In practice, a noticeable difference in hitch angles between the two images is generally preferred. At step 310, the controller 38 derives an edge map for both images by calculating the intensity gradient, or edge value, for each of their corresponding pixels. For purposes of illustration, FIG. 14 shows a first image 312 in which the hitch angle between the vehicle 14 and the trailer 12 is approximately zero whereas FIG. 15 shows a second image 314 in which the hitch angle between the vehicle 14 and the trailer 12 is approximately 5 degrees relative to the zero hitch angle position shown in FIG. 14. For purposes of clarity, the edge maps associated with both images 312, 314 are not shown. At step 320, the controller 38 defines a detection window 322 of variable height and width in both images 312, 314. Each detection window 322 is centered at a common pixel position, such as pixel position 324, which is located on a reference line 326 that extends vertically across the middle column of the corresponding image 312, 314. The reference line 326 is defined by the controller 38 and is assumed to coincide with the center longitudinal axis of an imaged drawbar 327 and intersect with an imaged hitch point 328 between the vehicle 14 and the trailer 12.


At step 330, the controller 38 determines an average pixel intensity of the pixels bounded by each detection window 322 when centered at the current pixel position, and at step 340, the controller 38 calculates an averaged pixel difference, which is defined herein as the absolute value of the difference between the average pixel intensities, as calculated at step 330. Once the average pixel difference has been calculated, at step 350, the controller 38 shifts each detection window 322 in an upward vertical direction (as specified by arrow 342) so that each detection window 322 is commonly centered at a new pixel position that is 1 or more pixel positions higher on the corresponding reference line 326 than the previous pixel position. Thus, by making multiple iterations of steps 330-350, the controller 38 may calculate averaged pixel differences for when each detection window 322 is commonly centered at a number of pixel positions along the reference line 326. Once this has been done, at step 360, the controller 38 extrapolates the location of the hitch point 328 based on variations in the calculated averaged pixel differences.


With respect to images 312 and 314, it is generally expected that little variation will occur between the calculated averaged pixel differences associated with pixel positions that coincide with the imaged drawbar 327 due in part to the imaged drawbar 327 appearing in a common fixed position in both images 312, 314. In contrast, it is generally expected that greater variation to occur between the calculated averaged pixel differences associated with pixel positions that are located on portions of the reference line 326 that extend beyond the imaged drawbar 327 due in part to the trailer 12 appearing in different positions in both images 312, 314. For purposes of illustration, a graph is shown in FIG. 16 illustrating calculated averaged pixel differences for a number of pixel positions along the vertical reference line 326. The pixel positions may fall within a predetermined range in which the imaged hitch point 328 is expected to be located, thereby negating the need to determine averaged pixel differences for pixel positions along the vertical reference line 326 that are unlikely to correspond to the imaged hitch point 328.


As shown in FIG. 16, the graph generally demonstrates a relatively constant averaged pixel difference between 2 and 3 when each detection window 322 is commonly centered at pixel positions 0-44 as those pixel positions coincide with the imaged drawbar 327. In contrast the graph in FIG. 15 generally demonstrates a sharp increase in averaged pixel differences when each detection window 322 is commonly centered at pixel positions 46-70 as those pixel positions are located on portions of the reference line 326 that extend past the imaged drawbar 327. Recognizing this, the controller 38 may select, as the imaged hitch point 328, one of the pixel positions (e.g., pixel position 44 or 45) leading up to the sharp increase in averaged pixel differences. According to one embodiment, the controller 38 may iterate steps 330-350 of the drawbar scan method until the averaged pixel difference meets or exceeds a predetermined threshold value (e.g., 3.5) and select, as the imaged hitch point, the pixel position associated with the calculated averaged pixel difference that meets or exceeds the threshold value. This threshold value may be determined based on a number of considerations including, but not limited to, the size of the detection window 322, properties of the imaging device 34, etc. Once identified, the selected imaged hitch point should closely mirror the actual imaged hitch point 328 and may be used for hitch angle detection pursuant to the template matching method or the centerline method.


Referring to FIG. 17, a kinematic model of the vehicle 14 and trailer 12 is shown and serves as the basis for determining hitch angle according to another method, referred to herein as “the steady state method” and described in greater detail below. As shown in FIG. 17, the kinematic model is based on various parameters associated with the vehicle 14 and the trailer 12. These parameters include:


δ: steering angle at steered wheels 60 of the vehicle 14;


α: yaw angle of the vehicle 14;


β: yaw angle of the trailer 12;


γ: hitch angle between the vehicle 14 and the trailer 12 (γ=β−α);


W: wheelbase length between a front axle 370 and a rear axle 372 of the vehicle 14;


L: drawbar length between the hitch point 32 and the rear axle 372 of the vehicle 14;


D: trailer length between the hitch point 32 and axle 20 of the trailer 12 or effective axle for a multiple axle trailer; and


v: vehicle longitudinal speed.


From the kinematic model shown in FIG. 17, the yaw rate of the vehicle 14 may be represented with the following equation:








d





α

dt

=


-

v
W



tan





δ





Furthermore, the yaw rate of the trailer 12 may be represented with the following equation:








d





β

dt

=



v
D


sin





γ

+


Lv
DW


cos





γ





tan





δ






Accordingly, when the yaw rate of the vehicle 14 and the trailer 12 become equal, the hitch angle γ and the steering angle δ will be constant. This condition, referred to herein as steady state, can occur when a steering command is steadily maintained during a backing maneuver such as when the trailer 12 is reversed in a straight line with the vehicle 14 or when the vehicle 14 and trailer 12 are turning at a constant curvature for at least a threshold period of time or over a threshold distance of motion. Under such steady state driving conditions, the resulting hitch angle γ can be described using the following equation:

c=a cos γ+b sin γ


This equation can be rewritten as follows:

c=a√{square root over (1−sin2γ)}+b sin γ


The above equation can be rearranged into quadratic form and rewritten as follows:

c2−a2−2bc sin γ+(b2+a2)sin γ=0


Solving the quadratic equation for the hitch angle γ yields the following hitch angle equation:






γ
=

arcsin



bc
±

a




b
2

+

a
2

-

c
2







b
2

+

a
2










Where
,





c
=


-

1
W



tan





δ








b
=

1
D







a
=


L
DW


tan





δ





Accordingly, for a particular vehicle and trailer combination, the trailer length D, the wheelbase length W, and the drawbar length L are constant and assumed known. Thus, when the steady state condition is satisfied, the hitch angle γ between the vehicle 14 and trailer 12 may be determined as a function of the trailer length D, the wheelbase length W, the drawbar length L, and the steering angle δ.


Referring to FIG. 18, the steady state method is shown according to one embodiment. The steady state method may be executed by the controller 38 of the trailer backup assist system 10 and is exemplarily shown as one embodiment of the hitch angle detection routine 44. The method includes determining a steering angle at step 400. The steering angle may be provided by the steering angle sensor 64 and may be compensated to remove any offsets associated therewith. Next, at step 410, a steering angle rate is calculated and is filtered to remove noise. At step 420, it is determined whether the absolute value of the filtered steering angle rate is less than a threshold steering angle rate (e.g., 0.5 degrees per second) required for hitch angle calculation. The method also includes obtaining a hitch angle between the vehicle 14 and trailer 12, as measured pursuant to any of the hitch angle detection methods described herein (e.g., the centerline method) at step 430. At step 440, the controller 38 calculates a filtered hitch angle rate and determines at step 450 whether the absolute value of the filtered hitch angle rate is less than a threshold hitch angle rate (e.g., 0.5 degrees per second) required for hitch angle calculation. The method further includes obtaining a vehicle speed (e.g., from speed sensor 50) at step 460, calculating a filtered vehicle speed at step 470, and then determining at step 480 whether the absolute value of the filtered vehicle speed is greater than a threshold vehicle speed (e.g., 3 kilometers per second) required for hitch angle calculation. If the conditions specified at steps 420, 450, and 480 are met at step 490, the controller 38 determines that the steady state condition has been satisfied and calculates a hitch angle at step 500 using the hitch angle equation described herein with respect to the kinematic model shown in FIG. 17. So long as the steady state condition is satisfied, the controller 38 may continue to determine hitch angle via the hitch angle equation.


Referring to FIG. 19, a method of initializing hitch angle detection is illustrated. The method, referred to herein as the “hitch angle initialization method,” may be executed by the controller 38 of the trailer backup assist system 10 and is exemplarily shown as one embodiment of the operating routine 46. The hitch angle initialization method includes selecting between the various hitch angle detection methods described previously herein, which include the template matching method, the centerline method, and the steady state method, for the purposes of hitch angle detection. As described in greater detail below, the foregoing hitch angle detection methods generally vary in processing time and reliability. Thus, the hitch angle initialization method is executed in a manner such that the “best available” hitch angle method is chosen based on considerations including the availability of a template image for the trailer 12 being towed and current driving conditions.


The hitch angle initialization method may begin at step 600, where the controller 38 determines whether a template image is available for the trailer 12 being towed. If so, the controller 38 proceeds to step 610 to determine the hitch angle via the template matching method. The template matching method may determine the hitch angle in approximately 1 second and is generally the most reliable when compared to the centerline method and the steady state method. So long as the template image remains available, the template matching method is selected as the best available hitch angle detection. In the event no template image is available or the template matching method is unable to be executed (e.g., system error), the controller 38 proceeds to step 620 to determine whether the vehicle 14 and trailer 12 are moving in a straight direction. According to one embodiment, the direction of the vehicle 14 and trailer 12 may be determined by obtaining a steering angle from the steering angle sensor 64 over a period of time. If it is determined that the vehicle 14 and trailer 12 are moving in a straight direction, the controller 38 proceeds to step 625 and processes images captured by the imaging device 34 to derive a template image of the trailer 12 before proceeding to step 610 to determine the hitch angle via the template matching method. Otherwise, the controller 38 proceeds to step 630 to determine the hitch angle via the centerline method. The centerline method may determine the hitch angle in less than 1 second but is generally less reliable when compared to the template matching method and the steady state method.


Once the centerline method is selected, the controller 38 will continue to determine the hitch angle via the centerline method until a steady state condition is satisfied at step 640. As described previously herein, the steady state condition may be satisfied when the vehicle 14 and trailer 12 are moving in a straight direction or moving along a path at constant curvature. Or in other words, the steady state condition is satisfied when the yaw rate of the vehicle 14 and the trailer 12 become equal, thereby resulting in the hitch angle and the steering angle becoming constant. If the steady state condition is satisfied, the controller proceeds to step 650, where it determines whether the hitch angle is substantially zero. In instances where the steady state condition is satisfied due to the vehicle 14 and trailer 12 moving in a straight direction at a constant zero hitch angle value (γ=0), the controller 38 proceeds to step 625 and processes images captured by the imaging device 34 to derive a template image of the trailer 12 before proceeding to step 610 to determine the hitch angle via the template matching method. Otherwise, in instances where the steady state condition is satisfied due to the vehicle 14 and trailer 12 moving along a path at a constant non-zero hitch angle value (λ≠0), the controller proceeds to step 660 to determine the hitch angle via the steady state method. The steady state method may determine the hitch angle in approximately 1-3 seconds and is generally less reliable than the template matching method but more reliable than the centerline method. So long as the steady state condition is satisfied, the controller 38 will select either the template matching method or the steady state method. If the steady state method is the currently selected hitch angle detection method and the steady state condition is no longer satisfied, the controller 38 returns to step 630 to determine the hitch angle via the centerline method.


It is to be understood that variations and modifications can be made on the aforementioned structures and methods without departing from the concepts of the present invention, and further it is to be understood that such concepts are intended to be covered by the following claims unless these claims by their language expressly state otherwise.

Claims
  • 1. A system for detecting a hitch angle between a vehicle and a trailer, comprising: an imaging device configured to capture images of the trailer; andcircuitry configured to: process the captured images;derive an edge map and identify trailer pixels therein;define a number of candidate lines projecting from a common projection point in the edge map; anddetermine the hitch angle based on an angle between a selected candidate line and a reference line wherein the candidate lines are derived based on a width of a rear vehicle bumper, a drawbar length, and a trailer length; wherein the reference line extends along a middle column of the edge map and intersects an imaged hitch point between the vehicle and the trailer.
  • 2. The system of claim 1, wherein the circuitry processes successive images captured by the imaging device to derive an averaged image and derives the edge map by calculating an edge value for each pixel in the averaged image.
  • 3. The system of claim 2, wherein the circuitry identifies trailer pixels by comparing the edge value of each pixel within the edge map to a threshold value and selecting those pixels exceeding the threshold value to correspond to trailer pixels.
  • 4. The system of claim 1, wherein the projection point corresponds to an imaged hitch point between the vehicle and the trailer.
  • 5. The system of claim 1, wherein the projection point corresponds to a default projection point and is located in the edge map at a position in which an imaged hitch point between the vehicle and the trailer is assumed to be located.
  • 6. The system of claim 1, wherein the reference line is indicative of a predetermined hitch angle.
  • 7. The system of claim 1, wherein the selected candidate line corresponds to whichever candidate line has approximately the same number of trailer pixels on each of its sides.
  • 8. A method of detecting a hitch angle between a vehicle and a trailer, comprising the steps of: capturing images of the trailer using an imaging device;providing circuitry for: processing the captured images;deriving an edge map and identifying trailer pixels therein;defining a number of candidate lines in the edge map; anddetermining the hitch angle based on an angle between a selected candidate line and a reference line indicating a predetermined hitch angle wherein the candidate lines are derived based on a width of a rear vehicle bumper, a drawbar length, and a trailer length; wherein the reference line extends along a middle column of the edge map and intersects an imaged hitch point between the vehicle and the trailer.
  • 9. The system of claim 8, wherein the circuitry processes successive images captured by the imaging device to derive an averaged image and derives the edge map by calculating an edge value for each pixel in the averaged image.
  • 10. The system of claim 9, wherein the circuitry identifies trailer pixels by comparing the edge value of each pixel within the edge map to a threshold value and selecting those pixels exceeding the threshold value to correspond to trailer pixels.
  • 11. The system of claim 8, wherein the candidate lines project from a common projection point.
  • 12. The system of claim 11, wherein the projection point corresponds to an imaged hitch point between the vehicle and the trailer.
  • 13. The system of claim 11, wherein the projection point corresponds to a default projection point and is located in the edge map at a position in which an imaged hitch point between the vehicle and the trailer is assumed to be located.
  • 14. The system of claim 8, wherein the selected candidate line corresponds to whichever candidate line has approximately the same number of trailer pixels on each of its sides.
US Referenced Citations (379)
Number Name Date Kind
3542390 Fikes et al. Nov 1970 A
3605088 Savelli Sep 1971 A
3787077 Sanders Jan 1974 A
3833928 Gavit et al. Sep 1974 A
3860257 Mesley Jan 1975 A
4040006 Kimmel Aug 1977 A
4042132 Bohman et al. Aug 1977 A
4122390 Kollitz et al. Oct 1978 A
4212483 Howard Jul 1980 A
4366966 Ratsko et al. Jan 1983 A
4735432 Brown Apr 1988 A
4752080 Rogers Jun 1988 A
4848449 Martinet et al. Jul 1989 A
4852901 Beasley et al. Aug 1989 A
4943080 Reimer Jul 1990 A
5001639 Breen Mar 1991 A
5056905 Jensen Oct 1991 A
5097250 Hernandez Mar 1992 A
5108123 Rubenzik Apr 1992 A
5108158 Breen Apr 1992 A
5132851 Bomar et al. Jul 1992 A
5152544 Dierker, Jr. et al. Oct 1992 A
5191328 Nelson Mar 1993 A
5244226 Bergh Sep 1993 A
5246242 Penzotti Sep 1993 A
5247442 Kendall Sep 1993 A
5282641 McLaughlin Feb 1994 A
5289892 Notsu Mar 1994 A
5290057 Pellerito Mar 1994 A
5455557 Noll et al. Oct 1995 A
5517419 Lanckton May 1996 A
5521633 Nakajima et al. May 1996 A
5523947 Breen Jun 1996 A
5541778 DeFlorio Jul 1996 A
5558350 Kimbrough et al. Sep 1996 A
5559696 Borenstein Sep 1996 A
5579228 Kimbrough et al. Nov 1996 A
5631656 Hartman et al. May 1997 A
5650764 McCullough Jul 1997 A
5670935 Schofield Sep 1997 A
5690347 Juergens et al. Nov 1997 A
5719713 Brown Feb 1998 A
5747683 Gerum et al. May 1998 A
5821852 Fairchild Oct 1998 A
5837994 Stam Nov 1998 A
5980048 Rannells, Jr. et al. Nov 1999 A
6041582 Tiede et al. Mar 2000 A
6042196 Nakamura et al. Mar 2000 A
6049171 Stam Apr 2000 A
6124709 Allwine Sep 2000 A
6151175 Osha Nov 2000 A
6198992 Winslow Mar 2001 B1
6217177 Rost Apr 2001 B1
6218828 Bates et al. Apr 2001 B1
6223104 Kamen et al. Apr 2001 B1
6223114 Boros et al. Apr 2001 B1
6268800 Howard Jul 2001 B1
6292094 Deng et al. Sep 2001 B1
6351698 Kubota et al. Feb 2002 B1
6472865 Tola et al. Oct 2002 B1
6480104 Wall et al. Nov 2002 B1
6483429 Yasui et al. Nov 2002 B1
6494476 Masters et al. Dec 2002 B2
6498977 Wetzel et al. Dec 2002 B2
6539288 Ishida et al. Mar 2003 B2
6568093 Kogiso et al. May 2003 B2
6577952 Geier et al. Jun 2003 B2
6587573 Stam Jul 2003 B1
6611610 Stam Aug 2003 B1
6612394 Wessman Sep 2003 B2
6668225 Oh et al. Dec 2003 B2
6693524 Payne Feb 2004 B1
6704653 Kuriya et al. Mar 2004 B2
6712378 Austin Mar 2004 B1
6801125 McGregor Oct 2004 B1
6806809 Lee et al. Oct 2004 B2
6820888 Griffin Nov 2004 B1
6838979 Deng et al. Jan 2005 B2
6854557 Deng et al. Feb 2005 B1
6857494 Kobayashi et al. Feb 2005 B2
6879240 Kruse Apr 2005 B2
6891563 Schofield May 2005 B2
6956468 Lee Oct 2005 B2
6959970 Tseng Nov 2005 B2
6999856 Lee et al. Feb 2006 B2
7005974 McMahon et al. Feb 2006 B2
7006127 Mizusawa et al. Feb 2006 B2
7008088 Pisciotti Mar 2006 B2
7028804 Eki et al. Apr 2006 B2
7032705 Zheng et al. Apr 2006 B2
7046127 Boddy May 2006 B2
7058493 Inagaki Jun 2006 B2
7089101 Fischer et al. Aug 2006 B2
7154385 Lee et al. Dec 2006 B2
7159890 Craig et al. Jan 2007 B2
7167785 Lohberg et al. Jan 2007 B2
7170285 Spratte Jan 2007 B2
7171330 Kruse et al. Jan 2007 B2
7204504 Gehring et al. Apr 2007 B2
7219913 Atley May 2007 B2
7225891 Gehring et al. Jun 2007 B2
7229139 Lu et al. Jun 2007 B2
7239958 Grougan et al. Jul 2007 B2
7269489 Deng et al. Sep 2007 B2
7272481 Einig et al. Sep 2007 B2
7295907 Lu et al. Nov 2007 B2
7401871 Lu et al. Jul 2008 B2
7405557 Spratte et al. Jul 2008 B2
7413266 Lenz et al. Aug 2008 B2
7425889 Widmann Sep 2008 B2
7447585 Tandy, Jr. et al. Nov 2008 B2
7451020 Goetting et al. Nov 2008 B2
7463137 Wishart et al. Dec 2008 B2
7504995 Lawrence et al. Mar 2009 B2
7532109 Takahama et al. May 2009 B2
7540523 Russell et al. Jun 2009 B2
7548155 Schutt et al. Jun 2009 B2
7568716 Dietz Aug 2009 B2
7623952 Unruh et al. Nov 2009 B2
7648153 Metternich et al. Jan 2010 B2
7690737 Lu Apr 2010 B2
7715953 Shepard May 2010 B2
7720580 Higgins-Luthman May 2010 B2
7731302 Tandy, Jr. et al. Jun 2010 B2
7793965 Padula Sep 2010 B2
7798263 Tandy, Jr. et al. Sep 2010 B2
7878545 Rhymer et al. Feb 2011 B2
7881496 Camilleri Feb 2011 B2
7904222 Lee et al. Mar 2011 B2
7905507 Perri Mar 2011 B2
7950751 Offerle et al. May 2011 B2
7953536 Katrak May 2011 B2
7974444 Hongo Jul 2011 B2
8010252 Getman Aug 2011 B2
8010253 Lundquist Aug 2011 B2
8036792 Dechamp Oct 2011 B2
8038166 Piesinger Oct 2011 B1
8044779 Hahn et al. Oct 2011 B2
8073594 Lee et al. Dec 2011 B2
8157284 McGhie et al. Apr 2012 B1
8165770 Getman et al. Apr 2012 B2
8167444 Lee et al. May 2012 B2
8170726 Chen et al. May 2012 B2
8174576 Akatsuka et al. May 2012 B2
8180543 Futamura et al. May 2012 B2
8190364 Rekow May 2012 B2
8191915 Freese, V et al. Jun 2012 B2
8192036 Lee et al. Jun 2012 B2
8215436 DeGrave et al. Jul 2012 B2
8223204 Hahn Jul 2012 B2
8244442 Craig et al. Aug 2012 B2
8260518 Englert Sep 2012 B2
8267485 Barlsen et al. Sep 2012 B2
8280607 Gatti et al. Oct 2012 B2
8308182 Ortmann et al. Nov 2012 B2
8326504 Wu et al. Dec 2012 B2
8342560 Albers et al. Jan 2013 B2
8380390 Sy et al. Feb 2013 B2
8380416 Offerle et al. Feb 2013 B2
8393632 Vortmeyer et al. Mar 2013 B2
8401744 Chiocco Mar 2013 B2
8427288 Schofield et al. Apr 2013 B2
8469125 Yu et al. Jun 2013 B2
8504243 Kageyama Aug 2013 B2
8548680 Ryerson et al. Oct 2013 B2
8548683 Cebon et al. Oct 2013 B2
8576115 Basten Nov 2013 B2
8626382 Obradovich Jan 2014 B2
8675953 Elwell et al. Mar 2014 B1
8755984 Rupp et al. Jun 2014 B2
8807261 Subrt et al. Aug 2014 B2
8825328 Rupp et al. Sep 2014 B2
8833789 Anderson Sep 2014 B2
8838353 Wu Sep 2014 B2
8886400 Kossira et al. Nov 2014 B2
8888120 Trevino Nov 2014 B2
8909426 Rhode Dec 2014 B2
8930140 Trombley et al. Jan 2015 B2
8939462 Adamczyk et al. Jan 2015 B2
8955865 Fortin et al. Feb 2015 B2
8972109 Lavoie et al. Mar 2015 B2
9008913 Sears et al. Apr 2015 B1
9026311 Pieronek et al. May 2015 B1
9042603 Elwart et al. May 2015 B2
9082315 Lin et al. Jul 2015 B2
9085261 Lu Jul 2015 B2
9102272 Trombley Aug 2015 B2
9108598 Headley Aug 2015 B2
9114832 Wang et al. Aug 2015 B2
9120358 Motts et al. Sep 2015 B2
9120359 Chiu et al. Sep 2015 B2
9126525 Lynam Sep 2015 B2
9156496 Greenwood et al. Oct 2015 B2
9164955 Lavoie et al. Oct 2015 B2
9180890 Lu et al. Nov 2015 B2
9187124 Trombley et al. Nov 2015 B2
9227474 Liu Jan 2016 B2
9233710 Lavoie Jan 2016 B2
9238483 Hafner et al. Jan 2016 B2
9248858 Lavoie et al. Feb 2016 B2
9264672 Lynam Feb 2016 B2
9283892 Trombley Mar 2016 B2
9296422 Lavoie Mar 2016 B2
9315212 Kyrtsos Apr 2016 B1
9321483 Headley Apr 2016 B2
9335162 Kyrtsos May 2016 B2
9340228 Xu et al. May 2016 B2
9393996 Goswami et al. Jul 2016 B2
9428188 Schwindt et al. Aug 2016 B2
9434414 Lavoie Sep 2016 B2
9446713 Lu Sep 2016 B2
9500497 Lavoie Nov 2016 B2
9558409 Pliefke Jan 2017 B2
9616923 Lavoie Apr 2017 B2
9623904 Lavoie et al. Apr 2017 B2
9676377 Hafner et al. Jun 2017 B2
9827818 Hu et al. Nov 2017 B2
20010024333 Rost Sep 2001 A1
20010037164 Hecker Nov 2001 A1
20020128764 Hecker et al. Sep 2002 A1
20030133014 Mendoza Jul 2003 A1
20030137586 Lewellen Jul 2003 A1
20030234512 Holub Dec 2003 A1
20040017285 Zielinski et al. Jan 2004 A1
20040021291 Haug et al. Feb 2004 A1
20040093139 Wildey et al. May 2004 A1
20040130441 Lee Jul 2004 A1
20040222881 Deng et al. Nov 2004 A1
20050000738 Gehring Jan 2005 A1
20050074143 Kawai Apr 2005 A1
20050128059 Vause Jun 2005 A1
20050206224 Lu Sep 2005 A1
20050206225 Offerle Sep 2005 A1
20050206229 Lu et al. Sep 2005 A1
20050206231 Lu et al. Sep 2005 A1
20050236201 Spannheimer et al. Oct 2005 A1
20050236894 Lu Oct 2005 A1
20050236896 Offerle Oct 2005 A1
20050237385 Kosaka Oct 2005 A1
20060041358 Hara Feb 2006 A1
20060050018 Hutzel Mar 2006 A1
20060071447 Gehring et al. Apr 2006 A1
20060076828 Lu et al. Apr 2006 A1
20060103511 Lee et al. May 2006 A1
20060103727 Tseng May 2006 A1
20060111820 Goetting et al. May 2006 A1
20060142936 Dix Jun 2006 A1
20060152351 Daura Luna Jul 2006 A1
20060155455 Lucas et al. Jul 2006 A1
20060244579 Raab Nov 2006 A1
20070027581 Bauer et al. Feb 2007 A1
20070090688 Haemmerling et al. Apr 2007 A1
20070132560 Nystrom et al. Jun 2007 A1
20070152424 Deng et al. Jul 2007 A1
20070285808 Beale Dec 2007 A1
20080143593 Graziano et al. Jun 2008 A1
20080147277 Lu et al. Jun 2008 A1
20080231701 Greenwood Sep 2008 A1
20080312792 Dechamp Dec 2008 A1
20090005932 Lee Jan 2009 A1
20090045924 Roberts, Sr. Feb 2009 A1
20090079828 Lee et al. Mar 2009 A1
20090085775 Otsuka et al. Apr 2009 A1
20090093928 Getman et al. Apr 2009 A1
20090198425 Englert Aug 2009 A1
20090228182 Waldbauer et al. Sep 2009 A1
20090248346 Fennel et al. Oct 2009 A1
20090300701 Karaoguz et al. Dec 2009 A1
20090306861 Schumann et al. Dec 2009 A1
20100063702 Sabelstrom et al. Mar 2010 A1
20100171828 Ishii Jul 2010 A1
20100332049 Sy et al. Dec 2010 A1
20110001825 Hahn Jan 2011 A1
20110018231 Collenberg Jan 2011 A1
20110022282 Wu et al. Jan 2011 A1
20110050903 Vorobiev Mar 2011 A1
20110087398 Lu et al. Apr 2011 A1
20110112721 Wang et al. May 2011 A1
20110125457 Lee et al. May 2011 A1
20110160956 Chung et al. Jun 2011 A1
20110257860 Getman et al. Oct 2011 A1
20120041658 Turner Feb 2012 A1
20120045112 Lundblad Feb 2012 A1
20120086808 Lynam et al. Apr 2012 A1
20120095649 Klier et al. Apr 2012 A1
20120109471 Wu May 2012 A1
20120112434 Albers et al. May 2012 A1
20120185131 Headley Jul 2012 A1
20120200706 Greenwood Aug 2012 A1
20120265416 Lu Oct 2012 A1
20120271512 Rupp et al. Oct 2012 A1
20120271514 Lavoie et al. Oct 2012 A1
20120271515 Rhode et al. Oct 2012 A1
20120271522 Rupp et al. Oct 2012 A1
20120283909 Dix Nov 2012 A1
20120283910 Lee et al. Nov 2012 A1
20120310594 Watanabe Dec 2012 A1
20130006472 McClain et al. Jan 2013 A1
20130024064 Shepard Jan 2013 A1
20130027195 Van Wiemeersch et al. Jan 2013 A1
20130082453 Padula Apr 2013 A1
20130158863 Skvarce et al. Jun 2013 A1
20130179038 Goswami et al. Jul 2013 A1
20130207834 Mizutani et al. Aug 2013 A1
20130226390 Luo et al. Aug 2013 A1
20130250114 Lu Sep 2013 A1
20130261843 Kossira et al. Oct 2013 A1
20130268160 Trombley et al. Oct 2013 A1
20140005918 Qiang Jan 2014 A1
20140025260 McClure Jan 2014 A1
20140052337 Lavoie et al. Feb 2014 A1
20140058614 Trombley et al. Feb 2014 A1
20140058622 Trombley et al. Feb 2014 A1
20140058655 Trombley et al. Feb 2014 A1
20140058668 Trombley et al. Feb 2014 A1
20140067154 Yu et al. Mar 2014 A1
20140067155 Yu et al. Mar 2014 A1
20140085472 Lu et al. Mar 2014 A1
20140088824 Ishimoto Mar 2014 A1
20140160276 Pliefke Jun 2014 A1
20140172232 Rupp et al. Jun 2014 A1
20140183841 Jones Jul 2014 A1
20140188344 Lavoie Jul 2014 A1
20140188346 Lavoie Jul 2014 A1
20140200759 Lu Jul 2014 A1
20140210456 Crossman Jul 2014 A1
20140218506 Trombley Aug 2014 A1
20140218522 Lavoie et al. Aug 2014 A1
20140222288 Lavoie et al. Aug 2014 A1
20140236532 Trombley et al. Aug 2014 A1
20140249691 Hafner Sep 2014 A1
20140267688 Aich et al. Sep 2014 A1
20140267689 Lavoie Sep 2014 A1
20140277941 Chiu et al. Sep 2014 A1
20140277942 Kyrtsos Sep 2014 A1
20140297128 Lavoie et al. Oct 2014 A1
20140297129 Lavoie et al. Oct 2014 A1
20140303847 Lavoie Oct 2014 A1
20140303849 Hafner Oct 2014 A1
20140309888 Smit et al. Oct 2014 A1
20140324295 Lavoie Oct 2014 A1
20140343795 Lavoie Nov 2014 A1
20140358429 Shutko et al. Dec 2014 A1
20140379217 Rupp et al. Dec 2014 A1
20150002670 Bajpai Jan 2015 A1
20150035256 Klank et al. Feb 2015 A1
20150057903 Rhode et al. Feb 2015 A1
20150066296 Trombley et al. Mar 2015 A1
20150066298 Sharma et al. Mar 2015 A1
20150105975 Dunn Apr 2015 A1
20150115571 Zhang et al. Apr 2015 A1
20150120141 Lavoie et al. Apr 2015 A1
20150120143 Schlichting Apr 2015 A1
20150134183 Lavoie et al. May 2015 A1
20150138340 Lavoie May 2015 A1
20150149040 Hueger et al. May 2015 A1
20150158527 Hafner et al. Jun 2015 A1
20150165850 Chiu et al. Jun 2015 A1
20150197278 Boos et al. Jul 2015 A1
20150203156 Hafner et al. Jul 2015 A1
20150210254 Pieronek et al. Jul 2015 A1
20150210317 Hafner et al. Jul 2015 A1
20150217693 Pliefke et al. Aug 2015 A1
20150269444 Lameyre et al. Sep 2015 A1
20160001705 Greenwood et al. Jan 2016 A1
20160009288 Yu Jan 2016 A1
20160039456 Lavoie et al. Feb 2016 A1
20160059780 Lavoie Mar 2016 A1
20160059888 Bradley et al. Mar 2016 A1
20160059889 Herzog et al. Mar 2016 A1
20160096549 Herzog et al. Apr 2016 A1
20160129939 Singh et al. May 2016 A1
20160152263 Singh et al. Jun 2016 A1
20160153778 Singh et al. Jun 2016 A1
20160229452 Lavoie et al. Aug 2016 A1
20160280267 Lavoie et al. Sep 2016 A1
20170073005 Ghneim et al. Mar 2017 A1
20170101130 Lavoie Apr 2017 A1
20170106796 Lavoie et al. Apr 2017 A1
Foreign Referenced Citations (96)
Number Date Country
102582686 Sep 2013 CN
3923676 Jan 1991 DE
3931518 Apr 1991 DE
9208595 Aug 1992 DE
19526702 Feb 1997 DE
10030738 Aug 2001 DE
10031244 Jan 2002 DE
10065230 Jul 2002 DE
10122562 Jul 2002 DE
10154612 May 2003 DE
10312548 May 2004 DE
10333998 Feb 2005 DE
102004050149 Apr 2006 DE
102005042957 Mar 2007 DE
102005043466 Mar 2007 DE
102005043467 Mar 2007 DE
102005043468 Mar 2007 DE
102006002294 Jul 2007 DE
102006048947 Apr 2008 DE
102006056408 Jun 2008 DE
102008020838 Nov 2008 DE
102007029413 Jan 2009 DE
102008045436 Mar 2010 DE
102006035021 Apr 2010 DE
102008043675 May 2010 DE
102009007990 Aug 2010 DE
102009012253 Sep 2010 DE
102009027041 Dec 2010 DE
102009038552 Feb 2011 DE
102010006323 Aug 2011 DE
102008004158 Oct 2011 DE
102008004159 Oct 2011 DE
102008004160 Oct 2011 DE
102010021052 Nov 2011 DE
102010029184 Nov 2011 DE
102010045519 Mar 2012 DE
102011104256 Jul 2012 DE
102011101990 Oct 2012 DE
102012005707 Oct 2012 DE
202012010517 Dec 2012 DE
102011108440 Jan 2013 DE
102011120814 Jun 2013 DE
102012006206 Oct 2013 DE
102012206133 Oct 2013 DE
102012019234 Apr 2014 DE
102013000198 Jul 2014 DE
0418653 Mar 1991 EP
0433858 Jun 1991 EP
1361543 Nov 2003 EP
1593552 Mar 2007 EP
1810913 Jul 2007 EP
2388180 Nov 2011 EP
2452549 May 2012 EP
2487454 Aug 2012 EP
2551132 Jan 2013 EP
2644477 Oct 2013 EP
2803944 Nov 2014 EP
2515379 Apr 1983 FR
2265587 Oct 1993 GB
2342630 Apr 2000 GB
2398048 Aug 2004 GB
2398049 Aug 2004 GB
2398050 Aug 2004 GB
61006458 Jan 1986 JP
6159491 Mar 1986 JP
6385568 Jun 1988 JP
01095980 Apr 1989 JP
01095981 Apr 1989 JP
09267762 Oct 1997 JP
09328078 Dec 1997 JP
10001063 Jan 1998 JP
11124051 May 1999 JP
11278319 Oct 1999 JP
2002012172 Jan 2002 JP
2002068032 Mar 2002 JP
2003034261 Feb 2003 JP
2003148938 May 2003 JP
3716722 Nov 2005 JP
2008027138 Feb 2008 JP
2012105158 May 2012 JP
2012166647 Sep 2012 JP
2014002056 Jan 2014 JP
8503263 Aug 1985 WO
0044605 Aug 2000 WO
2005005200 Jan 2005 WO
2005116688 Dec 2005 WO
2006042665 Apr 2006 WO
2012059207 May 2012 WO
2012103193 Aug 2012 WO
2014019730 Feb 2014 WO
2014037500 Mar 2014 WO
2014070047 May 2014 WO
2014092611 Jun 2014 WO
2014123575 Aug 2014 WO
2015074027 May 2015 WO
2015187467 Dec 2015 WO
Non-Patent Literature Citations (43)
Entry
“Ford Super Duty: Truck Technologies”, Brochure, Sep. 2011, 2 pages.
Kristopher Bunker, “2012 Guide to Towing”, Trailer Life, 2012, 38 pages.
A. Gonzalez-Cantos, “Backing-Up Maneuvers of Autonomous Tractor-Trailer Vehicles using the Qualitative Theory of Nonlinear Dynamical Systems,” International Journal of Robotics Research, Jan. 2009, vol. 28, 1 page.
L. Chu, Y. Fang, M. Shang, J. Guo, F. Zhou, “Estimation of Articulation Angle for Tractor Semi-Trailer Based on State Observer”, ACM Digital Library, ICMTMA '10 Proceedings of the 2010 International Conference on Measuring Technology and Automation, vol. 2, Mar. 2010, 1 page.
M. Wagner, D. Zoebel, and A. Meroth, “Adaptive Software and Systems Architecture for Driver Assistance Systems” International Journal of Machine Learning and Computing, Oct. 2011, vol. 1, No. 4, 7 pages.
F.W. Kienhöfer; D. Cebon, “An Investigation of ABS Strategies for Articulated Vehicles”, Cambridge University, Engineering Department, United Kingdom, date unknown, 13 pages.
C. Lundquist; W. Reinelt; O. Enqvist, “Back Driving Assistant for Passenger Cars with Trailer”, ZF Lenksysteme GmbH, Schwabisch Gmünd, Germany, 2006 (SAE Int'l) Jan. 2006, 8 pages.
Zhe Leng; Minor, M., “A Simple Tractor-Trailer Backing Control Law for Path Following”, IEEE, Intelligent Robots and Systems (IROS) IEEE/RSJ International Conference, Oct. 2010, 2 pages.
Kinjo, H.; Maeshiro, M.; Uezato, E.; Yamamoto, T., “Adaptive Genetic Algorithm Observer and its Application to Trailer Truck Control System”, IEEE, SICE-ICASE International Joint Conference, Oct. 2006, 2 pgs.
J. Roh; H. Lee; W. Chung, “Control of a Car with a Trailer Using the Driver Assistance System”, IEEE, International Conference on Robotics and Biomimetics; Phuket, Thailand, Dec. 2011, 6 pages.
A. Gonzalez-Cantos; J.I. Maza; A. Ollero, “Design of a Stable Backing Up Fuzzy Control of Autonomous Articulated Vehicles for Factory Automation”, Dept. of Systems Engineering and Automatic Control, University of Seville, Spain, 2001, 5 pages.
Altafini, C.; Speranzon, A.; Wahlberg, B., “A Feedback Control Scheme for Reversing a Truck and Trailer Vehicle”, IEEE, Robotics and Automation, IEEE Transactions, Dec. 2001, vol. 17, No. 6, 2 pages.
Zare, A. Sharafi; M. Kamyad, A.V., “A New Approach in Intelligent Trailer Parking”, IEEE, 2010 2nd International Mechanical and Electrical Technology (ICMET), Sep. 2010, 1 page.
Tanaka, K.; Sano, M., “A Robust Stabilization Problem of Fuzzy Control Systems and its Application to Backing up Control of a Truck-trailer”, IEEE Transactions on Fuzzy Systems, May 1994, vol. 2, No. 2, 1 page.
Sharafi, M. Zare; A. Kamyad; A.V. Nikpoor, S., “Intelligent Parking Method for Truck in Presence of Fixed and Moving Obstacles and Trailer in Presence of Fixed Obstacles: Advanced Fuzzy Logic Technologies in Industrial Applications”, IEEE, 2010 International Electronics and Information Engineering (ICEIE), Aug. 2010, vol. 2, 1 page.
Hodo, D. W.; Hung, J.Y.; Bevly, D. M.; Millhouse, S., “Effects of Sensor Placement and Errors on Path Following Control of a Mobile Robot—Trailer System”, IEEE, American Control Conference, Jul. 2007, 1 page.
Sharafi, M. Zare; A. Kamyad; A.V. Nikpoor, S., “Intelligent Parking Method for Trailers in Presence of Fixed and Moving Obstacles”, IEEE, 2010 3rd International Conference on Advanced Computer Theory and Engineering (ICACTE), Aug. 2010, vol. 6, 1 page.
Chieh Chen; Tomizuka, M., “Steering and Independent Braking Control for Tractor-Semitrailer Vehicles in Automated Highway Systems”, IEEE, Proceedings of the 34th IEEE Conference on Decision and Control, Dec. 1995, vol. 2, 1 page.
P. Bolzern, R.M. Desantis, A. Locatelli, “An Input-Output Linearization Approach to the Control of an n-Body Articulated Vehicle”, J. Dyn. Sys., Meas., Control, Sep. 2001, vol. 123, No. 3, 3 pages.
Dieter Zöbel, David Polock, Philipp Wojke, “Steering Assistance for Backing Up Articulated Vehicles”, Systemics, Cybernetics and Informatics; vol. 1, No. 5, date unknown, 6 pages.
J.R. Billing; J.D. Patten; R.B. Madill, “Development of Configurations for Infrastructure-Friendly Five- and Six-Axle SemiTrailers”, National Research Council of Canada and Ontario Ministry of Transportation, date unknown, 11 pages.
Jesus Morales, Anthony Mandow, Jorge L. Martinez, and Alfonso Garcia-Cerezo, “Driver Assistance System for Backward Maneuvers in Passive Multi-Trailer Vehicles”, IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Oct. 2012, 7 pages.
Cedric Pradalier and Kane Usher, “Experiments in Autonomous Reversing of a Tractor-Trailer System”, 6th International Conference on Field and Service Robotics, inria-00195700, Version 1, Dec. 2007, 10 pages.
Andri Riid, Alar Leibak, Ennu Rüstern, “Fuzzy Backing Control of Truck and Two Trailers”, Tallinn University of Technology; Tallinn, Estonia, date unknown, 6 pages.
Jane McGrath, “How to Avoid Jackknifing”, A Discovery Company, date unknown, 3 pages.
Claudio Altafini, Alberto Speranzon, and Karl Henrik Johansson, “Hybrid Control of a Truck and Trailer Vehicle”, Springer-Verlag Berlin Heidelberg, HSCC 2002, LNCS 2289; 2002, 14 pages.
Jujnovich, B.; Roebuck, R.; Odhams, A.; David, C., “Implementation of Active Rear Steering of a Tractor Semitrailer”, Cambridge University, Engineering Department; Cambridge, United Kingdom, date unknown, 10 pages.
A.M.C. Odhams; R.L. Roebuck; C. Cebon, “Implementation of Active Steering on a Multiple Trailer Long Combination Vehicle”, Cambridge University, Engineering Department; Cambridge, United Kingdom, date unknown, 13 pages.
Cedric Pradalier and Kane Usher, “Robust Trajectory Tracking for a Reversing Tractor-Trailer System”, (Draft), Field and Service Robotics Conference, CSIRO ICT Centre, Jul. 2007, 16 pages.
Stahn, R.; Heiserich, G.; Stopp, A., “Laser Scanner-Based Navigation for Commercial Vehicles”, IEEE, 2007 IEEE Intelligent Vehicles Symposium, Jun. 2007, 1 page.
Lee Yong H.; Weiwen Deng; Chin Yuen-Kwok Steve; McKay Neil, “Feasibility Study for a Vehicle-Trailer Backing Up Control”, Refdoc.fr, SAE Transactions, vol. 113, No. 6, 2004, 1 page.
A.M.C. Odhams; R.L. Roebuck; B.A. Jujnovich; D. Cebon, “Active Steering of a Tractor-Semi-Trailer” Proceedings of the Institution of Mechanical Engineers, Part D: Journal of Automobile Engineering, SAGE Journals, vol. 225, No. 7, Jul. 2011, 1 page.
Haviland, G S, “Automatic Brake Control for Trucks—What Good Is It?”, TRID, Society of Automotive Engineers, Sep. 1968, 1 page.
William E. Travis; David W. Hodo; David M. Bevly; John Y. Hung, “UGV Trailer Position Estimation Using a Dynamic Base RTK System”, American Institute of Aeronautics and Astronautics, date unknown, 12 pages.
“VSE Electronic Trailer Steering”, ETS for Trailers, version 2009, VSE Trailer Systems B.V., 2009, 28 pages.
“Telematics Past, Present, and Future,” Automotive Service Association, www.ASAshop.org, May 2008, 20 pages.
“Fully Automatic Trailer Tow Hitch With LIN Bus,” https://webista.bmw.com/webista/show?id=1860575499&lang=engb&print=1, date unknown, 5 pages.
“VBOX Yaw Rate Sensor With Integral Accelerometers,” Racelogic, www.racelogic.co.uk, date unknown, 2 pages.
P.D.C.R Jayarathna; J.V Wijayakulasooriya; S.R Kodituwakku, “Fuzzy Logic and Neural Network Control Systems for Backing up a Truck and a Trailer”, International Journal of Latest Trends in Computing, vol. 2, No. 3, Sep. 2011, 8 pages.
Olof Enqvist, “AFS-Assisted Trailer Reversing,” Institutionen för systemteknik Deartment of Electrical Engineering, Jan. 27, 2006, 57 pages.
Novak, Domen; Dovzan, Dejan; Grebensek, Rok; Oblak, Simon, “Automated Parking System for a Truck and Trailer”, International Conference on Advances in the Internet, Processing, Systems and Interdisciplinary Research, Florence, 2007, WorldCat.org, 13 pgs.
SH. Azadi, H.R. Rezaei Nedamani, and R. Kazemi, “Automatic Parking of an Articulated Vehicle Using ANFIS”, Global Journal of Science, Engineering and Technology (ISSN: 2322-2441), 2013, pp. 93-104, Issue No. 14.
F. Cuesta and A. Ollero, “Intelligent System for Parallel Parking of Cars and Tractor-Trailers”, Intelligent Mobile Robot Navigation, STAR, 2005, pp. 159-188, Springer-Verlag Berlin Heidelberg.
Related Publications (1)
Number Date Country
20170174128 A1 Jun 2017 US