The present invention generally relates to trailer backup assist systems, and, more particularly, to trailer backup assist systems employing trailer angle detection through image processing.
Reversing a vehicle while towing a trailer can be challenging for many drivers, particularly for drivers that drive with a trailer on an infrequent basis or with various types of trailers. Some systems used to assist a driver in backing a trailer rely on trailer angle measurements to determine the position of the trailer relative to the vehicle. Thus, the accuracy and reliability of the trailer angle measurements can be critical to the operation of the trailer backup assist system.
According to one aspect of the present invention, a trailer angle identification system is disclosed. The system comprises an imaging device configured to capture an image. An angle sensor is configured to measure a first angle of the trailer relative to a vehicle. A controller is configured to process the image in a neural network and estimate a second angle of the trailer relative to the vehicle based on the image. The controller is further configured to train the neural network based on a difference between the first angle and the second angle.
The system may further comprise one or more of the following elements alone or in various combinations. The additional elements may include the following:
According to another aspect of the present invention, a method identifying a trailer angle between a vehicle and a trailer is disclosed. The method comprises capturing a plurality of images in a field of view and detecting a first angle with an angle sensor in connection the vehicle or trailer. The method further comprises processing the images in a neural network and estimating a second angle of the interface based on each of the images. The method further comprises training the neural network based on the first angle and the second angle for each of the images.
The method may further comprise one or more of the following steps alone or in various combinations. The additional steps of the method may include the following:
According to yet another aspect of the present invention, a trailer angle identification system is disclosed. The system comprises an imaging device configured to capture an image. An angle sensor is configured to measure a first angle of the trailer relative to a vehicle. A controller is configured to crop the image generating a cropped image based on a location of an interface between the vehicle and the trailer in the image. The controller is further configured to process the cropped image in a neural network and estimate a second angle of the trailer relative to the vehicle based on the image. The controller is further configured to train the neural network based on a difference between the first angle and the second angle, wherein the training is configured to train the neural network to identify an actual angle between the vehicle and the trailer based on the image without the first angle from the angle sensor. The system may further comprise the imaging device configured to capture the image in a field of view directed at a connection interface of the trailer to the vehicle.
These and other features, advantages, and objects of the present invention will be further understood and appreciated by those skilled in the art by reference to the following specification, claims, and appended drawings.
In the drawings:
For purposes of description herein, it is to be understood that the disclosed trailer backup assist system and the related methods may assume various alternative embodiments and orientations, except where expressly specified to the contrary. It is also to be understood that the specific devices and processes illustrated in the attached drawings, and described in the following specification, are simply exemplary embodiments of the inventive concepts defined in the appended claims. While various aspects of the trailer backup assist system and the related methods are described with reference to a particular illustrative embodiment, the disclosed invention is not limited to such embodiments, and additional modifications, applications, and embodiments may be implemented without departing from the disclosed invention. Hence, specific dimensions and other physical characteristics relating to the embodiments disclosed herein are not to be considered as limiting, unless the claims expressly state otherwise.
As used herein, the term “and/or,” when used in a list of two or more items, means that any one of the listed items can be employed by itself, or any combination of two or more of the listed items can be employed. For example, if a composition is described as containing components A, B, and/or C, the composition can contain A alone; B alone; C alone; A and B in combination; A and C in combination; B and C in combination; or A, B, and C in combination.
Referring to
In particular, the disclosure provides for the detection of the trailer angle γ based on the image data captured by an imaging device 14. Based on the image data captured by the imaging device 14, the system 8 may identify various characteristics of the trailer 10 based on a variety of image processing techniques (e.g. edge detection, background subtraction, template matching etc.). However, due to variations related in the trailer 10 and the local environment (e.g. shadows, textured surfaces, noise, etc.), conventional image processing techniques may not be sufficiently robust to reliably and accurately monitor the trailer angle γ without the aid of additional sensors.
As discussed herein, the improved systems and methods may utilize neural networks to improve the reliability and accuracy of the identified trailer angle γ to improve operation of the system 8. The neural networks and related methods may be configured to learn how to accurately detect the trailer angle γ without human intervention such that the resulting neural network may accurately identify the trailer angle solely based on image data acquired from the imaging device 14. Accordingly, the methods and systems discussed herein, may detect the trailer angle γ reliably without the aid of additional sensors, patterned markers or visual cues, or other aids that may otherwise be required to enable accurate operation of the system 8.
As further discussed in reference to
A drawback related to the implementation of deep convolution neural networks may include the labor intensive involvement from human operators. For example, in order to provide feedback to the network, a human operator may be required to label and review thousands of samples to ensure accurate learning and operation of the network. Accordingly, in addition to providing the application of neural networks to identify the trailer angle γ, the disclosure also provides for methods of programming and training the neural networks discussed herein. These and other aspects of the disclosure are further detailed in the following description.
Referring still to
As shown in
δ: steering angle at steered wheels 40 of the vehicle 12;
α: yaw angle of the vehicle 12;
β: yaw angle of the trailer 10;
γ: trailer angle between the vehicle 12 and the trailer 10 (γ=β−α);
W: wheelbase length between a front axle 42 and a rear axle 44 of the vehicle 12;
L: drawbar length between the hitch point 32 and the rear axle 44 of the vehicle 12; and
D: trailer length between the hitch point 32 and axle 20 of the trailer 10 or effective axle for multiple axle trailers.
It should be appreciated that additional embodiments of the trailer 10 may alternatively couple with the vehicle 12 to provide a pivoting connection, such as by connecting with a fifth wheel connector. It is also contemplated that additional embodiments of the trailer 10 may include more than one axle and may have various shapes and sizes configured for different loads and items, such as a boat trailer or a flatbed trailer.
In some embodiments, the trailer backup assist system 8 may also include the imaging device 14 located at the rear of the vehicle 12 and configured to image a rear-vehicle scene. The imaging device 14 may be centrally located at an upper region of a vehicle tailgate 46 such that the imaging device 14 is elevated relative to the tongue 26 of the trailer 10. The imaging device 14 has a field of view 48 located and oriented to capture one or more images that may include the tongue 26 of the trailer 10 and the hitch ball 30, among other things. Captured images or image data may be supplied to a controller 50 of the trailer backup assist system 8. As discussed herein, the image data may be processed by the controller 50 to determine the trailer angle γ between the vehicle 12 and the trailer 10.
Referring now to
In an exemplary embodiment, the controller 50 of the trailer backup assist system 8 may be configured to communicate with a variety of vehicle equipment. The trailer backup assist system 8 may include a vehicle sensor module 60 that monitors certain dynamics of the vehicle 12. The vehicle sensor module 60 may generate a plurality of signals that are communicated to the controller 50 and may include a vehicle speed signal generated by a speed sensor 62 and a vehicle yaw rate signal generated by a yaw rate sensor 64. A steering input device 66 may be provided to enable a driver to control or otherwise modify the desired curvature of the backing path of the trailer 10.
The steering input device 66 may be communicatively coupled to the controller 50 in a wired or wireless manner. In this configuration, steering input device 66 may provide the controller 50 with information defining the desired curvature of the backing path of the trailer 10. In response, the controller 50 may process the information and generate corresponding steering commands that are supplied to a power assist steering system 68 of the vehicle 12. In some embodiments, the steering input device 66 may comprise a rotatable knob 70 operable to rotate to positions that may correspond to an incremental change to the desired curvature of a backing path of the trailer 10.
According to some embodiments, the controller 50 of the trailer backup assist system 8 may control the power assist steering system 68 of the vehicle 12 to operate the steered wheels 40 to direct the vehicle 12 in such a manner that the trailer 10 reacts in accordance with the desired curvature of the backing path of the trailer 10. The power assist steering system 68 may be an electric power-assisted steering (EPAS) system that includes an electric steering motor 74 for turning the steered wheels 40 to a steering angle δ based on a steering command generated by the controller 50. In this configuration, the steering angle δ may be sensed by a steering angle sensor 76 of the power assist steering system 68 and provided to the controller 50. The steering command may be provided for autonomously steering the vehicle 12 during a backup maneuver and may alternatively be provided manually via a rotational position (e.g., a steering wheel angle) of the steering input device wheel 66 or the rotatable knob 70.
In some embodiments, the steering input device 66 (e.g. steering wheel) of the vehicle 12 may be mechanically coupled with the steered wheels 40 of the vehicle 12, such that the steering input device 66 may move in concert with steered wheels 40 via an internal torque, thereby preventing manual intervention with the steering input device 66 during autonomous steering of the vehicle 12. In such instances, the power assist steering system 68 may include a torque sensor 80 that senses torque (e.g., gripping and/or turning) on the steering input device 66, which may not be expected from autonomous control of the steering input device 66. Such unexpected torque may be detected by the controller 50 to indicate manual intervention by the driver. In some embodiments, external torque applied to the steering input device 66 may serve as a signal to the controller 50 that the driver has taken manual control and for the trailer backup assist system 8 to discontinue autonomous steering functionality.
The controller 50 of the trailer backup assist system 8 may also communicate with a vehicle brake control system 82 of the vehicle 12 to receive vehicle speed information, such as individual wheel speeds of the vehicle 12. Additionally or alternatively, vehicle speed information may be provided to the controller 50 by a powertrain control system 84 and/or the speed sensor 62, among other conceivable means. It is conceivable that individual wheel speeds may be used to determine a vehicle yaw rate, which can be provided to the controller 50, in the alternative or in addition, to the vehicle yaw rate measured by the yaw rate sensor 64 of the vehicle sensor module 60. In some embodiments, the controller 50 may provide braking commands to the vehicle brake control system 82, thereby allowing the trailer backup assist system 8 to regulate the speed of the vehicle 12 during a backup maneuver of the trailer 10. It should be appreciated that the controller 50 may additionally or alternatively regulate the speed of the vehicle 12 via interaction with the powertrain control system 84.
Through interaction with the power assist steering system 68, the vehicle brake control system 82, and/or the powertrain control system 84 of the vehicle 12, the potential for unacceptable trailer backup conditions can be reduced. Examples of unacceptable trailer backup conditions include, but are not limited to, a vehicle over-speed condition, a high trailer angle rate, trailer angle dynamic instability, a trailer jackknife condition, sensor failure, and the like. In such circumstances, the driver may be unaware of the failure until the unacceptable trailer backup condition is imminent or already happening. In order to avoid such conditions, the controller 50 may be configured to accurately monitor the trailer angle γ thereby providing feedback to ensure accurate operation.
According to some embodiments, the controller 50 may communicate with one or more devices, including a vehicle alert system 86, which may prompt visual, auditory, and tactile warnings. For instance, vehicle brake lights 88 and vehicle emergency flashers may provide a visual alert and a vehicle horn 90 and/or speaker 92 may provide an audible alert. Additionally, the controller 50 and/or vehicle alert system 86 may communicate with a human machine interface (HMI) 84 of the vehicle 12. The HMI 84 may include a touchscreen vehicle display 96, such as a center-stack mounted navigation or entertainment display capable of displaying images indicating the alert. Such an embodiment may be desirable to notify the driver of the vehicle 12 that an unacceptable trailer backup condition is occurring. Further, it is contemplated that the controller 50 may communicate via wireless communication with one or more electronic portable devices, such as portable electronic device 98, which is shown embodied as a smartphone. The portable electronic device 98 may include a display for displaying one or more images and other information to a user. In response, the portable electronic device 98 may provide feedback information, such as visual, audible, and tactile alerts.
Referring now to
In general, the trailer angle detection apparatus 102 may be utilized by the controller 50 to train the neural network of the hitch angle detection routine 56. Accordingly, the trailer angle detection apparatus 102 may only be required for initial training stages of the neural network in order to generate labels identifying the trailer angle γ for each image or at least a sample of images captured by the imaging device 14. Accordingly, the hitch angle detection routine 56 may be trained by the system 8 by utilizing the trailer angle γ detected by the trailer angle detection apparatus 102. However, once the neural network is trained such that the trailer angle γ can successfully be identified by the hitch angle detection routine 56 within an acceptable or predetermined level of error, the system 8 may be configured to utilize the neural network to detect the trailer angle γ by utilizing only the image data captured by the imaging device 14.
In an exemplary embodiment, the trailer angle detection apparatus 102 may comprise a housing 104 fixed to the hitch ball 30 on the vehicle 12. An element 106 attached to the trailer 10 may rotate relative to the housing 104 about an axis 108 defined by the hitch ball 30. A connecting member 110 may secure the element 106 to the trailer 10 for rotating the element 106 in conjunction with angular movement of the trailer 10. A sensor 112 may be configured to detect rotational movement of the element 106 for determining the trailer angle γ. It is contemplated that the element 106 in other embodiments may be alternatively secured to the trailer 10 to rotate the element 106 relative to the sensor 112 upon angular movement of the trailer 10.
In various embodiments, the sensor 112 may be referred to as the hitch angle sensor 112 and may be implemented by utilizing a variety of sensors. For example, the hitch angle sensor 112 may be implemented as a proximity sensor, a potentiometer, Hall Effect sensor, encoder, or various other forms of sensors that may be configured to measure the rotation of the trailer 10 relative to the vehicle 12. As shown in
Referring now to
In reference to the
Referring first to the training process 56a, the microprocessor 52 or, more generally, the controller 50 may first receive image data from the imaging device 14. The controller 50 may first process the image data via a pre-processing module 124. The pre-processing module 124 may be configured to crop each image frame received from the imaging device 14. The cropping of the image data may be consistently processed based on the positional relationship of the hitch ball 30 in the field of view 48. For example, the hitch ball 30 may be registered or identified during an initial connection or setup of the hitch ball 30 with the vehicle 12. Once identified, the controller 50 may be configured to crop the image data from the imaging device 14 based on predetermined extents or a portion of the image data designated in relation to the location of the hitch ball 30 in the field of view 48. In this way, the image data supplied to the neural network 122 may be limited in positional variation and quality that may be apparent in raw data received from the imaging device 14.
In some embodiments, the controller 50 may further be configured to process the data via an image augmentation module 126. The image augmentation module 126 may be configured to augment the image data by a variety of techniques. For example, the cropped image data received from the image pre-processing module 124 may be augmented by the image augmentation module 126 by various techniques including, but not limited to, flipping, rotating, translating, scaling, color enhancing, histogram stretching, noise filtering, selective noise inclusion, etc. Following processing of the image data via the image pre-processing module 124 and/or the image augmentation module 126, the controller 50 may utilize the trailer angle γ to label each frame of the image data via a trailer angle labeling module 128. Effectively, the trailer angle labeling module 128 may be implemented as a data attributed to each of the frames of the image data that may be input into the neural network 122. In this way, the training process 56a may provide for the image data from the imaging device 14 to be processed and input into the neural network 122 with the trailer angle γ of each image frame identified in order to train the parameters of the neural network 122 to accurately identify the trailer angle γ from only the image data.
Once the image data is received by the neural network 122, a deep learning procedure may be implemented to regress or estimate the trailer angle γ. For example, the neural network 122 may be implemented as a deep convolutional network. The architecture of the neural network 122 may be a plurality of convolutional networks followed by activation functions. To help avoiding overfitting, dropout layers and other regularization techniques may be implemented. In an exemplary embodiment, fully connected layers at the end of the neural network 122 are responsible identifying that outputting the trailer angle γ. Since the object of the neural network 122 may be to perform a regression task, an activation function may not be utilized at the output.
In general, the neural network 122 may comprise a plurality of neurons 130, which may be arranged in a three-dimensional array comprising a width, a depth, and a height. The arrangement of the neurons 130 in this configuration may provide for each layer (e.g. dimensional cross-section of the array) to be connected to a small portion of the preceding layer. In this way, the network 122 may process the data through regression to reduce each image to a single vector to identify the trailer angle γ. In this way, the neural network 122 may transform each frame of the image data layer by layer from original pixel values to the final output. In general, the specific architecture of the neural network 122 may vary and as may be understood by those having ordinary skill in the art, the training process 56a may begin with a pre-trained model. In this way, the training process 56a may be utilized to fine-tune the pre-trained, convolutional neural network to accurately detect the trailer angle γ from the image data captured by the imaging device 14. Examples of pre-trained models that may be implemented for the training process 56a may include, but are not limited to, the following: LeNet, AlexNet, ZF Net, GoogLeNet, VGGNet, ResNet, etc.
Referring now to
As previously discussed, the neural network 122 may be configured to receive the image data from the pre-processing module 124 and/or the augmentation module 126. Upon completion of the analysis of each frame of the image data, the controller 50 may output the corresponding trailer angle γ as shown in
Referring now to
As shown,
For example, the controller 50 may be configured to group image data based on the error in order to identify trailer angles γ or environmental conditions that lead to the error between the detected trailer vector 150a and the measured trailer vector 150b. In this way, the system may be configured to detect ranges of trailer angles and categories of lighting/environmental conditions that lead to inaccurate identification of the detected trailer vector 150a. Based on the error identified in the categories and/or ranges, the system 8 may identify image data that needs to be captured to improve the training of the neural network 122. Accordingly, the system 8 may be configured to utilize additional image samples and test images to improve the accuracy of the detection of the trailer angle γ to apply deep learning to complete the training process 56a of the neural network 122.
Referring now to
It is to be understood that variations and modifications can be made on the aforementioned structures and methods without departing from the concepts of the present invention, and further it is to be understood that such concepts are intended to be covered by the following claims unless these claims by their language expressly state otherwise.
It will be understood by one having ordinary skill in the art that construction of the described device and other components is not limited to any specific material. Other exemplary embodiments of the device disclosed herein may be formed from a wide variety of materials, unless described otherwise herein.
For purposes of this disclosure, the term “coupled” (in all of its forms, couple, coupling, coupled, etc.) generally means the joining of two components (electrical or mechanical) directly or indirectly to one another. Such joining may be stationary in nature or movable in nature. Such joining may be achieved with the two components (electrical or mechanical) and any additional intermediate members being integrally formed as a single unitary body with one another or with the two components. Such joining may be permanent in nature or may be removable or releasable in nature unless otherwise stated.
It is also important to note that the construction and arrangement of the elements of the device as shown in the exemplary embodiments is illustrative only. Although only a few embodiments of the present innovations have been described in detail in this disclosure, those skilled in the art who review this disclosure will readily appreciate that many modifications are possible (e.g., variations in sizes, dimensions, structures, shapes and proportions of the various elements, values of parameters, mounting arrangements, use of materials, colors, orientations, etc.) without materially departing from the novel teachings and advantages of the subject matter recited. For example, elements shown as integrally formed may be constructed of multiple parts or elements shown as multiple parts may be integrally formed, the operation of the interfaces may be reversed or otherwise varied, the length or width of the structures and/or members or connector or other elements of the system may be varied, the nature or number of adjustment positions provided between the elements may be varied. It should be noted that the elements and/or assemblies of the system may be constructed from any of a wide variety of materials that provide sufficient strength or durability, in any of a wide variety of colors, textures, and combinations. Accordingly, all such modifications are intended to be included within the scope of the present innovations. Other substitutions, modifications, changes, and omissions may be made in the design, operating conditions, and arrangement of the desired and other exemplary embodiments without departing from the spirit of the present innovations.
It will be understood that any described processes or steps within described processes may be combined with other disclosed processes or steps to form structures within the scope of the present device. The exemplary structures and processes disclosed herein are for illustrative purposes and are not to be construed as limiting.
It is also to be understood that variations and modifications can be made on the aforementioned structures and methods without departing from the concepts of the present device, and further it is to be understood that such concepts are intended to be covered by the following claims unless these claims by their language expressly state otherwise.
The above description is considered that of the illustrated embodiments only. Modifications of the device will occur to those skilled in the art and to those who make or use the device. Therefore, it is understood that the embodiments shown in the drawings and described above is merely for illustrative purposes and not intended to limit the scope of the device, which is defined by the following claims as interpreted according to the principles of patent law, including the Doctrine of Equivalents.
Number | Name | Date | Kind |
---|---|---|---|
3542390 | Fikes et al. | Nov 1970 | A |
3605088 | Savelli | Sep 1971 | A |
3787077 | Sanders | Jan 1974 | A |
3833928 | Gavit et al. | Sep 1974 | A |
3860257 | Mesley | Jan 1975 | A |
4040006 | Kimmel | Aug 1977 | A |
4042132 | Bohman et al. | Aug 1977 | A |
4122390 | Kollitz et al. | Oct 1978 | A |
4212483 | Howard | Jul 1980 | A |
4366966 | Ratsko et al. | Jan 1983 | A |
4735432 | Brown | Apr 1988 | A |
4752080 | Rogers | Jun 1988 | A |
4848449 | Martinet et al. | Jul 1989 | A |
4852901 | Beasley et al. | Aug 1989 | A |
4943080 | Reimer | Jul 1990 | A |
5001639 | Breen | Mar 1991 | A |
5056905 | Jensen | Oct 1991 | A |
5097250 | Hernandez | Mar 1992 | A |
5108123 | Rubenzik | Apr 1992 | A |
5108158 | Breen | Apr 1992 | A |
5132851 | Bomar et al. | Jul 1992 | A |
5152544 | Dierker, Jr. et al. | Oct 1992 | A |
5191328 | Nelson | Mar 1993 | A |
5244226 | Bergh | Sep 1993 | A |
5246242 | Penzotti | Sep 1993 | A |
5247442 | Kendall | Sep 1993 | A |
5282641 | McLaughlin | Feb 1994 | A |
5289892 | Notsu | Mar 1994 | A |
5290057 | Pellerito | Mar 1994 | A |
5455557 | Noll et al. | Oct 1995 | A |
5521633 | Nakajima et al. | May 1996 | A |
5523947 | Breen | Jun 1996 | A |
5541778 | DeFlorio | Jul 1996 | A |
5558350 | Kimbrough et al. | Sep 1996 | A |
5559696 | Borenstein | Sep 1996 | A |
5579228 | Kimbrough et al. | Nov 1996 | A |
5631656 | Hartman et al. | May 1997 | A |
5650764 | McCullough | Jul 1997 | A |
5690347 | Juergens et al. | Nov 1997 | A |
5719713 | Brown | Feb 1998 | A |
5747683 | Gerum et al. | May 1998 | A |
5821852 | Fairchild | Oct 1998 | A |
5980048 | Rannells, Jr. et al. | Nov 1999 | A |
6041582 | Tiede et al. | Mar 2000 | A |
6042196 | Nakamura et al. | Mar 2000 | A |
6124709 | Allwine | Sep 2000 | A |
6151175 | Osha | Nov 2000 | A |
6198992 | Winslow | Mar 2001 | B1 |
6217177 | Rost | Apr 2001 | B1 |
6218828 | Bates et al. | Apr 2001 | B1 |
6223104 | Kamen et al. | Apr 2001 | B1 |
6223114 | Boros et al. | Apr 2001 | B1 |
6268800 | Howard | Jul 2001 | B1 |
6292094 | Deng et al. | Sep 2001 | B1 |
6351698 | Kubota et al. | Feb 2002 | B1 |
6472865 | Tola et al. | Oct 2002 | B1 |
6480104 | Wall et al. | Nov 2002 | B1 |
6483429 | Yasui et al. | Nov 2002 | B1 |
6494476 | Masters et al. | Dec 2002 | B2 |
6498977 | Wetzel et al. | Dec 2002 | B2 |
6539288 | Ishida et al. | Mar 2003 | B2 |
6568093 | Kogiso et al. | May 2003 | B2 |
6577952 | Geier et al. | Jun 2003 | B2 |
6668225 | Oh et al. | Dec 2003 | B2 |
6704653 | Kuriya et al. | Mar 2004 | B2 |
6712378 | Austin | Mar 2004 | B1 |
6801125 | McGregor et al. | Oct 2004 | B1 |
6806809 | Lee et al. | Oct 2004 | B2 |
6820888 | Griffin | Nov 2004 | B1 |
6838979 | Deng et al. | Jan 2005 | B2 |
6854557 | Deng et al. | Feb 2005 | B1 |
6857494 | Kobayashi et al. | Feb 2005 | B2 |
6879240 | Kruse | Apr 2005 | B2 |
6956468 | Lee et al. | Oct 2005 | B2 |
6959970 | Tseng | Nov 2005 | B2 |
6999856 | Lee et al. | Feb 2006 | B2 |
7005974 | McMahon et al. | Feb 2006 | B2 |
7006127 | Mizusawa et al. | Feb 2006 | B2 |
7008088 | Pisciotti | Mar 2006 | B2 |
7028804 | Eki et al. | Apr 2006 | B2 |
7032705 | Zheng et al. | Apr 2006 | B2 |
7046127 | Boddy | May 2006 | B2 |
7058493 | Inagaki | Jun 2006 | B2 |
7089101 | Fischer et al. | Aug 2006 | B2 |
7154385 | Lee et al. | Dec 2006 | B2 |
7159890 | Craig et al. | Jan 2007 | B2 |
7167785 | Lohberg et al. | Jan 2007 | B2 |
7170285 | Spratte | Jan 2007 | B2 |
7171330 | Kruse et al. | Jan 2007 | B2 |
7204504 | Gehring et al. | Apr 2007 | B2 |
7219913 | Atley | May 2007 | B2 |
7225891 | Gehring et al. | Jun 2007 | B2 |
7229139 | Lu et al. | Jun 2007 | B2 |
7239958 | Grougan et al. | Jul 2007 | B2 |
7269489 | Deng et al. | Sep 2007 | B2 |
7272481 | Einig et al. | Sep 2007 | B2 |
7295907 | Lu et al. | Nov 2007 | B2 |
7401871 | Lu et al. | Jul 2008 | B2 |
7405557 | Spratte et al. | Jul 2008 | B2 |
7413266 | Lenz et al. | Aug 2008 | B2 |
7425889 | Widmann et al. | Sep 2008 | B2 |
7447585 | Tandy, Jr. et al. | Nov 2008 | B2 |
7451020 | Goetting et al. | Nov 2008 | B2 |
7463137 | Wishart et al. | Dec 2008 | B2 |
7504995 | Lawrence et al. | Mar 2009 | B2 |
7532109 | Takahama et al. | May 2009 | B2 |
7540523 | Russell et al. | Jun 2009 | B2 |
7548155 | Schutt et al. | Jun 2009 | B2 |
7568716 | Dietz | Aug 2009 | B2 |
7623952 | Unruh et al. | Nov 2009 | B2 |
7648153 | Metternich et al. | Jan 2010 | B2 |
7690737 | Lu | Apr 2010 | B2 |
7715953 | Shepard | May 2010 | B2 |
7731302 | Tandy, Jr. et al. | Jun 2010 | B2 |
7793965 | Padula | Sep 2010 | B2 |
7798263 | Tandy, Jr. et al. | Sep 2010 | B2 |
7878545 | Rhymer et al. | Feb 2011 | B2 |
7904222 | Lee et al. | Mar 2011 | B2 |
7905507 | Perri | Mar 2011 | B2 |
7950751 | Offerle et al. | May 2011 | B2 |
7953536 | Katrak | May 2011 | B2 |
7974444 | Hongo | Jul 2011 | B2 |
8010252 | Getman et al. | Aug 2011 | B2 |
8010253 | Lundquist | Aug 2011 | B2 |
8036792 | Dechamp | Oct 2011 | B2 |
8038166 | Piesinger | Oct 2011 | B1 |
8044779 | Hahn et al. | Oct 2011 | B2 |
8073594 | Lee et al. | Dec 2011 | B2 |
8157284 | McGhie et al. | Apr 2012 | B1 |
8165770 | Getman et al. | Apr 2012 | B2 |
8167444 | Lee et al. | May 2012 | B2 |
8170726 | Chen et al. | May 2012 | B2 |
8174576 | Akatsuka et al. | May 2012 | B2 |
8180543 | Futamura et al. | May 2012 | B2 |
8190364 | Rekow | May 2012 | B2 |
8191915 | Freese, V et al. | Jun 2012 | B2 |
8192036 | Lee et al. | Jun 2012 | B2 |
8215436 | DeGrave et al. | Jul 2012 | B2 |
8223204 | Hahn | Jul 2012 | B2 |
8244442 | Craig et al. | Aug 2012 | B2 |
8260518 | Englert | Sep 2012 | B2 |
8267485 | Barlsen et al. | Sep 2012 | B2 |
8280607 | Gatti et al. | Oct 2012 | B2 |
8308182 | Ortmann et al. | Nov 2012 | B2 |
8326504 | Wu et al. | Dec 2012 | B2 |
8342560 | Albers et al. | Jan 2013 | B2 |
8380390 | Sy et al. | Feb 2013 | B2 |
8380416 | Offerle et al. | Feb 2013 | B2 |
8393632 | Vortmeyer et al. | Mar 2013 | B2 |
8401744 | Chiocco | Mar 2013 | B2 |
8427288 | Schofield et al. | Apr 2013 | B2 |
8469125 | Yu et al. | Jun 2013 | B2 |
8504243 | Kageyama | Aug 2013 | B2 |
8548680 | Ryerson et al. | Oct 2013 | B2 |
8548683 | Cebon et al. | Oct 2013 | B2 |
8576115 | Basten | Nov 2013 | B2 |
8626382 | Obradovich | Jan 2014 | B2 |
8675953 | Elwell et al. | Mar 2014 | B1 |
8755984 | Rupp et al. | Jun 2014 | B2 |
8807261 | Subrt et al. | Aug 2014 | B2 |
8825328 | Rupp et al. | Sep 2014 | B2 |
8833789 | Anderson | Sep 2014 | B2 |
8886400 | Kossira et al. | Nov 2014 | B2 |
8888120 | Trevino | Nov 2014 | B2 |
8909426 | Rhode et al. | Dec 2014 | B2 |
8930140 | Trombley et al. | Jan 2015 | B2 |
8939462 | Adamczyk et al. | Jan 2015 | B2 |
8955865 | Fortin et al. | Feb 2015 | B2 |
8972109 | Lavoie et al. | Mar 2015 | B2 |
9008913 | Sears et al. | Apr 2015 | B1 |
9026311 | Pieronek et al. | May 2015 | B1 |
9042603 | Elwart et al. | May 2015 | B2 |
9082315 | Lin et al. | Jul 2015 | B2 |
9108598 | Headley | Aug 2015 | B2 |
9114832 | Wang et al. | Aug 2015 | B2 |
9120358 | Molls et al. | Sep 2015 | B2 |
9120359 | Chiu et al. | Sep 2015 | B2 |
9156496 | Greenwood et al. | Oct 2015 | B2 |
9164955 | Lavoie et al. | Oct 2015 | B2 |
9180890 | Lu et al. | Nov 2015 | B2 |
9187124 | Trombley et al. | Nov 2015 | B2 |
9227474 | Liu | Jan 2016 | B2 |
9238483 | Hafner et al. | Jan 2016 | B2 |
9248858 | Lavoie et al. | Feb 2016 | B2 |
9296422 | Lavoie | Mar 2016 | B2 |
9315212 | Kyrtsos et al. | Apr 2016 | B1 |
9321483 | Headley | Apr 2016 | B2 |
9335162 | Kyrtsos et al. | May 2016 | B2 |
9340228 | Xu et al. | May 2016 | B2 |
9393996 | Goswami et al. | Jul 2016 | B2 |
9428188 | Schwindt et al. | Aug 2016 | B2 |
9434414 | Lavoie | Sep 2016 | B2 |
9437055 | Kuehnie et al. | Sep 2016 | B2 |
9500497 | Lavoie | Nov 2016 | B2 |
9610974 | Herzog et al. | Apr 2017 | B2 |
9616923 | Lavoie | Apr 2017 | B2 |
9623904 | Lavoie et al. | Apr 2017 | B2 |
9676377 | Hafner et al. | Jun 2017 | B2 |
9731568 | Wuergler et al. | Aug 2017 | B2 |
9798953 | Hu | Oct 2017 | B2 |
9802542 | Lu et al. | Oct 2017 | B2 |
9827818 | Hu et al. | Nov 2017 | B2 |
9836060 | Ghneim et al. | Dec 2017 | B2 |
9840278 | Lavoie et al. | Dec 2017 | B2 |
10046800 | Hu et al. | Aug 2018 | B2 |
10169678 | Sachdeva | Jan 2019 | B1 |
20010024333 | Rost | Sep 2001 | A1 |
20010037164 | Hecker | Nov 2001 | A1 |
20020128764 | Hecker et al. | Sep 2002 | A1 |
20040017285 | Zielinski et al. | Jan 2004 | A1 |
20040021291 | Haug et al. | Feb 2004 | A1 |
20040093139 | Wildey et al. | May 2004 | A1 |
20040130441 | Lee et al. | Jul 2004 | A1 |
20040222881 | Deng et al. | Nov 2004 | A1 |
20050000738 | Gehring et al. | Jan 2005 | A1 |
20050128059 | Vause | Jun 2005 | A1 |
20050206224 | Lu | Sep 2005 | A1 |
20050206225 | Offerle et al. | Sep 2005 | A1 |
20050206229 | Lu et al. | Sep 2005 | A1 |
20050206231 | Lu et al. | Sep 2005 | A1 |
20050236201 | Spannheimer et al. | Oct 2005 | A1 |
20050236896 | Offerle et al. | Oct 2005 | A1 |
20060041358 | Hara | Feb 2006 | A1 |
20060071447 | Gehring et al. | Apr 2006 | A1 |
20060076828 | Lu et al. | Apr 2006 | A1 |
20060103511 | Lee et al. | May 2006 | A1 |
20060111820 | Goetting et al. | May 2006 | A1 |
20060142936 | Dix | Jun 2006 | A1 |
20060155455 | Lucas et al. | Jul 2006 | A1 |
20060244579 | Raab | Nov 2006 | A1 |
20070027581 | Bauer et al. | Feb 2007 | A1 |
20070090688 | Haemmerling et al. | Apr 2007 | A1 |
20070132560 | Nystrom et al. | Jun 2007 | A1 |
20070152424 | Deng et al. | Jul 2007 | A1 |
20070285808 | Beale | Dec 2007 | A1 |
20080143593 | Graziano et al. | Jun 2008 | A1 |
20080147277 | Lu et al. | Jun 2008 | A1 |
20080231701 | Greenwood et al. | Sep 2008 | A1 |
20080312792 | Dechamp | Dec 2008 | A1 |
20090005932 | Lee et al. | Jan 2009 | A1 |
20090079828 | Lee et al. | Mar 2009 | A1 |
20090085775 | Otsuka et al. | Apr 2009 | A1 |
20090093928 | Getman et al. | Apr 2009 | A1 |
20090198425 | Englert | Aug 2009 | A1 |
20090228182 | Waldbauer et al. | Sep 2009 | A1 |
20090248346 | Fennel et al. | Oct 2009 | A1 |
20090300701 | Karaoguz et al. | Dec 2009 | A1 |
20090306861 | Schumann et al. | Dec 2009 | A1 |
20100063702 | Sabelstrom et al. | Mar 2010 | A1 |
20100171828 | Ishii | Jul 2010 | A1 |
20100332049 | Sy et al. | Dec 2010 | A1 |
20110001825 | Hahn | Jan 2011 | A1 |
20110018231 | Collenberg | Jan 2011 | A1 |
20110022282 | Wu et al. | Jan 2011 | A1 |
20110087398 | Lu et al. | Apr 2011 | A1 |
20110112721 | Wang et al. | May 2011 | A1 |
20110125457 | Lee et al. | May 2011 | A1 |
20110160956 | Chung et al. | Jun 2011 | A1 |
20110257860 | Getman et al. | Oct 2011 | A1 |
20120041658 | Turner | Feb 2012 | A1 |
20120086808 | Lynam et al. | Apr 2012 | A1 |
20120095649 | Klier et al. | Apr 2012 | A1 |
20120109471 | Wu | May 2012 | A1 |
20120112434 | Albers et al. | May 2012 | A1 |
20120185131 | Headley | Jul 2012 | A1 |
20120200706 | Greenwood et al. | Aug 2012 | A1 |
20120271512 | Rupp et al. | Oct 2012 | A1 |
20120271514 | Lavoie et al. | Oct 2012 | A1 |
20120271515 | Rhode et al. | Oct 2012 | A1 |
20120271522 | Rupp et al. | Oct 2012 | A1 |
20120283909 | Dix | Nov 2012 | A1 |
20120283910 | Lee et al. | Nov 2012 | A1 |
20120310594 | Watanabe | Dec 2012 | A1 |
20130006472 | McClain et al. | Jan 2013 | A1 |
20130024064 | Shepard | Jan 2013 | A1 |
20130027195 | Van Wiemeersch et al. | Jan 2013 | A1 |
20130082453 | Padula | Apr 2013 | A1 |
20130158863 | Skvarce et al. | Jun 2013 | A1 |
20130179038 | Goswami et al. | Jul 2013 | A1 |
20130207834 | Mizutani et al. | Aug 2013 | A1 |
20130226390 | Luo et al. | Aug 2013 | A1 |
20130250114 | Lu | Sep 2013 | A1 |
20130261843 | Kossira et al. | Oct 2013 | A1 |
20130268160 | Trombley et al. | Oct 2013 | A1 |
20140005918 | Qiang | Jan 2014 | A1 |
20140025260 | McClure | Jan 2014 | A1 |
20140052337 | Lavoie et al. | Feb 2014 | A1 |
20140058614 | Trombley et al. | Feb 2014 | A1 |
20140058622 | Trombley et al. | Feb 2014 | A1 |
20140058655 | Trombley et al. | Feb 2014 | A1 |
20140058668 | Trombley et al. | Feb 2014 | A1 |
20140067154 | Yu et al. | Mar 2014 | A1 |
20140067155 | Yu et al. | Mar 2014 | A1 |
20140085472 | Lu et al. | Mar 2014 | A1 |
20140088824 | Ishimoto | Mar 2014 | A1 |
20140160276 | Pliefke et al. | Jun 2014 | A1 |
20140172232 | Rupp et al. | Jun 2014 | A1 |
20140183841 | Jones | Jul 2014 | A1 |
20140188344 | Lavoie | Jul 2014 | A1 |
20140188346 | Lavoie | Jul 2014 | A1 |
20140210456 | Crossman | Jul 2014 | A1 |
20140218506 | Trombley et al. | Aug 2014 | A1 |
20140218522 | Lavoie et al. | Aug 2014 | A1 |
20140222288 | Lavoie et al. | Aug 2014 | A1 |
20140236532 | Trombley et al. | Aug 2014 | A1 |
20140249691 | Hafner et al. | Sep 2014 | A1 |
20140267688 | Aich | Sep 2014 | A1 |
20140267689 | Lavoie | Sep 2014 | A1 |
20140277941 | Chiu et al. | Sep 2014 | A1 |
20140277942 | Kyrtsos et al. | Sep 2014 | A1 |
20140297128 | Lavoie et al. | Oct 2014 | A1 |
20140297129 | Lavoie et al. | Oct 2014 | A1 |
20140303847 | Lavoie | Oct 2014 | A1 |
20140309888 | Smit et al. | Oct 2014 | A1 |
20140324295 | Lavoie | Oct 2014 | A1 |
20140343795 | Lavoie | Nov 2014 | A1 |
20140358429 | Shutko et al. | Dec 2014 | A1 |
20140379217 | Rupp et al. | Dec 2014 | A1 |
20150002670 | Bajpai | Jan 2015 | A1 |
20150035256 | Klank et al. | Feb 2015 | A1 |
20150057903 | Rhode et al. | Feb 2015 | A1 |
20150066296 | Trombley et al. | Mar 2015 | A1 |
20150066298 | Sharma et al. | Mar 2015 | A1 |
20150105975 | Dunn | Apr 2015 | A1 |
20150115571 | Zhang et al. | Apr 2015 | A1 |
20150120141 | Lavoie et al. | Apr 2015 | A1 |
20150120143 | Schlichting | Apr 2015 | A1 |
20150134183 | Lavoie et al. | May 2015 | A1 |
20150138340 | Lavoie | May 2015 | A1 |
20150149040 | Hueger et al. | May 2015 | A1 |
20150158527 | Hafner et al. | Jun 2015 | A1 |
20150165850 | Chiu et al. | Jun 2015 | A1 |
20150197278 | Boos et al. | Jul 2015 | A1 |
20150203156 | Hafner et al. | Jul 2015 | A1 |
20150210254 | Pieronek et al. | Jul 2015 | A1 |
20150210317 | Hafner et al. | Jul 2015 | A1 |
20150217693 | Pliefke et al. | Aug 2015 | A1 |
20150269444 | Lameyre et al. | Sep 2015 | A1 |
20160001705 | Greenwood et al. | Jan 2016 | A1 |
20160009288 | Yu | Jan 2016 | A1 |
20160039456 | Lavoie | Feb 2016 | A1 |
20160059780 | Lavoie | Mar 2016 | A1 |
20160059888 | Bradley et al. | Mar 2016 | A1 |
20160059889 | Herzog et al. | Mar 2016 | A1 |
20160096549 | Herzog et al. | Apr 2016 | A1 |
20160129939 | Singh et al. | May 2016 | A1 |
20160152263 | Singh et al. | Jun 2016 | A1 |
20160153778 | Singh et al. | Jun 2016 | A1 |
20160229452 | Lavoie et al. | Aug 2016 | A1 |
20160280267 | Lavoie et al. | Sep 2016 | A1 |
20170043806 | Muharemovic | Feb 2017 | A1 |
20170073005 | Ghneim et al. | Mar 2017 | A1 |
20170101130 | Lavoie | Apr 2017 | A1 |
20170106796 | Lavoie et al. | Apr 2017 | A1 |
20170174130 | Hu et al. | Jun 2017 | A1 |
20170278014 | Lessmann | Sep 2017 | A1 |
20170297619 | Lavoie et al. | Oct 2017 | A1 |
20170297620 | Lavoie et al. | Oct 2017 | A1 |
20170313351 | Lavoie | Nov 2017 | A1 |
20180012411 | Richey | Jan 2018 | A1 |
20180251153 | Li et al. | Sep 2018 | A1 |
20180253608 | Diessner et al. | Sep 2018 | A1 |
20180276839 | Diessner | Sep 2018 | A1 |
20190016264 | Potnis | Jan 2019 | A1 |
20190016382 | Zarco | Jan 2019 | A1 |
20190039633 | Li | Feb 2019 | A1 |
20190122384 | Ertle | Apr 2019 | A1 |
20190180502 | Englard | Jun 2019 | A1 |
20190279366 | Sick | Sep 2019 | A1 |
20190289282 | Briggs | Sep 2019 | A1 |
20190294966 | Khan | Sep 2019 | A1 |
20200130582 | Wong | Apr 2020 | A1 |
Number | Date | Country |
---|---|---|
102582686 | Sep 2013 | CN |
106250893 | Dec 2016 | CN |
3923676 | Jan 1991 | DE |
3931518 | Apr 1991 | DE |
9208595 | Aug 1992 | DE |
19526702 | Feb 1997 | DE |
10030738 | Aug 2001 | DE |
10031244 | Jan 2002 | DE |
10065230 | Jul 2002 | DE |
10122562 | Jul 2002 | DE |
10154612 | May 2003 | DE |
10312548 | May 2004 | DE |
10333998 | Feb 2005 | DE |
102004050149 | Apr 2006 | DE |
102005042957 | Mar 2007 | DE |
102005043466 | Mar 2007 | DE |
102005043467 | Mar 2007 | DE |
102005043468 | Mar 2007 | DE |
102006002294 | Jul 2007 | DE |
102006048947 | Apr 2008 | DE |
102006056408 | Jun 2008 | DE |
102008020838 | Nov 2008 | DE |
102007029413 | Jan 2009 | DE |
102008045436 | Mar 2010 | DE |
102006035021 | Apr 2010 | DE |
102008043675 | May 2010 | DE |
102009007990 | Aug 2010 | DE |
102009012253 | Sep 2010 | DE |
102009027041 | Dec 2010 | DE |
102009038552 | Feb 2011 | DE |
102010006323 | Aug 2011 | DE |
102008004158 | Oct 2011 | DE |
102008004159 | Oct 2011 | DE |
102008004160 | Oct 2011 | DE |
102010021052 | Nov 2011 | DE |
102010029184 | Nov 2011 | DE |
102010045519 | Mar 2012 | DE |
102011104256 | Jul 2012 | DE |
102011101990 | Oct 2012 | DE |
102012005707 | Oct 2012 | DE |
202012010517 | Dec 2012 | DE |
102011108440 | Jan 2013 | DE |
102011120814 | Jun 2013 | DE |
102012006206 | Oct 2013 | DE |
102012206133 | Oct 2013 | DE |
102012019234 | Apr 2014 | DE |
102013000198 | Jul 2014 | DE |
0418653 | Mar 1991 | EP |
0433858 | Jun 1991 | EP |
1361543 | Nov 2003 | EP |
1593552 | Mar 2007 | EP |
1810913 | Jul 2007 | EP |
2388180 | Nov 2011 | EP |
2452549 | May 2012 | EP |
2487454 | Aug 2012 | EP |
2551132 | Jan 2013 | EP |
2644477 | Oct 2013 | EP |
2803944 | Nov 2014 | EP |
2515379 | Apr 1983 | FR |
2265587 | Oct 1993 | GB |
2342630 | Apr 2000 | GB |
2398048 | Aug 2004 | GB |
2398049 | Aug 2004 | GB |
2398050 | Aug 2004 | GB |
61006458 | Jan 1986 | JP |
6159491 | Mar 1986 | JP |
6385568 | Jun 1988 | JP |
01095980 | Apr 1989 | JP |
01095981 | Apr 1989 | JP |
09267762 | Oct 1997 | JP |
09328078 | Dec 1997 | JP |
10001063 | Jan 1998 | JP |
11124051 | May 1999 | JP |
11278319 | Oct 1999 | JP |
2002012172 | Jan 2002 | JP |
2002068032 | Mar 2002 | JP |
2003034261 | Feb 2003 | JP |
2003148938 | May 2003 | JP |
3716722 | Nov 2005 | JP |
2008027138 | Feb 2008 | JP |
2012105158 | May 2012 | JP |
2012166647 | Sep 2012 | JP |
2014002056 | Jan 2014 | JP |
8503263 | Aug 1985 | WO |
0044605 | Aug 2000 | WO |
2005005200 | Jan 2005 | WO |
2005116688 | Dec 2005 | WO |
2006042665 | Apr 2006 | WO |
2012059207 | May 2012 | WO |
2012103193 | Aug 2012 | WO |
2014019730 | Feb 2014 | WO |
2014037500 | Mar 2014 | WO |
2014070047 | May 2014 | WO |
2014092611 | Jun 2014 | WO |
2014123575 | Aug 2014 | WO |
2015074027 | May 2015 | WO |
2015187467 | Dec 2015 | WO |
2017123880 | Jul 2017 | WO |
Entry |
---|
Novak, Domen; Dovzan, Dejan; Grebensek, Rok; Oblak, Simon, “Automated Parking System for a Truck and Trailer”, International Conference on Advances in the Internet, Processing, Systems and Interdisciplinary Research, Florence, 2007, WorldCat.org, 13 pgs. |
Commonly assigned U.S. Appl. No. 16/294,540, filed Mar. 6, 2019, entitled Trailer Angle Detection Using End-to-End Learning (FOR025 P2142). |
“Ford Super Duty: Truck Technologies”, Brochure, Sep. 2011, 2 pages. |
Kristopher Bunker, “2012 Guide to Towing”, Trailer Life, 2012, 38 pages. |
A. Gonzalez-Cantos, “Backing-Up Maneuvers of Autonomous Tractor-Trailer Vehicles using the Qualitative Theory of Nonlinear Dynamical Systems,” International Journal of Robotics Research, Jan. 2009, vol. 28, 1 page. |
L. Chu, Y. Fang, M. Shang, J. Guo, F. Zhou, “Estimation of Articulation Angle for Tractor Semi-Trailer Based on State Observer”, ACM Digital Library, ICMTMA '10 Proceedings of the 2010 International Conference on Measuring Technology and Automation, vol. 2, Mar. 2010, 1 page. |
M. Wagner, D. Zoebel, and A. Meroth, “Adaptive Software and Systems Architecture for Driver Assistance Systems” International Journal of Machine Learning and Computing, Oct. 2011, vol. 1, No. 4, 7 pages. |
F.W. Kienhöfer; D. Cebon, “An Investigation of ABS Strategies for Articulated Vehicles”, Cambridge University, Engineering Department, United Kingdom, date unknown, 13 pages. |
C. Lundquist; W. Reinelt; O. Enqvist, “Back Driving Assistant for Passenger Cars with Trailer”, ZF Lenksysteme GmbH, Schwäbisch Gmünd, Germany, 2006 (SAE Int'l) Jan. 2006, 8 pages. |
Zhe Leng; Minor, M., “A Simple Tractor-Trailer Backing Control Law for Path Following”, IEEE, Intelligent Robots and Systems (IROS) IEEE/RSJ International Conference, Oct. 2010, 2 pages. |
Kinjo, H.; Maeshiro, M.; Uezato, E.; Yamamoto, T., “Adaptive Genetic Algorithm Observer and its Application to Trailer Truck Control System”, IEEE, SICE-ICASE International Joint Conference, Oct. 2006, 2 pgs. |
J. Roh; H. Lee; W. Chung, “Control of a Car with a Trailer Using the Driver Assistance System”, IEEE, International Conference on Robotics and Biomimetics; Phuket, Thailand, Dec. 2011, 6 pages. |
A. Gonzalez-Cantos; J.I. Maza; A. Ollero, “Design of a Stable Backing Up Fuzzy Control of Autonomous Articulated Vehicles for Factory Automation”, Dept. of Systems Engineering and Automatic Control, University of Seville, Spain, 2001, 5 pages. |
Altafini, C.; Speranzon, A.; Wahlberg, B., “A Feedback Control Scheme for Reversing a Truck and Trailer Vehicle”, IEEE, Robotics and Automation, IEEE Transactions, Dec. 2001, vol. 17, No. 6, 2 pages. |
Zare, A. Sharafi; M. Kamyad, A.V., “A New Approach in Intelligent Trailer Parking”, IEEE, 2010 2nd International Mechanical and Electrical Technology (ICMET), Sep. 2010, 1 page. |
Tanaka, K.; Sano, M., “A Robust Stabilization Problem of Fuzzy Control Systems and its Application to Backing up Control of a Truck-trailer”, IEEE Transactions on Fuzzy Systems, May 1994, vol. 2, No. 2, 1 page. |
Sharafi, M. Zare; A. Kamyad; A.V. Nikpoor, S., “Intelligent Parking Method for Truck in Presence of Fixed and Moving Obstacles and Trailer in Presence of Fixed Obstacles: Advanced Fuzzy Logic Technologies in Industrial Applications”, IEEE, 2010 International Electronics and Information Engineering (ICEIE), Aug. 2010, vol. 2, 1 page. |
Hodo, D. W.; Hung, J.Y.; Bevly, D. M.; Millhouse, S., “Effects of Sensor Placement and Errors on Path Following Control of a Mobile Robot-Trailer System”, IEEE, American Control Conference, Jul. 2007, 1 page. |
Sharafi, M. Zare; A. Kamyad; A.V. Nikpoor, S., “Intelligent Parking Method for Trailers in Presence of Fixed and Moving Obstacles”, IEEE, 2010 3rd International Conference on Advanced Computer Theory and Engineering (ICACTE), Aug. 2010, vol. 6, 1 page. |
Chieh Chen; Tomizuka, M., “Steering and Independent Braking Control for Tractor-Semitrailer Vehicles in Automated Highway Systems”, IEEE, Proceedings of the 34th IEEE Conference on Decision and Control, Dec. 1995, vol. 2, 1 page. |
P. Bolzern, R.M. Desantis, A. Locatelli, “An Input-Output Linearization Approach to the Control of an n-Body Articulated Vehicle”, J. Dyn. Sys., Meas., Control, Sep. 2001, vol. 123, No. 3, 3 pages. |
Dieter Zöbel, David Polock, Philipp Wojke, “Steering Assistance for Backing Up Articulated Vehicles”, Systemics, Cybernetics and Informatics; vol. 1, No. 5, date unknown, 6 pages. |
J.R. Billing; J.D. Patten; R.B. Madill, “Development of Configurations for Infrastructure-Friendly Five- and Six-Axle SemiTrailers”, National Research Council of Canada and Ontario Ministry of Transportation, date unknown, 11 pages. |
Jesus Morales, Anthony Mandow, Jorge L. Martinez, and Alfonso Garcia-Cerezo, “Driver Assistance System for Backward Maneuvers in Passive Multi-Trailer Vehicles”, IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Oct. 2012, 7 pages. |
Cedric Pradalier and Kane Usher, “Experiments in Autonomous Reversing of a Tractor-Trailer System”, 6th International Conference on Field and Service Robotics, inria-00195700, Version 1, Dec. 2007, 10 pages. |
Andri Riid, Alar Leibak, Ennu Rüstern, “Fuzzy Backing Control of Truck and Two Trailers”, Tallinn University of Technology; Tallinn, Estonia, date unknown, 6 pages. |
Jane McGrath, “How to Avoid Jackknifing”, A Discovery Company, date unknown, 3 pages. |
Claudio Altafini, Alberto Speranzon, and Karl Henrik Johansson, “Hybrid Control of a Truck and Trailer Vehicle”, Springer-Verlag Berlin Heidelberg, HSCC 2002, LNCS 2289; 2002, 14 pages. |
Jujnovich, B.; Roebuck, R.; Odhams, A.; David, C., “Implementation of Active Rear Steering of a Tractor Semitrailer”, Cambridge University, Engineering Department; Cambridge, United Kingdom, date unknown, 10 pages. |
A.M.C. Odhams; R.L. Roebuck; C. Cebon, “Implementation of Active Steering on a Multiple Trailer Long Combination Vehicle”, Cambridge University, Engineering Department; Cambridge, United Kingdom, date unknown, 13 pages. |
Cedric Pradalier and Kane Usher, “Robust Trajectory Tracking for a Reversing Tractor-Trailer System”, (Draft), Field and Service Robotics Conference, CSIRO ICT Centre, Jul. 2007, 16 pages. |
Stahn, R.; Heiserich, G.; Stopp, A., “Laser Scanner-Based Navigation for Commercial Vehicles”, IEEE, 2007 IEEE Intelligent Vehicles Symposium, Jun. 2007, 1 page. |
Lee Yong H.; Weiwen Deng; Chin Yuen-Kwok Steve; McKay Neil, “Feasibility Study for a Vehicle-Trailer Backing Up Control”, Refdoc.fr, SAE Transactions, vol. 113, No. 6, 2004, 1 page. |
A.M.C. Odhams; R.L. Roebuck; B.A. Jujnovich; D. Cebon, “Active Steering of a Tractor-Semi-Trailer” Proceedings of the Institution of Mechanical Engineers, Part D: Journal of Automobile Engineering, SAGE Journals, vol. 225, No. 7, Jul. 2011, 1 page. |
Haviland, G S, “Automatic Brake Control for Trucks—What Good Is It?”, TRID, Society of Automotive Engineers, Sep. 1968, 1 page. |
William E. Travis; David W. Hodo; David M. Bevly; John Y. Hung, “UGV Trailer Position Estimation Using a Dynamic Base RTK System”, American Institute of Aeronautics and Astronautics, date unknown, 12 pages. |
“VSE Electronic Trailer Steering”, ETS for Trailers, version 2009, VSE Trailer Systems B.V., 2009, 28 pages. |
“Telematics Past, Present, and Future,” Automotive Service Association, www.ASAshop.org, May 2008, 20 pages. |
“Fully Automatic Trailer Tow Hitch With LIN Bus,” https://webista.bmw.com/webista/show?id=1860575499&lang=engb&print=1, date unknown, 5 pages. |
“VBOX Yaw Rate Sensor With Integral Accelerometers,” Racelogic, www.racelogic.co.uk, date unknown, 2 pages. |
P.D.C.R Jayarathna; J.V Wijayakulasooriya; S.R Kodituwakku, “Fuzzy Logic and Neural Network Control Systems for Backing up a Truck and a Trailer”, International Journal of Latest Trends in Computing, vol. 2, No. 3, Sep. 2011, 8 pages. |
Olof Enqvist, “AFS-Assisted Trailer Reversing,” Institutionen for systemteknik Deartment of Electrical Engineering, Jan. 27, 2006, 57 pages. |
SH. Azadi, H.R. Rezaei Nedamani, and R. Kazemi, “Automatic Parking of an Articulated Vehicle Using ANFIS”, Global Journal of Science, Engineering and Technology (ISSN: 2322-2441), 2013, pp. 93-104, Issue No. 14. |
F. Cuesta and A. Ollero, “Intelligent System for Parallel Parking of Cars and Tractor-Trailers”, Intelligent Mobile Robot Navigation, STAR, 2005, pp. 159-188, Springer-Verlag Berlin Heidelberg. |
Number | Date | Country | |
---|---|---|---|
20200164803 A1 | May 2020 | US |