The present invention pertains to agricultural vehicles and, more specifically, to a system for detecting and managing an implement which is attached to an agricultural vehicle.
Farmers utilize a wide variety of implements to prepare soil for planting. For example, a strip tillage implement is capable of collectively tilling soil in strips along the intended planting rows, moving residue to the areas in between rows, and preparing the seedbed of the strip in preparation for planting. As another example, a field cultivator is capable of simultaneously tilling soil and leveling the tilled soil in preparation for planting.
Some modern implements may automatically identify themselves to the control system of the agricultural vehicle upon being electrically coupled to the agricultural vehicle, for example by way of an ISOBUS connection. However, some implements may not include modern electronics. In such cases, the operator must manually identify the implement type within the control software of the agricultural vehicle in order to properly set operational parameters and log implement data. As can be appreciated, the operator may forget to identify or improperly identify the implement; thus, causing suboptimal operation and improper data collection. Therewith, the operator or an automatic guidance system may improperly conduct a turning maneuver if an implement is improperly identified. For instance, the operator or guidance system may conduct an overly narrow end-of-row turn, which may lead to the implement contacting the agricultural vehicle. If the contact between the implement and the agricultural vehicle is severe, then such contact may damage the implement or the agricultural vehicle
What is needed in the art is a system and method to automatically identify an implement and manage the operation thereof.
Exemplary embodiments provided according to the present disclosure include a method and an agricultural system for the automatic detection and management of an implement which is towed behind an agricultural vehicle. The agricultural vehicle automatically detects a type of implement, sets the operational parameters, and manages turning maneuvers based at least in part upon the sensed and real-time position of the implement relative to the agricultural vehicle.
In some exemplary embodiments provided in accordance with the present disclosure, a method for operating an agricultural vehicle is provided. The agricultural vehicle includes a controller and at least one image capturing device operably connected to the controller. The agricultural vehicle tows an implement. The method includes capturing, by the at least one image capturing device, image data of the implement, receiving, by the controller, the image data of the implement, and identifying the implement, by the controller, depending at least in part upon the image data of the implement. The method also includes setting, by the controller, at least one operational parameter of the implement, and managing, by the controller, a turning maneuver of the agricultural vehicle depending at least in part upon the image data of the implement.
In some exemplary embodiments provided in accordance with the present disclosure, an agricultural vehicle which is configured to tow an implement is provided. The agricultural vehicle includes a frame and at least one image capturing device connected to the frame. The at least one image capturing device is configured to capture image data of the implement. The agricultural vehicle also includes a controller operably connected to the at least one image capturing device. The controller is configured to receive the image data of the implement, identify the implement depending at least in part upon the image data of the implement, set at least one operational parameter of the implement, and manage a turning maneuver of the agricultural vehicle depending at least in part upon the image data of the implement.
One possible advantage that may be realized by exemplary embodiments provided according to the present disclosure is that an operator does not need to manually identify the type of the implement.
Another possible advantage that may be realized by exemplary embodiments provided according to the present disclosure is that turning maneuvers may be optimized, automatically without operator input, by reducing the turning radius to its minimum value without risking damage to the implement and the agricultural vehicle.
For the purpose of illustration, there are shown in the drawings certain embodiments of the present invention. It should be understood, however, that the invention is not limited to the precise arrangements, dimensions, and instruments shown. Like numerals indicate like elements throughout the drawings. In the drawings:
The terms “forward”, “rearward”, “left” and “right”, when used in connection with the agricultural vehicle and/or components thereof are usually determined with reference to the direction of forward operative travel of the towing vehicle, but they should not be construed as limiting. The terms “longitudinal” and “transverse” are determined with reference to the fore-and-aft direction of the towing vehicle and are equally not to be construed as limiting.
Referring now to the drawings, and more particularly to
The agricultural vehicle 12 may generally include a frame 16, a prime mover, a cab, and wheels and/or tracks 18. It is noted that only the rear wheels and/or tracks 18 are illustrated in
The agricultural vehicle 12 may also include at least one image capturing device 24. The at least one image capturing device 24 may capture pictures and/or videos of the implement 14 and the area surrounding the implement 14. The at least one image capturing device 24 may collect the image data before the implement 14 has been connected to the agricultural vehicle 12, during the connection process, and/or after the implement 14 has been connected to the agricultural vehicle 12. For instance, the image capturing device 24 may continually collect image data throughout a farming operation or selectively capture image data only during a connection process and a turning maneuver. Each image capturing device 24 may be connected to the frame 16. The at least one image capturing device 24 may be in the form of a camera, such as a backup camera.
The implement 14 may be pivotally connected to and towed by the agricultural vehicle 12. The implement 14 generally includes a main frame 26, a subframe 28, wheels connected to the main frame 26, various ground-engaging tools mounted to the frame 26 and/or the subframe 28, and a tongue or drawbar 30 which pivotally connects to the agricultural vehicle 12. Once connected to the agricultural vehicle 12, the longitudinal axis LA of the implement 14, e.g. the drawbar 30 thereof, may define an angle A1 relative to a transverse axis TA of the agricultural vehicle 12, e.g. an axis which is perpendicular to the forward direction of travel (
The autonomous or semi-autonomous agricultural system 10 may further include a controller 40 with a memory 42. The controller 40 may be incorporated into the agricultural vehicle 12. The controller 40 can be operably connected to the user interface 20, the steering system 22, and each image capturing device 24. The controller 40 may also be additionally connected to any other desired sensor, including a global positioning system (GPS) location sensor, a speed sensor, and/or an inclinometer.
The controller 40 may comprise the Case IH Advanced Farming System® (AFS), which may collectively and automatically control and record the operation of the agricultural vehicle 12 and the implement 14. The controller 40 may comprise one or more systems for identifying the implement 14, recording data relating to the agricultural vehicle 12 and/or the implement 14, and controlling the operation of the agricultural vehicle 12 and/or the implement 14. Therein, the controller 40 may include an automatic vehicle guidance system 44, which actively controls the steering system 22, and a data management system 46 for recording data relating to the agricultural vehicle 12 and/or the implement 14. Hence, the controller 40 can continually calculate a vehicle steering heading or turning maneuver by comparing vehicle position and directional heading to a desired travel path, and further by incorporating a determined minimum turning angle and/or interference zone Z (
Additionally, the controller 40 may automatically conduct implement detection and turn management of the agricultural vehicle 12. More particularly, the controller 40 may receive the image data from the image capturing device(s) 24, identify the implement 14 depending at least in part upon the image data, set at least one operational parameter, and manage a turning maneuver of the agricultural vehicle 12 depending at least in part upon the image data. As can be appreciated a turning maneuver may include any desired turning operation of the agricultural vehicle 12, such as an end-of-row turn in the headland area of the field. For instance,
In identifying the implement 14, the controller 40 can compare the image data collected by the image capturing device 24 to a database of implements and match the implement 14 to one implement of the database of implements. For instance, the controller 40 may compare one or more identifying characteristics of the implement 14, such as brand name or logo 32, with identifying characteristics of known implements. Therewith, in identifying the implement 14, the controller 40 may conduct a machine learning algorithm or other deep-learning artificial intelligence algorithm to, at least partially, create the database of implements and to identify the implement 14. It should be appreciated that the database of implements may comprise information of various types of implements and the identifying characteristics associated with the various types of implements. Such identifying characteristic information may include the brand name, model number, height and/or shape of the frame and/or subframe, QR code(s), accompanying tools, etc.
Additionally, the controller 40 may also streamline the implement selection process within the data management system 46. For instance, the controller 40 may populate a picklist with one or more possible implements from which the operator may choose. Also, for instance, the controller 40 may automatically select the appropriate implement in the data management system 46. Thereby, by way of example only, the operator may initially choose the implement 14 within the data management system 46, and the controller may subsequently automatically select the implement 14 by way of the machine learning algorithm. Thereafter, the controller 40 may automatically set the initial settings and/or operational parameters of the implement 14.
The controller 40 may optimize the turning maneuver by minimizing a turning radius of the agricultural vehicle 12. In managing the turning maneuver, the controller 40 may also artificially limit a maximum turning angle of the agricultural vehicle 12 to prevent interference between the implement 14 and the agricultural vehicle 12. Furthermore, the controller 40 may determine the angle A1 of the drawbar 30 relative to the agricultural vehicle 12 depending at least in part upon the image data such that the artificial limit of the maximum turning angle of the agricultural vehicle 12 depends upon the real-time angle A1 of the drawbar relative to the agricultural vehicle 12. It should be appreciated that the controller 40 may also monitor a portion of the agricultural vehicle 12, e.g. the wheels and/or tracks 18, relative to the position of the drawbar 30. Additionally or alternatively, the controller 40 may determine an interference zone Z of the drawbar 30. If a portion of the agricultural vehicle 12, e.g. the rear wheels and/or tracks 18, enters or occupies the interference zone Z, it may signify that a potential interference, i.e., contact, between the agricultural vehicle 12 and the implement 14 may occur. The controller 40 may determine the interference zone Z by defining an area which is a preselected distance away from each side of the drawbar 30. Thereafter, the controller 40 may set the artificial limit of the maximum turning angle of the agricultural vehicle 12 depending upon a position of the agricultural vehicle 12 relative to the interference zone Z of the drawbar 30. As used herein, the term interference zone refers to an area surrounding at least a portion of the drawbar 30.
Also, the controller 40 may manage the turning maneuvers by determining and sending a steer-limit output signal to the steering system 22. The steer-limit output signal may correspond to a predetermined minimum angle and/or desired interference zone Z. Upon receiving the output signal from the controller 40, the steering system 22 may prevent the agricultural vehicle 12 from turning beyond the maximum turning angle of the agricultural vehicle 12 which was artificially limited by the controller 40 for a specific implement type.
The controller 40 may also automatically steer the agricultural vehicle 12, via the automatic vehicle guidance system 44, during the turning maneuver. Therein, the controller 40 may automatically control the steering system 22 to minimize a turning radius of the agricultural vehicle 12 and prevent interference between the implement 14 and the agricultural vehicle 12.
The autonomous or semi-autonomous agricultural system 10 may also optionally include a network 50 which operably couples the agricultural vehicle 12 to one or more other agricultural vehicles 52. Thereby, the agricultural vehicle 12 can be part of a neural network comprising at least one other agricultural vehicle 52. The network 50 may operably connect the controller 40 to the controllers of the other agricultural vehicles 52. The network 50 may be configured to receive and transmit the image data of the implement 14. The network 50 may be any suitable network, including a wireless network having one or more processors or nodes. Additionally, the network 50 may broadly represent any combination of one or more data communication networks including local area networks, wide area networks, etc., using a wired or wireless connection.
Furthermore, the autonomous or semi-autonomous agricultural system 10 may also optionally include a remote machine learning or data center 54. In cooperation with the controller 40, the data center 54 may also be configured to receive, process, and record the image data of the implement 14. Additionally, the data center 54 may include one or more processors arranged to conduct a machine learning algorithm or other deep-learning artificial intelligence algorithm to, at least partially, create the database of implements and/or to identify the implement 14.
Referring now to
It is to be understood that the steps of the method 60 may be performed by the controller 40 upon loading and executing software code or instructions which are tangibly stored on a tangible computer readable medium, such as on a magnetic medium, e.g., a computer hard drive, an optical medium, e.g., an optical disc, solid-state memory, e.g., flash memory, or other storage media known in the art. Thus, any of the functionality performed by the controller 40 described herein, such as the method 60, is implemented in software code or instructions which are tangibly stored on a tangible computer readable medium. The controller 40 loads the software code or instructions via a direct interface with the computer readable medium or via a wired and/or wireless network. Upon loading and executing such software code or instructions by the controller 40, the controller 40 may perform any of the functionality of the controller 40 described herein, including any steps of the method 60 described herein.
The term “software code” or “code” used herein may refer to any instructions or set of instructions that influence the operation of a computer or controller. They may exist in a computer-executable form, such as machine code, which is the set of instructions and data directly executed by a computer's central processing unit or by a controller, a human-understandable form, such as source code, which may be compiled in order to be executed by a computer's central processing unit or by a controller, or an intermediate form, such as object code, which is produced by a compiler. As used herein, the term “software code” or “code” also includes any human-understandable computer instructions or set of instructions, e.g., a script, that may be executed on the fly with the aid of an interpreter executed by a computer's central processing unit or by a controller.
These and other advantages of the present invention will be apparent to those skilled in the art from the foregoing specification. Accordingly, it is to be recognized by those skilled in the art that changes or modifications may be made to the above-described embodiments without departing from the broad inventive concepts of the invention. It is to be understood that this invention is not limited to the particular embodiments described herein, but is intended to include all changes and modifications that are within the scope and spirit of the invention.
Number | Name | Date | Kind |
---|---|---|---|
4423883 | Wallentowitz | Jan 1984 | A |
5018757 | Kozuka | May 1991 | A |
5129477 | Hurlburt | Jul 1992 | A |
6933837 | Gunderson et al. | Aug 2005 | B2 |
6990399 | Hrazdera | Jan 2006 | B2 |
7689356 | Dix et al. | Mar 2010 | B2 |
8165407 | Khosla et al. | Apr 2012 | B1 |
9041806 | Baur et al. | May 2015 | B2 |
9085261 | Lu et al. | Jul 2015 | B2 |
9232688 | Kormann et al. | Jan 2016 | B2 |
9305216 | Mishra | Apr 2016 | B1 |
9446713 | Lu et al. | Sep 2016 | B2 |
9944318 | Hilnbrand et al. | Apr 2018 | B2 |
10046800 | Hu et al. | Aug 2018 | B2 |
10049298 | Kraeling et al. | Aug 2018 | B2 |
10266201 | Dang et al. | Apr 2019 | B2 |
10405535 | Rees | Sep 2019 | B2 |
10795372 | Runde | Oct 2020 | B2 |
10940727 | Billich | Mar 2021 | B2 |
11315258 | Anagnostopoulos | Apr 2022 | B1 |
20050055147 | Hrazdera | Mar 2005 | A1 |
20060048953 | Rozendaal | Mar 2006 | A1 |
20160313737 | Berkemeier | Oct 2016 | A1 |
20170112044 | Stratton | Apr 2017 | A1 |
20170297621 | Bunderson | Oct 2017 | A1 |
20170339821 | Ray | Nov 2017 | A1 |
20170344004 | Foster | Nov 2017 | A1 |
20180109762 | Aich et al. | Apr 2018 | A1 |
20180373256 | Runde | Dec 2018 | A1 |
20190016184 | Billich | Jan 2019 | A1 |
20190210418 | Hall et al. | Jul 2019 | A1 |
20200353977 | Davies | Nov 2020 | A1 |
20200387720 | Stanhope | Dec 2020 | A1 |
20210043085 | Kreiling | Feb 2021 | A1 |
20210201531 | Creaby | Jul 2021 | A1 |
Number | Date | Country |
---|---|---|
106960216 | Jul 2017 | CN |
208026252 | Oct 2018 | CN |
111754577 | Oct 2020 | CN |
102016212455 | Jan 2018 | DE |
20 2017 105 279 | Jan 2019 | DE |
3 299 996 | Mar 2018 | EP |
3603379 | Jul 2022 | EP |
2011172495 | Sep 2011 | JP |
5560762 | Jul 2014 | JP |
2020012323 | Jan 2020 | JP |
7058569 | Apr 2022 | JP |
WO-2004077926 | Sep 2004 | WO |
Entry |
---|
“CabCAM Cabled Rear-View System with 1 Camera”, A & I Products, Farmtronics Ltd.—Farm Electronics Online Store, Oct. 21, 2019 (5 pages). |
“Nonlinear Model Predictive Trajectory Control in Tractor-Trailer System for Parallel Guidance in Agricultural Field Operations”, J. Backman et al., IFAC AgriControl 2010 conference, Kyoto, Japan, 2010, pp. 133-138 (6 pages). |
“Development and Implementation of a Team of Robotic Tractors for Autonomous Peat Moss Harvesting”, David A. Johnson, et al., Journal of Field Robotics 26(6-7), pp. 549-571, 2009 (23 pages). |
“Driver Assistance System for Backward Maneuvers in Passive Multi-trailer Vehicles”, Jesus Morales et al., 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, Oct. 7-12, 2012, pp. 4853-4858 (6 pages). |
“Comparison of controls for a stereo camera based reversing assistance system for vehicle trailer combinations”, Ilja Stasewitsch, et al., Bornimer Agrartechnische Berichte, Heft 101, ISSN 0947-7314, (9 pages). |
“Implementation of Active Steering on A Multiple Trailer Long Combination Vehicle”, A.M.C. Odhams et al., Cambridge University, Engineering Department, pp. 1-13, 2010 (13 pages). |
Number | Date | Country | |
---|---|---|---|
20220201920 A1 | Jun 2022 | US |