Embodiments of the disclosure relate to providing a drone system for herding animals.
Animal husbandry involving the herding domestic animals such as cows or sheep is a socially complex activity typically involving communication and cooperation between three or four different types of social mammals: men; trained canines; the herded animals; and horses if the men are on horseback. The activity generally involves learned patterns of communications between the herd animals, attendant dogs, horses, and men, and their joint synchronized movement, often over relatively large distances and difficult terrain that are negotiated at least on part by land vehicle and/or aircraft. The activity regularly requires long hours of alert, but often monotonous, work and attention to detail, and may be a relatively expensive component of the costs of a financial return provided by the husbandry.
An aspect of an embodiment of the disclosure relates to providing a herd management (HeMan) system, also referred to simply as “HeMan”, comprising an optionally cloud based control and data hub and at least one drone with which the hub communicates that may operate autonomously or semi-autonomously to control movement of a herd of animals.
In an embodiment HeMan is configured to receive assignment of a herding task such as moving a herd of animals from a first location to a second location and operate to deploy at least one drone, optionally referred to as a “drone cowboy”, that may autonomously herd the animals from the first location to the second location. In an embodiment the HeMan hub comprises or has access to a telecommunications system for communication with the drone cowboy and for receiving location data transmitted from an animal in the herd that may be tagged with a radio transmitter and a GPS receiver.
In an embodiment the hub comprises a memory storing a terrain map of a geographical region of interest (GROI) in which the herd may be located and a drone herding gesture (DHG) data base. The DHG data base may comprise a library of herding gestures that a drone cowboy may perform to control movement of a herd. A herding gesture may comprise at least one or any combination of more than one of an aerobatic maneuver, an acoustic gesture, or an optical gesture that a drone cowboy may perform to control herd movement. An aerobatic maneuver comprises a formatted gesture flight pattern intended to elicit a particular type of movement by a herd or an animal in a herd. An acoustic gesture may comprise by way of example, a barking noise made by a herd dog, a vocalization made by a cowboy, or an artificial noise made to herd an animal or animals. An optical gesture may comprise a visual light stimulus that elicits a desired response from a herd or herd animal.
In an embodiment, the at least one drone cowboy comprises a radio transceiver for communicating with the HeMan hub, a GPS receiver and/or optionally an inertial measurement unit (IMU) for determining location of the drone, a camera, and a controller operable to control flight of the drone cowboy. In an embodiment, the controller comprises a memory for storing a terrain map of the GROI, coordinates of landmarks and/or locations of animals relevant to the herding assignment, and/or a herding flight plan, optionally at least partially preplanned, for carrying out the herding assignment. The flight plan optionally comprises a sequence of herding gestures to be synchronized with execution of the flight plan by the at least one drone cowboy.
In an embodiment a herding flight plan may be dynamically updated during execution of the herding assignment responsive to behaviour of the herded animals, unknown features in the GROI and/or changes in the GROI. To facilitate monitoring and real time updating of performance of the herding task, the at least one drone cowboy may be configured to image the animals being herded and/or the terrain in which the animals are located and process the images and/or transmit the images for processing by the HeMan hub to update the herding flight plan. A flight plan may be determined by a user, the HeMan hub, and/or the controller of the at least one drone cowboy.
In accordance with an embodiment of the disclosure, HeMan may comprise a neural network that learns to refine performance of HeMan in carrying out herding assignments based on HeMan experience in carrying out such assignments. For example, for a given herd of animals the neural network may learn which herding gestures, or features of herding gestures are advantageous in eliciting desires responses from herded animals. Additionally, or alternatively, the neural network may learn to distinguish particular features of the GROI landscape which are conducive to or interfere with efficient herding of the herded animals.
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter.
Non-limiting examples of embodiments of the invention are described below with reference to figures attached hereto that are listed following this paragraph. Identical features that appear in more than one figure are generally labeled with a same label in all the figures in which they appear. A label labeling an icon representing a given feature of an embodiment of the invention in a figure may be used to reference the given feature. Dimensions of features shown in the figures are chosen for convenience and clarity of presentation and are not necessarily shown to scale
In the discussion, unless otherwise stated, adjectives such as “substantially” and “about” modifying a condition or relationship characteristic of a feature or features of an embodiment of the disclosure are understood to mean that the condition or characteristic is defined to within tolerances that are acceptable for operation of the embodiment for an application for which the embodiment is intended. Wherever a general term in the disclosure is illustrated by reference to an example instance or a list of example instances, the instance or instances referred to, are by way of non-limiting example instances of the general term, and the general term is not intended to be limited to the specific example instance or instances referred to. The phrase “in an embodiment”, whether or not associated with a permissive, such as “may”, “optionally”, or “by way of example”, is used to introduce for consideration an example, but not necessarily a required configuration of a possible embodiment of the disclosure. Unless otherwise indicated, the word “or” in the description and claims is considered to be the inclusive “or” rather than the exclusive or, and indicates at least one of, or any combination of more than one of items it conjoins.
The static parameters of a DHG may include a gesture flight pattern “GFP” followed by the numerical ID of the DHG, and an intended gesture direction “GD” followed by the numerical ID of the DHG. The flight pattern, “GFP”, of a given DHG may comprise a set of executable instructions which when executed by an onboard controller of a HeMan drone cowboy cause the drone cowboy to engage in a particular flight pattern intended to elicit a particular response from a herd or herd animal. The intended gesture direction GD of a given DHG is a direction of motion of a herd or herd animal that the given gesture is intended to generate or affect. A GD is substantially fixed with respect to a geometry of the given gesture’s flight pattern and may be defined by a unit vector having a fixed direction relative to a direction of the gesture flight pattern GFP. In each figure the gesture direction GD is indicated by a patterned block arrow labeled “GESTURE DIRECTION (GD)” and indicated by a numerical label “21”. The associated gesture flight pattern is labeled “GFP” and indicated by a numerical label “22”.
A DHG for which the GFP comprises a sequence of distinct component flight movements is represented by a plurality of component DHG functions. Each component DHG function belonging to a same DHG is identified by a decimal ID number having a same number to the left of the decimal and a different number to the right of the decimal. The number to the left of the decimal is used to reference the DHG and generically reference the component DHG functions. The increasing order of the numbers to the right of the decimal indicate the sequence in which the distinct flight movements belonging to the DHG component functions are performed.
Dynamic arguments for a given DHG are shown in italicized script and may include a location of a waypoint “W” along a drone cowboy flight path at which a drone cowboy flying the flight path operates to gesture to a herd by performing the given DHG. The dynamic arguments include arguments that characterise location and movement of the herd gestured to and desired movement and/or location of the herd to be achieved by preforming the gesture. A centroid “Ch” determined for locations of herd animals in the herd and a measure of dispersion “σh” of the locations may be used to characterize location of the herd. A velocity “Vh” of the centroid may be used to characterize motion of the herd. A desired velocity of the centroid, “Vd”, may be used to characterize a desired movement of the herd and “σd” a desired spatial dispersion to be achieved by the DHG.
By way of example, DHG-1 shown in
In an embodiment HeMan may determine that σh ≥ ULM(σh) and monitor progress in clustering the herd based on processing data comprised in GPS locations received by the HeMan hub and/or the drone cowboy from herd animals and/or data in images of the herd acquired by a camera system that the drone cowboy may comprise. Processing data provided by the GPS locations and/or the herd images may be preformed by a processor or processors that the HeMan hub and/or drone cowboy comprises or has access to.
By way of another example, DHG-5 performed at a waypoint W by a HeMan drone cowboy to gesture to a herd to turn left optionally comprises component drone herding gestures DHG-5.1 and DHG-5.2 shown in
It is noted that whereas in the above description and
By way of example HeMan 30 is assigned with a task of herding cattle 120 to arrive at corral 106 by a desired time of arrival (TOA). In response to the assigned task, HeMan 30 operates to determine and execute a herding plan for performing the herding task.
In a block 201 of flow diagram 200 HeMan hub 30 receives the assignment to herd animals 120 in GROI 100 and receives or retrieves from memory 33 a terrain map for GROI 100, a location of corral 106 in the GROI, and/or a desired TOA of animals 120 at corral 106 and operates to determine locations of animals 120 in GROI 100. HeMan 30 may determine the locations of animal 120 by processing data comprised in signals transmitted by GPS transceivers attached to the animals to HeMan hub 32 via at least one communication tower 50. Additionally or alternatively, HeMan 30 may deploy a drone cowboy to scan and image GROI 100 and process images of the GROI received from the drone cowboy to determine the locations of the animals. Optionally in a block 203 HeMan 30 processes the determined locations of animal 120 to determine a centroid, Ch, dispersion σh, and velocity Vh for the herd.
In a block 205 HeMan 30 may use the determined values for Ch, σh, and Vh, the terrain map, and desired TOA of herd of animals 120 at corral 106, to determine a herding plan route to be traveled by animals 120 to reach corral 106.
To determine the herding route, HeMan 30 optionally identifies features of terrain 102 that are conducive to or present obstacles to movement of animals 120. For example, region 108 of terrain 102 is characterized by a steep terrain gradients may be difficult or dangerous for passage of herd animals 120 and may substantially slow movement of the herd animals. On the other hand, stream 100 may be conducive to herd movement and enable relatively rapid movement of herd animals along its banks while providing the animals with drinking water as they move. In an embodiment HeMan processor 33 may use a neural network to process data from the terrain map of terrain 102 to determine a herding plan route in GROI 100 for animals 120 to traverse to corral 106. Alternatively or additionally HeMan 30 may receive a suggested herding route from a user. Optionally HeMan comprises software executable to integrate route herding suggestions made by a user with herding route segments autonomously determined by HeMan to provide a herding route along which to drive animals 120 to corral 106.
In a block 207 HeMan 30 optionally determines a plurality of N waypoints, Wn(Ln,tn), 1≤n≤N, at locations Ln along the herding plan route at which herd animals 120 may be expected to require intervention and suitable gesturing by at least one drone cowboy dispatched by HeMan to arrive at locations Ln at times tn to control movement of the animals along the route. In an embodiment, a first waypoint W1(L1,t1) along the herding plan route is located at a starting location L1 of the route at an estimated time of arrival t1 of the dispatched drone cowboy to herd animals 120. A last waypoint WN(LN,tN) along the herding plan route is located at an end of the route substantially at the herding destination, corral 106, of animals 120 at a time tN substantially equal to the desired TOA of animals 120 at the corral. Optionally, in a block 209 HeMan determines a flight plan according to which the dispatched drone cowboy is expected to fly to reach waypoints Wn(Ln,tn) and drone herding gestures, DHGs, the drone cowboy is planned to gesture to the animals at the waypoints.
Optionally in a block 211 HeMan uploads the herding plan to at least one drone cowboy and in a block 213 optionally dispatches the at least one cowboy to arrive at waypoint W1(L1,t1).
Descriptions of embodiments are provided by way of example and are not intended to limit the scope of the disclosure. The described embodiments comprise different features, not all of which are required in all embodiments of the disclosure. Some embodiments utilize only some of the features or possible combinations of the features. Variations of embodiments of the disclosure that are described, and embodiments of the disclosure comprising different combinations of features noted in the described embodiments, will occur to persons of the art. The scope of the disclosure is limited only by the claims
In the description and claims of the present application, each of the verbs, “comprise” “include” and “have”, and conjugates thereof, are used to indicate that the object or objects of the verb are not necessarily a complete listing of components, elements or parts of the subject or subjects of the verb.
Descriptions of embodiments of the disclosure in the present application are provided by way of example and are not intended to limit the scope of the disclosure. The described embodiments comprise different features, not all of which are required in all embodiments of the disclosure. Some embodiments utilize only some of the features or possible combinations of the features. Variations of embodiments of the disclosure that are described, and embodiments of the disclosure comprising different combinations of features noted in the described embodiments, will occur to persons of the art. The scope of the invention is limited only by the claims.
The present application claims the benefit under 35 U.S.C. 119(e) of U.S. Provisional Application 63/048,816 filed on Jul. 7, 2020 the disclosure of which is incorporated herein by reference.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/IL2021/050836 | 7/7/2021 | WO |
Number | Date | Country | |
---|---|---|---|
63048816 | Jul 2020 | US |