This invention relates in general to precision targeting, and more particularly to a system and method for precision collaborative targeting.
Far Target Location (FTL) sensor systems, such as the Long Range Advanced Scout Surveillance System (LRAS3) and the Improved Target Acquisition System (ITAS), are long-range sensor systems used for target detection, acquisition, and recognition. Azimuth determination error is the largest error component in the Circular Error Probability (CEP) calculation when targeting an object using an FTL sensor equipped with the GPS Interferometer Subsystem (GPSIS). The addition of expensive, external Mission Equipment Packages (MEP) consisting of Inertial Measurement Units (IMU) or North Finding Modules (e.g., gyrocompasses) and precision position pick-off measurement capability of sensor yoke to the sensor vehicle platform previously attempted and have been incorporated to try and solve this problem.
From the foregoing, it may be appreciated by those skilled in the art that a need has arisen to improve accuracy of target location in a cost effective manner. In accordance with the embodiments of the present invention, a system and method for precision collaborative targeting are provided that substantially eliminate or greatly reduce disadvantages and problems associated with conventional target location systems.
According to an embodiment of the present invention, there is provided a method for precision collaborative targeting that includes determining information indicative of a current location and of a known remote location. A first distance from the current location to the known remote location is determined. A sensor is stewed towards an unknown remote location, the sensor being initially directed from the current location to the known remote location. The distance from the current location to the unknown remote location is measured. The information indicative of the location of the unknown remote location is determined from the amount of slewing and the distance from the current location to the unknown remote location.
Certain embodiments of the present invention provide various technical advantages over conventional target location systems. These technical advantages include allowing a user operating an FTL sensor to be able to more accurately determine the location of a distant target without needing to incorporate additional expensive equipment. Accordingly, by leveraging the hardware and software already installed on the FTL sensor, the user may be able to accurately determine the location of a distant target.
Other technical advantages will be readily apparent to one skilled in the art from the following figures, descriptions, and claims. Moreover, while specific advantages have been enumerated above, various embodiments may include all, some, or none of the enumerated advantages.
For a more complete understanding of particular embodiments of the present invention and the advantages thereof, reference is now made to the following description, taken in conjunction with the accompanying drawings wherein like reference numerals represent like parts, in which:
Some of the several possibilities for determining a known surrogate target location may include: (1) utilizing self-surveyed waypoints left along a route traveled; (2) collaborating with another sensor/vehicle via a network 130 and obtaining its accurate GPS self position; (3) locating a fixed surrogate target position on a geo-registered map; or (4) taking advantage of other precision FTL mechanisms such as the Precision Strike Suite for Special Operations Forces (PSS-SOF) used for fixed site FTL.
The amount of separation in distance of the surrogate target 110 from the sensor platform 100 may increase performance. The methodology of particular embodiments may be valid for near term targeting position determination as the methodology assumes the received GPS position errors of the sensor platform 100 and the surrogate target 110 are correlated.
The sensor then determines or receives the GPS-measured position of a surrogate target at block 210. The surrogate target may include any vehicle, building, or landmark having a known GPS-measured position. In the case of a vehicle, the vehicle may pass its GPS-measured position to the sensor performing the FTL termination via a network. The network may be any suitable combination and arrangement of components supporting communications between devices. In some embodiments the information may be passed via a geo-registered map. In the case of a building or landmark, the position of the surrogate target may be determined from a map or other such device. Then the range from the sensor performing the FTL determination to the surrogate target may be determined at block 220. The range may be determined via a laser range finder. At block 230, this information may be used to calculate an accurate azimuth pointing angle to the surrogate target.
The user may then, at block 240, slew a viewfinder initially trained on the surrogate target having a known location to the desired target having an unknown location. The viewfinder may be interfaced with an imaging module (e.g., SBESS). In particular embodiments, the imaging module may, in essence, be counting the number of pixels between the surrogate target and the desired target. At block 250, the imaging module may be able to use the number of pixels to determine an accurate azimuth pointing angle from slewing the viewfinder from the surrogate target to the desired target. The determination of the azimuth may be performed real-time (e.g., as the user slews the viewfinder) or once the user stops slewing the viewfinder. The user may then determine the range to the desired target at block 260. For example, the user may use a laser range finder to determine the range to the desired target.
This information may then be used to accurately calculate a precise GPS position of the desired target at block 270. More specifically, by using the location and range to the surrogate target and the azimuth angle and range to a desired target having an unknown location, the FTL sensor may be able to determine the location of the desired target.
Some of the steps discussed with respect to
Gathering and determining the data needed to perform the FTL determination and performing the FTL determination may involve a combination of hardware, software, and computer logic encoded on a computer readable medium. The hardware may include at least a processor, an interface, and a viewfinder. Some embodiments may further include memory. The components may be coupled together in such a way that they may provide any necessary communications, data, instructions and/or processing capabilities to provide the above identified functionality. In addition, any of these components may be provided as separate external components or located within a common entity or device where appropriate. Particular embodiments contemplate great flexibility in the arrangement of these elements as well as their internal components.
Processor 310 may be a microprocessor, controller, or any other suitable computing device, resource, or combination of hardware, software, and/or encoded logic operable to provide, either alone or in conjunction with other device 300 components (e.g., memory 320), precision targeting functionality. Such functionality may include providing various features discussed herein to a user. One feature that certain embodiments may provide may include determining the azimuth between a known location and an unknown location based on the amount of movement of a viewfinder 340. In particular embodiments the amount of movement may be determined, in essence, by counting the number of pixels between the known location and the unknown location. The pixels may be generated by an imaging algorithm, such as SBESS. By combining the coordinates or position of the known location and the range to that location, with the azimuth to the unknown location and the range to the unknown location, processor 310 may be able to determine the coordinates of the unknown location.
Memory 320 may be any form of volatile or non-volatile memory including, without limitation, magnetic media, optical media, random access memory (RAM), read-only memory (ROM), removable media, or any other suitable local or remote memory component. Memory 320 may store any suitable data or information, including software and encoded logic, utilized by device 300 in determining precise targeting data. For example, memory 320 may maintain a listing, table, or other organization of information used in determining the coordinates/positions of several known locations. Memory 320 may also store any logic needed to perform any of the functionality described herein. For example, memory 320 may store one or more algorithms that may be used to determine the azimuth angle from the number of pixels between the known location and the unknown location.
Interface 330 may comprise any hardware, software, and/or encoded logic needed to be able to send and receive information between other components, such as a surrogate target and other sensors. For example, interface 330 may receive messages indicating the location of a particular surrogate target. In particular embodiments, interface 330 may be able to send and receive Join Variable Message Format messages over a Tactical Network for Army use.
Viewfinder 340 may comprise any optical or electrical viewfinder operable to focus on or aim at another location, such as a surrogate target or an enemy target, in order to provide precision targeting. For example, the viewfinder may be interfaced with an imaging module (e.g., SBESS) in order to determine an accurate azimuth pointing angle from slewing the viewfinder from the surrogate target to the desired target. In particular embodiments, the imaging module may be counting the number of pixels between the surrogate target and the desired target for use by the processor 310 to calculate the azimuth pointing angle to the desired target.
In summary, embodiments of the present invention, use a combination of networking of sensors for sharing of GPS positioning information for a targeting sensor platform and SBESS processing for measuring precise angle offset of actual target from a known collaborative target in order to determine a location of the actual target. Other embodiments in known surveyed landmarks as collaborative target positions and then SBESS to measure an angle to the actual target. Error correlation of known positions can improve accuracy.
Several different embodiments and features have been presented. Particular embodiments may combine one or more of these features depending on operational needs and/or component limitations. This may allow for great adaptability to the needs of various organizations and users. For example, a particular embodiment may use multiple surrogate targets and/or determine the position of multiple enemy targets. While various implementations and features are discussed with respect to multiple embodiments, it should be understood that such implementations and features may be combined in various embodiments. Numerous other changes, substitutions, variations, alterations, and modifications may be ascertained by those skilled in the art and it is intended that the present invention encompass all such changes, substitutions, variations, alterations, and modifications as falling within the spirit and scope of the appended claims.
This application claims the benefit to U.S. Provisional Application No. 60/987,974, filed Nov. 14, 2007.
Number | Name | Date | Kind |
---|---|---|---|
5969676 | Tran | Oct 1999 | A |
6323942 | Bamji | Nov 2001 | B1 |
6338011 | Furst | Jan 2002 | B1 |
6377211 | Hsiung | Apr 2002 | B1 |
6535816 | Smith | Mar 2003 | B1 |
7728264 | De Sa | Jun 2010 | B2 |
20030218546 | Lemp, III | Nov 2003 | A1 |
20050055417 | Reich et al. | Mar 2005 | A1 |
20050060092 | Hablani | Mar 2005 | A1 |
20050225854 | McWilliams | Oct 2005 | A1 |
20060238860 | Baun et al. | Oct 2006 | A1 |
20070016371 | Waid et al. | Jan 2007 | A1 |
20070050097 | Lo | Mar 2007 | A1 |
20070057840 | Thomas et al. | Mar 2007 | A1 |
20070103671 | Ash | May 2007 | A1 |
20070282565 | Bye et al. | Dec 2007 | A1 |
20080059071 | Meckauskas et al. | Mar 2008 | A1 |
20080163504 | Smith | Jul 2008 | A1 |
20080316021 | Manz et al. | Dec 2008 | A1 |
20090021447 | Austin et al. | Jan 2009 | A1 |
20100013917 | Hanna | Jan 2010 | A1 |
Number | Date | Country |
---|---|---|
2 394 376 | Apr 2004 | GB |
WO-2009064982 | May 2009 | WO |
Entry |
---|
PCT Notification of Transmittal of the International Search Report and The Written Opinion of the International Searching Authority, or the Declaration, PCT/US2008/083550; dated Feb. 25, 2009, 13 pages. |
Communication pursuant to Article 94(3) EPC from European Patent Office; 3 pages; dated Oct. 29, 2010 and reported Nov. 26, 2010. |
“European Application Serial No. 08848589.1, Response filed May 9, 2011 to Office Action mailed Oct. 29, 2010”, 15 pgs. |
“European Application Serial No. 08848589.1, Amended Claims filed May 26, 2010 in response to the PCT Written Opinion”, 6 pgs. |
“European Application Serial No. 08848589.1, Office Action mailed Jun. 23, 2010”, 2 pgs. |
“European Application Serial No. 08848589.1, Response filed Jul. 30, 2010 to Office Action mailed Jun. 23, 2010”, 1 pg. |
“International Application Serial No. PCT/US2008/083550, International Preliminary Report on Patentability mailed May 18, 2010”, 7 pgs. |
Number | Date | Country | |
---|---|---|---|
20090216432 A1 | Aug 2009 | US |
Number | Date | Country | |
---|---|---|---|
60987974 | Nov 2007 | US |