The invention relates to a shooting simulation system and method for training personnel in targeting visual and non-line of sight targets.
Military, security, and law enforcement personnel conduct training in order to experience and learn from mistakes prior to a “real world” event. Small arms and vehicle marksmanship training involves a mix of techniques, including firing live ammunition on a firearm range. An important training technique is live, force-on-force training. In such training, participants in a field environment employ tactics and their full range of firearm systems against each other. An important component of such training is proper employment of the trainees' firearms while reinforcing proper tactics, techniques, and procedures.
Current state of the art employs laser emitters on the Shooters' firearms and laser sensors on the targets. An exemplar system of this type is the Multiple Integrated Laser Engagement System, or MILES. In laser engagement systems an emitter mounted on the firearm generates a laser signal when the firearm's trigger is pulled and a blank cartridge creates the appropriate acoustic, flash, and/or shock signature. These types of laser engagement systems suffer many drawbacks that collectively provide “negative training”, that is training that results in incorrect results or behaviors. The present invention addresses each of these drawbacks.
The first major drawback to laser engagement systems is that they cannot be used to engage partially occluded targets, such as a target that is partially hidden behind a bush. Terrain features that would not stop an actual projectile block lasers. There is evidence that in exercises involving laser engagement systems participants incorrectly learn to take cover behind terrain that would not stop a bullet, resulting in higher casualties in their initial firefights. Similarly, obscurants, such as smoke or fog, may block a laser, stopping participants from successfully engaging legitimate targets.
Proper marksmanship techniques involve aiming slightly ahead of or leading a moving target. The second major drawback of laser engagement systems is that participants are penalized for leading moving targets. Lasers travel in a straight line and are nearly instantaneous. When engaging a moving target with a laser engagement system, participants must—incorrectly—aim at the target, not ahead of it. This is another source of negative training.
Bullets travel in a parabolic trajectory, not a straight line. The sights of firearms are aligned with the barrel of the firearm so that the path of the bullet intersects the line of sight at specified distances, such as 25 and 250 meters, based on how the weapon is bore sighted. At different ranges the bullet's trajectory may be above or below the line of sight so that when firing at shorter ranges the Shooter may have to aim below the center of mass of the target and at longer ranges the Shooter may have to aim above the center of mass. With laser engagement systems, employing these proper marksmanship techniques often results in incorrect misses being recorded, which is yet another source of negative training.
Laser engagement systems project a beam from the emitter toward the target, where one or more detectors worn by the target sense the beam. The beam has a wider diameter as it travels farther due to diffraction. This results in anomalous situations. At short ranges, the beam may be so small that it does not trigger any detectors even though the beam strikes the center of mass of the target. At longer distances, the beam may be so wide that it triggers a detector even though the center of the beam is far from the intended target. Again, these phenomena result in negative training.
Lasers travel in a straight line. This makes laser engagement systems incapable of representing high-trajectory, or non-line of sight, firearms, such as grenade launchers and rifle grenades. As these firearms often represent a significant percent of a military unit's firepower, the inability to simulate them has a negative impact on training. Small unit leaders do not have the opportunity to train to employ these firearms as part of their actions in contact with an enemy and the operators of those firearms do not get a chance to employ them as part of a tactical situation.
Lasers are instantaneous. Armed forces often employ relatively slow moving weapons like anti-tank guided missiles (ATGMs) whose time of flight between the Shooter and the target can be a few seconds. With these systems, it is important for the Shooter to maintain his sight picture of the target throughout the time of flight. Since lasers strike the target almost instantaneously with the pull of the trigger, these slower weapons are not represented realistically in live, force-on-force training.
Finally, laser engagement systems rely on a laser signal striking detectors. Participants who want to win the training event often go to some length to obscure or cover the detectors. A solution that does not rely on a signal striking a detector would be advantageous.
State of the art for mixed and augmented reality technologies has proven insufficient to address live, force-on-force training, largely because they rely on very precise tracking of the participants' locations and the orientations of their firearms. Current tracking technologies used to estimate participant and firearm location and orientation are insufficient to support long-range direct fire. Tracking solutions developed for augmented reality (AR) only support engagements at ranges of approximately 50 meters, but military personnel are trained to fire at targets at 375 meters.
Techniques have been proposed that involve active emitters on the targets to make them easier to sense; however, many military, security, and law-enforcement personnel wear night vision devices. An emitter that is visible in night vision devices is another source of negative training as it may make targets unrealistically easy to detect in the environment.
Other techniques have been proposed which rely on indicia to properly identify targets and compute hits and misses. Techniques involving indicia suffer from many of the same drawbacks as laser engagement systems, namely that they do not enable non-line of sight engagements and they do not permit firing through obscurants and terrain features like bushes and tall grass.
A technology that addresses the shortcomings of laser engagement systems would be advantageous to military, security, and law enforcement professionals and might even be applied to entertainment uses. A solution that permits firing through obscurants and fire at partially occluded targets would improve live, force-on-force training. A solution that takes into account the ballistic characteristics of the simulated projectile with respect to the projectiles trajectory as well as time of flight would enable participants to properly elevate their firearm based on the range to the target and to lead moving targets. If such a system also permitted high-trajectory or non-line of sight fire, that would be advantageous. It would also be advantageous for a system to require no indicia, emitter, or beacons. Finally such a system should enable accurate credit for a hit or miss out to realistic ranges, based on the firearm system being simulated.
Shooting simulation systems may be seen in the Carter U.S. Pat. Nos. 8,888,491 and 8,459,997 and 8,678,824. These patents teach an optical recognition system for simulated shooting using a plurality of firearms with each firearm held by a separate player. Each player has a computer and an optical system associated with the firearm for capturing an image. The image provides information on a trajectory of a simulated bullet fired from a shooting firearm and is used to determine a hit or miss of the targeted player. Each player is wearing some type of indicia such as color codes, bar codes, helmet shape for identification which does not allow non-line of sight engagements and does not permit firing through obscurants and terrain features like bushes and tall grass.
The Sargent U.S. Pat. No. 8,794,967 is for a firearm training system for actual and virtual moving targets. A firearm has a trigger initiated image capturing device mounted thereon and has a processor and a display. The Lagettie et al. U.S. Patent Application Publication No. 2011/0207089 is for a firearm training system which uses a simulated virtual environment. The system includes a firearm having a scope and a tracking system and a display and a processor.
A firearm simulation system has a plurality of participants each having a firearm capable of use with direct and non-line of sight shooting. The Shooter can be a person with a direct fire small arm, such as a rifle or submachine gun or with an indirect fire or high-trajectory firearm, such as a grenade launcher or an unmanned system, or an unmanned ground vehicle or unmanned aerial vehicle. The simulation system includes a plurality of firearms, each firearm having a trigger sensor and one firearm being held by each of a plurality of participants in the simulation. Each participant carries a computer and a position location sensor for determining his location, orientation and movement information. Each firearm has an orientation sensor for recording the orientation of the firearm with respect to a known three-dimensional coordinate system, and has an optical system aligned to the sights of the firearm for capturing the sight picture at the time the trigger sensor is activated to provide image information about the aim point of the Shooter participant's firearm with respect to an intended target participant. A remote computer server has an entity server database and a target resolution module. The remote computer server is wirelessly coupled to each participant to periodically receive and store each participant's position location, orientation and speed information in the server entity state database. The stored data is then used by the remote computer server receiving the captured image and the orientation of the Shooter participant's firearm at the time the trigger sensor is activated for use by the computer server target resolution module for identifying the target participant. The computer server stores reported information on each of a plurality of participants' location, orientation and speed and remotely determines the identification of the target participant of the Shooter participant upon activation of the Shooter Participant's trigger sensor.
A method of simulating firearm use between a plurality of participants includes equipping each of a plurality of participants with a firearm having a trigger sensor and an orientation sensor for recording the orientation of the firearm with respect to a known three-dimensional coordinate system, and an optical system aligned to the sights of the firearm for capturing the sight picture at the time the trigger sensor is activated to provide image information about the aim point of the Shooter participant's firearm with respect to an intended target participant. Equipping each of the plurality of participants with a computer and a position location sensor for determining the location, orientation and movement information of the participant. A remote server is selected having an entity state database and a target resolution module and periodically communicates and stores each participant's location, orientation and movement information to the remote server's entity state database. The captured image and the orientation of the Shooter participant's firearm is received at the remote server at the time the trigger sensor is activated in the computer server. The remote computer server determines which participant is a Shooter participant, which activating a firearm's trigger sensor and which participant is the target participant of the Shooter participant with the remote computer server target resolution module using information stored in the entity state database and the received captured image and the orientation of the Shooter participant's firearm. The remote computer server stores the reported periodic information on each of a plurality of participant's location, orientation and movement for computing the remote identification of a target participant of a Shooter participant.
The accompanying drawings, which are included to provide further understanding of the invention are incorporated in and constitute a part of the specification, and illustrate an embodiment of the invention and together with the description serve to explain the principles of the invention.
The present invention is a system for simulating live, force-on-force simulated firearms engagements at realistic ranges. The Shooter can be a person with a direct fire small arm, such as a rifle or submachine gun or with an indirect fire or high-trajectory firearm, such as a grenade launcher, or an unmanned ground vehicle or unmanned aerial vehicle. The invention simulates a plurality of firearms. The system is symmetrical and homogenous in that a Shooter can also be a target, and vice versa.
In
The Shooter 10 aims his firearm at his Target 11 and pulls the trigger which activates a trigger sensor. The Shooter's location, firearm orientation, and sight image are transmitted to the wireless relay. The sight image is a digital representation of the Shooter's view through his firearm's sight when he pulls the trigger. The location and orientation of the Shooter 10 and his sight image are transmitted to the Remote Server 14 and to the Interaction Manager 16. The Interaction Manager queries the target Resolution Module 17, which produce a list of possible targets from the Entity State Database based on the firearm location, orientation, known position sensor error, and known orientation sensor error. This list of possible targets is provided to the Hit Resolution Module 18.
The Hit Resolution Module 18 runs the multiple, multi-spectral algorithms to find targets in the sight image. Multiple algorithms may be used based on environmental conditions and other factors that influence which algorithms will be the most successful. This step includes processing the sight image to locate targets and determining the relationship between the aim point and the target based on the sight image. For instance, did the Shooter aim high, low, left, or right of center of mass of the target.
The Hit Resolution Module 18 calls the Target Reconciliation Module 20, which reconciles results from the computer vision computation with information from the Entity State Database. This step identifies which targets from the Target Resolution Module 20 correspond to targets identified by the computer vision algorithm. This step is purely based on the results of employing a plurality of computer vision (CV) algorithms and does not rely on any artificial indicia in the scene. The CV algorithms use a plurality of algorithms to construct a silhouette around the target; however, if the CV algorithms cannot construct a full silhouette, they then construct a bounding box around the targets in the scene.
The Hit Resolution Module 18 queries the Munitions Fly-out Module 21 for the flight time of the projectile and adjustments to the trajectory of the round. These adjustments can be based on range (e.g., drop of the round over distance), atmospheric effects, weather, wind, interactions with the terrain, and other factors as required to accurately predict the trajectory of the round. The system uses a representation of the terrain in the area of interest to compute whether the simulated projectile struck the target.
The Hit Resolution Module 18 computes whether the trajectory of the round intersects the target determined by the Target Reconciliation Module 17 based on the adjusted trajectory, time of flight, and relative velocity of the target. Relative velocity accounts for movement of the target, the Shooter, and the Shooter's firearm. If the round strikes the projected target location at time of impact, the Hit Resolution Module 18 calls the Damage Effects Module 22. This module computes the damage to the target based on the firearms characteristics, the munitions characteristics, and location of the calculated impact point in the target's calculated silhouette. Damage effects indicate the extent of damage to the target, such as whether the target was killed, sustained a minor wound or major wound, the location of the wound, and the like.
A near miss is reported through the wireless relay 12 and retransmitted to the Target 11 and the Shooter 10, respectively, who are informed of the near-miss results via audio and visual effects similar to the existing MILES system. A hit result is reported through the wireless relay 12 and re-transmitted to the Target 11 and the Shooter 10, respectively. The Shooter is notified of a hit, and the Target is notified that he was hit, with what firearm or round he was hit, and the severity of the damage.
When the participant pulls the trigger on his training rifle, the Trigger Pull Sensor 25 sends a message to the Participant-Worn Computing Device 24. The Participant-Worn Computing Device 24 captures the trigger-pull events. The Firearm Orientation Sensor 26 returns the firearm orientation to the Participant-Worn Computing Device 24. Similarly, the Image Capture Device 27 provides the sight image as seen by the Shooter 10 to the Participant-Worn Computing Device. The Image Capture Device 27 may provide:
The Participant-Worn Computing Device 24 sends the location and orientation of the firearm as well as the sight images via the Wireless Relay 12 to the Remote Server 14.
The target is not augmented with indicia or beacons. Other than the participant-worn subsystem, the target includes only his operational equipment.
In
The Orientation Sensor 26 provides three-dimensional orientation with respect to the geomagnetic frame of reference. This three-dimensional representation can be in the form of a quaternion; yaw, pitch, and roll; or other frame of reference, as appropriate. The Orientation Sensor 26 is calibrated to the fixed coordinate system when the system is turned on, and it can be periodically recalibrated during a simulation event as necessary. The orientation sensor may employ a plurality of methods to determine three-dimensional orientation. There is no minimum accuracy requirement for the Orientation Sensor 26; although, a more accurate orientation sensor reduces the burden on the Target Reconciliation Module 17.
The Location Sensor 23 provides the Shooter's location with respect to a fixed reference frame. In the current embodiment, this is provided as latitude and longitude, but other coordinate representation methods may be employed. The participant's speeds may be measured directly by the position sensor or may be inferred through the collection of several position reports over time.
The location, orientation, and velocity updates are transmitted 13 to a Remote Server 14, where they are stored in the Entity State Database 15 for later use, as shown in
As depicted in
As shown in
The sight image is a digital representation of the Shooter's view through his firearm's sight when he pulls the trigger. The image capture device 27 is aligned with the barrel and sights of the simulated firearm so that the image captured from the device is an accurate representation of the Shooter's sight picture when the trigger was pulled. In the first embodiment of the invention, the image capture device 27 is the same scope through which the Shooter is aiming the firearm, but the image capture device may be separate form the weapon sights. The image capture device 27 may provide:
The Position Location Sensor 23 provides periodic updates of the participant's location, orientation, and speed to the Participant-Worn Computing Device 24. The Participant-Worn Computing Device transmits these updates to the wireless relay 12.
The location and orientation of the Shooter 10 and his sight image are transmitted from the Wireless relay 12 to the Remote Server 14 and the Interaction Manager 16. Any communication means with sufficient bandwidth may be used in this step of the process. The Participant-Worn Computing Device 24 may perform preprocessing of the captured sight picture to reduce bandwidth requirements. Pre-processing includes, but is not limited to, cropping the image, reducing the resolution of the image, compressing the image, and/or adjusting the tint, hue, saturation, or other attributes of the image.
In
The Target Resolution Module 17 provides this list of possible targets to the Hit Resolution Module 18. In
In
In
Having determined the intended target, in
In
In
The Munitions Fly-Out Module 21 accounts for weapon systems that detonate based on range to the target, distance from the firearm, or other factors, by determining when the detonation occurs. As an example, but not a limitation of the invention, if a Shooter fires simulated munitions from his firearm that explode at a pre-sent distance, the Munitions Fly-Out Module 21 computes the trajectory of the munitions to their points of detonation. The locations where the munitions detonated are then passed to the Damage Effects Module 22 to compute damage to any nearby participants.
In
In
In
The system records information from the Remote Server 14 to assist in reviewing the training event. Information such as, but not limited to, participant's locations over time, sight pictures when triggers were pulled, sight pictures after the CV algorithms have processed them, results from the Target Reconciliation Module 20, and status of participant-worn devices may be displayed to an event controller during and after the training event.
This invention is equally applicable to high-trajectory or non-line of sight shooting. In the case of high-trajectory fire, the image from the Image Capture Device 27 is not necessary. The modified process for non-line of sight and high-trajectory shooting is depicted in
In Step 206, the Target Resolution Module 17 queries the Entity State Database 15 to determine whether any participants, friendly or enemy, are within the burst radius of the simulated munitions. In Step 207, the Munitions Fly-Out Module 21 predicts the locations of those participants at the time of impact or detonation of the simulated munitions. In Step 208, for each participant within the burst radius of the munitions, the Damage Effects Module 22 determines if the participant is hit, where the target was hit, and the severity of the damage, just as described in Step 115,
In Step 209, if a participant received a hit from a high-trajectory shot, in Step 212, the target is notified of the results, including location(s) and severity of wounds. The Shooter 10 may be notified that he has hit his target as well. In an augmented reality situation, this notification might come in the form of a depiction of an explosion near the target(s). If the high-trajectory shot is a miss or near miss, in Step 210, this is reported to the target. The Shooter 10 may also be notified in Step 211. The reporting of hits and misses can be configured based on different training situations. For instance in one training mode, the system sends feedback to the Shooter 10 after each shot so that the Shooter may learn from each shot and improve his marksmanship. In another training mode, such as simulating a firefight, this constant feedback from the system to the Shooter 10 may be both distracting and inappropriate. In such a situation, the messages to the Shooter 10 may be suppressed during the event and reported afterward.
It should be clear at this time that a shooting simulation system for personnel, unmanned systems, and vehicles has been provided that enables non-line of sight engagements and permits firing through obscurants and terrain features like bushes and tall grass. However the present invention is not to be considered limited to the forms shown which are to be considered illustrative rather than restrictive.
This a continuation of application Ser. No. 15/141,114, filed Apr. 28, 2016. This prior application is incorporated by reference in its entirety.
This invention was made with government support under contract no. W911QX-13-C-0026 awarded by the United States Army—Army Contracting Command (ACC) Aberdeen Proving Ground (APG). The government has certain rights in the invention.
Number | Name | Date | Kind |
---|---|---|---|
20070190494 | Rosenberg | Aug 2007 | A1 |
20140178841 | Carter | Jun 2014 | A1 |
Number | Date | Country | |
---|---|---|---|
20190249955 A1 | Aug 2019 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15141114 | Apr 2016 | US |
Child | 16390189 | US |