Aspects of the invention relate generally to firearm training systems and, more particularly, to firearm training systems that utilize a simulated virtual environment.
Training for marksmen and snipers, including military and law enforcement personnel, is often critical for the development of targeting and accuracy skills. However, live training often requires relatively expensive ammunition and the maintenance of shooting range facilities. Additionally, live training typically fails to provide a variety of targets, including moving targets. In order to provide additional avenues for firearms training, various shooting simulators have been developed. These conventional shooting simulators typically present images via projection screens or other image displays. Simulated projectiles or lasers are then typically identified as shots when contact is made with the display surface.
However, due to the utilization of two-dimensional surfaces for shot recognition and capture, accuracy is typically compromised in these conventional shooting simulators. These conventional shooting simulators also fail to accurately represent ballistic physics during the capture of shooting events. Additionally, these conventional shooting simulators fail to effectively utilize scopes to represent digital imagery within a virtual environment. For example, if a user were to look through a scope at an image display utilized in a conventional shooting simulator, the user will be presented with a pixilated magnified view of the image display. Accordingly, improved firearm training systems and associated methods are desirable.
Some or all of the above needs and/or problems may be addressed by certain embodiments of the invention. Embodiments of the invention may include firearm training systems and methods of using the same. In one embodiment, a firearm training system may be provided. The firearm training system may include a firearm, a scope, a tracking system, and at least one controller. The scope may be mounted to the firearm, and the scope may include a display. The tracking system may be configured to collect information associated with a position and an orientation of the firearm. The at least one controller may be configured to (i) determine, based at least in part upon the information collected by the tracking component, a viewpoint of the firearm, (ii) determine, based at least in part upon the viewpoint of the firearm, image data to be rendered by the display, and (iii) direct output of the image data for receipt by the display.
In accordance with another embodiment of the invention, a system for modifying a firearm for use in a simulation environment may be provided. The system may include a display, a tracking component, and at least one controller. The display may be configured to be attached to a scope associated with the firearm. The tracking component may be configured to be attached to the firearm, and the tracking component may be further configured to collect information associated with a position and an orientation of the firearm. The at least one controller may be configured to (i) determine, based at least in part upon the information collected by the tracking system, a viewpoint of the firearm, (ii) determine, based at least in part upon the viewpoint of the firearm, image data to be rendered by the display, and (iii) direct output of the image data for receipt by the display.
In accordance with yet another embodiment of the invention, a firearm scope may be provided. The firearm scope may include a housing, a display, and a video receiver. The housing may be configured to be mounted onto a firearm. The display may be mounted at least partially within the housing. The video receiver may be in communication with the display and configured to output video to be rendered by the display. The video receiver may receive the video from at least one controller configured to (i) determine a viewpoint of the firearm, (ii) adjust a position of the firearm within a virtual environment, (iii) determine the video to be rendered, and (iv) direct communication of the video to the video receiver.
In accordance with yet another embodiment of the invention, a method for providing a firearm simulation may be provided. Motion of a firearm may be tracked via a tracking component attached to the firearm. Based upon information collected by the tracking component, a viewpoint of the firearm may be determined by a controller, such as a controller including one or more computers. Based at least in part upon the determined viewpoint, the controller may determine or identify image data to be rendered by a display associated with a scope mounted onto the firearm. The determined image data may be output by the controller for receipt by the display, and the image data may be presented by the display.
Additional systems, methods, apparatus, features, and aspects may be realized through the techniques of various embodiments of the invention. Other embodiments and aspects of the invention are described in detail herein with reference to the description and to the drawings and are considered a part of the claimed invention.
Reference will now be made to the accompanying drawings, which are not necessarily drawn to scale, and wherein:
Embodiments of the invention now will be described more fully hereinafter with reference to the accompanying drawings, in which embodiments of the invention are shown. This invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the invention to those skilled in the art. Like numbers refer to like elements throughout.
Embodiments of the invention may include firearm training systems, apparatus, and methods for using the same. In certain embodiments, a scope may be mounted to a firearm, and a display may be associated with the scope. For example, the display may be integrated into, incorporated into, or at least partially situated within the scope. As another example, the display may be attached to the scope. A wide variety of suitable displays may be utilized as desired in various embodiments of the invention, such as a microdisplay, a liquid crystal display (“LCD”), a light emitting diode (“LED”) display, an organic light emitting diode (“OLED”) display, or other suitable display. The display may facilitate the presentation of image data, such as video, to a user of the firearm. For example, in a virtual environment, one or more displays external to the firearm (e.g., projection screens, etc.) may be utilized to present a virtual environment to a user. When a user looks through the scope, an enlarged representation of a portion of the virtual environment may be presented to the user via the display.
Additionally, at least one tracking system may be associated with the firearm. The tracking system may be configured to track motion, position, and/or orientation data associated with the firearm. A wide variety of suitable tracking systems may be utilized as desired in various embodiments of the invention, including but not limited to, a laser tracking system, an electromagnetic tracking system, a magnetic tracking system, an optical tracking system, an infrared tracking system, an ultrasonic tracking system, an inertia tracking system, a mechanical tracking system, a Global Positioning System (“GPS”) tracking system, or a combination of tracking systems. As desired in various embodiments, motion and/or movement of the firearm and/or scope may be tracked along any number of axes and/or in a wide variety of different directions. For example, movement may be tracked along three axes (e.g., the x, y, and z axes) with anywhere from two to six degrees of freedom.
Information collected by the tracking system may be utilized to determine a viewpoint associated with the firearm. For example, information collected by the tracking system may be provided to one or more controllers configured to process the information and calculate a viewpoint for the firearm and/or the firearm scope. In this regard, image data to be presented via the display may be determined. As desired in various embodiments of the invention, a position of the firearm may be adjusted or corrected prior to determining the image data to be presented. For example, an external display (e.g., one or more projection screens, etc.) may be configured to present a virtual environment from the perspective of a single calibration point. Prior to determining a viewpoint of the firearm, the position of the firearm may be corrected to the calibration point. In this regard, the image data presented via the display may correspond to that presented to the user via the external display, regardless of the user's movement within a simulation environment. This position correction may assist in facilitating a simulation environment in which multiple firearms are utilized.
A trigger mechanism or trigger actuation detection mechanism may also be associated with the firearm. For example, a trigger mechanism that facilitates the detection of a trigger pull event may be incorporated into a firearm. As another example, a trigger detection mechanism, such as a false or inert round of ammunition that detects actuation of a tiring pin, may be inserted into a firearm. Based upon the detection of a trigger actuation, a firing event may be identified, and an indication of the firing event may be communicated to one or more controllers. The controllers may utilize a wide variety of information associated with the firearm, scope, simulated ammunition, and/or simulated conditions (e.g., windage, elevation, etc.), as well as one or more ballistic models, to determine a ballistic path for the shot within the virtual environment. As a result of calculating a shot trajectory and/or a ballistic path, the accuracy of the simulated shot may be enhanced. Following the determination of information associated with a simulated shot, the image data output for presentation via the scope display may be updated.
System Overview
An example system 100 for facilitating firearms training will now be described illustratively with respect to
With reference to
Additionally, in certain embodiments, a plurality of firearms may be utilized in association with a training system 100. Each of the firearms may include or be associated with respective simulation components (e.g., a display, one or more motion tracking systems, etc.). In the event that the system 100 includes one or more display devices situated external to the firearms, such as the one or more projection screens 130 illustrated in
The scope 110 may be any suitable scope device, modified scope device, or replica scope device suitable for use in a shooting or training simulation. For example, in certain embodiments, the scope 110 may be a replica scope that has been modified to include or otherwise be associated with a display 115 and/or other simulation components. In other embodiments, the scope 11.0 may be an actual scope (i.e., a telescopic sight) that has been modified to include or otherwise be associated with a display and/or other simulation components. According to an aspect of the invention, a display 115 may be associated with the scope 110. In certain embodiments, the display 115 may be at least partially mounted within a housing of the scope 110. For example, a display 115 may be integrated into a scope 110 utilized in association with a modified simulation firearm. In other embodiments, such as embodiments in which a user utilizes a personal firearm, the display 115 may be removably attached to the scope. In yet other embodiments, a display 11.5 may be at least partially mounted within the housing of a scope 110 that is attached to an actual firearm when the firearm is utilized in conjunction with the simulation system 100.
The display 115 may be any suitable display device configured to present image data, such as video data, to a user of the firearm 105. Examples of suitable display devices include, but are not limited to, microdisplays, liquid crystal displays (“LCD's”), light emitting diode (“LED”) displays, organic light emitting diode (“OLED”) displays, active matrix OLED (“AMOLED”) displays, pico front or rear projection displays, stereoscopic displays, holographic displays, and/or cathode ray tube (“CRT”) displays. In certain embodiments, the display 115 may be configured to present high-definition imagery. Additionally, in certain embodiments, the display 115 may be at least partially mounted within the housing of the scope 110. In other embodiments, the display 115 may be removably attached to the scope 110. For example, the display may be attached to either end of the scope 110.
In certain embodiments, a video receiver 135 may be provided to route image and video signals to the display 115 for presentation to a user. For example, a video receiver 135 may receive image data from the one or more controllers 125, and the video receiver 135 may process and/or format the received image data. The processed and/or formatted image data may then be provided to the display 115. The video receiver 135 may be either a wired or wireless video receiver. For example, the video receiver 135 may be connected to the one or more controllers 125 via a video graphics array (“VGA”) cable, an Ethernet cable, a Universal Serial Bus (“USB”) cable, a mini-USB cable, a Firewire cable, a digital video interface (“DVI”) cable, a high definition multimedia interface (“HDMI”) cable, or other suitable cable or wired connection. As another example, the video receiver 135 may be connected to the one or more controllers 125 via a suitable Bluetooth connection, wireless Ethernet connection, Wi-Fi connection, or other wireless connection. In certain embodiments, one or more suitable communications interfaces (e.g., wired interfaces and/or wireless interfaces) 140 may facilitate communications between the video receiver 135 and the one or more controllers 125. Additionally, in certain embodiments, the video receiver 135 may be incorporated and/or integrated into the firearm 105 or the scope 110. In other embodiments, the video receiver 135 may be removably attached to the firearm 105.
As desired, an audio video (“AV”) receiver may be provided, and the AV receiver may process both audio and video data. The video data may be provided to the display 110, and the audio data may be provided to one or more suitable audio presentation devices (e.g., speakers, headphones, etc.). In certain embodiments, the AV receiver may receive separate audio and video signals. For example, separate audio and video signals may be received from the one or more controllers 125. As another example, a video signal may be received from the one or more controllers 125, and a separate audio signal may be received from one or more other controllers associated with the generation and presentation of a primary or main simulation scenario, such as a scenario presented via the one or more projection screens 130.
As desired, any number of suitable motion tracking systems 120 may be utilized in various embodiments of the invention. A motion tracking system 120 may be configured to monitor and/or track the motion, position, and/or orientation of the firearm 105 and/or the scope 110. In this regard, a viewpoint or perspective of the firearm 105 and/or the scope 110 may be determined, and image data may be determined and/or generated for presentation by the display 115. Examples of suitable motion tracking systems 120 that may be utilized include, but are not limited to, a laser tracking system, an electromagnetic tracking system, an optical tracking system, an infrared tracking system, an ultrasonic tracking system, an inertia tracking system, a mechanical tracking system, and/or a Global Positioning System (“GPS”) tracking system. In certain embodiments, a single motion tracking system may be utilized. In other embodiments, a plurality or combination of motion tracking systems may be utilized. In this regard, positional and/or orientation information for the firearm 105 and/or the scope 110 may be verified and/or accuracy may be improved.
A laser tracking system may include a laser emitter that is attached to or incorporated into the firearm 105 or the scope 110. For example, a laser emitter may be inserted into the barrel of the firearm 105. One or more laser detection devices, such as a laser camera or other laser detector, may also be attached to the firearm 105 and/or at one or more other positions suitable for viewing an area in which the laser emitter is pointed. For example, one or more laser detection devices may be utilized to identify an emitted laser that contacts and/or that is reflected off of the one or more projection screens 130. Based upon the capture of the emitted laser, positional and/or orientation data for the firearm 105 may be determined or calculated.
An electromagnetic tracking system may include one or more electromagnetic sources attached to the firearm 105 and one or more sensors utilized to monitor a calibrated area for position and motion of the electromagnetic sources. In this regard, positional and/or orientation data for the firearm 105 may be determined or calculated. In an optical tracking system, a video device may be calibrated to one or more unique optical markers attached to the firearm 105 and/or to one or more visual identifiers inherent to the firearm 105. Motion of the firearm 105 may be monitored by the video device, and information associated with the monitored motion may be translated into positional and/or orientation data for the firearm 105. With an infrared tracking system, a series of LEDs and/or reflective markers may be attached to the firearm 105. The LEDs may be utilized to emit infrared signals that are reflected by the markers. One or more infrared cameras may be utilized to monitor a field of view encompassing the firearm, such as a simulation area. As the firearm 105 is moved, the infrared camera may identify movement of the LEDs and/or reflective markers, and information associated with the identified movement may be translated into positional and/or orientation data for the firearm 105. In certain embodiments, movement of the LEDs and/or reflective markers may be identified relative to one another in order to provide a perspective for the firearm 105.
An ultrasonic tracking system may include one or more ultrasonic transmitters that are attached to the firearm 105. One or more audio sensors situated in predetermined positions may be utilized to measure the time-of-flight and/or the phase of the ultrasonic sound emitted by the ultrasonic transmitters. The measurements data may then be processed in order to determine distances between the sensors and the firearm, and the distances may be utilized to determine positional and/or orientation data for the firearm 105. With an inertia tracking system, one or more inertia sensors including gyro meters and/or accelerometers may be attached to the firearm 105. Motion or movement of the firearm 105 may be measured as inertia and translated into positional and/or orientation data for the firearm 105. A mechanical tracking system may include one or more mechanical devices and/or mechanical sensors that are attached to the firearm 105 such that movement of the firearm 105 directly affects the mechanical devices. Movement of the gun may be translated into movement of portions of the mechanical devices and may be translated into positional and/or orientation data. With a GPS tracking system, one or more GPS receivers may be attached to the firearm 105. A GPS receiver may precisely measure the timing of signals output by orbital GPS satellites. As the firearm 105 moves, the GAS receiver continually calculates positional information in real-time and tracks movement of the firearm 105. The movement data may then be translated into positional and/or orientation information for the firearm 105.
According to an aspect of the invention, movement and/or motion of the firearm 105 may be tracked along any number of axes utilizing two to six degrees of freedom. For example, motion may be tracked along an x, y, and/or z axis. Additionally, two, three, four, five, or six degrees of freedom may be utilized to track or monitor motion of the firearm 105. In this regard, motion may be selectively tracked as the firearm is moved forward and backward (e.g., closer to or further away from a projection screen), left or right, and/or up or down. Additionally, yaw, pitch, and/or roll of the firearm 105 may be tracked. The tracking of the firearm may be utilized to determine a viewpoint and/or perspective of the firearm in order to determine image data for presentation via the display 115.
As desired in certain embodiments, a wide variety of other components may be incorporated into, attached to, and/or utilized in conjunction with the firearm 105. For example, one or more communications interfaces 140 may be incorporated into and/or attached to the firearm 105 or the scope 110. The communications interfaces may include wired and/or wireless communications devices (e.g., transceivers, communications cards, etc.) that facilitate communication with the controllers 125. In this regard, data and/or information collected by simulation components attached to the firearm 105 (e.g., the motion tracking systems 120, a camera, a trigger recognition device, a keypad, etc.) may be communicated to the controllers 125. Additionally, data output by the controllers 125 (e.g., image data, calibration data etc.) may be received and distributed to the various simulation components.
In certain embodiments, a camera 145 may be integrated into or incorporated into the firearm 105 or the scope 110. The camera 145 may be utilized to detect and/or monitor grid patterns projected onto or output by an external display surface, such as the one or more projection screens 130, The data captured by the camera 145 may be processed in order to determine an orientation and/or perspective of the firearm 105. In this regard, the motion tracking of the firearm 105 may be verified. Additionally, a determination may be made as to whether the firearm 105 is positioned correctly.
Additionally, one or more trigger mechanisms 150 and/or trigger detection devices may be incorporated into, inserted into, and/or attached to the firearm 105. These mechanisms and/or devices may facilitate the detection of a trigger pull or a trigger actuation. Once a trigger actuation has been detected, an indication of the detected actuation may be communicated to the controllers 125, and a shot event may be registered. A wide variety of trigger mechanisms 150 and/or trigger detection devices may be utilized as desired in various embodiments of the invention. For example, a firearm 105 may be modified to include a trigger mechanism that includes one or more actuation sensors. As another example, an inert or dummy round may be inserted into a firearm 105, and the dummy round may include one or more suitable electrical components that detect actuation of the firing pin. An inert round is explained in greater detail below with reference to
A wide variety of other sensors may be incorporated into and/or attached to the firearm 105 as desired in various embodiments. For example, one or more sensors may be attached to the scope 110 in order to detect user adjustment made to the scope, such as windage adjustments, elevation adjustments, focus adjustments, zoom adjustments, and/or mode adjustments (e.g., a standard mode, a thermal imaging mode, a night vision mode, etc.). As another example, one or more sensors may be utilized to monitor the removal and/or insertion of magazines. As desired, information collected by the sensors may be communicated to the controllers 125 and utilized during the presentation of audio and/or image data during a simulation.
With continued reference to
With reference to
The data files 160 may include any suitable data that facilitates the operation of the controller 125 and/or the provision of a firearms simulation. For example, the data files 160 may include, but are not limited to, received motion and/or movement data for one or more firearms 105, calibration information for a simulation environment, information associated with types of rounds or ammunition to be simulated, information associated with desired scope settings, information associated with desired simulation conditions (e.g., wind, etc.), one or more ballistic models, stored image and/or audio data, and/or information utilized to generate and/or format image and/or audio data for presentation during a simulation.
The OS 162 may be a suitable software module that controls the general operation of the controller 125 and/or the execution of other software modules, for example, the tracking module 164, the video module 166, and/or the ballistics module 168. The tracking module 164 may be a suitable software module or application that facilitates the determination of a viewpoint of the firearm 105. The viewpoint of the firearm 105 may then be utilized to determine video to be presented via the display 115. As desired, the viewpoint of the firearm 105 may take a wide variety of position, orientation, and/or perspective information into account, as well as various settings associated with the scope 110 (e.g., a magnification setting, an operational mode, etc.).
In operation, the tracking module 164 may receive measurements and/or tracking data from the one or more motion tracking systems 120, and the tracking module 164 may utilize at least a portion of the received data to determine a viewpoint of the firearm 105. For example, received measurements data may be translated into positional and/or orientation information for the firearm 105. The positional and/or orientation information may then be utilized to determine a viewpoint for the firearm 105. For example, an area of the one or more projection screens 130 at which the firearm 105 is aimed or pointed may be determined. In certain embodiments, various tilts and/or angles of the firearm (e.g., pitch, yaw, roll) may be taken into consideration when determining the viewpoint. Additionally, in certain embodiments, a distance between the firearm 105 and a relevant projection screen 130 may be taken into consideration when determining the viewpoint.
In certain embodiments, the tracking module 164 may receive tracking data from a plurality of motion tracking systems 120, and the tracking module 164 may process the respective data to verify the positional and/or orientation information for the firearm 105. As desired, positional and/or orientation information may be averaged or otherwise processed in the event that tracking data is received from a plurality of sources. Additionally, in certain embodiments, grid pattern information may be received from the camera 145 and utilized to verify a viewpoint of the firearm 105.
In certain embodiments, such as embodiments in which a plurality of firearms are utilized in conjunction with a simulation, the tracking module 164 may correct or adjust respective positions of the one or more firearms. In this regard, image data that is determined or generated for display via the firearm scopes may correspond to the perspective illustrated via the external projection screens 130, which is the same for all users within the simulation environment. In other words, prior to determining a viewpoint of a firearm 105, the position of the firearm 105 may be corrected to a calibration point for the simulation environment. In this regard, the image data presented via the display 115 may correspond to that presented to the user via the projection screens 130, regardless of the user's movement within a simulation environment. In other embodiments, such as embodiments with a single user or embodiments in which each user has a personal external display (e.g., a personal monitor, a headset, etc.), a simulation scenario illustrated on the one or more external displays may be modified as the user moves within a simulation environment.
The video module 166 may be a suitable software module or application configured to determine and/or generate image data for presentation via the display 115. For example, the video module 166 may receive viewpoint information for the firearm 105, and the video module 166 may utilize the viewpoint information to generate image data for communication to the video receiver 135 and/or the display 115. As desired, the video module 166 may additionally utilize one or more settings associated with the scope (e.g., a magnification, a display mode, etc.) and/or an indication of a shot event to determine or generate image data for presentation. A wide variety of image data may be generated by the video module 166, including simulation scenario imagery, enhanced or enlarged imagery corresponding to a portion or section of the imagery displayed on the projection screens 130, imagery associated with a simulated view through the scope 110 (e.g., a reticle, etc.), and/or imagery associated with a simulated shot. In certain embodiments, the video module 166 may also include an audio component configured to determine or generate audio for presentation to a user of the firearm 105.
The ballistics module 168 may be a suitable software module or application configured to calculate or determine a simulated trajectory of a shot within a virtual environment. Once a shot event has been identified, information associated with the viewpoint and/or orientation of the firearm 105 may be provided to the ballistics module 168, and the ballistics module 168 may calculate a trajectory for the shot. As desired, the ballistics module 168 may utilize a wide variety of other information to calculate a trajectory for the shot. For example, information associated with the firearm 105 (e.g., caliber information, etc.), information associated with a simulated ammunition, elevation information and/or wind information may be provided to any number of suitable ballistic models to calculate or determine a trajectory for the simulated shot. Results of the simulated shot may then be provided to the video module 166 to facilitate the generation of image data associated with the simulated shot. As a result of calculating shots by the ballistics module 168 within a virtual environment, shot accuracy may be improved or enhanced. In conventional training systems, such as systems that detect a shot by detecting a laser projected onto a two-dimensional surface, the size of the laser point or other shot may be bigger than the target aimed at, thereby reducing accuracy and ballistic fidelity. By simulating shots directly within a virtual environment, external shot detection mechanisms may be bypassed, and overall accuracy may be improved.
A wide variety of suitable operations may be performed by the simulation components associated with a firearm 105 and/or by the one or more controllers 125 to facilitate provision of a training simulation. The various operations described above for a controller 125 are provided by way of example only. Another example of the operations that may be performed by the controllers 125 is described in greater detail below with reference to
With continued reference to the controller 125 of
With continued reference to
Additionally, in certain embodiments, one or more simulation controllers 175 may be provided. A simulation controller 175 may be a suitable processor-driven device configured to control the presentation of a simulation via the one or more projection screens 130. As desired, a simulation controller 175 may include components similar to those described above for the controller 125. For example, a simulation controller 175 may include one or more processors and/or memory devices, and simulation software may be executed by the processors to facilitate the generation of image data to be. displayed via the projection screens 130. In certain embodiments, a simulation controller 175 may be separate from one or more controllers 125 associated with individual firearms. In other embodiments, the simulation controller 175 may be combined with one or more other controllers 125.
Those of ordinary skill in the art will appreciate that the system 100 shown in and described with respect to
As set forth above, a wide variety of suitable techniques may be utilized as desired to modify firearms for use in a simulation environment.
With continued reference to
Although
As desired, the communications interfaces 315 associated with the scope 300 may also be utilized to facilitate communication of measurements data to one or more controllers. For example, the communications interfaces 315 may be in communication with a trigger mechanism 320, a motion tracking system 325, and/or a camera 330. In this regard, shot indication data, motion data, and/or grid pattern data may be communicated to one or more controllers for processing. Although
Indeed, a wide variety of different firearm configurations may be utilized in accordance with various embodiments of the invention. These configurations should not be limited to those described in
As shown in
During a simulation, a user may move a firearm, such as firearm 610a, within the simulation environment 600. As the firearm is moved, the firearm 610a may be pointed at different portions of the external displays 605. Accordingly, the viewpoint of the firearm 610a may change as the firearm 610a is moved. During the generation or determination of image data to be presented via a scope-mounted display associated with the firearm 610a, the position of the firearm 610a within the simulation environment 600 may be autocorrected to the calibration point 615. For example, an area or portion of the external displays 605 at which the firearm 610a is aimed may be determined. A viewpoint of that area or portion from the perspective of the calibration point may then be determined as the viewpoint of the firearm 610a. In other words, regardless of the user's position within the simulation environment, it may be assumed that the user is positioned at the calibration point 615. Accordingly, the imagery illustrated via the scope-mounted display may correlate or correspond to the imagery presented to the user via the external displays 605.
Operational Overview
At block 705, communication may be initiated with a host controller, such as a host computer. In this regard, motion tracking information may be communicated to the host controller, and the image data may be received from the host controller for presentation via the display associated with the scope.
At block 710, interaction with one or more tracking systems (e.g., a laser tracking system, an electromagnetic tracking system, an optical tracking system, etc.) may be initiated. Each tracking system may be configured to track motion and/or movement of the firearm 105 within a simulation environment. In certain embodiments, a tracking system may be configured to determine position and/or orientation information for the firearm 105. For example, as illustrated in optional block 715, position and/or orientation information for the firearm 105 may be determined, and the position and/or orientation information may be communicated to the host controller at block 720.
In other embodiments; tracking information collected by a tracking system may be provided or communicated to the host controller at block 720, and the host controller may determine position and/or orientation information for the firearm 105. Regardless of the component or system that determines the position and/or orientation information for the firearm 105, a viewpoint of the firearm 105 or the firearm scope may be determined by the host controller, and the host controller may generate image data in accordance with the determined viewpoint. The image data may be output by the host controller for receipt by the firearm 105, and the firearm 105 may receive the image data at block 725. The received image data, such as a received image feed, may then be output by the display for presentation to the user at block 730. Accordingly, when a user views the scope associated with the firearm 105, the user may view the image data output by the host controller.
At block 735, a determination may be made as to whether the image is calibrated. In other words, a determination may be made as to whether the firearm 105 has been properly calibrated within the simulation system and/or the virtual environment. If it is determined at block 735 that the image is not calibrated, then operations may continue at block 740. At block 740, the firearm 105 and/or the position of the firearm 105 may be calibrated within the virtual environment. For example, a user may be prompted to utilize a suitable calibration routine (e.g., aiming the firearm 105 at the four corners of an external display, etc.) to calibrate the firearm 105. Operations may then continue at block 725 described above.
If, however, it is determined at block 735 that the image is calibrated, then operations may continue at block 745. At block 745, a determination may be made as to whether a trigger associated with the firearm 105 has been pulled or actuated by a user. In other words, a determination may be made as to whether a shot has occurred. As described in greater detail above with reference to
At block 750, a shot indication may be communicated from the firearm 105 to the host controller. The shot indication may be processed by the host controller to simulate a shot within a virtual environment. For example, one or more ballistic models may be utilized to determine a trajectory of a shot within the virtual environment. Image data (e.g., an image feed) may then be updated by the host controller to reflect the shot, and the updated image data may be output by the host controller for receipt by the firearm 105. The firearm 105 may receive the updated image data at block 755 and present information associated with a virtual shot via the display at block 760. Operations may then continue at block 765.
At block 765, a determination may be made as to whether a position and/or orientation of the firearm 105 has changed. For example, a determination may be made as to whether the firearm 105 has been moved or whether a tracking system has detected movement of the firearm 105. If it is determined at block 765 that the position and/or orientation of the firearm 105 has not changed, then operations may continue at block 730, and image data may continue to be output. If, however, it is determined at block 765 that the position and/or orientation of the firearm 105 has changed, then operations may continue at block 715 (or block 720). In this regard, a new viewpoint of the firearm 105 may be determined, and the image data output by the display may be updated to reflect the new viewpoint.
The method 700 may end following the completion of a simulation scenario in which the firearm 105 is utilized.
At block 805, the controller 125 may initiate communication with a firearm and/or a scope, such as the firearm 105 and/or the scope 110 illustrated in
At block 810, one or more firearm and/or scope parameters may be determined. In certain embodiments, the one or more parameters may be determined or identified within suitable ballistics simulation software, such as the ballistics module 168 illustrated in
At block 815, the controller 125 may initiate communication with one or more suitable tracking systems associated with the firearm 105 and/or the scope 110, such as the motion tracking system 120 illustrated in
At block 825, which may be optional in certain embodiments of the invention, a position of the firearm 105 within the virtual environment may be corrected or adjusted. For example, the position of the firearm 105 may be set to a predetermined position, such as a predetermined calibration point. In this regard, a determined viewpoint for the firearm 105 may correspond with the viewpoint presented via one or more external (i.e., non-firearm-mounted) displays.
At block 830, an image feed for display via a scope-mounted display may be determined and/or generated based at least in part upon the determined viewpoint for the firearm 105. The image feed may then be output at block 835 for receipt by the firearm 105 and presentation to the user via the scope-mounted display. At block 840, a determination may be made as to whether the image communicated to the firearm 105 is calibrated. In other words, a determination may be made as to whether the firearm 105 has been properly calibrated within the simulation system and/or the virtual environment. If it is determined at block 840 that the image is not calibrated, then operations may continue at block 845. At block 845, the firearm 105 and/or the position of the firearm 105 may be calibrated within the virtual environment. For example, a user may be prompted to utilize a suitable calibration routine (e.g., aiming the firearm 105 at the four corners of an external display, etc.) to calibrate the firearm 105. Operations may then continue at block 830 described above.
If, however, it is determined at block 840 that the image is calibrated, then operations may continue at block 850. At block 850, a determination may be made as to whether a shot event has occurred. For example, a determination may be made as to whether a shot indication has been received from the firearm 105. If it is determined at block 850 that a shot event has not occurred, then operations may continue at block 865 described in greater detail below. lf, however, it is determined at block 850 that a shot event has occurred, then operations may continue at block 855.
At block 855, the shot event may be registered within the ballistic simulation software, and a trajectory of the shot may be calculated or determined. As desired, a wide variety of ballistic factors, such as weather conditions, elevation, ammunition type, firearm type, etc., may be taken into consideration during the calculation or determination of a shot trajectory. Once a shot trajectory has been calculated, operations may continue at block 860, and the image feed may be adjusted or modified to correspond with the registered shot. Image information associated with the simulated shot may then be output for receipt via the firearm 105 and presentation via the scope-mounted display. Operations may then continue at block 865.
At block 865, a determination may be made as to whether a position and/or orientation of the firearm 105 has changed. For example, a determination may be made as to whether motion or movement information has been received from one or more tracking systems. As another example, a determination may be made as to whether updated position and/or orientation information has been received. If it is determined at block 865 that the position and/or orientation of the firearm 105 has not changed, then operations may continue at block 835, and image data may continue to be output for receipt by the firearm 105. If, however, it is determined at block 865 that the position and/or orientation of the firearm 105 has changed, then operations may continue at block 820, and a new viewpoint for the firearm 105 may be determined. In this regard, the image data output for display by the scope may be updated to reflect the new viewpoint.
The method 800 may end following the completion of a simulation scenario associated with the controller 125.
The operations described and shown in the methods 700 and 800 of
Various block and/or flow diagrams of systems, methods, apparatus, and/or computer program products according to example embodiments of the invention are described above. It will be understood that one or more blocks of the block diagrams and flow diagrams, and combinations of blocks in the block diagrams and flow diagrams, respectively, can be implemented by computer-executable program instructions. Likewise, some blocks of the block diagrams and flow diagrams may not necessarily need to be performed in the order presented, or may not necessarily need to be performed at all, according to some embodiments of the invention.
These computer-executable program instructions may be loaded onto a special purpose computer or other particular machine, a processor, or other programmable data processing apparatus to produce a particular machine, such that the instructions that execute on the computer, processor, or other programmable data processing apparatus create means for implementing one or more functions specified in the flow diagram block or blocks. These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means that implement one or more functions specified in the flow diagram block or blocks. As an example, embodiments of the invention may provide for a computer program product, comprising a computer-usable medium having a computer-readable program code or program instructions embodied therein, said computer-readable program code adapted to be executed to implement one or more functions specified in the flow diagram block or blocks. The computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational elements or steps to be performed on the computer or other programmable apparatus to produce a computer-implemented process such that the instructions that execute on the computer or other programmable apparatus provide elements or steps for implementing the functions specified in the flow diagram block or blocks.
Accordingly, blocks of the block diagrams and flow diagrams support combinations of means for performing the specified functions, combinations of elements or steps for performing the specified functions and program instruction means for performing the specified functions. It will also be understood that each block of the block diagrams and flow diagrams, and combinations of blocks in the block diagrams and flow diagrams, can be implemented by special purpose, hardware-based computer systems that perform the specified functions, elements or steps, or combinations of special purpose hardware and computer instructions.
Many modifications and other embodiments of the invention set forth herein will be apparent having the benefit of the teachings presented in the foregoing descriptions and the associated drawings. Therefore, it is to be understood that the invention is not to be limited to the specific embodiments disclosed and that modifications and other embodiments are intended to be included within the scope of the appended claims. Although specific terms are employed herein, they are used in a generic and descriptive sense only and not for purposes of limitation.
The present application claims the benefit of U.S. Provisional Application Ser. No. 61/308,169, filed Feb. 25, 2010, and entitled “Firearm Scope and System for Simulated Environments,” the contents of which are incorporated by reference herein in their entirety.
Number | Date | Country | |
---|---|---|---|
61308169 | Feb 2010 | US |