BACKGROUND
Shooting and target practice ranges are known in the art. In a typical shooting range, a user is presented with a target, fires a series of rounds, then has to retrieve the target to determine their accuracy. Some improvements have been made, including more immediate feedback on accuracy after each round fired and targets that are remote controlled or on a time-delay system.
SUMMARY
An aspect of the disclosure relates to a multichannel controller for controlling a target in a target system. In one embodiment, a multichannel controller is configured to control a target system and includes a user input interface that receives a user input for the multichannel controller, wherein the user input is a command to control one or more targets in the target system and a processor that generates the command to send to the one or more targets in the target system and a command translation unit that relays the command to the one or more targets, wherein the command comprises a motion sequence.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a view of a multi-target system, in accordance with an embodiment of the present invention.
FIG. 2 is a diagrammatic view of a target controller in accordance with an embodiment of the present invention.
FIG. 3 is a view of a multichannel controller in accordance with an embodiment of the present invention.
FIG. 4A is a view of a control mode selector screen on the multichannel controller in accordance with an embodiment of the present invention.
FIG. 4B is a view of a stored motions selector screen on the multichannel controller in accordance with an embodiment of the present invention.
FIG. 5 is a diagrammatic view of a multichannel controller accessing an applications store over a network in accordance with an embodiment of the present invention.
FIG. 6 is a view of a multi-target system controlled by a multichannel controller in accordance with an embodiment of the present invention.
FIG. 7 is a flow chart depicting a motion creation process in accordance with an embodiment of the present invention.
FIGS. 8A-F are views of a plurality of control modes for a multichannel controller in accordance with an embodiment of the present invention.
FIGS. 9A-D are views of a plurality of sub motion sequences in accordance with an embodiment of the present invention.
FIGS. 10A-C are views of a plurality of device control setting screens on a multichannel controller in accordance with an embodiment of the present invention.
FIGS. 11A-D are views of a plurality of profile management screens on a multichannel controller in accordance with an embodiment of the present invention.
FIGS. 12A-D are views of a plurality of motion management screens on a multichannel controller in accordance with an embodiment of the present invention.
FIGS. 13A-D are views of a plurality of customize new motion screens on a multichannel controller in accordance with an embodiment of the present invention.
DETAILED DESCRIPTION OF ILLUSTRATIVE EMBODIMENTS
Multichannel controllers are commonly used to control a wide variety of systems. For example, a multichannel controller can be used to control a target, such as a target in a shooting range or as used to train police recruits. In such a case, one channel of the multichannel controller may be used to control side-to-side or front-to-back motion of the target system, and another channel of the multichannel controller may be used to trigger the target to pop into view for the shooter. One method of providing multichannel control has included using controllers with physical joysticks. Positioning of the physical joysticks causes signals to be sent to the system being controlled.
FIG. 1 shows, in one embodiment, a target shooting range 100 with two target systems 102 and 104. Target range 100 may, in an alternative embodiment, contain only one target system, for example just target 102 or more than two target systems. In one embodiment, target range 100 contains a movable target system 102 that includes a bulls-eye target structure 106 that is on top of a stand 108 with a height measuring length 110. While a bulls-eye target structure 106 is shown in FIG. 1, as an exemplary embodiment, another image or shape could be used. For example, a user of the target range 100 may want to have target structure with images of ducks or deer for shooting, in another embodiment.
Movable target system 102 may also comprise, in one embodiment, a motorized base 112 with a communicator 114 and four wheels 116.
The communicator 114 may, in one embodiment, be a Wi-Fi wireless communication system. In another embodiment, the communicator may be an alternate RF-based or NFC-based communication system. The communicator 114 receives communications from a controller, either user-input commands or preprogrammed commands, which indicate directions of movement for the moveable target system 102.
The target range 100 may also include a fixed target system 104 that includes a bulls-eye target structure 106 that is attached to an expandable base 120. The expanadable base 120 moves the target closer to a wall 118 or further from a wall 118 along an expandable range 124. The fixed target system 104 is fixed at a fixed point 122 on the wall 118. The target is able to move back and forth along expansion range 124 but may not move along the wall 118 beyond the fixed point 122. However, in another embodiment, the fixed target system 104 could be fixed at a fixed point 122 on a wall 118 but able to rotate in a semi-circle by movement of a fixed support structure. Other target structure movement and fixing means that are appropriate could also be used in the target range 100.
In one embodiment, target range 100 contains controllable target systems, such as moveable target system 102 and fixed target system 104 that are controlled by a controller, such as controller 200 (shown in FIG. 2). In one embodiment, controller 200 contains a communications interface 202 that communicates with the processor 204. The processor 204 communicates with a touch screen 206. The processor 204 also communicates with the user interface 208 via the touch screen 206. In one embodiment, processor 204 also contains memory 210. Memory 210 may store, in one embodiment, applications 212, control modes, 214 and motions 216. Additionally, in another embodiment, memory 210 may store other applications not related to the target control application. In one embodiment, the controller 200 communicates with a command translation unit 218 that sends commands to target systems 220-1, 220-2 and any other targets that may be in communication with the controller 200, up through and including, target system N 220-n. These commands may be sent using Wi-Fi communication, other RF communication or NFC communication means. Additionally, any other appropriate communication techniques may also be employed in accordance with the embodiments of the present invention.
FIG. 3 shows controller 200 with touch screen 302 and main portion 304 where a user may enter commands, or otherwise interact, with the touch screen 302. Controller 200 also contains an icons portions 306 with multiple icons 308. These multiple icons 308 may, for example in one embodiment, connect the user to different targets and show the status of different targets. Additionally, the multiple icons 308 may connect the user to different pre-programmed movement sequences. Further, the user may, in another embodiment, be able to program the multiple icons 308 such that they comprise a combination of targets and movement sequences. In one embodiment, more than the depictured five icons may be stored in the icons portions 306 and arrow 310 shows the left and right motion capability of the icons portions such that a user of the controller 200 can scroll back and forth along the directions of the arrow 310 to reveal more icons 308.
Depending on the user's preference, they may, as shown in FIG. 4A, choose a different control mode with which to control the multiple targets in the target system 100. For example, FIG. 4A shows that a user may be able to select a touch pad 402, a joystick 404, a trackball 406, a touchpad slider 408, a touchpad wheels 410, a joystick wheels 412, or a trackball slider 414. Additionally, a user may be able to download additional control modes not shown in FIG. 4A by clicking on the app store icon 416. The control modes shown in FIG. 4A are only exemplary and the user could use additional means of controlling the targets in another embodiment. For example, in a further embodiment, the user could use a voice activated control mode where the user could issue commands for the movement of targets via voice input. In a voice input embodiment the multichannel controller may also comprise a microphone portion or it may comprise an input portion for the user to add an external microphone. In another embodiment, the user could use the movement of the controller itself to issue commands to a target, for example by the use of gyroscopes and accelerometers in the multichannel controller, for example where the multichannel controller is a mobile phone.
FIG. 4B shows a depiction of stored motions that the user may have on the controller 200. For example, the user may record motions generated by the use of any of the control modes shown in FIG. 4A, or other control modes not envisioned in FIG. 4A. Additionally, the user may control the targets by preprogrammed motions, for example as shown in FIG. 4B, a pop-up 418 motion, a pop-out 420 motion, or a side-to-side 422 motion. In one embodiment, the pop-up 418 motion triggers the target to pop-up from ground level to a shooting height or flip up from ground level to a shooting height. In one embodiment, the pop-out 420 motion may be used for a target fixed to a wall wherein the target pops out from the wall in a sideways manner such that it becomes available for a user to shoot at. In one example, the side-to-side 422 motion might indicate to a target that the user wants to move to engage the motor and move in a left to right or forward to backward motion across a target field. In another embodiment, the side-to-side motion may, for a fixed target, indicate that the user wants the target to sway from side to side from the fixed position therefore creating a more difficult target for a user to shoot at. As also shown in FIG. 4B, the user can, in one embodiment, select app store icon 424 in order to obtain additional motions. For example, a user may download motions recorded by other users or preprogramed by other users through the app store, accessed by app store icon 424.
FIG. 5 shows, in one embodiment, how a user may interact with an application store 502 using their controller 200. In one embodiment, the application store 502 is accessible over a network 500. The network 500 may, in one embodiment, be accessed by use of the Internet. The application store 502 may, in one embodiment, include a motions database 504 and a control mode database 506. The motions database 504 may, in one embodiment, offer the user a selection of movement sequences for purchase or download to the controller 200. The motions database 504 may, in one embodiment, consist of movements created by a manufacturer of controllers or targets, or they may be created by other users of the target system 100 or of the controller 200. Similarly, in another embodiment, the control mode database 506 contains a series of control modes that are created by users of the target system 100 or controller 200 or created by the manufacturer for the target system 100 or controller 200 available for the user to download or purchase. Additionally, in another embodiment, the application store 502 may contain access to purchase modes and motions 508 where purchase modes and motions 508 include control modes and motions that the user has already purchased for their controller 200. In one embodiment, the availability to re-download these purchase control modes and motions allows a user to recover modes and motions lost in the event that their controller loses functionality and needs to be reset to factory conditions. In another embodiment, the application store 502 contains a cloud storage portion where the user can store their saved motions and modes 510 that they created for use on their controller 200 or accessible on another controller 200, for example by entering a user name and password.
FIG. 6 shows a target environment 600 that has multiple targets that interact with the controller 200. In one embodiment, the target system 600 provides a multiuser environment where one user can control the actions of the one or more targets using control modes on the controller 200 and another user can interact with the targets using shooting mechanism 610. In one embodiment, shooting mechanism 610 is a gun (for example, a pistol or a rifle). However, in another embodiment, shooting mechanism 610 may be a NERF® gun or other toy pistol for shooting with targets. In a further embodiment, the shooting mechanism 610 could be a bow and arrow or any other suitable weapon or replica thereof. Embodiments of the present invention may also be used with any other form of targets and shooting, for example such as a dart and target board system.
Additionally, while FIG. 6 only shows bullseye targets, any other appropriate target could be used. For example, in a target system designed for hunters, targets may comprise images of animals for sport. In another embodiment, for example, wherein the target system 600 is used as a training operation for policemen, the targets may comprise images of criminals and/or bystanders such that policemen can be trained to recognize targets from non-target items in a short span of time. FIG. 6 also shows that the controller can distinguish from fresh target 602 and hit targets 604. For example, the controller shows fresh target 602 as an empty box on controller. Whereas hit target 604-1 and 604-2 are shown as “hit” on the controller 200. This may be accomplished, in one embodiment, by a communicator sending an indication from the target to the controller indicative of a hit registered to the target.
Additionally, as shown in FIG. 6, the system can identify where the hits have occurred on the targets and thus, in a multiuser system, may be able to keep score for different shooters. Further, the system may be able to, in another embodiment, identify when the hits occurred either relative to a sequence of hits or absolutely relative to a time sequence on the controller. The target system 600 may comprise multiple targets that either move or are fixed either along walls or along the floor or otherwise movable throughout the system. For example, arrow 606 shows a movement indication of target 604-2 moving from right to left across the target range 600. Additionally, movement indication 608 shows that the target 604-2 has flipped from an upright position to a downward position after, for example, being hit by a user. Additionally, target 604-1 shows a movement indication 606 showing that target 604-1 has moved rapidly from being engaged with an adjoining wall out into the target range. In one embodiment, the controller 200 shows target representations 614 on the touch screen of the target. These target representations could be of, as shown in FIG. 6, a rectangular shape just indicating the existence of a target. Additionally, target representation 614 could also be a visual representation of the target itself. For example, in the police target system discussed previously, the indication on the touch screen of the target representation 614 could comprise different images for criminals and/or bystanders, for example.
Additionally, controller 200 may, in another embodiment, comprise indications of hit feedback 616. One embodiment shown in FIG. 6 comprises hit feedback 616 that is consistent with the target representation 614 changing color to indicate that a hit has been delivered to the target. However, hit feedback 616 could also comprise, in another embodiment, a flashing light or, in a further embodiment, indication of a number of points or an accuracy representation showing how accurately a user hit the target. In a further embodiment, the hit feedback could include an indication of where on the target a user successfully hit. Additionally, the hit feedback 616 could show an indication of the time that it took for a user to hit the target or the time between successful hits to a target. For example, as discussed above with police training, this indication of how accurately and how long it took to hit a target may be important in determining feedback or training for an individual officer. In a further embodiment, the feedback 616 is not hit feedback, but is feedback concerning the activity of the targets, for example that targets 604-1 and 604-2 have been interacted with by a user, for example that target 604-1 has popped out from the wall and that target 604-2 has already popped up and back down.
FIG. 7 depicts a flow chart where a user may use the controller 200 to give movement instructions to a target. The user starts in box 702 by selecting a target. The user then enters a new action sequence 704. The action sequence may consist of purchasing a motion 708, using a stored motion 710, or creating a new motion 706. In the event that a user wants to enter more than one action sequence for a specific target, the user can then enter another action 712 and repeat the process of boxes 704-710. In this way, the user can enter a series of modular motions to generate a unique target motion sequence. Once the user finishes entering motions for a specific target, the user can then add additional motion configurations as shown in block 714. In one embodiment, an additional motion configuration 714 may comprise changing a preset speed at which a target may move or may consist of entering a repeat sequence, for example, for a target to sway from left to right repeatedly. Once the user has finalized an action sequence for a specific target, they may go on to enter motions for another target 716. Once the user has entered all of the motions for all of the targets they may then save the motion 718. Additionally, in a further embodiment, another motion configuration may comprise triggering a motion based on an event in the target environment. For example, in one embodiment, a movement on a second target may be triggered by a user successfully hitting the first target.
As described above, the user may preprogram action sequences for use with the controller. However, in another embodiment, the user may choose to select a control mode for manual control of the target. For example, in a multiuser system, one user may actively control the targets while another user attempts to shoot the targets. FIGS. 8A-8F depict multiple control mode options for this manual control of the target. Additionally, in another embodiment, the user may use any of control modes 8A-8F to record motions (for example, for sale in the application store 502 or through later usage with the controller 200). FIG. 8A shows one embodiment of a control mode that comprises a joystick 800. FIG. 8B shows another embodiment of the control mode that comprises a trackpad 802.
FIG. 8C shows, in another embodiment, a control mode comprising a touchpad slider mode with a forward/backward slider slot 804 and a movable slider icon 806 and a left/right slider icon 808 with a movable slider icon 812. The touchpad slider mode may also contain, in one embodiment, an up/down trigger button 810 that might trigger the target to pop up from a lying down position into an upright position. However, in another embodiment, this up/down trigger button 810 could be replaced by the user touching the touch screen to trigger the springing motion of a target.
FIG. 8D shows a touchpad wheel control mode for use with the controller 200 that comprises a forward/backward wheel 814 and a left/right wheel 816. The touchpad wheel mode also contains an up/down trigger button 818. In another embodiment, the up/down motion of a target could be triggered by a touch on the touchscreen by a user. FIG. 8E shows a joystick and wheels control mode with a joystick 820 and a forward/backward wheel 822 and an up/down trigger button 826. FIG. 8E also shows an up/down trigger button 826, however, the up/down motion of a target could also be triggered by touching the touch screen or could also be triggered by the joystick 820. FIG. 8F shows a trackball and slider mode with a trackball 828 and a forward/backward slider slot 830 on a movable slider icon 832, and a left/right slider slot 834 with a movable slider icon 832. The trackball slider mode also contains an up/down trigger button 838.
FIG. 9 shows a series of motion sequences that a user might encounter when programming their own motion sequences. FIG. 9A, for example, shows a forward/backward motion 902. FIG. 9B shows a left/right motion 904. FIG. 9C shows an up/down motion 906. In another embodiment, the motion of FIG. 9C is a trigger motion which might trigger a target to spring from a lying down position to an upright position. FIG. 9D shows an arc or a turn motion 908 that a user may use to indicate to a target that it should turn to the left or the right or sway to the left or the right in an arc motion. These basic motions of FIG. 9 may, in one embodiment, be combined by a process such as that outlined in FIG. 7, described above, to generate a unique motion sequence for a user.
FIG. 10A shows a screen where the user can set the device orientation of controller 200 where a user may select to have the icons portion at the bottom of the screen as shown by icon 1004, or on the right side of the screen as shown in icon 1006, or on the top of the screen as shown in icon 1008, or on the left of screen as shown in icon 1010.
FIG. 10B shows a speed setting screen 1012 wherein the user can set a maximum speed of left to right movement 1014 and the user may also set a maximum forward to backward speed 1016. The user can set these maximum speeds with respect to the maximum speed of the motor indicated by the 0-100% bar.
FIG. 10C shows a position lock screen 1018 where a user may lock the position of a target such that, for example, it cannot move from left to right 1020, it cannot move forward to backward 1022, or it cannot move up or down 1024.
FIG. 11A shows, in one embodiment, an ability to manage profiles using the controller 200. For example, on a manage profile screen 1100 the user may save profile 1102, load a profile 1104, or delete a profile 1106.
FIG. 11B shows a save profile screen 1108 wherein a user may save their current settings as a profile 1110, or save current settings as an existing profile 1112. The user may then, as shown in FIG. 11C, load saved profiles using load saved profiles screen 1114 and could select any of profiles 1116-1 to 1116-5. In another embodiment, the user may have more or fewer profiles than shown in FIG. 11C. The user may also delete saved profiles as shown in FIG. 11D using the delete saved profiles screen 1118. In this screen, the user may delete any of profiles 1120-1 to 1120-5 that the user no longer needs or wants on their controller 200.
FIGS. 12A-12D show a series of manage motion screens. Manage motion screen 1122 shows that a user may use the controller 200 to create motion 1124, perform save motion 1126, or delete saved motion 1128. Create motion screen 1200 allows a user to create a new motion as described, for example in FIG. 7, or by using the record motion tab 1202. The ability to record a motion 1202 using any of the control modes previous discussed, or program a motion as described in FIG. 1206 gives the user an ability to come up with the exact motion sequence they desire. Once the user comes up with a new motion sequence, the user may then save motion using save motion icon 1204. FIG. 12C shows that the user may perform a series of motions using perform save motion page 1206. The user may select any of motions 1208-1 to 1208-5 for performance. Upon selecting a saved motion to use, the controller 200 will then communicate with a series of targets on a target range. For example, in one embodiment, target range 600 which will then perform the save motions. Additionally, the user may delete motions that they no longer wish to use as shown in FIG. 12D on delete saved motion page 1210. The user may delete any of displayed motions 1212-1 to 1212-5 to remove it from the controller 200.
FIGS. 13A-13D show another embodiment comprising an interface for creating a new motion, for example, by programming a new motion according to flow chart 700. The user will start with a create new motion page 1300 where the user will need to choose a template that they can either click to choose a template 1302 or start from scratch 1304. If the user chooses to choose a template 1302, the user may start from a downloaded template from the application store, or one of a series of preprogrammed templates within the controller 200. The user may then alter the details of these templates, or add an additional motion to the template, or change the settings of chosen template.
However, in another embodiment the user may choose to create a brand new motion. In which case, the user may, in one embodiment, encounter, as shown in FIG. 13B, an add motion type screen 1306. On the add motion type screen 1306, the user can select from a series of basic motions as shown in FIG. 9 (for example forward or backward icon 1308, left or right icon 1310, up or down icon 1312, or arc or turn icon 1314). Once the user has selected a motion, the user moves to the customize motion screen 1316 shown in FIG. 13C. FIG. 13C shows that the user has, for example, chosen the arc or turn motion and now can customize specifically how that motion will command a target to move across an area. The user will see a left to right axis 1320 and a forward to backward axis 1322. In FIG. 13C, these axes are shown from going 0-10 feet. However, in another embodiment where the target range comprises, for example, several hundred yards, these axes can be resized (for example, by touching axes 1320 or 1322) to set the parameters of the target field. In another embodiment, the user may import settings from a different profile where the user already programmed target sequence for a specific target environment of their choice.
The user may also see, in another embodiment, the current motion 1324 as depicted in the current motion sequence. The user may then change a series of motion axes in order to get exactly the right arc motion that they want, in one embodiment. For example, if the user wants a target to move further on the forward to backward axis 1322 than on the left to right axis 1320, the user may change motion axes 1328 to pull the arc forward or backwards. Additionally, if the user wants to have the arc move further on the axis 1320 than on the axis 1322, they user may engage motion axis 1326 to pull the arc either to the left or the right. Additionally, if the user wants to change the depth of the arc, the user may engage motion axis 1330 to make the arc either deeper or wider according to their preferences. This customized motion screen 1316 thus allows the user to get exactly the right motion that they want for the target of their choice. Customized motion screen 1316 only shows left/right axes 1320 and forward/backward axes 1322. However, in another embodiment, the screen may also show a three dimensional representation that includes an up/down axes or may allow the user to select a point during the motion where the user target will be triggered to move up or down.
FIG. 13D shows that the user can program a series of submotions comprising the motions shown in FIG. 9 to create a customized target motion sequence of their choice. Current motion screen 1322, as shown in FIG. 13D, shows that the user has only indicated one motion in their motion sequence an arc to the back and to the right 1334. At this point, the user may choose to save their motion sequence 1335 at another submotion 1336 or discard the motion 1338.
Although the present invention has been described with reference to preferred embodiments, workers skilled in the art will recognize that changes may be made in form and detail without departing from the spirit and scope of the invention.