This subject invention relates to mobile, remotely controlled robots, and weaponized robots.
Mobile, remotely controlled robots are often equipped with new technologies and engineered to carry out some missions in a more autonomous manner.
iRobot, Inc. (Burlington, Mass.) and the Boston University Photonics Center (Boston, Mass.), for example, demonstrated a robot equipped with sensors that detect a gunshot. The robot head, upon detection of a shot, swiveled and aimed two clusters of bright-white LEDs at the source of the shot. See “Anti-Sniper/Sniper Detection/Gunfire Detection System at a Glance”, by David Crane, defensereview.com, 2005, incorporated herein by this reference. See also U.S. Pat. Nos. and Published Patent Applications Nos. 5,241,518; 7,121,142; 6,999,881; 5,586,086; 7,139,222; 6,847,587; 5,917,775; 4,514,621; and 2006/0149541, all of which incorporated herein by this reference.
The assignee hereof has devised a robot with a weapon which can be fired by the operator controlling the weapon. See, e.g., U.S. patent application Ser. No. 11/543,427 entitled “Safe And Arm System For A Robot”, filed on Oct. 5, 2006, incorporated by reference herein. The following co-pending patent applications by the assignee of the applicants hereof are hereby incorporated by this reference: U.S. patent application Ser. Nos. 12/316,311, filed Dec. 11, 2008; 11/543,427 filed Oct. 5, 2006; 11/732, 875 filed Apr. 5, 2007; 11/787,845 filed Apr. 18, 2007; and 12/004,173 filed Dec. 19, 2007.
The inventors have discovered that its robots, when deployed in hostile environments, are often fired upon. Therefore, it is insufficient for the robot to merely detect a gunshot or other sound. Instead, the robot must be capable of detecting a gunshot, targeting the origin of the gunshot, maneuvering, and maintaining the targeted origin as the robot moves. Requiring an operator controlling the robot to maintain the target origin while maneuvering the robot significantly increases the workload requirements of the operator.
It is therefore an object of this invention to provide a robot which can both pinpoint the origin of a sound, such as a gunshot, and also maneuver while targeting the origin.
It is a further object of this invention to provide such a robot which is less likely to suffer damage from unfriendly fire.
It is a further object of this invention to provide such a robot which can return fire.
It is a further object of this invention to provide such a robot which reduces the work load requirements faced by the robot operator.
The subject invention results from the realization that a new robot which pinpoints the origin of a sound, such as a gunshot or similar type sound, aims a device, such as a weapon, at the origin of the sound, and maneuvers and maintains the aim of the device at the origin while maneuvering is effected by a turret on the robot in combination with a turret drive, a set of sensors, and processing electronics which control the turret drive to orient the turret to aim a device, such as a weapon mounted to the turret, at the origin of the sound and to maintain the aim as the robot moves.
The subject invention, however, in other embodiments, need not achieve all these objectives and the claims hereof should not be limited to structures or methods capable of achieving these objectives.
This invention features a mobile, remotely controlled robot including a robot drive subsystem for maneuvering the robot, a turret on the robot, and a turret drive for moving the turret. A noise detection subsystem detects the probable origin of a noise. The robot includes a robot position and movement sensor subsystem, and a turret position sensor subsystem. One or more processors are responsive to the noise detection subsystem, the robot position and movement sensor subsystem, and the turret position sensor subsystem and are configured to control the turret drive to orient the turret to aim a device mounted thereto at the origin of the noise and to maintain said aim as the robot moves.
In one embodiment, the noise detection subsystem may include a gunshot detection subsystem configured to detect the origin of a gunshot and to provide the coordinates of the origin to the one or more processors. An initiation subsystem may activate a device may be mounted to the turret and the one or more processors may be configured to provide an output to the initiation subsystem to activate the device upon receiving a signal from the detection subsystem. The device mounted to the turret may include a source of illumination, a lamp, or a laser. The device mounted to the turret and may include a weapon. The system may include a weapon fire control subsystem for firing the weapon. The system may include an operator control unit for remotely controlling the robot. The one or more processors may include a central processing unit responsive to the noise detection subsystem, the robot position and movement sensor subsystem, and the turret position sensor subsystem configured to calculate the movement of the turret required to keep the device aimed at the origin of the noise, and a turret drive controller responsive to the central processing unit and configured to control the turret drive. A turret drive controller responsive to the robot position and movement sensor subsystem and may be configured to control the turret drive between updates provided by the one or more processors. The robot position and movement sensor subsystem may include a GPS receiver and motion sensors. The turret drive may include motors for rotating and elevating the turret. The turret position sensor subsystem may include encoders. The processing electronics may include one or more of a GPS receiver, a rate gyro, a fiber optic gyro, a 3-axis gyro, a single axis gyro, a motion controller, and an orientation sensor. The system may include a directional communication subsystem for providing communication between the operator control unit and the robot.
The subject invention also features a mobile, remotely controlled gunshot detection stabilized turret robot including a robot drive subsystem for maneuvering the robot, a turret on the robot, and a turret drive for moving the turret. A gunshot detection subsystem detects the origin of a gunshot and provides the coordinates thereof. The robot includes a robot position and movement sensor subsystem, and a turret position sensor subsystem. One or more processors are responsive to the noise detection subsystem, the robot position and movement sensor subsystem, and the turret position sensor subsystem and are configured to control the turret drive to orient the turret to aim a device mounted thereto at the origin of the noise and to maintain said aim as the robot moves.
Other objects, features and advantages will occur to those skilled in the art from the following description of a preferred embodiment and the accompanying drawings, in which:
Aside from the preferred embodiment or embodiments disclosed below, this invention is capable of other embodiments and of being practiced or being carried out in various ways. Thus, it is to be understood that the invention is not limited in its application to the details of construction and the arrangements of components set forth in the following description or illustrated in the drawings. If only one embodiment is described herein, the claims hereof are not to be limited to that embodiment. Moreover, the claims hereof are not to be read restrictively unless there is clear and convincing evidence manifesting a certain exclusion, restriction, or disclaimer.
In this way, robot 10 of this invention not only detects the origin of a gunshot or similar type sound and aims weapon 12 at the origin of the sound, robot 10 also maintains the aim at the origin of sound as robot 10 maneuvers. This allows a user, when maneuvering robot 10 from position C to D, for example, to fire weapon 12 to location of the origin of the sound. Because robot 10 continues to maneuver while weapon 12 is aimed at location of the origin of the sound, e.g., O-13, the likelihood that robot 10 will be damaged by fire from that location is reduced and robot 10 can then continue on its mission. Robot 10 can fire upon the location of the origin of the sound automatically or under the control of an operator. Further, robot 10 can communicate wirelessly with robot 11 at location E and provide robot 11 with data concerning location of the origin of the sound so robot 11 can aim its weapon 13 at that location.
One example of the primary subsystems associated with a robot 10 is shown in
Turret 20 is preferably rotatable and configured to elevate the device mounted thereto under the control of turret drive 22. Turret position sensor subsystem 40 detects, e.g., using encoders, inclinometers, and the like, discussed in detail below, and outputs the position of the turret and the device (e.g., angles θ and γ,
The position of the robot, e.g., robot 10 at positions A-D,
Processing electronics subsystem 46 preferably includes one or more processors, e.g., CPU 47, and/or CPU 49. Processing electronics subsystem 46 is responsive to the outputs of noise detection subsystem 42, robot position and movement sensor subsystem 44, and turret position sensor subsystem 40 and is configured to control turret drive 22 to orient turret 20 and aim a device mounted thereto at the origin of the gunshot or similar type noise and to maintain that aim as robot 10 maneuvers. Subsystem 46 can be configured, upon receipt of a signal from noise detection subsystem 42, to signal device initiation subsystem 30 to activate a device mounted to turret 20. In this way, a laser, for example, is automatically turned on and aimed at a target. Or, a weapon can be aimed and then automatically fired.
Preferably, processing electronics 46, turret drive 22, turret 20, and turret position sensor subsystem 40 are all integrated in a single modular unit.
The subject invention brings together several capabilities that have not previously been integrated into a single ground system for use in the real world. These capabilities include a proven unmanned ground vehicle or robot capable of operating in tactically significant environments, a tightly integrated 360° turret and elevation axis capable of carrying payloads up to 30 lb, a stabilized weapon/payload turret on the robot, the ability to maintain weapon/payload pointed at the point of origin of a gunshot or similar sound at all times, the ability to autonomously navigate using a sensor fused robotic vehicle state estimate based on GPS, robotic vehicle orientation, rates of motion, and odometry, and an overhead map vehicle location feedback and waypoint and target input. Robot 10 automates tasks that would otherwise completely consume the attention of the operator. Using robot 10, the operator can act more as a commander than a driver or gunner. The operator can command robot 10 to proceed along a path to a specified location while maintaining the weapon/payload pointed at the location of the origin of the gunshot or similar sound. This level of automation of the basic robot tasks of robot 10 allows a single user to operate multiple robots 10.
The turret 20 is preferably designed for interfacing with a small, highly mobile robotic vehicle, e.g., robot 10,
Robot 10 of this invention may stabilize the payload in one of several ways: gyro stabilization, stabilization about a heading and an elevation, or stabilization about a GPS coordinate. In gyro stabilization mode, turret 20,
Robot 10 is ideally suited for carrying small payload into rapidly changing and hostile environments. Turret 20 is preferably designed to be capable of >180°/s slew rates, allowing the payload pointing direction to be changed rapidly. Camera systems can be slewed to observe a threat, reducing the chance of robot 10 being taken by surprise. Small weapon systems can be slewed rapidly, keeping enemy forces or bystanders in urban combat scenarios away from robot 10.
As a reconnaissance platform, robot 10 of invention can be used in either leading or supporting roles. Robot 10 can be driven out in front of the combat unit by the operator. In the reconnaissance role, robot 10 may include high powered zoom cameras, FLIR cameras, or directional audio sensors. Robot 10 can be used to clear a room prior to entry by the squad. Robot 10 may be outfitted with flash-bangs or non-lethal weapons to allow it to engage an enemy in a less-than-lethal manner.
The commander of robot 10 may have a target designator in the form of either an encoded laser, a range finder, or laser pointer. The operator can drive robot 10 into a hostile area and using high powered zoom cameras and FLIR systems can designate targets for the human element of the squad to engage. Sniper detection is one example of such a mission. Robot 10 may be driven into an open or danger area and the operator uses the sensors mounted thereto to seek and detect enemy snipers. When a sniper is detected, an infrared laser pointer is used to mark the location of the sniper. The troops can use night vision goggles to detect the location of the laser dot and can engage the target location as they see fit.
In the automated response role, robot 10 may be either a sentinel with motion detection systems or robot 10 may use a threat recognition/localization subsystem to home in on the enemy autonomously. In the sentinel role, robot 10 may be parked outside a perimeter. When the motion detection system recognizes an incoming threat, the turret will swing a response payload toward the target and either engage or alert the operator.
A sniper detection system may be mounted on the turret. When a shot is detected and localized, the turret can swing a camera or a weapon in the direction of the sniper, and can either engage the area or alert the operator. If a shot is detected, the turret would swing a camera payload to observe the sniper location, providing an immediate image to the passenger in the vehicle of the sniper's location.
Robot 10 may also be designed to point the payload at a certain location in space. A long range radio link may be established between two robot 10 and robot 11 by putting YAGI style antennas on the turret and having those turrets remain pointed at each other. Each robot sends its location to the other, e.g., robot 10,
In one embodiment, directional communication subsystem 51, maintains a link automatically between robot 10 and robot 11, without human intervention. The chance of interception of the communications is drastically reduced due to the directionality of the link. Anyone outside the projection cone will not be able to eavesdrop on the link.
Multi-robot systems, e.g., such as those which employ robots of this invention, will likely play a critical role in tomorrow's battlefield. Squads of robots may be deployed to engage an enemy or perform reconnaissance. These robots must have exceptional self awareness and awareness of the whereabouts of the rest of the team. They must be able to engage targets designated by the commander vehicle (as described above) in a rapid and fluid way.
Directional communication subsystem 51,
In one design, turret 20,
Turret 20,
As robot 10 is turning, the aimpoint may change, requiring turret 20 and weapon or other device attached thereto to slew even faster than the robot slew. In one example, turret drive 22 provides about 200°/s to provide about 90°/s turret motion in the direction opposite the slew direction of robot 10. This maximum slew rate allows robot 10 to achieve any new aimpoint within 2 seconds regardless of vehicle motion. In one example, a 5° accuracy is a preferred accuracy with which turret 20 can maintain a payload pointed at a target location. The dynamic accuracy of 5° ensures that turret 20 can maintain a target within the middle third of the field of view of, e.g., a 30° FOV camera or within the beam-width of a YAGI directional antenna.
In one example, the pointing resolution is less than about <0.01° may be used to ensure that the aimpoint can be adjusted to within about 15.24 cm at 1000 m. A 360° continuous slew is preferably used for proper stabilization.
Processing electronics 46,
Processing electronics 46 ideally controls the motion of turret 20 via turret drive 22 and the motion of robot 10. Processing electronics 46 also preferably logs mission data, measures/estimates system localization information (e.g., GPS coordinate, vehicle orientation, vehicle dynamics), and the like, and also provides a payload interface that includes both power and communication. Processing electronics 46 may also provide processing power for targeting and/or fire solution calculation. In one design, the processing electronics 46 may integrate with a TALON® 36V power bus and use a TALON® communication component. Processing electronics 46 preferably utilizes PC-104 standard components, e.g., PC-104 stack 73,
The one or more processors, e.g., CPU 47, and/or CPU 49, forms the primary intelligence of robot 10, allowing robot 10 to run several software processes simultaneously, to handle inputs and output, and to perform high level control of system components.
In one example, turret position sensor subsystem 40,
CPU 47, the serial interface, and motion controller 258 preferably communicate over the PC-104 bus, e.g., bus 99,
Robot 10 preferably uses power and communication systems, e.g., as disclosed in U.S. patent application Ser. Nos. 11/543,427, cited supra. OCU 26 provides a well known and intuitive interface to the robot. Directional communication subsystem 51,
Self-awareness sensors 51,
In one example, turret 20 may include two RS-232 ports, four Digital I/O (for trigger actuator, firing circuit, arming circuit, and the like), two analog outputs, and 36V, 2 A current draw.
The equations of motion, in state-space notation, for the simulation shown in
Equation (1) allows for simulation of the behavior of robot 10 in virtual space. The model may be built in Matlab®/Simulink (www.mathworks.com) and responses to inputs are simulated.
As shown in
In one example, turret position and sensor subsystem 40,
In this example, stabilization loop 102 controls the velocity of turret 20. The rate feedback acts essentially as a low pass filter, damping out higher frequency vibrations, but not affecting the lower frequencies.
A PID position controller is preferably implemented to give the robot 10 a strong response at low frequencies. Such a controller maintains the pointing direction of the turret, and works in conjunction with the stabilization loop to maintain a steady aimpoint, e.g., at the point of origin of a sound, such as a gunshot.
Position feedback loop 120 of controller 100 significantly improves the response of subsystem 40,
In one embodiment, control system 100,
By proactively counteracting the effects of a disturbance on robot 10, the effects of the disturbance can be virtually eliminated. If subsystem 40,
The mechanical and electromechanical design of robot 10 preferably uses modeling of the mechanical and servo systems to specify motors and amplifiers that would satisfy the requirements of robot 10. Preferably the servo system is able to accelerate a 1250 lb-in2 payload to 180°/s in 0.2 seconds. Such a rate of change allows for sufficiently rapid motion to allow for stabilization of the payload.
In one example, the azimuth drive motor 78,
In one example, the motor amplifiers for motors 78, 79 may be Advance Motion Controls (AMC) ZB12A8 brushless motor amplifiers. These amplifiers have a maximum output of 12 A, and well suited for driving the Kollmorgen AKM22E motors utilized in the turret. Commutation is controlled by the amplifier using hall effect measurements from the motors. The amplifiers convert a +/−10V control signal from the motion controller and to a current proportional to this input signal.
Robot 10 typically includes a large number of sensors, e.g., as shown in
In one example, robot 10 may employ fiber optics gyro 254, e.g., a KVH DSP-3000 fiber optic gyro to improve low rate stabilization performance.
In one design, robot 10 may include orientation sensor 262, e.g., a 3DM-G orientation sensor to provide an absolute measurement of the orientation of robot 10 in space. Orientation sensor 262 typically consists of 3 gyros, 3 accelerometers, and 3 magnetometers. The outputs of the three sensor sets are fused onboard sensor 262 to provide an estimate of the true orientation of robot 10 in space. Orientation sensor 262 works through the entire 360° range of orientations and has an accuracy of 5°.
Motion controller 258,
The software architecture used for robot 10 is preferably a multi-process architecture running on Linux windows, or similar type platform. Each process running on the robot 10 is responsible for a logical task, e.g., turret control, radio communications, vehicle communication and control, localization process, navigation, payload control, and sensor drivers, and the like.
Turret component 204 typically handles all the control details for the turret 20 and turret drive 22 and also provides turret state information to any other system component. In practice this means that the turret component 204 handles all communications to a motion controller 258, e.g., DMC1220 motion controller, (or similar type motion control) that is to be used for controlling the servo motors 78, 79,
LogServer 202,
ProcessManager 200 preferably launches all the other system components shown in
In one example, 3DMG orientation process 210, Garmin 15 GPS process 212, and DSP3000 gyro process 214 gather information from orientation 3DMG sensor 262, Garmin 15 GPS receiver 250, and DSP300 gyro sensor 254, respectively.
KalmanFilter process 222,
Navigator component 226,
In order to minimize the burden on CPU 47,
Motion controller 258 typically receives a command from CPU 47 indicating which motion mode the system is in. The possible motion modes of operation may include: 1) fully manual: no automatic motion control is conducted and turret 20 simply follows the joystick commands from the operator, 2) gyro stabilized: turret 20 maintains the payload pointed along a vector in space, relying on the gyros to detect motion of the payload and counteracting these motions through appropriate motor commands, or 3) stabilized about a GPS target location: the payload is kept pointed at a location in space, designated as a GPS coordinate. As the robot 10 moves, the payload pointing direction is updated to maintain the aimpoint, as robot 10 moves, e.g., a discussed above with reference to
In fully manual mode, turret motors 78, 79,
In gyro stabilized mode, turret 78, 79 motors will counteract the motion of robot 10. The joystick commands passed to the motion controller indicate the rate at which the turret should move in absolute space. Therefore, if the joystick is neutral, the turret will attempt to remain pointed in a given direction even if the vehicle is moving. A joystick command will move the turret relative to the global coordinate system, regardless of the vehicle dynamics.
Targeting refers to the ability of system to “focus” the turret or a user defined target or on the origin of the noise or gunshot, e.g., O-13,
In one example, the targeting system 45,
Turret component 204 is constantly being updated by the localization process and the command router 206 as to the location and orientation of the robot 10 and the desired target point, respectively. Using these two pieces of information, robot 10 can calculate the desired position of the two turret axes, as described below.
When stabilized about a target location, orientation sensor 262,
The gains on the encoder count error are preferably set fairly low to ensure smooth operation. Over short time intervals, the gyros, e.g., gyros 252, 254, and/or 260,
The user can change the stabilization mode mid-mission as needed. For example, the user can switch to fully manual mode from GPS stabilized when the user needs to fine-aim the weapon or payload, and resume stabilization when firing or payload actuation is complete.
In one example, motion controller 258 may be a DMC1220 motion controller designed for a dedicated motion control DSP and is specifically designed to handle low level control functions. Functions such as position control or velocity control are very easily implemented.
Due to the simplicity of velocity and position control implementation on the motion controller 258, robot 10 leverages these functions to eliminate the need for CPU 47 to perform low level motion control. In one example, motion controller 258 can accept up to 8 analog inputs, sufficient for both rate feedback and vehicle rate feedforward. Motion controller 258 also interfaces with the servo motor encoders, reducing the amount of required hardware development.
In one example, velocity of the motors 78, 79,
The following are advantages of control system 350 lower computational burden on CPU 47, allowing CPU 47 to service other tasks in a more timely manner, simplified implementation since low level control methods are available onboard motion controller 258,
The feedforward stabilization and control system 350,
{dot over (θ)}elev=C{dot over (θ)}1,ƒw sin(ψ)+C{dot over (θ)}2,ƒw cos(ψ) (2)
where {dot over (θ)}elev is the commanded elevation rate, {dot over (θ)}1,ƒw and {dot over (θ)}2,ƒw are the roll and pitch rates of robot 10, respectively, and ψ is the azimuth location of turret 20 with respect to the forward direction. Therefore, if turret 20 is pointed forward, robot 10 will cause little or no movement of the elevation axis, while pitching motion will be entirely counteracted. If turret 20 is pointed to the side, the roll behavior will be counteracted, but not the pitch behavior. Roll and pitch will be both counteracted if turret 20 is off-axis (i.e. not exactly forward or exactly to the side). This feedforward stabilization algorithms works well for small angle deviations.
Preferably, CPU 47,
These variables are used by controller 258 software to specify the behavior of turret 20. As these variables are updated, turret 20 reacts appropriately. As more capability is added, additional data can sent to the motion controller in a similar manner.
Dual axis stabilization may be implemented. The feedforward loop shown in
When in stabilized mode, turret 20 and robot 10 act essentially independently. Turret 20 will slew at the desired rate in the global reference regardless of the vehicle slew rate and robot 10.
To avoid noise issues associated with feedback gyros and the drift in the horizontal feedforward gyros, the stabilization algorithm was reduced to simply azimuth feedforward. This provides the most useful stabilization performance since the drift is reduced significantly and the noise in the feedback gyros is eliminated from the control loop.
In one embodiment, fiber optic gyro 254,
One approach to calculating the turret pointing direction begins by determining the vector, P, from robot 10 to the target. Both the target and robot 10 location are preferably given in a NED coordinate system. The pointing vector is calculated as,
Once the P vector is known, it is transformed from the NED coordinate system to the vehicle coordinate system. Once the vector is known in vehicle coordinates, the turret angles required to achieve the P-designated pointing direction are found using simple trigonometry.
The localization Kalman filter provides vehicle pitch/roll/yaw information. Pitch, roll, and yaw are preferably transformed to a 3×3 orientation (or transformation) matrix by the turret component. The transformation matrix is used to transform a vector from one reference frame to another, without changing the vector location or orientation in space.
The transformation matrix output, M3DM-GNED,actual, is used to define the P vector in vehicle coordinates:
P′=M3DM-GNED,actualP (4)
Once the P′ vector is known (i.e. the vector pointing to the target defined in the vehicle reference frame), the vector must be mapped to turret coordinates. The commanded (desired) azimuth angle (AZDEZ) is calculated as:
And commanded (desired) elevation angle (ELDES) is calculated as,
The two values are passed to the turret motion controller 258,
Although specific features of the invention are shown in some drawings and not in others, this is for convenience only as each feature may be combined with any or all of the other features in accordance with the invention. The words “including”, “comprising”, “having”, and “with” as used herein are to be interpreted broadly and comprehensively and are not limited to any physical interconnection. Moreover, any embodiments disclosed in the subject application are not to be taken as the only possible embodiments. Other embodiments will occur to those skilled in the art and are within the following claims.
In addition, any amendment presented during the prosecution of the patent application for this patent is not a disclaimer of any claim element presented in the application as filed: those skilled in the art cannot reasonably be expected to draft a claim that would literally encompass all possible equivalents, many equivalents will be unforeseeable at the time of the amendment and are beyond a fair interpretation of what is to be surrendered (if anything), the rationale underlying the amendment may bear no more than a tangential relation to many equivalents, and/or there are many other reasons the applicant can not be expected to describe certain insubstantial substitutes for any claim element amended.
This application hereby claims the benefit of and priority to U.S. Provisional Application No. 61/123,299, filed Apr. 7, 2008, under 35 U.S.C. §§119, 120, 363, 365, and 37 C.F.R. §1.55 and §1.78, incorporated by reference herein.
This invention was made with U.S. Government support under Contract No. W15QKN-04-C-1013 awarded by the U.S. Army. The Government may have certain rights in the invention.
Number | Name | Date | Kind |
---|---|---|---|
4316218 | Gay | Feb 1982 | A |
4386848 | Clendenin et al. | Jun 1983 | A |
4514621 | Knight et al. | Apr 1985 | A |
5123327 | Alston et al. | Jun 1992 | A |
5241518 | McNelis et al. | Aug 1993 | A |
5586086 | Permuy et al. | Dec 1996 | A |
5917775 | Salisbury | Jun 1999 | A |
6467388 | Malakatas | Oct 2002 | B1 |
6535793 | Allard | Mar 2003 | B2 |
6701821 | Lundqvist et al. | Mar 2004 | B2 |
6847587 | Patterson et al. | Jan 2005 | B2 |
6999881 | Permuy et al. | Feb 2006 | B2 |
7121142 | Chevret et al. | Oct 2006 | B2 |
7139222 | Baxter et al. | Nov 2006 | B1 |
7210392 | Greene et al. | May 2007 | B2 |
7584045 | Peretz | Sep 2009 | B2 |
7600462 | Quinn | Oct 2009 | B2 |
7650826 | Son et al. | Jan 2010 | B2 |
7654348 | Ohm et al. | Feb 2010 | B2 |
7962243 | DeGuire et al. | Jun 2011 | B2 |
7974738 | Bruemmer et al. | Jul 2011 | B2 |
20040068415 | Solomon | Apr 2004 | A1 |
20060149541 | Jaklitsch et al. | Jul 2006 | A1 |
20060271263 | Self et al. | Nov 2006 | A1 |
20070057842 | Coleman et al. | Mar 2007 | A1 |
20080063400 | Hudson et al. | Mar 2008 | A1 |
20080071480 | Lin | Mar 2008 | A1 |
20080083344 | Deguire et al. | Apr 2008 | A1 |
20080121097 | Rudakevych et al. | May 2008 | A1 |
20090164045 | Deguire et al. | Jun 2009 | A1 |
20100212482 | Morin et al. | Aug 2010 | A1 |
20100263524 | Morin et al. | Oct 2010 | A1 |
20110005847 | Andrus et al. | Jan 2011 | A1 |
Entry |
---|
Thoren—RedOwl.pdf (Dr. Glenn Thoren, From Hollywood to Homes (and to Defense and Security): Robots for Today the RedOwl Project, Boston University, Mar. 31, 2006, 24 pages). |
www.defensereview.com—anti-snipersniper-detectiongunfire-detection-systems-at-a-glance.pdf (David Crane Anti-Sniper/Sniper Detection/Gunfire Detection Systems at a glance, Defense Review, published Jul. 19, 2006, pp. 1-7, cache on Google on Oct. 17, 2011). |
Number | Date | Country | |
---|---|---|---|
20090281660 A1 | Nov 2009 | US |
Number | Date | Country | |
---|---|---|---|
61123299 | Apr 2008 | US |