This application is related to US Patent Applications entitled “Static Wireless Data Glove For Gesture Processing/Recognition and Information Coding/Input”, Ser. No. 12/323,986, filed Nov. 26, 2008, and “Wireless Haptic Glove for Language and Information Transference”, Ser. No. 12/325,046, filed Nov. 28, 2008, both of which are assigned to the same assignee as the present application, the contents of both of which are fully incorporated by reference herein.
Aircraft marshaling is visual signaling between ground personnel and aircraft pilots on an aircraft carrier, airport or helipad. Marshaling is a one-on-one visual communication technique between an aircraft marshal and the pilot, and may be an alternative to, or additional to, radio communications between the aircraft and air traffic control. The usual attire of the aircraft marshal is a reflecting safety vest, a helmet with acoustic earmuffs, and illuminated beacons or gloves. The beacons are known as marshaling wands to provide pilots with visual gestures indicating specific instructions.
For instance, an aircraft marshal, using well known arm gesture motions, signals the pilot to keep turning, slow down, stop, and the like, leading the aircraft to its parking location, or to the runway at an airport, or to a launch position on an aircraft carrier.
The marshaling wands currently in use frequently have different colored lights to signal a pilot with marshaling instructions, such as using a yellow light with appropriate arm motions for general instructions such as turn, slow down, and the like, and then switching to a red light with appropriate arm motions to signal the pilot to stop the aircraft. Other color configurations can be used as well, such as blue, green, and amber. However, such marshaling wands do typically not provide radio communications between the aircraft marshal and the pilot. There are limitations to such marshaling wands, particularly when used on an aircraft carrier, where the very limited space and time between take-offs and landing makes radio communications between the aircraft marshal and the pilot a difficult alternative.
In one preferred embodiment, an aircraft marshaling wand controller displays aircraft marshaling instructions to a pilot on a video display monitor on-board an aircraft, such as an aircraft on an aircraft carrier. When an aircraft marshal uses arm motion gestures to form aircraft marshaling instructions for the pilot on the aircraft, the wand controller of the present invention senses or detects those gesture motions, and generates digitized command signals representative of those gesture motions made by the aircraft marshal. A wireless transceiver then transmits those digitized command signals to the aircraft for display on the video monitor for viewing by the pilot.
Throughout the several views, like elements are referenced using like reference numerals, wherein:
One purpose of the present invention is to provide an input device and method for recognition of hand waves and gestures. In one embodiment, the device or apparatus can input data to personal digital assistants or computers. Also, one embodiment of the present invention provides network enabled devices to monitor gestures or motions of aircraft carrier marshaling signals, as used by landing signal officers.
One objective of the wand controller 10 shown in
Typical aircraft marshalling signals are shown on the left hand portion of
“PROCEED TO NEXT MARSHALER”, “STOP”, or ‘SLOW DOWN” signals to a pilot on an airplane, such as on a Navy aircraft carrier. There are many other gesture signals well know to the aircraft community, whether on an aircraft carrier, or a land tarmac at an airport.
The present invention provides, among other features, the capability to visually display the marshaling signals such as shown in
As shown in
The “PROCEED TO NEXT MARSHALER”, “STOP” and “SLOW DOWN” signals shown in
In
The discrete data is then converted into vector quantities to determine the spatial points. All of these data are processed by the microcontroller 20 through mathematical algorithms. The microcontroller 20 processes the vector quantities by calculating and translating to proper commands/words or letters.
In one embodiment, the processor or microcontroller 20 can compare the processed vector quantities with stored predetermined gesture information data in memory 22 which is representative of various command instructions, such as the “STOP”, “SLOW DOWN”, and “PROCEED TO NEXT MARSHALER” instructions shown in
The result is transmitted (sent) via transceiver 42 of
In
Referring again to
Each of the gyroscope sensors 34 are 3-axis or three-dimensional (XYZ) sensors to measure the angular rate of a gesture motion over a period of time. These angular gesture motions can be computed and yield a rotation angle, representative of the gesture motion rotation such as would occur in
Each of the accelerometer sensors 36 shown in
Each sensor of the 3 axis (3D) magnetometer sensor 30 allows the present invention to capture the motion of the wand controller shown in
u=g×h
v=w×u
w=−g
where g is unit vector of G, h is unit vector H, u is unit vector parallel with the sensor x-axis, v is unit vector parallel with the sensor x-axis, and w is unit vector parallel with the sensor x-axis.
Computer calculation: Each of sensor values is read into the processor is processed as followed:
Magnetic Hx=Read in Magnetic Hx−Midpoint Hx
Magnetic Hy=Read in Magnetic Hy−Midpoint Hy
Magnetic Hz=Read in Magnetic Hz−Midpoint Hy
Acceleration Ax=Read in Acceleration Ax−Midpoint Ax
Acceleration Ay=Read in Acceleration Ay−Midpoint Ay
Acceleration Az=Read in Acceleration Az−Midpoint Ay
where Midpoint Hx, Midpoint Hy, Midpoint Hz, Midpoint Ax Midpoint Ay and Midpoint Az are the calibration data at static state.
Scaling these scalars to be Magnetic Hx, Magnetic Hy, Magnetic Hz, Acceleration Ax Acceleration Ay and Acceleration Az.
Normalizing all the above vectors to be the same size or magnitude of unit one vector
In
ex=[1,0,0]
ey=[0,1,0]
ez=[0,0,1],
where ex refers to a bearing of North, ey refers to a bearing of East, and ez refers to an orientation of “up.”
N=[Nx,Ny,Nz)=[u·ez,v·ez,w·ez],
which provides the scalar components for the sensor's “upward” orientation.
Next in
P=[Px,Py,Pz]=[u·ey,v·ey,w·ey],
which provides the scalar components for the sensor's “eastward” orientation.
Next, in
Q=[Qx,Qy,Qz]=[u·ex,vex,w·ex],
which provides the scalar components for the sensor's “northward” orientation.
In using the N, P and Q vectors, we can calculate the absolute orientation angle of the sensor with respect to the earth global coordinate system. Accordingly as shown in
Pitch=sin−1(Pz)
Roll=sin−1(Qz)
Heading=tan−1(Py/Px).
With all combination of vectors derived from the above sensors are obtained and processed by the microcontroller 20 yields a relational motion of devices over a period of time. With the mathematical calculation within the microcontroller 20, the wand controller 10 determines the orientation of the device and predicts possible gestures as sequences of digitized points in space, in terms of command and alphanumerical characters.
Also, the vector relationship between sensors on each wand controller shown in
In other embodiments, the wand controller of the present invention can include additional features.
For instance, a speaker controlled by audio indicator controller 46, to produce an audible sound representative of what a completed gesture sequence meant. For instance, an audible command could be received from another wand controller according to the present invention. In another instance, the “STOP” signal could be audibly sent to a pilot in an aircraft as a still additional safety measure.
A vibration motor, such as haptic feedback 48, which is controlled by ON-OFF pulse generated by microcontroller 20 to indicate the gesture sequence.
A touch keypad, such as keypad area 40 shown in
As seen in
As the wand controller is moving in the 3-D or the air, sensors are acquiring data representative of the gesture motions. The sensed analog data is combined and processed to detect (generate) alpha-numerical characters, A . . . Z, and including 0, 1 . . . 9. The motion detection mechanism of the wand controller is also decoding proper gestures into meaningful commands. The generated data can then be sent to over the wireless network to a personal digital assistant (PDA), or including a computer, where it may be further processed or displayed.
The hardware unit is designed or integrated into many shapes and sizes to serve various applications, and can be designed to be compatible to personal digital devices (PDD), laptop or desktop computers.
A pair of wand controllers can be used for directing (marshalling) an airplane while on an aircraft carrier or land tarmac. These wand controller pairs are designed to send gesture signals directly to an airplane pilot via wireless link onto a cockpit display (monitor) to enable the pilot to visually see and couple both wand marshalling signaler and cockpit information for the extra safety measure of airplane maneuver over the aircraft carrier or tarmac.
In
In
As shown in
Another embodiment of the invention is to embed the wand controller onto a surgical scalpel. The scalpel-wand controller would be used in training medical students or aid the surgeon in their precision with incisions during surgery. Information on incision depths and locations on the body can all be wirelessly transmitted back to the surgeon as a feedback system.
In
The sensed gesture motion information would correspond to the three dimensional sensor information detected by gyroscope 34, accelerometer 36 and magnetic sensor 30, as has been previously described in conjunction with the block diagram of a wand controller 10 shown in
In
From the above description, it is apparent that various techniques may be used for implementing the concepts of the present invention without departing from its scope. The described embodiments are to be considered in all respects as illustrative and not restrictive. It should also be understood that system is not limited to the particular embodiments described herein, but is capable of many embodiments without departing from the scope of the claims.
This invention (Navy Case No. 100,271) is assigned to the United States Government and is available for licensing for commercial purposes. Licensing and technical inquiries may be directed to the Office of Research and Technical Applications, Space and Naval Warfare Systems Center, Pacific, Code 72120, San Diego, Calif., 92152; voice (619) 553-2778; email T2@spawar.navy.mil.
Number | Name | Date | Kind |
---|---|---|---|
5036442 | Brown | Jul 1991 | A |
5392203 | Harris, Jr. | Feb 1995 | A |
5622423 | Lee | Apr 1997 | A |
5642931 | Gappelberg | Jul 1997 | A |
5714698 | Tukloka | Feb 1998 | A |
6293684 | Riblett | Sep 2001 | B1 |
6294985 | Simon | Sep 2001 | B1 |
6494882 | Lebouitz et al. | Dec 2002 | B1 |
6561119 | Rigitano | May 2003 | B1 |
6577299 | Schiller | Jun 2003 | B1 |
6747599 | McEwan | Jun 2004 | B2 |
6903730 | Mathews | Jun 2005 | B2 |
7050606 | Paul | May 2006 | B2 |
7257255 | Pittel | Aug 2007 | B2 |
7267453 | Chang | Sep 2007 | B2 |
7279646 | Xu | Oct 2007 | B2 |
7287874 | Irisawa | Oct 2007 | B2 |
7289645 | Yamamoto | Oct 2007 | B2 |
7397469 | Vablais | Jul 2008 | B2 |
7460011 | Liau et al. | Dec 2008 | B1 |
7500917 | Barney et al. | Mar 2009 | B2 |
7606411 | Venetsky | Oct 2009 | B2 |
7737867 | Arthur et al. | Jun 2010 | B2 |
8058975 | Barnardo et al. | Nov 2011 | B2 |
8240599 | Edelson et al. | Aug 2012 | B2 |
20040118945 | Russell | Jun 2004 | A1 |
20040143512 | Sturr | Jul 2004 | A1 |
20040179352 | Anderson et al. | Sep 2004 | A1 |
20060279549 | Zhang | Dec 2006 | A1 |
20070176898 | Suh | Aug 2007 | A1 |
20070268278 | Paratore | Nov 2007 | A1 |
20090265671 | Sachs | Oct 2009 | A1 |
20100013944 | Venetsky | Jan 2010 | A1 |