Headset computer that uses motion and voice commands to control information display and remote devices

Information

  • Patent Grant
  • 11947387
  • Patent Number
    11,947,387
  • Date Filed
    Tuesday, November 2, 2021
    2 years ago
  • Date Issued
    Tuesday, April 2, 2024
    a month ago
Abstract
A wireless hands-free portable headset computer with a micro display arranged near but below a wearer's eye in a peripheral vision area not blocking the wearer's main line of sight. The headset computer can display an image or portions of an image, wherein the portions can be enlarged. The headset computer also can be equipped with peripheral devices, such as light sources and cameras that can emit and detect, respectively, visible light and invisible radiation, such as infrared radiation and ultraviolet radiation. The peripheral devices are controllable by the wearer by voice command or by gesture. The headset computer also can be broken down into component parts that are attachable to another article worn by an individual, such as a helmet or respirator mask.
Description
BACKGROUND

The present application relates to human/computer interfaces and more particularly to a wearable headset computer that accepts voice commands, tracks hand gestures and/or detects head movements to provide inputs to control software running within the headset computer and/or peripheral devices.


Small, portable electronic devices capable of storing and displaying large amounts of high resolution computer graphic information and video content continue to be increasingly popular. Devices such as the Apple iPhone™, Google Android™ and other smartphones represent a significant trend in convergence among mobile telephones, portable computers and digital media players (iPhone is a trademark of Apple Computer, Inc. and Android is a trademark of Google, Inc.). Although these smartphones typically include a display screen, the visual experience of a high-resolution, large format display cannot easily be replicated of because physical size limitations in the handheld form factor.


Other devices which provide improved functionality over smart phones are known by various names such as headset computers, video eyewear, head mounted displays with embedded computer processors, and the like. These devices include a frame or other support mechanism that is worn about the face and/or head, similar to a pair of eyeglasses and/or headphones. The frame houses a small, high-resolution microdisplay, optical lenses and other components needed to present an electronic image to the wearer. Circuitry in the headset computer can include display drivers, wireless interface(s), and fully functional personal computer systems. For further information regarding such headset computers, refer to corresponding patent application entitled “Handheld Wireless Display Devices Having High-Resolution Display Suitable for Use as a Mobile Internet Device”, PCT International Application Number PCT/US09/38601 filed Mar. 27, 2009, the entire contents of which is hereby incorporated by reference.


Such devices are also further described in U.S. Application No. 61/300,611, filed on Feb. 2, 2010, titled “Head Mounted Video Eyewear With Accessory Mount;” Ser. No. 12/774,179, filed on May 5, 2010, titled “Remote Control Of Host Application Using Motion And Voice Commands;” 61/176,662, filed on May 8, 2009, titled “Remote Control Of Host Application Using Tracking And Voice Commands;” 61/237,884, filed on Aug. 28, 2009, titled “Remote Control Of Host Application Using Motion And Voice Commands;” Ser. No. 12/008,114, filed on Jan. 8, 2008, titled “Monocular Display Device;” and Ser. No. 12/008,104, filed on Jan. 8, 2008, titled “Monocular Display Device;” the contents of each of which are incorporated by reference in their entirety.


SUMMARY

A headset computer includes a microdisplay, multiple input devices such as a head tracking accelerometer and/or camera to detect movements such as head movements, hand motions and/or gestures, and audio processing circuits to detect voice commands. These inputs provide control over and operation of an application program running within the headset computer itself and/or peripherals associated with the headset computer.


In one implementation, voice, head motion, and/or hand gesture inputs are received from sensors located within the headset computer. A field of view into a 3-D virtual space is then determined from the voice, head motion and/or hand gesture inputs. Data representing the 3-D virtual space may be maintained either by a processor local to the headset computer device and/or a remote processor. The 3-D virtual space contains data representing one or more graphical objects. The graphical objects may include various elements such as computer desktops, application windows, digital images, photographs, 3-D models or other image data. The voice, head motion, and/or hand gesture inputs may determine both the field of view and scale factor to determine a viewpoint into the 3-D virtual space. Image data selected from the 3-D virtual space determined based on the field of view and scale factor, and then presented on the microdisplay.


The hand gesture, head motion and/or voice commands can be used not only to set the field of view and scale factor but also to select which of several graphic objects within the 3-D virtual space are selected for presentation on the microdisplay, in whole or in part. Thus by using these input commands the wearer of the headset may navigate through a large format 3-D space and completely control which portions of the 3-D space are seen on the microdisplay.


The head set computer may also include a light source and camera. The spectral properties of the light source and camera can be manipulated by the input commands to provide a synthetic vision function. In particular, the scene illuminated by the light source is detected by the camera, and in turn presented on the microdisplay. The light source and camera preferably operate in invisible electromagnetic portion of the spectrum including infrared, near infrared, ultraviolet, shortwave infrared or other invisible wavelengths. In this manner, the wearer of the headset computer has the ability to view scenes in an invisible portion of the spectrum.


The voice, head motion and/or hand gestures can control an operating wavelength for the light source and the camera, intensity of emissions from the light source, sensitivity of the camera, or other aspects of presentation of the synthetic vision function on the micro display.


The light source also can be used to determine range information. In particular, the light source may emit a high precision light, such as a laser light. A reflection of the light by an object can then be detected by a camera or other sensor capable of determining a round-trip time delay for the light. The headset computer can then determine a range to one or more points in physical space. This can provide not only a range to a given object but also, for example, a distance between two objects. In a case where the distances to more than two points are determined, a volume of space can be estimated.


The headset computer can be packaged to fit headgear such as a helmet. In one implementation, the processors and other electronic components can be disposed in a first housing, a second housing may carry the microdisplay on a boom and a third housing may include a power supply. One or more signal and/or power connectors are then provided between the various housings. The housings are separately attachable and detachable from the headgear. This permits retrofitting of a headset computer to helmets or other headgear that safety, security, and military personnel are accustomed to wearing. This then eliminates the need for them to fit completely assembled headsets inside of or on top of their headgear. In certain embodiments the housing for the micro display and/or processor may include noise cancellation circuits that may assist with operations in a noisy environment such as with a rebreather apparatus.


The headset computer may also control a remote vehicle and receive and display images returned from a camera on the remote vehicle. Control commands derived from voice, head motion, and hand gesture inputs can be used as a remote control for an associated vehicle. In other embodiments, the control command can alter position and/or other operating characteristics of a camera located on the vehicle. The image information returned from the camera on the remote vehicle may be displayed on the micro display.


A wearer of the headset computer can therefore experience using the remote vehicle as if he were a miniature pilot traveling on the vehicle, operating the vehicle while having a view of the scene around the vehicle displayed on the micro display. A wireless interface between the headset computer and the vehicle can be used to detect control commands for the vehicle.


The control commands may also control apparatus that are mounted on the vehicle such as payloads and other devices to be delivered by vehicle. In certain aspects, the voice command can generate complex commands for the vehicle such as “return to base”, “circle at a specific altitude”, and so forth.





BRIEF DESCRIPTION OF THE DRAWINGS

The foregoing will be apparent from the following more particular description of example embodiments of the disclosure, as illustrated in the accompanying drawings in which like reference characters refer to the same parts throughout the different views. The drawings are not necessarily to scale, emphasis instead being placed upon illustrating various embodiments.



FIG. 1 shows a headset computer that implements several functions described herein.



FIG. 2 depicts a viewport into a 3-D virtual space that may be manipulated with voice, head motions, or hand gestures.



FIGS. 3A and 3B depict a synthetic enhanced vision mode and range finding mode.



FIGS. 4A and 4B illustrate the synthetic vision mode in more detail, with FIG. 4A being a black screen seen with the light source off, and FIG. 4B a scene that is in view when the light source is enabled.



FIGS. 5A and 5B illustrate an implementation where components of the headset computer are individually attached to a helmet.



FIG. 6 illustrates a wireless joystick and mouse controller that can be used with the headset computer to control another device such as a vehicle.



FIGS. 7A, 7B and 7C illustrate more examples of vehicles that can be controlled by the headset computer operating components contained thereon.



FIG. 8 shows a wireless joystick used with the headset computer.



FIG. 9 is a more detailed view of the headset computer.



FIG. 10 is a high-level block diagram of the functions performed by the headset computer.



FIG. 11 is a more detailed block diagram of the components of the headset computer.





DETAILED DESCRIPTION


FIG. 1 illustrates a headset computer 100 that is worn on the head of a person. The headset computer 100 is also known as video eyewear, head mounted display (HMD) with embedded computer, and by various other names. In this example, the headset computer 100 consists of apparatus, processor(s), and software that control the presentation of objects on microdisplay 120 and peripheral devices. The microdisplay 120 may be enclosed in a display housing 125 supported by a boom 132 that attaches to a frame 140. The frame 140 may additionally contain a housing 150 to enclose further electronic components; in the example shown here the housing 150 is positioned on the back of the wearer's head. One or more speakers 135 deliver audio signals to the wearer's ears and similarly one or more microphones 137 either contained in the housing 125 and/or in other areas detect audio signals. A peripheral such as a camera 160 may be attached to a peripheral port 165. The headset computer 100 may be further supported, for example, by a strap 170 on the wearer's head.


As will be explained in detail below in connection with FIGS. 9, 10 and 11, the headset computer 100 is a completely contained personal computer system including one or more data processor(s) for generating images on the microdisplay 120 and performing other functions. The processor(s) located in the headset computer 100 are capable of interpreting voice commands, detecting hand movements of the wearer (such as through the camera) 160 and/or detecting the wearer's head movements through accelerometers or other motion sensors. These inputs are then interpreted as commands to either the headset computer 100 and/or remote devices with which the headset computer can communicate, such as over wireless interfaces.


Viewport into 3-D Virtual Space


One function performed by the headset computer 100 is to provide a graphical viewport and/or window into a 3-D virtual space. The graphical viewpoint determines which information is presented on the microdisplay 120. In this mode, for example, a movement by the wearer's head can bring a different section of that 3-D virtual space into view on the microdisplay 120.



FIG. 2 illustrates this in some detail. Here the wearer 200 is shown relative to the virtual 3-D space 300. The virtual 3-D space 300 has a number of windows 310, 320, 330, 340, 350 located within it and maintained by the processor(s) within the headset computer 100 or the external host. However, only a single one of the windows 340 is shown without shading, to indicate that it is an active window. The other windows 310, 320, 330 and 350 are shown grayed out. Thus while these other windows exist in the 3-D virtual space, the wearer 200 does not see the entire 3-D virtual space—rather the user 200 sees only the window 340 (or more typically, a selected portion thereof) on the microdisplay 120. The window 340 may be simply a viewport into for example, a graphic window in a windowed operating system. In the example shown here, the graphical object is a digital map image and the wearer is viewing only a portion of the map and not the whole map.


It will be understood that the 3-D virtual space may include various elements such as computer desktops, application windows, photographs, 3-D object models or any other type of digital image objects. It should be further understood that these image objects can be positioned next to, overlaid on or behind or beside one another in the 3-D virtual space.


The user can manipulate the various image objects by giving commands using the headset computer 100. In one example, the user can ask for a level of enlargement of a particular area of interest within one of the objects. The location and size of the window area may be selected by the tracking of head motions, voice commands and/or hand gestures. For example, the user may specify a position and magnification and/or zoom level to be applied to a particular application software window. The result is similar to using a magnifying glass to look at something seamlessly over a large area, but by using the head tracker/gesture detector/voice input detection to zoom into an area being seen on the microdisplay 120 and at what level of magnification. Thus, using this feature the user can move his head left, right, up or down and then select a particular one of the image objects 300, 310, 320 through 340 to be active. In one example, the user 200 might from the position shown in FIG. 2, turn his head to the right. This motion would then cause a new window 330 to then become the active window, with window 340 then becoming deactivated.


The user 200 can also issue commands to retain a piece of a large image that he wishes to magnify, freezing that portion on the screen and setting it aside and then going back and looking at another area of that image or even requesting another level of magnification for that other area. In this way, the user can view the same portions of an image at different levels of magnification and/or view different bits or pieces of a larger image at different levels of magnification and then switch between them by merely moving his head left or right, up or down


In yet another example, the wearer may issue voice commands to manipulate the position of the various image objects in the 3-D virtual space. For example, he may select an image object such as by moving his head, but then issue a voice command such as to “move object up” or “move object A behind object B”. This causes the head tracker to then control the relative position of the selected image object(s) within the 3-D virtual space, rather than allowing him to navigate among a given single object within the 3-D space.


It will be understood that the wearer 200 thus has access to a virtual desktop that is in any form factor that can be represented in a 3-D virtual space, i.e. he may be working in a 360° surface that wraps around his head or may be given the impression that he is working in a 3-D space with a long depth of field.


In another example, the user 200 may turn his head to the lower left causing the window 350 to become active. This window may be a 3-D model of an object such as an engine. The user may then proceed to manipulate this 3-D model using voice, head tracking and/or hand gesture commands to manipulate the viewpoint in 3-D space. The wearer may also issue a command to manipulate the model itself, such as to say, “rotate object 90° horizontal” causing the representation of the motor to rotate in 3-D space.


The view of the displayed image on the microdisplay 120 does not require the user to be physically oriented as if he were looking in any particular direction. For example, the user may remotely view any image being virtually generated in a sitting or standing position as might be projected on a wall in a room, but yet that wearer may be himself physically oriented in other positions such as laying down.


Hands-Free Synthetic Vision



FIGS. 3A and 3B illustrate other modes and functions provided by the headset computer 100. In this example, an emitter such as a light source 195 is located within the headset computer 100 typically within the same small housing as the camera 160. If the wavelength of the camera 1600 and light source 195 are coordinated such that the camera is sensitive at the same wavelength emitted by the light source(s) then a synthetic vision function can result. For example, the light source may be infrared or ultraviolet, and if the camera is similarly sensitive in these wavelength regions, the resulting image from the camera can be seen on to the microdisplay 120. The processor in headset computer 100 can also cause not just the image to be captured by the camera and directly viewed on the microdisplay 120, but to also be stored in memory and/or be sent to a remote system or display.


Using the headset computer 100, the wearer can thus experience hands-free synthetic vision that combines a synthetic view that is, for example, a far infrared view showing heat signatures of individuals or objects on the other side of a wall or other obstruction. An example of the same is shown in FIGS. 4A and 4B. With the light source 195 off, the image on the microdisplay is completely blank. However, with the light source 195 on, the infrared camera picks up the emission of the infrared and see an image that would not otherwise be visible. Using the headset computer 100, with an integrated infrared light source and camera, the wearer is then able to see what is in the dark environment on the microdisplay 120, but himself remain unseen by those relying only on natural non-enhanced sight. The individual 350 in the scene would not be able to detect the wearer's presence, because the infrared illumination from the light source would not be visible to the unaided eye.


As shown in FIG. 3B, the light source may also include a laser range finder 198. The laser 198 can be aimed by the user either by moving his head and/or by using voice commands. The range finder can be used for various functions such as finding the distance to an object in physical space, determining relative distance between two things. In the latter example, the user may aim his head at one object and measure a distance to that first object, and then move his head to otherwise aim the laser at second object. The user can then ask the headset computer 100 to solve the triangulation equation between his position and the two objects, thereby estimating a distance between the two objects.


In a further example, a volume of space can be estimated by the wearer aiming the laser at three or more points and asking the headset computer to figure out the distances between them. These functions can be useful in uses such as surveying or material estimating necessary. This can now be accomplished without the wearer actually moving about or by using measuring implements other than the laser range finder as built into the headset computer 100.


Components Retrofittable to Helmet



FIGS. 5A and 5B illustrate another example of the headset computer 100 packaged in a particular way to be retrofit onto existing headgear. In this implementation, headset computer electronics (including peripheral devices such as a camera and battery power source), the main processor and so forth may be packaged into one component housing 502, and the boom with an integrated optical/visual/audio part packaged as another component housing 501. The components 501, 502 are individually attachable to an existing headgear familiar to the user. In this example shown, a helmet 500 may have two Velcro™ pads, with the electronics component 502 attached to one pad Velcro™, and a battery 503 separately packaged and attached to the other Velcro™ pad (Velcro is a trademark of the Velcro Corporation). The boom element 501 is also attached to the helmet, such as via a mechanical clip or fastener; the boom of course including the integrated microdisplay and microphones. A cabling system can connect the microdisplay and boom 501 to the electronics 502 and battery 503. The boom 501 may be fixed and/or bolted to the left side or right side of the helmet, depending on user preference and also depending upon other equipment that the user may be operating.


Camera(s), laser(s), and other peripherals can also be mounted to the helmet 500. Instead of requiring the wearer to wear a dedicated headset under the helmet, this packaging approach can implement a headset computer functionality without the user having to become comfortable with new headgear. In addition, operation with certain types of headgear (such as a rebreather) is not affected. This particular end use may be improved if the on board electronics also provide for noise cancellation. For example, if the wearer is using a rebreather, the rebreather tends to make a lot of background noise that would otherwise interfere with voice inputs or sound recording. The on-board electronics may include noise cancellation circuits or programming that eliminate the background noise of the rebreather. A similar approach can be used to cancel out other background noises to allow for clearer recording of voices or other sounds.


Headset Computer Controls Remote Vehicle, Receives and Displays Images from and to the Remote Vehicle


In yet another implementation, the voice, head motion and/or hand gesture inputs received from the sensors located within the headset computer 100 can be used to derive a remote control command. That control command can then be sent over a wireless interface to control a remote vehicle robot, or other object. In this end use, the input device may also further include a wireless joystick and/or mouse to provide further inputs to control the vehicle.


In one example, a voice input to the headset computer can generate a control command to control the path of the vehicle. Voice commands, such as “turn right”, “turn left”, “move forward”, “move backward”, “stop” and so forth can be included in the processing capabilities of the headset computer 100. Similarly, head tracking inputs can generate a control command to control the path of the vehicle, or more commonly the direction of the camera on the vehicle. In this way, the user can obtain an experience that he is physically located on the vehicle. This is accomplished by having the camera on the vehicle transmitting video preferably wirelessly back to the headset computer. The video received at the remote vehicle can then be displayed on the display within the headset computer.


In yet another example, a wireless handheld controller 610 such as that shown in FIG. 6 can be used with the headset computer 100 to control the path position, attitude and/or direction of the vehicle 620 more naturally.


Using this arrangement, a person can control a vehicle such as an unmanned aerial vehicle (FIG. 7A), unmanned ground vehicle (FIG. 7B) or a toy (FIG. 7C) and so forth. This eliminates problems with prior art that simply operates a remote vehicle 620 with a videogame type of controller that requires almost total user attention and both hands to operate. By utilizing wireless joystick controller 610 in combination with the wearable headset computer 100 that can obtain head motion, voice and hand tracking commands, the control and electronic processing capabilities of the headset computer can give anyone control such as a soldier, policeman, fire or industrial worker control over one or more remote systems or vehicles 620 in simple and natural way.


In the absence of a separate user input device, the camera on the headset computer 100 may detect the user's hand gestures as control inputs. The wearer can also give speech commands to give the vehicle certain commands. For example, if the wearer says “freeze”, that can be detected by the headset computer which then translates the spoken command into one or more commands to control the flight path of the unmanned aerial vehicle, to stop doing everything else and simply hover or follow a circular flight path around a current point of interest.


In other examples a voice command such as “return to base” can cause the vehicle to follow a complex programmed flight path. Another example can be “circle at a specific altitude” which can cause the vehicle to generally follow a geo-stable circle around its present location. This can alleviate the user from tediously having to continuously provide commands via the handheld controller.


Other voice commands and hand held commands can be used to control other aspects of the vehicle's capabilities, performance and/or path of travel.


In one embodiment, the vehicle 620 may itself contain a camera that transmits its video output wirelessly back to the headset computer 100. Video carried back to the headset computer 100 is then displayed on the microdisplay 120. The wearer's head movements and/or gestures may then be used in a natural way to control the position, attitude, pan, zoom, magnification, light spectral sensitivities or other capabilities of the camera on the remote vehicle. The user's head movements can then be tracked by the on board electronics of the headset computer 100 and translated by the headset computer into commands that are sent back to aim the camera of the unmanned vehicle. As an example, if the wearer looks to the left, that motion is detected by the head tracker in the headset computer, translated into a camera “move left” command. That “move left” command is then sent wirelessly to the remote vehicle, causing the camera on the remote vehicle to pan to the left.


By returning the video stream back from the vehicle and displaying it on the microdisplay gives the wearer a visual experience as if he were, for example, a miniature pilot inside an unmanned aerial vehicle.


In yet another function, the user can, for example, use speech commands to control other peripherals that the vehicle itself might contain. An unmanned aerial vehicle such as shown in FIG. 8, may carry a payload such as a camera or other sensor to be dropped at a remote location. These payloads, weapons or other objects that the vehicle is capable of delivering can be controlled by the user of the headset computer 100. Control over these payloads can be implemented regardless of what the vehicle itself is being commanded to do.


System Description



FIG. 9 shows a wireless headset computer 100 (also referred to as a video eyewear device 100) that incorporates a high resolution (VGA or better) microdisplay element and other features described below. Headset computer 100 typically includes many different types of integrated circuits including a microprocessor (single or multi-core), one or more wireless interfaces, associated memory or other storage devices, one or more cameras (optical sensors) and/or various sensors. These sensors may include audio input and/or output devices, such as one or more microphone(s) input and output speaker(s) the sensors may include geo-positional sensing, 3 axis to 9 axis degrees of freedom orientational sensors (such as a digital magnetometer), atmospheric sensors, health condition sensors, GPS, digital compass, pressure sensors, environmental sensors, energy sensors, acceleration, position, attitude, motion, velocity or optical sensors, and cameras (visible, infrared, etc.). Further circuits such as additional wireless radios, auxiliary lighting, range finders, or the like, and/or an array of sensors may be embedded in and/or attached to the device. Also typically located within the device 100 are a peripheral mount or mounts such as a “hot shoe” (not shown in FIG. 9) for mounting optional accessories such as cameras or additional sensors. The camera(s), motion sensor(s) and/or sensor(s) are used to track the motion and/or position of the user's head, hands and/or body in at least a first axis 111 (horizontal), but preferably also a second (vertical), a third (depth), a fourth (pitch), a fifth (roll) and a sixth (yaw).


The headset computer device 100 can be used in various ways. It can be used as a completely contained, head-mounted fully functional portable personal computer/smart phone with full connectivity to external computers and networks through a short and/or long-range wireless links such as Bluetooth, WiFi, cellular, LTE, WiMax or other wireless radios.


Device 100 can be also used as a remote display for a streaming video signal provided by a remote host computer. The host may be, for example, a laptop, cell phone, Blackberry, iPhone™, or other computing device having lesser or greater computational complexity than the device 100 itself. The host then provides information to the device 100 to be displayed. The device 100 and host are connected via one or more suitable wireless connections such as provided by the Bluetooth WiFi, cellular, LTE, WiMax or other wireless radio link. The host may itself be further connected to other networks such as through a wired or wireless connection to the Internet.


While what is shown in FIG. 9 is a monocular microdisplay presenting a single fixed display element supported on the face of the user with a cantilevered boom, it should be understood that other mechanical configurations for various video eyewear devices 100 are possible.


In the FIG. 9 implementation, headset computer 100 includes generally a frame 1000, a strap 1002, a back section 1004, speaker 1006, cantilever or arm 1008, and microdisplay subassembly 1010. On one side of the device 100 opposite the cantilever arm 1008 is a peripheral port 1020. The peripheral port 1020 provides corresponding connections to one or more peripheral devices, so a user can removably attach various accessories to the device 100. As an example port 1020 provides a mechanical and electrical accessory mount such as a hot shoe. Wiring carries electrical signals from port 1020 through, for example, the back portion 1004 to circuitry disposed therein. Hot shoe 1020 can operate much like the hot shoe on a camera, automatically providing connections to power the accessory and carry signals to and from the rest of device 100. Various types of accessories can thus be used with port 1020 to provide the hand movements, head movements, and or vocal inputs to the system, such as but not limited to microphones, positional, orientation and other previously described sensors, cameras, and the like.



FIG. 1 was a view of the headset computer 100 as worn on the head of a user where an accessory 1060 has been placed in the hot shoe port 1020. This accessory 1060 is a self-contained camera (or other motion sensor) assembly. The camera 1060 can include both audio and video sensing, recording, and light emission capabilities in a package similar to a “bullet cam”. It can be connected to the remaining components in device 100 via built in wiring in the back section 1004 (as in the case of the speaker previously described) or can be wirelessly connected via a Bluetooth™ or WiFi™ connection. The camera 1060 may not necessarily be a video camera, but may also detect infrared, ultraviolet, or other wavelengths. The camera 1060 can also include a user adjustable auxiliary light source. With the light source, the camera 1060 can also be used as a light source as desired without activating the camera portion.


The camera, motion tracking and audio inputs to the device 100 are interpreted as user commands in various ways to control operation of the local processor, the microdisplay, or the external host.


Head movement tracking and/or vocal commands can also be provided by the user 1050 to manipulate the settings of camera 1060. For example, a user vocal command, such as “zoom” or “pan”, can be recognized by the local processor and cause the camera 1060 to zoom in or telephoto out.



FIG. 10 is a block diagram showing more detail of the device 100, optional host 200 and the data that travels between them. The device 100 receives audio signals input via the microphone, hand movements or body gestures via positional and orientation sensors, the camera or optical sensor(s), and head movement inputs via the head tracking circuitry such 3 axis to 9 axis degree of freedom orientation sensors. These are translated by software in a processor local to the device 100 into commands. These commands may then be interpreted by a processor internal to the device 100 to control aspects of the presentation of information on the microdisplay or other objects such as a peripheral or remote vehicle. The commands may also be sent over the Bluetooth or other wireless interface 150 to the host 200. The host 200 then interprets these translated commands in accordance with its own operating system/application software to perform various functions, and or returns information to the device 100. In one example, the device 100 and/or host 200 maintain a 3-D virtual space into what graphic objects are stored.


Among the commands that can be carried out on the local processor and/or the remote host 200 is one to select a field of view 300 within the virtual display. Thus, it should be understood that a very large format virtual display area might be associated with operating system or application software running on the device 100 or on the host 200. However, only a portion of that large virtual display area within the field of view is returned to and actually displayed by the remote control display device 120 as selected by the voice, hand gestures, or head motion commands.



FIG. 11 is a simplified high level block diagram of a non-limiting example embodiment of the headset computer device 100. The system electronics can be placed on or in the frame in an appropriate location (such as back section 1004) and include an Open Media Application Platform (OMAP) as the local processor 4110, a power/audio companion chip 4102, a display driver 4104, a head tracker circuit board 4106, and wireless LAN/Bluetooth interface 4108. Also located in the housing is a power source, such as a lithium ion battery 4200.


The device 100 may also include an eye pod assembly 4000 that includes the aforementioned microdisplay 4010 (e.g. the microdisplay 1010 and boom 1008 of FIG. 2A), and one or more microphones 4020. One or more speakers 4030 are positioned in the HMD housing earpiece near the user's ear (item 1006 in FIG. 9). The head tracker circuitry 4106 may include circuits to determine head movements and gestures detected by sensors in the device 100, such as lateral movements along and rotation gestures around the X, Y and Z axes using Hall effect sensors, MIM diodes, accelerometers, gyros and/or transducers or other sensors as mentioned above.


Device system 100 may also receive inputs from external input devices such as a wireless mouse, track ball, or keyboard that may be wirelessly connected through the Bluetooth interface 4108.


Software in the WLAN/BT front end 4108, the OMAP 4100 and/or host 200 may be used to interpret hand gestures detected by the camera or other sensors. A camera board 4060 may optionally provide video input, as well.


The OMAP processor 4100 may include a central processing unit, and on-chip memory such as Random Access Memory (RAM) that may include non volatile memory and/or Read Only Memory (ROM). The OMAP may be a Texas Instruments model OMAP 3530 processor or newer version sold by Texas Instruments, Inc. and using a multimedia processor. The OMAP 4100 may typically execute an embedded system such as operating a particular version of MicroSoft Windows®. The OMAP 4100 is generally a more powerful, and more power consuming processor than the WLAN/BT interface 4108.


In this example, a TPS 65950 power/audio companion chip, also available from Texas Instruments, provides audio, USB, keypad control and battery charging functions to the system.


The WLAN/BT interface 4108 may be a model LBEE 1W8 NEC-interface circuit, a Bluetooth circuit such as available from CSR, Ltd. of Cambridge, United Kingdom or other radio module with similar or greater capabilities.


The display driver may be a model KCD-A 910 display driver available from Kopin Corporation of Westborough, Massachusetts.


The microdisplay 4010, also available from Kopin, can include models CyberDisplay 230K, WQVGA, VGA, WVGA, SVGA or other manufactures' acceptable microdisplays.


An NCS module 4400 takes raw microphone signal data as input, and outputs audio data with background noise removed. It produces an audio signal to the audio companion chip 4102 and from there to the OMAP processor 4100. Voice recognition is performed in software on the OMAP processor 4100, using the cleaned up microphone signals as fed in by the NCS 4400.


The teachings of all patents, published applications and references cited herein are incorporated by reference in their entirety.


While this disclosure has described several example embodiments, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the scope of the invention encompassed by the appended claims.

Claims
  • 1. A headset computer system comprising: a first housing including a processor;a second housing including a boom with an integrated microdisplay and microphone;a third housing including a power supply, each of the housings separately packaged and individually attachable to a headgear;one or more signal and power connections between the first, second and third housings; andfasteners for separately attaching and detaching the first, second and third housing to and from the headgear; wherein the headgear has two pads disposed thereon, wherein the first housing is attachable to the headgear via a first pad of the two pads, and wherein the third housing is attachable to the headgear via a second pad of the two pads.
  • 2. The headset computer system of claim 1, wherein the headgear is a helmet.
  • 3. The headset computer system of claim 1, wherein the fasteners include hook and loop fasteners.
  • 4. The headset computer system of claim 1, wherein the first housing further encloses noise cancellation circuits.
  • 5. The headset computer system of claim 4, wherein the noise cancellation circuits are configured to reduce background noise of a rebreather.
  • 6. The headset computer system of claim 1 wherein the boom is configured to support the microdisplay.
  • 7. The headset computer system of claim 1, wherein the fasteners include a mechanical clip configured to attach the boom to the headgear.
  • 8. The headset computer system of claim 1, wherein the first, second, and third housings are configured to be retrofittable to the headgear.
  • 9. The headset computer system of claim 1, wherein the first housing further includes a camera, light source, or combination thereof.
  • 10. A method comprising: retrofitting headgear to include headset computer functionality, the retrofitting including implementing the headset computer functionality via a first housing, second housing, and third housing, each of the housings separately packaged and individually attachable to the headgear,the implementing employing a processor, a boom with an integrated microdisplay and microphone, and a power supply,the first housing including the processor, the second housing including the boom with the integrated microdisplay and microphone, the third housing including the power supply,the implementing further employing one or more signal and power connections between the first, second and third housings, andthe retrofitting further including employing fasteners for separately attaching and detaching the first, second and third housing to and from the headgear.
RELATED APPLICATIONS

This application is a continuation of U.S. application Ser. No. 16/748,668, entitled “Headset Computer That Uses Motion And Voice Commands To Control Information Display And Remote Devices,” filed on Jan. 21, 2020, which is a continuation of U.S. application Ser. No. 13/468,207, entitled “Headset Computer That Uses Motion And Voice Commands To Control Information Display And Remote Devices,” filed on May 10, 2012, now U.S. Pat. No. 10,627,860, which claims the benefit of U.S. Provisional Patent Application No. 61/484,464, entitled “Hands-Free Enhanced Digital Eyewear,” filed on May 10, 2011. The entire teachings of the above applications are incorporated herein by reference.

US Referenced Citations (289)
Number Name Date Kind
4567479 Boyd Jan 1986 A
5005213 Hanson et al. Apr 1991 A
5208449 Eastman May 1993 A
5594469 Freeman et al. Jan 1997 A
5689619 Smyth Nov 1997 A
5698834 Worthington Dec 1997 A
5742263 Wang et al. Apr 1998 A
5818455 Stone Oct 1998 A
5990793 Beiback Nov 1999 A
6010216 Jesiek Jan 2000 A
6084556 Zwern Jul 2000 A
6108197 Janik Aug 2000 A
6192343 Morgan et al. Feb 2001 B1
6198462 Daily et al. Mar 2001 B1
6204974 Spitzer Mar 2001 B1
6313864 Tabata Nov 2001 B1
6325507 Jannard et al. Dec 2001 B1
6346929 Fukushima et al. Feb 2002 B1
6369952 Rallison Apr 2002 B1
6408257 Harrington et al. Jun 2002 B1
6505055 Kahn et al. Jan 2003 B1
6532446 King Mar 2003 B1
6538676 Peters et al. Mar 2003 B1
6741836 Lee et al. May 2004 B2
6778906 Hennings et al. Aug 2004 B1
6798391 Petersen, III Sep 2004 B2
6853293 Swartz et al. Feb 2005 B2
6900777 Hebert et al. May 2005 B1
6922184 Lawrence et al. Jul 2005 B2
6956614 Quintana et al. Oct 2005 B1
6965862 Schuller Nov 2005 B2
6966647 Jannard et al. Nov 2005 B2
7004582 Jannard et al. Feb 2006 B2
7013009 Warren Mar 2006 B2
7082393 Lahr Jul 2006 B2
7147324 Jannard et al. Dec 2006 B2
7150526 Jannard et al. Dec 2006 B2
7213917 Jannard et al. May 2007 B2
7216973 Jannard et al. May 2007 B2
7219994 Jannard et al. May 2007 B2
7231038 Warren Jun 2007 B2
7249846 Grand et al. Jul 2007 B2
7278734 Jannard et al. Oct 2007 B2
7331666 Swab et al. Feb 2008 B2
7445332 Jannard et al. Nov 2008 B2
7452073 Jannard et al. Nov 2008 B2
7458682 Lee Dec 2008 B1
7461936 Jannard Dec 2008 B2
7494216 Jannard et al. Feb 2009 B2
7501995 Morita et al. Mar 2009 B2
7512414 Jannard et al. Mar 2009 B2
7522880 Stieber et al. Apr 2009 B2
7620432 Williams et al. Nov 2009 B2
7620433 Bodylet Nov 2009 B2
7682018 Jannard Mar 2010 B2
7740353 Jannard Jun 2010 B2
7744213 Jannard et al. Jun 2010 B2
7753520 Fuziak, Jr. Jul 2010 B2
7760898 Howell et al. Jul 2010 B2
7798638 Fuziak, Jr. Sep 2010 B2
7806525 Howell et al. Oct 2010 B2
7918556 Lewis Apr 2011 B2
7959084 Wulff Jun 2011 B2
7966189 Le et al. Jun 2011 B2
7967433 Jannard et al. Jun 2011 B2
7969383 Eberl et al. Jun 2011 B2
7969657 Cakmakci et al. Jun 2011 B2
7969918 Takatori et al. Jun 2011 B2
7976480 Grajales et al. Jul 2011 B2
7988283 Jannard et al. Aug 2011 B2
7997723 Pienimaa et al. Aug 2011 B2
8010156 Warren Aug 2011 B2
8020989 Jannard et al. Sep 2011 B2
8025398 Jannard Sep 2011 B2
8072393 Riechel Dec 2011 B2
8092011 Sugihara et al. Jan 2012 B2
8098439 Amitai et al. Jan 2012 B2
8099046 Helferich Jan 2012 B2
8108143 Tester Jan 2012 B1
8123352 Matsumoto et al. Feb 2012 B2
8140197 Lapidot et al. Mar 2012 B2
8170262 Liu May 2012 B1
8184983 Ho et al. May 2012 B1
8212859 Tang et al. Jul 2012 B2
8327295 Ikeda Dec 2012 B2
8577427 Serota Nov 2013 B2
8838075 Basir Sep 2014 B2
8855719 Jacobsen et al. Oct 2014 B2
8862186 Jacobsen et al. Oct 2014 B2
8885719 Kondo et al. Nov 2014 B2
8929954 Jacobsen et al. Jan 2015 B2
9118875 Ida Aug 2015 B2
9122307 Jacobsen et al. Sep 2015 B2
9235262 Jacobsen et al. Jan 2016 B2
9294607 Jacobsen et al. Mar 2016 B2
9301085 Parkinson et al. Mar 2016 B2
9316827 Lindley et al. Apr 2016 B2
9369760 Jacobsen et al. Jun 2016 B2
9507772 Parkinson et al. Nov 2016 B2
9817232 Lindley et al. Nov 2017 B2
10013976 Woodall et al. Jul 2018 B2
10627860 Jacobsen et al. Apr 2020 B2
11237594 Jacobsen et al. Feb 2022 B2
20010003712 Roelofs Jun 2001 A1
20010035845 Zwern Nov 2001 A1
20020015008 Kishida et al. Feb 2002 A1
20020030649 Zavracky et al. Mar 2002 A1
20020044152 Abbott, III et al. Apr 2002 A1
20020065115 Lindholm May 2002 A1
20020094845 Inasaka Jul 2002 A1
20020130818 Viertl Sep 2002 A1
20020154070 Sato et al. Oct 2002 A1
20020158815 Zwern Oct 2002 A1
20030016253 Aoki et al. Jan 2003 A1
20030035266 Ng Feb 2003 A1
20030046401 Abbott et al. Mar 2003 A1
20030065805 Barnes Apr 2003 A1
20030067536 Boulanger et al. Apr 2003 A1
20030068057 Miller et al. Apr 2003 A1
20030169339 Allen et al. Sep 2003 A1
20030222917 Trantow Dec 2003 A1
20040100389 Naito May 2004 A1
20040102967 Levin May 2004 A1
20040113867 Tominc Jun 2004 A1
20040189675 Pretlove et al. Sep 2004 A1
20040193413 Wilson et al. Sep 2004 A1
20040210852 Balakrishnan et al. Oct 2004 A1
20040267527 Creamer et al. Dec 2004 A1
20050047629 Farrell et al. Mar 2005 A1
20050108643 Schybergson et al. May 2005 A1
20050114140 Brackett et al. May 2005 A1
20050237296 Lee Oct 2005 A1
20050245292 Bennett et al. Nov 2005 A1
20050261890 Robinson Nov 2005 A1
20050264527 Lindholm Dec 2005 A1
20060007056 Ou Jan 2006 A1
20060010368 Kashi Jan 2006 A1
20060012884 Snorteland et al. Jan 2006 A1
20060028400 Lapstun et al. Feb 2006 A1
20060061544 Min et al. Mar 2006 A1
20060061551 Fateh Mar 2006 A1
20060074624 Sahashi Apr 2006 A1
20060109234 Hong et al. May 2006 A1
20060109237 Morita et al. May 2006 A1
20060132382 Jannard Jun 2006 A1
20060166705 Seshadri et al. Jul 2006 A1
20060178085 Sotereanos et al. Aug 2006 A1
20060221266 Kato et al. Oct 2006 A1
20060238877 Ashkenazi et al. Oct 2006 A1
20070009125 Frerking et al. Jan 2007 A1
20070030174 Randazzo et al. Feb 2007 A1
20070030259 Karttunen Feb 2007 A1
20070053544 Jhao et al. Mar 2007 A1
20070093279 Janik Apr 2007 A1
20070103388 Spitzer May 2007 A1
20070121423 Rioux May 2007 A1
20070180979 Rosenberg Aug 2007 A1
20070220108 Whitaker Sep 2007 A1
20070238475 Goedken Oct 2007 A1
20070265495 Vayser Nov 2007 A1
20070291974 Eisenbraun Dec 2007 A1
20080027591 Lenser et al. Jan 2008 A1
20080052643 Ike et al. Feb 2008 A1
20080055194 Baudino et al. Mar 2008 A1
20080084992 Peddireddy et al. Apr 2008 A1
20080120141 Kariathungal et al. May 2008 A1
20080144854 Abreu Jun 2008 A1
20080154613 Haulick Jun 2008 A1
20080169998 Jacobsen Jul 2008 A1
20080180640 Ito Jul 2008 A1
20080198324 Fuziak Aug 2008 A1
20080200774 Luo Aug 2008 A1
20080201634 Gibb et al. Aug 2008 A1
20080211768 Breen et al. Sep 2008 A1
20080239080 Moscato Oct 2008 A1
20090002640 Yang et al. Jan 2009 A1
20090018419 Torch Jan 2009 A1
20090079839 Fischer et al. Mar 2009 A1
20090093304 Ohta Apr 2009 A1
20090099836 Jacobsen et al. Apr 2009 A1
20090117890 Jacobsen et al. May 2009 A1
20090128448 Riechel May 2009 A1
20090154719 Wulff et al. Jun 2009 A1
20090180195 Cakmakci et al. Jul 2009 A1
20090182562 Claire et al. Jul 2009 A1
20090204410 Mozer et al. Aug 2009 A1
20090213071 Wang et al. Aug 2009 A1
20090240488 White Sep 2009 A1
20090244048 Yamanaka Oct 2009 A1
20090251409 Parkinson Oct 2009 A1
20090318163 George Dec 2009 A1
20100001699 Dragojevic Jan 2010 A1
20100020229 Hershey et al. Jan 2010 A1
20100033830 Yung Feb 2010 A1
20100041447 Graylin Feb 2010 A1
20100053069 Tricoukes et al. Mar 2010 A1
20100073201 Holcomb et al. Mar 2010 A1
20100106497 Phillips Apr 2010 A1
20100117930 Bacabara et al. May 2010 A1
20100119052 Kambli May 2010 A1
20100121480 Stelzer et al. May 2010 A1
20100128626 Anderson et al. May 2010 A1
20100141554 Devereaux et al. Jun 2010 A1
20100156812 Stallings et al. Jun 2010 A1
20100164990 Van Doorn Jul 2010 A1
20100169073 Almagro Jul 2010 A1
20100171680 Lapidot et al. Jul 2010 A1
20100182137 Pryor Jul 2010 A1
20100204981 Ribeiro Aug 2010 A1
20100225734 Weller et al. Sep 2010 A1
20100235161 Kim et al. Sep 2010 A1
20100238184 Janicki Sep 2010 A1
20100238396 Jannard Sep 2010 A1
20100245585 Fisher et al. Sep 2010 A1
20100250231 Almagro Sep 2010 A1
20100271587 Pavlopoulos Oct 2010 A1
20100277563 Gupta et al. Nov 2010 A1
20100289817 Meier et al. Nov 2010 A1
20100295772 Alameh et al. Nov 2010 A1
20100302137 Benko et al. Dec 2010 A1
20100302282 Dobbie Dec 2010 A1
20100306711 Kahn et al. Dec 2010 A1
20100309295 Chow Dec 2010 A1
20100315329 Previc Dec 2010 A1
20110001699 Jacobsen et al. Jan 2011 A1
20110089207 Tricoukes et al. Apr 2011 A1
20110090135 Tricoukes et al. Apr 2011 A1
20110092157 Clark et al. Apr 2011 A1
20110092825 Gopinathan et al. Apr 2011 A1
20110134910 Chao-Suren et al. Jun 2011 A1
20110187640 Jacobsen et al. Aug 2011 A1
20110214082 Osterhout et al. Sep 2011 A1
20110221656 Haddick et al. Sep 2011 A1
20110221669 Shams Sep 2011 A1
20110221671 King, III et al. Sep 2011 A1
20110227812 Haddick et al. Sep 2011 A1
20110227813 Haddick et al. Sep 2011 A1
20110238405 Pedre Sep 2011 A1
20110248904 Miyawaki Oct 2011 A1
20110254698 Eberl et al. Oct 2011 A1
20110254865 Yee et al. Oct 2011 A1
20110255050 Jannard et al. Oct 2011 A1
20110273662 Hwang et al. Nov 2011 A1
20120013843 Jannard Jan 2012 A1
20120026071 Hamdani et al. Feb 2012 A1
20120056846 Zaliva Mar 2012 A1
20120062445 Haddick et al. Mar 2012 A1
20120068914 Jacobsen et al. Mar 2012 A1
20120075177 Jacobsen et al. Mar 2012 A1
20120089392 Larco et al. Apr 2012 A1
20120092208 LeMire et al. Apr 2012 A1
20120992208 LeMire et al. Apr 2012
20120105740 Jannard et al. May 2012 A1
20120110456 Larco et al. May 2012 A1
20120114131 Tricoukes et al. May 2012 A1
20120166203 Fuchs Jun 2012 A1
20120088245 Rotter et al. Jul 2012 A1
20120173100 Ellis Jul 2012 A1
20120188245 Hyatt Jul 2012 A1
20120218303 Nakada Aug 2012 A1
20120236025 Jacobsen et al. Sep 2012 A1
20120287284 Jacobsen et al. Nov 2012 A1
20120302288 Born et al. Nov 2012 A1
20130019182 Gil et al. Jan 2013 A1
20130041670 Morgan et al. Feb 2013 A1
20130070930 Johnson Mar 2013 A1
20130174205 Jacobsen et al. Jul 2013 A1
20130231937 Woodall et al. Sep 2013 A1
20130239000 Parkinson et al. Sep 2013 A1
20130274985 Lee et al. Oct 2013 A1
20130288753 Jacobsen et al. Oct 2013 A1
20130289971 Parkinson Oct 2013 A1
20130290882 Cotte Oct 2013 A1
20130300649 Parkinson et al. Nov 2013 A1
20140003616 Johnson et al. Jan 2014 A1
20140059263 Rosenberg et al. Feb 2014 A1
20140093103 Breece et al. Apr 2014 A1
20140111427 Lindley et al. Apr 2014 A1
20140223299 Han Aug 2014 A1
20140235169 Parkinson et al. Aug 2014 A1
20140334644 Selig Nov 2014 A1
20140368412 Jacobsen et al. Dec 2014 A1
20150039311 Clark et al. Feb 2015 A1
20150072672 Jacobsen et al. Mar 2015 A1
20150279354 Gruenstein Oct 2015 A1
20150346489 Lindley et al. Dec 2015 A1
20180277114 Woodall et al. Sep 2018 A1
20190279636 Woodall et al. Sep 2019 A1
20200159283 Jacobsen et al. May 2020 A1
Foreign Referenced Citations (81)
Number Date Country
1335702 Feb 2002 CN
1698048 Nov 2005 CN
1702607 Nov 2005 CN
1735019 Feb 2006 CN
1797299 Jul 2006 CN
1815437 Aug 2006 CN
1828586 Sep 2006 CN
1960670 May 2007 CN
101185016 May 2008 CN
101196793 Jun 2008 CN
101243392 Aug 2008 CN
101349944 Jan 2009 CN
101444087 May 2009 CN
101581969 Nov 2009 CN
101599267 Dec 2009 CN
101620511 Jan 2010 CN
201393284 Jan 2010 CN
101706850 May 2010 CN
101755299 Jun 2010 CN
101809651 Aug 2010 CN
101986255 Mar 2011 CN
102541438 Jul 2012 CN
102724353 Oct 2012 CN
102736827 Oct 2012 CN
102812417 Dec 2012 CN
103620527 Mar 2014 CN
103 44 062 Apr 2005 DE
2 207 164 Jul 2010 EP
02-084686 Mar 1990 JP
09-034895 Feb 1997 JP
10-020867 Jan 1998 JP
2000-215022 Aug 2000 JP
2001-100878 Apr 2001 JP
2001-506389 May 2001 JP
2001-202175 Jul 2001 JP
2001-216069 Aug 2001 JP
2002-525769 Aug 2002 JP
2003-241880 Aug 2003 JP
2004-233117 Aug 2004 JP
2005-012377 Jan 2005 JP
2005-352619 Dec 2005 JP
2007-079978 Mar 2007 JP
2007-213501 Aug 2007 JP
2008-052590 Mar 2008 JP
2008-278536 Nov 2008 JP
2009-179062 Aug 2009 JP
2010-102163 May 2010 JP
2011-511935 Apr 2011 JP
2011-198150 Oct 2011 JP
2012-002568 Jan 2012 JP
2012-500526 Jan 2012 JP
2012-044429 Mar 2012 JP
2012-056568 Mar 2012 JP
2012-174149 Sep 2012 JP
2012-533055 Dec 2012 JP
2013-541092 Nov 2013 JP
2004-54879 Feb 2014 JP
2015-515701 May 2015 JP
WO 1995021408 Aug 1995 WO
WO 1995023994 Sep 1995 WO
WO 1999001838 Jan 1999 WO
WO 2000017848 Mar 2000 WO
WO 2000079327 Dec 2000 WO
WO 2005017729 Feb 2005 WO
WO 2006082764 Aug 2006 WO
2009017797 Feb 2009 WO
WO 2009076016 Jun 2009 WO
WO 2009091639 Jul 2009 WO
WO 2009120984 Oct 2009 WO
WO 2010019634 Feb 2010 WO
WO 2010129679 Nov 2010 WO
WO 2011004026 Jan 2011 WO
WO 2011051660 May 2011 WO
WO 2011097226 Aug 2011 WO
WO 2011114149 Sep 2011 WO
WO 2012025956 Mar 2012 WO
WO 2012040107 Mar 2012 WO
WO 2012040386 Mar 2012 WO
WO 2012154938 Nov 2012 WO
WO 2013101438 Jul 2013 WO
WO 2013162908 Oct 2013 WO
Non-Patent Literature Citations (8)
Entry
International Preliminary Report on Patentability for PCT/US2011/023337 dated Aug. 16, 2012; entitled “Wireless Hands-Free Computing Headset with Detachable Accessories Controllable by Motion, Body Gesture and/or Vocal Commands”, 8 pages.
International Preliminary Report on Patentability for PCT/US2012/037284, Headset Computer that Uses Motion and Voices to Control Information Display and Remote Devices, dated Nov. 21, 2013, 7 pages.
International Preliminary Report on Patentability for PCT/US2013/041070 dated Jul. 16, 2015; entitled “Context Sensitive Overlays in Voice Controlled Headset Computer Displays”.
International Search Report and Written Opinion for PCT/US2012/037284 dated Oct. 1, 2012, entitled “Headset Computer That Uses Motion and Voices to Control Information Display and Remote Devices”.
International Search Report and Written Opinion for PCT/US2012/068686, dated Mar. 25, 2013; entitled Wireless Hands-Free Computing Head Mounted Video Eyewear for Local/Remote Diagnosis And Repair, 11 pages.
International Search Report and Written Opinion for PCT/US2013/065927 dated Mar. 21, 2014, entitled, “Improved Headset Computer Operation Using Vehicle Sensor Feedback for Remote Control Vehicle.”
International Search Report and Written Opinion for PCT/US2013/065927 dated Jul. 16, 2015, entitled, “Improved Headset Computer Operation Using Vehicle Sensor Feedback for Remote Control Vehicle.”
Morphew, M. E., et al., “Helmet, Mounted Displays for Unmanned Aerial Vehicle Control,” Proceedings of SPIE, vol. 5442 (2004).
Related Publications (1)
Number Date Country
20220057831 A1 Feb 2022 US
Provisional Applications (1)
Number Date Country
61484464 May 2011 US
Continuations (2)
Number Date Country
Parent 16748668 Jan 2020 US
Child 17516962 US
Parent 13468207 May 2012 US
Child 16748668 US