The present invention relates to imaging and, more particularly, to hands-free imaging.
Frequently, an individual in a mobile environment encounters visual information that he or she wants to record. A conventional way to record the visual information is to take a picture using a hand-held camera, which may be a conventional digital camera or a cellular phone camera.
Unfortunately, hand-held cameras cannot be used by individuals who do not have hands or who are afflicted by a neuromuscular disorder that causes involuntary hand movement. Moreover, using a hand-held camera in a mobile environment can be a time-intensive and clumsy practice. Before taking the picture, the camera must be powered-up or enabled. This requires pressing a button or a selector switch and may also require opening a clamshell assembly, all of which involve use of one and potentially both hands. The individual must also frame the object of interest. This requires that the individual move the camera or the object, or both, and possibly adjust a lens focus control by hand. The individual must also press the shutter button to capture the image, which is normally done by hand. Additional use of hands is often required to offload the captured image to a personal computer or server for processing and/or printing. These requirements leave hand-held cameras ill-suited to many real-world imaging applications, such as a research application where an individual wants to quickly image select pages of a manuscript that he or she is holding and reviewing or a security application where a government official wants to image passports of inbound travelers for rapid validation.
The present invention, in a basic feature, provides a voice activated headset imaging system and elements thereof that enable hands-free imaging. The system is well suited to a mobile environment but can be used in a stationary (e.g. desktop) environment as well.
In one aspect of the invention, a headset assembly for a voice activated headset imaging system comprises a head frame, a microphone assembly having microphone logic coupled with the head frame and a camera assembly having camera logic coupled with the head frame, wherein the camera logic is adopted to execute a control instruction generated in response to a voice command received by the microphone logic.
In some embodiments, the camera assembly has a camera and an adjustable arm and the camera is coupled with the head frame via the adjustable arm.
In some embodiments, the camera assembly has a camera and an object pointer coupled with the camera and the object pointer is directionally disposed to illuminate an object within a field of view of the camera.
In some embodiments, the headset assembly further comprises a wireless network interface adopted to transmit the voice command to a wireless handset and receive the control instruction from the wireless handset.
In some embodiments, the headset assembly further comprises a system on chip adapted to receive the voice command from the microphone logic, generate the control instruction and transmit the control instruction to the camera logic.
In some embodiments, execution of the control instruction awakens the camera assembly from a power-saving state.
In some embodiments, execution of the control instruction causes the camera assembly to enter a power-saving state.
In some embodiments, execution of the control instruction activates the object pointer.
In some embodiments, execution of the control instruction captures an image within a field of view of the camera.
In some embodiments, execution of the control instruction deletes a captured image.
In some embodiments, execution of the control instruction downloads an image from the camera assembly to the wireless handset.
In some embodiments, execution of the control instruction captures an image within a field of view of the camera and downloads the image from the camera assembly to the wireless handset for processing and emailing to a predetermined address.
In another aspect of the invention, a wireless handset for a voice activated headset imaging system comprises a processor and a wireless network interface communicatively coupled with the processor, wherein the wireless handset receives from a headset assembly via the wireless network interface a voice command and in response to the voice command under control of the processor generates and transmits to the headset assembly via the wireless network interface an imaging control instruction.
In some embodiments, the wireless handset further receives from the headset assembly via the wireless network interface an image and under control of the processor stores the image on the wireless handset.
In some embodiments, the wireless handset further receives from the headset assembly via the wireless network interface an image and under control of the processor enhances the image.
In some embodiments, the wireless handset further receives from the headset assembly via the wireless network interface an image and under control of the processor routes the image to a predetermined address.
In yet another aspect of the invention, a method for hands-free imaging in a mobile environment comprises the steps of receiving by a headset assembly a voice command and executing by the headset assembly an imaging control instruction generated in response to the voice command.
In some embodiments, the method further comprises the steps of transmitting by the headset assembly via a wireless network interface the voice command and receiving by the headset assembly via the wireless network interface the imaging control instruction.
In some embodiments, execution of the imaging control instruction involves at least one operation selected from the group consisting of activating an object pointer, capturing an image, deleting a captured image and downloading a captured image.
In some embodiments, the method further comprises the step of executing by the headset assembly an audio control instruction to output status information on the voice command.
These and other aspects of the invention will be better understood by reference to the following detailed description taken in conjunction with the drawings that are briefly described below. Of course, the invention is defined by the appended claims.
Head frame 120 has been illustrated as a sealed head frame that represents a particularly robust type of head frame that provide sonic isolation and a high degree of structural stability. In other embodiments, an open-air head frame may be employed that has smaller over-the-ear earpieces held in place by a light head band that allow the user's ears to remain partially exposed to the ambient environment and provides a high degree of comfort over an extended period of use. In still other embodiments, an earbud-type head frame may be used in which the earpieces fit into the outer ear of the user and are held in place by a light head band or attachment clips. In still other embodiments, a canal-type head frame may be used in which the earpieces fit into the user's ear canals and are held in place by a light head band or attachment clips.
If the next voice command is a POINTER-ON or POINTER-OFF command, the system turns-on or turns-off, respectively, object pointer 290 (630) and returns to the listening state (610). When microphone logic 310 receives a POINTER-ON voice command, microphone logic 310 transmits to wireless handset 120 via wireless interface 340 the voice command in digital form for interpretation and processing. In response to the POINTER-ON voice command, wireless handset 120 generates and returns to camera logic 320 via wireless interface 420 a control instruction that camera logic 320 executes to activate object pointer 290. When microphone logic 310 receives a POINTER-OFF voice command, microphone logic 310 transmits to wireless handset 120 via wireless interface 340 the voice command in digital form for interpretation and processing. In response to the POINTER-OFF voice command, wireless handset 120 generates and returns to camera logic 320 via wireless interface 420 a control instruction that camera logic 320 executes to deactivate object pointer 290.
If the next voice command is a CAPTURE command, the system captures the image within the present field of view of camera 280 (640) and returns to the listening state (610). When microphone logic 310 receives a CAPTURE voice command, microphone logic 310 transmits to wireless handset 120 via wireless interface 340 the voice command in digital form for interpretation and processing. In response to the CAPTURE voice command, wireless handset 120 generates and returns to camera logic 320 via wireless interface 420 a control instruction that camera logic 320 executes to actuate image capture. The image capture actuated by the CAPTURE command may be one of single frame capture (i.e. still imaging), burst frame capture or sequential frame capture at a predetermined rate (i.e. full motion video). Where full motion video is captured, the system may also capture via microphone 250 and store audio synchronized with the full motion video. The full motion video and accompanying audio may be captured for a predetermined time, or may continue until a second voice command indicating to terminate image capture (e.g. STOP) is processed.
If the next voice command is an IGNORE command, the system deletes the most recently captured image (650) and returns to the listening state (610). When microphone logic 310 receives an IGNORE voice command, microphone logic 310 transmits to wireless handset 120 via wireless interface 340 the voice command in digital form for interpretation and processing. In response to the IGNORE voice command, wireless handset 120 generates and returns to camera logic 320 via wireless interface 420 a control instruction that camera logic 320 executes to delete from camera image store 330 the most recently captured image.
If the next voice command is a CLEAR command, the system deletes all images from camera image store 330 (660) and returns to the listening state (610). When microphone logic 310 receives a CLEAR voice command, microphone logic 310 transmits to wireless handset 120 via wireless interface 340 the voice command in digital form for interpretation and processing. In response to the CLEAR voice command, wireless handset 120 generates and returns to camera logic 320 via wireless interface 420 a control instruction that camera logic 320 executes to delete all images from camera image store 330.
If the next voice command is an EXPORT command, the system downloads to wireless handset 120 all images from camera image store 330 (670) and returns to the listening state (610). When microphone logic 310 receives an EXPORT voice command, microphone logic 310 transmits to wireless handset 120 via wireless interface 340 the voice command in digital form for interpretation and processing. In response to the EXPORT voice command, wireless handset 120 generates and returns to camera logic 320 via wireless interface 420 a control instruction that camera logic 320 executes to download to wireless handset 120 via wireless interface 340 all images presently stored in camera image store 330.
If the next voice command is an EMAIL command, the system performs a multifunction workflow operation in which an image is captured, downloaded, processed and emailed (680) before returning to the listening state (610). When microphone logic 310 receives an EMAIL voice command, microphone logic 310 transmits to wireless handset 120 via wireless interface 340 the voice command in digital form for interpretation and processing. In response to the EMAIL voice command, wireless handset 120 generates and returns to camera logic 320 via wireless interface 420 a control instruction that camera logic 320 executes to actuate image capture and download the captured image to wireless handset 120. Wireless handset 120 then performs image processing on the captured and downloaded image (e.g. compression) and emails the image to a predetermined email address associated with client device 160.
Other imaging control instructions are possible. For example, in some embodiments a RECORD command is supported that when received by microphone logic 310 causes the system to store in camera store 330 in association with a recently captured still image audio information spoken into microphone 250 by the user. The RECORD command can be invoked to add voice annotation to the still image.
In addition to processing imaging control instructions, the system is operative to execute audio control instructions. Camera logic 320 transmits to wireless handset 120 via wireless interface 340 periodic or event-driven status information in response to which wireless handset 120 issues audio control instructions to speaker logic 350 that speaker logic 350 executes to inform the user via audible output on the loudspeaker of the status of voice commands and camera image store 330. Such audible output may, for example, notify the user that a voice command was or was not understood or has or has not been carried-out, or that camera image store 330 is at or near capacity. Such audible output may be delivered in the form of predefined tones or prerecorded messages, for example.
It will be appreciated by those of ordinary skill in the art that the invention can be embodied in other specific forms without departing from the spirit or essential character hereof. For example, in other embodiments, a headset assembly includes dual cameras and an object is imaged by the two cameras simultaneously. The image pairs are used to compute depth information, or increase the effective imaging area and/or effective imaging resolution. The present description is therefore considered in all respects to be illustrative and not restrictive. The scope of the invention is indicated by the appended claims, and all changes that come with in the meaning and range of equivalents thereof are intended to be embraced therein.