This Application claims priority of Taiwan Patent Application No. 96150828, filed on Dec. 28, 2007, the entirety of which is incorporated by reference herein.
1. Field of the Invention
The invention relates to electronic devices and related operation methods, and more particularly, to operation methods for use in an electronic device having an image capture unit.
2. Description of the Related Art
Driven by user requirements, more and more electronic devices, especially handheld or portable electronic devices such as smart phones, personal digital assists (PDAs), and tablet PCs or Ultra Mobile PCs (UMPCs), comprise various peripherals such as a video camera, for improving user convenience.
In general, a user issues a control instruction to the electronic device through the input unit such as a keyboard, a mouse or a touch-sensitive screen. A user may also issue a control instruction to the electronic devices by voice-controlling. The recognition capability of voice-controlling, however, depends on environmental noise at the time of recognition. Thus, should environmental noise be great enough, the recognition capability of voice-controlling would be relatively low and a wrong instruction may be interpreted or an instruction may not be executable. In addition, some electronic devices may use a video camera for performing video talk over a network or photographing images.
A simple operation method for use in an electronic device having an image capture unit is provided to intuitively and quickly issue commonly used or complex control instructions by users, so as to improve convenience in use and control of the electronic device.
An embodiment of an operation method for operating an electronic device having an image capture unit is provided. The method comprises the following steps. (a) An input image of an object having first and second sub images are acquired from the image capture unit. (b) The acquired first and second sub images are recognized to obtain information regarding outlines or positions of the first and second sub images. (c) A corresponding control instruction is generated according to the relative relationship of the outlines or positions of the first and second sub images. (d) At least one operation corresponding to the generated control instruction is performed.
An embodiment of an electronic device is also provided. The electronic device comprises an image capture unit, a recognition unit and a processing unit. The image capture unit acquires an input image of an object having first and second sub images. The recognition unit recognizes positions of the acquired first and second sub images and generates a corresponding control command according to the relative relationship between the positions of the first and second sub images. The processing unit performs at least one operation corresponding to the generated control instruction.
Another embodiment of an operation method for operating an electronic device having an image capture unit is further provided. The method comprises the following steps. A first image and a second image different from the first image are first acquired from the image capture unit. The acquired first and second images are then recognized and a corresponding control instruction is generated according to a variational relationship between the first and second images. At least one operation corresponding to the generated control instruction is accordingly performed.
The invention can be more fully understood by reading the subsequent detailed description and examples with reference to the accompanying drawings, wherein:
The following description is of the best-contemplated mode of carrying out the invention. This description is made for the purpose of illustrating the general principles of the invention and should not be taken in a limiting sense. The scope of the invention is best determined by reference to the appended claims.
Embodiments of the invention provide an operation method for use in an electronic device having an image capture unit (e.g. camera or video camera), such as a smart phone, a personal digital assist (PDA), and a tablet PC or an Ultra Mobile PC (UMPC). An input image is acquired by the image capture unit and the acquired image is then recognized. Then, the recognized result of the acquired image (still or motion image) is converted to generate a corresponding control instruction based on a predefined relationship corresponding thereto so as to direct the electronic device to perform at least one operation corresponding to the generated control instruction.
The embodiments provide intuitive operation methods to issue control instructions to the electronic devices by analyzing interaction between a user's hand and his (her) face or by a motion of the user's hand. An image corresponding to a control instruction may be inputted to the electronic device through the image capture unit (e.g. video camera), and then the inputted image would be recognized and the control instruction corresponding to the recognized image would then be obtained. Accordingly, an operation corresponding to the obtained control instruction would be performed. Therefore, simplifying the operation process.
The image capture unit 110 acquires an input image of an object and transfers the acquired image to the recognition unit 120 for image recognition of the acquired image. The input image at least comprises still and motion images in which the still image includes first and second sub images while the motion image represents a motion image performing a specific motion on a 2D (two-dimension) or 3D (three-dimension) plane. The recognition unit 120 may effectively perform an image recognition operation to recognize positions and outlines of the acquired first and second sub images, and obtain the relationship corresponding to the positions of the first and second sub images. The method for recognizing the outlines and positions of the first and second sub images by the recognition unit 120 are detailed in the following. Note that only one image capture unit is utilized to acquire the first and second sub images in this embodiment. However, in other embodiments, more than one image capture unit may be utilized to acquire the first and second sub images. In this case, sub images may be acquired by different image capture units and all of the acquired sub images may be sent to the database 150 for comparison at the same time.
The processing unit 140 may obtain the relative relationship between positions of the first and second sub images based on the recognized result from the recognition unit 120, and then inspect the database 150 to generate a corresponding control instruction and perform at least one operation corresponding to the generated control instruction. In other embodiments, operations for obtaining the relative relationship between positions of the first and second sub images and further inspecting the database 150 to generate a corresponding control instruction may be performed by the recognition unit 120 instead of the processing unit 140. As a result, the processing unit 140 may perform operations corresponding to the control instruction generated by the recognition unit 120. The database 150 may comprise a plurality of predetermined images in which each predetermined image would include the first and second sub images, and the relative positions of the first and second sub images in one predetermined image may be different from those in another predetermined image. Each predetermined image corresponds to a control instruction and the control instruction may be pre-inputted by users. For example, a user interface may be provided for allowing a user to input each of the required control instructions and setup an image representing an instruction corresponding thereto in advance and the inputted data would be stored into the database 150. Therefore, while receiving an acquired image, the recognition unit 210 may compare the acquired image to images pre-stored in the database 150 and inspect whether any matched image is found, and if so, output the control instruction corresponding to the found image as the corresponding control instruction to the processing unit 140. An image is normally recognized as being matched, if positions and outlines of the sub images therein are the same as those of the acquired image.
For example, a user may easily make a motion of putting his (her) forefinger on his (her) lips to remotely issue an instruction for turning on the mute function. This motion would be recognized and converted to a mute instruction and the mute instruction would then be sent to the processing unit 140. Upon receiving the mute instruction, the processing unit 140 would perform related operations for turning the mute function, such as turning off the volume of the speaker.
The electronic device 100 may further comprise a display unit (not shown) for providing various displays for user operation. In this embodiment, the display unit may display a message indicator which corresponds to the control instruction represented by the input image for user indication. The user may determine and ensure whether the control instruction received by the electronic device 100 is correct or not based on the displayed message indicator.
In this embodiment, a user interface capable of performing the operation method of the invention may be activated by the user by a specific method such as hot key activation from a hardware source, automatic activation, voice control activation or key activation from a software source. Activation of the user interface may be user defined or based on customer requirement. Hot key activation from a hardware source may be achieved by pressing at least one specific key or button for activation of a desired function. Automatic activation may be activated or enabled when detecting a specific motion of a user's hand and disabled when the detected specific motion of the user's hand has disappeared. Voice control activation may be achieved or disabled by a user using his (her) voice to issue an instruction or cancel the instruction. Key activation from a software source may be achieved by control from a software source.
It is to be noted that, for illustration purpose, in the following embodiments, the electronic device 100 is assumed to be a personal computer (PC) and the image capture unit 110 is assumed to be a video camera, but it is not limited thereto.
After the locating procedure is completed, bench marked images of the hand and face images can be obtained. Then, a motion corresponding to a required operation may be made in front of the video camera. Note that it is assumed that the mute function required is to turn on the mute function and that the motion corresponding to turning on the mute function is a motion of putting his (her) forefinger on his (her) lips. Next, the user makes a motion by putting his (her) forefinger on his (her) lips to issue an instruction to turn on the mute function. Then, the computer acquires the input image through the video camera and sends the acquired image to the recognition unit 120 for image recognition. The motion of putting the user's forefinger on the user's lips, will be recognized by the recognition unit 120 and a corresponding control instruction for turning on the mute function would be generated. Thus, a message indicator “whether to turn on the mute function?” corresponding to the generated control instruction will be displayed in the display unit. At the same time, the user may determine whether the control instruction received by the electronic device is correct based on the displayed message indicator. If the user requires to cancel the proposed execution of turning on the mute function or a recognition result for the acquired image is incorrect, a specific key [SPACE] may be pressed to inform the computer to cancel the operation and revert back to the previous step for allowing the user to once again acquire the image. On the other hand, a confirmation instruction may be inputted according to a predefined rule if the user requires the mute function to be turned on. In this embodiment, the user may stand in front of the video camera for three seconds without any motion to ensure and implement the operation of turning on the mute function. Thereafter, the computer would perform related operations for turning on the mute function, such as turning off the volume of the speaker, according to the control instruction.
Note that outlines of different hand images (e.g. hand gesture) may represent different control instructions even if the position of the hand image and face image are overlapped. Therefore, the outline of the hand image has to be recognized during the recognition operation. For example, the user may make a motion of keeping his five fingers opened in front of his mouth to issue a control instruction to turn off the mute function which is slightly similar to the turning on the mute function.
Moreover, the user may make a dynamic motion to issue a specific control instruction. In this case, the database 150 may comprise a plurality of predefined motion images. Each of the predefined motion images correspond to a control instruction that is pre-inputted by the user.
When the input image is a motion image performing a specific motion on a 2D or 3D plane, a recognition result for the acquired motion image is sent to the processing unit 140 by the recognition unit 120. After receiving the recognized motion image, the processing unit 140 sends the recognized motion image to the motion analyzer unit 130 for motion determination. The motion analyzer unit 130 may compare the recognized motion image to predefined motion images pre-stored in the database 150 and inspect whether a matched motion image is found, and if so, output the control instruction corresponding to the found motion image.
A motion may be recognized as a 2D plane relative image or a 3D plane relative image based on a difference in the static and dynamic range of the video camera. The motion of a 2D plane relative image may be achieved by making a simple motion without considering the layers presented on the screen. The motion of a 3D plane relative image may, however, consider layers presented on the screen and a distance between the video camera and the hand may be detected and determined, to obtain more than two kinds of level distances for response to layer relation of files or folders presented on the display unit.
Referring to
In some embodiments, if the user makes a motion on a 3D plane, after the video camera position has been located, a distance between the video camera and the object (e.g. user's hand) may also be utilized to determine whether the object is far away or close to the screen based on an image size of the acquired image, so as to determine layers of the displayed screen in a display unit. In some embodiments, a distance between the video camera and the object and a fixed hand gesture may be combined to issue a set of control instructions.
In some embodiments, image size of the input image acquired by the video camera may be utilized to obtain a distance between the video camera and the acquired image and be further utilized to input a specific control instruction. Referring to
In some embodiments, inputting of the control instruction may be achieved by analyzing image size of the input image acquired by the video camera to obtain a distance between the video camera and the acquired image along with variations in hand gestures. Referring to
After the processing unit 140 obtains the control instruction, in step S530, related operations corresponding to the obtained control instruction will then be performed.
In summary, with the operation method of the invention, the user can intuitively make a motion (such as in combination with his hand and face or moving his hand) to issue an instruction to the electronic device through the image capture unit when needed, thereby effectively simplifying use for users when issuing required instructions and improving user convenience.
While the invention has been described by way of example and in terms of preferred embodiment, it is to be understood that the invention is not limited thereto. To the contrary, it is intended to cover various modifications and similar arrangements (as would be apparent to the skilled in the art). Therefore, the scope of the appended claims should be accorded to the broadest interpretation so as to encompass all such modifications and similar arrangements.
Number | Date | Country | Kind |
---|---|---|---|
TW96150828 | Dec 2007 | TW | national |