1. Technical Field
This disclosure relates to recognition systems and related methods, and, in particularly, to a gesture recognition system and a related method that recognizes a bodily gesture or a hand-gesture of a user.
2. Description of Related Art
Currently, motion recognition of the user is generally implemented by an external camera or a plurality of motion sensors distributed across the body of the user.
The external camera may be a depth-sensing camera and enabled reliable body tracking. However, the external camera usually has a limited working distance owing to the angle-of-view of the camera.
On the other hand, although the motion sensors may be wearable devices that are put on arms, legs, shoulders, fingers and so on, wearing these motion sensors is commonly inconvenient.
Therefore, it is an urgent issue in the art to provide a recognition system and a recognition method that can recognize a bodily gesture or a hand-gesture of a user so as to improve the above defects.
The present disclosure provides a gesture recognition system, comprising: a camera configured to capture a user to obtain one or more raw images, wherein the camera is mounted on the user and has a wide angle-of-view; a transmitter electrically connected to the camera to transmit the one or more raw images; and a processor configured to receive and process the one or more raw images, transform the processed one or more raw images into a corresponding gesture image, and build a recognition module according to a plurality of gesture images, such that the processor recognizes a gesture of the user through the recognition module when one or more new raw images are captured by the camera.
The present disclosure further provides a method for recognizing a gesture, comprising: mounting a camera having a wide angle-of-view on a central portion of a body of a user; capturing a sequence of raw images of a limb of the user by the camera; receiving and processing the sequence of raw images; transforming the sequence of raw images into a corresponding gesture image; building a recognition module according to a plurality of gesture images; and recognizing a bodily gesture of the limb of the user through the recognition module when a new sequence of raw images is captured by the camera.
The present disclosure also provides a method for recognizing a gesture, comprising: mounting a camera having a wide angle-of-view on a finger of a user; capturing a raw image of a hand of the user by the camera; receiving and processing the raw image; transforming the raw image into a corresponding gesture image; building a recognition module according to the gesture image; and recognizing a hand gesture of the user through the recognition module when a new raw image is obtained by the camera.
The disclosure can be more fully understood by reading the following detailed descriptions of the embodiments, with reference made to the accompanying drawings, wherein:
In the following detailed description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the disclosed embodiments. It will be apparent, however, that one or more embodiments may be practiced without these specific details. In other instances, well-known structures and devices are schematically shown in order to simplify the drawings.
The camera 11 includes an image sensor and one or more lenses, and is configured to capture an image or a sequence of images of a user. The image sensor may be provided with a charge-coupled device (CCD) or a complementary metal-oxide-semiconductor (CMOS) type device that converts the received light intensities into corresponding electrical signals. The one or more lenses may have a fish-eye lens or a lenses-assembly with a wide angle-of-view. In an embodiment, the camera 11 is equipped with a wide-angle lens and may have an angle-of-view more than 180 degree, such as 185 degree or 235 degree.
In an embodiment, the camera 11 further includes an emitter attached around the one or more lenses, such that the camera emits light to the user and receive the light reflected from the user to obtain depth-like information related to the images. Accordingly, the processor 15 can distinguish the user from the background based on the depth-like information. The emitter may be implemented by infrared LEDs to provide uniform illumination, such that the user looks brighter than a background object. The camera 11 with the attached infrared LEDs can detect both visible and infrared light. To facilitate extracting the user from images, a filter is included in the camera 11 to block visible light and allow only infrared reflection from foreground objects, such as the body, to pass. Alternatively, the camera 11 may be a time-of-flight depth camera which can capture images with depth information.
The transmitter 13 is electrically connected to the camera 11 to transmit images to the processor 15. The transmitter 13 can also be mounted on the user or combined with the camera 11.
In an embodiment, the camera 11 is mounted on a central portion of the body of a user, for example, the camera 11 may be mounted on the chest. In an embodiment, the camera 11 and the transmitter 13 can be integrated into a single-piece wearable device. As shown in
The processor 15 is configured to receive and process one or more raw images from the transmitter 13. As such, a sequence of raw images can be processed and transformed into a corresponding gesture image that represents a bodily gesture or a hand gesture. In an embodiment, the processor 15 may employ a plurality of gesture images to build a recognition module stored in the memory 17. Accordingly, when a new raw image is captured by the camera 11, the processor 15 generates a corresponding new gesture image and recognizes a bodily gesture or a hand gesture with the recognition module according to the new gesture image. In an embodiment, the processor 15 and the memory 17 may be incorporated into a computer or a processing unit. The details will be described later.
In step 202, a camera having a wide angle-of-view is mounted on a central portion of a body of a user. In an embodiment, the camera has an angle-of-view of more than 180 degrees, such as 235 degrees.
In step 204, the camera captures a sequence of raw images of at least one limb of the user. In an embodiment, the camera further obtains depth information related to the sequence of raw images.
In step 206, the processor receives and processes the sequence of raw images. The processed sequence of raw images distinguish the at least one limb of the user from a background object. For example, the at least one limb of the user is distinguished from the background object with or without the depth information and marked as foreground objects.
In step 208, the processor generates a gesture image according to the processed sequence of raw images. The gesture image has spatial and temporal information of the at least one limb of the user, such that the gesture image represents a bodily gesture. Subsequently, the processor builds a recognition module given a plurality of gesture images. For example, the recognition module is trained by a plurality of gesture images with corresponding known bodily gesture(s), such that the trained recognition module is capable of recognizing one or more bodily gestures. In an embodiment, the recognition module is stored in a memory.
In step 210, when the user performs a bodily gesture, the camera captures a new sequence of raw images of the limb(s) of the user. The processor transforms the new sequence of raw images into a corresponding new gesture image and recognizes the bodily gesture performed by the user through the recognition module according to the new gesture image.
In an embodiment, the camera and the transmitter are integrated into a single-piece ring-style wearable device and worn on a user. The camera has an angle-of view of 185 degrees and is equipped with a fish-eye lens.
In step 302, a camera having a wide angle-of-view is mounted on a finger of a user. In an embodiment, the camera has an angle-of-view of more than 180 degrees, such as 185 degrees.
In step 304, the camera captures a raw image of a hand of the user, where a portion of the hand, such as the fingers or a part of the palm, may be captured.
In step 306, the processor receives and processes the raw image. The processed raw image distinguishes the fingers and palm of the user from a background object by using the color information in which the fingers and palm of the user are considered as foreground objects.
In step 308, the processor generates a gesture image according to the processed raw image. Subsequently, the processor builds a recognition module given a plurality of gesture images. For example, the recognition module is trained by a plurality of gesture images with corresponding known hand gesture(s), such that the trained recognition module is capable of recognizing one or more hand gestures. In an embodiment, the recognition module is stored in a memory.
In step 310, when the user performs a hand gesture, the camera captures a new raw image of the hand of the user. The processor transforms the new raw image into a corresponding new gesture image and recognizes the hand gesture performed by the user through the recognition module according to the new gesture image.
In an embodiment, the memory further pre-stores at least one activation gesture image corresponding to at least one interaction mode. Accordingly, the processor is operated in the interaction mode when a new gesture image matches the activation gesture image.
In an embodiment, the camera and the transmitter are integrated into a single-piece wearable device, and can be worn as a ring. The camera has an angle-of view of 185 degrees and is equipped with a fish-eye lens.
In an embodiment, the camera 11 can be positioned on a central portion of the hand of the user. In an embodiment, as shown in
In an embodiment, a plurality of activation gesture images are stored in the memory. As such, when a new gesture image representing a new hand gesture matches one of the activation gesture images, the processor enters the corresponding interaction mode. For example, the user may bend his thumb to enable a writing input mode, such that the user can use his index finger of one hand to write on the palm of another hand.
It will be apparent to those skilled in the art that various modifications and variations can be made to the disclosed embodiments. It is intended that the specification and examples be considered as exemplary only, with a true scope of the disclosure being indicated by the following claims and their equivalents.