Field of the Invention
The present invention relates generally to user interface devices, and more specifically to touchless user interface devices that use hand gestures.
Background of the Invention
Humans have been using hand gestures to communicate for as long as there have been humans. The human hand is a versatile and highly expressive communication organ. As technology gets more and more omnipresent in our lives, the idea of using hand gestures to communicate with the various technological objects we use becomes more and more appealing.
There are some attempts to create a gesture interface to communicate with a computer or tablet. Such attempts tend to leverage the computer's power and resources to receive images of a person's hand, process those images, and interpret the person's gestures. However, all of this is very resource-intensive and takes up a lot of computing power.
A need exists for a gesture interface for simpler devices that have little or no computing power—light switches, lamps, appliances, and so on. While motion sensors are currently used for such devices, a motion sensor cannot communicate detailed information such as may be needed to operate an appliance or even a light with a dimmer switch. However, the no-touch nature of a motion sensor may be desirable for some applications—for example, in an operating room of a hospital where touch may compromise sterility.
An object of the present invention is to provide a gesture interface for a device with minimal computing power that is self-contained, simple, and cheap.
Another object of the present invention is to provide a system and method for identifying hand positions and motions.
The method of the present invention preferably comprises illuminating a hand using a first frequency of light (preferably infrared), taking a first image of the hand using a camera, turning off illumination and taking a second image of the hand, subtracting the second image from the first image to obtain a clean image, and analyzing the clean image of the hand.
In an embodiment, the clean image of the hand is analyzed to determine the pose of the hand; this is preferably done by creating a classification tree to classify each hand pose according to at least one category and at least one subcategory, and then determining a category and subcategory for the clean image of the hand.
In an embodiment, the hand is illuminated again and a third image is taken; then the illumination is turned off and a fourth image is taken. The fourth image is subtracted from the third image to produce a second clean image of the hand. Both the clean image of the hand and the second clean image of the hand are then processed using an adaptive threshold to generate a shape, and circles are inscribed into the shape. The circles are preferably greater in diameter than a predetermined number. Then, the first shape is subtracted from the second shape, and all the circles are overlaid on top of the image. Each circle is evaluated for whether or not it contains any non-black pixels. If at least one circle containing only black pixels is below the difference image, the system concludes that the hand is moving up; if it is above the difference image, the hand is moving down; if it is to the left of the difference image, the hand is moving to the right; and if it is to the right of the difference image, the hand is moving to the left.
In an embodiment, the system also evaluates the distance between the difference image and the furthest circle containing only black pixels. That is used to estimate the speed of motion of the hand.
In an embodiment, the steps are repeated to generate a trajectory for the hand.
In the preferred embodiment, the first frequency of light is infrared and the camera is an infrared camera.
In an embodiment, the illumination is turned on and off at a regular frequency of 120 Hz and the camera takes images at a regular frequency of 240 Hz.
In an embodiment, the image is cropped to just the image of the hand to remove unnecessary blank space. This is preferably done by using an adaptive filter to binary the image, determining pixel intensity in the image, and cropping the image to just the rectangular area where pixel intensity is nonzero.
The results of the analyzing step may be used to control any device; examples include a light switch, a music player, a toilet, a water faucet, a shower, a thermostat, or medical equipment.
A system of the present invention preferably comprises a camera, a light source, and a processor that performs the above functions.
Several embodiments of the present invention are described below. It will be understood that the present invention encompasses all reasonable equivalents to the below-described embodiments, as is evident to a person of reasonable skill in the art.
A block diagram of an embodiment of the system of the present invention is shown in
In the preferred embodiment, the infrared LED flashes on and off at regular intervals, and the camera is triggered to take a photographic image of the hand with and without the infrared illumination.
After the two images are taken (one with IR illumination and one without), the processor preferably subtracts one image from the other to remove the background.
In an embodiment, the image is cropped to remove extraneous blank space and to save memory. This is preferably done in a manner shown in
The image is then analyzed to interpret the position or motion of the hand. In one embodiment, the static hand pose is identified. For example, the hand pose could be a fist, an open palm, a thumbs-up sign, and so on. To identify the hand pose, the shape of the hand is compared with images of various hand poses stored in memory. In one embodiment, the images stored in memory are classified according to at least one classification and at least one sub-classification to form a classification tree. For example, some classifications could be “open hand”, “closed hand” —then the “open hand” poses could be further classified into “palm forward” or “palm back” and the “closed hand” poses could be further classified into “thumb out” or “thumb in”, and so on.
In an alternate embodiment, the motion of the hand is identified. This embodiment is illustrated in
After a blob analysis is performed on each image, the two images are subtracted from each other and all the circles are overlaid on the resulting image. This is known as trace analysis. The processor then looks for circles that are empty and circles that are partially filled and circles that are entirely filled. The relative locations and distances of the empty circles and the partially or entirely filled circles is what determines the speed and direction of motion of the hand. In the example shown in the Figure, the empty circles are below the filled circles; this means that the hand is moving upward.
In an embodiment, only the four cardinal directions (up, down, left, right) can be determined. In another embodiment, the system can also evaluate the angle at which the hand is moving, calculating the “center of mass” of the sum total of the empty circles and the “center of mass” of the sum total of the filled or partially filled circles.
In an embodiment, the speed of motion of the hand is also determined. This is estimated based on the distance between the furthest empty circle from the filled or partially filled circle.
The applications of the present invention can be numerous. For example, a user could turn on a water faucet with a hand gesture, and change the water temperature with another hand gesture. A user could use a hand gesture to turn on a music player and another hand gesture to control its volume. A hand gesture could be used to flush a toilet and different hand gestures could be used to trigger the toilet to perform other functions, such as bidet functions, heating the seat, air-drying functions, and so on. In a hospital setting, different hand gestures could be used to control various medical equipment without touching it and thus compromising sterility. Due to the present invention's simplicity, it could be built into a device easily without increasing its footprint or energy usage, or it could be a separate standalone module that could be connected to a device wirelessly or by a cable.
As used herein any reference to “one embodiment” or “an embodiment” means that a particular element, feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment. The appearances of the phrase “in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment.
Some embodiments may be described using the expression “coupled” and “connected” along with their derivatives. For example, some embodiments may be described using the term “coupled” to indicate that two or more elements are in direct physical or electrical contact. The term “coupled”, however, may also mean that two or more elements are not in direct contact with each other, but yet still co-operate or interact with each other. The embodiments are not limited in this context.
As used herein, the terms “comprises”, “comprising”, “includes”, “including,” “has”, “having” or any other variation thereof, are intended to cover a non-exclusive inclusion. For example, a process, method, article, or apparatus that comprises a list of elements is not necessarily limited to only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Further, unless expressly stated to the contrary, “or” refers to an inclusive or and not to an exclusive or. For example, a condition A or B is satisfied by any one of the following: A is true (or present) and B is false (or not present), A is false (or not present) and B is true (or present), and both A and B are true (or present).
In addition, use of the “a” or “an” are employed to describe elements and components of the embodiments herein. This is done merely for convenience and to give a general sense of the invention. This description should be read to include one or at least one and the singular also includes the plural unless it is obvious that it is meant otherwise.
Upon reading this disclosure, those of ordinary skill in the art will appreciate still additional alternative structural and functional designs through the disclosed principles of the embodiments. Thus, while particular embodiments and applications have been illustrated and described, it is to be understood that the embodiments are not limited to the precise construction and components disclosed herein and that various modifications, changes and variations which will be apparent to those skilled in the art may be made in the arrangement, operation and details of the method and apparatus disclosed herein without departing from the spirit and scope as defined in the appended claims.
The present application takes priority from Provisional App. No. 62/276,967, filed Jan. 11, 2016, and Provisional App. No. 62/276,969, filed Jan. 11, 2016, which are herein incorporated by reference.
Number | Date | Country | |
---|---|---|---|
62276967 | Jan 2016 | US | |
62276969 | Jan 2016 | US |