(NOT APPLICABLE)
The invention relates to augmented reality devices and, more particularly, to touchless gesture recognition using a built-in camera for executing an action or program.
As augmented reality devices become more of a reality, touchless methods of interacting with software become more important. While voice recognition is one possible way of interaction, the built-in camera in many portable devices has not been used to recognize gestures.
Gestures provide an opportunity for users to signal to the device their interests with limited potential distortions, increasing accuracy, as well as limiting the possible interference by outside noise, thereby increasing reliability.
An important component for success of the interface is an easily recognizable gesture. A “duckbill” gesture is exemplary. Shaping a user's hand into the shape of the bill of the duck, the user can point to point of interest elements in the field of vision. The system would treat the point as an analog to a mouse pointer. Using the duckbill gesture as an exemplary user gesture, to enact a mouse click, the user would “explode” his fingers.
In a merchant funded debit rewards program or the like, this could be used to point to a restaurant or merchant, and “click” to bring up a detailed view of the location.
In an exemplary embodiment, a method of recognizing a touchless interaction is performed with a computer including a processor, a display, a memory storing computer programs executable by the processor, and a camera capturing digital images and delivering signals representative of the digital images to the processor. The method includes the steps of (a) the processor identifying a user pre-gesture based on the digital images from the camera; (b) the processor associating a point of interest on the display according to one of a shape and a position of the user pre-gesture; (c) the processor identifying a change from the user pre-gesture to a user gesture based on the digital images from the camera; and (d) the processor executing an appropriate one of the computer programs based on the user gesture.
The user pre-gesture may include a convergence point, where step (b) is practiced by associating the convergence point with an object on the display. Step (a) may be practiced by capturing an image with the camera of the user's hand with at least two fingers of the user's hand converged such that tips of the at least two fingers are in proximity to or are touching one another, the convergence point being where the at least two fingers of the user's hand are converged. Step (c) may be practiced by identifying the change from the user pre-gesture to the user gesture by identifying an image of the user's hand captured by the camera of the at least two fingers spreading out from the convergence point.
Step (a) may be practiced by capturing an image with the camera of the user's hand with all fingers of the user's hand converged such that the tips of the fingers are in proximity to or are touching one another, the convergence point being where the fingers of the user's hand are converged. In this context, step (c) may be practiced by identifying the change from the user pre-gesture to the user gesture by identifying an image of the user's hand captured by the camera of the fingers spreading out from the convergence point.
The method may further include, after step (c), a step of executing the change from the user pre-gesture to the user gesture as a mouse click.
Steps (a) and (c) may be practiced in a field of vision of the camera without the user touching the display.
In another exemplary embodiment, a computer system for recognizing a touchless interaction includes a processor, a display, a memory storing computer programs executable by the processor, and a camera capturing digital images and delivering signals representative of the digital images to the processor. The processor is programmed to identify a user pre-gesture based on the digital images from the camera and to associate a point of interest on the display according to one of a shape and a position of the user pre-gesture. The processor is programmed to identify a change from the user pre-gesture to a user gesture based on the digital images from the camera, and the processor is programmed to execute an appropriate one of the computer programs based on the user gesture.
In yet another exemplary embodiment, a method of recognizing a touchless interaction with a computer includes the steps of identifying a duckbill gesture of a user's hand based on the digital images of the user's hand from the camera; associating a point of interest on the display according to one of a shape and a position of the duckbill gesture; identifying a change from the duckbill gesture to an exploded gesture based on the digital images of the user's hand from the camera; and executing an appropriate one of the computer programs based on the change.
These and other aspects and advantages will be described in detail with reference to the accompanying drawings, in which:
With reference to the drawings, the system and methodology according to preferred embodiments of the invention utilize a computer's built-in video camera to recognize a dynamic hand gesture. The dynamic hand gesture may be used to point to a point of interest on the computer display in the field of vision of the camera. The system treats the point and hand gesture as an analog to a mouse click.
With reference to
In the exemplary embodiment shown in the figures, the user pre-gesture is a “duckbill” gesture made by the user's hand where at least two fingers of the user's hand are converged such that tips of the at least two fingers are in proximity to or are touching one another. The convergence point 18 is where the at least two fingers of the user's hand are converged. In the embodiment shown in
With reference to
With reference to
The system described with reference to
1. The user runs a web browser program on his/her computer.
2. The user connects to the server computer (e.g., via the Internet). Connection to the server computer may be conditioned upon the correct entry of a password as is well known.
3. The user requests a page from the server computer. The user's browser sends a message to the server computer that includes the following:
the transfer protocol (e.g., http://); and
the address, or Uniform Resource Locator (URL).
4. The server computer receives the user's request and retrieves the requested page, which is composed, for example, in HTML (Hypertext Markup Language).
5. The server then transmits the requested page to the user's computer.
6. The user's browser program receives the HTML text and displays its interpretation of the requested page.
Thus, the browser program on the user's computer sends requests and receives the data needed to display the HTML page on the user's computer screen. This includes the HTML file itself plus any graphic, sound and/or video files mentioned in it. Once the data is retrieved, the browser formats the data and displays the data on the user's computer screen. Helper applications, plug-ins, and enhancements such as Java™ enable the browser, among other things, to play sound and/or display video inserted in the HTML file. The fonts installed on the user's computer and the display preferences in the browser used by the user determine how the text is formatted.
If the user has requested an action that requires running a program (e.g., a search), the server loads and runs the program. This process usually creates a custom HTML page “on the fly” that contains the results of the program's action (e.g., the search results), and then sends those results back to the browser.
Browser programs suitable for use in connection with the system and methodology of the present invention include Mozilla Firefox® and Internet Explorer available from Microsoft® Corp.
While the above description contemplates that each user has a computer running a web browser, it will be appreciated that more than one user could use a particular computer terminal or that a “kiosk” at a central location (e.g., a cafeteria, a break area, etc.) with access to the system server could be provided.
It will be recognized by those in the art that various tools are readily available to create web pages for accessing data stored on a server and that such tools may be used to develop and implement the system described below and illustrated in the accompanying drawings.
A number of program modules may be stored on the hard disk 211, removable magnetic disk 215, optical disk 219 and/or ROM 252 and/or RAM 254 of the system memory 205. Such program modules may include an operating system providing graphics and sound APIs, one or more application programs, other program modules, and program data. A user may enter commands and information into computer system 201 through input devices such as a keyboard 227 and a pointing device 229. Other input devices may include a microphone, joystick, game controller, satellite dish, scanner, or the like. A digital camera 236 may also be coupled with the system bus 207 as an input device. These and other input devices are often connected to the processing unit 203 through a serial port interface 231 that is coupled to the system bus 207, but may be connected by other interfaces, such as a parallel port interface or a universal serial bus (USB). A monitor 233 or other type of display device is also connected to system bus 207 via an interface, such as a video adapter 235.
The computer system 201 may also include a modem or broadband or wireless adapter 237 or other means for establishing communications over the wide area network 239, such as the Internet. The modem 237, which may be internal or external, is connected to the system bus 207 via the serial port interface 231. A network interface 241 may also be provided for allowing the computer system 201 to communicate with a remote computing device 250 via a local area network 258 (or such communication may be via the wide area network 239 or other communications path such as dial-up or other communications means). The computer system 201 will typically include other peripheral output devices, such as printers and other standard peripheral devices.
As will be understood by those familiar with web-based forms and screens, users may make menu selections by pointing-and-clicking using a mouse, trackball or other pointing device, or by using the TAB and ENTER keys on a keyboard or using the touchless gesture recognition of the described embodiments. For example, menu selections may be highlighted by positioning the cursor on the selections using a mouse or by using the TAB key. The mouse may be left-clicked to select the selection or the ENTER key may be pressed. Other selection mechanisms including voice-recognition systems, touch-sensitive screens, etc. may be used, and the invention is not limited in this respect.
While the invention has been described in connection with what is presently considered to be the most practical and preferred embodiments, it is to be understood that the invention is not to be limited to the disclosed embodiments, but on the contrary, is intended to cover various modifications and equivalent arrangements included within the spirit and scope of the appended claims.
This application claims the benefit of U.S. Provisional Patent Application Ser. No. 61/926,543, filed Jan. 13, 2014, the entire content of which is herein incorporated by reference.
Number | Date | Country | |
---|---|---|---|
61926543 | Jan 2014 | US |