The present disclosure relates to the field of human-computer interaction techniques, and in particular, to a human-computer interaction method, and a related device and system.
Human-computer interaction techniques generally refer to techniques for implementing dialogues between people and a terminal device effectively by using an input/output device of the terminal device (for example, a computer or a smart phone). The techniques include that the terminal device provides a large quantity of related information, prompts, requests, and the like for people by using an output device or a display device, and that people input a related operation instruction into the terminal device by using an input device, to control the terminal device to execute a corresponding operation. The human-computer interaction techniques are one of important parts in computer user interface designing and are closely associated with subject areas such as cognition, human engineering, and psychology.
The human-computer interaction techniques have evolved their input manners gradually from the primary keyboard input and mouse input to touch screen input and finger gesture input. The gesture input has advantages such as direct operation and high user experience and is increasingly favored by people. However, in practical applications, the finger gesture input generally is implemented by directly capturing and interpreting a finger gesture by using an ordinary camera. Through practices, it is found that directly capturing and interpreting a finger gesture by using an ordinary camera has a poor anti-interference performance, thereby causing low operational accuracy.
In the existing technology, directly capturing and interpreting a finger gesture by using an ordinary camera has a poor anti-interference performance and causes low operational accuracy.
In view of the above, the present disclosure provides a human-computer interaction method and related device and system, which are capable of improving an anti-interference performance of a finger gesture input, thereby improving the operational accuracy.
According to one aspect of the present disclosure, the human-computer interaction method, which is at a terminal device having one or more processors and memory for storing program modules to be executed by the one or more processors, includes:
capturing, using a camera module, an image including an auxiliary light source formed by a finger gesture on an auxiliary light screen located in front of the camera module;
processing the image to acquire an image that only displays the auxiliary light source formed by the finger gesture on the auxiliary light screen;
determining a position and/or a motion track of the auxiliary light source in the image captured by the camera module; and
executing a corresponding operation instruction according to the position and/or the motion track.
Correspondingly, according to another aspect of the present disclosure, a terminal device has one or more processors, memory, and one or more program modules stored in the memory and to be executed by the one or more processors, the one or more program modules further comprising:
a camera module, configured to capture an image including an auxiliary light source formed by a finger gesture on an auxiliary light screen located in front of the camera module;
a processing module, configured to process the image to acquire an image that only displays the auxiliary light source formed by the finger gesture on the auxiliary light screen;
a determining module, configured to determine a position and/or a motion track of the auxiliary light source in an image captured by the camera module; and
an executing module, configured to execute a corresponding operation instruction according to the position and/or the motion track.
Correspondingly, according to another aspect of the present disclosure, a human-computer interaction system, comprising an auxiliary light screen, a camera, and a terminal device, the camera being built into the terminal device or being connected to the terminal device in a wired or wireless manner, and a photographing area of the camera covering a working coverage area of the auxiliary light screen;
the auxiliary light screen being touched by a finger so as to form an auxiliary light source;
the camera capturing the auxiliary light source formed by the finger gesture on the auxiliary light screen; and
the terminal device further including:
As can be known from the foregoing technical solutions, in the described aspects of the present disclosure, the terminal device can capture, by using the camera module, an auxiliary light source formed by a finger gesture on an auxiliary light screen, determine a position and/or a motion track of the auxiliary light source in an image captured by the camera module, further query for a code corresponding to the position and/or the motion track, and execute an operation instruction corresponding to the code. It can be seen from the above that, the present disclosure implements a human-computer interaction on the basis of an auxiliary light source, which not only achieves a very good anti-interference performance and higher operational accuracy, but also has a great commercial value.
To describe the technical solutions of the embodiments of the present application or the existing technology more clearly, the following briefly introduces the accompanying drawings required for describing the embodiments or the existing technology. Apparently, the accompanying drawings in the following description show only some embodiments of the present application, and a person of ordinary skill in the art may still derive other drawings from these accompanying drawings without creative efforts.
The following describes in details the respective embodiments of the present application with reference to the accompanying drawings in the embodiments of the present application. Apparently, the described embodiments are only some of the embodiments of the present application rather than all of the embodiments. All other embodiments obtained by a person of ordinary skill in the art based on the embodiments of the present application without creative efforts shall fall within the protection scope of the present disclosure.
The embodiments of the present application provide a human-computer interaction method, and a related device and system. In the human-computer interaction method, a terminal device captures, by using a camera module, an auxiliary light source formed by a finger gesture on an auxiliary light screen, and determines a position and/or a motion track of the auxiliary light source in an image captured by the camera module. Further, the terminal device executes a corresponding operation instruction according to the position and/or the motion track. The human-computer interaction method of the embodiments of the present application can improve an anti-interference performance of a finger gesture input, thereby improving operational accuracy. The embodiments are described in detail separately below.
Step 101: A terminal device captures, by using a camera module, an auxiliary light source formed by a finger gesture on an auxiliary light screen located in front of the camera module.
In an embodiment for implementing the present disclosure, the terminal device for implementing the human-computer interaction method may be a computer, a smart phone, or a television set in which the control software is installed and that has a computing capability, or may also be a household intelligent device, a commercial intelligent device, an office intelligent device, a mobile Internet device (MID), or the like, which is not specifically limited in this embodiment of the present application.
In this embodiment, the camera module may be built into the terminal device, which includes but is not limited to: a terminal device such as a notebook computer, a tablet computer, a smart phone, or a personal digital assistant (PDA), for example, a camera built in a terminal device such as a camera-equipped computer, a smart phone, a tablet computer, or a PDA. The camera module may also be disposed by being externally connected to the terminal device. For example, the camera module may be connected to the terminal device by using a universal serial bus (USB), the camera module may be connected to the terminal device by using a remote network (Wide Area Network, WAN), or the camera module may also be connected to the terminal device in a wireless manner such as Bluetooth, Wi-Fi, and infrared rays. In an embodiment of the present application, the camera may be built in the human-computer interaction terminal, externally connected to the human-computer interaction terminal, or disposed by combining the two manners. The connection manner between the camera and the human-computer interaction terminal may be: wired connection, wireless connection, or a combination of the two connection manners.
In an embodiment of the present application, the terminal device may capture, by using the camera module, the image that includes the auxiliary light source formed by the finger gesture on the auxiliary light screen, and process the image so as to acquire an image that only displays the auxiliary light source formed by the finger gesture on the auxiliary light screen, so that step 101 is implemented.
The embodiments of the present application will subsequently describe in detail, by using examples, the specific implementation procedures of processing the image by using the camera module, so as to acquire the image that only displays the auxiliary light source formed by the finger gesture on the auxiliary light screen, which are not described herein.
In an embodiment of the present application, the camera module may be an infrared-light camera. Correspondingly, the auxiliary light screen may be an infrared-light auxiliary light screen. In this case, the auxiliary light source formed by the finger gesture on the auxiliary light screen is a highlighted auxiliary light source.
In another embodiment of the present application, the camera module may be a visible-light camera. Correspondingly, the auxiliary light screen may be a visible-light auxiliary light screen. In this case, the auxiliary light source formed by the finger gesture on the auxiliary light screen is a dark auxiliary light source.
The embodiments of the present application will subsequently describe in detail the specific implementation of the auxiliary light screen, which are not described herein.
Step 102: The terminal device determines a position and/or a motion track of the auxiliary light source in an image captured by the camera module.
In an embodiment of the present application, if the finger touches the auxiliary light screen by means of tapping so as to form the auxiliary light source, the terminal device may determine a block number indicating where the auxiliary light source falls into the image captured by the camera module, and use the block number indicating where the auxiliary light source falls into the image captured by the camera module as a position of the auxiliary light source in the image captured by the camera module; and if the finger touches the auxiliary light screen by means of sliding so as to form the auxiliary light source, the terminal device may determine a quantity of blocks that the auxiliary light source goes through in the image captured by the camera module and a direction of the auxiliary light source, and use the quantity of the blocks that the auxiliary light source goes through in the image captured by the camera module and the direction of the auxiliary light source as a motion track of the auxiliary light source in the image captured by the camera module.
In this embodiment, the image captured by the camera module may be evenly divided into a plurality of blocks by the terminal device by using a certain corner (for example, an upper left corner) as an original point.
Step 103: The terminal device queries for a code corresponding to the position and/or the motion track.
In an embodiment of the present application, the control software of the terminal device may query, according to the block number indicating where the auxiliary light source falls into the image captured by the camera module, a stored mapping between blocks and codes for a code corresponding to the block number indicating where the auxiliary light source falls into the image captured by the camera module.
In another implementation of the present application, the control software of the terminal device may query, according to the quantity of the blocks that the auxiliary light source goes through in the image captured by the camera module and the direction of the auxiliary light source, a stored mapping among quantities of blocks, directions, and codes for a code corresponding to the quantity of the blocks that the auxiliary light source goes through in the image captured by the camera module and the direction of the auxiliary light source.
The embodiments of the present application will subsequently describe in detail the mapping between the blocks and the codes, and the mapping among the quantities of the blocks, the directions, and the codes, which are not described herein.
Step 104: The terminal device acquires an operation instruction corresponding to the code from a stored mapping between codes and operation instructions according to the found code, and executes the operation instruction corresponding to the code.
The embodiments of the present application will subsequently describe in detail the mapping between the codes and the operation instructions, which are not described herein.
In an embodiment of the present application, the operation instruction may be a computer operation instruction (for example, a mouse operation instruction such as opening, closing, zooming in, or zooming out) or a television remote control instruction (for example, a remote control operation instruction such as turning on, turning off, increasing volume, decreasing volume, switching to a lower channel number, switching to a higher channel number, or muting).
In an embodiment of the present application, the auxiliary light screen overlaps or is parallel to a display screen. When the auxiliary light screen is parallel to the display screen, the auxiliary light screen an infrared-light auxiliary light screen superposed with one visible-light light screen, and the visible-light light screen is used to indicate a position of the auxiliary light screen.
In the human-computer interaction method described in
The foregoing describes in detail the human-computer interaction method according to some embodiments of the present application.
According to yet another embodiment of the present application, a human-computer interaction method is further provided.
Step 201: A terminal device can capture, by using a camera module, an auxiliary light source formed by a finger tap on an auxiliary light screen.
In an embodiment of the present application, reference is made to
In an embodiment of the present application, a mobile phone may also be used to illuminate the screen so as to implement the auxiliary light screen. This manner is simple and effective and also has a low cost.
In an embodiment of the present application, it is assumed that the laser in
Reference is made to
Step 202: The terminal device determines a position of the auxiliary light source in an image captured by the camera module.
In an embodiment of the present application, because the finger touches the auxiliary light screen by means of tapping so as to form the auxiliary light source, the terminal device may determine a block number indicating where the auxiliary light source falls into the image captured by the camera module, and use the block number indicating where the auxiliary light source falls into the image captured by the camera module as the position of the auxiliary light source in the image captured by the camera module.
In an embodiment of the present application, as shown in
Step 203: The terminal device queries for a code corresponding to the position.
In an embodiment of the present application, the control software of the terminal device may query, according to the block number indicating where the auxiliary light source falls into the image captured by the camera module, a mapping, between blocks and codes, that is stored in a code library for a code corresponding to the block number indicating where the auxiliary light source falls into the image captured by the camera module.
In an embodiment of the present application, the mapping, between the blocks and the codes, that is stored in the code library is shown in table 1.
In an embodiment of the present application, Table 1 shows that the terminal device evenly divides the image captured by the camera module into nine blocks by using the upper left corner of the image captured by the camera module as the original point. A person skilled in the art should understand that table 1 is only an embodiment and a user may also evenly divide the image captured by the camera module into more blocks according to the preference of the user and self-define more codes, so as to enrich operations on the terminal device.
For example, assuming that the block parameters of the block number indicating where the auxiliary light source falls into the image captured by the camera module is “left border=0, right border=image width/3, upper border=0, lower border=image height/3”, the control software of the terminal device may find, according to the block (indicated by “left border=0, right border=image width/3, upper border=0, lower border=image height/3”) that the auxiliary light source falls into the image captured by the camera module, that a code corresponding to the block number indicating where the auxiliary light source falls into the image captured by the camera module is code A from the mapping, between the blocks and the codes, that is stored in the code library shown in Table. 1.
In an embodiment of the present application, assuming that the block parameters of the block number indicating where the auxiliary light source falls into the image captured by the camera module is “left border=image width*⅔, right border=image width, upper border=image height*⅔, lower border=image height”, the control software of the terminal device may find, according to the block (indicated by “left border=image width*⅔, right border=image width, upper border=image height*⅔, lower border=image height”) that the auxiliary light source falls into the image captured by the camera module, that a code corresponding to the block number indicating where the auxiliary light source falls into the image captured by the camera module is I from the mapping, between the blocks and the codes, that is stored in the code library shown in Table. 1.
Step 204: The terminal device acquires, according to the found code, an operation instruction corresponding to the code from a mapping, between codes and operation instructions, which is stored in a code and instruction mapping library, and executes the operation instruction corresponding to the code.
In an embodiment of the present application, with reference to the mapping, between the blocks and the codes, that is stored in the code library shown in table 1, it is assumed that the mapping, between codes and the operation instructions, that is stored in the code and instruction mapping library is shown in table 2.
In an embodiment of the present application, table 3 shows that: the mapping, between the codes and the operation instructions, that is stored in the code and instruction mapping library shown in table 2 is displayed on the nine blocks formed by evenly dividing the image captured by the camera module in table 1.
In the embodiment of the human-computer interaction method described in
The foregoing describes in detail the human-computer interaction method according to another embodiment of the present application.
According to another embodiment of the present application, another human-computer interaction method is further provided.
Step 601: A terminal device captures, by using a camera module, an auxiliary light source formed by a finger sliding on an auxiliary light screen.
In an embodiment of the present application, the specific implementation of the auxiliary light screen is introduced in detail in the preceding embodiments, which is not described again in this embodiment.
In an embodiment of the present application, the terminal device can capture, by using the camera module, an image that includes a highlighted auxiliary light source formed by a finger sliding on the auxiliary light screen, and processes the image so as to acquire an image that only displays the highlighted auxiliary light source formed by the finger sliding on the auxiliary light screen.
Step 602: The terminal device determines a motion track of the auxiliary light source in an image captured by the camera module.
In an embodiment of the present application, the terminal device may perform, by using the control software, continuous recognition on a sequence of images that only display highlighted auxiliary light sources formed after the hand slides the auxiliary light screen, so that the motion track of the auxiliary light source in the image captured by the camera module can be determined
In an embodiment of the present application, because the finger touches the auxiliary light screen by means of sliding so as to form the auxiliary light source, the terminal device may determine a quantity of blocks that the auxiliary light source goes through in the image captured by the camera module and a direction of the auxiliary light source, and use the quantity of the blocks that the auxiliary light source goes through in the image captured by the camera module and the direction of the auxiliary light source as the motion track of the auxiliary light source in the image captured by the camera module.
Step 603: The terminal device queries for a code corresponding to the motion track.
In an embodiment of the present application, a code library of the terminal device may pre-store a mapping among the quantities of the blocks that the auxiliary light source goes through in the image captured by the camera module, the directions of the auxiliary light source, and the codes, as shown in table 4. With reference to the attached table 4 below, assuming that the image captured by the camera module is evenly divided into a plurality of blocks shown in
As shown in
Step 604: The terminal device acquires, according to the found code, an operation instruction corresponding to the code from a mapping, between codes and operation instructions, which is stored in a code and instruction mapping library, and executes the operation instruction corresponding to the code.
In an embodiment of the present application, with reference to the mapping, shown in table 4, among the quantities of the blocks that the auxiliary light source goes through in the image captured by the camera module, the directions of the auxiliary light source, and the codes, the code and instruction mapping library stores a mapping between codes and operation instructions, as shown in table 5.
In an embodiment of the present application, when the control software of the terminal device finds, according to the motion track of the auxiliary light source in the image captured by the camera module, that the motion track of the auxiliary light source in the image captured by the camera module is corresponding to the code a from table 4, the control software of the terminal device may further acquire that the operation instruction is “scroll down content” from table 5. In this case, the terminal device may be informed of executing the operation instruction to scroll down the content.
In the embodiment of the human-computer interaction method described in
The foregoing describes in detail the human-computer interaction method according to some embodiments of the present application.
According to another embodiment of the present application, a terminal device is further provided.
The camera module 801 captures an auxiliary light source formed by a finger gesture on an auxiliary light screen. For example, the camera module 801 captures, using a camera module, an image including the auxiliary light source formed by the finger gesture on the auxiliary light screen located in front of the camera module and then processes the image to acquire an image that only displays the auxiliary light source formed by the finger gesture on the auxiliary light screen.
The determining module 802 determines a position and/or a motion track of the auxiliary light source in an image captured by the camera module 801.
The executing module 803 executes a corresponding operation instruction according to the position and/or the motion track.
In an embodiment of the present application, the camera module 801 captures an image that includes the auxiliary light source formed by the finger gesture on the auxiliary light screen, and processes the image so as to acquire an image that only displays the auxiliary light source formed by the finger gesture on the auxiliary light screen.
In an embodiment of the present application, the determining module 802 determines a block that the auxiliary light source falls into in an image captured by the camera module 801; and/or determines a quantity of blocks that the auxiliary light source goes through in an image captured by the camera module 801 and a direction of the auxiliary light source. The image captured by the camera module 801 is evenly divided into a plurality of blocks (for example, by using an upper left corner as an original point).
As shown in
The query submodule 80321 queries for a code corresponding to the position and/or the motion track.
The acquiring submodule 80322 acquires an operation instruction corresponding to the code from a stored mapping between codes and operation instructions according to the code, and executes the operation instruction corresponding to the code.
In an embodiment of the present application, the query submodule 80321 queries, according to the block number indicating where the auxiliary light source falls into the image captured by the camera module 801, a stored mapping between blocks and codes for a code corresponding to the block number indicating where the auxiliary light source falls into the image captured by the camera module 801.
The query submodule 80321 may further query, according to a quantity of blocks that the auxiliary light source goes through in the image captured by the camera module 801 and a direction of the auxiliary light source, a stored mapping among quantities of blocks, directions, and codes for a code corresponding to the quantity of the blocks and the direction.
In an embodiment of the present application, the operation instruction may be either a computer operation instruction or a television remote control instruction, which is not limited in this embodiment.
According to an embodiment of the present application, the human-computer interaction method shown in
According to another embodiment of the present application, the units of the terminal device shown in
According to another embodiment of the present application, a computer program (including program codes) capable of executing the human-computer interaction method shown in
The terminal device shown in
The foregoing describes in detail the terminal device according to some embodiments of the present application.
According to another embodiment of the present application, a human-computer interaction system is further provided.
The auxiliary light screen 901 is to be touched by a finger so as to form an auxiliary light source.
The camera 902 captures an auxiliary light source formed by a finger gesture on the auxiliary light screen 901.
The terminal device 903 includes: a determining module 9031 and an executing module 9032.
The determining module 9031 determines a position and/or a motion track of the auxiliary light source in an image captured by the camera 902.
The executing module 9032 executes a corresponding operation instruction according to the position and/or the motion track.
In an embodiment of the present application, the camera 902 specifically may capture an image that includes the auxiliary light source formed by the finger gesture on the auxiliary light screen, and process the image so as to acquire an image that only displays the auxiliary light source formed by the finger gesture on the auxiliary light screen.
In an embodiment of the present application, the determining module 9031 of the terminal device 903 specifically may determine a block that the auxiliary light source falls into in an image captured by the camera 902; and/or determine a quantity of blocks that the auxiliary light source goes through in an image captured by the camera 902 and the direction of the auxiliary light source. The image captured by the camera 902 is evenly divided into a plurality of blocks.
In this embodiment, the executing module 9032 of the terminal device 903 includes: a query submodule 90321 and an acquiring submodule 90322.
The query submodule 90321 queries for a code corresponding to the position and/or the motion track.
The acquiring submodule 90322 acquires an operation instruction corresponding to the code from a stored mapping between codes and operation instructions according to the code and executes the operation instruction corresponding to the code.
In an embodiment of the present application, the query submodule 90321 queries, according to the block number that the auxiliary light source falls into the image captured by the camera 902, a stored mapping between blocks and codes for a code corresponding to the block number that the auxiliary light source falls into the image captured by the camera 902.
The query submodule 90321 may further query, according to a quantity of blocks that the auxiliary light source goes through in the image captured by the camera 902 and a direction of the auxiliary light source, a stored mapping among quantities of blocks, directions, and codes for a code corresponding to the quantity of the blocks in the captured image and the direction.
In an embodiment of the present application, the operation instruction may be a computer operation instruction or a television remote control instruction, which is not limited in this embodiment.
In an embodiment of the present application, the camera 902 may be an infrared-light camera. Correspondingly, the auxiliary light screen 901 may be an infrared-light auxiliary light screen. The camera 902 may further be a visible-light camera. Correspondingly, the auxiliary light screen 901 may be a visible-light auxiliary light screen.
According to an embodiment of the present application, the human-computer interaction method shown in
According to another embodiment of the present application, the units of the human-computer interaction system shown in
According to another embodiment of the present application, a computer program (including program codes) capable of executing the human-computer interaction method shown in
The storage media may include: a flash disk, a read-only memory (ROM), a random access memory (RAM), a magnetic disk, or an optical disc.
In the human-computer interaction system described in
To sum up, by using the human-computer interaction method, and the related device and system according to the embodiments of the present application, when the auxiliary light screen is deployed on a desktop, the auxiliary light screen needs to be disposed in parallel with the desktop at a certain distance. Otherwise, a light trace may be formed, thereby affecting recognition. Certainly, the auxiliary light screen may be deployed on a wall surface or a desktop, or be deployed on a facade surface in air, so that a user can touch the auxiliary light screen in the air, thereby implementing a human-computer interaction operation. In addition, a dual-light screen including a visible-light light screen and an infrared-light light screen may be used as the auxiliary light screen, so that when the finger touches the auxiliary light screen, a finger is illuminated by the visible light and human eyes receive a feedback, and in addition, a camera captures a light spot (that is, the auxiliary light source) formed between the infrared-light auxiliary light screen and the finger.
The foregoing describes in detail the human-computer interaction method, and the related device and system provided in the embodiments of the present application. The principles and implementation manners of the present disclosure are described with specific examples to illustrate the above embodiments of the present application. However, the embodiments are not intended to limit the scope of the present disclosure and the scope of the present disclosure is defined by the appended claims. Any modification, equivalent replacement, or improvement made without departing from the spirit and principle of the present disclosure shall fall within the protection scope of the claims.
Number | Date | Country | Kind |
---|---|---|---|
201210388925.9 | Oct 2012 | CN | national |
This patent application is a continuation application of PCT Patent Application No. PCT/CN2013/080324, entitled “HUMAN-COMPUTER INTERACTION METHOD, AND RELATED DEVICE AND SYSTEM” filed on Jul. 29, 2013, which claims priority to Chinese Patent Application No. 201210388925.9, entitled “HUMAN-COMPUTER INTERACTION METHOD, AND RELATED DEVICE AND SYSTEM” filed on Oct. 15, 2012, both of which are incorporated by reference in their entirety.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/CN2013/080324 | Jul 2013 | US |
Child | 14677883 | US |