1. Field of the Invention
The present invention relates to a method of operating an electronic device and more particularly, to an air gesture recognition type electronic device operating method for inputting control signals into the electronic device without direct contact.
2. Description of the Related Art
Following fast development of the modern technology and electronic industry, many different kinds of consumer electronics, such as computer, display device and digital TV have entered into our daily life. Further, an electronic device is generally equipped with an input device, such as mouse, keyboard or key switches for data input. When using a mouse to operate a computer, the mouse must be placed on the desk or a flat surface. Due to the constraint of the mouse, the user must sit in front of the computer when using the computer. The use of the input device increases the installation cost of the computer and limits the operational flexibility of the computer.
In view of the operational inconvenience due to the use of an input device in an electronic system, many advanced electronic devices operable without any attached input device are created. For example, touch panel is an electronic visual output that can detect the presence of location of the touch or contact to the display area by a finger, hand, pen or other passive objects. A touch panel enables one to interact with what is displayed directly on the hand, and lets one do so without requiring any intermediate device. A touch panel can be attached to computers or any of a variety of other electronic devices. Nowadays, touch panel has been intensively used in communication equipment, household appliances, entertainment appliances, IA products, medical instruments and etc.
However, when operating a touch panel, the user needs to touch the screen of the touch panel directly. Thus, the user must stand closer to the electronic device so that the user's finger or hand can touch the screen of the touch panel of the electronic device to input a command. This operating method still brings inconvenience. In order to eliminate this drawback, vision-based human computers are created. These vision-based human computers detect hand gesture, body language interpretation and/or facial expression interpretation. Researches continue to improve point gesture recognition technology. Air gesture recognition technology has been effectively used in TV, computer or projector as an effective human machine interface system.
At present, the application of air gesture recognition technology requires a camera to pick up the gesture of the user's hand or body for analyzing the moving direction of the user's hand or body and running a corresponding input procedure subject to the result of the computation. The use of the camera requires an extra cost.
Therefore, it is desirable to provide an air gesture recognition type electronic device that is operable without any attached input device or camera.
The present invention has been accomplished under the circumstances in view. It is one object of the present invention to provide an air gesture recognition type electronic device operating method, which enables the user to input control signals into the electronic device within a predetermined range without direct contact, enhancing operational flexibility. It is another object of the present invention to provide an air gesture recognition type electronic device operating method, which achieves air gesture recognition without any camera, saving the hardware cost. It is still another object of the present invention to provide an air gesture recognition type electronic device operating method, which saves power consumption when the electronic device is not operated.
To achieve these and other objects of the present invention, an air gesture recognition type electronic device operating method, which enables a user to operate an electronic device that has multiple sensors in each of multiple peripheral sides thereof by: approaching an object to the sensors to produce sensing signals and determining whether or not the object has been continuously sensed, and then determining whether or not the moving direction and moving speed of the object match a respective predetermined value, and then coupling and computing all received sensing signals to produce an operating parameter for running an air gesture application procedure. Thus, a user can operate the electronic device without direct contact or the use of any camera or input media, saving the hardware cost and enhancing the operational flexibility.
In an alternate form of the present invention, the air gesture recognition type electronic device operating method includes the step of preparing an electronic device having multiple sensors in each of multiple peripheral sides thereof and then providing multiple objects for approaching the sensors of the electronic device to produce sensing signals, the step of determining whether or not at least one object is approaching, the step of determining whether or not multiple sensing signals have been produced and whether or not these sensing signals are continuous sensing signals, the step of determining whether or not the moving directions of the sensed objects are different and whether or not the moving direction/speed of each sensed object matches a predetermined value, the step of coupling and computing all received sensing signals to produce an operating parameter, and the step of running an air gesture application procedure subject to the produced operating parameter.
Further, when one object is sensed by one sensor, the electronic device is switched from a power-saving mode to an operating mode. This wakeup mode saves power consumption when the electronic device is not operated.
Referring to
According to the present preferred embodiment, the electronic device 1 is a display device. The display device 1 has four peripheral sides, namely, the first peripheral side 11, the second peripheral side 12, the third peripheral side 13 and the fourth peripheral side 14 disposed around a rectangular screen 10 thereof, and a plurality of sensor means mounted in each of the four peripheral sides 11-14. The first peripheral side 11 and the third peripheral side 13 are disposed opposite to each other. The second peripheral side 12 and the fourth peripheral side 14 connected between the first peripheral side 11 and the third peripheral side 13 at two opposite lateral sides. The sensors can be capacitive sensor or infrared sensors. Exemplars of the electrical switching means can be seen in U.S. Pat. Nos. 7,498,749; 7,443,101; 7,336,037.
The electronic device operating method uses an object 3 for input control. Further, as an example of the present invention, a first sensor 21, a second sensor 22 and a third sensor 23 are installed in the first peripheral side 11 of the display device 1 and electrically connected to a control module 20 at a circuit board 2 inside the display device 1. The air gesture recognition type electronic device operating method in accordance with the first embodiment of the present invention includes the steps of:
When one object 3, for example, the user's fingers enter the set sensing range of the first sensor 21 in the first peripheral side 11 of the display device 1, for example, the range X within 10-25 cm, the first sensor 21 senses the presence of the object 3 and produces a sensing signal. When the object 3 moves away from the set sensing range of the first sensor 21 into the set sensing range of the second sensor 22 in the first peripheral side 11 of the display device 1, for example, the range X within 10-25 cm, the second sensor 22 senses the presence of the object 3 and produces a sensing signal. Subject to the sensing signal from the first sensor 21 and the sensing signal from the second sensor 22, a continuous sensing is confirmed by the control module 20. Further, the control module 20 stores the sensing signals received from the first sensor 21 and the second sensor 22 in a built-in memory or an external memory that is electrically connected to the control module 20.
Thereafter, the control module 20 determines whether or not the moving direction and speed of the sensed object 3 matched a respective predetermined value that is stored in the built-in memory or the external memory that is electrically connected to the control module 20. When matched, the control module 20 couples and analyzes all the received sensing signals to produce an operating parameter. The sensing signal produced by each sensor comprises the data of, but not limited to, distance, direction and speed. The computation is made subject to the formula of:
Ag=S1{f(d),f(t)}·S2{f(d),f(t)} . . . Sy{f(d),f(t)}
where:
Ag (air gesture operation)=the operating parameter;
S=sensor;
S1=the first sensor;
S2=the second sensor;
Sy=the yth sensor;
f(d)=the distance between the sensed object 3 and the sensor sensing the object 3;
f(t)=the moving time from one sensor to a next sensor.
Calculation of the moving time is made by: defining the time of the first contact to be the first time point t1 and the time of the last contact to be the second time point t2, and then obtaining the moving time by the formula of t2−t1. Thus, the control module 20 can couple and analyze the sensing signals received from the sensors to produce an operating parameter. According to the present preferred embodiment, the operating parameter comprises the data of, but not limited to, the moving direction of the sensed object 3, the distance between the sensed object 3 and the respective sensor, and the moving speed of the sensed object 3. Subject to the operating parameter thus produced, an air gesture application program is performed.
The arrangement of the first sensor 21, second sensor 22 and third sensor 23 in the first peripheral side 11 of the display device 1 is simply an example of the present invention. However, this example is simply for the purpose of illustration only but not for use as a limitation. According to the aforesaid operation flow, the control module 20 determines whether or not the object 3 has been continuously sensed by the first sensor 21, the second sensor 22 and the third sensor 23. When the object 3 is continuously sensed by the first sensor 21, the second sensor 22 and the third sensor 23, it is judged to be a continuous sensing status. Thereafter, the control module 20 determines the moving direction of the object 3 subject to the sequence of the sensing signals received. Subject to the calculation formula Ag=S1{f(d), f(t)}·S2{f(d), f(t)} . . . Sy{f(d), f(t)}, it is known that the object 3 moves relative to the first peripheral side 11 from the left toward the right. Thereafter, the distance between the object 3 and the first sensor 21 and the distance between the object 3 and the second sensor 22 are determined subject to f(d). Thereafter, subject to f(t), the moving speed of the object 3 is determined to be in conformity with the set value or not. For example, if the time period from the first time point t1 to the second time point t2 is 5-6 seconds and the distances between the object 3 and the first sensor 21, second sensor 22 and third sensor 33 are equal and all to be 5 cm, it is determined to be an operation for volume control. On the other hand, when the control module 20 received sensing signals from the first sensor 21, the second sensor 22 and the third sensor 23 within a predetermined time period, the time period from the first time point t1 to the second time point t2 during movement of the object 3 is shorter than one second, and the distances between the object 3 and the first sensor 21, second sensor 22 and third sensor 33 are equal and all to be 5 cm, thus it is determined to be an operation for turning to the next page. However, it is to be understood that the above explanation is simply an example of the present invention and shall not be considered to be limitations of the invention.
According to the present invention, the electronic device 1 has stored therein multiple operating parameters, such as the value defined for next page operation, the parameter for volume control, or the parameter for picture rotation. Further, the invention uses the control module 20 to receive sensing signals from the sensors, and uses a formula to calculate the content of the sensing signals. If the content of one sensing signal obtained through calculation matches one pre-set operating parameter, the control module 20 executes the corresponding application program and operating software procedure. Thus, the user can input control signals into the electronic device 1 within a predetermined range without direct contact, enhancing operational flexibility.
According to the embodiment shown in
Referring to
When one object 3 enters the set sensing range of the sensors, the procedure as explained in the aforesaid first embodiment is performed. This second embodiment will detect whether or not a second one of the objects 3 enters the set sensing range of the sensors after detection of the approaching of a first one of the objects 3. By means of the control module 20, it determines whether or not the moving directions of the sensed objects 3 are different. If the moving directions of the sensed objects 3 are different, it is indicated that multiple objects 3 are simultaneously appeared to operate the electronic device 1.
For example, when a first object 31 and a second object 32 enter the sensing range of the first sensor 21, second sensor 22, third sensor 23, seventh sensor 27 and eighth sensor 28, these sensors 21;22;23;27;28 respectively provide a respective sensing signal to the control module 20. At this time, the control module 20 makes judgment. When the first object 31 enters the sensing range of the first sensor 21, second sensor 22 and third sensor 23 and then moves from the set sensing range of the first sensor 21 into the set sensing range of the third sensor 23, the sensing is judged to be a continuous sensing, and the movement of the first object 31 is judged by the control module 20 to be from the left toward the right. Relatively, when the second object 32 enters the sensing range of the seventh sensor 27 and the eighth sensor 28 and moves from the set sensing range of the seventh sensor 27 into the set sensing range of the eighth sensor 28, the sensing is judged to be a continuous sensing, and the movement of the second object 32 is judged by the control module 20 to be from the right toward the left. Subject to the relative relationship of movement of the sensed objects 3, the control module 20 judges that multiple objects 3 are in movement, and then determines whether or not the moving direction and speed of each sensed object 3 match respective predetermined values, and then couples and analyzes all the received sensing signals to produce an operating parameter, for example, zoom in, and then runs the zoom-in application procedure. Thus, multiple objects 3 are applicable for controlling the operation of the electronic device 1.
Further, when one object 3 enters a predetermined range relative to the electronic device 1, the sensors provide a respective sensing signal to the control module 20, causing the control module 20 to start up power supply for the other modules of the electronic device 1, putting the other modules of the electronic device 1 into standby mode. Thus, power consumption is minimized when the electronic device 1 is not operated.
In conclusion, the invention provides an air gesture recognition type electronic device operating method, which has advantages and features as follows:
Although particular embodiments of the invention have been described in detail for purposes of illustration, various modifications and enhancements may be made without departing from the spirit and scope of the invention. Accordingly, the invention is not to be limited except as by the appended claims.