This application claims priority under 35 U.S.C. §119(a) to Korean Patent Application No. 10-2014-0048216, which was filed in the Korean Intellectual Property Office on Apr. 22, 2014, the content of which is incorporated herein by reference.
1. Field of the Invention
The present invention relates generally to a method and an apparatus for recognizing a gesture in an electronic device.
2. Description of the Prior Art
An electronic device may recognize a gesture by detecting intensity of light (e.g., an infrared ray) reflected by an object through an infrared sensor and determining a distance and an operation of the gesture. Through a hovering technology, a Touch Screen Panel (TSP) of an electronic device may detect the motion of an object (for example, a finger of a user or a stylus pen) on a screen and recognize a hovering gesture. Specifically, the user inputs a gesture in a certain direction based on locations of various sensors (for example, a gesture recognition sensor) to perform a specific instruction corresponding to the gesture. For example, when the user inputs a circular pattern, the electronic device may perform an operation corresponding to the circular pattern by accurately inputting the circular pattern with reference to a sensor.
However, when the user desires to input a specific pattern to perform a specific instruction while the electronic device is positioned above the user's hand, (for example, when the electronic device is located on a desk), it may be uncomfortable for the user to input the specific pattern while raising their arm and bending their wrist.
The present invention has been made to address at least the problems and/or disadvantages described above and to provide at least the advantages described below.
Accordingly, an aspect of the present invention is to provide a method and an apparatus for effectively recognizing an operation intended by a user.
In accordance with an aspect of the present invention, a gesture recognition method is provided, which includes extracting one or more vector values from an input gesture; generating a pattern of a vector based on the extracted one or more vector values; comparing the generated pattern to one or more patterns of stored vectors; and determining a type of the input gesture based on the comparing.
In accordance with another aspect of the present invention, a gesture recognition method is provided, which includes detecting two or more multiple proximity inputs; extracting one or more first vector values from the detected two or more multiple proximity inputs; detecting a motion of the detected two or more multiple proximate inputs; extracting one or more second vector values for the motion; analyzing loci of the first vector values and the second vector values; determining whether a pinch gesture is generated, based on the analyzed loci; and if the pinch gesture is generated, performing a function corresponding to the pinch gesture.
In accordance with another aspect of the present invention, a gesture recognition apparatus is provided, which includes a gesture recognition device; and a controller that detects an input of a gesture through the gesture recognition device, extracts one or more vector values from the detected gesture, generates a pattern of a vector based on the extracted one or more vector values, compares the generated pattern of the vector to one or more patterns of stored vectors, and determines a type of the gesture, based on the comparison.
In accordance with another aspect of the present invention, a gesture recognition apparatus is provided, which includes a gesture recognition device; and a controller that detects two or more multiple proximity inputs through the gesture recognition device, extracts one or more first vector values for the multiple proximity inputs, detects a motion generated by the multiple proximate inputs, extracts one or more second vector values from the motion, analyzes loci of the first vector values and the second vector values, determines whether a pinch gesture is generated, based on the analyzed loci, and performs a function corresponding to the pinch gesture, when the pinch gesture is generated.
The above and other aspects, features, and advantages of certain embodiments of the present invention will be more apparent from the following detailed description in conjunction with the accompanying drawings, in which:
Various embodiments of the present invention will now be described more fully in conjunction with the accompanying drawings. The present invention may have various embodiments, and modifications and changes may be made therein. Therefore, the present invention will be described in detail with reference to particular embodiments shown in the accompanying drawings. However, it should be understood that there is no intent to limit various embodiments of the present invention to the particular embodiments disclosed, but the present invention should be construed to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the various embodiments of invention. In connection with descriptions of the drawings, similar components are designated by the same reference numeral.
It will be understood that the expressions “comprises” and “may comprise” are used to indicate the presence of a disclosed function, operation, component, etc., but do not preclude the presence of one or more functions, operations, components, etc. It will be further understood that the terms “comprises” and/or “has” when used in this specification, specify the presence of stated feature, number, step, operation, component, element, or a combination thereof, but do not preclude the presence or addition of one or more other features, numbers, steps, operations, components, elements, or combinations thereof. In the present disclosure, the expression “and/or” is taken as specific disclosure of each and any combination of enumerated things. For example, A and/or B is to be taken as specific disclosure of each of A, B, and A and B.
As used herein, terms such as “first,” “second,” etc., are used to describe various components; however, it is obvious that the components should not be defined by these terms. For example, the terms do not restrict the order and/or importance of the corresponding components. The terms are used only for distinguishing one component from another component. For example, a first component may be referred to as a second component and likewise, a second component may also be referred to as a first component, without departing from the teaching of the inventive concept.
It will be understood that when an element or layer is referred to as being “on”, “connected to” or “coupled to” another element or layer, it can be directly on, connected or coupled to the other element or layer or intervening elements or layers may be present. In contrast, when an element is referred to as being “directly on,” “directly connected to” or “directly coupled to” another element or layer, there are no intervening elements or layers present.
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the disclosure. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise.
Unless otherwise defined herein, all terms including technical or scientific terms used herein have the same meanings as commonly understood by those skilled in the art to which the present invention belongs. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the specification and relevant art and should not be interpreted in an idealized or overly formal sense unless expressly so defined herein.
An electronic device according to various embodiments of the present invention may be an apparatus including a gesture recognition function, and may also include devices having an operation support function. For example, an electronic device may include a smartphone, a table Personal Computer (PC), a mobile phone, a video phone, an electronic book (e-book) reader, a desktop PC, a laptop PC, a netbook computer, a Personal Digital Assistant (PDA), a Portable Multimedia Player (PMP), an MP3 player, a mobile medical appliance, a camera, and a wearable device (e.g., head-mounted device (HMD) such as an electronic glasses, electronic clothing, an electronic bracelet, an electronic necklace, an electronic appcessory, an electronic tattoo, a smartwatch, etc.).
Additionally, an electronic device may be a smart home appliance having an operation support function, such as a television, a Digital Video Disk (DVD) player, an audio player, a refrigerator, an air-conditioner, a vacuum cleaner, an electric oven, a microwave oven, a laundry machine, an air cleaner, a set-top box, a TV box (e.g., Samsung HomeSync®, apple TV®, and google TV®), a game console, an electronic dictionary, an electronic key, a camcorder, an electronic frame, etc.
Other examples of an electronic device include a medical device (e.g., a Magnetic Resonance Angiography (MRA) device, a Magnetic Resonance Imaging (MRI) device, a Computed Tomography (CT) device), a Navigation device, a Global Positioning System (GPS) receiver, an Event Data Recorder (EDR), a Flight Data Recorder (FDR), a car infotainment device, a maritime electronic device (e.g., a maritime navigation device and a gyro compass), an aviation electronic device (avionics), a security device, a vehicle head unit, an industrial or home robot, an Automatic Teller Machine (ATM) of a financial institution, a Point Of Sales (POS) device, etc.
Other examples of an electronic device may include furniture and buildings/structures having a communication function, an electronic board, an electronic signature receiving device, a projector, and a metering device (e.g., water, electric, gas, and/or electric wave metering devices).
Additionally, the electronic device may be a flexible device.
Further, an electronic device may be any combination of the aforementioned devices.
It is obvious to those skilled in the art that an electronic device is not limited to the aforementioned examples.
Herein, the term “user” may denote a person or a device (e.g., an artificial intelligent electronic device) using an electronic device.
Referring to
The bus 110 connects the aforementioned components to each other and may be a circuit for exchanging signals (e.g., control messages) among the components.
For example, the processor 120 receives a command from any of the aforementioned components (e.g., the memory 130, the input/output interface 140, the display 150, the communication interface 160, and the gesture recognition module 170) through the bus 110, interprets the command, and executes operation or data processing according to the decrypted command.
The memory 130 stores the command or data received from the processor 120 or other components or generated by the processor 120 or other components. The memory 130 stores program modules including a kernel 131, middleware 132, an Application Programming Interface (API) 133, applications 134, etc. Herein, each programming module may be implemented as software, firmware, hardware, and any combination thereof.
The kernel 131 controls or manages the system resources (e.g. bus 110, processor 120, and memory 130) for use in executing an operation or a function implemented with the middleware 132, the API 133, and/or the application 134. The kernel 131 also provides an interface for the middleware 132, API 133, and/or application 134 to access the components of the electronic device 101 to control or manage.
The middleware 132 may work as a relay of data communicated between the API 133 or application 134 and the kernel 131. The middle 132 may also execute control of the task requests from the applications 134 in such a way of assigning priority for use of the system resources (e.g., bus 110, processor 120, and memory 130) of the electronic device 101 to at least one of the applications 134.
The API 133 is an interface for the applications 134 to control the function provided by the kernel 131 or the middleware 132, and may include at least one interface or function (e.g., a command) for file control, window control, image control, or text control.
For example, the applications 134 may include a Short Messaging Service/Multimedia Messaging Service (SMS/MMS) application, an email application, a calendar application, an alarm application, a health care application (e.g., an application for measuring motion or a blood sugar level), and an environmental information application (e.g., atmospheric pressure, humidity, and/or temperature applications). Additionally or alternatively, the application 134 may be an application related to information exchange between the electronic device 101 and other external electronic devices (e.g., an electronic device 104 or a server 106). Examples of the information exchange application may include a notification relay application for relaying specific information to the external electronic device and a device management application for managing the external electronic device.
For example, the notification relay application may be provided with a function of relaying the alarm information generated by the other applications (e.g., an SMS/MMS application, an email application, a health care application, and an environmental information application) of the electronic device 101 to the electronic device 104.
Additionally or alternatively, the notification relay application may provide the user with the notification information received from the electronic device 104. The electronic device application may manage (e.g., install, delete, or update) a function of the electronic device 104 (e.g. turn-on/off of the electronic device 104 (or a component thereof) or adjust the brightness (or resolution) of another display), which communicates with the electronic device 101, or manage a service (e.g., communication or messaging service) provided by the electronic device 104.
For example, the applications 134 may include an application designated according to a property (e.g., a type) of an external electronic device (e.g., the electronic device 104).
If the electronic device 104 is an MP3 player, the application 134 may include a music playback application. Similarly, if the electronic device 104 is a mobile medical appliance, the application 134 may include a heal care application.
As another example, the application 134 may include at least one of an application designated to the electronic device 101 or an application received from an external electronic device (e.g., the server 106 or the electronic device 104).
The input/output interface 140 delivers a command or data input by a user through an input/output device (e.g., a sensor, a keyboard, and/or a touch screen) to the processor 120, memory 130, communication interface 160, and/or gesture recognition module 170 through the bus 110. For example, the input/output interface 140 provides the processor 120 with data corresponding to a touch made by a user on the touch screen.
The input/output interface 140 may output a command or data, which is received from the processor 120, memory 130, communication interfaced 160, and/or the gesture recognition module 170 through the bus 110, through an input/output device (e.g., a speaker and/or a display). For example, the input/out interface 140 may output the voice data processed by the processor 120 to the user through a speaker.
The display 150 displays various information (e.g., multimedia data and text data) to the user.
The communication interface 160 establishes a communication connection of the electronic device 101 with an external device (e.g., the electronic device 104 and/or the server 106). For example, the communication interface 160 connects to a network 162 through a wireless or wired link for communication with the electronic device 104. Examples of the wireless communication technology may include wireless fidelity (Wi-Fi), Bluetooth (BT), Near Field Communication (NFC), Global Positioning System (GPS), and cellular communication technology (e.g., Long Term Evolution (LTE), LTE-Advanced (LTE-A), Code Division Multiple Access (CDMA), Wideband CDMA (WCDMA), Universal Mobile Telecommunication System (UMTS), Wireless-Broadband (WiBro), and General System for Mobile communications (GSM)). Examples of the wired communication technology may include Universal Serial Bus (USB), High Definition Multimedia Interface (HDMI), Recommended Standard 232 (RS-232), and Plain Old Telephone Service (POTS).
The network 162 may be a telecommunication network including at least one of a computer network, the Internet, the Internet of Things, and/or a telephone network.
A communication protocol (e.g., a transport layer protocol, a data link layer protocol, and/or a physical layer protocol) between the electronic device 101 and an external device may be supported by at least one of the applications 134, the API 133, the middleware 132, the kernel 131, and the communication interface 160.
The server 106 may support driving of the electronic device 101 by performing at least one of operations (or functions) implemented by the electronic device 101. For example, the server 106 may include a gesture recognition server module 108 that may support a gesture recognition module 170 realized in the electronic device 101. For example, the gesture recognition server module 108 may include at least one element of the gesture recognition module 170 to perform handle at least one of the operations performed by the gesture recognition module 170.
The gesture recognition module 170 may process information acquired from other elements (e.g., the processor 120, the memory 130, the input/output interface 140, and the communication interface 160), and may provide the processed information to a user through various methods. For example, the gesture recognition module 170 may control at least some functions of the electronic device 101 such that the electronic device 101 interworks with another electronic device (e.g., the electronic device 104 or the server 106) using the processor 120 or independently from the processor 120.
Referring to
The controller 210 generates a vector pattern for a generated gesture, and compares a vector pattern for the gesture with a stored vector pattern to determine the type of the gesture and perform a function for the gesture.
Accordingly, in a module operation of the controller 210, the vector value extraction module 211 may extract a vector value for a generated gesture. The vector value storage module 212 may store a vector value extracted through the vector value extraction module 211. The vector pattern comparison module 213 may compare a vector pattern for the generated gesture with a stored vector pattern in the memory 130. The gesture type determination module 214 may determine the type of the gesture through the comparison operation.
The memory 130 may store a vector pattern of a comparison target to determine the type of the gesture. The memory 130 may store the vector pattern for the detected gesture.
The input/output interface 140 may include, for example, an input unit (for example, an input/output interface 140) such as a touch panel or a key button panel. The touch panel may include a touch screen such that the touch screen is integral with the display 150, and may detect inputs touched on the display 150.
For example, the controller 210 may include the gesture recognition module 170, and may determine a gesture of the user through a gesture sensor such as an IR sensor, hovering through TSP, and/or an image sensor. The gesture recognition module 170 may determine an object (for example, a finger of the user or a stylus pen) that is a motion target, when a gesture of the user is detected, and may detect a motion of the object. Thereafter, when a motion of the object is detected, the gesture recognition module 170 may detect a motion on the z-axis corresponding to depth, and motions on the x and y axes, when the motion of the object is detected. That is, when a motion is detected while the gesture is generated, a vector value for the motion in the form of (x, y, z) axes may be extracted.
In accordance with an embodiment of the present invention, the vector value may include at least one phase value. Motion information in which the extracted vector values in the form of (x, y, z) are accumulated may be compared with a predefined motion (stored in the 3-dimensional form of (x, y, z). Specifically, a distribution of vector values for a predefined motion may be compared with a distribution for vector values for a gesture.
Further, a progress direction in which vector values for a predefined motion are generated may be compared with a progress direction in which vector values for a gesture are generated. Through the comparison operations, the type of gesture of the user may be determined, and a function corresponding to the type of gesture may be performed.
When a gesture is detected while the electronic device is horizontal, vertical, or inclined, gestures in various states may be recognized as the same gesture by comparing the distribution of the vector values for the gesture and the progress direction in which the vector values are generated.
The controller 210 may control to extract at least one vector value for a detected gesture, to generate a pattern of a vector based on the vector value, to compare the pattern of the at least one vector stored in advance with the pattern of the generated vector, and to determine the type of the gesture based on the comparison process.
When a pattern of the vector is generated, the controller 210 may classify the pattern of the vector through at least one of variance, deviation, and average of the vector values.
The controller 210 may compare the distribution of the vector values and the progress direction in which the vector values are generated.
When a new gesture is detected, the controller 210 may extract at least one vector value for the new gesture.
As the new gesture is detected, the controller 210 may accumulate and store vector values for the new gestures.
The controller 210 may detect the gesture through at least one of an image sensor, an IR sensor, and a touch panel.
The controller 210 may also control to, when at least two multiple proximity inputs are detected, extract at least one first vector value for the multiple proximity inputs, to extract at least one second vector value for the motion, if a motion is generated while the multiple proximity inputs are detected, to analyze motion loci of the first vector value and the second vector values, to determine generation of a pinch gesture (pinch in or pinch out), and to perform a function corresponding to the pinch gesture.
The controller 210 may perform at least one of functions of enlarging and reducing an image and adjusting a selection range of materials arranged on the Z-axis, when the pinch gesture is generated.
Referring to
Referring to
Referring to
As illustrated in
For example, if the input gesture is viewed only through the x and y-axes, a sensor may recognize the gesture as being close to a circle as indicated by reference numeral 309 of
If a finger of the user applies pressure to a display, as indicated by reference numeral 313 of
According to another embodiment of the present invention, z-axis information may be sensed through input by a stylus pen (S-Pen).
Referring to
By sensing x, y, and z information for a gesture, a unique property of a pattern may not be impaired, even if the location of the electronic device changes (for example, the electronic device is horizontal or inclined) or the same gesture is input several times.
For example, when a star-shaped pattern is input, it may be recognized that the same star-shaped pattern is input, even if the starting point is different, as illustrated in
As another example, when a circular pattern is input, the electronic device may recognize the same circular pattern input, regardless of whether the user draws a circle once or three times, as illustrated in
Referring to
In step 403, the electronic device extracts a vector value for the gesture. In step 405, the electronic device stores the extracted vector value.
In step 407, the electronic device determines whether a new gesture is detected. If a new gesture is generated, the operation returns to steps 403 and 405 to repeat extracting and storing a vector value for the new gesture.
However, if a new gesture is not generated in step 407, the electronic device classifies the stored vector values. For example, classifying the vector values may be performed through the variance, deviation, and average of the vector values, and a pattern of the vector for the gesture may be formed through the classifying operation.
In step 411, the electronic device compares a pattern of a stored vector with a pattern of a vector for the detected gesture. For example, in the comparison operation, a distribution of the stored vector pattern stored may be compared with a distribution of a vector pattern for the detected gesture. Further, a sequence (e.g., a progress direction) in which the stored vector pattern is generated may be compared with the sequence in which a vector pattern for the gesture is generated.
In step 413, the electronic device determines a type of the gesture, based on the comparison of the vector to the stored vectors.
In step 415, the electronic device performs a function corresponding to the type of the gesture.
The electronic device determines whether the function will end in step 417. For example, if an end instruction is generated, the electronic device detects the end instruction and ends the function. However, if an end instruction is not generated, the operation returns to step 401, and the electronic device detects if a gesture is detected.
Referring to
Thereafter, a new gesture, i.e., a second gesture 2 is generated, as indicated by reference numeral 501, a group of vectors for the second gesture may be generated in another range as indicated by “2” of reference numeral 503. Thereafter, as a third gesture 3 is generated, as indicated by reference numeral 501, a group of vectors for the third gesture may be generated as indicated by “3” of reference numeral 503, as a fourth gesture 4 is generated, as indicated by reference numeral 501, a group of vectors for a fourth gesture as indicated by reference 501 may be generated as indicated by “4” of reference numeral 503, and as a fifth gesture 5 is generated, as indicated by reference numeral 501, a group of vectors for a fifth gesture as indicated by reference numeral 501 may be generated as indicated by “5” of reference numeral 503. Further, as a sixth gesture 6 is generated, as indicated by reference numeral 501, a group of vectors for the sixth gesture may be generated as indicated by “6” of reference numeral 503.
When the group of vectors for the sixth gesture returns to a group of vectors of “1”, a pattern of the vectors for the gesture may be generated. The group of vectors for a pentagonal pattern may be defined, such that a variance thereof is a predetermined level or higher, and a progress direction of the group of the generated vectors may be constant. That is, the vector values for the gesture may be distributed, as indicated by reference numeral 503, and when a progress direction in which the vector values are generated is 1, 2, 3, 4, 5, and 6 (1), the gesture may be determined to be a pentagonal pattern.
Referring to
When the group of vectors for the sixth gesture returns to the group of vectors of “1”, a pattern of the vectors for the gesture may be generated. The group of vectors for a star-shaped pattern may be defined, such that a variance thereof is a predetermined level or higher, and a progress direction of the group of the generated vectors may be constant. That is, the vector values for the gesture may be distributed, as indicated by reference numeral 507, and when a progress direction in which the vector values are generated is 1, 4, 2, 5, 3, and 6 (1), the gesture may be determined to be a star-shaped pattern. Although the distribution of the vector values for a pentagonal pattern, as indicated by reference numeral 503 of
Referring to
In step 603, the electronic device extracts vector values for the gesture from an image sensor. Specifically, the vector values for the motion may be extracted by determining presence of a motion of an object (for example, a finger of the user or a stylus pen) detected by an image sensor.
In step 605, the electronic device stores the extracted vector values. In step 607, the electronic device determines whether a new gesture is detected. If a new gesture is generated, the operation returns to steps 603 and 605 to repeat the extracting and storing of a vector value for the new gesture from the image sensor.
However, if a new gesture is not generated in step 607, the electronic device classifies the stored vector values in step 609. For example, classifying the vector values may be performed through the variance, deviation, and average of the vector values, and a pattern of the vector for the gesture may be formed through the classifying operation.
In step 611, the electronic device compares a pattern of a stored vector stored with a pattern of a vector for the detected gesture. For example, in the comparison operation, a distribution of the stored vector pattern may be compared with a distribution of a vector pattern for the detected gesture, and a sequence (e.g., progress direction) in which the stored vector pattern is generated may be compared with a sequence in which a vector pattern for the detected gesture is generated. In step 613, the electronic device determines a type of the detected gesture, based on the comparison.
In step 615, the electronic device performs a function corresponding to the type of the gesture.
In step 617, the electronic device determines whether the function is completely performed. For example, if an end instruction is generated, the electronic device detects the end instruction and ends the function. However, if an end instruction is not generated in step 617, the operation returns to step 601.
Referring to
Referring to
Additionally, as the sensing period becomes shorter, the vector information has attributes corresponding to an acceleration of motion.
Specifically, as illustrated in
Referring to
In step 807, the electronic device determines whether a new proximity input is detected. If a new proximity input is generated, the operation returns to steps 803 and 805 to repeat the extracting and storing of a vector value for the new proximity input.
However, if a new proximity input is not generated in step 807, the electronic device classifies the stored vector values in step 809. For example, classifying the vector values may be performed through the variance, deviation, and average of the vector values, and a pattern of the vector for the proximity input may be formed through the classifying operation.
In step 811, the electronic device compares a pattern of a stored vector with a pattern of a vector for the detected proximity input. For example, in the comparison operation, a distribution of the stored vector pattern may be compared with a distribution of a vector pattern for the detected proximity input. Further, a sequence (e.g., a progress direction) in which the stored vector pattern is generated may be compared with a sequence in which the vector pattern for the detected proximity input is generated.
In step 813, the electronic device determines a type of the proximity input, based on the comparison, and in step 815, performs a function corresponding to the proximity input, based on the determined type.
In step 817, the electronic device determines whether the function is completely performed. For example, if an end instruction is generated in step 817, the electronic device detects the end instruction and ends the function. However, if an end instruction is not generated in step 817, the operation returns to step 801.
Referring to
Although the figure 8-shaped pattern is two attached circles as illustrated in
Further, the figure 8-shaped pattern may be identified from a clockwise motion of a unit vector and a counterclockwise motion as indicated by reference numeral 905 of
Referring to
If a touch gesture is generated, the electronic device extracts a coordinate value (x,y) for the gesture from the touch panel in step 1003. In step 1005, the electronic device extracts a depth value (z) for the gesture, e.g., from an IR sensor. For example, the electronic device may detect a distance according to a degree by which an infrared ray projected from a light emitting unit is reflected by a finger and is introduced into a light receiving unit through the IR sensor, and may determine a motion of an object (e.g., a finger of the user or a stylus pen) using an intensity of the reflected light.
In step 1007, the electronic device stores the extracted vector values.
In step 1009, the electronic device determines whether a new gesture is detected. If a new gesture is detected in step 1009, the operation returns to steps 1003, 1005, and 1007 to extract a coordinate value and a depth for the new gesture and store the extracted values.
However, if a new gesture is not generated in step 1009, the electronic device classifies the stored vector values in step 1011. For example, classifying the vector values may be performed through the variance, deviation, and average of the vector values, and a pattern of the vector for the gesture may be formed through the classifying operation.
In step 1013, the electronic device compares a pattern of a stored vector with a pattern of a vector for the detected gesture. For example, in the comparison operation, a distribution of the stored vector pattern may be compared with a distribution of a vector pattern for the detected gesture, and a sequence (e.g., progress direction) in which the stored vector pattern was generated may be compared with a sequence (progress direction) in which the vector pattern for the detected gesture is generated.
In step 1015, the electronic device determines a type of the detected gesture, based on the comparison.
In step 1017, the electronic device performs a function corresponding to the gesture, based on the determined type.
In step 1019, the electronic device determines whether the function is completely performed. For example, if an end instruction is generated, the electronic device detects the end instruction and ends the function. However, if an end instruction is not generated in step 1019, the operation returns to step 1001.
Referring to
Further, the electronic device may identify that the vector values for the four-sided pattern are partially distributed, as illustrated in
Referring to
In step 1205, the electronic device determines whether a motion is generated while the multiple proximity inputs are generated. If a motion is generated in step 1205, the electronic device extracts and stores a second vector value for the motion in step 1207.
In step 1209, the electronic device analyzes motion loci of the first vector value and the second vector value.
In step 1211, based on the analysis of the motion loci, the electronic device determines if the gesture is a pinch gesture. For example, if a motion in the reverse direction is detected, while two multiple proximity inputs are generated between threshold values, the electronic device determines the multiple proximity inputs as a gesture.
When the pinch gesture is determined in step 1211, the electronic device performs a function for the pinch gesture in step 1213.
However, when the pinch gesture is not determined in step 1211, i.e., another gesture is input, the electronic device performs a corresponding function in step 1215.
In step 1217, the electronic device determines whether the function is completely performed. For example, if an end instruction is generated in step 1217, the electronic device detects the end instruction and ends the function. However, if an end instruction is not generated in step 1217, the operation returns to step 1201.
Referring to
For example,
0<threshold 1<α<threshold 2 (1)
The gesture may be recognized as the same pinch as an X-Y axis based pinch gesture according to vector information of the gesture, and for example, a function of enlarging an image displayed on a screen may be performed.
Referring to
Referring to
Specifically, the wearable device may determine a gesture of the user through a gesture sensor, such as an IR sensor, hovering through a TSP, or an image sensor. The wearable device determines an object (for example, a finger of the user or a stylus pen) that is a motion target, when a gesture of the user is detected, and detects a motion of the object. When motion of the object is detected, the wearable device, e.g., the gesture recognition module 170 therein, detects motion on the z-axis corresponding to depth, and motions on the x and y axes. That is, when motion is detected while the gesture occurs, a vector value for the motion in the form of (x, y, z) may be extracted. The type of the gesture may be determined by comparing the vector value for a motion extracted through recognition of the gesture with a stored vector value, and a function corresponding to the gesture may then be performed, based on the determined type.
A gesture recognition method according to an embodiment of the present invention may include an operation of extracting one or more vector values for the gesture, an operation of generating a pattern of the vector based on the vector values, an operation of comparing one or more vector patterns with the generated pattern of the vector, and an operation of determining a type of the gesture, based on the comparison.
In accordance with an embodiment of the present invention, the vector value may include at least one phase value.
In accordance with an embodiment of the present invention, the operation of generating the pattern of the vector may include an operation of classifying the vector values through at least one of a variance, a deviation, and an average of the vector values.
In accordance with an embodiment of the present invention, in a comparison operation, the distribution of the vector values and the progress direction in which the vector values are generated may be compared.
In accordance with an embodiment of the present invention, in an extraction operation, when a new gesture is detected, one or more vector values for the new gesture may be extracted. As the new gesture is detected, the vector values for the new gesture may be accumulated and stored. The gesture may be detected through at least one of an image sensor, an IR sensor, and a touch panel.
A gesture recognition method according to an embodiment of the present invention may include operations of, when at least two multiple proximity inputs are detected, extracting at least one first vector value for the multiple proximity inputs, to, if a motion is generated while the multiple proximity inputs are detected, extracting at least one second vector value for the motion, analyzing motion loci of the first vector value and the second vector values, determining occurrence of a pinch gesture, and performing a function corresponding to the pinch gesture.
In the operation of performing the function corresponding to the pinch gesture at least one of functions of enlarging and reducing an image and adjusting a selection range of materials arranged in the Z-axis when the pinch gesture occurs may be performed.
Referring to
The AP 1510 may operate an Operating System (OS) and/or application programs to control a plurality of hardware and/or software components connected to the AP 1510 and perform data-processing and operations on multimedia data. For example, the AP 1210 may be implemented in the form of a System on Chip (SoC). According to an embodiment, the AP 1510 may include a Graphic Processing Unit (GPU).
The communication module 1520 (similar to the communication interface 160) may perform data communication with other electronic devices through a network. According to an embodiment, the communication module 1520 may include a cellular module 1521, a Wi-Fi module 1523, a BT module 1525, a GPS module 1527, an NFC module 1528, and a Radio Frequency (RF) module 1529.
The cellular module 1521 is responsible for voice and video communication, text messaging, and Internet access services through a communication network (e.g. LTE, LTE-A, CDMA, WCDMA, UMTS, WiBro, and GSM networks). The cellular module 1521 may perform identification and authentication of electronic devices in the communication network using the SIM card 1524. According to an embodiment, the cellular module 1521 may perform at least one of the functions of the AP 1510. For example, the cellular module 1521 may perform at least a part of the multimedia control function.
The cellular module 1521 may include a Communication Processor (CP). The cellular module 1521 may be implemented in the form of an SOC. Although the cellular module 1521, the memory 1530, and the power management module 1595 are depicted as independent components separated from the AP 1510, the present invention is not limited thereto, but may be embodied such that the AP includes at least one of the other components.
Each of the AP 1510 and the cellular module 1521 may load a command or data received from at least one of the components on a non-volatile or volatile memory and process the command or data. The AP 1210 or the cellular module 1521 may store the data received from other components or generated by at least one of other components in the non-volatile memory.
Each of the Wi-Fi module 1523, the BT module 1525, the GPS module 1527, and the NFC module 1528 may include a processor for processing the data it transmits/receives. Although the cellular module 1521, the Wi-Fi module 1523, the BT module 1525, the GPS module 1527, and the NFC module 1528 are depicted as independent blocks; at least two of these components may be integrated in the form of an SoC.
The RF module 1529 is responsible for data communication, e.g., transmitting/receiving RF signals. Although not illustrated, the RF module 1529 may include a transceiver, a Power Amp Module (PAM), a frequency filter, and a Low Noise Amplifier (LNA). The RF module 1529 also may include the elements for transmitting/receiving electric wave in free space, e.g. conductor or conductive wire. Although
The SIM card 1524 may be designed so as to be inserted into a slot formed at a predetermined position of the electronic device. The SIM card 1524 may store unique identity information (e.g. Integrated Circuit Card Identifier (ICCID)) or subscriber information (e.g. International Mobile Subscriber Identity (IMSI)).
The memory 1530 (e.g. memory 130) includes the internal memory 1532 and an external memory 1534. The internal memory 1532 may include at least one of a volatile memory (e.g. Dynamic Random Access Memory (DRAM), Static
RAM (SRAM), Synchronous Dynamic RAM (SDRAM) or a non-volatile memory (e.g. One Time Programmable Read Only Memory (OTPROM), Programmable ROM (PROM), Erasable and Programmable ROM (EPROM), Electrically Erasable and Programmable ROM (EEPROM), mask ROM, flash ROM, NAND flash memory, and NOR flash memory).
The internal memory 1532 may be a Solid State Drive (SSD). The external memory 1534 may be a flash drive such as Compact Flash (CF), Secure Digital (SD), micro-SD, Mini-SD, extreme Digital (xD), and Memory Stick. The external memory 1534 may be connected to the electronic device 1501 through various interfaces functionally. The electronic device 1501 may include a storage device (or storage medium) such as hard drive.
The sensor module 1540 may measure physical quantity or check the operation status of the electronic device 1501 and convert the measured or checked information to an electric signal. The sensor module 1540 includes a gesture sensor 1540A, Gyro sensor 1540B, atmospheric pressure sensor 1540C, magnetic sensor 1540D, acceleration sensor 1540E, grip sensor 1540F, proximity sensor 1540G, color sensor 1540H (e.g. Red, Green, Blue (RGB) sensor), biometric sensor 15401, temperature/humidity sensor 1540J, illuminance sensor 1540K, and Ultra Violet (UV) sensor 1540M. Additionally or alternatively, the sensor module 1540 may include E-nose sensor, Electromyography (EMG) sensor (not shown), Electroencephalogram (EEG) sensor, Electrocardiogram (ECG) sensor, Infrared (IR) sensor, iris sensor, and fingerprint sensor. The sensor module 1540 may further include a control circuit for controlling at least one of the sensors included therein.
The input device 1550 includes a touch panel 1552, (digital) pen sensor 1554, keys 1556, and an ultrasonic input device 1558. The touch panel 1552 may be one of capacitive, resistive, infrared, microwave type touch panel. The touch panel 1552 may include a control circuit. In the case of the capacitive type touch panel, it is possible to detect physical contact or approximation. The touch panel 1552 may further include a tactile layer. In this case, the touch panel 1552 may provide the user with haptic reaction.
The (digital) pen sensor 1554 may be implemented with a sheet with the same or similar way as touch input of the user or a separate recognition sheet. The keys 1556 may include physical buttons, optical key, and keypad. The ultrasonic input device 1558 is a device capable of checking data by detecting sound wave through a microphone 1588 and may be implemented for wireless recognition. The electronic device 1501 may receive the user input made by means of an external device (e.g. computer or server) connected through the communication module 1520.
The display module 1560 (similar to the display 150) includes a panel 1562. a hologram device 1564, and a projector 1566 The panel 1562 may be a Liquid Crystal Display (LCD) panel or an Active Matrix Organic Light Emitting Diodes (AMOLED) panel. The panel 1562 may be implemented so as to be flexible, transparent, and/or wearable. The panel 1562 may be implemented as a module integrated with the touch panel 1552. The hologram device 1564 may present 3-dimensional (3D) image in the air using interference of light. The projector 1566 may project an image to a screen. The screen may be placed inside or outside the electronic device. The display module 1560 may include a control circuit for controlling the panel 1562, the hologram device 1564, and the projector 1566.
The interface 1570 includes a High-Definition Multimedia Interface (HDMI) 1572, a Universal Serial Bus (USB) 1574, an optical interface 1576, and a D0subminiature (D-sub) 1578. The interface 1570 may include the communication interface 160, as illustrated in
The audio module 1580 may convert sound to electric signal and vice versa. At least a part of the audio module 1580 may be included in the input/output interface 140 as illustrated in
The camera module 1591 is a device capable of taking still and motion pictures and, according to an embodiment, includes at least one image sensor (e.g. front and rear sensors), a lens, and Image Signal Processor (ISP), and a flash (e.g. LED or xenon lamp).
The power management module 1595 may manage the power of the electronic device 1501. Although not illustrated, the power management module 1595 may include a Power Management Integrated Circuit (PMIC), a charger Integrated Circuit (IC), a battery, and a battery gauge.
The PMIC may be integrated into an integrated circuit or SoC semiconductor. The charging may be classified into wireless charging and wired charge. The charger IC may charge the battery and protect the charger against overvoltage or overcurrent. According to an embodiment, the charger IC may include at least one of wired charger and wireless charger ICs. Examples of the wireless charging technology includes resonance wireless charging and electromagnetic wave wireless charging, and there is a need of extra circuit for wireless charging such as coil loop, resonance circuit, and diode.
The battery gauge may measure the residual power of the battery 1596, charging voltage, current, and temperature. The battery 1596 may store or generate power and supply the stored or generated power to the electronic device 1501. The battery 1596 may include a rechargeable battery or a solar battery.
The indicator 1597 may display operation status of the electronic device 1501 or a part of the electronic device, booting status, messaging status, and charging status. The motor 1598 may converts the electronic signal to mechanical vibration. Although not shown, the electronic device 1501 may include a processing unit (e.g., GPU) for supporting mobile TV. The processing unit for supporting the mobile TV may be able to processing the media data abiding by the broadcast standards such Digital Multimedia Broadcasting (DMB), Digital Video Broadcasting (DVB), and media flow.
Each of the components of the electronic device according to the present disclosure may be implemented by one or more components and the name of the corresponding component may vary depending on a type of the electronic device. The hardware according to an embodiment of the present disclosure may include at least one of the above-described elements. Some of the above-described elements may be omitted from the hardware, or the hardware may further include additional elements. Further, some of the components of the electronic device according to the present disclosure may be combined to be one entity, which can perform the same functions as those of the components before the combination.
The term “module” used in the present disclosure may refer to, for example, a unit including one or more combinations of hardware, software, and firmware. The “module” may be interchangeably used with a term, such as unit, logic, logical block, component, or circuit. The “module” may be the smallest unit of an integrated component or a part thereof. The “module” may be the smallest unit that performs one or more functions or a part thereof. The “module” may be mechanically or electronically implemented. For example, the “module” according to the present disclosure may include at least one of an Application-Specific Integrated Circuit (ASIC) chip, a Field-Programmable Gate Arrays (FPGA), and a programmable-logic device for performing operations which has been known or are to be developed hereinafter.
According to various embodiments, at least some of the devices (for example, modules or functions thereof) or the method (for example, operations) according to the present invention may be implemented by a command stored in a computer-readable storage medium in a programming module form. When the command is executed by one or more processors (for example, the controller 210), the one or more processors may execute a function corresponding to the command. The computer-readable storage medium may be, for example, the memory 130. At least a part of the programming module may be implemented (for example, executed) by, for example, the controller 210. At least some of the programming modules may include, for example, a module, a program, a routine, and a set of instructions or a process for performing one or more functions.
The computer-readable recording medium may include magnetic media such as a hard disk, a floppy disk, and a magnetic tape, optical media such as a Compact Disc Read Only Memory (CD-ROM) and a DVD, magneto-optical media such as a floptical disk, and hardware devices specially configured to store and perform a program instruction (for example, programming module), such as a
ROM, a RAM, a flash memory and the like. In addition, the program instructions may include high class language codes, which can be executed in a computer by using an interpreter, as well as machine codes made by a compiler. The aforementioned hardware device may be configured to operate as one or more software modules in order to perform the operation of the present invention, and vice versa.
The programming module according to the present invention may include one or more of the aforementioned components or may further include other additional components, or some of the aforementioned components may be omitted. Operations executed by a module, a programming module, or other component elements according to various embodiments of the present invention may be executed sequentially, in parallel, repeatedly, or in a heuristic manner. Further, some operations may be executed according to another order or may be omitted, or other operations may be added.
The present invention also provides a recording medium that is implemented to recognize a gesture, in which a program for extracting one or more vector values for a gesture, generating a pattern of a vector based on the vector values, comparing one or more patterns of vectors stored in advance with the generated pattern of the vector, and determining a type of the gesture based on the comparison is stored.
The electronic device according to an embodiment of the present invention can easily perform a function corresponding to a gesture by recognizing the same input at any angle, even if the user bends their wrist or corrects their posture to input a pattern. Because the user can input a pattern without considering a state of the electronic device, the input can be recognized at a high rate and can be made even in situations in which the input may not be easily made.
While the present invention has been particularly shown and described with reference to certain embodiments thereof, it will be understood by those of ordinary skill in the art that various changes in form and details may be made therein without departing from the spirit and scope of the present invention as defined by the following claims and their equivalents.
Number | Date | Country | Kind |
---|---|---|---|
10-2014-0048216 | Apr 2014 | KR | national |