The present invention relates to a human machine interface of an electronic device and a method of commanding an electronic device through a human machine interface.
People familiar with UI design understand that underneath a user interface are layers in hierarchy. The layers define pages where various functional buttons are disposed thereon to control operations of an electronic device. That is, through a selection of a functional button, a user can control the corresponding operation of the electronic device. However, sometimes when a user interface is poorly designed, the user may have to go through pages and pages to find the operation he/she intends. For instance, assuming a user wants to change the background theme of a touchscreen-based media displayer. In this case, the user may firstly press the functional button on the main page where the system configuration located, then find the change of theme button in the subordinated layer, and subsequently select the background theme he/she wishes in the next inferior layer, and lastly confirm the selection and the setting. The above manual operations may be feasible as far as handheld devices, such as a mobile phone, and personal computers are concerned. However, when it comes to a driver who is driving a vehicle, any slightly more complicated operation will pose a great threat and is extremely dangerous. Thus, there is a need to provide a method for drivers to control facilities installed in vehicle in safer manners.
According to an embodiment of the present disclosure, a method of commanding an electronic device is provided. The method includes the following actions: providing a user interface where a plurality of functional objects are presented on a screen. Each of the functional objects has an underlying function. The method also includes: detecting at least one user motion; translating the user motion to at least one instruction, wherein the instruction includes a selection of one functional object; and instructing the electronic device to perform the underlying function of the selected functional object.
According to another embodiment of the present disclosure, an electronic device having a hand-free user interface is provided. The electronic device includes: a screen displaying a page containing a plurality of functional objects, wherein each of the functional objects has an underlying function; a sensor configured to receive at least one user motion made by a user; and a processor configured to translate the user motion into at least one instruction, wherein the instruction includes a selection of one functional object; wherein, upon a confirmation of the selection, the processor causes the electronic device to perform the underlying function of the selected functional object.
According to another embodiment of the present disclosure, an electronic device having a user interface upon which a user command the electronic device through a plurality of user motions associated with instructions to the electronic device is further disclosed. The electronic device includes: a screen configured to display a page containing a plurality of functional objects to the user, wherein each of the functional objects corresponds to an underlying function; a sensor configured to detect at least an eye-related motion; and a processor configured to translate the eye-related motion into an instruction operable to move an indicator on the screen for a selection of one of the functional objects; wherein the eye-related motion is determined based on the user's gaze vector and gaze point on the screen; wherein, upon a confirmation of the selection, the processor causes the electronic device to perform the underlying function of the selected functional object.
These and other objectives of the present invention will no doubt become obvious to those of ordinary skill in the art after reading the following detailed description of the preferred embodiment that is illustrated in the various figures and drawings.
The present disclosure provides a human-machine interface (HMI) operable to an electronic device installed in a vehicle. The disclosed HMI is operated through binding user motions including head-related, facial-related and/or eye-related motions. By exercising one user motion exampled above, the electronic device perform a binding function accordingly.
The electronic device 10 includes a user interface 100 and a processor 110. The user interface 100 provides a medium in between a user and the electronic device 10. The user interface 100 includes a screen 102 and a sensor 104. In an embodiment, the screen 102 may be any type of display device such as a liquid crystal display (LCD) or an organic light emitting diode display (OLED display); and additionally, the screen 102 may be a touch screen of any sort.
In an embodiment, the sensor 104 may be any type of sensor capable of obtaining and appreciating user data. The sensor 104 may be a visible light camera, infrared camera, light sensor, or combinations of multiple types of sensors. In an alternative embodiment, the sensor 104 may be independent to the user interface 100.
In the instant disclosure, the user data includes various of user motions that are perceived by the sensor 104. In an embodiment, the sensor 104 detects head-related motions constituted by, for instance, a change of head pose, a change of head position, and/or any combinations of the above. The head position may be defined by a relative position of the user's head in view of the sensor 104, or by a relative position of the user's head in relation to the screen 102. The head-related motions may also include head poses such as a head nod, a head shake, and a head tilt. Further, others such as a yaw rotation, a pitch rotation and a roll rotation are also considered as head-related motions in the embodiment.
In an embodiment, the sensor 104 detects eye-related motions including, for example, a change of gaze vector, a change gaze point and/or any combinations of the above. The gaze vector refers to a gaze from the user's eyes toward the direction where the user is looking at. The gaze vector may be derived from at least one of the head position, the head pose, the light reflected from the iris, the iris position and/or the pupil position. The gaze vector may be represented with reference to a world coordinate system, a facial coordinate system, or a display coordinate system. The gaze point is the point of interest on a display screen where the user stares at. The gaze point may be derived by reference to the gaze vector and the positions of two eyes.
In an embodiment, the sensor 104 detects facial-based motions including, for instance, a change of facial expression, a series of facial motions, and/or any combinations of the above. A facial-based motion may be a static facial expression, or a dynamic facial motion. The facial-related motion in accordance with the present disclosure may include a smile, a laugh, an opened mouth, an eye blink or a wrinkled nose and so on.
In an embodiment, the processor 110 may be a central processing unit (CPU), a microprocessor, a micro controller unit (MCU) or an application-specific integrated circuit (ASIC). The processor may be implemented with hardware and/or software.
In the present disclosure, numbers of user motions are predefined and associated with instructions and functions. Upon receiving a user motion from the sensor 104 via the user interface 100, the processor 110 translates the user motion into a corresponding instruction given to the electronic device 10 to perform an underlying function accordingly. For instance, if a functional object namely “System Configuration” is selected and confirmed, through several operations of user motions, the electronic device will carry the function for the user to change settings of the electronic device 10.
In one embodiment, the head-related motions are relevant to the control the way in which the functional objects are laid out in a page and displayed on the screen 102. Exemplarily, a head-related motion may instruct the screen 102 to pan the image displayed thereon; and another head-related motion may cause the screen 102 to zoom in the page. Based on the similarly concept, various user motions can be defined and associated with different instructions.
Sometimes a page may be too larger to fit into a display screen. One way to show the entire page is by reducing its resolution to accommodate the screen size. But the content on the page may be suffered and become difficult to read. Alternatively, one may choose to only display the page partially, and meanwhile allow the viewer to pan, zoom-in and zoom-out the page manually. As iterated, those manual operations are impractical when it comes to drivers.
In another implementation, there are user motions designed and designated to zoom in and out the page on the screen 102. Conceptually speaking, the user may simply lean to the screen to zoom in the page; on the other hand, is the user is moving his/her head away from the screen, the page on the screen 102 is zoomed out. As shown in
Although the above examples are separately discussed, it should be understood that one user motion may be operated in conjunction another or others. Continuing to the above two examples, the user may firstly instruct the screen 102 to zoom in the page by leaning the head toward the screen, and subsequently instruct the screen 102 to display the unseen portion on the right-hand side by turning the head to the left. By the operations, the user can see, for instance, an enlarged size of the blocks C and D on the screen.
In another embodiment, the user interface 100 may accept commands made by eye-related motions. Assuming an indicator is designed underneath the user interface 100 to indicate which functional object on the screen 102 is being pointed to. The indicator functions like a cursor but may be in any other forms. For instance, instead of having a pointer, the indictor may be in the form of color-highlighting, frame-highlighting, various curving effects, etc. to show what functional object on the screen 102 is being pointed to.
In yet another embodiment, the user may command the electronic device 10 via the user interface 10 to perform a function in response to a facial-related function.
User interface 100 may include functional objects, once selected by the user, will cause the electronic device 10 to perform an underlying function. For instance, under a conventional operation on a mobile device, a click on an icon of an application on the screen will cause the electronic device 10 to run the application. Under the design of the instant disclosure, the series of operations are made through user motions. For instance, following the above example, assuming the object “B” is an icon of a map application. To instruct the electronic device 10 to execute the map application, firstly the user turns his/her eyesight toward the direction where the object “B” locates. As discussed above, the sensor 104 retrieves the positional change and the processor 110 consequently moves the cursor to the relevant position to indicate the object “B” is being selected. Additionally, as discussed above, a user motion may be predefined. When the user exercises the motion, a binding function is performed. In the instant embodiment, the binding function is a confirmation command for the selection. In the realm of conventional computer operation, it is like a click on the icon. Continuing to the example, once the object “B” is locked, the user may perform the binding user motion (e.g. a smile) to confirm the selection of the object “B”. In effect, the sensor 104 senses the facial motion and then causes the processor 110 to execute the map application. From the user's perspective, the only two things he/she needs to do are: look at the object “B” and smile, as simple as that and without any manual operations, the map application will then be launched.
Aside from receiving user motion-related commands via the user interface 100, instructions may be manually input to the electronic device 10 through the user input device 120. The user input device 120 may be, for instance, physical buttons, or sensors for sensing, hand and/or palm gestures. Although the user interface 100 and the user input device 120 are separately depicted in
In action 302, a user interface where a plurality of functional objects are presented on a screen is provided.
In action 304, a sensor detects a gaze vector of a user and determines a gaze point based on the gaze vector.
The gaze vector refers to the direction where the user is looking toward. The gaze vector may be obtained based on the head position, the head pose, the light reflected from the iris, the iris position and/or the pupil position, etc. The gaze vector may be defined by reference to a world coordinate system, a facial coordinate system, or a display coordinate system. The gaze point is a point of interest on the screen. It can be calculated based on the gaze vector.
In action 306, a processor moves an indicator on the screen to the position where the user's gaze point falls.
In action 308, the sensor detects a user motion.
The user motion may be a facial-related motion, an eye-related motion, or a head-related motion. In any case, the user motion must be predefined and associated with an instruction for performing a binding function.
In action 310, the processor translates the user motion into an associated instruction.
In another instance, the instruction is a made by the user via pressing a physical button, inputting a specific hand gesture, or inputting with a touch screen.
In action 312, the processor commands the electronic device to execute the underlying function in response to the instruction.
In some implementations, the sensor 104 may continuously retrieve head-related motions. Precisely, the sensor 104 may obtain the real-time position of the user's head. The movement of the user's head may be used by the processor 110 to a control a page's up and/or down on the screen 102. For instance, the page is scrolled down if the user moves his/her head down, and scrolled up if the user turns the head upward. Similarly, if the user turns the head to the left, the sensor 104 detects the movement and the processor 110 consequently turns the page to the next one; or if the users turns the head to right, the screen 102 displays the previous page instead. Alternatively, the page's size may be largely designed and therefore exceed the size of the screen 102. In the instance, the processor 110 may decide to partially display the page on the screen 102 while leave some portions unseen (i.e. hidden part) to the user. To see the hidden part, the user may turn the head right to move the page right to see the hidden part on the left; or turn the head left to move the page to the left accordingly to see the unseen portion on the right. As mentioned, the head-related motions are predefined and associated with the aforementioned instructions. When the sensor 104 detects a head motion, the processor 110 translates the motion into the corresponding instruction and control the display of the page accordingly on the screen 102. In the convention way that users may have to manually go through pages over pages through arrangements likes a mouse or a touch pad. However, under the design of the instant disclosure, none of the actions are hand-operated. Users, especially the driver in a vehicle, can therefore focus on driving without distraction.
In yet embodiment, the page may be zoomed in or out depending on the movement of the user's head. Specifically, the page is enlarged if the user leans toward screen 102. On the other hand, the page is zoomed out if the user moves the head away from the screen 102. As mentioned, the sensor 102 retrieves the positional change and the processor 110 translates the motion into the corresponding instruction and controls electronic device 10 to show the portion of the page on the screen 102 accordingly. Apart from the above, the head-related motions may further include head-tilted and rotations. For example, if the user tiles the head to the left, the page is instructed to pan left as well. And if the user tilts the head to the right-hand side, the page is instructed to pan right. However, the present disclosure is not limited thereto. In the present disclosure, the page can be displayed in response to the given instruction issued by the user through user motions.
In the instant embodiment, a page 400 containing three functional objects are designed. The three items include: “System Configuration”, “Information Mode”, and “GPS Mode”. Each of the functional objects carries an underlying function as stated by its label. As shown in
In this embodiment, the functional objects on the page are arranged in a row; while in another embodiment, the functional objects may be arranged in other layouts, e.g., an arc. In such condition, when a rightward movement is detected, the processor 110 instructs the page to move clockwise. Similarly, when a leftward movement is detected, the processor 110 instructs the page to move counterclockwise.
In one embodiment, the electronic device 10 may be an automotive electronic device, and operated by the driver. For instance, the screen 102 may be a screen of a dashboard, and the page 400 displayed thereon contains abundant of vehicle information. For example, if the “Information Mode” is selected, the screen 102 may show vehicle information such as the weather, time and travel route(s). If the “GPS Mode” is selected, the screen 102 may show GPS navigation information such as a map and destinations. Further, if the “System Configuration” is selected, the screen 102 may show the configuration menu for the dashboard's settings. It should be noted that the functional objects on the page 400 are not limited to these three; and in other embodiments, there could be other functional objects such as “a speedometer information”, “a dashboard camera information”, “connecting to remote mobile device mode”. The display and the selection of any of the functional objects may be conducted through a user motion as discussed previously.
Apart from the head-related motions, the sensor 104 may also detect the eye-related motions. Similarly, several eye-related motions are predefined and associated with instructions. In one embodiment, the instructions may include an instruction to move an indicator on the screen in a designated direction. The indicator may be realized in various manners, such as a cursor, an icon, an outline, or a specific color to indicate what functional object or an area of the screen is being selected.
In the example of
Once the indicator is on the desired position where a functional object is being selected, the processor 110 determines whether a confirmation command is received. In one embodiment, the confirmation command may be issued through the user input device 120. When the selection is confirmed, the processor 110 controls the electronic device 10 to perform the underlying function in response to the selected functional object. In another embodiment, the confirmation command may be given by facial-related motions detectable by the sensor 104. The confirmation command (regardless how it is performed, through a facial-related motion or the user input device 120) is in effect similar to a click operation of a mouse or a touch pad.
The facial-related motions may be in various forms. For example, they may be any facial expressions such as a smile, a laugh, an opened mouth, or a wrinkled nose. The facial expression may further be a series of dynamic facial motions. The facial expression may be enrolled and associated with instructions in advance. For example, a user may associate a smile (i.e., the rising of the mouth corners) with a performance of a binding function (e.g. a confirmation command for a selection). When the sensor 104 detects that the user smiles (i.e. the user's mouth corners rise), the processor interprets the smile and translates it into a confirmation command.
An exemplary operation of performing an underlying function according to the facial-relate motion is described with reference to
In another embodiment, the identity of the user may also be considered. The function is critical in shared car scenarios. That is, even for the same user motion, different person may associate it with different functions. For instance, for a first user, a smile may be associated with an instruction to confirmation; while for a second user, a smile may be an instruction to quit the selection. The details as to how to distinguish users based on facial features are discussed in the '053 application and will be skipped in the present disclosure.
The above detections and recognition of head-, eye- and facial-related motions may be achieved through computer vision technologies. For instance, the sensor 104 obtains a head photo of the user. The processor 110 then removes irrelevant parts, down-samples the photo, and then retrieves the head image via some sort of facial detection/recognition technologies to delineate the boundary of the head image (i.e. a bounding box). The operation of down sampling is optional to the present disclosure; the upside is it will reduce the data volume and therefore increases the processing speed. The processor 110 may then retrieve the image contents within the bounding box to further processing. The detections of head-, eye- and facial-related motions may be obtained from the image contents within the bounding box.
The processor 110 may apply facial landmark detection technologies to obtain the landmark position of facial features such as eyes, mouth and noise. The processor 110 obtains the head-related motions of the user to determine the head pose such as the head position, movement, or whether the head tilts or rotates. The processor 110 may further determine the iris position, the pupil position and/or the light reflected from the iris from an eye image, to calculate the gaze vector and the gaze point of the user. The processor 110 may further determine the facial expression according to variations on a specific area or a part of a facial feature, or according to a series of dynamic facial variations. For example, a smile is determined when it is detected that the corners of mouth are rised. The position(s) of each facial features may further be considered to determine the landmarks. The above are mere examples to illustrate how a head-, eye-, and/or a facial-related motion may be detected, and under no circumstance become a limitation to the present disclosure.
In the present disclosure, a hand-free user machine interface is disclosed. Those skilled in the art may make modifications and alternations accordingly. For example, in the above embodiments, the head, eyes and facial motions are associated with various of display instructions. By performing those motions, the user can easily adjust the content and/or the position of a page shown on the screen. Further, the user motions may also be associated with various instructions/functions. Once a motion is made, it will lead the electronic device to perform the underlying function.
It should be noted that the present disclosure may also be working in conjunction with other hand-free user interface, such as voice commands, gesture commands, etc.
Additionally, the user interface of the present disclosure may also be implemented on electronic devices of any kind. The electronic devices may include, an in-vehicle electronic system, a video game console, a handheld device, an e-book reader, etc.
Those skilled in the art will readily observe that numerous modifications and alterations of the device and method may be made while retaining the teachings of the invention. Accordingly, the above disclosure should be construed as limited only by the metes and bounds of the appended claims.
Number | Date | Country | Kind |
---|---|---|---|
201711099567.9 | Nov 2017 | CN | national |
201721487383.5 | Nov 2017 | CN | national |
201711482541.2 | Dec 2017 | CN | national |
201721900090.5 | Dec 2017 | CN | national |
201810237040.6 | Mar 2018 | CN | national |
201820387687.2 | Mar 2018 | CN | national |
This patent application is a continuation-in-part application and claims the benefit of U.S. patent application Ser. No. 15/935,053 (“'053 application”) filed on 25 Mar. 2018 which is incorporated by reference herein.
Number | Date | Country | |
---|---|---|---|
Parent | 15935053 | Mar 2018 | US |
Child | 16234579 | US |