The present disclosure generally relates to a control system, a vehicle and a method for controlling multiple facilities.
Modern vehicles equip with various facilities, such as air conditioners, infotainment, to make drivers and passengers comfortable and entertained. Additionally, other facilities, such as rear-view mirrors, are used to facilitate safe driving. Those facilities are usually managed independently. For example, the buttons or switches disposed on an air conditioner can only control the air conditioner. Similarly, equipment like rear-view mirrors, wipers, and infotainment systems all have its dedicated controller. If a driver intends to control or adjust one of the facilities, such control or adjustment can only be conducted through that particular buttons or switches.
Another issue regarding the conventional arrangement is that those controlling interface such as knob, buttons, or switches are disposed in various places in the vehicle cockpit. The driver has to change his/her position or lean his/her body in order to reach the related buttons or switches, which adds potential safety concerns. In one scenario, assuming a driver wants to increase the temperature setting in the vehicle, he/she may need to take several seconds to find the correct control in a dashboard, which is a distracted behavior during driving. Alternatively, if the controls are made through a single touch panel, they may be arranged in different menu and hierarchies. While the level of the controlling interface complexity increases, the driver is likely to be distracted more when he/she attempts to operate the controlling interface.
Several technologies, such as a voice control, or a gesture control, have been provided to overcome this issue. However, neither of them is ideal. Voice control techniques require a preamble before it can work properly, let alone the surrounding noises could affect the recognition rate or falsely turn on the voice assistance. As for gesture control, since the driver would have to keep at least one hand on the wheel when driving the vehicle, the instructions can only be given through a bare hand and thus it is limited.
SUMMARY
In one exemplary embodiment disclosed herein, a control system is provided. The control system includes an image capturing unit, an input interface, and a processing unit. The image capturing unit is configured to capture a plurality of images of a user. The input interface is configured to receive an input command from the user, and the processing unit is configured to recognize a facial feature from the captured images; calculate a gaze point of the user according to the facial feature; determine a target facility among multiple facilities according to the gaze point of the user; receive a confirmation signal; configure the target facility as a facility subject to control when the confirmation signal is received; and control the facility subject to control in response to a control signal received from the input interface.
In another exemplary embodiment disclosed herein, a method for controlling multiple facilities is provided. The method includes the following actions. The image capturing unit captures a plurality of images of a user. The processing unit identifies a facial feature from the captured images; calculates a gaze point of the user according to the facial feature; determines a target facility according to the gaze point of the user; receives a confirmation signal; configures the target facility as a facility subject to control when the confirmation signal is received; and controls the facility subject to control in response to a control signal received from the input interface.
In yet another exemplary embodiment disclosed herein, a vehicle is provided. The vehicle includes multiple facilities and a control system. The control system includes an image capturing unit, an input interface and a processing unit. The image capturing unit captures a plurality of images of a user. The processing unit recognizes a facial feature from the captured images; calculates a gaze point of the user according to the facial feature; determines a target facility according to the gaze point of the user; receives a confirmation signal; configures the target facility as a facility subject to control when the confirmation signal is received; and controls the facility subject to control in response to a control signal received from the input interface.
Additional features and advantages of the disclosed technology will be made apparent from the following detailed description of embodiments that proceeds with reference to the accompanying drawings.
The following description contains specific information pertaining to exemplary implementations in the present disclosure. The drawings in the present disclosure and their accompanying detailed description are directed to merely exemplary implementations. However, the present disclosure is not limited to merely these exemplary implementations. Other variations and implementations of the present disclosure will occur to those skilled in the art. Unless noted otherwise, like or corresponding elements among the figures may be indicated by like or corresponding reference numerals. Moreover, the drawings and illustrations in the present disclosure are generally not to scale, and are not intended to correspond to actual relative dimensions.
In view of the foregoing, it is important to develop a system and a method for a user to control or adjust facilities in a non-distractive way; particularly when the user is operating a machine, e.g. a vehicle, which requires a hundred-percent concentration.
The image capturing unit 10 is configured to capture images of a user. In one implementation, the image capturing unit 10 may be, but not limited to, one or more cameras or any device capable of capturing images. The image capturing unit 10 may be a depth-sensing camera with a depth sensor. The camera may be an RGB color camera or an infrared (IR) camera. In some embodiments, the image capturing unit 10 further includes a light source (e.g., an IR illuminator or a visible light illuminator) enabling instant profiling of the hand/face/head/body of the user. With the light source and high dynamic range (HDR) imaging, the object detection and biometric recognition may be adapted to a darker environment.
The input interface 12 is configured to receive an input signal from the user. The input interface 12 may be, but not limited to, a microphone to record the sound, voice or speech of the user, a button, a switch, a knob, a touch panel, a keyboard. In another embodiment, the input interface 12 is coupled to an electronic device to receive data or instructions. For instance, the electronic device may include, but not limited to, a tablet, a voice receiving/recognition device, a wearable device or a mobile device. Thus, the input interface 12 may receive various types of input signal via different kinds of input interface.
The processing unit 14 is coupled to the image capturing unit 10 and the input interface 12. The processing unit 14 may process data and instructions. In one embodiment, the processing unit 14 may be a hardware module comprising one or more central processing unit (CPU), microcontroller(s), ASIC, or a combination of above but is not limited thereof. The processing unit 14 may perform image recognition and image processing. In one embodiment, the processing unit 14 is configured to detect a face from the captured images. In some embodiment, the processing unit 14 detect the head pose and head position from the captured images. In some embodiments, the processing unit 14 detects the gaze vector and project the gaze vector to the 3D coordinate. In some embodiments, the processing unit 14 further performs facial recognition, voice recognition and other biometric recognition. In some embodiments, the processing unit 14 configures the facilities in response to the input signal or control signal received from the input interface 12.
The method will be described as follow based on a scenario where the control system 100 is installed in a vehicle. However, the present invention is not limited thereto. In another example, the control system 100 may also be equipped in a room having multiple facilities to be controlled. In one example, the control system 100 may be installed in a vehicle where various facilities, such as a instrument cluster, an air conditioner, an infotainment system, rearview mirrors, and wipers are installed and subject to users' command through the control system 100. The controls to the various facilities are all made through the input interface 12. One or more parts of the input interface may be arranged on the steering wheel, beside the seat, on the roof, near the dashboard or anywhere reachable by the user. To be able to control/adjust various types of facilities, the processing unit 14 may have different configurations to accommodate the facilities. Precisely, given the nature of the facility being subject to the user's control, the control system 100 changes the configurations of the processing unit 14 to adapt to the functions of the facility. For instance, when a first configuration is applied to control an infotainment device, the user may issue various control signals to control the infotainment device via the configured input interface 12. The control signals may include tune the channel, turn up/down the volume, streaming music, etc. A first control signal, e.g. a double click on a first button of an input interface 12 will trigger a first function of the infotainment. Similarly, if a second configuration is applied to control a rear mirror, the user may issue various control signals to control the rear mirror via the configured input interface 12. The control signal should be able to adjust the position of the rear mirror. The first control signal, e.g. a double click on the first button of the input interface 12 will trigger a first function of the rear mirror. Thus, the settings of the processing unit 14 should be able to conduct various types of controls via the control signals from the input interface 12.
Inside a vehicle where the control system 100 is installed, a three-dimensional (3D) world coordinate system is established accordingly to locate the positions of a user and various facilities. The arbitrary point of the world coordinate system may be any dot inside the vehicle, preferably, a fixed point. Because the facilities are fixtures, their locations are ascertained in view of the world coordinate system. In one example, a location of a facility may be presented by the coordinate of the facility's reference point. The reference point may be any point on the facility; one typical example is its center point.
The operation of the control system 100 is disclosed as follows. To begin with, the image capturing unit 10 captures facial images of the user. The images should at least contain information of the user's eyes, so the image capturing unit 10 may be placed in the location where the necessary information can be conveniently obtained. In one example, the image capturing unit 10 is disposed right in front of the user when the user is facing directly straight. In any case, the position of the image capturing unit 10 is ascertained in view of the world coordinate system. Moreover, the image capturing unit 10 may capture the user's facial images continuously. Based on the continuously obtained images, the control system 100 can track the positional changes of the user's dynamic gaze point, i.e. the gaze trajectory, in real-time.
After the images of a user are captured by the image capturing unit 10, the processing unit 14 recognizes a facial feature of the user from the captured images. For instance, a face bounding box, an eye region, a head pose, a head position, a facial landmark, a gaze vector is recognized, and a gaze point is inferred and could be represented in light of the world coordinate system. As exampled in
Afterwards, the processing unit may infer which facility that the user is looking at once the user's position, particularly, the positions of the user's eyes, is ascertained and determines it as the target facility. For example, the processing unit estimates a gaze point of the user according to the gaze vector, the position of the user, and the positions of the facilities, to determine whether the user is looking at one of the facilities. Plainly, the gaze point P1 is the point where the user's left-eye vector V1 (i.e. representing the viewing direction from the user's left-eye) and right-eye vector V2 (i.e. representing the viewing direction from the user's right-eye) come across, as depicted in
If it is determined that the gaze point is on a target facility (e.g., their coordinates correlate), the processing unit configures the target facility as the facility subject to control. For example, a state of service signal may be issued to change the configurations of the input interface 12 or the processing unit 14 to adapt to the functions of the target facility. Consequently, the user may exercise the input interface 12 including buttons, switches, knobs, or a touch panel arranged somewhere nearby the user to command/control the target facility.
Alternatively, the input interface 12 may be coupled to a standalone controller such as a mobile device capable of controlling facilities. In such scenario, a communication link is established among the control system 100, the facilities and the standalone controller via a binding communication protocol, such as Wi-Fi, Bluetooth. In this way, the standalone controller, though is independent to the control system 100, can still be able to control the facilities through the communication link. In one embodiment, assuming the control is made through a mobile device, the user may even be able to cast contents, e.g. music, video, from the mobile device to the infotainment device via control signals or input signals which is under the user's control.
Additionally, those skilled persons would understand that all the variants including the coordinates of the user's, other facial landmarks such as the nose tip, the eyebrow, the lip, may also be considered as face features. to determine the user's gaze direction, head pose, viewing-angle, and therefore to conclude the gaze point. In some embodiments, the features for recognizing the gaze vector and gaze point may include the non-facial features such as the body skeleton. The above are mere examples without intending to limit the practices of the present invention.
In another example of the present disclosure, every facility may have a predefined operational zone to determine whether the user stares at the facility. The operational zone is delineated by expanding a reference point of the facility outward to a certain distance, or it may be the boundary of the facility with or without a buffer space. If the processing unit 14 determines that the user's gaze point falls within the operational zones, the corresponding facility will then be subject to the user's control through the operation. In some embodiment, the control system 100 may adopt more than one manner in one occasion to define operational zones of the various facilities. Delineations of operational zones will be disclosed in detail by reference to
As exemplarily shown in
As illustrated in
Optionally, each of the facility may have an operational zone upon which the processing unit 14 can determine whether a facility is subject to the user's control. In one embodiment, as depicted in
Assuming it is concluded that the gaze point Pa is on the infotainment system 3 or within the operational zone MZ3, the processing unit 14 issues a state of service to change the configurations of the input interface 12 or the processing unit 14 to accommodate the functions of the infotainment system 3. Subsequently, the user can exercise the input interface 12 arranged nearby to control the infotainment system 3, e.g. change the soundtracks, and/or increase or reduce the volume. As discussed, the input interface 12 may be a set of buttons/switches arranged somewhere nearby the user. In the previous example, the user may change the soundtrack displaying on the infotainment system 3 by, for instance, pressing the corresponding button/switch. Alternatively, if the input interface 12 is coupled to a standalone controller, the user will also be able to change the soundtrack and/or cast contents to the infotainment system 3 through the standalone controller provided a communication link has been established.
As disclosed, based on the nature of the facility subject to operate, the processing unit 14 of the control system 100 issues a state of service to configure the input interface 12 to accommodate the facility. For instance, the control signals of the infotainment system 3 may include change the channel, turn on/off the volume, streaming a music, etc. Similarly, if the control is made to the left rearview mirror 4, the configurations should be able to adjust the positions of the left rearview mirror 4. On the other hand, if the control is to the air conditioner 2, the control signals may include turn on/off, increase/decrease the indoor temperature, etc. In either case, the configurations should be able to accommodate the various functions and therefore to control different types of facilities.
It should be noted that in the
It should be noted that in some cases, probably because the facilities are too close or because the operational zones are not well defined, there is a chance that the user's current gaze point may be inferred on more than one facilities, and thus may cause a confusion as to what exactly the facility is subject to control.
To resolve the issue, a probability correlation level between the gaze point and the facilities may be employed to determine the target facility. To be precise, the correlation is characterized by a probability distribution value, in which a closer distance between the gaze point and the facility operation zone will result in higher probability. Alternatively, the user's confirmation may be used to help defining the facility subject to control. For instance, when a confirmation signal made by the user is received from the input interface 12, the processing unit 14 configures the target facility as the facility subject to control. In another instance, when the user stares at the target facility more than a period of time (e.g., 2 seconds), the target facility is confirmed to be the facility subject to control. In some other instances, a facial gesture of the user is used as a confirmation of the target facility. After the confirmation, the user may take control over the facility. In this case, even when the user looks away from the confirmed target facility, the user may still be able to control the facility via the input interface 12. In other words, the driver doesn't need to look at the facility to make commands, the driver could keep his/her eyes on the road while driving and make commands to the facility at the same time and thus reduce the distraction. Therefore, the present disclosure provides a convenient and safer way for the driver to control multiple facilities.
Alternatively, a point of interest may be used to define the facility subject to control. The point of interest differs than the gaze point is that the point of interests is a conclusion of feature behaviors from the continuous image frames and patterns. In one implementation, the point of interest is represented by a probability model, which will be established to determine which facility, more probable than not, that the user would like to control. The probability model may be determined based on various factors, such as, a gaze point of the user, a position of the user, positions of the facilities, a z-depth, a nature of the facility, and a user's preference (e.g., the user's habit or previous behaviors). The model is preloaded into the control system 100 for determining the point of interest. Through some sorts of deep-learning, machine learning or similar techniques familiar by the skilled persons, the point of interest could be estimated. Following the above example and as shown in
Sometimes a facility may take a while to initiate before it can actually be controlled. Under such circumstance, the user may have to wait for the initiation. To shorten the latency, the processing unit 14 may predict in what direction that the user would be looking toward and therefore prepare/pre-initiate the facilities alongside with the predictive path before the user stares at them. For instance, the processing unit 14 predicts a gaze trajectory according to a variation of the facial feature, such as the on-going positional change of the user's eyes and/or head pose, and maybe the user's previous habits.
In another embodiment, the processing unit 14 may predict a facility subject to control according to various factors, for example, the vehicle's status, the point of interest, and the user's preference. For instance, if the user tends to adjust the right rearview mirror after an adjustment to the left rearview mirror. The processing unit 14 anticipates the tendency and prepares the right rearview mirror while the user is adjusting the left rearview mirror. In this way, the latency would be reduced. Furthermore, the processing unit 14 initiate the infotainment system since it is estimated as the point of interest, or when the vehicle stops.
In some embodiment, the control system 100 pre-initiate one of the facilities beforehand.
It should however be noted that just because some facilities are pre-initiated, it does not mean that they will ultimately be controlled by the user. Under the design of the present disclosure, it is only when the user's gaze point falls on a target facility (i.e. the user actually stares at the target facility) that a control through the input interface 12 can be made to it.
In some embodiments, to better understand what facility is being controlled, an indicator, such as a LED, may be disposed on each of the facilities to show the status. In a further example, the LED may have different colors to indicate the status of a facility. For example, the facility is inactive (i.e. not subject to control), the facility is currently under the user's control through the input interface 12, and the facility is initiated and ready to be controlled by the user when the user stares at it. In yet another example, if multiple displays are operated, a cursor, for instance, may appear on the display subject to control so that the user will understand he/she is controlling which display.
In some embodiments, there could be many people in the same space want to control the facilities. Assuming the control system 100 is installed in a vehicle, either the driver or the passenger(s), may attempt to control/adjust the facilities. As a result, it is important to know who has the authority and the priority to command. Moreover, since there could be other standalone controllers capable of controlling the facilities, the control system 100 should also be able to distinguish the dominating controller over the others. In one embodiment, the control system 100 may provide a log-in service to verify users. The verification may be conducted through, for example, facial recognition. For instance, a user may be required to register his/her face IDs before he/she can command the facilities through the control system 100. The face ID may be stored in a storage of the control system 100. Since there can be more than one user whose face IDs are registered, the control system 100 may further prioritize the registered users.
In a scenario, a user may engage the control system 100 by, for example, appearing in front of the image capturing unit 10, the image capturing unit 10 captures the user's facial features to obtain his/her face ID, and then the processing unit 14 verifies if the user is one of the registered users by comparing the face ID against the data stored in the storage. If the face ID is verified, the user can command the facilities connected to the control system 100 through the operations of the input interface 12. Unless and until the user exits the control system 100, he/she is the super user dominating all the controls through the control system 100. Alternatively, a user may log in to the control system 100 through his/her own mobile device which has been registered to the control system 100.
In another occasion, if, for example, a driver and a passenger in a vehicle are both registered users and are both attempting to control the connected devices. In this instance, depending on the ranking, the control system 100 will determine who, either the driver or the passenger, has the priority over the other. Once decided, the facilities connected to the control system 100 are subject to the super user's command. In the same manner, only the standalone controller holding by the super user can be used to control/adjust the subject facility. The control system 100 may recognize the standalone controller by some sort of biometric data, e.g. face ID, also stored therein.
There are many ways to authenticate a user identity; facial recognition is a mere example to explain how the authentication can be achieved. It should however under no circumstance become a limitation to the present invention.
As discussed previously, the control system is provided to control various types of the facilities. Thus, the method of the present disclosure further includes modifying the configurations of the input interface to accommodate to the functions of the target facility. Further, the determination of the gaze point may be based on the user's eye vectors, his/her head pose, and nose tip, etc. Besides, the method of the present disclosure may also include defining an operational zone to each of the facilities and determining if the gaze point falls within any of the operational zones to conclude the target facility. Additionally, the method of the present disclosure may further include notifying the user through an indicator indicating that the target facility is subject to control.
Apart from the above, the method of the present disclosure may further include predicting a predictive path indicating the direction where the user would be looking toward and pre-initiating the facilities alongside with the predictive path. The predictive path may be obtained given the on-going positional/directions changes of the user's eyes and/or head pose, and the user's previous behavior.
Last but not least, the method of the present disclosure may also include requesting users to register and consequently log in to the system; and prioritizing the registered users to determine whose control over the facilities dominates.
In summary, the proposed invention could reduce the distraction and simplify the procedures that a user may have when it comes to operate facilities. Embodiments of the invention may include various operations as set for the above or fewer operations or operations in an order that is different from the order described. The operations may be embodied in machine-executable instructions that cause a general-purpose or special-purpose processor to perform certain operations. That is, the techniques may be carried out in a computer system or other data processing system in response to its processor, such as a microprocessor, executing sequences of instructions. Such as a computer program may be stored or transmitted in a machine-readable medium. A machine-readable medium includes any mechanism that provides (i.e. store, and/or transmits) information in a form accessible through a machine (e.g., a computer, network device, personal digital assistant, and processors). For example, a machine-readable medium includes recordable/non-recordable media such as, but not limited to, a machine-readable storage medium (e.g. any type of disk including floppy disks, optical disks, CD-ROMs, and magnetic-optical disks, read-only memories (ROMs), random access memories (RAMs), EPROMs, EEPROMs, flash memory, magnetic or optical cards, or any type of media suitable for storing electronic instructions), or a machine-readable transmission medium such as, but not limited to, any type of electrical, optical acoustical or other form of propagated signals (e.g. carrier waves, infrared signals, digital signals).
Modifications, additions, or omissions may be made to the systems, apparatuses, and methods described herein without departing from the scope of the disclosure. For example, the component of the systems and apparatuses may be integrated or separated. Moreover, the operations of the systems and apparatuses disclosed herein may be performed by more, fewer, or other components and the methods described may include more, fewer, or other steps. Additionally, steps may be performed in any suitable order. As used in this document, “each” refers to each member of a set or each member of a subset of a set.
Based on the above, several control systems and methods for controlling multiple facilities, and vehicles having a control system are provided in the present disclosure. The implementations shown and described above are only examples. Even though numerous characteristics and advantages of the present technology have been set forth in the foregoing description, together with details of the structure and function of the present disclosure, the disclosure is illustrative only, and changes may be made in the detail, including in matters of shape, size and arrangement of the parts within the principles of the present disclosure up to, and including, the full extent established by the broad general meaning of the terms used in the claims.
This application claims the benefit and priority to of U.S. Provisional Application Ser. No. 62/788,942, filed on Jan. 6, 2019, and entitled “SYSTEM AND METHOD FOR CONTROLLING FACILITIES”, which is incorporated herein by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
62788942 | Jan 2019 | US |