This disclosure generally relates to a mobile robot and, more particularly, to a mobile robot that performs the obstacle avoidance, positioning and object recognition according to image frames captured by the same optical sensor corresponding to lighting of different light sources.
The smart home is one part of developing a smart city, and a cleaning robot has almost become one standard electronic product in a smart home. Generally, the cleaning robot is arranged with multiple functions to improve the user experience, e.g., including mapping of an operation area, obstacle detection and avoidance during operation. The current cleaning robot is employed with multiple types of sensors to perform these different detecting functions.
For example, the cleaning robot includes a sensor arranged at a top surface thereof to implement the visual simultaneous localization and mapping (VSLAM) by capturing images above the path by which the cleaning robot passes. In addition, the cleaning robot further adopts a front sensor to implement the obstacle detection and avoidance by capturing images in front of a moving direction of the mobile robot.
That is, the conventional cleaning robot needs multiple sensors to perform different detecting functions.
Accordingly, the present disclosure provides a mobile robot that performs the obstacle avoidance, positioning and object recognition according to the image frames captured by the same one optical sensor corresponding to lighting of different light sources.
The present disclosure provides a mobile robot that performs the obstacle avoidance according to the image frame captured by an optical sensor when a laser diode is emitting light, and performs the visual simultaneous localization and mapping (VSLAM) according to the image frame captured by the optical sensor when a light emitting diode is emitting light.
The present disclosure further provides a mobile robot that determines a region of interest according to the image frame captured by an optical sensor when a laser diode is emitting light, and performs the object recognition in the region of interest of the image frame captured by the optical sensor when a light emitting diode is emitting light to reduce the computation loading and power consumption as well as improve the recognition correctness.
The present disclosure provides a mobile robot including a first optical sensor, a light source, a second optical sensor and a processor. The first optical sensor has a field of view toward a bottom opening of the mobile robot, and is configured to capture first image frames using an exposure time determined by an auto exposure. The light source is configured to project a light section toward a moving direction of the mobile robot. The second optical sensor is configured to capture a second image frame containing the light section. The processor is electrically coupled to the first optical sensor and the second optical sensor, and configured to predict a time interval during which the field of view of the first optical sensor contains a low reflectivity region according to the second image frame, and keep the exposure time of the first optical sensor not changed within the time interval.
The present disclosure provides a mobile robot including a first optical sensor, a light source, a second optical sensor and a processor. The first optical sensor has a field of view toward a bottom opening of the mobile robot, and is configured to capture first image frames using an exposure time determined by an auto exposure. The light source is configured to project a light section toward a moving direction of the mobile robot. The second optical sensor is configured to capture a second image frame containing the light section. The processor is electrically coupled to the first optical sensor and the second optical sensor, and configured to predict a time interval during which the field of view of the first optical sensor contains a low reflectivity region according to the second image frame, and select a predetermined exposure time, which is between 1.5 and 3 times of an exposure time determined prior to the time interval, of the first optical sensor within the time interval.
The present disclosure provides a mobile robot including a first optical sensor, a light source, a second optical sensor and a processor. The first optical sensor has a field of view toward a bottom opening of the mobile robot, and is configured to capture first image frames using an exposure time determined by an auto exposure. The light source is configured to project a light section toward a moving direction of the mobile robot. The second optical sensor is configured to capture a second image frame containing the light section. The processor is electrically coupled to the first optical sensor and the second optical sensor, and configured to predict a time interval during which the field of view of the first optical sensor contains a low reflectivity region according to the second image frame, and control the first optical sensor to have identical exposure times at a start and an end of the time interval.
The present disclosure further provides a mobile robot including a first laser light source, a second laser light source, a LED light source and an optical sensor. The first laser light source is configured to project a transverse light section toward a moving direction at a first time interval. The second laser light source is configured to project a longitudinal light section toward the moving direction at a second time interval. The LED light source is configured to illuminate a front area of the moving direction at a third time interval. The optical sensor is configured to respectively capture a first image frame, a second image frame and a third image frame within the first time interval, the second time interval and the third time interval.
In the present disclosure, the mobile robot realizes multiple detecting functions by using a single optical sensor incorporating with different light sources activating at different times.
Other objects, advantages, and novel features of the present disclosure will become more apparent from the following detailed description when taken in conjunction with the accompanying drawings.
It should be noted that, wherever possible, the same reference numbers will be used throughout the drawings to refer to the same or like parts.
The mobile robot of the present disclosure is to operate using a single optical sensor incorporating with different light sources. The linear light source is used to find an obstacle and measure a distance of the obstacle as a reference for turning a moving direction of the robot. The illumination light source is used to illuminate a front area for the visual simultaneous localization and mapping (VSLAM) and the object recognition.
Referring to
Please referring to
The first light source LS1 includes, for example, a laser light source and a diffractive optical element. The diffractive optical element causes light emitted by the laser light source to generate a transverse projecting light after passing thereby such that the first light source LS1 projects a transverse light section toward a moving direction. The moving direction is along a side arranging the first light source LS1, the second light sources LS21 and LS22, the third light source LS3 and the optical sensor 11.
The second light sources LS21 and LS22 respectively include, for example, a laser light source and a diffractive optical element. The diffractive optical element causes light emitted by the laser light source to generate a longitudinal projecting light after passing thereby such that the second light sources LS21 and LS22 respectively project a longitudinal light section toward the moving direction.
In the present disclosure, the laser light source is, for example, an infrared laser diode (IR LD).
The third light source LS3 is, for example, an IR light emitting diode (LED), and used to illuminate a front area of the moving direction. An area illuminated by the third light source LS3 is preferably larger than or equal to a field of view of the optical sensor 11. In the present disclosure, when the third light source LS3 is lighted up, the first light source LS1 as well as the second light sources LS21 and LS22 are turned off.
Please referring to
The optical sensor 11 is, for example, a CCD image sensor or a CMOS image sensor that captures a first image frame, a second image frame and a third image frame respectively within the first time interval T1, the second time interval T2 and the third time interval T3 using a sampling frequency. When the first image frame contains an obstacle, the first image frame has a broken line as shown in
It is appreciated that as the second light sources LS21 and LS22 project two parallel light sections on a moving surface, in the second image frame captured by the optical sensor 11, two parallel light sections present tilted lines. In addition,
The position of broken line in the image frame reflects a position of the obstacle in front of the mobile robot 100. As long as the relationship between the position of broken line in the image frame and the actual distance of obstacles is previously recorded, a distance of one obstacle from the mobile robot 100 is obtainable when an image frame containing a broken line is captured.
As shown in
As shown in
The processor 13 is electrically coupled to the first light source LS1, the second light sources LS21 and LS22, the third light source LS3 and the optical sensor 11, and used to control ON/OFF of light sources and the image capturing. The processor 13 further performs the range estimation according to the first image frame (e.g.,
Referring to
Although
In one aspect, the optical sensor 11 includes a pixel array. All pixels of the pixel array receive incident light via an IR light filter. For example,
In another aspect, the pixel array of the optical sensor 11 includes a plurality of first pixels PIR and a plurality of second pixels Pmono, as shown in
In the aspect including two pixel types, the first image frame and the second image frame mentioned above are formed by pixel data generated by the plurality of first pixels PIR. That is, the processor 13 performs the range estimation only according to pixel data generated by the plurality of first pixels PIR. The third image frame mentioned above is formed by pixel data generated by both the plurality of first pixels PIR and the plurality of second pixels Pmono since the first pixels PIR and the second pixels Pmono both detect infrared light when the third light source LS3 is emitting light. The processor 13 is arranged to process the pixel data corresponding to the lighting of different light sources.
In one aspect, the plurality of first pixels PIR and the plurality of second pixels Pmono of the pixel array are arranged as a chessboard pattern as shown in
In the aspect that the first pixels PIR and the second pixels Pmono are arranged in a chessboard pattern, the processor 13 further performs the pixel interpolation on the first image frame and the second image frame at first so as to fill interpolated data at positions in the first image frame and the second image frame corresponding the second pixels Pmono. After the pixel interpolation, the range estimation is performed.
When the pixel array of the optical sensor 11 is arranged as the chessboard pattern, the mobile robot 100 of the present disclosure may operate in another way to increase the frame rate of the range estimation and positioning (e.g., using VSLAM). In the aspect of
Referring to
The pixel array of the optical sensor 11 captures a first image frame, a second image frame and a third image frame respectively within the first time interval T1, the second time interval T2 and a third time interval T3 between the first time interval T1 and the second time interval T2. That is, when the pixel array of the optical sensor 11 captures the third image frame, all light sources are not turned on. In
The processor 13 performs the range estimation (e.g., including finding an obstacle and calculating a distance therefrom) according to the first image frame and the second image frame, wherein the first image frame and the second image frame are formed by pixel data generated by the plurality of first pixels PIR. That is, when the first light source LS1 as well as the second light sources LS21 and LS22 are lighted up, pixel data associated with the first pixels PIR is not influenced by other colors of light, and thus the processor 13 is arranged to perform the range estimation according to the pixel data only associated with the plurality of first pixels PIR.
In this embodiment, the third image frame is formed by pixel data generated by the plurality of second pixels Pmono.
Similarly, the processor 13 further performs the pixel differencing between the first image frame and the pixel data in the third image frame associated with the first pixels PIR, and performs the pixel differencing between the second image frame and the pixel data in the third image frame associated with the first pixels PIR so as to eliminate background noises.
Similarly, when the first pixels PIR and the second pixels Pmono are arranged in the chessboard pattern, before performing the range estimation, the processor 13 further performs the pixel interpolation on the first image frame and the second image frame to fill interpolated data at positions in the first image frame and the second image frame corresponding to the second pixels Pmono at first. Then, the range estimation is performed.
In the second embodiment, the processor 13 performs the VSLAM according to pixel data in the third image frame associated with the second pixels Pmono In this embodiment, the third light source LS3 is not lighted (e.g., the third light source LS3 may be omitted). Since the pixel data generated by the first pixels PIR exclude components outside IR spectrum, the third image frame of this embodiment is formed by pixel data generated by the plurality of second pixels Pmono. In addition, before performing the VSLAM according to the third image frame, the processor 13 further performs the pixel interpolation on the third image frame so as to fill interpolated data at positions in the third image frame corresponding to the first pixels PIR.
It is seen from
However, when ambient light is not enough, the processor 13 may not able to correctly perform the VSLAM without lighting the third light source LS3. To solve this problem, the processor 13 further identifies ambient light strength according to the third image frame, e.g. comparing with a brightness threshold. When identifying that the ambient light is weak, the processor 13 further changes the lighting timing of the first light source LS1 as well as the second light sources LS21 and LS22. For example, the processor 13 controls the lighting of light sources and the image capturing as shown in
The present disclosure further provides a mobile robot that performs the ranging estimation and obstacle recognition according to images captured by the same optical sensor 11. When identifying that one obstacle is a specific object, e.g., a wire or socks, the mobile robot 100 directly moves across the obstacle; whereas when identifying that one obstacle is an electronic device, e.g., a cell phone, the mobile robot 100 dodges the electronic device without moving across it. The obstacle that can be moved across is determined previously according to different applications.
The mobile robot 100 of this embodiment is also shown as
As mentioned above, to cancel the interference from ambient light, the optical sensor 11 further captures a first dark image frame, for differencing with the first image frame, within a first dark interval (e.g., T3 in
In this embodiment, the pixel array of the optical sensor 11 receives incident light via the light filter 15.
The processor 13 identifies an obstacle according to the first image frame and the second image frame, wherein the method of identifying the obstacle has been described above and thus details thereof are not repeated herein. After the obstacle is found, the processor 13 controls the third light source LS3 to light up within a third time interval (e.g., T3 in
In this embodiment, before appearance of the obstacle is identified by the processor 13, the third light source LS3 is not lighted up, and thus the operational timing of the mobile robot 100 is shown as
After receiving the third image frame from the optical sensor 11, the processor 13 determines a region of interest (ROI) in the third image frame according to a position of obstacle (i.e. the position of broken line), e.g., shown in
In one non-limiting aspect, the ROI has a predetermined image size. That is, when the position (e.g., center or gravity center, but not limited to) of one obstacle is determined, the processor 13 determines a region of interest having the predetermined size at the position.
In another aspect, a size of the ROI is determined by the processor 13 according to the first image frame and the second image frame. In this case, when the obstacle is larger, the ROI is larger; on the contrary, the ROI is smaller.
The processor 13 then recognizes an object type of the obstacle in the ROI using a pre-trained learning model (e.g., embedded in the processor 13 by means of ASIC or firmware). As the learning model does not recognize (e.g., not calculating convolution) rest region in the third image frame outside the ROI, the computation loading, time and power consumption are significantly reduced. Meanwhile, as the ROI contains a small number of object images, the recognition is not interfered by other object images to improve the recognition correctness.
In addition, to further improve the recognition correctness, the processor 13 further identifies a height of obstacle according to the second image frame, e.g., taking a length H of the broken line in
In one aspect, the object height is used as the learning material by the data network architecture (e.g., including neural network learning algorithm, deep learning algorithm, but not limited to) together with the ground truth image in a training phase to generate the learning model.
In another aspect, in the training phase, the data network architecture only uses the ground truth image to generate the learning model. In operation, when the learning model calculates the probability of several possible objects, the height is used to filter some possible objects. For example, if the height of one object type categorized by the learning model exceeds the height identified according to the second image frame, even though this one object type has the highest probability, the learning model still excludes this object type.
The method of categorizing the object in an image by the learning model is known to the art, and thus details thereof are not described herein. Meanwhile, the incorporation between the learning model and the object height to recognize the obstacle is not limited to that described in the present disclosure.
In one aspect, as a capturing frequency of the optical sensor 11 is higher than a moving speed of the mobile robot 100, the processor 13 further controls the first light source LS1, the second light sources LS21 and LS22, and the third light source LS3 to turn off for a predetermined time interval after the third time interval T3 (i.e. after capturing one third image frame) till the obstacle leaves the projection range of the first light source LS1. In this way, it is able to prevent repeatedly recognizing the same obstacle. The predetermined time interval is determined according to, for example, the moving speed of the mobile robot 100 and the height determined according to the second image frame.
Referring to
In this embodiment, the linear light includes, for example, the first light source LS1 as well as the second light source LS21 and LS22 mentioned above. The illumination light includes, for example, the third light source LS3 mentioned above. It is appreciated that positions of every light source shown in
Step S51: The processor 13 respectively controls the first light source LS1 as well as the second light source LS21 and LS22 to light up, for example, at the first time interval T1 and the second time interval T2 as shown in
Step S52: When identifying that the first image frame contains the broken line as shown in
When identifying that the first image frame or the second image frame contains the broken line, the processor 13 further records (e.g., in the memory) a position of broken line as the object position.
Step S53: The processor 13 then controls the third light source LS3 to turn on, e.g., at the third time interval T3 shown in
Step S54: The processor 13 then determines the ROI in the third image frame. The ROI is at the object position determined in the Step S52. As mentioned above, a size of the ROI is determined previously or determined according to a width W of the broken line in the first image frame (as shown in
Steps S55-S56: Finally, the processor 13 recognizes the object image within the ROI using the learning model trained before shipment to identify an object type.
Step S57: To increase the recognition correctness, when identifying an obstacle in the Step S52, the processor 13 further identifies an object height according to the second image frame, e.g., according to H in
After the object type is recognized, the processor 13 bypasses or dodges specific obstacles or directly moves across some obstacles according to previously determined rules. The operation after the object type being recognized is set according to different applications without particular limitations.
It should be mentioned that although the above embodiments are described in the way that the second light sources LS21 and LS22 are turned on and off together, the present disclosure is not limited thereto. In other aspects, LS21 and LS22 are turned on sequentially (and optical sensor capturing images correspondingly) as long as LS21 and LS22 respectively project a longitudinal light section toward the moving direction.
The first light source 720, the first optical sensor 730, the second light source 740 and the second optical sensor 750 are disposed in the housing 710. In this embodiment, the first light source 720 and the first optical sensor 730 are disposed at or close to a bottom surface of the housing 710. The second light source 740 and the second optical sensor 750 are disposed at a side surface of the housing 710 as shown in
In this embodiment, the first light source 720 illuminates the working surface G1 via a bottom opening of the housing 710 of the mobile robot 700, wherein the first light source 720 is an inviable light source, for example an infrared light source, e.g., infrared LED. The first optical sensor 730 has a field of view θ1 toward the bottom opening of the mobile robot 700 to sense the working surface G1 illuminated by the first light source 720 via the bottom opening using an exposure time, wherein the first optical sensor 730 is an image sensor, for example a CMOS or a CCD sensor and the exposure time thereof is determined by an auto exposure. The first optical sensor 730 (with its own processor) provides a displacement of the mobile robot 700 with respect to the working surface G1 according to the captured image frame (e.g., first image frame IF1) of the working surface G1, e.g., comparing image frames captured at different times. Or, the first optical sensor 730 provides first image frames IF1 to the processor 760, and the processor 760 calculates the displacement.
When a type of the working surface G1 is a ceramic tile, reflectivity of the gap between ceramic tiles is significantly lower than the ceramic tile, and the gap between ceramic tiles will be hard to reflect light from the first light source 720. In detail, for maintaining the brightness of the first image frame IF1 sensed by the first optical sensor 730, the exposure time of the first optical sensor 730 is increased automatically, by the auto exposure, when the mobile robot 700 goes through the gap between ceramic tiles as shown in
For solving the above issue, the mobile robot 700 of this embodiment identifies whether there is a gap between ceramic tiles or not by using the forward second optical sensor 750 and the forward second light source 740 as shown in
When the image frame (e.g., second image frame IF2) sensed by the second optical sensor 750 contains an image in
In this embodiment, since the information of the gap between ceramic tiles in front of the mobile robot 700 can be sensed by the second optical sensor 750, the exposure time of the first optical sensor 730 is rapidly shortened to a proper exposure time at time T2 as shown in
In one aspect, the processor 760 recognizes whether the second image frame IF2 contains an image of tiles, e.g., using the machine learning. Then, the processor 760 predicts a time interval, e.g., T1 to T2 shown in
In this aspect, the processor 760 controls the first optical sensor 730 to capture the first image frame IF1 during the time interval between T1 and T2 using the exposure time being used right before the time interval, i.e. E1 determined at or prior to time T1. After the time interval, i.e. at or behind time T2, the processor 760 turns on the auto exposure of the first optical sensor 730 again. In this way, the exposure time of the first optical sensor 730 is operated following a dashed line section B as shown in
In another aspect, the exposure time within the time interval between T1 and T2 is selected as slightly higher than E1, e.g., 1.5E1, 2E1, 2.5E1 or 3E1, since the first image frame IF1 captured by the first optical sensor 730 within the time interval between T1 and T2 contains a gap image having lower reflectivity.
In an alternative aspect, the processor 760 recognizes whether the second image frame IF2 contains an image of tiles, predicts a time interval, e.g., T1 to T2 shown in
Referring to
Step S111: As mentioned above, the mobile robot 700 has at least one light source, e.g., LS1, LS21 and LS22 in
Step 112: The front optical sensor 750 captures second image frames IF2 at a predetermined frame rate.
Step 113: The processor 760 receives the second image frames IF2 from the front optical sensor 750 and recognizes whether the second image frames IF2 contain an image of tiles, e.g., using model trained by machine learning.
Step S114: If the second image frames IF2 are recognized to contain an image of tiles, the processor 760 predicts a time interval, e.g., T1 to T2 shown in
Step S115: In the time interval between T1 and T2, if the auto exposure is used by the bottom optical sensor 730, the exposure time is sharply increased as a curve A shown in
The Steps S111 to S115 in
In one aspect, the processor 760 controls the bottom optical sensor 730 to capture a bottom image frame during the time interval between T1 and T2 and at an end of the time interval, e.g., at or after time T2 using the exposure time, e.g., E1, being used right before the time interval, e.g., at or before time T1. Within the time interval between T1 and T2, the auto exposure of the bottom optical sensor 730 is closed or deactivated.
In another aspect, the processor 760 records the exposure time, e.g., E1, being used by the bottom optical sensor 730 before the time interval, e.g., at or before time T1, in a memory. Then, the processor 760 controls the bottom optical sensor 730 to capture a bottom image frame at an end of the time interval, i.e. at or after time T2, using the recorded exposure time. In this aspect, the bottom optical sensor 730 still captures, within the time interval between T1 and T2, the bottom image frame using an exposure time determined by the auto exposure.
In the present disclosure, the auto exposure of the first optical sensor 730 is turned off and turned on according to whether a gap between tiles will be captured by the first optical sensor 730. However, the second optical sensor 750 is not controlled to turned on/off the auto exposure thereof.
It should be mentioned that although the above embodiment is illustrated by using the ceramic tile as an example, the present disclosure is not limited thereto. The tiles may have other materials and are not limited to ceramic tiles.
In addition, a number of first light source, the second light source and the third light source is not limited to those shown in
In the present disclosure, the “transverse” is referred to substantially parallel to a moving surface (e.g., the ground), and the “longitudinal” is referred to substantially perpendicular to the moving surface. The object on the moving path is called the obstacle.
As mentioned above, the conventional cleaning robot adopts multiple types of sensors to respectively implement different detecting functions, and has the issues of high computation loading, time and consumption power as well as low recognition correctness. Accordingly, the present disclosure further provides a mobile robot suitable to smart home (e.g.
Although the disclosure has been explained in relation to its preferred embodiment, it is not used to limit the disclosure. It is to be understood that many other possible modifications and variations can be made by those skilled in the art without departing from the spirit and scope of the disclosure as hereinafter claimed.
The present application is a continuation application of U.S. patent application Ser. No. 17/064,776 filed on Oct. 7, 2020, which is a continuation-in-part application of U.S. patent application Ser. No. 16/929,232 filed on Jul. 15, 2020, which is a continuation-in-part application of U.S. patent application Ser. No. 16/425,955 filed on May 30, 2019, which is a continuation-in-part application of U.S. patent application Ser. No. 15/841,376 filed on Dec. 14, 2017, which claims the priority benefit of U.S. Provisional Application Ser. No. 62/514,349, filed on Jun. 2, 2017, the disclosures of which are hereby incorporated by reference herein in their entirety. To the extent any amendments, characterizations, or other assertions previously made (in this or in any related patent applications or patents, including any parent, sibling, or child) with respect to any art, prior or otherwise, could be construed as a disclaimer of any subject matter supported by the present disclosure of this application, Applicant hereby rescinds and retracts such disclaimer. Applicant also respectfully submits that any prior art previously considered in any related patent applications or patents, including any parent, sibling, or child, may need to be re-visited.
Number | Date | Country | |
---|---|---|---|
62514349 | Jun 2017 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 17064776 | Oct 2020 | US |
Child | 18226503 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16929232 | Jul 2020 | US |
Child | 17064776 | US | |
Parent | 16425955 | May 2019 | US |
Child | 16929232 | US | |
Parent | 15841376 | Dec 2017 | US |
Child | 16425955 | US |