The present disclosure and claimed embodiments relate to a mobile wheeled robotic vehicle designed to navigate indoor environments, and, in particular, a vehicle acting as an intelligent cleaning robot, and controlling methods therefor.
Examples of mobile cleaning robots, and, in particular, robots for use in the home, are well known in the art. Current models of robots are able to traverse a room to perform vacuum operations on carpeted surfaces or mopping operations on hard surfaces. Simple embodiments of home cleaning robots employ bump sensors and/or proximity sensors for obstacle avoidance. More advanced embodiments employ a single camera or other optical sensor for mapping an environment and determining location of the robot within the environment, typically utilizing a visual SLAM (simultaneous localization and mapping) method.
Several problems exist with even the more advanced embodiments of home cleaning robots. One problem is that they are unable to detect obstacles on the ground, such as to perform obstacle avoidance, without engaging with the obstacle, for example, by running into the obstacle and engaging a bump sensor on the device or utilizing a proximity sensor. Thus, the navigational capabilities and mapping capabilities of current state-of-the-art home cleaning robots is extremely limited. Given a region, a typical home cleaning robot will traverse the region in such a manner that some areas of the region will be covered multiple times while other areas of the region will be completely missed.
Described herein is an intelligent autonomous cleaning robot for domestic use. The robot autonomously maps and cleans multiple rooms in a house in an intelligent manner. The described embodiments include hardware and software for sensing the environment, planning actions, and executing actions.
The intelligent cleaning robot uses stereo cameras with a static projected light pattern to generate 3D data used for the purposes of mapping, localization, and obstacle avoidance. The system also uses optical sensors in various locations around the robot for additional sensing for mapping, localization, and obstacle avoidance. One embodiment of the robot uses laser light striping paired with a vertically positioned camera at the front corners of the robot facing backwards (similar to side mirrors in a car) for additional 3D data for mapping, localization, and obstacle avoidance. Another embodiment of the robot uses a downward facing camera housed within the robot that tracks features on the ground, providing visual odometry from which a motion estimate that is independent of wheel slip may be derived.
The front facing stereo camera uses a static projected light pattern that is pulsed and can alternate with a uniform light pattern. The front facing stereo camera can also track features to perform visual odometry and provide a motion estimate independent of wheel slip. The system also uses multiple time-of-flight (ToF) laser distance sensors oriented in various directions for cliff detection, wall following, obstacle detection, and general perception.
The base station for the intelligent cleaning robot uses a beacon to allow the robot to identify the bearing and presence of the base station. IR (infrared) illuminated patterns on the base station allow the robot to identify orientation and location of the base station, which allows for autonomous docking. The orientation and location are also used for mapping and localization. The robot can also use patterns on the base station and/or the known physical shape of the base station for the purposes of calibrating the stereo cameras.
The term microcontroller, as used herein, may mean a dedicated hardware device, circuitry, an ASIC, an FPGA, a single or distributed microprocessor(s) running software or any other means known in the art. It is further understood that the microcontroller will include connections to sensors, active light sources, motor controllers and audio/visual output devices for receiving data, sending control signals and providing user feedback. The invention is not intended to be limited to one method of implementing the functions of the controller.
As used herein, the terms camera and sensor are used interchangeably.
A robot suitable for domestic or commercial floor cleaning functions is described. The robot body may comprise a chassis having a longitudinal axis in a fore/aft direction and a transverse axis oriented perpendicular to the longitudinal axis. The body, from a top view, may be round in shape, square or rectangle or in shape, or a combination of both. For example, the front of the robot may be squared off, while the rear of the robot is rounded. Preferably the profile of the robot will be as low as possible to allow it to navigate under objects, such as furniture. The robot will be fitted with a motive system for generally moving the robot along the longitudinal axis. The motive system may comprise, for example, motor driven wheels mounted on opposite sides of the robot and a caster wheel supporting the front of the robot. In alternate embodiments, the motive system may comprise treads mounted on opposite sides of the robot (also referred to herein as wheels). The robot may be fitted with a vacuum system and one or more rotating brushes. The brushes may be oriented such as to rotate along an axis parallel to the transverse axis or may be oriented to rotate about an axis orthogonal to both the longitudinal axis and the transverse axis. The robot may be fitted with a series of sensors of various types, as described herein, for performing localization, object detection, mapping, etc. A microcontroller and memory storing software and mapping data may be provided. The microcontroller may control the robot in accordance with instructions implemented in the software, including, for example, controlling the motive system, determining routes, reading and processing data from the sensors, etc.
Stereo cameras 104 are positioned at the front of the robot, as shown in
In a preferred embodiment, multiple camera types can be used in in the stereo configuration. Preferably stereo cameras 104 are RGB/IR imagers having pixels that are sensitive to only visible (red, green and blue specific pixels) spectrum and also pixels which are sensitive to only infrared (IR specific pixels) spectrum, to allow the projection of a pattern in infrared such as not to be visible to the naked eye. The RGB/IR type of camera allows the capture of color (RGB) and IR data simultaneously. Preferably the infrared pixels will be used for 3D mapping while the RGB pixels will be used for visual odometry and other techniques which rely on visible light, for example, place recognition, object rate recognition, object classification and floor type recognition.
In alternate embodiments, the stereo cameras 104 may also be monochrome, having pixels that are sensitive to visible and IR, but unable to discriminate between both. Additional filtering may be used to limit the sensitivity to only visible or IR. Color cameras may also be used, which have RGB sensitive pixels. In some cases, the pixels in the color camera can also be sensitive in the interim infrared spectrum.
In a preferred embodiment, one or more laser modules are used as the active stereo illuminator 105 to project the random pattern. Preferably, the one or more lasers (with diffractive optical elements) are oriented such that their combined fields of view cover an entire 120° field of view. In preferred embodiments, two lasers with 60° fields of view are used. The overlapping patterns are shown in
In some embodiments, the robot may also be equipped with a visible and/or an infrared light source 114 to illuminate the scene. Preferably, the visible and infrared light sources would be white and IR LEDs respectively. The visible infrared light sources may be housed in close proximity to the stereo cameras 104 and active stereo illuminator 105 and may be positioned to point in a generally forward direction.
Several strategies exist for cycling both the active stereo illuminator 105 and the visible and infrared illuminators 114. In one embodiment, the dot pattern projection and diffuse IR illumination may be alternated. The dot pattern augments the scene by adding texture to the scene, in the form of the random dot pattern. The diffuse IR illuminator illuminates the scene without adding features that move with the robot, which may mislead visual odometry or tracking algorithms. In an alternate embodiment the diffuse visible/infrared illuminator 114 would be kept on when needed, for example, in a dark environment or under furniture, and the pattern projector 105 would be cycled on and off with a specific pattern, for example, odd or even video frames. In yet another embodiment, the diffuse visible/infrared illuminator 114 is kept on when needed and the pattern projector 105 is cycled on or off as needed (as determined by the software algorithm which controls the pattern projector) instead of in a repeating pattern.
Preferably, stereo cameras 104 and pattern projectors 105 are mounted in a housing which is positioned at the front of the robot, as shown in
The cleaning robot may come in different form factors, as shown in
Preferably two of the range sensors 106 form a crisscrossed pair at the front of the robot. These sensors maybe mounted in the same housing as stereo cameras 104 and active stereo illuminator 105 and their fields-of-view are shown in
An additional pair of rangefinders having fields-of-view shown in
An optional third pair of rangefinders, shown by having fields-of-view shown in
In some embodiments, the robot may also optionally be provided with a proximity sensor 116 for cliff detection. The proximity sensor 116, as shown in
In other embodiments, the robot may also optionally be provided with side cameras, as shown in
Preferably, the robot will be equipped with an inertial measurement unit (IMU) 108 to measure the motion of the robot inertially, consisting of at least 6-axes, 3-axes of gyros to measure rotation and 3-axes of accelerometers to measure translation. IMU 108 provides can be used to provide an estimate of 3D acceleration, 3D rate of rotation, and 3D magnetic fields, and elevation. This is used for estimating attitude/orientation, and for estimating wheel slip. Optionally, an additional 3-axes magnetometer 112 may be used to measure the magnetic field. Magnetometer 112 may be, in some embodiments, a Hall Effect or magnetic strip sensors and may be used to indicate the presence of a magnetic field. This is useful for identifying the presence of a magnetic strip used to partition areas where the robot should not navigate.
Additional embodiments include wheel encoders 110 to measure the translational speed of the robot. In preferred embodiments, each wheel may be equipped with an encoder to maximize accuracy and to allow measurement of direction by the difference in the wheel speeds. Wheel encoders 110 may provide signed (typically quadrature) or unsigned tick count of wheel or wheel motor rotation, which is useful for calculating estimated translational velocity and for determining wheel slip.
Additional sensor data and data generated by a low-level controller are used as inputs to the navigation module 100.
Bump sensors 113 provide data indicating when the robot bumper depresses, usually as a result of contacting an object.
Cliff sensors 116 provide data indicating a vertical drop that could be hazardous to the robot. Cliff sensors 116 can be time-of-flight (TOF) or proximity sensors. Cliff detection events are used as input to the robot's map, so the robot does not repeatedly go near cliffs.
Wheel drop sensors 118 provide data indicating if a wheel (typically on a spring-loaded suspension) moves away from the robot body, typically indicating that the wheel is no longer in contact with the ground.
Robot button status 120 indicates what button the user has pressed on the robot and may be used to change the state of the robot, for example, from idle to cleaning, or cleaning to dock.
Battery voltage data 122 provides data regarding the charge level of the robot's battery. This is used to trigger return to dock behavior when the level drops below a specified threshold. Additionally, the level can be combined with an estimate of charge needed to traverse the path back to the dock.
Current data 124 provides information regarding the electrical current used by the suction, brush, and side brush motors. The brush current and side brush motor current can be used, in some embodiments, to estimate the type of floor the robot is on. Current data 124 can also be used to estimate the amount of slip the wheels could experience by correlating current with floor type and slip models.
Charge presence data 126 indicates if the robot is on the docking station, which provides charging for the battery. This is useful for detecting that the robot has successfully docked for the purposes of charging.
IR dock receiver data 128 provides an indication of the presence of an encoded IR signal emitted from the dock. This is used to help approximate the dock's location when the robot is navigating to the dock.
Commanded wheel speed data 130 comprises desired wheel speed commands 146 sent by the navigation module to a lower-level controller for a desired speed.
Low-level error data 132 indicates if a low-level controller has detected any errors, which may include, but is not limited to, critical battery level, over-current conditions, cliff detection events and wheel drop events.
Light ring 134 is a part of the user interface of the robot and is used to communicate status to the user.
Cliff escape intent data 136 allows a low-level module to communicate intended motions for escaping a cliff event to the navigation module 100.
Over the Air (OTA) data 138 is a data package which may be received wirelessly that can be used to update the low-level controller and/or the navigation module software.
Commands from App 140 allows the user to send commands to the robot, which may include, but are not limited to clean, dock, or stop, in addition to setting cleaning aggressiveness, querying for robot location, or querying for robot map. Commends from App 140 may be generated by an App running on a mobile device.
Robot map data and cleaning report data 142 provides grid map and/or vectors describing the environment previously mapped, cleaned, and navigated through by the robot. Robot map data and cleaning report data 142 may also include statistics on how much area was cleaned.
Robot autonomous status 144 indicates if the robot is cleaning, exploring, idle, docking, etc.
Desired vacuum, brush, and side brush mode data 148 allows navigation module 100 to request that a low-level controller turn on/off or modify the intensity of the vacuum, brush, and/or side brush.
Desired audio 150 allows navigation module 100 to request a low-level module playback audio over the robot's speaker.
The calibration correction module 204 uses images 104 and disparity 107 of the dock to correct potential errors in calibration. The module can use disparity data 107 to determine if the dock (which has known dimensions) has the expected dimensions sensed from the disparity information. If the error is greater than a threshold, then the module can use the images to adjust the principal point of each stereo imager to optimize correlation of feature points and to match the expected geometry of the dock. This calibration/adjustment could also occur whenever the dock is visible irrespective of any error detected.
Wheel drop sensors 118 and current sensors 124 allow the navigation module 100 to determine, at 402, if it is beached or stuck on an object. One way of detection is to observe when motor currents from the wheels goes very high for a longer than normal period of time. Once detected, navigation module 100 performs an escape motion to move away from the suspected location of the object. Additionally, navigation module 100 can add the location of the object as a potential obstacle to the robot's map. The module uses the location of the object based on the beginning of the interaction.
Local path planning module 404 uses local grid maps 193. Regional path planning module 406 uses regional maps 194 and local maps 193. Global path planning module 408 uses global maps 195, regional maps 194, and local maps 193.
Once a desired path 413 is created by a module, the motion engine module 414 sends wheel speeds commands 146 to attempt to get the robot to follow the desired path while using a local scrolling map to avoid obstacles.
A tight spaces module 416 detects when the robot is in confined space. This typically occurs when the path planner cannot find a free path away from a small area. The module proceeds to use the bumper sensors 113 and rangefinders 106 to find a way out, interpreting the local map as untrustworthy.
A perimeter behavior module 418 controls the robot motion to go around the perimeter of a region to map the region. The behavior tends to be conservative when avoiding obstacles or following walls. This generally helps minimize interacting with obstacles. The perimeter behavior attempts to complete a loop of the region. When making a perimeter, the module attempts to make a perimeter around only areas that have not been previous cleaned. The module limits the size of the region it makes a perimeter around to a maximum dimension. The perimeter behavior attempts to identify the type of surface the robot is on more often than when cleaning. The perimeter behavior attempts to use natural barriers (like walls or doorways) to help define the perimeter it creates.
The cleaning behavior module 420 controls the robot motion when cleaning a region. The robot attempts to clean an entire region before exploring to find other regions to clean. The cleaning behavior module 420 attempts to create a cleaning pattern consisting of parallel lines that are also parallel/perpendicular to the area's principal direction (result is typically seeing cleaning paths that are parallel to walls). If the robot cannot get to an area it thinks it should be able to clean (typically due to bumper hits, cliff detection, or magnetic strip), then it marks the area as inaccessible and no longer attempts to clean the area.
The lost behavior module 422 detects when the robot is lost and controls the robot motion when lost. When lost, the robot will first attempt to re-establish positioning/localization in place. Next, the robot will execute a path that moves the robot around to try and re-establish positioning/localization. Finally, the lost behavior module 422 will switch the robot to a behavior that attempts to re-explore the environment to try and find the dock and then return to the dock.
The exploration and navigation module 424 directs the robot to areas that have not been cleaned or had a perimeter created.
The return to dock and docking module 426 monitors the battery voltage 122 to determine if the robot should return to dock based on a simple voltage threshold, or by periodically planning a path back to the dock and estimating the amount of charge needed to return to dock to trigger a return to dock behavior. The return to dock and docking module 426 will plan a path back to the dock and re-plan if needed. If the dock is not in the expected location, the return to dock and docking module 426 will control the robot to explore in the area nearby the expected location to try and acquire the dock position (usually using the IR dock receivers). Whether from IR dock receivers or the dock being in the expected location, the return to dock and docking module 426 turns off the projected laser projector 105 and uses the IR LED illumination 114 to illuminate the dock, which preferably has a retroreflective tag behind IR transparent material. The return to dock and docking module 426 then uses features from the tag to determine the dock location, which is used for guiding the robot back to the dock. The robot stops attempting to dock once the charge presence 126 is detected and consistent.
The docking station 800 shown in
In a preferred embodiment, images of the pattern 802 may be taken by either or both cameras 104 during the docking process and may be used by the robot to calculate the extent of the misalignment of the cameras, which may then be compensated for by factoring in the misalignment when processing images from the cameras 104. In a preferred embodiment, the calculation of the compensation for the misalignment may be performed while the robot is charging, based on images taken will the robot is approaching the dock.
In alternative embodiments, the docking station 800 may be passive and recognizable to the robot via scanning of a logo which appears on the docking station. The logo may be recognized utilizing a sophisticated detection matching algorithm and the logo must have salient features, for example, corners. To make the pattern easily recognizable and visible to the robot, the pattern can have a backlight or be retroreflective. As an added feature, the pattern could be covered by an IR transparent material, make the pattern in visible to a human, but visible to the robot in the IR spectrum. The algorithms for recognizing the pattern are commonly available.
Obstacle detection, performed by obstacle detection module 1102, can be accomplished with 2D and 3D data and can be calculated with data collected from the stereo camera 104 or from ToF rangefinders 106. The obstacle detection module 1102 can be used to perform the identification of small objects, for example, cords, Legos, pet messes, etc. through the use of color and shape. Machine learning may be used to perform the identification. The system can then use room designations to vary detection thresholds. For example, in the kitchen, a small yellow object may be identified as a Cheerio, whereas a small yellow object in the bedroom may be classified as a Lego. The system may adjust “shyness” or “cleaning aggressiveness” based on a user setting or which may be defined automatically by predefined room types. The cleaning aggressiveness setting would indicate which objects should be picked up by robot 100 or left as they are and avoided. The cleaning aggressiveness setting could also be used beyond obstacle classification to include other features, for example, modifying cleaning patterns and strategies. For example, in the kitchen or entryway, robot 100 may overlap cleaning areas to pick up more dirt, while in other areas there may be little overlap to improve efficiency. The aggressiveness setting could also be used to determine, for example, to run the brush roll at a higher speed or increase suction power.
Mapping module 1104, as shown in
Wheel odometry, which estimates position by measuring how far the drive wheels turn, is a common way to estimate translation and/or rotational movement of a vehicle, and, in particular, is especially effective with a differentially driven vehicle. Various surfaces may, however, cause this measurement to be an inaccurate representation of the robot's pose (translation and rotation). Inaccuracies typically result from wheel slip. For example, on a thick plush rug, the wheel odometry may indicate a translation of the robot 1.4 m, where the actual translation was only 1.0 m. Slip can be defined as the difference between the distance thought to be traveled by the wheels, based on odometry readings, and the actual distance traveled by the robot. Slip typically occurs on carpets or rugs were additional resistance occurs due to the caster wheel, the robot body, or the wheel interaction, typically from the sides when turning in place. Slip may also be caused by the wheels needing to move the fibers of the carpet or rug. It has been that found that slip on carpets or rugs typically has directionality. That is, the slip ratio changes depending on the direction of travel. Further, slip ratios vary between carpet types. The interaction of the carpet fibers with the wheels is different between carpet types. Independent of linear slip, some carpets have high lateral resistance while others have a much lower lateral resistance.
The wheel encoders 110, providing wheel odometry, are the primary sensor for measuring linear motion, whereas the IMU 108 is used primarily for orientation and heading. Slip results in an inaccurate estimate of linear motion when using the wheel encoders. The inaccurate estimate of linear motion leads to an inaccurate pose estimation and therefore inaccurate maps, positioning, motion control and obstacle avoidance.
One way to mitigate the effects of slip is to use other estimates of linear motion in addition to the wheel odometry. Scan matching matches a point cloud 199 generated from the stereo data to prior point clouds, typically lagged between 100 ms and 2,000 ms, or an occupancy map to determine a transformation between the current position and the local map, which may serve as a linear motion estimate. After mapping, localization may provide a position estimate that helps mitigate the error from slip. To allow for compensation of slip, the estimate of how much the robot is slipping is necessary to correct the motion estimate. Scan matching should work well independent of the surface or carpet type but requires suitable data to match against. This means that the method may not work when the robot is very close to walls or in wide open featureless spaces.
Special motions may also be used to measure the slip at 1206. Spinning the robot in place allows a differential measurement of the rate of turn from the IMU 108 and the rate of turn estimated by the wheels. Further, comparing the IMU 108 results with the wheel turn estimate yields both an estimate of slip as well as the directionality of the slip, by correlating the slip to heading. A rotational slip model with directionality may be transformed into a linear slip model with directionality. An advantage of this method of estimating slip is that it will work in almost any location relatively quickly. However, the method may not be wholly effective across various carpet types because some carpet types do not yield observable slip when turning. There may be significant potential variation translating rotational slip to linear slip.
Turning the robot in a figure-8 or turning on one wheel at various speeds is another special motion that may be used to measure slip. Spinning in a figure-8 or around one wheel may yield a more accurate estimate of the slip and directionality across carpet types, using the same principles as spinning in place. However, this method requires more space to operate and takes potentially more time. Aesthetically, the motion is also more noticeable by a human.
Because the robot builds a map of the environment, the robot can also build a map of the slip models using any of the previously mentioned approaches. The result is that, if the robot is in an area that already has a slip estimate, the robot does not need to perform any additional maneuvers or matching to account for slip. However, at some point the robot would still be required to estimate the slip and must localize the estimate of the slip.
A scaling factor may be applied to the wheel odometry values to correct for slip. The scaling factor can be estimated by one of the methods mentioned above. Alternatively, the scaling factor could be continuously estimated while driving on the surface using, for example, an EKF. A surface estimator sensor can be used to differentiate when the robot is on a new or different surface. As an example, measuring the current used by the brush to detect when the robot is on a different thickness of rug or on hardwood would be one method of determining when the robot transitions to a different surface. The method could also use visual or 3D sensors to make this determination. Surface types could be recorded on the map of the environment and the scaling factors for each surface could be tracked for different surfaces in different parts of the map. The vehicle may continuously reference updated scaling factors from the map.
Because the robot builds a map of the environment, the robot can also build a map of the slip models using any of the above approaches. The result is that if the disrobot is in an area that already has a slip estimate, the robot does not need to perform additional maneuvers or matching to account for slip. The map of slip models may be mapped to the local map 193, the regional map 194 or the global map 195 at 1212.
In the case of using an MCL particle filter, the algorithm also has some features like outlier rejection, where, after localization/tracking is established, measurements that are more different than a set threshold are not used in the scoring process. Another feature for the MCL process is that when tracking is lost, the EKF location of the robot (driven completely by local pose once tracking in the MCL is lost) seeds the guess poses used by the particle filter to establish tracking and localization again. Tracking is determined as lost once the score from the best particle drops below a defined threshold. Tracking is determined as established once the iterations for initialization complete and the best score is above a defined threshold. Outlier rejection is typically not used during initialization.
If the robot is not performing a perimeter map of the space, the module attempts to align the snapshot of the local map 193 to the current regional map 194. If alignment succeeds, then the module updates the current regional map 194. If the update is in a previously visited region and the robot has entered the region recently, the alignment is used to create an additional constraint to the global graph between the region just recently exited and the region the module successfully aligned and updated. The global graph is then updated and the global map 195 rebuilt by combining the regional maps 194 using the updated global graph information.
In some embodiments, the module can also match entire regions to prior regions to create constraints for the global graph.
In other embodiments, the module also calculates a limit for the size of area explored during the perimeter mapping phase. The limit is used by behaviors to ensure that the robot perimeter stays within a predetermined size.
In yet other embodiments, the module also calculates a combined limit for exploration combining size limits along with areas already cleaned so that behaviors explore a perimeter of an area while minimizing overlap with areas already cleaned.
The point cloud 199 is segmented into points below 199a and above 199b a height threshold. The low-height point cloud 199a is used in conjunction with the TOF sensor data 106 in a ray tracer 1502 that updates both free and occupied cells in occupancy maps (at least one scrolling map and one local map). The high-height point cloud is used by a ray tracer 1504 to update an occupancy map used for a high-height local map 1506. Both ray trace modules model the noise for the sensors when updating the occupancy map to account for uncertainty from the different sensors. Bump sensors 113, cliff sensors 116, and Hall Effect sensors 116 (detect magnetic strip boundary) are also used to update the local scrolling map 1508 and low-height local map 1510 based on per sensor models that estimate uncertainty.
Both height maps can be combined to a single local map 1512 that represents both sets of information.
In a preferred embodiment, if the robot is trying to navigate to an area, but is unable reach the area, the area is marked as inaccessible in the robot maps, so the robot does not attempt to repeatedly try and clean the area it cannot get to.
An example of a global map as shown in
In some embodiments, maps may include, at varying levels of resolution, additional information, for example, information regarding occupancy confidence, per sensor occupancy confidence, virtual walls, region designation, clean and/or covered status, dirt status, floor type, risk for getting stuck, and room labels. It should be realized that additional information could also be included within the map. In other embodiments, maps may also include a path defined as a sequence of locations and uncertainties between locations. This is different than a grid defining the robot path.
To those skilled in the art to which the invention relates, many modifications and adaptations of the invention will suggest themselves. Implementations provided herein, including implementations using various components or arrangements of components should be considered exemplary only and are not meant to limit the invention in any way. As one of skill in the art would realize, many variations on implementations discussed herein which fall within the scope of the invention are possible. Accordingly, the exemplary methods and apparatuses disclosed herein are not to be taken as limitations on the invention, but as an illustration thereof.
This application is a national phase filing under 35 U.S.C. § 371 claiming the benefit of and priority to International Patent Application No. PCT/US2018/066517, filed on Dec. 19, 2018, which claims the benefit of U.S. Provisional Patent Application Ser. No. 62/607,775, filed Dec. 19, 2017, U.S. Provisional Patent Application Ser. No. 62/620,898, filed Jan. 23, 2018, U.S. Provisional Patent Application Ser. No. 62/702,656, filed Jul. 24, 2018 and U.S. Provisional Patent Application Ser. No. 62/752,052, filed Oct. 29, 2018. The entire contents of these applications are incorporated herein by reference.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US2018/066517 | 12/19/2018 | WO |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2019/126332 | 6/27/2019 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
5040116 | Evans, Jr. | Aug 1991 | A |
5187663 | Kamimura | Feb 1993 | A |
5461478 | Sakakibara | Oct 1995 | A |
5794166 | Bauer | Aug 1998 | A |
6629028 | Paromtchik | Sep 2003 | B2 |
6732826 | Song et al. | May 2004 | B2 |
7162056 | Burl et al. | Jan 2007 | B2 |
7164118 | Anderson | Jan 2007 | B2 |
7177737 | Karlsson et al. | Feb 2007 | B2 |
7206677 | Hulden | Apr 2007 | B2 |
7239105 | Lim et al. | Jul 2007 | B2 |
7386163 | Sabe | Jun 2008 | B2 |
7539557 | Yamauchi | May 2009 | B2 |
7801645 | Taylor | Sep 2010 | B2 |
7860301 | Se et al. | Dec 2010 | B2 |
7963736 | Takizawa | Jun 2011 | B2 |
8386081 | Landry | Feb 2013 | B2 |
8428778 | Landry | Apr 2013 | B2 |
8463018 | Chung et al. | Jun 2013 | B2 |
8508388 | Karlsson et al. | Aug 2013 | B2 |
8532368 | Se et al. | Sep 2013 | B2 |
8577538 | Lenser et al. | Nov 2013 | B2 |
8788092 | Casey | Jul 2014 | B2 |
8798840 | Fong | Aug 2014 | B2 |
8830091 | Karlsson et al. | Sep 2014 | B2 |
8874263 | Kwak et al. | Oct 2014 | B2 |
8874264 | Chiappetta et al. | Oct 2014 | B1 |
8880271 | Jeon | Nov 2014 | B2 |
8897947 | Nakano | Nov 2014 | B2 |
8970693 | Chang | Mar 2015 | B1 |
8996292 | Park | Mar 2015 | B2 |
9002511 | Hickerson | Apr 2015 | B1 |
9020641 | Jeong et al. | Apr 2015 | B2 |
9104204 | Jones | Aug 2015 | B2 |
9218003 | Fong et al. | Dec 2015 | B2 |
9250081 | Gutmann et al. | Feb 2016 | B2 |
9286810 | Eade | Mar 2016 | B2 |
9339163 | Noh et al. | May 2016 | B2 |
9404756 | Fong et al. | Aug 2016 | B2 |
9453743 | Lee et al. | Sep 2016 | B2 |
9456725 | Kim et al. | Oct 2016 | B2 |
9508235 | Suessemilch | Nov 2016 | B2 |
9517767 | Kentley | Dec 2016 | B1 |
9519289 | Munich et al. | Dec 2016 | B2 |
9582889 | Shpunt et al. | Feb 2017 | B2 |
9632505 | Hickerson | Apr 2017 | B2 |
9701239 | Kentley | Jul 2017 | B2 |
9785148 | Yun | Oct 2017 | B2 |
9840003 | Szatmary | Dec 2017 | B2 |
10025886 | Rublee | Jul 2018 | B1 |
10089778 | Moule | Oct 2018 | B2 |
10325336 | Kotake | Jun 2019 | B2 |
10422648 | Afrouzi | Sep 2019 | B2 |
10609862 | Wu | Apr 2020 | B2 |
10612929 | Afrouzi | Apr 2020 | B2 |
10732639 | Palanisamy | Aug 2020 | B2 |
10935383 | Ebrahimi Afrouzi | Mar 2021 | B1 |
20040148057 | Breed | Jul 2004 | A1 |
20050000543 | Taylor | Jan 2005 | A1 |
20050010330 | Abramson | Jan 2005 | A1 |
20050055792 | Kisela | Mar 2005 | A1 |
20050065662 | Reindle | Mar 2005 | A1 |
20050166355 | Tani | Aug 2005 | A1 |
20050182518 | Karlsson | Aug 2005 | A1 |
20050251292 | Casey | Nov 2005 | A1 |
20050253555 | Lee | Nov 2005 | A1 |
20050278888 | Reindle | Dec 2005 | A1 |
20060012493 | Karlsson | Jan 2006 | A1 |
20060025888 | Gutmann | Feb 2006 | A1 |
20060191096 | Sudo | Aug 2006 | A1 |
20060237037 | Kim | Oct 2006 | A1 |
20060293794 | Harwig | Dec 2006 | A1 |
20070042716 | Goodall | Feb 2007 | A1 |
20070156286 | Yamauchi | Jul 2007 | A1 |
20070192910 | Vu | Aug 2007 | A1 |
20070201863 | Wilson | Aug 2007 | A1 |
20070234492 | Svendsen | Oct 2007 | A1 |
20070257910 | Gutmann | Nov 2007 | A1 |
20080015738 | Casey | Jan 2008 | A1 |
20080027591 | Lenser | Jan 2008 | A1 |
20080065266 | Kim | Mar 2008 | A1 |
20080088702 | Linsenmaier | Apr 2008 | A1 |
20080091305 | Svendsen et al. | Apr 2008 | A1 |
20080151233 | Blanke | Jun 2008 | A1 |
20080273791 | Lee | Nov 2008 | A1 |
20090082879 | Dooley | Mar 2009 | A1 |
20090194137 | Friedman | Aug 2009 | A1 |
20090252580 | Takizawa | Oct 2009 | A1 |
20090276092 | Yoon | Nov 2009 | A1 |
20090281661 | Dooley | Nov 2009 | A1 |
20100049365 | Jones | Feb 2010 | A1 |
20100082193 | Chiappetta | Apr 2010 | A1 |
20100152945 | Park | Jun 2010 | A1 |
20100217528 | Sato | Aug 2010 | A1 |
20100257691 | Jones | Oct 2010 | A1 |
20110054689 | Nielsen | Mar 2011 | A1 |
20110166737 | Tanaka | Jul 2011 | A1 |
20110202175 | Romanov | Aug 2011 | A1 |
20110204209 | Barrows | Aug 2011 | A1 |
20110316695 | Li | Dec 2011 | A1 |
20120095619 | Pack | Apr 2012 | A1 |
20120121161 | Eade | May 2012 | A1 |
20120125363 | Kim | May 2012 | A1 |
20120185094 | Rosenstein et al. | Jul 2012 | A1 |
20120197464 | Wang | Aug 2012 | A1 |
20120247510 | Chen | Oct 2012 | A1 |
20130204483 | Sung | Aug 2013 | A1 |
20130325244 | Wang | Dec 2013 | A1 |
20140043309 | Go | Feb 2014 | A1 |
20140207282 | Angle | Jul 2014 | A1 |
20140247354 | Knudsen | Sep 2014 | A1 |
20140350839 | Pack | Nov 2014 | A1 |
20150000068 | Tsuboi | Jan 2015 | A1 |
20150052703 | Lee | Feb 2015 | A1 |
20150157182 | Noh | Jun 2015 | A1 |
20150290795 | Oleynik | Oct 2015 | A1 |
20150375395 | Kwon | Dec 2015 | A1 |
20160059412 | Oleynik | Mar 2016 | A1 |
20160070268 | Ko | Mar 2016 | A1 |
20160147230 | Munich et al. | May 2016 | A1 |
20160148433 | Petrovskaya | May 2016 | A1 |
20160158942 | Augenbraun | Jun 2016 | A1 |
20160188977 | Kearns | Jun 2016 | A1 |
20160375592 | Szatmary | Dec 2016 | A1 |
20160378117 | Szatmary | Dec 2016 | A1 |
20170010623 | Tang | Jan 2017 | A1 |
20170036349 | Dubrovsky et al. | Feb 2017 | A1 |
20170039756 | Moule | Feb 2017 | A1 |
20170105592 | Fong et al. | Apr 2017 | A1 |
20180125003 | Wu | May 2018 | A1 |
20180246520 | Martinson | Aug 2018 | A1 |
20190114798 | Afrouzi | Apr 2019 | A1 |
20190120633 | Afrouzi | Apr 2019 | A1 |
Number | Date | Country |
---|---|---|
103033189 | Apr 2013 | CN |
2013235870 | Nov 2013 | JP |
WO-2005055796 | Jun 2005 | WO |
WO-2019126332 | Jun 2019 | WO |
Entry |
---|
International Search Report and Written Opinion for International application No. PCT/US2018/066517, mailed on May 29, 2019, 12 pages. |
International Preliminary Report on Patentability for International application No. PCT/US2018/066517, mailed on May 29, 2019, 9 pages. |
Number | Date | Country | |
---|---|---|---|
20200409382 A1 | Dec 2020 | US |
Number | Date | Country | |
---|---|---|---|
62752052 | Oct 2018 | US | |
62702656 | Jul 2018 | US | |
62620898 | Jan 2018 | US | |
62607775 | Dec 2017 | US |