This document relates to autonomous driving systems. In particular, described herein are systems and methods for providing visual alerts to vehicles following an autonomous vehicle as well as to other road users sharing environment with the autonomous vehicle.
Self-driving or autonomous vehicles can be autonomously controlled to navigate along a path to a destination. Autonomous driving generally requires sensors and processing systems that take in the environment surrounding an autonomous vehicle and make decisions that ensure the safety of the autonomous vehicle and surrounding vehicles as well as other objects, both moving and stationary, around the autonomous vehicle. For example, these sensors include cameras and light detection and ranging (LiDAR) sensors that use light pulses to measure distances to various objects surrounding the autonomous vehicle.
Systems and methods described herein include features allowing an autonomous vehicle to create visual or audio signals for vehicles around the autonomous vehicle, e.g., those vehicles that are tailgating the autonomous vehicle or are in a blind spot of the autonomous vehicle such that maneuvers of the autonomous vehicle might affect their safety.
The above and other aspects and features of the disclosed technology are described in greater detail in the drawings, the description, and the claims.
Autonomous driving systems (also referred to as autonomous driving vehicles or autonomous vehicles) should safely accommodate all types of road configurations and conditions, including weather conditions (e.g., rain, snow, wind, dust storms, etc.), traffic conditions, and behaviors of other road users (e.g., vehicles, pedestrians, construction activities, etc.). Autonomous driving systems should make decisions about the speed and distance of traffic as well as about obstacles, including obstacles that obstruct the view of the autonomous vehicle's sensors. For example, an autonomous vehicle should estimate the distances between it and other vehicles, as well as the speeds and/or accelerations of those vehicles (e.g., relative to the autonomous vehicle and/or relative to each other; vehicle speed or acceleration can be determined in a certain system of coordinates, for example). Based on that information, the autonomous vehicle can decide whether or not it is safe to proceed along a planned path, when it is safe to proceed, and it also can make corrections to the planned path, if necessary. In various embodiments, speeds or velocities are determined and locations of objects or distances to the objects are determined. For simplicity, the following description uses speed, but velocity could also be determined, where velocity is a speed and a direction (is a vector). Also, although distance is used below, location (e.g., in a 2D or a 3D coordinate system) can be used as well.
Examples of road configurations where these determinations and decisions should be made include so-called “T” intersections, so-called “Y” intersections, unprotected left turns, intersections with a yield where an autonomous vehicle (e.g., an autonomous truck) does not have the right-of-way, a roundabout, an intersection with stop signs where all traffic has to stop, and an intersection with 4 road sections and two stop signs where the autonomous vehicle must stop and other vehicles are not required to stop (e.g., cross-traffic does not stop), as well as many other road configurations. Examples of traffic conditions can include a vehicle tailgating the autonomous vehicle at a distance from the autonomous vehicle that the autonomous vehicle determines to be unsafe or potentially unsafe. For example, an autonomous vehicle may determine that a distance is unsafe if the distance is below a threshold value, which may be a predetermined distance value or a distance value determined by the autonomous vehicle based on traffic conditions, road conditions, speed of the vehicle following the autonomous vehicle relative to the speed of the autonomous vehicle, and/or a weight of the autonomous vehicle including the weight of goods (e.g., lumber, cars, furniture, corn, etc.) loaded in/on a trailer coupled to and transported by the autonomous vehicle. In some examples, the weight of the load and the trailer being transported/hauled by an autonomous vehicle can impact the performance of the autonomous vehicle. For example, an engine/motor of the vehicle drive subsystems 142 of
For all of the foregoing road configurations and traffic conditions, the autonomous vehicle must decide how it can safely proceed. To increase safety of the autonomous vehicle operation, the autonomous vehicle can provide visual and/or audio indications to other users on the road to help them to ensure to stay a safe distance from the autonomous vehicle, for example. According to some example embodiments, because of a non-compliant driver (e.g., a driver that drives erratically between the lanes) ahead of the autonomous vehicle, the autonomous vehicle might anticipate that it will need to apply its brakes suddenly or perform another sudden or aggressive maneuver at some point along its projected path. In such a situation, the autonomous vehicle can present a visual sign or signs on one or more locations on its body (e.g., back and/or sides) that alert other vehicles around it (e.g., that alert drivers of those vehicles) that the autonomous vehicle anticipates an upcoming situation where it might need to perform an aggressive/sudden maneuver (e.g., a sudden braking, an aggressive lane change, an aggressive acceleration, an aggressive deceleration, etc.) that can affect those vehicles. The visual sign can, for example, stay on until the potentially dangerous situation is eliminated, as determined by the autonomous vehicle. According to some example embodiments, the autonomous vehicle can also present visual signs to other vehicles indicating that they are in a spot with limited sensor reception by the autonomous vehicle (e.g., a “blind spot”) or that they are about to move into an area (e.g., sides or behind) around the autonomous vehicle where sensor perception of the autonomous vehicle is limited. Some implementations can provide signaling of the autonomous vehicle's intent via, e.g., an external visual indicator of what the autonomous vehicle's intent is (e.g., whether the vehicle is about to brake, change lanes, or perform some other maneuver).
Also, according to some example embodiments, when the autonomous vehicle is stopped at a traffic stop near a pedestrian crossing, it can generate visual and/or audio signals to acknowledge a pedestrian crossing the street along the pedestrian crossing (by, e.g., playing a pre-recorded message announcing that the autonomous vehicle is aware of the pedestrian's presence). Additionally, based on sensor data from sensors (e.g., scanning forward and rearward for other vehicles) on/in the AV, the AV may indicate to pedestrians that are on a sidewalk awaiting to cross the street that it may be safe to cross the street. For example, the AV may play a pre-recorded message announcing that those pedestrians can proceed crossing the street.
According to some example embodiments, an autonomous vehicle can display a sign for a tailgating vehicle indicating that the tailgating vehicle is too close to the autonomous vehicle. In some example embodiments, the sign can include a message indicating a distance at which it would be considered safe for other vehicles to follow the autonomous vehicle. Because the autonomous vehicle can typically obtain information about the surrounding environment a long distance ahead, it can instruct the vehicles following it (e.g., via displaying a visual sign for them) to keep at a safe distance from the autonomous vehicle. That safe distance can be determined by the autonomous vehicle based on, for example, information obtained by the autonomous vehicle from the surrounding environment using one or more sensors of the autonomous vehicle, a speed of the autonomous vehicle, a relative speed of the autonomous vehicle and another vehicle, or it can be a preset distance value. According to example embodiments, the safe distance can be updated by the autonomous vehicle (e.g., in a periodic manner). In some embodiments, visual indicators used by the autonomous vehicle (such as visual cue alerts for vehicles following the autonomous vehicle) may be based on sensor data of the autonomous vehicle. Visual indicators for a tailgating vehicle can be displayed on a rear part/surface of the autonomous vehicle, for example.
In the tailgating scenario, the autonomous vehicle can also display another sign indicating that the autonomous vehicle encourages the tailgating vehicle to pass the autonomous vehicle. For example, the AV may access data from its sensors indicating that there are no vehicles approaching the AV from a direction opposite from the AV. In some implementations, after a predefined time of opportunity for the tailgating vehicle to pass or to increase its distance from the autonomous vehicle has elapsed, if that vehicle continues to tailgate the autonomous vehicle, the autonomous vehicle can, for example, change lanes, or reduce or increase its speed within a proper speed limit to prevent a potential collision. In some implementations, the autonomous vehicle may display a sign for another vehicle only after that vehicle follows the autonomous vehicle at a distance less than a safe distance for a predetermined amount of time. In some example embodiments, the sign displayed by the autonomous vehicle to a tailgating vehicle can be a yellow light displayed on the back/rear side of the autonomous vehicle (or the back of the trailer that is connected to or is a part of the autonomous vehicle, back of a tractor, back of a passenger vehicle, etc.) indicating that the tailgating vehicle should increase its distance from the autonomous vehicle. That yellow light can turn green when the other vehicle increases its distance from the autonomous vehicle to a safe distance (predetermined or dynamically changing according to sensor data collected by the autonomous vehicle, for example). Providing the indicators to the vehicles following the autonomous vehicle may avoid rear-end collisions between the autonomous vehicles and other vehicles, for example.
In some example embodiments, when the autonomous vehicle is a tractor-trailer (e.g., a class 8 or other class vehicles), the autonomous vehicle can generate visual and/or audio warning signs when it performs a wide right turn to warn other vehicles following behind it in the same and/or other lanes of a road.
According to some example embodiments, when the autonomous vehicle is stopped at a traffic stop, it can indicate that it believes it is its turn to leave the stop by displaying a corresponding visual sign for other vehicles in the traffic stop area.
In some example embodiments, the autonomous vehicle can acknowledge that the autonomous vehicle understands directions given by a traffic controller at a road construction site or a police officer at an accident site by displaying corresponding visual signs and/or playing pre-recorded and/or ad hoc synthesized audio messages.
According to some example embodiments, an autonomous vehicle can display a sign or provide other means of visual indication that it is operating in the autonomous mode. Such information can be helpful for other vehicles and/or their drivers that share the road with the autonomous vehicle.
In certain example embodiments, the autonomous vehicle can use visual indicators in addition to the standard turn indicators when it is about to do an aggressive lane change (e.g., when the projected amount of time between the start of the standard turn indication and the actual turn is less than a predetermined threshold value).
The types of means or devices that can be used by an autonomous vehicle to provide visual signs, icons, indicators or cues to vehicles (both autonomous and human-operated) as well as other road users (e.g., pedestrians, construction workers, law enforcement persons, etc.) around the autonomous vehicle according to various embodiments include but are not limited to: one or more light sources (also referred to as lights), e.g., static or flashing; a group or an array or a series of light sources (e.g., light-emitting diodes (LEDs)) that can display a sequence of lights varying in position, intensity and/or color; one or more liquid crystal displays (LCDs) that can display both static and dynamic visual information (e.g., animations). Audio signals, cues, and indicators, of varying intensity, according to the disclosed technology can be generated by one or more speakers that can be positioned at any location on or in the autonomous vehicle. In some embodiments, if an autonomous vehicle provides warning signals to a vehicle that is following the autonomous vehicle too closely, and that vehicle does not increase its distance from the autonomous vehicle, the autonomous vehicle may increase the intensity or frequency of the warning signals or provide different warning signals. For example, the warning signals may become brighter in color or luminosity and/or become louder in audio. The autonomous vehicle can obtain information about its surrounding environment using various sensors and devices including but not limited to video cameras, LiDAR or RADAR (Radio Detection and Ranging) sensors, accelerometers, gyroscopes, inertial measurement units (IMUs), etc.
Vehicle sensor subsystems 144 can include sensors for general operation of the autonomous truck 105. The sensors for general operation of the autonomous vehicle may include cameras, a temperature sensor, an inertial sensor (IMU), a global positioning system, a light sensor, a LIDAR system, a radar system, and wireless communications.
The vehicle control subsystems 146 may be configured to control operation of the autonomous vehicle, or truck, 105 and its components. Accordingly, the vehicle control subsystems 146 may include various elements such as an engine power output subsystem, a brake unit, a navigation unit, a steering system, and an autonomous control unit. The engine power output may control the operation of the engine, including the torque produced or horsepower provided, as well as provide control of the gear selection of the transmission. The brake unit can include any combination of mechanisms configured to decelerate the autonomous vehicle 105. The brake unit can use friction to slow the wheels in a standard manner. The brake unit may include an anti-lock brake system (ABS) that can prevent the brakes from locking up when the brakes are applied. The navigation unit may be any system configured to determine a driving path or route for the autonomous vehicle 105. The navigation unit may additionally be configured to update the driving path dynamically while the autonomous vehicle 105 is in operation. In some embodiments, the navigation unit may be configured to incorporate data from a GPS device and one or more predetermined maps so as to determine the driving path for the autonomous vehicle 105. The steering system may represent any combination of mechanisms that may be operable to adjust the heading of the autonomous vehicle 105 in an autonomous mode or in a driver-controlled mode.
The autonomous control unit may represent a control system configured to identify, evaluate, and avoid or otherwise negotiate potential obstacles in the environment of the autonomous vehicle 105. In general, the autonomous control unit may be configured to control the autonomous vehicle 105 for operation without a driver or to provide driver assistance in controlling the autonomous vehicle 105. In some embodiments, the autonomous control unit may be configured to incorporate data from the GPS device, the RADAR, the LiDAR, the cameras, and/or other vehicle subsystems to determine the driving path or trajectory for the autonomous vehicle 105.
An in-vehicle control computer 150, which may be referred to as a vehicle control unit or VCU, can include, for example, any of: a vehicle subsystem interface 160, a driving operation module 168, one or more processors 170, a meta-perception module 165, a memory 175, an external signaling module 167, or a network communications subsystem 178. This in-vehicle control computer 150 may control many operations of the autonomous truck 105 in response to information available from the various vehicle subsystems 140. The one or more processors 170 execute the operations associated with the meta-perception module 165 that, for example, allow the system to determine confidence in perception data indicating a hazard, determine a confidence level of a regional map, and to analyze the behavior of agents of interest (also referred as targets) surrounding the autonomous vehicle 105. According to some example embodiments, an agent of interest or a target can be one of: another vehicle, a vehicle following the autonomous vehicle 105, a vehicle in a vicinity of the autonomous vehicle 105, a pedestrian, a construction zone, or a vehicle proximate to the autonomous vehicle 105. For example, the target may be within an intended maneuver zone around the autonomous vehicle. Data from vehicle sensor subsystems 144 may be provided to the meta-perception module 165 so that the course of action may be appropriately determined. Alternatively, or additionally, the meta-perception module 165 may determine the course of action in conjunction with another operational or control module, such as the driving operation module 168 or the external signaling module 167. According to some example embodiments, the external signaling module 167 can be configured to control signaling behaviors of the autonomous vehicle 105. According to some example embodiments, the signaling behaviors of the autonomous vehicle can be determined by the external signaling module 167 using, e.g., information provided by one or more sensors of the vehicle sensor subsystems 144. Example signaling behaviors of the autonomous vehicle 105 are described below.
The memory 175 may contain additional instructions as well, including instructions to transmit data to, receive data from, interact with, or control one or more of the vehicle drive subsystems 142, the vehicle sensor subsystems 144, or the vehicle control subsystems 146. The in-vehicle control computer (VCU) 150 may control the function of the autonomous vehicle 105 based on inputs received from various vehicle subsystems (e.g., the vehicle drive subsystems 142, the vehicle sensor subsystems 144, and the vehicle control subsystems 146). Additionally, the VCU 150 may send information to the vehicle control subsystems 146 to direct the trajectory, velocity, signaling behaviors, and the like, of the autonomous vehicle 105. The autonomous control vehicle control subsystems 146 may receive a course of action to be taken from one or more modules of the VCU 150 and consequently relay instructions to other subsystems to execute the course of action.
In some embodiments, the intended maneuver zone 330 may change in size and shape based on an intended maneuver of the AV 210. As shown in
According to some example embodiments, the autonomous vehicle 210 includes an intention indicator 340, which may be located on the back surface 310 of the trailer 212 the vehicle 105 is towing, on the sides of the autonomous tractor 105, on the sides of the trailer 212, or on the back of the autonomous tractor 105 without a trailer 212. In some embodiments, when a trailer 212 is hooked up with an autonomous tractor 105, the intention indicator 340 on the back of the autonomous tractor 105 may be deactivated and the intention indicator 340 on the back surface 310 of the trailer 212 may be activated. When the trailer 212 is unhooked from the autonomous tractor 105, the intention indicator 340 on the back of the autonomous tractor 105 is reactivated.
As shown in
Various technical solutions that may be implemented by some embodiments include:
A method (e.g., method 400) of operating an autonomous vehicle, including determining, by the autonomous vehicle, whether a target is in an intended maneuver zone around the autonomous vehicle; generating, by the autonomous vehicle, a signal in response to determining that the target is within the intended maneuver zone around the autonomous vehicle; determining, by the autonomous vehicle and based on perception information acquired by the autonomous vehicle, whether the target has left the intended maneuver zone around the autonomous vehicle; and determining, by the autonomous vehicle, that it is safe to perform the intended maneuver in response to determining, by the autonomous vehicle, that the target is not in the intended maneuver zone or in response to determining, by the autonomous vehicle, that the target has left the intended maneuver zone.
The method of operating an autonomous vehicle further includes performing, by the autonomous vehicle, the intended maneuver.
The method of operating an autonomous vehicle further includes performing, by the autonomous vehicle, an alternative maneuver or delaying the intended maneuver in response to determining that the target is within the intended maneuver zone.
In the method of operating an autonomous vehicle the signal is generated, by the autonomous vehicle, via a light source and/or an image screen.
In the method of operating an autonomous vehicle the signal is an intended maneuver signal including a time sequence of lights, an image, an icon, and/or an animation.
In the method of operating an autonomous vehicle the target includes a vehicle following the autonomous vehicle, a pedestrian, a construction zone, and/or a vehicle in a vicinity of the autonomous vehicle.
In the method of operating an autonomous vehicle the signal includes one or more warning signals generated, by the autonomous vehicle, in response to determining, by the autonomous vehicle, that a distance between a vehicle following the autonomous vehicle and the autonomous vehicle is below a threshold value.
In the method of operating an autonomous vehicle the threshold value is based on a predetermined value or based on a value determined, by the autonomous vehicle, based on traffic conditions, road conditions, speed of the vehicle following the autonomous vehicle relative to the speed of the autonomous vehicle, and/or a weight of the autonomous vehicle including a weight of goods being transported by the autonomous vehicle.
In the method of operating an autonomous vehicle the one or more warning signals include one or more variable-intensity visual and/or audio signals, and wherein the one or more variable-intensity visual signals are presented, by the autonomous vehicle, at one or more external locations on the autonomous vehicle, and/or the one or more variable-intensity audio signals are presented, by the autonomous vehicle, via one or more audio devices in or on the autonomous vehicle.
A system for autonomous driving operation, including an autonomous vehicle that includes a plurality of subsystems configured to determine, by at least one of the plurality of subsystems, whether a target is in an intended maneuver zone around the autonomous vehicle; generate, by at least one of the plurality of subsystems, a signal in response to determining that the target is within the intended maneuver zone around the autonomous vehicle; determine, by at least one of the plurality of subsystems, perception information indicating whether the target has left the intended maneuver zone around the autonomous vehicle; and determine, by at least one of the plurality of subsystems, that it is safe for the autonomous vehicle to perform the intended maneuver in response to determining, by at least one of the plurality of subsystems, that the target is not in the intended maneuver zone or in response to determining, by at least one of the plurality of subsystems, that the target has left the intended maneuver zone.
In the system for autonomous driving operation at least one of the plurality of subsystems causes the autonomous vehicle to perform the intended maneuver.
In the system for autonomous driving operation at least one of the plurality of subsystems causes the autonomous vehicle to perform an alternative maneuver or delay the intended maneuver in response to determining that the target is within the intended maneuver zone.
In the system for autonomous driving operation the signal is generated, by at least one of the plurality of subsystems, via a light source and/or an image screen.
In the system for autonomous driving operation the signal is an intended maneuver signal including a time sequence of lights, an image, an icon, and/or an animation.
In the system for autonomous driving operation the target includes a vehicle following the autonomous vehicle, a pedestrian, a construction zone, and/or a vehicle in a vicinity of the autonomous vehicle.
In the system for autonomous driving operation the signal includes one or more warning signals generated, by at least one of the plurality of subsystems, in response to determining, by at least one of the plurality of subsystems, that a distance between a vehicle following the autonomous vehicle and the autonomous vehicle is below a threshold value.
In the system for autonomous driving operation the threshold value is based on a predetermined value or based on a value determined, by at least one of the plurality of subsystems, based on traffic conditions, road conditions, speed of the vehicle following the autonomous vehicle relative to the speed of the autonomous vehicle, and/or a weight of the autonomous vehicle including a weight of goods being transported by the autonomous vehicle.
In the system for autonomous driving operation the one or more warning signals include one or more variable-intensity visual and/or audio signals, and wherein the one or more variable-intensity visual signals are presented, by at least one of the plurality of subsystems, at one or more external locations on the autonomous vehicle, and/or the one or more variable-intensity audio signals are presented, by at least one of the plurality of subsystems, via one or more audio devices in or on the autonomous vehicle.
A non-transitory machine-useable storage medium embodying instructions which, when executed by a machine, cause the machine to determine, by an autonomous vehicle, whether a target is in an intended maneuver zone around the autonomous vehicle; generate, by the autonomous vehicle, a signal in response to determining that the target is within the intended maneuver zone around the autonomous vehicle; determine, by the autonomous vehicle and based on perception information acquired by the autonomous vehicle, whether the target has left the intended maneuver zone around the autonomous vehicle; and determine, by the autonomous vehicle, that it is safe to perform the intended maneuver in response to determining, by the autonomous vehicle, that the target is not in the intended maneuver zone or in response to determining, by the autonomous vehicle, that the target has left the intended maneuver zone.
In the non-transitory machine-useable storage medium the signal includes one or more warning signals generated, by the autonomous vehicle, in response to determining, by the autonomous vehicle, that a distance between a vehicle following the autonomous vehicle and the autonomous vehicle is below a threshold value.
Implementations of the subject matter and the functional operations described in this document can be implemented in various systems, semiconductor devices, ultrasonic devices, digital electronic circuitry, or in computer software, firmware, or hardware, including the structures disclosed in this specification and their structural equivalents, or in combinations of one or more of them. Implementations of aspects of the subject matter described in this specification can be implemented as one or more computer program products, e.g., one or more modules of computer program instructions encoded on a tangible and non-transitory computer readable medium for execution by, or to control the operation of, data processing apparatus. The computer readable medium can be a machine-readable storage device, a machine-readable storage substrate, a memory device, a composition of matter effecting a machine-readable propagated signal, or a combination of one or more of them. The term “data processing unit” or “data processing apparatus” encompasses all apparatus, devices, and machines for processing data, including by way of example a programmable processor, a computer, or multiple processors or computers. The apparatus can include, in addition to hardware, code that creates an execution environment for the computer program in question, e.g., code that constitutes processor firmware, a protocol stack, a database management system, an operating system, or a combination of one or more of them.
A computer program (also known as a program, software, software application, script, or code) can be written in any form of programming language, including compiled or interpreted languages, and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A computer program does not necessarily correspond to a file in a file system. A program can be stored in a portion of a file that holds other programs or data (e.g., one or more scripts stored in a markup language document), in a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub programs, or portions of code). A computer program can be deployed to be executed on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a communication network.
The processes and logic flows described in this specification can be performed by one or more programmable processors executing one or more computer programs to perform functions by operating on input data and generating output. The processes and logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application specific integrated circuit).
Processors suitable for the execution of a computer program include, by way of example, both general and special purpose microprocessors, and any one or more processors of any kind of digital computer. Generally, a processor will receive instructions and data from a read only memory or a random-access memory or both. The essential elements of a computer are a processor for performing instructions and one or more memory devices for storing instructions and data. Generally, a computer will also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto optical disks, or optical disks. However, a computer need not have such devices. Computer readable media suitable for storing computer program instructions and data include all forms of nonvolatile memory, media and memory devices, including by way of example semiconductor memory devices, e.g., EPROM, EEPROM, and flash memory devices. The processor and the memory can be supplemented by, or incorporated in, special purpose logic circuitry.
In this disclosure, LiDAR and LIDAR are used to refer to light detection and ranging devices and methods, and alternatively, or additionally, laser detection and ranging devices and methods. The use of these acronyms does not imply limitation of the described devices, systems, or methods to the use of one over the other.
While this document contains many specifics, these should not be construed as limitations on the scope of any invention or of what may be claimed, but rather as descriptions of features that may be specific to particular embodiments of particular inventions. Certain features that are described in this document in the context of separate embodiments can also be implemented in combination in a single embodiment. Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable subcombination. Moreover, although features may be described above as acting in certain combinations and even initially claimed as such, one or more features from a claimed combination can in some cases be excised from the combination, and the claimed combination may be directed to a subcombination or variation of a subcombination.
Similarly, while operations are depicted in the drawings in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all illustrated operations be performed, to achieve desirable results. Moreover, the separation of various system components in the embodiments described in this document should not be understood as requiring such separation in all embodiments.
Only some implementations and examples are described, and other implementations, enhancements and variations can be made based on what is described and illustrated in this document.
This patent document claims the priority to and the benefits of U.S. Provisional Application No. 63/233,108 entitled “SYSTEM AND METHOD FOR AN AUTONOMOUS VEHICLE” filed on Aug. 13, 2021. The entire disclosure of the aforementioned application is hereby incorporated by reference as part of the disclosure of this application.
Number | Date | Country | |
---|---|---|---|
63233108 | Aug 2021 | US |