Priority is claimed on Japanese Patent Application No. 2021-089427, filed May 27, 2021, the content of which is incorporated herein by reference.
The present invention relates to a mobile object control device, a mobile object control method, and a storage medium.
In recent years, research on automated driving that automatically controls the traveling of a vehicle has been conducted. In this regard, technology for estimating virtual markings on the basis of positions of previously detected markings when the present state changes from a state in which markings of a lane in which vehicles are traveling are detected to a state in which no markings are detected and controlling traveling of a host vehicle so that the host vehicle is at a prescribed position with respect to the estimated virtual markings is known (for example, PCT International Publication No. WO 2018/012179).
However, because positions of previously detected markings do not always continue as they are, there is a case where markings different from the actual ones are recognized.
Aspects of the present invention have been made in consideration of such circumstances and the present invention provides a mobile object control device, a mobile object control method, and a storage medium capable of further improving the accuracy of recognition of markings for dividing an area through which a mobile object passes.
A mobile object control device, a mobile object control method, and a storage medium according to the present invention adopt the following configurations.
(1): According to an aspect of the present invention, there is provided a mobile object control device including: a recognizer configured to recognize a surrounding situation of a mobile object on the basis of an output of an external sensor; and a marking recognizer configured to recognize markings for dividing an area through which the mobile object passes on the basis of the surrounding situation recognized by the recognizer, wherein the marking recognizer extracts a prescribed area from the surrounding situation when it is determined that marking recognition accuracy has been lowered, extracts an edge within the extracted prescribed area, and recognizes the markings on the basis of an extraction result.
(2): In the above-described aspect (1), the mobile object control device further includes a storage controller configured to cause a storage to store information about the markings before the marking recognizer determines that the marking recognition accuracy has been lowered, wherein the marking recognizer extracts marking candidates on the basis of the edge extracted from the prescribed area and recognizes the markings for dividing the area through which the mobile object passes on the basis of degrees of similarity between information about the extracted marking candidates and the information about the markings stored in the storage.
(3): In the above-described aspect (2), the information about the markings includes at least one of positions, directions, and types of the markings.
(4): In the above-described aspect (2), the storage further stores map information and the marking recognizer extracts the prescribed area on the basis of information about the markings acquired from the map information on the basis of position information of the mobile object or the information about the markings before it is determined that the marking recognition accuracy has been lowered stored in the storage when it is determined that the marking recognition accuracy has been lowered.
(5): In the above-described aspect (1), the prescribed area is set on left and right sides in a traveling direction of the mobile object.
(6): In the above-described aspect (1), the marking recognizer extracts the edge in the surrounding situation and determines that the marking recognition accuracy has been lowered when at least one of a length, reliability, and quality of the extracted edge is less than a threshold value.
(7): According to an aspect of the present invention, there is provided a mobile object control method including: recognizing, by a computer, a surrounding situation of a mobile object on the basis of an output of an external sensor; recognizing, by the computer, markings for dividing an area through which the mobile object passes on the basis of the recognized surrounding situation; extracting, by the computer, a prescribed area from the surrounding situation when it is determined that marking recognition accuracy has been lowered; extracting, by the computer, an edge within the extracted prescribed area; and recognizing, by the computer, the markings on the basis of an extraction result.
(8): According to an aspect of the present invention, there is provided a computer-readable non-transitory storage medium storing a program for causing a computer to: recognize a surrounding situation of a mobile object on the basis of an output of an external sensor; recognize markings for dividing an area through which the mobile object passes on the basis of the recognized surrounding situation; extract a prescribed area from the surrounding situation when it is determined that marking recognition accuracy has been lowered; extract an edge within the extracted prescribed area; and recognize the markings on the basis of an extraction result.
According to the above-described aspects (1) to (8), it is possible to further improve the accuracy of recognition of markings for dividing an area through which a mobile object passes.
Embodiments of a mobile object control device, a mobile object control method, and a storage medium of the present invention will be described below with reference to the drawings. Hereinafter, a vehicle is used as an example of a mobile object. The vehicle is, for example, a vehicle such as a two-wheeled vehicle, a three-wheeled vehicle, or a four-wheeled vehicle, and a drive source thereof is an internal combustion engine such as a diesel engine or a gasoline engine, an electric motor, or a combination thereof. The electric motor operates using electric power generated by a power generator connected to the internal combustion engine or electric power when a secondary battery or a fuel cell is discharged. Hereinafter, an embodiment in which the mobile object control device is applied to an automated driving vehicle will be described as an example. For example, automated driving is a process of executing driving control by automatically controlling one or all of steering, acceleration, and deceleration of the vehicle. The driving control of the vehicle may include, for example, various types of driving assistance control such as adaptive cruise control (ACC), auto lane changing (ALC), a lane keeping assistance system (LKAS), and traffic jam pilot (TJP). Some or all driving of the automated driving vehicle may be controlled according to manual driving of an occupant (a driver). The mobile object may include, in addition to (or instead of) a vehicle, for example, a ship, a flying object (including, for example, a drone, an aircraft, etc.) and the like.
For example, the camera 10 is a digital camera using a solid-state imaging element such as a charge coupled device (CCD) or a complementary metal oxide semiconductor (CMOS). The camera 10 is attached to any position on the vehicle (hereinafter, a vehicle M) in which the vehicle system 1 is mounted. When the view in front of the vehicle M is imaged, the camera 10 is attached to an upper part of a front windshield, a rear surface of a rearview mirror, or the like. For example, the camera 10 periodically and iteratively images the surroundings of the vehicle M. The camera 10 may be a stereo camera.
The radar device 12 radiates radio waves such as millimeter waves around the vehicle M and detects at least a position (a distance to and a direction) of a physical object by detecting radio waves (reflected waves) reflected by the physical object. The radar device 12 is attached to any position on the vehicle M. The radar device 12 may detect a position and a speed of the physical object in a frequency modulated continuous wave (FM-CW) scheme.
The LIDAR sensor 14 radiates light (or electromagnetic waves of a wavelength close to that of light) to the vicinity of the vehicle M and measures scattered light. The LIDAR sensor 14 detects a distance to an object on the basis of a time period from light emission to light reception. The radiated light is, for example, pulsed laser light. The LIDAR sensor 14 is attached to any position on the vehicle M. When the SONAR sensor is provided on the vehicle M, the SONAR sensor is provided, for example, on a front end and a rear end of the vehicle M, and is installed on a bumper or the like. The SONAR sensor detects a physical object (for example, an obstacle) within a prescribed distance from an installation position.
The physical object recognition device 16 performs a sensor fusion process on detection results from some or all of the components of the external sensor ES (the camera 10, the radar device 12, the LIDAR sensor 14, and the SONAR sensor) to recognize a position, a type, a speed, and the like of a physical object. The physical object recognition device 16 outputs recognition results to the automated driving control device 100. The physical object recognition device 16 may output detection results of the external sensor ES to the automated driving control device 100 as they are. In this case, the physical object recognition device 16 may be omitted from the vehicle system 1.
The communication device 20 communicates with another vehicle in the vicinity of the vehicle M using, for example, a cellular network, a Wi-Fi network, Bluetooth (registered trademark), dedicated short range communication (DSRC), or the like, or communicates with various types of server devices via a radio base station.
The HMI 30 outputs various types of information to the occupant of the vehicle M and receives input operations from the occupant. The HMI 30 includes, for example, various types of display devices, a speaker, a buzzer, a touch panel, a switch, a key, a microphone, and the like. The various types of display devices are, for example, a liquid crystal display (LCD), an organic electro luminescence (EL) display device, and the like. The display device is provided, for example, near the front of the driver's seat (the seat closest to a steering wheel) on an instrument panel, and is installed at a position where the occupant can perform visual recognition from the steering wheel gap or through the steering wheel. The display device may be installed at the center of the instrument panel. The display device may be a head up display (HUD). The HUD projects an image onto a part of the front windshield in front of the driver's seat so that the eyes of the occupant sitting in the driver's seat can see the virtual image. The display device displays an image generated by the HMI controller 170 to be described below. The HMI 30 may include an operation changeover switch or the like that mutually switches between automated driving and manual driving by the occupant.
The vehicle sensor 40 includes a vehicle speed sensor configured to detect the speed of the vehicle M, an acceleration sensor configured to detect acceleration, a yaw rate sensor configured to detect an angular speed around a vertical axis, a direction sensor configured to detect a direction of the vehicle M, and the like. The vehicle sensor 40 may include a position sensor that acquires a position of the vehicle M. The position sensor is, for example, a sensor that acquires position information (longitude/latitude information) from a Global Positioning System (GPS) device. The position sensor may be a sensor that acquires position information using a global navigation satellite system (GNSS) receiver 51 of the navigation device 50.
For example, the navigation device 50 includes the GNSS receiver 51, a navigation HMI 52, and a route determiner 53. The navigation device 50 retains first map information 54 in a storage device such as a hard disk drive (HDD) or a flash memory. The GNSS receiver 51 identifies a position of the vehicle M on the basis of a signal received from a GNSS satellite. The position of the vehicle M may be identified or corrected by an inertial navigation system (INS) using an output of the vehicle sensor 40. The navigation HMI 52 includes a display device, a speaker, a touch panel, keys, and the like. The navigation HMI 52 may be partly or wholly shared with the above-described HMI 30. For example, the route determiner 53 determines a route (hereinafter referred to as a route on a map) from the position of the vehicle M identified by the GNSS receiver 51 (or any input position) to a destination input by the occupant using the navigation HMI 52 with reference to the first map information 54. The first map information 54 is, for example, information in which a road shape is expressed by a link indicating a road and nodes connected by the link. The first map information 54 may include curvature of a road, point of interest (POI) information, and the like. The route on the map is output to the MPU 60. The navigation device 50 may perform route guidance using the navigation HMI 52 on the basis of the route on the map. The navigation device 50 may be implemented, for example, according to a function of a terminal device such as a smartphone or a tablet terminal possessed by the occupant. The navigation device 50 may transmit a current position and a destination to a navigation server via the communication device 20 and acquire a route equivalent to the route on the map from the navigation server.
For example, the MPU 60 includes a recommended lane determiner 61 and retains second map information 62 in a storage device such as an HDD or a flash memory. The recommended lane determiner 61 divides the route on the map provided from the navigation device 50 into a plurality of blocks (for example, divides the route every 100 [m] in a traveling direction of the vehicle), and determines a recommended lane for each block with reference to the second map information 62. The recommended lane determiner 61 determines in what lane numbered from the left the vehicle will travel. For example, the recommended lane determiner 61 determines the recommended lane so that the vehicle M can travel along a reasonable route for traveling to a branching destination when there is a branch point in the route on the map.
The second map information 62 is map information which has higher accuracy than the first map information 54. The second map information 62 includes, for example, road marking information such as positions, directions, and types of road markings (hereinafter simply referred to as “markings”) that divide one or more lanes included in a road, information of the center of the lane or information of the boundary of the lane based on the marking information, and the like. The second map information 62 may include information about protective barriers such as guardrails and fences, chatter bars, curbs, median strips, road shoulders, sidewalks, and the like provided along an extension direction of the road or the like. The second map information 62 may include road information (a type of road), legal speeds (a speed limit, a maximum speed, and a minimum speed), traffic regulation information, address information (an address/postal code), facility information, telephone number information, and the like. The second map information 62 may be updated at any time when the communication device 20 communicates with another device.
The driving operation elements 80 include, for example, an accelerator pedal, a brake pedal, a shift lever, and other operation elements in addition to the steering wheel. A sensor for detecting an amount of operation or the presence or absence of an operation is attached to the driving operation element 80 and a detection result is output to the automated driving control device 100 or some or all of the travel driving force output device 200, the brake device 210, and the steering device 220. The operation element does not necessarily have to be annular and may be in the form of a variant steering wheel, a joystick, a button, or the like.
The automated driving control device 100 includes, for example, a first controller 120, a second controller 160, an HMI controller 170, a storage controller 180, and a storage 190. Each of the first controller 120, the second controller 160, and the HMI controller 170 is implemented, for example, by a hardware processor such as a central processing unit (CPU) executing a program (software). Some or all of the above components may be implemented by hardware (including a circuit; circuitry) such as a large-scale integration (LSI) circuit, an application-specific integrated circuit (ASIC), a field-programmable gate array (FPGA), or a graphics processing unit (GPU) or may be implemented by software and hardware in cooperation. The program may be prestored in a storage device (a storage device including a non-transitory storage medium) such as an HDD or a flash memory of the automated driving control device 100 or may be stored in a removable storage medium such as a DVD or a CD-ROM and installed in the HDD or the flash memory of the automated driving control device 100 when the storage medium (the non-transitory storage medium) is mounted in a drive device. A combination of the action plan generator 140 and the second controller 160 is an example of a “driving controller.” The HMI controller 170 is an example of an “output controller.”
The storage 190 may be implemented by the above-described various types of storage devices or a solid-state drive (SSD), an electrically erasable programmable read-only memory (EEPROM), a read-only memory (ROM), a random-access memory (RAM), or the like. The storage 190 stores, for example, recognized marking information 192, a program, various other types of information, and the like. Details of the recognized marking information 192 will be described below. The above-described map information (first map information 54 and second map information 62) may be stored in the storage 190.
The recognizer 130 recognizes space information indicating a surrounding situation of the vehicle M, for example, on the basis of information input from the external sensor ES. For example, the recognizer 130 recognizes states of positions, speeds, acceleration, and the like of physical objects (for example, other vehicles or other obstacles) near the vehicle M. For example, the position of the physical object is recognized as a position on absolute coordinates with a representative point (a center of gravity, a driving shaft center, or the like) of the vehicle M as the origin and is used for control. The position of the physical object may be represented by a representative point such as a center of gravity or a corner of the physical object or may be represented by an area. The “state” of a physical object may include acceleration or jerk of another vehicle or an “action state” (for example, whether or not a lane change is being made or intended) when the physical object is a mobile object such as another vehicle.
The recognizer 130 recognizes, for example, a lane where the vehicle M is traveling (a traveling lane) from a surrounding situation of the vehicle M. The lane recognition is executed by the marking recognizer 132 provided in the recognizer 130. The details of the function of the marking recognizer 132 will be described below. The recognizer 130 recognizes an adjacent lane adjacent to the traveling lane. The adjacent lane is, for example, a lane where the vehicle M can travel in the same direction as the traveling lane. The recognizer 130 recognizes a temporary stop line, an obstacle, red traffic light, a toll gate, a road sign, and other road events.
When the traveling lane is recognized, the recognizer 130 recognizes a position or an orientation of the vehicle M with respect to the traveling lane. For example, the recognizer 130 may recognize a gap of a reference point of the vehicle M from the center of the lane and an angle formed with respect to a line connected to the center of the lane in a traveling direction of the vehicle M as a relative position and an orientation of the vehicle M related to the traveling lane. Alternatively, the recognizer 130 may recognize a position of the reference point of the vehicle M related to one side end (a marking or a road boundary) of the traveling lane or the like as a relative position of the vehicle M related to the traveling lane. Here, the reference point of the vehicle M may be the center of the vehicle M or the center of gravity. The reference point may be an end (a front end or a rear end) of the vehicle M or may be a position where one of a plurality of wheels provided in the vehicle M is present.
The action plan generator 140 generates a future target trajectory along which the vehicle M automatedly travels (independently of the driver's operation) so that the vehicle M can generally travel in the recommended lane determined by the recommended lane determiner 61 and cope with a surrounding situation of the vehicle M. For example, the target trajectory includes a speed element. For example, the target trajectory is represented by sequentially arranging points (trajectory points) at which the vehicle M is required to arrive. The trajectory points are points at which the vehicle M is required to arrive for each prescribed traveling distance (for example, about several meters [m]) along a road. In addition, a target speed and target acceleration for each prescribed sampling time (for example, about several tenths of a second [sec]) are generated as parts of the target trajectory. The trajectory point may be a position at which the vehicle M is required to arrive at the sampling time for each prescribed sampling time. In this case, information about the target speed or the target acceleration is represented by an interval between the trajectory points.
The action plan generator 140 may set an automated driving event (function) when a target trajectory is generated. Automated driving events include a constant-speed traveling event, a low-speed tracking event, a lane change event, a branch point-related movement event, a merge point-related movement event, a takeover event, and the like. The action plan generator 140 generates a target trajectory according to an activated event.
The second controller 160 controls the travel driving force output device 200, the brake device 210, and the steering device 220 so that the vehicle M passes along the target trajectory generated by the action plan generator 140 at the scheduled times.
The second controller 160 includes, for example, an acquirer 162, a speed controller 164, and a steering controller 166. The acquirer 162 acquires information of a target trajectory (trajectory points) generated by the action plan generator 140 and causes a memory (not shown) to store the information. The speed controller 164 controls the travel driving force output device 200 or the brake device 210 on the basis of a speed element associated with the target trajectory stored in the memory. The steering controller 166 controls the steering device 220 in accordance with a degree of curvature of the target trajectory stored in the memory. The processes of the speed controller 164 and the steering controller 166 are implemented by, for example, a combination of feedforward control and feedback control. As an example, the steering controller 166 executes feedforward control according to the curvature of the road in front of the vehicle M and feedback control based on a deviation from the target trajectory in combination.
The HMI controller 170 notifies the driver of the vehicle M of prescribed information using the HMI 30. The prescribed information includes, for example, driving assistance information. The driving assistance information includes, for example, information such as a speed of the vehicle M, an engine speed, the remaining amount of fuel, a radiator water temperature, a traveling distance, a state of a shift lever, a marking, a lane, other vehicles, and the like recognized by the physical object recognition device 16, the automated driving control device 100, and the like, a lane in which the vehicle M should travel, and a future target trajectory. The driving assistance information may include information indicating the switching of the driving mode to be described below and a driving state (for example, a type of automated driving in operation such as LKAS or ALC) in the driving assistance process and the like. For example, the HMI controller 170 may generate an image including the above-described prescribed information, cause the display device of the HMI 30 to display the generated image, generate a sound indicating the prescribed information, and cause the generated sound to be output from the speaker of the HMI 30. The HMI controller 170 may output the information received by the HMI 30 to the communication device 20, the navigation device 50, the first controller 120, and the like.
The travel driving force output device 200 outputs a travel driving force (torque) for enabling the vehicle M to travel to driving wheels. For example, the travel driving force output device 200 includes a combination of an internal combustion engine, an electric motor, a transmission, and the like, and an electronic control unit (ECU) that controls the internal combustion engine, the electric motor, the transmission, and the like. The ECU controls the above-described components in accordance with information input from the second controller 160 or information input from the driving operation element 80.
For example, the brake device 210 includes a brake caliper, a cylinder configured to transfer hydraulic pressure to the brake caliper, an electric motor configured to generate hydraulic pressure in the cylinder, and a brake ECU. The brake ECU controls the electric motor in accordance with the information input from the second controller 160 or the information input from the driving operation element 80 so that brake torque according to a braking operation is output to each wheel. The brake device 210 may include a mechanism configured to transfer the hydraulic pressure generated according to an operation on the brake pedal included in the driving operation elements 80 to the cylinder via a master cylinder as a backup. The brake device 210 is not limited to the above-described configuration and may be an electronically controlled hydraulic brake device configured to control an actuator in accordance with information input from the second controller 160 and transfer the hydraulic pressure of the master cylinder to the cylinder.
For example, the steering device 220 includes a steering ECU and an electric motor. For example, the electric motor changes a direction of steerable wheels by applying a force to a rack and pinion mechanism. The steering ECU drives the electric motor in accordance with the information input from the second controller 160 or the information input from the steering wheel 82 of the driving operation element 80 to change the direction of the steerable wheels.
Hereinafter, a marking recognition process according to the embodiment will be described. Hereinafter, a scene in which LKAS control based on automated driving is executed will be described. The LKAS control is, for example, a control process of controlling the steering of the vehicle M so that the vehicle M travels near the center of the traveling lane while recognizing the markings of the traveling lane, and supporting lane keeping of the vehicle M.
The marking recognizer 132 recognizes markings of the lane L1 in which the vehicle M travels, for example, on the basis of space information indicating a surrounding situation within the recognizable range RA of the external sensor ES. The recognition of the markings is repeatedly executed at prescribed timings. The prescribed timings may be, for example, prescribed cycle times or timings based on a speed or a traveling distance of the vehicle M.
For example, the marking recognizer 132 extracts an edge from a captured image of the recognizable range RA photographed by the camera 10 and recognizes a position of a marking on the basis of an extraction result. The edge includes, for example, a pixel (or a pixel group) in which a pixel value difference from pixels around the edge is larger than a reference value, i.e., a characteristic pixel. The marking recognizer 132 extracts the edge using, for example, a prescribed differential filter or an edge extraction filter such as a Prewitt filter or a Sobel filter with respect to a luminance value of each pixel within the image. The above-described edge extraction filter is merely an example, and the marking recognizer 132 may extract an edge on the basis of another filter or algorithm.
When there is a line segment (for example, a straight line or a curve) of an edge having a length longer than or equal to a first threshold value, the marking recognizer 132 recognizes the line segment as a marking on the basis of an edge extraction result. The marking recognizer 132 may connect edge line segments whose positions and directions are similar. The term “similar” indicates that a difference in a position or direction of the edge is within a prescribed range. The term “similar” may indicate that a degree of similarity is greater than or equal to a prescribed value.
The marking recognizer 132 may recognize that the line segment is a curve even if the line segment is greater than or equal to a first threshold value and the line segment is not a marking when the curvature thereof is greater than or equal to a prescribed value (a radius of curvature is less than or equal to a prescribed value). Thereby, a line segment that is clearly not a marking can be excluded and the accuracy of recognition of the marking can be improved.
The marking recognizer 132 may derive one or both of the reliability and quality of the edge in addition to (or instead of) the length of the edge. For example, the marking recognizer 132 derives the reliability of the edge as a marking in accordance with a degree of continuity and a degree of scattering of the extracted edge. For example, the marking recognizer 132 increases the reliability as the continuity of the line segment of the edge increases or the scattering in the extension direction of the edge decreases. The marking recognizer 132 may compare edges extracted from the left and right sides with respect to the position of the vehicle M and increase the reliability of the marking of each edge as a degree of similarity of the continuity or the scattering of the edge increases. The marking recognizer 132 increases the quality of the marking obtained from the edge as the number of extracted edges increases. For example, the quality may be replaced with an index value (a quality value) so that the value increases as the quality increases.
The marking recognizer 132 determines whether or not the marking recognition accuracy has been lowered. For example, the marking recognition accuracy is lowered when the actual marking is scratched, dirty, or the like or lowered by road surface reflection due to light from outside in the vicinity of the exit of a tunnel, road surface reflection due to outdoor lights in rainy weather or lights of an oncoming lane, performance deterioration of the external sensor ES due to the weather such as heavy rain, or the like. For example, when the length of the line segment of the edge extracted in an edge extraction process is less than the first threshold value, the marking recognizer 132 determines that the marking recognition accuracy has been lowered.
When the reliability or the quality value of the edge is derived, the marking recognizer 132 may determine that the marking recognition accuracy has been lowered if the reliability is less than a second threshold value or if the quality value is less than a third threshold value. That is, the marking recognizer 132 may determine that the marking recognition accuracy has been lowered, for example, when at least one of the length, reliability, and quality of the edge is less than the threshold value on the basis of results of the edge extraction process. Thereby, it is possible to more accurately determine whether or not the marking recognition accuracy has been lowered using a plurality of conditions. The marking recognizer 132 may determine whether the marking has been recognized normally using a criterion similar to the above-described determination criterion instead of determining whether or not the marking recognition accuracy has been lowered.
When the marking recognizer 132 determines that the marking recognition accuracy has not been lowered, the storage controller 180 stores the marking recognition result of the marking recognizer 132 (information about the marking) in recognized marking information 192. The information about the marking includes, for example, information about a state of the marking.
When it is determined that the marking recognition accuracy has been lowered, the marking recognizer 132 extracts a prescribed area from a surrounding situation recognized by the recognizer 130, extracts an edge with respect to the extracted prescribed area, and recognizes a marking of the lane L1 in which the vehicle M is traveling on the basis of an extraction result. In the scene of
The marking recognizer 132 may extract a road on which the vehicle M travels from the position information of the vehicle M with reference to the second map information 62 on the basis of, for example, the position information of the vehicle M, in place of (or in addition to) the recognized marking information 192 and extract an area (a marking presence area) where there is a marking that divides the traveling lane from marking information of the extracted road. The marking recognizer 132 may extract the final marking presence area on the basis of the marking presence area extracted from the recognized marking information 192 and the marking presence area extracted from the second map information 62.
The marking recognizer 132 sets, for example, marking presence areas on the left and right sides of the vehicle M with respect to a traveling direction (a forward direction) of the vehicle M. The marking recognizer 132 may set a size or a shape of the marking presence area according to a position of other vehicles in a nearby area, the presence/absence and shape of a physical object OB such as a guardrail, and the like. In the example of
The marking recognizer 132 extracts, for example, edges with respect to areas that are the marking presence areas LLA and CLA included in an image captured by the camera 10. In this case, the marking recognizer 132 extracts edges on the basis of the various types of edge extraction filters described above and other filters or algorithms. The marking presence areas LLA and CLA have higher likelihoods of presence of the markings than the other areas of the recognizable range RA. Thus, the marking recognizer 132 extracts edges using a filter or an algorithm that extracts the edges more easily than an edge extraction process of the marking recognizer 132. Thereby, the edge can be extracted more reliably within the marking presence area.
The marking recognizer 132 extracts a line segment of an edge whose length is greater than or equal to a fourth threshold value included in the marking presence areas LLA and CLA as a marking candidate. The fourth threshold value may be the first threshold value or may be smaller than the first threshold value. By making the threshold value smaller than the first threshold value, more marking candidates can be extracted. The marking recognizer 132 may connect line segments of edges whose positions or directions are similar.
Next, the marking recognizer 132 acquires marking state information associated with the vehicle position closest to the position information (in other words, marking state information recognized finally in a state in which recognition accuracy has not been lowered in the marking recognition process) with reference to the vehicle position of the recognized marking information 192 using the position information of the vehicle M, compares the acquired marking state information with the marking candidate information, and recognizes a marking of the traveling lane from the marking candidates.
In the example of
The marking recognizer 132 may not recognize the marking candidate as a marking when the degree of similarity is less than a prescribed value. The marking recognizer 132 may recognize the marking in each of the marking presence areas LLA and CLA.
The operation controller (the action plan generator 140 and the second controller 160) executes LKAS control on the basis of the marking recognized by the marking recognizer 132.
As described above, in the embodiment, even if the marking recognition accuracy is lowered, it is possible to improve the recognition accuracy or reliability of the marking by extracting an edge of an area having a high probability of presence of the marking and recognizing the marking on the basis of the extracted edge. The number of processing resources can be reduced by recognizing the markings in a limited area.
For example, when a traveling state of the vehicle M is output to the HMI 30 in a driving assistance process or the like, the HMI controller 170 may cause the display device of the HMI 30 to display an image related to the marking recognized by the marking recognizer 132. In this case, the HMI controller 170 may cause markings recognized in a state in which the recognition accuracy has been lowered and a state in which the recognition accuracy has not been lowered to be displayed in different display modes (for example, a color change mode, a blink display mode, a pattern change mode, and the like). The HMI controller 170 may cause information indicating that the marking recognition accuracy has been lowered to be output from the HMI 30. Thereby, the occupant can be notified of the state of the vehicle M more accurately.
In the example of
Subsequently, the marking recognizer 132 determines whether or not the marking recognition accuracy has been lowered (step S104). When it is determined that the marking recognition accuracy has been lowered, the marking recognizer 132 extracts a marking presence area having a high likelihood of presence of the marking as an example of a prescribed area within the recognizable range RA of the external sensor ES (step S106). In the processing of step S106, the marking recognizer 132 may extract a marking presence area on the basis of the position of the marking, for example, before the marking recognition accuracy is lowered, or may extract a marking presence area with reference to a high-precision map (the second map information 62). The final marking presence area may be extracted on the basis of each extracted marking presence area.
Subsequently, the marking recognizer 132 captures an image of an area including the marking presence area with the camera 10 and extracts an edge within the marking presence area from the captured image (step S108). Subsequently, the marking recognizer 132 extracts marking candidates on the basis of an edge extraction result (step S110) and recognizes a marking on the basis of degrees of similarity between the extracted marking candidates and a marking recognition result acquired before the marking recognition accuracy is lowered (step S112).
When it is determined that the marking recognition accuracy has not been lowered after the processing of step S112 or in the processing of step S104, the driving controller (the action plan generator 140 and the second controller 160) executes a driving control process such as LKAS on the basis of the recognized marking (step S114). Thereby, the process of the present flowchart ends.
In the above-described embodiment, the marking recognizer 132 may perform the above-described marking recognition process even if the driving control process other than LKAS is being executed. For example, when an ALC control process is performed, the marking recognizer 132 may recognize not only markings for dividing the traveling lane but also markings for dividing an adjacent lane that is a lane change destination. When a marking presence area is extracted, the marking recognizer 132 may extract a marking presence area with respect to the traveling lane on the basis of a position, a direction, or the like of the marking of the lane (the lane L2) other than the traveling lane (the lane L1) of the vehicle M, for example, as shown in
Although an edge included in an image captured by the camera 10 is mainly extracted and a marking is recognized on the basis of an extraction result in the above-described example, an edge extraction process may be performed on the basis of a detection result (LIDAR data) of the LIDAR sensor 14 included in the external sensor ES in addition to (or instead of) the above-described edge extraction process. If there is a physical object with irregularities such as a protective barrier of a guardrail or the like, a chatter bar, a curb, or a median strip, the marking may be recognized or the marking presence area may be extracted on the basis of detection results of the radar device 12 and/or the SONAR sensor. Thereby, the marking can be recognized with higher accuracy.
The marking recognizer 132 may recognize a traveling lane by comparing a marking pattern (for example, an arrangement of solid lines and broken lines) obtained from the second map information 62 with a marking pattern near the vehicle M recognized from the image captured by the camera 10. The marking recognizer 132 may recognize the traveling lane by recognizing a traveling path boundary (a road boundary) including a marking, a road shoulder, a curb, a median strip, a guardrail, and the like as well as the marking. In this recognition, a position of the vehicle M acquired from the navigation device 50 and a processing result by the INS may be added.
When the marking recognition accuracy has been lowered and the marking recognizer 132 can recognize no marking, the HMI controller 170 may cause the HMI 30 to output information indicating that no marking can be recognized or may cause the HMI 30 to output information for prompting the occupant of the vehicle M to perform manual driving by ending the LKAS control process.
According to the above-described embodiment, the automated driving control device 100 (an example of a mobile object control device) includes the recognizer 130 configured to recognize a surrounding situation of the vehicle M (an example of a mobile object) on the basis of an output of the external sensor ES; and the marking recognizer 132 configured to recognize markings for dividing an area through which the vehicle M passes on the basis of the surrounding situation recognized by the recognizer 130, wherein the marking recognizer 132 extracts a prescribed area from the surrounding situation when it is determined that marking recognition accuracy has been lowered, extracts an edge within the extracted prescribed area, and recognizes the markings on the basis of an extraction result, whereby it is possible to further improve the accuracy of recognition of markings for dividing an area through which the vehicle M passes.
Specifically, according to the embodiment, for example, it is possible to recognize a marking with high accuracy more efficiently by extracting only a prescribed area from an image captured by the camera on the basis of the vicinity of a previous marking recognition result, the vicinity of a boundary of segmentation, and the vicinity of a position where there is a marking of high-precision map information and extracting an edge only in the extracted area. According to the embodiment, it is possible to improve the certainty of a marking by collating information of a learning-based marking recognized in a state in which the recognition accuracy has not been lowered with information of the marking recognized in the edge extraction process. According to the embodiment, it is possible to improve the accuracy or reliability of lane recognition and limit erroneous detection of markings by selecting a marking similar to a marking used in a previous control process with respect to state information of marking candidates obtained in the edge extraction process.
The embodiment described above can be represented as follows.
A mobile object control device including:
a storage device storing a program; and
a hardware processor,
wherein the hardware processor executes the program stored in the storage device to:
recognize a surrounding situation of a mobile object on the basis of an output of an external sensor;
recognize markings for dividing an area through which the mobile object passes on the basis of the recognized surrounding situation;
extract a prescribed area from the surrounding situation when it is determined that marking recognition accuracy has been lowered;
extract an edge within the extracted prescribed area; and
recognize the markings on the basis of an extraction result.
While preferred embodiments of the invention have been described and illustrated above, it should be understood that these are exemplary of the invention and are not to be considered as limiting. Additions, omissions, substitutions, and other modifications can be made without departing from the spirit or scope of the present invention. Accordingly, the invention is not to be considered as being limited by the foregoing description, and is only limited by the scope of the appended claims.
Number | Date | Country | Kind |
---|---|---|---|
2021-089427 | May 2021 | JP | national |