The present application is based upon and claims priority to Chinese Patent Application No. 201811015653.1, filed on Aug. 31, 2018, the entirety contents of which are incorporated herein by reference.
The present disclosure relates to a technical field of intelligent transportation, and more particularly, to an intelligent roadside unit.
Currently, an intelligent roadside unit is an important support for autonomous driving. As requirements on intelligence level of the intelligent roadside unit increases, demands on the sensing capability of the intelligent roadside unit is becoming higher and higher, such that various sensing detectors need to be disposed on the intelligent roadside unit to improve the active sensing of the intelligent roadside unit. However, since the intelligent roadside unit needs to process a large amount of data, the reliability may be decreased due to the high temperature and heavy processing burden, causing a security risk in the navigation of the intelligent driving vehicle when a malfunction of the intelligent roadside unit occurs.
The present disclosure aims at solving at least one of the technical problems in the related art to some extent.
To this end, an objective of the present disclosure is to provide an intelligent roadside unit.
To achieve the above objectives, embodiments of the present disclosure propose an intelligent roadside unit, including: a radar configured to detect an obstacle within a first preset range of the intelligent roadside unit; a camera configured to capture an image of a second preset range of the intelligent roadside unit; a master processor coupled to the radar and the camera, and configured to generate a point cloud image according to information on the obstacle detected by the radar and the image detected by the camera; and a slave processor coupled to the radar and the camera, and configured to generate a point cloud image according to the information on the obstacle detected by the radar and the image detected by the camera, wherein the slave processor checks the master processor, and when the master processor breaks down, it is switched from the master processor to the slave processor.
Additional aspects and advantages of the present disclosure will be given in the following description, some of which will become apparent from the following description or be learned from practices of the present disclosure.
The foregoing and/or additional aspects and advantages of the present disclosure become obvious and easily understood in descriptions of the embodiments with reference to the following accompanying drawings.
Embodiments of the present disclosure are described below in detail, examples of the embodiments are shown in accompanying drawings, and reference numerals that are the same or similar from beginning to end represent the same or similar components or components that have the same or similar functions. The embodiments described below with reference to the accompanying drawings are exemplary, are merely used to explain the present disclosure, and should not be construed as a limit to the present disclosure.
Hereinafter, the intelligent roadside unit according to embodiments of the present disclosure will be described with reference to the drawings.
The intelligent roadside unit provided by embodiments of the present disclosure may be applied to an automatic vehicle.
The radar 10 is configured to detect an obstacle within a first preset range of the intelligent roadside unit.
The camera 20 is configured to capture an image of a second preset range of the intelligent roadside unit.
The master processor 30 coupled to the radar 10 and the camera 20 is configured to generate a point cloud image according to information on the obstacle detected by the radar 10 and the image detected by the camera 20.
The slave processor 40 coupled to the radar 10 and the camera 20 is configured to generate a point cloud image according to the information on the obstacle detected by the radar 10 and the image detected by the camera 20. The slave processor 40 checks the master processor 30, and when the original master processor 30 breaks down, it is switched from the master processor to the slave processor.
Specifically, one or more radars 10 may be arranged according to practical application requirements to detect an obstacle within the first preset range of the intelligent roadside unit. The first preset range may be set according to practical application requirements.
In a possible implementation, the intelligent roadside unit according to embodiments of the present disclosure is applied to a vehicle A, and the radar 10 may detect an obstacle (for example, another vehicle, a pedestrian or a building that is less than one meter from the vehicle A) within a certain range from the vehicle A.
In a possible implementation, as shown in
In practical applications, the first radar 101 is configured to detect a distant obstacle while the second radar 102 is configured to detect a nearby obstacle. A detection range of the first radar 101 may be arranged broader than a detection range of the second radar 102, thereby further improving the reliability of the radar on detecting an obstacle and lowering the costs of the intelligent roadside unit.
In a possible implementation, the intelligent roadside unit according to embodiments of the present disclosure is applied to the vehicle A. The second radar 102 detects an obstacle such as another vehicle, a pedestrian and a building of which the distance from the vehicle A is less than one meter; and the first radar 101 detects an obstacle such as another vehicle, a pedestrian and a building that is of which the distance from the vehicle A is greater than one meter but less than two meters.
It should be noted that the intelligent roadside unit according to embodiments of the present disclosure is applied to a field of automatic drive. The second radar 102 is mainly applied to blind zone detection (blind zone monitoring), lane keeping and lane changing assist, collision alarm or anti-collision protection of a rear radar, parking assist, traffic monitoring at an intersection and on the like.
It should be noted that the intelligent roadside unit according to embodiments of the present disclosure is applied to the field of automatic drive. The first radar 101 is applied to braking assist, autonomous distance control and on the like.
It should be noted that types of the first radar 101 and the second radar 102 may be chosen according to practical application requirements to further improve the obstacle detection ability. Examples are illustrated as follows.
In a first example, the first radar 101 and the second radar 102 are both laser radars. A laser radar is mainly used for measuring a distance from a fixed or movable object.
The first radar 101 is a 64-channel laser radar, and the second radar 102 is a 16-channel laser radar.
As a second example, the first radar 101 is a laser radar, and the second radar 102 is a millimeter-wave radar. A millimeter-wave radar is the only sensor that works around the clock, and its accuracy of speed measurement and ranging is much higher than that of vision.
It should be noted that positions of the radar 10 and the camera 20 may be arranged according to practical application requirements. In a possible implementation, the first radar 101 may be arranged above the camera 20, and the second radar 102 may be arranged below the camera 20.
Specifically, one or more cameras 20 may be arranged according to practical application requirements to capture an image of the second preset range of the intelligent roadside unit. The second preset range may be set according to practical application requirements.
In a possible implementation, the intelligent roadside unit according to embodiments of the present disclosure is applied to the vehicle A, and the camera 20 may capture an image (for example, another vehicle, a pedestrian or an intersection indicator that is less than one meter from the vehicle A) of a certain range of the vehicle A.
As an example, the at least one camera 20 may be a plurality of cameras 20, i.e., the intelligent roadside unit may include a plurality of cameras 20. The plurality of cameras 20 correspond to intersections monitored by the intelligent roadside unit respectively. By setting cameras 20 at different positions and in different directions, the accuracy and reliability of the point cloud image may be further improved.
Specifically, both the master processor 30 and the slave processor 40 are coupled to the radar 10 and the camera 20. It may be switched from the master processor 30 to the slave processor 40 when the original master processor 30 breaks down, and the point cloud image may be generated according to the information on the obstacle detected by the radar 10 and the image detected by the camera 20.
The slave processor 40 may adopt a variety of methods to detect the master processor 30. In a possible implementation, the slave processor 40 is configured to send a heartbeat detection message to the master processor 30, and when the slave processor 40 does not receive a response message fed back by the master processor 30 within a preset period, it is determined that the master processor 30 breaks down. In another possible implementation, a breakdown detection sensor is arranged to check the master processor 30, and the breakdown detection sensor sends a check result to the slave processor 40, such that the slave processor 40 may check out whether the master processor 30 breaks down according to the check result.
With the intelligent roadside unit according to embodiments of the present disclosure, the radar is configured to detect an obstacle within a first preset range of the intelligent roadside unit; the camera is configured to capture an image of a second preset range of the intelligent roadside unit; the master processor is coupled to the radar and the camera, and is configured to generate a point cloud image according to information on the obstacle detected by the radar and the image detected by the camera; and the slave processor is coupled to the radar and the camera, and is configured to generate a point cloud image according to the information on the obstacle detected by the radar and the image detected by the camera. The slave processor checks the master processor, and when the original master processor breaks down, it is switched from the master processor to the slave processor. By setting a main processor and a secondary processor (i.e., a master processor and a slave processor, or a primary processor and a backup processor), it may be guaranteed that when one processor breaks down, the other processor may be switched on timely for subsequent processing, thereby improving the reliability of the intelligent roadside unit.
The communicator 50 is coupled to the master processor 30 and the slave processor 40. The currently active processor 30 is configured to send the point cloud image to the self-driving vehicle or the server close to the intelligent roadside unit through the communicator 50.
The communicator 50 may be an antenna or a wireless connection device.
Specifically, the master processor 30 is configured to control the traffic lights 60 according to the point cloud image. Consequently, the applicability of the intelligent roadside unit is further improved.
In an embodiment of the present disclosure, the intelligent roadside unit further includes a shielding layer wrapping at least a part of the camera 20.
As a possible implementation, the shielding layer wraps the camera 20 excluding the lens and the heat dissipation portion.
Consequently, wrapping the camera 20 except for the lens and the heat dissipation portion with a shielding material may eliminate interferences caused by components such as the radar 10 and the antenna on the camera 20, thereby increasing an imaging clarity of the camera 20 and further improving the reliability of the intelligent roadside unit.
In the description of the present disclosure, reference throughout this specification to “an embodiment,” “some embodiments,” “example,” “a specific example,” or “some examples,” means that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the present disclosure. In the specification, the terms mentioned above are not necessarily referring to the same embodiment or example of the present disclosure. Furthermore, the particular features, structures, materials, or characteristics may be combined in any suitable manner in one or more embodiments or examples. In addition, different embodiments or examples and features of different embodiments or examples described in the specification may be combined by those skilled in the art without contradictions.
In addition, terms such as “first” and “second” are used herein for purposes of description and are not intended to indicate or imply relative importance or significance. Thus, the feature defined with “first” and “second” may comprise one or more this feature. In the description of the present disclosure, “a plurality of” means at least two, for example, two or three, unless specified otherwise.
Any process or method described in a flow chart or described herein in other ways may be understood to include one or more modules, segments or portions of codes of executable instructions for achieving specific logical functions or steps in the process, and the scope of a preferred embodiment of the present disclosure includes other implementations, which should be understood by those skilled in the art.
The logic and/or step described in other manners herein or shown in the flow chart, for example, a particular sequence table of executable instructions for realizing the logical function, may be specifically achieved in any computer readable medium to be used by the instruction execution system, device or equipment (such as the system based on computers, the system comprising processors or other systems capable of obtaining the instruction from the instruction execution system, device and equipment and executing the instruction), or to be used in combination with the instruction execution system, device and equipment. As to the specification, “the computer readable medium” may be any device adaptive for including, storing, communicating, propagating or transferring programs to be used by or in combination with the instruction execution system, device or equipment. More specific examples of the computer readable medium comprise but are not limited to: an electronic connection (an electronic device) with one or more wires, a portable computer enclosure (a magnetic device), a random access memory (RAM), a read only memory (ROM), an erasable programmable read-only memory (EPROM or a flash memory), an optical fiber device and a portable compact disk read-only memory (CDROM). In addition, the computer readable medium may even be a paper or other appropriate medium capable of printing programs thereon, this is because, for example, the paper or other appropriate medium may be optically scanned and then edited, decrypted or processed with other appropriate methods when necessary to obtain the programs in an electric manner, and then the programs may be stored in the computer memories.
It should be understood that each part of the present disclosure may be realized by the hardware, software, firmware or their combination. In the above embodiments, a plurality of steps or methods may be realized by the software or firmware stored in the memory and executed by the appropriate instruction execution system. For example, if it is realized by the hardware, likewise in another embodiment, the steps or methods may be realized by one or a combination of the following techniques known in the art: a discrete logic circuit having a logic gate circuit for realizing a logic function of a data signal, an application-specific integrated circuit having an appropriate combination logic gate circuit, a programmable gate array (PGA), a field programmable gate array (FPGA), etc.
It would be understood by those skilled in the art that all or a part of the steps carried by the method in the above-described embodiments may be completed by relevant hardware instructed by a program. The program may be stored in a computer readable storage medium. When the program is executed, one or a combination of the steps of the method in the above-described embodiments may be completed.
In addition, individual functional units in the embodiments of the present disclosure may be integrated in one processing module or may be separately physically present, or two or more units may be integrated in one module. The integrated module as described above may be achieved in the form of hardware, or may be achieved in the form of a software functional module. If the integrated module is achieved in the form of a software functional module and sold or used as a separate product, the integrated module may also be stored in a computer readable storage medium.
The storage medium mentioned above may be read-only memories, magnetic disks or CD, etc. Although explanatory embodiments have been shown and described, it would be appreciated by those skilled in the art that the above embodiments cannot be construed to limit the present disclosure, and changes, alternatives, and modifications can be made in the embodiments without departing from scope of the present disclosure.
Number | Date | Country | Kind |
---|---|---|---|
201811015653.1 | Aug 2018 | CN | national |
Number | Name | Date | Kind |
---|---|---|---|
5289183 | Hassett | Feb 1994 | A |
6466862 | DeKock | Oct 2002 | B1 |
6489920 | Anders | Dec 2002 | B1 |
6580385 | Winner | Jun 2003 | B1 |
7460951 | Altan | Dec 2008 | B2 |
7541943 | Manor | Jun 2009 | B2 |
7573400 | Arnold | Aug 2009 | B2 |
7889098 | Arnold | Feb 2011 | B1 |
8692690 | Dalal | Apr 2014 | B2 |
9684064 | Nagy | Jun 2017 | B2 |
9791557 | Wyrwas | Oct 2017 | B1 |
10436898 | Kurata | Oct 2019 | B2 |
10816654 | Rohani | Oct 2020 | B2 |
20020145541 | Matsui | Oct 2002 | A1 |
20050102562 | Shinohara et al. | May 2005 | A1 |
20060101201 | Vageline | May 2006 | A1 |
20080175438 | Alves | Jul 2008 | A1 |
20090219172 | Wilbrod | Sep 2009 | A1 |
20100097262 | Hong | Apr 2010 | A1 |
20100328140 | Hong | Dec 2010 | A1 |
20110102237 | Hong | May 2011 | A1 |
20110291876 | Hong | Dec 2011 | A1 |
20120188115 | Hong | Jul 2012 | A1 |
20130151135 | Aubrey | Jun 2013 | A1 |
20140089731 | Lee | Mar 2014 | A1 |
20140195138 | Stelzig | Jul 2014 | A1 |
20140240167 | Cho | Aug 2014 | A1 |
20140307087 | Evanitsky | Oct 2014 | A1 |
20140312749 | Paetzoldt | Oct 2014 | A1 |
20140333469 | Nagy | Nov 2014 | A1 |
20150360565 | Goto et al. | Dec 2015 | A1 |
20170090015 | Breen | Mar 2017 | A1 |
20180017675 | Hong | Jan 2018 | A1 |
20180059236 | Wodrich | Mar 2018 | A1 |
20180182243 | Baba | Jun 2018 | A1 |
20180278919 | Lee | Sep 2018 | A1 |
20190235504 | Carter | Aug 2019 | A1 |
20190256091 | Lepp | Aug 2019 | A1 |
20190339382 | Hess | Nov 2019 | A1 |
20190353774 | Chondro | Nov 2019 | A1 |
20190377086 | Rogan | Dec 2019 | A1 |
20200025575 | Weissman | Jan 2020 | A1 |
20200298842 | Takamatsu | Sep 2020 | A1 |
Number | Date | Country |
---|---|---|
102571443 | Jul 2012 | CN |
107945540 | Apr 2018 | CN |
108010360 | May 2018 | CN |
108010360 | May 2018 | CN |
108182817 | Jun 2018 | CN |
2003302470 | Oct 2003 | JP |
2006085246 | Mar 2006 | JP |
2009199532 | Sep 2009 | JP |
2009201098 | Sep 2009 | JP |
2012531138 | Dec 2012 | JP |
2017000189 | Jan 2017 | JP |
Entry |
---|
Office Action for Japanese Application No. 2019-158725, dated Oct. 13, 2020, 6 pages. |
Extended European Search Report for EP Application No. 19191467.0, dated Feb. 14, 2020, 8 pages. |
Office Action for Chinese Application No. 201811015653.1, dated Jul. 15, 2021, 8 pages. |
Decision on Rejection for Japanese Application No. 2019-158725, dated Apr. 27, 2021, 3 pages. |
Office Action for European Application No. 19191467.0, dated Oct. 29, 2021, 6 pages. |
Number | Date | Country | |
---|---|---|---|
20200072962 A1 | Mar 2020 | US |