Automatic driving control device

Information

  • Patent Grant
  • 10518783
  • Patent Number
    10,518,783
  • Date Filed
    Friday, March 4, 2016
    8 years ago
  • Date Issued
    Tuesday, December 31, 2019
    4 years ago
Abstract
Control can be continued so that control of automatic driving is not canceled even in a non-average specific road environment. An automatic driving control device, which controls an actuator of a vehicle in order to automatically travel on a road without driver operation through the use of an external surrounding recognition sensor for recognizing external surroundings and information from the external surrounding recognition sensor, has a learning recording device for recording external surrounding recognition information and vehicle status information when the driver is driving, an output comparison device for comparing the information recorded in the learning recording device with the processing results when the information recorded in the recording device is processed by the automatic driving control device, and a control parameter setting device for setting control parameters so that the results compared by the output comparison device approach the driving of the driver.
Description
TECHNICAL FIELD

The present invention relates to an automatic driving control device, and especially relates to setting control parameters.


BACKGROUND ART

PTL 1 describes an exemplary method for setting parameters of an automatic driving control device. For example the control device learns the parameters for controlling lane changes. The control device compares the frequency of lane changes on the market with the frequency of lane changes made while a driver manually drives a vehicle, and learns the comparison in order to bring the parameters close to the frequency of the driver's lane changes.


An automatic driving control device described in PTL 2 compares a feature value generated from a model driver in the automatic driving control device such as vehicle speed, a vehicle-to-vehicle distance, or the manipulated variable of the gas pedal with the feature value generated while the driver drives the vehicle and learns the comparison to optimize the parameter of the model driver in order to maximize the probability of agreement between the feature value of the model driver and the feature value of the driving of the driver.


CITATION LIST
Patent Literature



  • PTL 1: JP 2008-180591 A

  • PTL 2: JP 2007-176396 A



SUMMARY OF INVENTION
Technical Problem

However, it may be impossible for the automatic driving control devices in PTL 1 and PTL 2 to respond to each individual road environment although the automatic driving control devices perform control so that a feature value such as the frequency of lane changes, the vehicle speed, the vehicle-to-vehicle distance, or the manipulated variable of the gas pedal when the driver manually drives the vehicle in a certain interval agrees with the feature value when the driving is automated.


For example, when the driver commutes by car, the road environment changes between the alley in front of the driver's house, a main road in a city area, a highway between cities, a main road in an industrial area, and an alley near the factory.


If the parameters are adjusted based on the average values or precisions of the feature value, or the parameters are adjusted so that the precision of the feature value is agreed with such a road environment, the vehicle is properly controlled on the highway between cities with the longest travel distance. On the other hand, the vehicle is not properly controlled or the driver feels discomfort on the other roads of the road environment.


Especially, when the vehicle is not properly controlled while the driving is automated, it is necessary to cancel the control. For example, there are some places in which the control is canceled during the commute. This means that it may be impossible to use automatic driving on a daily basis. This spoils the value of the automatic driving device.


An objective of the present invention is to continue control so that control of automatic driving is not canceled even in a non-average specific road environment.


Solution to Problem

An automatic driving control device according to the present invention is, for example, an automatic driving control device that controls an actuator of a vehicle in order to cause the vehicle to automatically travel according to an external surrounding recognition sensor that recognizes external surroundings and information from the external surrounding recognition sensor without driver operation, and the automatic driving control device includes: a learning recording device that records external surrounding recognition information and vehicle condition information when the driver drives the vehicle; an output comparison device that compares a result of a process in which the automatic driving control device processes the information recorded in the recording device with the information recorded in the learning recording device; and a control parameter setting device that sets a control parameter so that the result compared by the output comparison device approaches driving of the driver.


Advantageous Effects of Invention

According to the invention, control can be continued so that control of automatic driving is not canceled even in a non-average specific road environment.





BRIEF DESCRIPTION OF DRAWINGS


FIG. 1 is a block diagram illustrating an embodiment of the present invention.



FIG. 2 is a block diagram of the control of automatic driving.



FIG. 3 is a schematic diagram of a road for describing exemplary control according to the present invention.



FIG. 4 is a diagram illustrating transition from a normal automatic control mode to a learning mode.



FIG. 5 is a data flow diagram while a driver drives a vehicle.



FIG. 6 is a flowchart of a process for evaluating an automatic driving parameter.



FIG. 7 is a block diagram of control for using the data when the driver drives the vehicle.



FIG. 8 is a flowchart of a process for calculating the position of the driver's vehicle when the driving is automated according to the data when the driver drives the vehicle.



FIG. 9 is a diagram for describing a resetting point on a map.



FIG. 10 is a flowchart of a process for learning the driver's driving.



FIG. 11 is a diagram of ranges that an external surrounding recognition sensor recognizes.





DESCRIPTION OF EMBODIMENTS

The whole configuration of a control device according to the present invention will be described with referent to FIG. 1. An automatic driving ECU 105 obtains the information about the road environment around the driver's vehicle from an external surrounding recognition sensor to be described below.


The external surrounding recognition sensor includes a front radar 1011, a right front radar 1012, a left front radar 1013, a right rear radar 1015, a left rear radar 1014, a front camera 102, and a surrounding camera 1021. The sensing range of each of the radars and cameras is illustrated in FIG. 11. The sensing range of the front radar 1011 is indicated with 1107. The sensing range of the right front radar 1012 is indicated with 1105. The sensing range of the left front radar 1013 is indicated with 1105. The sensing range of the right rear radar 1015 is indicated with 1102. The sensing range of the left rear radar 1014 is indicated with 1101. The sensing range of the front camera 102 is indicated with 1106. The sensing range of the surrounding camera 1021 is indicated with 1103.


The position of an object around the driver's vehicle and the relative speed are obtained from the front radar 1011, the right front radar 1012, the left front radar 1013, the right rear radar 1015, the left rear radar 1014, and the surrounding camera 1103.


The position of an object in front of the driver's vehicle, the relative speed, the position of a white line on the road, the position of the boundary between the road and another area, the type and position of a traffic sign, and the color and position of a traffic signal are obtained from the front camera 102.


The road shape information that is not obtained from the front camera 102, for example, a radius R and length of a curve in front of the driver's vehicle, the inclination and length of a steep, the increase or decrease in the number of lanes, the road regulation information, the position of an intersection, or the position of a turnpike is obtained from a map ECU 103. The map ECU 103 is connected to a GPS 104 to obtain the current position of the driver's vehicle.


The position and speed of a peripheral object, the position and color of a traffic signal, and the position and type of a traffic sign that are not obtained from the front camera 102 and the radars 1011 to 1015 are obtained from a vehicle-to-vehicle/road-to-vehicle communication ECU 117.


The automatic driving control device (automatic driving ECU) 105 controls the driver's vehicle by controlling an actuator to be described below.


Controlling an engine ECU 110 controls acceleration of the driver's vehicle. Controlling a brake ECU 111 controls deceleration of the driver's vehicle. Controlling a steering ECU 113 controls the direction in which the driver's vehicle travels.


The automatic driving ECU 105 further obtains the conditions of the driver's vehicle and a driving state in which the driver drives the vehicle from sensors to be described below. The opening angle of the throttle valve opened by the driver is obtained from a throttle valve sensor 106. The yaw rate of the driver's vehicle is obtained from a yaw rate sensor 107. The speed of the driver's vehicle is obtained from a vehicle speed sensor 108. The manipulated variable of the steering wheel operated by the driver (the steering angle) is obtained from a steering angle sensor 109.


A process performed in the automatic driving ECU 105 in a normal control will be described with reference to FIG. 2. An object determination unit 203 extracts an object for which the driver's vehicle is likely to be controlled from sensor data 201 obtained from the radars 1011 to 1015, the front camera 102, the map ECU 103, and the C2X ECU 117 illustrated in FIG. 1. The sensor data includes the information that is not related to the control of the driver's vehicle so much, such as noise. Thus, the object determination unit 203, for example, removes a pebble-sized object from the objects for which the driver's vehicle is controlled to stop or curve.


Next, the object for which the driver's vehicle is likely to be controlled, which is extracted by the object determination unit 203, and the peripheral object information obtained from the vehicle-to-vehicle/road-to-vehicle communication data 204 are mapped on the map information obtained from the map data 202 in 205.


The weighting unit 207 assigns a weight to the object for which the driver's vehicle is to be controlled, for example, according to the position mapped on the map, the type of the object, and the traveling direction in which the driver's vehicle travels.



FIG. 3 illustrates an exemplary operation of the weighting unit 207. There is a traffic signal 301 on a road 305 in the direction in which the driver's vehicle travels and a bicycle 302 approaches the driver's vehicle on a road 306 crossing the road 305. In addition, a pedestrian 303 walks in the same direction as the direction in which the driver's vehicle travels. In such a case, the weight assigned to the traffic signal 301 is the heaviest, the weight assigned to the bicycle 302 is the second heaviest, and the weight assigned to the pedestrian 303 is the lightest.


The weight assigned to each object at each place is set as a parameter in a control parameter 2016. For example, the traffic signal 301 has a weight of 10. The weight of the bicycle 302 entering a crosswalk on a sidewalk is set according a Time-to-collision (TTC) table 1. In this example, the driver's vehicle 304 is controlled according to the traffic signal until the TTC of the bicycle 302 becomes one second.









TABLE 1







TTC table













TTC
10
3
2
1







Weight
5
6
8
11










According to the weights assigned by the weighting unit 207, an object for which the driver's vehicle is to be controlled determination unit 208 determines an object for which the driver's vehicle is to be controlled. A control unit 209 calculates a control quantity relative to the object for which the driver's vehicle is to be controlled, which the object for which the driver's vehicle is to be controlled determination unit 208 determines as described above, in order to control the actuator 210.


The control unit 209 determines the distance to the object for which the driver's vehicle is to be controlled, the speed of the driver's vehicle, the acceleration and deceleration of the driver's vehicle. Such control quantities are determined according to the control parameter 2016.


The process performed in the normal control has been described. The modes other than the normal control will be described with reference to FIG. 4.


There are an OFF mode 401 in which the device does not control the driver's vehicle, a standby mode 402 for waiting the start of control, a normal control mode 403, and a learning mode 406 of the device according to the present invention.


The OFF mode 401 is a state in which the driver drives the vehicle and the automatic driving is not performed. The standby mode 402 is a state in which the driver drives the vehicle similarly to 401. At the same time, the standby mode 402 is also a state in which the fact that automatic driving control is kept on standby is displayed on a display device such as a meter and, once the driver operates, for example, a switch SW, the device can be shifted to the normal control 403 or the learning mode 406. The OFF mode 401 and the standby mode 402 are provided as described above in order to prevent an error in the driver's switch operation or the like to suddenly shift the device to the normal control 403.


The learning mode 406 includes a data recording mode 404 and a parameter optimizing mode 405. These modes will be described with reference to FIGS. 5 and 6.


In the data recording mode 404, all of the sensor data 201 and the vehicle condition information 502 are recorded on a data recording medium 501 as illustrated in FIG. 5.


The vehicle condition 502 is the information obtained, for example, from the throttle valve sensor 106, the yaw rate sensor 107, the vehicle speed sensor 108, and the steering angle sensor 109. A sensor information item about the inside of the driver's vehicle may also be recorded as the vehicle condition 502.


The data recording mode 404 are maintained from the time the driver requests the transition from the standby mode 402 to the learning mode 406 to the time the driver requests the termination of the learning mode 406.


For example, in order to perform the learning of control on the road of the driver's commute route, the driver needs to shift the mode to the learning mode 406 at the garage of the driver's house and terminate the learning mode 406 at the parking lot of the driver's office.


The process performed in the parameter optimizing mode 405 will be described next with reference to FIG. 6.


Here, the traveling data 501 recorded while the driver drives the vehicle is compared with prediction data that predicts the traveling data when automatic driving is performed in the same conditions and the control parameters are set so that the comparison result is lower than or equal to a threshold. A specific setting process will be described below.


The data recorded on the data recording medium 501 in the data recording mode 404 is input into an automatic driving control quantity calculation unit 200 and used to calculate the automatic driving control quantity when automatic driving is performed under the same conditions as the conditions under which the driver drives the vehicle. The automatic driving control quantity calculation unit 200 calculates the quantity of control during the automatic driving. As illustrated in FIG. 7, the traveling data 501 is used for the calculation instead of the sensor data 201, the map data 202, and the C2X data 203 illustrated in FIG. 2.


Next, an external surrounding recognition information prediction calculation unit 601 calculates the predicted traveling data when automatic driving is performed according to the control quantity calculated by the automatic driving control quantity calculation unit 200.


A method for calculating the predicted traveling data will be described with reference to FIG. 8.


The control quantity from the automatic driving control quantity calculation unit 200, and target acceleration and target steering angle 811 are input to a model vehicle 801 so that predicted speed, predicted steering angle, and predicted opening angle of the throttle valve 812 are calculated. The model vehicle 801 includes a dead time and a first-order time delay coefficient with respect to the control quantity. A mechanism to make the vehicle respond to a target physical quantity, such as the opening angle of the throttle valve is determined, for example, according to the map. For example, the predicted speed is calculated by integrating a value that is the target acceleration added to the dead time and the first-order time delay, and adding the integrated value to a default speed. Similarly, the predicted steering angle is obtained by adding the dead time and first-order time delay to the target steering angle. The predicted opening angle of the throttle valve is obtained by subtracting the map of opening angle of the throttle valve from the target acceleration and the corresponding gear ratio.


Next, a vehicle vector calculation unit 802 calculates driver's vehicle speeds in a longitudinal direction and a lateral direction 813 according to the predicted speed and predicted steering angle. A traveling angle θ of the traveling direction in which the driver's vehicle travels is calculated from the predicted steering angle. From this calculation, the driver's vehicle speed in a longitudinal direction=the predicted speed×sin θ and the driver's vehicle speed in a lateral direction=the predicted speed×cos θ can be found.


Next, a vehicle position difference calculation unit 803 calculates: the difference between the predicted position that is the value of integral of the driver's vehicle speeds in a longitudinal direction and a lateral direction 813 and the position that is the value of integral of the driver's vehicle speeds in a longitudinal direction and a lateral direction 821 provided from the traveling data 501; and deviations of the driver's vehicle in a longitudinal direction and a lateral direction 814.


Next, a relative position calculation unit 804 calculates the relative positions of a peripheral object to the driver's vehicle in a longitudinal direction and a lateral direction when the driving is automated from the deviations of the driver's vehicle in a longitudinal direction and a lateral direction 814 and the position of the peripheral object 822 provided from the traveling data 501.


The relative positions of a peripheral object to the driver's vehicle in a longitudinal direction and a lateral direction are the position of the peripheral object with respect to the driver's vehicle. For example, when a position 910 of the driver's vehicle when the driving is automated is different from a position 900 of the driver's vehicle when the driver drives the vehicle by +0.5 m in a lateral direction 912 and −1.0 min a longitudinal direction 911 as illustrated in FIG. 9, the relative positions of peripheral objects 902 to 904 are slid by +0.5 m in a lateral direction and −1.0 m in a longitudinal direction. The slid positions are set as the relative positions in a lateral direction and in a longitudinal direction. The external surrounding recognition information prediction calculation unit 601 performs such processes.


The value of integral of the driver's vehicle speeds in a longitudinal direction and a lateral direction 821, which is used to find the difference between the positions of the driver's vehicle, is reset when the position of the driver's vehicle on the map recorded on the traveling data 501 reaches a certain point. For example, when the traveling data illustrated in FIG. 9 is recorded as the traveling data 501, the value of integral is reset every time the driver's vehicle 900 reaches a point 901 on the map. The positions of the peripheral objects 902 to 905 taken from the traveling data 501 are also reset and changed to the positions in the traveling data 901 when the driver's vehicle in the traveling data 501 reaches a point 901 on the map. In other words, when the driver's vehicle in the traveling data 501 reaches a point 901 on the map, both of the position of the driver's vehicle and the positions of the peripheral objects are reset and changed to the same positions as the positions when the driver drives the vehicle. Without such resetting, for example, the vehicle fails to recognize the pedestrian 902 illustrated in FIG. 9 when the driving is automated. An objective of the process is not to find an accurate position of the driver's vehicle when the driving is automated, but to determine the control parameters mainly with respect to the driver's vehicle and the peripheral objects with reference to the relative positions when the driver drives the vehicle. Thus, such resetting is performed.


Next, an automatic driving evaluation function 603 will be described. An exemplary evaluation function 603 will be shown as an equation 1. In this equation, the J is an evaluation result of the evaluation function. The smaller the value of J is, the better. The intervals to be integrated are shown as is to that are the time when the driver's vehicle reaches a point 9011 at which resetting in FIG. 9 is performed and the time when the driver's vehicle reaches a point 9012 at which the next resetting is performed. The wn is a weight to be assigned to a variable to be evaluated. The weight is previously set for each variable. The weight may be used as a control parameter and be changed with the evaluation result by the evaluation function. The vd is the speed at which the driver drives the vehicle and obtained from the traveling data 501. The va is a predicted speed of the driver's vehicle when the driving is automated and obtained from the external surrounding recognition information prediction calculation unit 601. Similarly, the dyd is the longitudinal distance between the driver's vehicle and a peripheral object when the driver drives the vehicle and obtained from the traveling data 501. The dya is the predicted longitudinal distance between the driver's vehicle and a peripheral object when the driving is automated and obtained from the external surrounding recognition information prediction calculation unit 601. Each subtraction such as (Vd−Va) is squared. This prevents the result of comparison of the data when the driver drives the vehicle with the data when the driving is automated to have a negative value, and correctly integrates the difference between the data when the driver drives the vehicle and the data when the driving is automated. As described above, the differences in all or a part of the data obtained when the driver drives the vehicle and the data obtained when the driving is automated are integrated. Then, the driving by the driver and the automatic driving are compared. Not only the obtained data but also the differentiated values may be used for the comparison. For example, the vehicle speeds are differentiated so that the accelerations can be compared.

[Expression 1]
J=∫tetsw1(vd−va)2+w2(dd−da)2 . . . dx  Equation 1:


Next, a control parameter adjustment unit 603 will be described with reference to FIG. 10.


First, the control parameter adjustment unit 603 determines in 1002 whether the result of the evaluation with the evaluation function is lower than or equal to a threshold. If the result is lower than or equal to the threshold, the control parameter adjustment unit 603 records the parameter on the map in 1006. The “recording on the map” means that the control parameters, which are used when control is performed between the points at which the resetting in FIG. 9 is performed such as the points 9011, 9012, and 9013, are associated with the positions of the driver's vehicle on the map and recorded on the map so that the parameters can be used when the driver's vehicle reaches an interval between the points. For example, a parameter is calculated according to the traveling data 501 obtained between the resetting points 9011 and 9012 in FIG. 9. The calculation result is evaluated with the evaluation function equation 1. If the evaluation result is lower than or equal to the threshold, the driver's vehicle is controlled with the parameter between the time when the driver's vehicle reaches the resetting point 9011 in an upward direction of FIG. 9 and the time when the driver's vehicle reaches the resetting point 9012 while the driving is automated.


The parameters recorded on the map may be recorded, for example, on a server installed outside the driver's vehicle and used for controlling another vehicle. Using the parameters recorded on such a server enables the driver's vehicle to be automated with reference to the operation of another vehicle even on a road through which the driver has not passed before.


If the evaluation result is not lower than or equal to the threshold in 1002, the control parameter adjustment unit 603 determines in 1003 whether the evaluations of all the parameters are completed. If the evaluations of all the parameters are completed, this means that the parameters are not properly set. Thus, the information indicating that automatic driving control is not allowed is recorded on the map in 1007. When automatic driving is performed and the vehicle reaches the interval where the information indicating that automatic driving control is not allowed is recorded, the control parameter adjustment unit 603 notifies the driver of the fact that automatic driving control is not allowed and cancels the automatic driving control.


If the control parameter adjustment unit 603 determines in 1003 that the evaluations of all the parameters are not completed, the control parameter adjustment unit 603 newly sets the other parameter in 1004 and returns to the automatic driving control quantity calculation 200. Then, the automatic driving control quantity calculation 200 and the external surrounding recognition information prediction calculation 601 are repeated with the newly set parameter. Then, the control parameter adjustment unit 603 determines whether the evaluation result of the evaluation function equation 1 is lower than or equal to the threshold.


After the parameter is recorded on the map in 1006, or after the information indicating that the automatic driving control is not allowed is recorded on the map in 1007, it is determined whether the intervals between all the resetting points in the traveling data 501 are processed. If all the intervals are not processed, the process goes to the next interval in 1009. If all the intervals are processed, the process is terminated in 10010.


As described above, the parameters are changed and evaluated at the intervals between all the resetting points in the traveling data 501. This optimizes the control parameters.


REFERENCE SIGNS LIST




  • 1011 front radar


  • 1012 right front radar


  • 1013 left front radar


  • 1015 right rear radar


  • 1014 left rear radar


  • 102 front camera


  • 1021 surrounding camera


  • 105 automatic driving control device


  • 501 data recording medium (learning recording device)


  • 602 automatic driving evaluation function (output comparison device)


  • 603 control parameter adjustment unit (control parameter setting unit)


Claims
  • 1. An automatic driving control device that controls an actuator of a vehicle in order to cause the vehicle to automatically travel according to an external surrounding recognition sensor that recognizes external surroundings and information from the external surrounding recognition sensor without driver operation, the automatic driving control device comprising: a learning recording device that records external surrounding recognition information and vehicle condition information when the driver drives the vehicle;an output comparison device that compares a result of a process in which the automatic driving control device processes the information recorded in the recording device with the information recorded in the learning recording device; anda control parameter setting device that sets a control parameter so that the result compared by the output comparison device approaches driving of the driver,wherein the control parameter that is set by the control parameter setting device is associated with a position of the vehicle on a map, andwherein a value of an integral of a speed of the vehicle in a longitudinal direction and a speed of the vehicle in a lateral direction is reset when the position of the vehicle on the map reaches a predetermined point on the map.
  • 2. The automatic driving control device according to claim 1, wherein the control parameter set on the map is read again from the map and control is performed.
  • 3. An automatic driving control device, comprising: an external surrounding recognition sensor that recognizes external surroundings;a learning recording device that records external surrounding recognition information and vehicle condition information when a driver drives a vehicle; andan output comparison device that compares a result of a process in which the automatic driving control device processes the information recorded in the recording device with the information recorded in the learning recording device,wherein the automatic driving control device controls an actuator of the vehicle in order to cause the vehicle to automatically travel according to information from the external surrounding recognition sensor without driver operation, andchanges a weight assigned to a recognized object which is recognized by the external surrounding recognition sensor of the vehicle so that the result compared by the output comparison device approaches driving of the driver,wherein the weight assigned to the recognized object is recorded as a control parameter on a map and is associated with a position of the vehicle on the map, andwherein a value of an integral of a speed of the vehicle in a longitudinal direction and a speed of the vehicle in a lateral direction is reset when the position of the vehicle on the map reaches a predetermined point on the map.
  • 4. An automatic driving control device, comprising: an external surrounding recognition sensor that recognizes external surroundings;a learning recording device that records external surrounding recognition information and vehicle condition information when a driver drives a vehicle; andan output comparison device that compares a result of a process in which the automatic driving control device processes the information recorded in the recording device with the information recorded in the learning recording device,wherein the automatic driving control device controls an actuator of the vehicle in order to cause the vehicle to automatically travel according to information from the external surrounding recognition sensor without driver operation,wherein a control parameter is associated with a position of the vehicle on a map, andwherein a value of an integral of a speed of the vehicle in a longitudinal direction and a speed of the vehicle in a lateral direction is reset when the position of the vehicle on the map reaches a predetermined point on the map.
Priority Claims (1)
Number Date Country Kind
2015-070616 Mar 2015 JP national
PCT Information
Filing Document Filing Date Country Kind
PCT/JP2016/056696 3/4/2016 WO 00
Publishing Document Publishing Date Country Kind
WO2016/158197 10/6/2016 WO A
US Referenced Citations (62)
Number Name Date Kind
7974748 Goerick Jul 2011 B2
9308914 Sun Apr 2016 B1
9610949 Healey Apr 2017 B2
9619638 Dow Apr 2017 B2
9766625 Boroditsky Sep 2017 B2
9950708 Cullinane Apr 2018 B1
10002531 Lynch Jun 2018 B2
10037699 Toyoda Jul 2018 B1
10053093 Thomas Aug 2018 B2
10077056 Fields Sep 2018 B1
10146224 Tafti Dec 2018 B2
20030093210 Kondo May 2003 A1
20080243312 Nakamura Oct 2008 A1
20090234552 Takeda Sep 2009 A1
20100241303 Taguchi Sep 2010 A1
20120083960 Zhu Apr 2012 A1
20120083964 Montemerlo Apr 2012 A1
20120277947 Boehringer Nov 2012 A1
20130304334 Fuehrer Nov 2013 A1
20140371981 Nordbruch Dec 2014 A1
20150073620 Matsumura Mar 2015 A1
20150149017 Attard May 2015 A1
20150166062 Johnson Jun 2015 A1
20150266455 Wilson Sep 2015 A1
20150294422 Carver Oct 2015 A1
20150308837 Cudak Oct 2015 A1
20150309512 Cudak Oct 2015 A1
20160325753 Stein Nov 2016 A1
20160357188 Ansari Dec 2016 A1
20160357262 Ansari Dec 2016 A1
20160358477 Ansari Dec 2016 A1
20170174129 Chin Jun 2017 A1
20170225677 Yoshida Aug 2017 A1
20170282930 Kochhar Oct 2017 A1
20170297586 Li Oct 2017 A1
20170305434 Ratnasingam Oct 2017 A1
20170369073 Huber Dec 2017 A1
20180032082 Shalev-Shwartz Feb 2018 A1
20180061237 Erickson Mar 2018 A1
20180074493 Prokhorov Mar 2018 A1
20180074497 Tsuji Mar 2018 A1
20180089563 Redding Mar 2018 A1
20180143639 Singhal May 2018 A1
20180170392 Yang Jun 2018 A1
20180194280 Shibata Jul 2018 A1
20180203443 Newman Jul 2018 A1
20180253963 Coelho de Azevedo Sep 2018 A1
20180259956 Kawamoto Sep 2018 A1
20180267558 Tiwari Sep 2018 A1
20180281812 Tochioka Oct 2018 A1
20180284774 Kawamoto Oct 2018 A1
20180292222 Lin Oct 2018 A1
20180292830 Kazemi Oct 2018 A1
20180307228 Smith Oct 2018 A1
20180322715 Toyoda Nov 2018 A1
20180328750 Yun Nov 2018 A1
20180336424 Jang Nov 2018 A1
20180339709 Tiwari Nov 2018 A1
20180345981 Ferguson Dec 2018 A1
20180346021 Wang Dec 2018 A1
20180348763 Jiang Dec 2018 A1
20180349802 Jiang Dec 2018 A1
Foreign Referenced Citations (10)
Number Date Country
10 2013 205 840 Oct 2014 DE
H05-004544 Jan 1993 JP
H07-108849 Apr 1995 JP
2003-211999 Jul 2003 JP
2005-092285 Apr 2005 JP
2006-282073 Oct 2006 JP
2007-176396 Jul 2007 JP
2008-180591 Aug 2008 JP
2009-137410 Jun 2009 JP
WO-2009057701 May 2009 WO
Non-Patent Literature Citations (4)
Entry
Japanese Office Action and machine translation thereof issued in corresponding application No. 2017-509441 dated Aug. 21, 2018.
Japanese Office Action and machine translation thereof issued in corresponding application No. 2017-509441 dated May 15, 2018.
International Search Report with English translation and Written Opinion issued in corresponding application No. PCT/JP2016/056696 dated Jun. 7, 2016.
Extended European Search Report dated Oct. 30, 2018 in Application No. 16772081.2.
Related Publications (1)
Number Date Country
20180050702 A1 Feb 2018 US