The present invention relates to a system for detecting surrounding conditions of a moving body that notifies a moving body driver of conditions around the moving body, in particular, surrounding conditions that can be used to facilitate driving of the moving body.
To assist driving of a vehicle (i.e., moving body), a device that notifies a vehicle driver of surrounding conditions regarding traveling of the vehicle is known see, for example, Patent Document 1). The device notifies the surrounding conditions by displaying on an on-vehicle display unit a video obtained by video-recording surroundings of the vehicle with an on-vehicle camera.
However, there is a problem; to check contents displayed on the on-vehicle display unit, the driver needs to move his visual line from a travelling direction of the moving body to the display unit.
Patent Document 1: Japanese Patent Application Kokai No. 2000-285394
The present invention has been developed in order to solve the above-described problem, and its object is to provide a system for detecting surrounding conditions of a moving body (e.g., a vehicle, a ship, and an aircraft) that can notify a driver of surrounding conditions of the moving body without causing the driver to move his visual line.
A system for detecting surrounding conditions of a moving body according to claim 1 is a system for detecting surrounding conditions of a moving body that includes a plurality of lighting devices (illumination devices), a detector, and a controller. The illumination devices are disposed in a room (cabin) of the moving body. Each of the illumination devices has a switch for switching between a lighting state (or a turn on state) and an unlit state (or a turn off state). The detector is configured to detect an object to which attention must be paid around the moving body. The controller is configured to change a state of at least one of the lighting devices upon detecting the attention-required object.
When a system for detecting surrounding conditions of a moving body according to the present invention detects a “pay attention” object to which attention must be paid during traveling, in the surroundings of the moving body, then the system forcibly changes a state of at least one illumination device among a plurality of illumination devices installed in a cabin of the moving body (change from a unlit state to a lighting or blinking state, change in emission color, or change in luminance). Accordingly, even if a driver of the moving body does not visually observe an attention-required object present near the moving body, a change in the state of the lighting device (illumination device) enters the driver's field of vision. This enables the driver to recognize that the attention-required object approaches the moving body or the moving body approaches the attention-required object. To this end, changed is the state of the illumination device(s) installed at a position corresponding to a direction where the attention-required object is present. This allows the moving body driver to recognize the direction where the attention-required object is present without moving the driver's visual line. In the case where the attention-required object is lighting of a rear lamp of a forward vehicle (the rear lamp may be a lamp assembly that includes at least one of a stop lamp, a backup lamp, a turn signal lamp, and a hazard warning lamp), an illumination panel (light-emitting panel) installed in front of a driver's seat is caused to light or blink (flash) in a color corresponding to a lighting color of the rear lamp assembly. Accordingly, even if the driver does not look straight the rear lamp assembly of the forward vehicle, the driver can recognize whether the lighting rear lamp assembly of the forward vehicle is the backup lamp, the stop lamp, the turn signal lamp, or the hazard warning lamp without moving the direction of his visual line. In the case where the attention-required object is a traffic light, the illumination panel installed above the driver's seat in the moving body is caused to light in a color corresponding to a lighting color of the traffic light. Accordingly, even if the driver does not look at the traffic light ahead, the driver can recognize the lighting color of the traffic light without moving the direction of his visual line. In the case where the attention-required object is an avoidance-required obstacle which the vehicle equipped with the surrounding condition detecting system must avoid, such as pedestrian, a bicycle, a wall, an edge stone, an obstruction, or an emergency vehicle, then an illumination panel installed in a direction that the avoidance-required obstacle is present is forced to light or blink. An emission color, luminance, and/or a blinking speed of the illumination panel are also altered according, to a distance and a relative speed between the moving body and the avoidance-required obstacle. When the avoidance-required obstacle is an obstruction that hinders the traveling of the vehicle equipped with the surrounding condition detecting system, then the emission color and the luminance of the lighting panel (illumination panel) are changed according to a size of the obstruction. This allows the driver to intuitively recognize a hazard of the approaching avoidance-required obstacle.
Embodiments
In
The illumination panels 21 to 2n have the same, internal configuration.
When the controller 1 sends a lighting or a blinking command signal of a certain color (red, white, yellow, green, blue, purple, indigo or orange) to a lighting on/off switch 21 of an illumination panel 2 via the control bus BUS, the lighting on/off switch 21 turns on and allows a power supply voltage V1 to be supplied to an electrolumineseence (EL) panel driver 22. When the controller 1 supplies an turn off command signal to the on/off switch 21 via the control bus BUS, the switch 21 turns off, and stops the supply of the power supply voltage V1 to the EL panel driver 22 The EL panel driver 22 generates various light emission drive voltages based on the lighting or blinking command signal of a desired color (red, white, yellow, green, blue, purple, indigo or orange) received from the controller 1 via the control bus BUS while the power supply voltage V1 is being supplied through the switch 21, and supplies the resulting light emission drive voltages to a touch-sensor-integrated panel 23. When, for example, the EL panel driver 22 receives a red lighting command signal from the controller 1, the EL panel driver 22 supplies a light emission drive voltage to the EL panel 23 to cause the EL panel to light in red. When the EL panel driver 22 receives a yellow blinking command signal from the controller 1, the EL panel driver 22 supplies a light emission drive voltage to the EL panel 23 to cause the EL pane 23 to blink in yellow.
As shown in
The organic EL light emitting panel ELP as a plane EL light emitting element lights or blinks in an emission color corresponding to the light emission drive voltage supplied from the EL panel driver 22. The light is emitted to outside via the touch sensor panel TP. The organic EL light emitting panel ELP is in the off state (unlit state) when the light emission drive voltage is not supplied. The touch sensor panel TP is a switch that switches between lighting and extinction (no lighting) of the organic EL light emitting panel ELP. When a person touches a touch sensor face of the touch sensor panel TP, the touch sensor panel TP generates a contact detection signal TC to turn on or off the organic EL light emitting panel ELP while the contact is being made. The contact detection signal TC is supplied to the controller 1 via the control bus BUS.
With the above-described configuration, each of the illumination panels 21 to 2n lights or blinks in the emission color indicated by the lighting or blinking command signal that represents a desired color (red, white, yellow, green, blue, purple, indigo, or orange) received from the controller 1. When, for example, the controller 1 sends a white lighting command to the lighting panel 21, the illumination panel 21 emits a white light. When the controller 1 sends a white blinking command, the lighting panel 21 blinks in white. When the controller 1 sends a red lighting command to the illumination panel 21, the illumination panel 21 emits a red light. When the controller 1 sends a red blinking command, the illumination panel 21 blinks in red. Supply of a turn off command signal from the controller 1 can turn off each of the light-emitting panels 21 to 2n. Each of the light-emitting panels 21 to 2n generates the contact detection signal TC each time the touch sensor panel TP is made contact. The contact detection signal TC is sent to the controller 1.
The front camera 3 is installed at a front portion of a vehicle exterior. The front camera 3 supplies an image signal V1, which is obtained by video-recording the scene forward of the vehicle, to an image recognition processor 4.
The rear camera 5 is installed at a rear portion of the vehicle exterior. The rear camera 5 supplies an image signal V2, which is obtained by video-recording; the scene backward of the vehicle, to the image recognition processor 4.
The side cameras 6 are installed at lateral parts of the vehicle exterior, such as right and left side-view mirrors of the vehicle, respectively. The side cameras 6 supply image signals V3, which are obtained by video-recording the scene on both sides of the vehicle, to the image recognition processor 4.
The image recognition processor 4 performs an image recognition process on the image signals V1 to V3 to detect in these images an attention-required object that the driver needs to care during driving. The attention-required object is, for example, a lighting state of a rear lamp assembly of a forward vehicle, a traffic light, and an avoidance-required obstacle which the vehicle equipped with the surrounding condition detecting system must avoid. The avoidance-required obstacle is, for example, a pedestrian, a bicycle, an emergency vehicle, an obstruction that becomes an obstacle in traveling, a wall at a road with poor visibility, and an edge stone.
When light emission of the ear lamp assembly of a preceding vehicle is detected in an image obtained from the image signal V1, the image recognition processor 4 generates rear lamp detection data RD. The rear lamp detection data RD indicates whether the emission color of the rear lamp assembly is red, white, or yellow. The image recognition processor 4 supplies the rear lamp detection data RD to the controller 1 via the control bus BUS. When the emission color of the rear lamp assembly is yellow, that is, when a turn signal lamp or a hazard warning lamp is activated, direction indication information is included in the rear lamp detection data RD. The direction indication information indicates whether the rear lamp assembly is blinking only on the right side, blinking only on the left side, or blinking on both right and left sides.
When the traffic light is detected in the image obtained from on the image signal V1, the image recognition processor 4 generates traffic-light detection data SD. The traffic-light detection data SD indicates a lighting color of the traffic light. The image recognition processor 4 supplies the traffic-light detection data SD to the controller 1 via the control bus BUS. In the case where a traffic light shows (indicates, displays) a direction with an arrow, the image recognition processor 4 incorporates, in the traffic-light detection data SD, information representing a fact that this is an arrow indication and also representing the pointing direction of the arrow.
When a pedestrian or a bicycle is found in an image or images obtained from the image signals V1 to V3, the image recognition processor 4 generates pedestrian bicycle detection data HD. The pedestrian/bicycle detection data HD indicates the direction of, distance to, and relative speed of the pedestrian or bicycle, and the data is supplied to the controller 1 via the control bus BUS.
When the image recognition processor 4 detects an attention-required road zone such as a road zone with poor visibility or a route in a parking area (parking lot) in the images derived from the image signals V1 to V3 and also detects an attention-required object, such as a wall, an edge stone, and another vehicle on the attention-required road, then the image recognition processor 4 generates attention-required zone detection data ED. The attention-required zone detection data ED includes information representing the direction of, distance to, and relative speed of the attention-required object. The image recognition processor 4 supplies the attention-required zone detection data ED to the controller 1 via the control bus BUS. The road zone with poor visibility is a zone where the vehicle equipped with the surrounding condition detecting system enters a wider road from a narrower road without effective right/left visibility due to, for example, a blockade, such as a house wall.
When the image recognition processor 4 detects an emergency vehicle such as a patrol car, a police car, ambulance, and a fire engine in the images obtained from the image signals V1 to V3, the image recognition processor 4 generates emergency vehicle detection data QD. The emergency vehicle detection data QD indicates the direction of, distance to, and relative speed of the emergency vehicle. The image recognition processor 4 supplies the emergency vehicle detection data QD to the controller 1 via the control bus BUS.
When the image recognition processor 4 detects an obstruction on the road in the images obtained from the image signals V1 to V3, the image recognition processor 4 generates obstruction detection data GD. The obstruction detection data GD indicates the direction of, distance to, and a size of the obstruction. The image recognition processor 4 supplies the obstruction detection data GD to the controller 1 via the control bus BUS.
The controller 1 sends the turn on command (i.e., lighting command signal) or the turn off command to the illumination panel 2, which is a delivery source of the contact detection signal TC, in response to the contact detection signal TC received from the illumination panel 21 to 2n. This allows the illumination panel 2 to be lighted by simply touching the light emitting face of the illumination panel 2 with a fingertip in the case where a person in the vehicle equipped with the surrounding condition detecting system needs interior lighting. Touching the light emitting face of the illumination panel 2 with the fingertip again turns off the illumination panel 2.
The controller 1 forces the light-emitting panel 2 to light or blink in order to notify the person in the vehicle of the surrounding conditions by performing the following surrounding condition notification control.
After the execution of Step S6, the controller 1 returns to Step S1 and repeats the processing of Steps S1 to S6.
In
After the execution of Step S19, S20, or S21, the controller 1 exits the forward vehicle rear lamp notification subroutine shown in
In
After the execution of Step S40, S43, S44, or S45, the controller 1 exits the traffic light notification subroutine shown in
In
Next, the controller 1 sets emission conditions (luminance a color, and a blinking speed) corresponding to a distance to and a speed relative to the pedestrian or the bicycle represented by the pedestrian/bicycle detection data HD (Step S53). The emission condition is set, for example, as follows. The smaller the distance to the pedestrian/bicycle, the higher the luminance is. The smaller the distance, the more the emission color transitions from a cool color to a warm color (e.g., purple→indigo→blue→green→yellow→orange→red). The smaller the distance, the higher the blinking speed is. The higher the relative speed, the higher the luminance is. The higher the relative speed, the more the emission color transitions from a cool color to a warm color. The higher the relative speed, the higher the blinking speed is. It should be noted that the above-mentioned emission conditions based on the distance and relative speed may be combined.
Next, the controller 1 supplies the lighting or blinking command signal to the attention-required direction indication panel (Step S54). The lighting or blinking command signal causes the attention-required direction indication panel selected among the illumination panels 29 to 211 and 212 to 214 (
When Step S51 determines that the pedestrian/bicycle detection data HD is not received, the controller 1 supplies the turn off command signal to the attention-required direction indication panel in order to turn off the attention-required direction indication panel (Step S55).
After the execution of Step S54 or S55, the controller 1 exits the pedestrian/bicycle detection subroutine shown in
In
Next, the controller 1 sets light emission conditions (luminance a color, and a blinking speed) corresponding to a distance to and a speed relative to the attention- required object represented by the attention-required zone detection data ED (Step S63). The light emission condition is set, for example, as follows. The smaller the distance, the higher the luminance is. The smaller the distance, the more the emission color transitions from a cool color to a warm color (purple→indigo→blue→green→yellow→orange→red). The smaller the distance, the higher the blinking speed is. The higher the relative speed, the higher the luminance is. The higher the relative speed, the more the emission color transitions from a cool color to a warm color. The higher the relative speed, the higher the blinking speed is. It should be noted that the above-described light emission conditions based on the distance and relative speed may be combined.
Next, the controller 1 supplies the lighting or blinking command signal to the attention-required direction (cautionary direction) indication panel (Step S64). The lighting or blinking command signal causes the attention-required direction indication panel, selected among the illumination panels 29 to 211 and 212 to 214 (
When Step S61 determines that the attention-required zone detection data ED is not received, the controller 1 supplies the turn off command signal to the attention-required direction indication panel in order to turn off the attention-required direction indication panel (Step S65).
After the execution of Step S64 or S65, the controller 1 exits the attention-required zone notification subroutine shown in
In
Next, the controller 1 sets light emission conditions (luminance, a color, and a blinking speed) corresponding to a distance to and a speed relative to the emergency vehicle represented by the emergency vehicle detection data QD (Step S73). The light emission condition is set, for example, as follows. The smaller the distance, the higher the luminance is. The smaller the distance, the more the emission color transitions from a cool color to a warm color (purple→indigo→blue→green→yellow→orange→red). The smaller the distance, the higher the blinking speed is. The higher the relative speed, the higher the luminance is. The higher the relative speed, the more the emission color transitions from a cool color to a warm color. The higher the relative speed, the higher the blinking speed is. It should be noted that the light emission conditions based on the distance and relative speed may be combined.
Next, the controller 1 supplies the lighting or blinking command signal to the attention-required direction indication panel (Step S74). The lighting or blinking command signal causes the attention-required direction indication panel, selected among the light emission panels 29 to 211 and 212 to 214 (
When Step S71 determines that the emergency vehicle detection data QD is not received, the controller 1 supplies the turn off command signal to the attention-required direction indication panel in order to turn off the attention-required direction indication panel (Step S75).
After the execution of Step S74 or S75, the controller 1 exits the emergency vehicle notification subroutine shown in
In
Next, the controller 1 sets light emission conditions (luminance, a color, and a blinking speed) corresponding to a distance to and a size of the obstruction represented by the obstruction detection data GD (Step S83). The light emission condition is set, for example, as follows. The smaller the distance, the higher the luminance is. The smaller the distance, the more the emission color transitions from a cool color to a warm color (purple→indigo→blue→green→yellow→orange→red). The smaller the distance, the higher the blinking speed is. The larger the obstruction size, the higher the luminance is. The larger the obstruction size, the more the emission color is inclined toward a warm color. The larger the obstruction size, the higher the blinking speed is. It should be noted that the light emission conditions based on the distance and size may be combined.
Next, the controller 1 supplies the lighting or blinking command signal to the attention-required direction indication panel (Step S84). The lighting or blinking command signal causes the attention-required direction indication panel, selected among the illumination panels 29 to 211 and 212 to 211 (
When it is determined in Step S81 that the obstruction detection data GD is not received, then the controller 1 supplies the turn off command signal to the attention-required direction indication panel in order to turn off the attention-required direction indication panel (Step S85).
After the execution of Step S84 or S85, the controller 1 exits the obstruction notification subroutine shown in
Operations and advantages obtained upon execution of the processing shown in
When a stop lamp of a vehicle traveling forward of the vehicle equipped with the surrounding condition detecting system emits a red light, the illumination panel 21 emits a red light in accordance with the lighting stop lamp. The illumination panel 21 is installed at the center of the steering wheel SW, which is a shaded portion in
That is, when an image captured by the front camera 3 includes a lighting rear lamp of the forward vehicle, the illumination panels 21 to 24, which are installed at the steering wheel SW, reproduce the lighting or blinking state with a color corresponding to the emission color of the rear lamp assembly of the forward vehicle. Accordingly, even if the driver's visual line is out of the forward vehicle, the lighting/blinking state of the illumination panels 21 to 24 installed at the steering wheel SW enters the driver's field of vision. Thus, even if the driver does not look at the forward vehicle, the driver can recognize a state of the rear lamp assembly of the forward vehicle (the backup lamp, the stop lamp, the turn signal lamp, or the hazard warning lamp) without moving the direction of his visual line. Accordingly, in the case where heavy braking is applied to the forward vehicle while the driver of the vehicle equipped with the surrounding condition detecting system is checking, for example, a traffic sign, the driver of the vehicle equipped with the surrounding condition detecting system can immediately take avoidance behavior.
When a traffic light appears forward while the vehicle equipped with the surrounding condition detecting system is traveling, and the emission color of the traffic light is red, then the illumination panel 25 lights in red. The illumination panel 25 is installed above the windshield FG, which is a shaded portion in
Thus, when an image captured by the front camera 3 includes the traffic light, the illumination panels 25 to 28 and 223, which are installed above the windshield FO in front of the driver's seat, reproduce the lighting state of the traffic light with a color corresponding to the lighting color of the traffic light. Accordingly, even if the driver cannot visually observe a color of the traffic light due to sunlight, the lighting color of the traffic light can be checked. Even if the driver's visual line is out of the traffic light, lighting states of the illumination panels 25 to 28 and 223 installed above the windshield FG enters the driver's field of vision. This allows the driver to recognize the color of the traffic light without moving an eye direction of the driver. When the traffic light displays a turn-arrow, the direction of the turn arrow displayed can be recognized by the relationship between a position of one illumination panel 2 that emits a red light and a position of another illumination panel 2 that emits a blue light.
When the avoidance-required obstacle such as a pedestrian, a bicycle, a wall, an edge stone, an obstruction, and an emergency vehicle is present around the vehicle equipped with the surrounding condition detecting system, that illumination panel 2 which is installed at a position corresponding to a direction of the avoidance-required obstacle lights or blinks. For example, when the avoidance-required obstacle is present right front of the vehicle equipped with the surrounding condition detecting system, then the illumination panel 212, which is a shaded portion in
That is, when an image captured by the front camera 3, the rear camera 5, or the side camera 6 includes the avoidance-required obstacle, i.e., when the avoidance-required obstacle is present near the vehicle equipped with the surrounding condition detecting system, then a state of at least one of the light emitting panels 2 provided in the vehicle cabin is forcibly changed (change from unlit state to lighting or blinking state, changes in emission color or change in luminance). Accordingly, even if the driver of the vehicle equipped with the surrounding condition detecting system does not visually observe the attention-required object around the vehicle equipped with the surrounding condition detecting system, the change in the state of the light emitting panel(s) 2 enters the driver's field of vision. This allows the driver to recognize that the attention required object approaches his vehicle or his vehicle approaches the attention-required object. The light emitting panel 2 installed at a position corresponding to a direction that the attention-required object is present is caused to light or blink (flash). This allows the driver to recognize the direction where the attention-required object (cautionary object) is present without moving the driver's visual line.
Accordingly, even if the avoidance-required obstacle present near the vehicle is out of the driver's field of vision during backward driving, entering a road zone with poor visibility, or exiting a parking area or lot, light (blinking or no blinking) of the luminescence panel 2 located at a position corresponding to a direction of the avoidance-required obstacle enters the driver's field of vision. This enables the driver to recognize the direction where the avoidance-required obstacle is present without moving the driver's visual line. Because the emission color, the luminance, and the blinking speed of the luminescence panel 2 change in accordance with the distance and the relative speed between the vehicle and the avoidance-required obstacle, the driver is able to intuitively recognize danger or hazard associated with the approaching avoidance-required obstacle.
Steps S73 and S74 in the emergency vehicle notification subroutine shown in
Although in the above-described embodiment, as shown in
In the case where, for example, an avoidance-required object approaches from rightward of the vehicle equipped with the surrounding condition detecting system at an intersection with poor visibility where entry is from a narrow road without effective right/left visibility into a wide road, a luminescence panel 215 mounted on a dashboard on the driver's seat side shown in
1 controller
2
1 to 2n light emitting panels
3 front camera
4 image recognition processor
5 rear camera
6 side cameras
This application claims benefit under 35 USC § 120 as a Continuation of U.S. patent application Ser. No. 16/119,905, filed Aug. 31, 2018, which is a Continuation of U.S. patent application Ser. No. 16/034,257, filed Jul. 12, 2018 (U.S. Pat. No. 10,384,596, issued Aug. 20, 2019), and U.S. patent application Ser. No. 16/034,283, filed Jul. 12, 2018 (U.S. Pat. No. 10,625,666, issued Apr. 21, 2020), which are a Continuation of U.S. patent application Ser. No. 15/711,988 filed Sep. 21, 2017, which is a Continuation of U.S. patent application Ser. No. 15/489,174, filed on Apr. 17, 2017 (U.S. Pat. No. 9,802,532, issued Oct. 31, 20171), which is a Continuation of U.S. patent application Ser. No. 14/110,118, filed on Dec. 17, 2013 (U.S. Pat. No. 9,649,972, issued May 16, 2017), which is a National Stage entry under 35 USC § 371 of expired PCT/JP2011/058803, filed Apr. 7, 2011, the contents of which are incorporated by reference in their entirety for all purposes.
Number | Name | Date | Kind |
---|---|---|---|
3393344 | Engelmann | Jul 1968 | A |
4626850 | Chey | Dec 1986 | A |
4692798 | Seko et al. | Sep 1987 | A |
4694295 | Miller et al. | Sep 1987 | A |
4967319 | Seko et al. | Oct 1990 | A |
5786772 | Schofield et al. | Jul 1998 | A |
5823666 | Kingsolver | Oct 1998 | A |
6072391 | Suzuki et al. | Jun 2000 | A |
6420975 | DeLine et al. | Jul 2002 | B1 |
6470273 | Halsted et al. | Oct 2002 | B2 |
6536928 | Hein et al. | Mar 2003 | B1 |
6774988 | Stam et al. | Aug 2004 | B2 |
6894608 | Gunderson | May 2005 | B1 |
6985073 | Doan | Jan 2006 | B1 |
7038577 | Pawlicki et al. | May 2006 | B2 |
7104664 | Sugimoto et al. | Sep 2006 | B2 |
7299892 | Radu et al. | Nov 2007 | B2 |
7344277 | Anderson, Jr. et al. | Mar 2008 | B2 |
7477976 | Horii et al. | Jan 2009 | B2 |
7492281 | Lynam et al. | Feb 2009 | B2 |
7526103 | Schofield et al. | Apr 2009 | B2 |
7589883 | Varaprasad et al. | Sep 2009 | B2 |
7616781 | Schofield et al. | Nov 2009 | B2 |
7800483 | Bucher | Sep 2010 | B2 |
8045760 | Stam et al. | Oct 2011 | B2 |
8362922 | Kushi et al. | Jan 2013 | B2 |
8433473 | Perkins et al. | Apr 2013 | B2 |
8482430 | Szczerba | Jul 2013 | B2 |
8542129 | Taniguchi et al. | Sep 2013 | B2 |
8645001 | Basson | Feb 2014 | B2 |
8648535 | Tatara | Feb 2014 | B2 |
8862380 | Jung | Oct 2014 | B2 |
8892301 | Park et al. | Nov 2014 | B2 |
9533615 | Moberg et al. | Jan 2017 | B2 |
9649972 | Kato | May 2017 | B2 |
9718405 | Englander et al. | Aug 2017 | B1 |
9934685 | Bernhardt | Apr 2018 | B1 |
10232713 | Kim | Mar 2019 | B2 |
20020003571 | Schofield et al. | Jan 2002 | A1 |
20020159270 | Lynam et al. | Oct 2002 | A1 |
20030117728 | Hutzel et al. | Jun 2003 | A1 |
20040114379 | Miller et al. | Jun 2004 | A1 |
20040114381 | Salmeen et al. | Jun 2004 | A1 |
20050036660 | Otsuka et al. | Feb 2005 | A1 |
20060017583 | Davenport et al. | Jan 2006 | A1 |
20060074540 | Braunberger et al. | Apr 2006 | A1 |
20060125919 | Camilleri et al. | Jun 2006 | A1 |
20060125959 | Yoshizawa et al. | Jun 2006 | A1 |
20060164230 | DeWind et al. | Jul 2006 | A1 |
20060177097 | Fujimura et al. | Aug 2006 | A1 |
20070058257 | Lynam | Mar 2007 | A1 |
20070126564 | Lee et al. | Jun 2007 | A1 |
20070268118 | Watanabe et al. | Nov 2007 | A1 |
20070290823 | Watanabe et al. | Dec 2007 | A1 |
20080101077 | Watanabe | May 2008 | A1 |
20080122799 | Pryor | May 2008 | A1 |
20080316054 | Lynam et al. | Dec 2008 | A1 |
20090058621 | Yen | Mar 2009 | A1 |
20090063053 | Basson et al. | Mar 2009 | A1 |
20090268947 | Schaufler | Oct 2009 | A1 |
20100045797 | Schofield et al. | Feb 2010 | A1 |
20100100268 | Zhang et al. | Apr 2010 | A1 |
20100188343 | Bach | Jul 2010 | A1 |
20100194888 | McElroy et al. | Aug 2010 | A1 |
20100214455 | Simon | Aug 2010 | A1 |
20100214791 | Schofield | Aug 2010 | A1 |
20100315216 | Hada | Dec 2010 | A1 |
20110043481 | Bruwer | Feb 2011 | A1 |
20110128139 | Tauchi et al. | Jun 2011 | A1 |
20110187518 | Strumolo et al. | Aug 2011 | A1 |
20110298605 | Lu | Dec 2011 | A1 |
20120025969 | Dozza | Feb 2012 | A1 |
20120062743 | Lynam et al. | Mar 2012 | A1 |
20120123637 | Funayama et al. | May 2012 | A1 |
20120140080 | Taylor et al. | Jun 2012 | A1 |
20120147613 | Goldschmidt et al. | Jun 2012 | A1 |
20120176035 | Alloway et al. | Jul 2012 | A1 |
20120212353 | Fung et al. | Aug 2012 | A1 |
20120253553 | Nagata | Oct 2012 | A1 |
20120283894 | Naboulsi | Nov 2012 | A1 |
20120314434 | Park et al. | Dec 2012 | A1 |
20130039080 | Yamazaki et al. | Feb 2013 | A1 |
20130051042 | Nordbruch | Feb 2013 | A1 |
20130057397 | Cutler et al. | Mar 2013 | A1 |
20130128599 | Park et al. | May 2013 | A1 |
20130155704 | Takagaki | Jun 2013 | A1 |
20130177202 | Dierks et al. | Jul 2013 | A1 |
20130226408 | Fung et al. | Aug 2013 | A1 |
20140056012 | Yamazaki | Feb 2014 | A1 |
20140069015 | Salter et al. | Mar 2014 | A1 |
20140071702 | Faber et al. | Mar 2014 | A1 |
20140098230 | Baur | Apr 2014 | A1 |
20140098551 | Krauss et al. | Apr 2014 | A1 |
20140118130 | Chang et al. | May 2014 | A1 |
20140169010 | Imaeda | Jun 2014 | A1 |
20140218946 | Fritz | Aug 2014 | A1 |
20140307179 | Zhao et al. | Oct 2014 | A1 |
20140320656 | Foley | Oct 2014 | A1 |
20140350836 | Stettner et al. | Nov 2014 | A1 |
20140355280 | Fujiyoshi | Dec 2014 | A1 |
20150003087 | Futamura et al. | Jan 2015 | A1 |
20150042225 | Fukayama | Feb 2015 | A1 |
20150042226 | Hibino | Feb 2015 | A1 |
20150051797 | Ehlgen et al. | Feb 2015 | A1 |
20150062937 | Kasai et al. | Mar 2015 | A1 |
20150077826 | Beckman | Mar 2015 | A1 |
20150081168 | McWithey et al. | Mar 2015 | A1 |
20150081169 | Pisz | Mar 2015 | A1 |
20150105976 | Shikii et al. | Apr 2015 | A1 |
20150109450 | Walker | Apr 2015 | A1 |
20150202939 | Stettner et al. | Jul 2015 | A1 |
20150206439 | Marsden et al. | Jul 2015 | A1 |
20150239392 | Mizuno et al. | Aug 2015 | A1 |
20150241013 | Mochizuki et al. | Aug 2015 | A1 |
20150242953 | Suiter | Aug 2015 | A1 |
20150251586 | Imaeda et al. | Sep 2015 | A1 |
20150324972 | Hayakawa et al. | Nov 2015 | A1 |
20150336500 | Herntrich et al. | Nov 2015 | A1 |
20150342007 | Anderson et al. | Nov 2015 | A1 |
20160001781 | Fung et al. | Jan 2016 | A1 |
20160046298 | DeRuyck et al. | Feb 2016 | A1 |
20160297356 | Dellock et al. | Oct 2016 | A1 |
20160325676 | Yamada | Nov 2016 | A1 |
20160375768 | Konet et al. | Dec 2016 | A1 |
20170162047 | Garcia Lopez | Jun 2017 | A1 |
20170180707 | Hsu et al. | Jun 2017 | A1 |
20170240185 | Li | Aug 2017 | A1 |
20170268896 | Bai et al. | Sep 2017 | A1 |
20170327035 | Keiser | Nov 2017 | A1 |
20180079306 | Kim et al. | Mar 2018 | A1 |
20180174450 | Im et al. | Jun 2018 | A1 |
20180201184 | Lay et al. | Jul 2018 | A1 |
20190001874 | Verwys et al. | Jan 2019 | A1 |
20190283753 | Tatourian et al. | Sep 2019 | A1 |
Number | Date | Country |
---|---|---|
S62059880 | Mar 1987 | JP |
S63195045 | Aug 1988 | JP |
2000285394 | Oct 2000 | JP |
2003291689 | Oct 2003 | JP |
2005018235 | Jan 2005 | JP |
2005047383 | Feb 2005 | JP |
2006064968 | Jun 2006 | JP |
2006172315 | Jun 2006 | JP |
2006192972 | Jul 2006 | JP |
2008087832 | Jul 2008 | JP |
2008158963 | Jul 2008 | JP |
2009009320 | Jan 2009 | JP |
2010092171 | Apr 2010 | JP |
Entry |
---|
International Search Report and Written Opinion for related International Application No. PCT/JP2011/058803 dated Jul. 12, 2011; English translation of ISR provided; 8 pages. |
Number | Date | Country | |
---|---|---|---|
20210146829 A1 | May 2021 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16119905 | Aug 2018 | US |
Child | 17127783 | US | |
Parent | 16034283 | Jul 2018 | US |
Child | 16119905 | US | |
Parent | 16034257 | Jul 2018 | US |
Child | 16034283 | US | |
Parent | 15711988 | Sep 2017 | US |
Child | 16034257 | US | |
Parent | 15489174 | Apr 2017 | US |
Child | 15711988 | US | |
Parent | 14110118 | Dec 2013 | US |
Child | 15489174 | US | |
Parent | PCT/JP2011/058803 | Apr 2011 | US |
Child | 14110118 | US |