This application is a U.S. National Phase of International Patent Application No. PCT/JP2020/037455 filed on Oct. 1, 2020, which claims priority benefit of Japanese Patent Application No. JP 2019-198491 filed in the Japan Patent Office on Oct. 31, 2019 Each of the above-referenced applications is hereby incorporated herein by reference in its entirety.
The present technology relates to an image display apparatus.
There has heretofore been proposed an image display apparatus by which a three-dimensional object appears to exist in a space of a front (hereinafter, also referred to as a “front space”) of a display screen (e.g., see PTL 1). An image display apparatus described in PTL 1 allows a finger to be inserted into the front space to detect a position of a fingertip, thereby allowing a particular position to be specified in the front space and an intuitive operation to be performed.
In addition, in recent years, there has heretofore been proposed an image display apparatus in which a cylindrical screen and a substantially disc-shaped mirror are stacked on upper side of a cylindrical base, and a color projector that emits laser light corresponding to each color of RGB upward is disposed at the center of the base (e.g., see PTL 2). In the image display apparatus described in PTL 2, the color projector emits the laser light radially upward, and the emitted laser light is reflected radially toward the entire circumference of a screen by a reflective mirror. Then, a three-dimensional image, by which a three-dimensional object is visually recognizable from a plurality of circumferential directions, is displayed by the screen on which the laser light is reflected to the entire circumference, within a space partitioned by the screen.
However, in an image display apparatus described in PTL 2, it may not be possible to insert a detection target such as a finger into a space in which a three-dimensional object is visually recognized. Therefore, it may not be possible to cause a user to perform an intuitive operation by applying the technique described in PTL 1.
An object of the present disclosure is to provide an image display apparatus that enables an intuitive operation even when a detection target is not able to be inserted into a first space in which a three-dimensional object is visually recognized.
An image display apparatus of the present disclosure includes: (a) a display unit that displays a three-dimensional image, by which a three-dimensional object and a pointer appear to exist within a first space partitioned by a member configuring an outwardly convex outer surface, to be visually recognizable from a plurality of circumferential directions; (b) a position detection section that detects a position of a detection target; and (c) a display control section that moves a display position of the pointer displayed by the display unit on a basis of a position of the detection target that exists within a predetermined second space not overlapping the first space among the position of the detection target detected by the position detection section.
Hereinafter, description is given, with reference to
1. First Embodiment: Image Display Apparatus
1-1 Overall Configuration of Image Display Apparatus
1-2 Contents of Control Processing
1-3 Modification Examples
2. Second Embodiment: Image Display Apparatus
Description is given of an image display apparatus according to a first embodiment of the present disclosure. An image display apparatus 1 according to the first embodiment is a display apparatus that enables a three-dimensional image to be visually recognized from a plurality of circumferential directions. The image display apparatus 1 is sized to be small enough to be held with one hand as illustrated in
As illustrated in
The display unit 2 is formed by stacking, on upper side of the base 5 having a cylindrical shape with an upper surface being open and a lower surface being closed, a cylindrical screen 6, and a substantially disc-shaped reflective mirror 7 in this order.
At the center of the inside of the base 5, there is disposed an emission unit 9 with an optical axis 8 being vertically upward. As the emission unit 9, for example, there may be employed a color projector that displays each pixel by scanning laser light (hereinafter, also referred to as “image light 10”) corresponding to each color of RGB. Then, the emission unit 9 emits the image light 10 to configure a three-dimensional image radially upward in accordance with a signal from the controller 4 to reflect the emitted image light 10 on the screen 6 with the reflective mirror 7.
As the three-dimensional image, for example, there may be employed an image by which a three-dimensional object 12 and a pointer 13 (see
It is to be noted that, although the first embodiment gives an example in which a space partitioned by the cylindrical screen 6 is used as the first space 11, other configurations may also be employed. It is sufficient for the first space 11 to be a space partitioned by a member configuring an outwardly convex outer surface; for example, the first space 11 may be a space partitioned by the screen 6 having a semicylindrical shape, or may be a space partitioned by the screen 6 having a square tubular shape.
In addition, a space (hereinafter, also referred to as a “second space 14”) to allow a user to input a move destination of the pointer 13 is provided around the first space 11. That is, the second space 14 is a space not overlapping the first space 11. As the second space 14, for example, there may be employed a cylindrical space surrounding the circumference of the first space 11. In the second space 14, there is set a cylindrical coordinate system r2θ2z2, with the z1 axis of the first space 11 being a z2 axis and an r1 axis of the first space 11 being an r2 axis. Here, r2 is a radial component, θ2 is an azimuthal component, and z2 is a height component.
As the screen 6, for example, there may be employed a hologram screen on which a three-dimensional image is displayed by emitting the image light 10 that is incident in a diffusing manner. For example, it may be possible to employ a transmissive hologram that enables the screen 6 to be seen through, as described in International Publication No. WO2018/163945. According to the transmissive hologram, as illustrated in
As the reflective mirror 7, there may be employed a mirror having below a reflective surface that reflects the image light 10 emitted from the emission unit 9. Then, the reflective mirror 7 reflects the image light 10 emitted radially upward from the emission unit 9 radially toward the entire circumference of the screen 6.
In the display unit 2 having the above configuration, as illustrated in
The plurality of information acquisition units 3 are evenly spaced on a sidewall portion of the base 5 or a sidewall portion of the reflective mirror 7, and obtain information concerning a position of a detection target existing around the display unit 2. A detection result of the information acquisition unit 3 is outputted to the controller 4. As the information acquisition unit 3, for example, there may be employed a TOF (Time Of Flight) sensor that shoots a range image around the display unit 2, a camera that shoots an image around the display unit 2, an ultrasonic sensor that measures time from emission of an ultrasonic wave to the circumference of the display unit 2 to return of a reflected wave, a capacitance sensor that measures an electrostatic capacitance around the display unit 2 to thereby detect a distance to a body around the display unit 2, or a magnetic sensor by measuring a change in detection of a distance to the body around the display unit 2 by measuring a magnetic flux density around the display unit 2. In the following description of the first embodiment, the TOF sensor disposed on the sidewall portion of the base 5 is employed as the information acquisition unit 3, and the “information acquisition unit 3” is also referred to as a “TOF sensor 3”.
It is to be noted that in a case where the capacitance sensor or the magnetic sensor is employed as the information acquisition unit 3, a configuration may be adopted in which a transparent or semi-transparent material is used to dispose the sensor on the entire circumferential surface of the outer peripheral surface of the screen 6.
The controller 4 is disposed inside the base 5, and includes hardware resources such as a storage device 15 and a processor 16 as illustrated in
The storage device 15 stores a control program of the image display apparatus 1 executable by the processor 16. In addition, the storage device 15 stores various types of data necessary for executing the control program.
The processor 16 implements a display control section 17 and a position detection section 18 in accordance with the control program stored in the storage device 15. Then, the display control section 17 and the position detection section 18 are used to output a signal to control the image light 10 to the emission unit 9 on the basis of data of the range image outputted from the TOF sensor 3, and to execute control processing to move a display position of the pointer 13 represented by the three-dimensional image.
Next, description is given of control processing to be executed by the display control section 17 and the position detection section 18, which are implemented by the processor 16. The control processing is executed when an unillustrated power source switch of the image display apparatus 1 is brought into an ON state.
As illustrated in
Subsequently, in step S102, the position detection section 18 detects a position of a predetermined detection target. As the detection target, for example, there may be employed a fingertip, a stylus pen, and a marker. Suppose that, in the following description of the first embodiment, the fingertip is employed as the detection target. Specifically, the position detection section 18 first acquires data of a moving image outputted from each of the plurality of TOF sensors 3. Subsequently, a position of a fingertip of a person who exists around the display unit 2 is detected on the basis of the acquired data, i.e., each range image around the display unit 2. In the detection of the position of the fingertip, the cylindrical coordinate system r2θ2z2 is used to detect coordinates (r2′, θ2′, z2′) at the position of the fingertip. As the detection method, for example, there may be employed a method of performing pattern matching or the like on the range image.
Next, the processing proceeds to step S103, in which the display control section 17 determines whether a position of the fingertip detected in step S102 exists within the second space 14. Then, in a case where determination is made that the position of the fingertip is not within the second space 14 (No), the processing returns to step S102. Thus, the pointer 13 displayed by the display unit 2 does not move to remain at the current position. Meanwhile, in a case where determination is made that a position of the fingertip exists within the second space 14 (Yes), the processing proceeds to step S104.
In step S104, the display control section 17 outputs, to the emission unit 9, a signal to move the display position of the pointer 13 displayed by three-dimensional image. Specifically, first, coordinates (r1′, θ1′, z1′) of a move destination of the pointer 13 are calculated on the basis of the coordinates (r2′, θ2′, z2′) at the position of the fingertip existing within the second space 14, among coordinates of the position of the fingertip detected in step S102. For example, as illustrated in
In addition, for example, r1′ is decided, as illustrated in
Subsequently, the display control section 17 creates image data of a three-dimensional image in which the pointer 13 has been moved to the coordinates (r2′, θ2′, z2′) of a move destination of the decided pointer 13. Subsequently, the created image data is converted into data of the image light 10, and a signal to irradiate the image light 10 indicated by the data obtained by the conversion is outputted to the emission unit 9; then the processing returns to step S102. Thus, in the image display apparatus 1, as illustrated in
It is to be noted that, in a case where a plurality of positions of the fingertips detected in step S102 exist within the second space 14, a configuration may be adopted in which the coordinates (r1′, θ1′, z1′) of the move destination of the pointer 13 are calculated using the coordinates (r2′, θ2′, z2′) at the position of the fingertip having entered the second space 14 at first among the positions of the plurality of fingertips. That is, the display position of the pointer 13 is moved only on the basis of the position of the fingertip having entered the second space 14 at first among the positions of the plurality of fingertips existing within the second space 14.
In addition, in a case where the pointer 13 is in contact with the three-dimensional object 12 at the move destination of the pointer 13, a configuration may be adopted in which, as image data of a three-dimensional image, the display control section 17 brings the pointer 13 into contact with the three-dimensional object 12, and image data is created in which the three-dimensional object 12 is changed in response to the contact of the pointer 13. As the change of the three-dimensional object 12 in response to the contact of the pointer 13, for example, there may be employed movement of the display position due to pressing by the pointer 13, or deformation at a pressed location due to the pressing by the pointer 13.
As described above, in the image display apparatus 1 according to the first embodiment of the present disclosure, a position of the detection target is detected, and the display position of the pointer 13 displayed by the display unit 2 is moved on the basis of a position of the detection target existing within the predetermined second space 14 not overlapping the first space 11, among the position of the detected detection target. Therefore, the user inserts the detection target into the second space 14 and moves the position of the detection target to thereby be able to move the display position of the pointer 13 to a position corresponding to the position of the detection target. Therefore, it is possible to provide the image display apparatus 1 that enables an intuitive operation even when the detection target is not able to be inserted into the first space 11.
In addition, in the image display apparatus 1 according to the first embodiment of the present disclosure, the second space 14 is set as a space surrounding the circumference of the first space 11, the position of the pointer 13 is able to be altered by moving the detection target around the screen 6, thus making it possible to operate the position of the pointer 13 more intuitively.
In addition, in the image display apparatus 1 according to the first embodiment of the present disclosure, the display unit 2 includes the cylindrical screen 6 on which the three-dimensional image is displayed, and the first space 11 is set as a space of the hollow portion inside the screen 6, thus making it possible to display the three-dimensional object 12 in the screen 6.
In addition, in the image display apparatus 1 according to the first embodiment of the present disclosure, the cylindrical coordinate system r2θ2z2 with the center axis of the second space 14 being the z2 axis is set in the second space 14, and the cylindrical coordinate system r2θ2z2 is used to detect the coordinates (r2′, θ2′, z2′) at the position of the detection target. Therefore, the coordinate system is suitable for the shape of the second space 14, thus making it possible to represent the position of the detection target more appropriately.
In addition, in the image display apparatus 1 according to the first embodiment of the present disclosure, the cylindrical coordinate system r1θ1z1 with the center axis of the first space 11 being the z1 axis is set in the first space 11. In addition, the pointer 13 displayed by the display unit 2 is moved to allow the values of θ2′ and z2′ of the coordinates (r2′, θ2′, z2′) at the position of the detection target detected by the position detection section 18 and the values of θ1′ and z1′ of the coordinates (r1′, θ1′, z1′) of the pointer 13 in the cylindrical coordinate system r1θ1z1 to be the same. Thus, it is possible to operate the position of the pointer 13 more intuitively.
In addition, in the image display apparatus 1 according to the first embodiment of the present disclosure, the pointer 13 displayed by the display unit 2 is moved to allow r1′ of the coordinates (r1′, θ1′, z1′) of the pointer 13 in the cylindrical coordinate system r1θ1z1 to be “0” in a case where r2′ of the coordinates (r2′, θ2′, z2′) at the position of the detection target detected by the position detection section 18 is the same as the value R1 of the radius of the first space 11, and to allow r1′ of the coordinates (r1′, θ1′, z1′) to be the same as the value R1 of the radius of the first space 11 in a case where r2′ of the coordinates (r2′, θ2′, z2′) is the same as the value R2 of the radius of the second space 14. Thus, it is possible to operate the position of the pointer 13 more intuitively.
In addition, the image display apparatus 1 according to the first embodiment of the present disclosure includes the information acquisition unit 3 that is disposed at an upper portion or a lower portion of the display unit 2 and acquires information concerning the position of the detection target. Then, on the basis of a detection result of the information acquisition unit 3, the position of the detection target is detected. Thus, it is possible to operate the position of the detection target more appropriately.
In addition, in the image display apparatus 1 according to the first embodiment of the present disclosure, the TOF sensor, the camera, a touch panel, the capacitance sensor, or the magnetic sensor is employed as the information acquisition unit 3. Therefore, it is possible to acquire information concerning the position of the detection target more appropriately, and thus to detect the position of the detection target more appropriately.
In addition, in the image display apparatus 1 according to the first embodiment of the present disclosure, a fingertip of a person is set as the detection target. In addition, in a case where a plurality of positions of fingertips detected by the position detection section 18 exist within the second space 14, the display position of the pointer 13 displayed by the display unit 2 is moved only on the basis of the position of the fingertip having entered the second space 14 at first among the plurality of positions of the fingertips. Therefore, in a case where a plurality of fingertips result in erroneously entering the second space 14, it is possible to prevent the pointer 13 from moving to a position not intended by the user.
(1) It is to be noted that, in the first embodiment, the example has been given in which, in a case where a plurality of positions of fingertips are detected within the second space 14, the display control section 17 moves the display position of the pointer 13 only on the basis of the position of the fingertip having entered the second space 14 at first; however, other configurations may also be employed. For example, a configuration may be adopted in which, in a case where a plurality of positions of the fingertips detected by the position detection section 18 exist within the second space 14, the three-dimensional image displayed by the display unit 2 is changed on the basis of the plurality of positions of the fingertips. As the change in the three-dimensional image on the basis of the plurality of positions of the fingertips, for example, there may be employed reduction by pinch-in, enlargement by pinch-out, and movement and discarding of the three-dimensional object 12 by two-finger swipe. This makes it possible to move the display position of the pointer 13 on the basis of the position of a fingertip having entered the second space 14 at first, and to perform an operation other than the movement of the display position of the pointer 13 on the basis of the position of a fingertip having entered the second space 14 later. Therefore, it is possible to perform various operations on the three-dimensional image with an intuitive operation in the same manner as the operation of the movement of the display position of the pointer 13.
In addition, in a case where the configuration is adopted in which the three-dimensional image is changed on the basis of the positions of the plurality of fingertips, a configuration may further be adopted in which the display control section 17 sets a content of the change in the three-dimensional image on the basis of movement speed or movement acceleration of the plurality of fingertips. For example, a configuration may be adopted in which the three-dimensional object 12 is moved in a case where the movement speed or the movement acceleration of the fingertips at the time of execution of the two-finger swipe is equal to or less than a predetermined threshold value, whereas the three-dimensional object 12 pointed by the pointer 13 is discarded or a currently displayed three-dimensional image is switched to a next three-dimensional image in a case where the movement speed or the movement acceleration of the fingertips is larger than the threshold value. This makes it possible to perform more operations on the three-dimensional image by an intuitive operation in the same manner as the operation of the movement of the display position of the pointer 13.
(2) In addition, for example, as illustrated in
(3) In addition, although the example has been given in which the pointer 13 is not moved to remain at the current position, for example, in a case where determination is made that the position of the fingertip detected by the position detection section 18 does not exist within the second space 14, other configurations may also be employed. For example, a configuration may be adopted to erase the pointer 13.
(4) In addition, although, in the first embodiment, the example has been given in which the controller 4 is disposed inside the base 5, and the controller 4 is integrated with the display unit 2, other configurations may also be employed. For example, the controller 4 may be disposed outside the base 5, and the controller 4 may be separated from the display unit 2. In a case where the controller 4 is separated from the display unit 2, the controller 4 may be sized to be larger.
(5) In addition, although, in the first embodiment, the example has been given in which the cylindrical screen 6 is irradiated with the image light 10 to display the three-dimensional image, other configurations may also be employed. For example, a configuration may be adopted to display the three-dimensional image using an apparatus in which a liquid crystal display is disposed on each surface of a cube.
(6) In addition, although, in the first embodiment, the example has been given in which the display unit 2 that enables the three-dimensional image to be visually recognized from 360° around the display unit 2 is used, other configurations may also be employed. For example, a configuration may be adopted to use the display unit 2 in which an angular range enabling the three-dimensional image to be visually recognized is narrower than 360°.
Next, description is given of an image display apparatus according to a second embodiment of the present disclosure. An overall configuration of the image display apparatus of the second embodiment is similar to that illustrated in
The image display apparatus 1 of the second embodiment differs from the first embodiment in that the second space 14 is set as a cylindrical space provided at a position spaced apart from the first space 11. Specifically, in the second embodiment, as illustrated in
The shape of the second space 14 of the second embodiment is similar to the shape of the first space 11. That is, the ratio of radius R1:height H1 of the first space 11 is equal to the ratio of radius R2:height H2 of the second space 14. In the second space 14, there is set the cylindrical coordinate system r2θ2z2 with the center axis of the second space 14 being the z2 axis and an intersection of a bottom surface of the second space 14 and the z2 axis being the origin.
In addition, in the second embodiment, as illustrated in
In addition, in the second embodiment, a method for calculating the coordinates (r2′, θ2′, z2′) of the move destination of the pointer 13 in step S104 of the control processing illustrated in
As described above, in the image display apparatus 1 according to the second embodiment of the present disclosure, the second space 14 is set as a space provided at a position spaced apart from the first space 11, thus making it possible to operate the position of the pointer 13 more intuitively even when the detection target is not able to be inserted into the first space 11.
In addition, in the image display apparatus 1 according to the second embodiment of the present disclosure, the shape of the second space 14 is similar to the shape of the first space 11. Therefore, even in a case where the display unit 2 is sized to be large enough to be equal to or more than the height of a person or in a case where the display unit 2 is installed at a position away from an operator, for example, it is possible to alter the position of the pointer 13 with an intuitive operation.
In addition, in the image display apparatus 1 according to the first embodiment of the present disclosure, the display unit 2 includes the cylindrical screen 6 on which the three-dimensional image is displayed, and the first space 11 is set as a space of a hollow portion inside the screen 6, thus making it possible to display the three-dimensional object 12 in the screen 6.
In addition, in the image display apparatus 1 according to the second embodiment of the present disclosure, the cylindrical coordinate system r2θ2z2 with the center axis of the second space 14 being the z2 axis is set in the second space 14 to detect the coordinates (r2′, θ2′, z2′) at the position of the detection target using the cylindrical coordinate system r2θ2z2. Therefore, the coordinate system is suitable for the shape of the second space 14, thus making it possible to represent the position of the detection target more appropriately.
It is to be noted that the present technology may have the following configurations.
(1)
An image display apparatus including:
a display unit that displays a three-dimensional image, by which a three-dimensional object and a pointer appear to exist within a first space partitioned by a member configuring an outwardly convex outer surface, to be visually recognizable from a plurality of circumferential directions;
a position detection section that detects a position of a detection target; and
a display control section that moves a display position of the pointer displayed by the display unit on a basis of a position of the detection target that exists within a predetermined second space not overlapping the first space among the position of the detection target detected by the position detection section.
(2)
The image display apparatus according to (1), in which the second space includes a space surrounding a circumference of the first space.
(3)
The image display apparatus according to (2), in which
the display unit includes a cylindrical screen on which the three-dimensional image is displayed,
the first space includes a space of a hollow portion inside the screen, and
the second space includes a cylindrical space surrounding the circumference of the first space.
(4)
The image display apparatus according to (3), in which
a cylindrical coordinate system r2θ2z2 with a center axis of the second space being a z2 axis is set in the second space, and
the position detection section detects coordinates (r2′, θ2′, z2′) of the position of the detection target using the cylindrical coordinate system r2θ2z2.
(5)
The image display apparatus according to (4), in which
a cylindrical coordinate system r1θ1z1 with a center axis of the first space being a z1 axis is set in the first space, and
the display control section moves the display position of the pointer displayed by the display unit to allow values of θ2′ and z2′ of the coordinates (r2′, θ2′, z2′) detected by the position detection section and values of θ1′ and z1′ of coordinates (r1′, θ1′, z1′) of the pointer in the cylindrical coordinate system r1θ1z1 to be the same.
(6)
The image display apparatus according to (5), in which the display control section moves the display position of the pointer displayed by the display unit
to allow r1′ of the coordinates (r1′, θ1′, z1′) of the pointer in the cylindrical coordinate system r1θ1z1 to be “0” in a case where r2′ of the coordinates (r2′, θ2′, z2′) detected by the position detection section is same as a value of a radius of the first space, and
to allow r1′ of the coordinates (r1′, θ1′, z1′) to be same as the value of the radius of the first space in a case where r2′ of the coordinates (r2′, θ2′, z2′) is same as a value of a radius of the second space.
(7)
The image display apparatus according to any one of (1) to (6), including an information acquisition unit that is disposed at an upper portion or a lower portion of the display unit and acquires information concerning the position of the detection target, in which
the position detection section detects the position of the detection target on a basis of a result of detection of the information acquisition unit.
(8)
The image display apparatus according to (7), in which the information acquisition unit includes a TOF sensor, a camera, a touch panel, a capacitance sensor, or a magnetic sensor.
(9)
The image display apparatus according to any one of (1) to (8), including a reporting section that performs a predetermined reporting operation when the position of the detection target detected by the position detection section has entered the second space.
(10)
The image display apparatus according to (9), in which the reporting section performs, as the reporting operation, at least one of lighting of a light, blinking of the three-dimensional image, color change in the three-dimensional image, or output of a notification sound.
(11)
The image display apparatus according to any one of (1) to (10), in which
the detection target includes a fingertip of a person, and
the display control section moves, in a case where a plurality of positions of fingertips detected by the position detection section exist within the second space, the display position of the pointer displayed by the display unit only on a basis of a position of a fingertip having entered the second space at first among the plurality of the positions of the fingertips.
(12)
The image display apparatus according to any one of (1) to (10), in which
the detection target includes a fingertip of a person, and
the display control section changes, in a case where a plurality of positions of fingertips detected by the position detection section exist within the second space, the three-dimensional image displayed by the display unit on a basis of movements of the plurality of fingertips.
(13)
The image display apparatus according to (12), in which the display control section sets a content of the change in the three-dimensional image on a basis of movement speed or movement acceleration of the plurality of fingertips.
(14)
The image display apparatus according to (1), in which the second space includes a space provided at a position spaced apart from the first space.
(15)
The image display apparatus according to (14), in which a shape of the second space is similar to a shape of the first space.
(16)
The image display apparatus according to (1), in which
the display unit includes a cylindrical screen on which the three-dimensional image is displayed,
the first space includes a space of a hollow portion inside the screen, and
the second space includes a cylindrical space spaced apart from the first space.
(17)
The image display apparatus according to (16), in which
a cylindrical coordinate system r2θ2z2 with a center axis of the second space being a z2 axis is set in the second space, and
the position detection section detects coordinates (r2′, θ2′, z2′) of the position of the detection target using the cylindrical coordinate system r2θ2z2.
Number | Date | Country | Kind |
---|---|---|---|
2019-198491 | Oct 2019 | JP | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2020/037455 | 10/1/2020 | WO |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2021/085028 | 5/6/2021 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
20030142144 | Balakrishnan | Jul 2003 | A1 |
20040207599 | Kurtenbach | Oct 2004 | A1 |
20150355471 | Yoneno | Dec 2015 | A1 |
20160070356 | Aguirre | Mar 2016 | A1 |
20190138266 | Takechi et al. | May 2019 | A1 |
20190235643 | Kuribayashi | Aug 2019 | A1 |
20200389691 | Yoshizawa | Dec 2020 | A1 |
Number | Date | Country |
---|---|---|
105278112 | Jan 2016 | CN |
108140383 | Jun 2018 | CN |
109643206 | Apr 2019 | CN |
110199251 | Sep 2019 | CN |
2012-208705 | Oct 2012 | JP |
2013-105084 | May 2013 | JP |
2015-232633 | Dec 2015 | JP |
2018-014575 | Jan 2018 | JP |
2019-133395 | Aug 2019 | JP |
10-2018-0030767 | Mar 2018 | KR |
2018003862 | Jan 2018 | WO |
2018016095 | Jan 2018 | WO |
2018142524 | Aug 2018 | WO |
2018163945 | Sep 2018 | WO |
Entry |
---|
International Search Report and Written Opinion of PCT Application No. PCT/JP2020/037455, dated Nov. 10, 2020, 11 pages of ISRWO. |
Number | Date | Country | |
---|---|---|---|
20220365658 A1 | Nov 2022 | US |