The present invention relates to a photographing system, a photographed-portion setting apparatus, a photographing apparatus, and a photographing method and specifically relates to a photographing system, a photographed-portion setting apparatus, a photographing apparatus, and a photographing method for photographing a specific photographed portion set in a part of a photographed subject.
As a technique for supporting an inspection of a structure, such as a bridge, a technique in which damage, such as a crack, is automatically detected from an image obtained by photographing the surface of the structure is known. In this case, as an image for inspection, a panoramic composite image is often used.
In a case of generating a panoramic composite image, it is necessary to photograph a subject without omission. However, in a case where the subject extends over a broad area, an omission of photographing may occur or an image of low image quality may be captured. In such cases, the corresponding region is re-photographed.
However, it is difficult to accurately identify the portion to be re-photographed in a broad site. Specifically, in a case where a person giving an instruction for re-photographing is different from a person actually performing photographing on site, it is difficult to properly communicate necessary information to the photographing person side and the re-photographing operation requires considerable effort.
JP2012-160904A proposes a technique for supporting such a re-photographing operation in which support information necessary for photographing a specific portion is generated and communicated to a photographing person. The support information at least includes information about the photographing position and the photographing direction, and the photographing person adjusts the photographing position and the photographing method on the basis of the support information to perform photographing.
However, the technique of JP2012-160904A has a disadvantage that photographing needs to be performed while the photographing position and the photographing direction are accurately measured. To realize this, sophisticated measuring means is required on the photographing apparatus side, which is also a disadvantage.
The present invention has been made in view of the above-described circumstances, and an object thereof is to provide a photographing system, a photographed-portion setting apparatus, a photographing apparatus, and a photographing method for facilitating photographing of a specific portion of a photographed subject.
Means for addressing the above-described issue are as follows.
(1) A photographing system including: a photographed-portion setting apparatus that includes a first display unit, a photographed-subject image acquisition unit that acquires an image of a photographed subject, a display control unit that displays the image of the photographed subject on the first display unit, a photographed-portion setting unit that sets a photographed portion in the image of the photographed subject, a location information generation unit that generates information indicating a location of the set photographed portion, and a reference image generation unit that generates a reference image by cutting an image including a surrounding area of the photographed portion from the image of the photographed subject; and a photographing apparatus that includes a photographing unit, a photographing information acquisition unit that acquires the information indicating the location of the photographed portion and the reference image, a first guide unit that provides a guide for photographing the photographed portion on the basis of the information indicating the location of the photographed portion, a reference image detection unit that detects in real time, the reference image from an image captured by the photographing unit, and a second guide unit that provides a guide for photographing the photographed portion on the basis of a result of detection by the reference image detection unit.
(2) The photographing system of (1) above, in which the photographing apparatus further includes a second display unit, and the second guide unit displays the image captured by the photographing unit on the second display unit in real time as a live preview image and in a case where the reference image is detected, superimposes on the live preview image and displays information indicating a region of the reference image.
(3) The photographing system of (2) above, in which in a case where the reference image is detected, the second guide unit superimposes on the live preview image and displays a frame indicating the region of the reference image.
(4) The photographing system of (3) above, in which the second guide unit displays the frame in different display forms in a case where part of the reference image is detected and in a case where the entire reference image is detected.
(5) The photographing system of (4) above, in which the second guide unit displays the frame in different colors in a case where part of the reference image is detected and in a case where the entire reference image is detected.
(6) The photographing system of any one of (2) to (5) above, in which the first guide unit displays the information indicating the location of the photographed portion on the second display unit.
(7) The photographing system of any one of (1) to (6) above, in which the photographing apparatus further includes an unattended vehicle that is equipped with the photographing unit, and a remote operation unit that remotely operates the unattended vehicle and the photographing unit.
(8) The photographing system of (7) above, in which the first guide unit generates guide information for guiding the unattended vehicle to move to a position from which the photographed portion is photographable, on the basis of the information indicating the location of the photographed portion, and the photographing apparatus further includes a movement control unit that controls movement of the unattended vehicle in accordance with the guide information generated by the first guide unit.
(9) The photographing system of (7) or (8) above, in which the second guide unit generates guide information for guiding the unattended vehicle to a position from which the photographed portion is photographed, on the basis of the result of detection by the reference image detection unit, and the photographing apparatus further includes an automatic adjustment unit that automatically adjusts a position of the unattended vehicle in accordance with the guide information generated by the second guide unit.
(10) The photographing system of (7) or (8) above, in which the unattended vehicle includes a photographing direction change unit that changes a photographing direction of the photographing unit, the second guide unit generates guide information for providing a guide to a position of the unattended vehicle and/or the photographing direction of the photographing unit, on the basis of the result of detection by the reference image detection unit, and the photographing apparatus further includes an automatic adjustment unit that automatically adjusts the position of the unattended vehicle and/or the photographing direction of the photographing unit in accordance with the guide information generated by the second guide unit.
(11) The photographing system of any one of (1) to (10) above, in which the photographed-portion setting apparatus further includes a photographed-portion specifying unit that specifies the photographed portion on a screen of the first display unit on which the image of the photographed subject is displayed, and the photographed-portion setting unit sets a portion specified by the photographed-portion specifying unit as the photographed portion.
(12) The photographing system of (11) above, in which the photographed-portion specifying unit specifies the photographed portion by specifying a center of the portion that is set as the photographed portion.
(13) The photographing system of (11) above, in which the photographed-portion specifying unit specifies the photographed portion by specifying a region including the portion that is set as the photographed portion.
(14) The photographing system of any one of (1) to (10) above, in which the photographed-portion setting apparatus further includes a re-photographed portion detection unit that detects a portion that is to be re-photographed from the image of the photographed subject, and the photographed-portion setting unit sets the portion detected by the re-photographed portion detection unit as the photographed portion.
(15) The photographing system of (14) above, in which the re-photographed portion detection unit detects a region whose image is partially missing and/or a region having an image quality less than or equal to a specified criterion as the portion that is to be re-photographed.
(16) The photographing system of any one of (1) to (15) above, in which the reference image generation unit generates the reference image by cutting an image of a certain area centered on the photographed portion from the image of the photographed subject.
(17) The photographing system of any one of (1) to (16) above, in which the location information generation unit generates an image by marking the photographed portion on the image of the photographed subject, as the information indicating the location of the photographed portion.
(18) The photographing system of (17) above, in which the location information generation unit generates an image by marking the photographed portion on the image of the photographed subject and reduces the generated image so as to have a specified size, as the information indicating the location of the photographed portion.
(19) The photographing system of any one of (1) to (16) above, in which in a case where the photographed subject is divided into a plurality of zones and each zone is assigned a lot number, the location information generation unit generates information of a lot number at which a portion that is set as the photographed portion is present, as the information indicating the location of the photographed portion.
(20) The photographing system of any one of (1) to (16) above, in which in a case where the photographed subject is a structure, the location information generation unit generates information about a position, on the globe, of a portion set as the photographed portion, as the information indicating the location of the photographed portion.
(21) The photographing system of any one of (1) to (20) above, in which the image of the photographed subject acquired by the photographed-subject image acquisition unit is an image obtained by performing a panoramic composition for a plurality of images.
(22) A photographed-portion setting apparatus including: a first display unit; a photographed-subject image acquisition unit that acquires an image of a photographed subject; a display control unit that displays the image of the photographed subject on the first display unit; a photographed-portion setting unit that sets a photographed portion in the image of the photographed subject; a location information generation unit that generates information indicating a location of the set photographed portion; and a reference image generation unit that generates a reference image by cutting an image including a surrounding area of the photographed portion from the image of the photographed subject.
(23) A photographing apparatus including: a photographing unit; a photographing information acquisition unit that acquires information indicating a location of a photographed portion set for a photographed subject and a reference image generated by cutting an image including a surrounding area of the photographed portion from an image of the photographed subject; a first guide unit that provides a guide for photographing the photographed portion on the basis of the information indicating the location of the photographed portion; a reference image detection unit that detects in real time, the reference image from an image captured by the photographing unit; and a second guide unit that provides a guide for photographing the photographed portion on the basis of a result of detection by the reference image detection unit.
(24) A photographing method including: a step of acquiring information indicating a location of a photographed portion set for a photographed subject and a reference image generated by cutting an image including a surrounding area of the photographed portion from an image of the photographed subject; a step of providing a guide for photographing the photographed portion on the basis of the information indicating the location of the photographed portion; a step of detecting in real time, the reference image from a captured image; and a step of providing a guide for photographing the photographed portion on the basis of a result of detection of the reference image.
(25) The photographing method of (24) above, further including: a step of acquiring the image of the photographed subject; a step of setting the photographed portion in the image of the photographed subject; a step of generating the information indicating the location of the set photographed portion; and a step of generating the reference image by cutting the image including the surrounding area of the photographed portion from the image of the photographed subject.
The present invention can facilitate photographing of a specific portion of a photographed subject.
Hereinafter, preferred embodiments of the present invention will be described with reference to the attached drawings.
The photographing system of this embodiment is configured as a system for photographing a photographed portion that is set in a part of a photographed subject with pinpoint accuracy. Specifically, the photographing system of this embodiment is configured as a system for manual photographing.
As illustrated in
The photographed-portion setting apparatus 10 and the digital camera 100 are configured so as to enable mutual communication. Communication performed between the photographed-portion setting apparatus 10 and the digital camera 100 is short-range wireless communication (for example, communication conforming to a short-range wireless communication standard, such as the NFC (Near Field Communication) standard, Bluetooth (registered trademark), or WiFi (Wireless Fidelity)). Alternatively, the communication is performed via a communication network. The communication network is formed of, for example, a wireless communication network and base stations conforming to 3G (3rd Generation), WiMAX (Worldwide Interoperability for Microwave Access), or LTE (Long Term Evolution) and the Internet.
The photographed-portion setting apparatus 10 has a function of accepting specification of a photographed portion from a user and setting the photographed portion and a function of generating information necessary for photographing the set photographed portion with the digital camera 100.
The photographed-portion setting apparatus 10 is formed of a computer. That is, the computer executes a specific program (photographed-portion setting program) to thereby function as the photographed-portion setting apparatus.
As illustrated in
In the photographed-portion setting apparatus 10 of this embodiment, the display 15 is an example of a first display unit, and the touch panel 16 is an example of a photographed-portion specifying unit.
The photographed-portion setting apparatus 10 has functions of a photographed-subject image acquisition unit 30A that acquires an image of a photographed subject, a display control unit 30B that controls display on the display 15, a photographed-portion setting unit 30C that sets a photographed portion, a location information generation unit 30D that generates information indicating the location of the set photographed portion, a reference image generation unit 30E that generates a reference image, and a photographing information transmission processing unit 30F that transmits the information indicating the location of the photographed portion and the reference image to the digital camera 100. The functions of the units are implemented by the CPU 11 executing a specific program (photographed-portion setting program).
The photographed-subject image acquisition unit 30A acquires an image of a photographed subject. This image is an image of the entire photographed subject. In a case where the photographed subject is photographed in sections and panoramic composition is performed, an image obtained as a result of panoramic composition is acquired. The photographed-subject image acquisition unit 30A acquires an image of a photographed subject from an external device (for example, a server) via the communication unit 21 or the short-range wireless communication unit 22. Alternatively, the photographed-subject image acquisition unit 30A acquires an image of a photographed subject from the memory card 25 via the media drive 24.
The display control unit 30B controls display on the display 15. The display control unit 30B is an example of a display control unit. The display control unit 30B displays the image of the photographed subject acquired by the photographed-subject image acquisition unit 30A on the display 15.
The photographed-portion setting unit 30C sets a photographed portion in the image of the photographed subject. The set photographed portion is a part of the photographed subject. The photographed-portion setting unit 30C sets a photographed portion on the basis of input from the touch panel 16. That is, the photographed-portion setting unit 30C sets a portion touched on the screen of the display 15 on which the image of the photographed subject is displayed as a photographed portion.
The photographed portion is specified as a point or a region. For example, in a case where a relatively small region is set as a photographed portion, the photographed portion is specified as a point. On the other hand, in a case where a relatively large region is specified as a photographed portion, the photographed portion is specified as a region. In a case of specification as a point, a user touches the center (including an area that is substantially recognized as the center) of a region that is set as a photographed portion to specify the photographed portion. In a case of specification as a region, a user touches a region that is set as a photographed portion so as to surround the region to specify the photographed portion.
The location information generation unit 30D generates information indicating the location of the photographed portion set by the photographed-portion setting unit 30C. In this embodiment, as information indicating the location of the photographed portion, the location information generation unit 30D generates an image by marking the photographed portion on the image of the photographed subject. The form of marking is not specifically limited. Any form may be employed as long as the set photographed portion can be recognized in the image. For example, a form can be employed in which the set photographed portion is pointed by an arrow or the set photographed portion is surrounded with a frame to indicate the location of the set photographed portion.
The reference image generation unit 30E generates a reference image by cutting an image including the set photographed portion and its surrounding area from the image of the photographed subject. For example, in a case where the photographed portion is specified as a point, the reference image generation unit 30E cuts an image of a certain area centered on the specified point (for example, an image of a rectangular region having a certain size and centered on the specified point) from the image of the photographed subject to generate a reference image. For example, in a case where the photographed portion is specified as a region, the reference image generation unit 30E cuts an image of a rectangular region that includes the specified region from the image of the photographed subject to generate a reference image. Accordingly, a reference image is generated so as to include a proper image of the surrounding area (an image that is successfully captured and has a high resolution).
The photographing information transmission processing unit 30F transmits the information indicating the location of the photographed portion generated by the location information generation unit 30D and the reference image generated by the reference image generation unit 30E to the digital camera 100. The photographing information transmission processing unit 30F transmits the information indicating the location of the photographed portion and the reference image to the digital camera 100 via the communication unit 21 or the short-range wireless communication unit 22.
The digital camera 100 has a function of providing a guide for photographing a photographed portion set by the photographed-portion setting apparatus 10 in addition to usual photographing and playback functions.
As illustrated in
The imaging lens 110 forms an optical image of a photographic subject on a light-receiving surface of the image sensor 112. The imaging lens 110 has a focus adjusting function and a zoom function. The focus adjusting function and the zoom function are driven by a lens driving unit 110A and executed. The imaging lens 110 has a diaphragm and a shutter. The diaphragm and the shutter are driven by the lens driving unit 110A and operated. The lens driving unit 110A drives each part of the imaging lens 110 in accordance with an instruction from the camera microcomputer 140.
The image sensor 112 is formed of, for example, a two-dimensional solid-state imaging element, such as a CCD (charge-coupled device) image sensor or a CMOS (complementary metal-oxide semiconductor) image sensor. The image sensor 112 is driven by a sensor driving unit 112A and operated. The sensor driving unit 112A drives the image sensor 112 in accordance with an instruction from the camera microcomputer 140.
In the digital camera 100 of this embodiment, the imaging lens 110 and the image sensor 112 form a photographing unit.
The analog signal processing unit 114 takes in an analog image signal for each pixel output from the image sensor 112, performs specific signal processing (for example, correlative double sampling processing and amplification processing) for the analog image signal, and outputs a digitized signal.
The digital signal processing unit 116 takes in a digital image signal output from the analog signal processing unit 114, performs specific signal processing (for example, gradation transformation processing, white balance correction processing, gamma-correction processing, demosaicing processing, and YC conversion processing) for the digital image signal, and generates image data.
The media drive 118 reads and writes data from and to a memory card 120. The media drive 118 reads and writes data from and to the memory card 120 in accordance with an instruction from the camera microcomputer 140.
The back-side monitor 122 is formed of, for example, a liquid crystal display (LCD). The back-side monitor 122 is provided on the back side of the camera body. The back-side monitor 122 is an example of a second display unit. Display on the back-side monitor 122 is controlled by the camera microcomputer 140. The camera microcomputer 140 controls display on the back-side monitor 122 via a monitor driver 122A.
The short-range wireless communication unit 124 uses an antenna 124A to perform short-range wireless communication with an external device (for example, the photographed-portion setting apparatus 10).
The communication unit 126 uses an antenna 126A to wirelessly communicate with a nearby base station.
The flash 128 includes, for example, a xenon tube or an LED (light emitting diode) as a light source and causes the light source to emit light to irradiate a photographic subject with flash light. Light emission by the flash 128 is controlled by the camera microcomputer 140 via a flash light emission control unit 128A.
The microphone 130 collects external sound. The speaker 132 externally outputs sound. The audio signal processing unit 134 performs specific signal processing for an audio signal input from the microphone 130 and outputs a digitized signal. The audio signal processing unit 134 performs specific signal processing for audio data fed from the camera microcomputer 140 and outputs resulting sound via the speaker 132.
The camera operation unit 136 includes various operating members including a power button, a release button, a zoom key, a menu button, an OK button, a cancel button, a cross-shaped button, and a mode switching switch and a signal processing circuit for the operating members and outputs a signal based on an operation of each operating member to the camera microcomputer 140.
The camera microcomputer 140 is a control unit that centrally controls overall operations of the digital camera 100. The camera microcomputer 140 is formed of a microcomputer that includes a CPU, a ROM, and a RAM and executes a specific program to thereby function as, for example, a control unit of each part.
As described above, the digital camera 100 has the function of providing a guide for photographing a photographed portion set by the photographed-portion setting apparatus 10 in addition to usual photographing and playback functions. The usual photographing and playback functions are the same as those of a typical digital camera. Therefore, the guide function is described below.
As illustrated in
The photographing information acquisition unit 140A acquires location information and a reference image from the photographed-portion setting apparatus 10. The photographing information acquisition unit 140A, for example, communicates with the photographed-portion setting apparatus 10 via the communication unit 126 and acquires location information and a reference image from the photographed-portion setting apparatus 10. Alternatively, the photographing information acquisition unit 140A performs short-range wireless communication with the photographed-portion setting apparatus 10 via the short-range wireless communication unit 124 and acquires location information and a reference image from the photographed-portion setting apparatus 10. The acquired location information is fed to the first guide unit 140B, and the acquired reference image is fed to the reference image detection unit 140C.
The first guide unit 140B provides a guide for photographing a photographed portion on the basis of the location information. As described above, in the photographing system of this embodiment, location information is formed of a specific marking image. That is, location information is formed of an image obtained by marking the photographed portion on an image of the photographed subject. The first guide unit 140B displays this marking image on the back-side monitor 122 to provide a guide for photographing the photographed portion. The photographing person views the marking image displayed on the back-side monitor 122 to grasp the approximate position of the photographed portion.
The reference image detection unit 140C detects in real time, a region that matches the reference image from an image captured by the image sensor 112 through the imaging lens 110. A technique for this type of image recognition is publicly known, and therefore, a detailed description thereof is omitted (for example, various methods can be employed including a method in which feature points are extracted from the reference image and the captured image and compared with each other to detect a region of the reference image from the captured image).
The second guide unit 140D provides a guide for photographing the photographed portion on the basis of the result of detection by the reference image detection unit 140C. In this embodiment, the image captured by the image sensor 112 through the imaging lens 110 is displayed on the back-side monitor 122 in real time as a live preview image and in a case where the reference image is detected, information indicating the region of the reference image is superimposed on the live preview image and displayed to thereby provide a guide for photographing the photographed portion. More specifically, a frame (detection frame) that indicates the region of the reference image is superimposed on the live preview image and displayed to thereby provide a guide for photographing the photographed portion. At this time, the second guide unit 140D displays the frame in different display forms in a case where part of the reference image is detected from the captured image and in a case where the entire reference image is detected from the captured image. More specifically, in the case where part of the reference image is detected from the captured image, the second guide unit 140D displays a red frame. On the other hand, in the case where the entire reference image is detected from the captured image, the second guide unit 140D displays a blue frame. The photographing person adjusts the photographing angle and so on such that the frame fits in the angle of view and performs photographing. Accordingly, the specified photographed portion can be accurately photographed.
Now, a photographing method using the photographing system of this embodiment is described.
Here, an example case where a photographed subject is photographed in sections and panoramic composition is performed and where re-photographing of a part of the photographed subject is performed is described. Here, an example case where a floor slab of a bridge is photographed is described. Specifically, an example case where a floor slab is photographed for the purpose of inspection is described.
First, a bridge is described.
A bridge 1 of this example has a three-dimensional construction including main girders 2, cross girders 3, sway bracing 4, lateral bracing 5, and a floor slab 6 and has a construction such that these members are coupled with bolts and rivets, by welding, and so on. Over the main girders 2 and so on, the floor slab 6 on which vehicles, and so on are traveling is formed by pouring concrete. The floor slab 6 is typically built of reinforced concrete. The main girders 2 are members that are laid between abutments or piers to bear the load of vehicles and so on traveling on the floor slab 6. The main girders 2 have a plane (plane in the vertical direction) orthogonal to a plane (horizontal plane) of the floor slab 6. The cross girders 3 are members that each couple the main girders 2. The cross girders 3 are provided such that the plurality of main girders 2 bear the load. The sway bracing 4 and the lateral bracing 5 are members that couple the main girders 2 with each other. The sway bracing 4 and the lateral bracing 5 are provided to resist lateral loads imposed by winds and earthquakes.
Photographing of the floor slab 6 for the purpose of inspection is performed on a per panel basis. A panel is one division of the floor slab 6 partitioned by the main girders 2 and the cross girders 3. Therefore, in photographing of the floor slab 6 for the purpose of inspection, a panel is a photographed subject, photographing in sections is performed on a per panel basis, and panoramic composition is performed.
Now, a method for photographing a panel in a case of panoramic composition is described.
In a case of panoramic composition, a photographed subject is divided into a plurality of regions and the plurality of regions are separately photographed (photographed in sections). In
The photographing person (inspection technician) basically faces the floor slab 6 and performs photographing from a constant distance. Further, the photographing person basically performs photographing such that adjacent photographed regions partially overlap.
A plurality of images acquired by photographing are subjected to panoramic composition to form a single image. A technique for panoramic composition is publicly known, and therefore, a detailed description thereof is omitted. For example, correspondence points in images are detected to concatenate the images and compose a single image. At this time, each image is subjected to corrections including a sizing correction, a tilt correction, and a rotation correction as necessary.
Such an omission of photographing is usually found after panoramic composition. In a case where an omission of photographing is found, the portion P omitted from photographing is re-photographed, and a composition process is re-performed. A procedure for re-photographing a portion omitted from photographing by using the photographing system of this embodiment is described below.
A process for re-photographing is divided into setting of a photographed portion and actual photographing. Setting of a photographed portion is performed by using the photographed-portion setting apparatus 10, and photographing is performed by using the digital camera 100.
As described above, setting of a photographed portion is performed by using the photographed-portion setting apparatus 10.
First, an image of a photographed subject is acquired (step S1). An image of a photographed subject is acquired from an external device (for example, a server) via the communication unit 21 or the short-range wireless communication unit 22. Alternatively, an image of a photographed subject is acquired from the memory card 25 via the media drive 24. An image of a photographed subject here is an image of a panel and is the panoramic composite image PI acquired by photographing the panel in sections.
Next, the acquired image of the photographed subject is displayed on the display 15 (step S2).
Next, specification of a photographed portion is accepted on the screen of the display 15 on which the image of the photographed subject is displayed (step S3). Specification of a photographed portion is performed by touching a portion that is set as a photographed portion on the screen of the display 15 with a finger. In this example, the portion P omitted from photographing is re-photographed, and therefore, the portion P omitted from photographing is touched on the screen to specify a photographed portion. In this example, the portion P omitted from photographing is small, and therefore, the photographed portion is specified as a point. In this case, the center of the portion that is set as the photographed portion (the portion omitted from photographing) is touched with a finger to specify the photographed portion.
Next, the specified portion is set as the photographed portion (step S4). The specified portion is detected by the touch panel 16. Therefore, the photographed portion is set on the basis of the result of detection by the touch panel 16.
Next, information indicating the location of the set photographed portion (location information) is generated (step S5). In this embodiment, as the location information, an image obtained by marking the photographed portion on the image of the photographed subject (marking image) is generated.
Next, a reference image is generated (step S6). A reference image is generated by cutting an image that includes the set photographed portion and its surrounding area from the image of the photographed subject.
The series of steps described above is performed and the process for setting a photographed portion is completed. Subsequently, actual photographing is performed on the basis of the set information.
Photographing is performed by using the digital camera 100. The digital camera 100 acquires location information (in this example, a marking image) and a reference image from the photographed-portion setting apparatus 10 and provides a guide for photographing the photographed portion on the basis of the acquired information. The guide is provided while the mode of the digital camera 100 is set to a specific mode (photographing guide mode).
First, location information and a reference image are acquired (step S11). Location information and a reference image are acquired from the photographed-portion setting apparatus 10. The digital camera 100 communicates with the photographed-portion setting apparatus 10 and acquires location information and a reference image from the photographed-portion setting apparatus 10.
Next, the acquired location information is displayed on the back-side monitor 122 (step S12). As described above, the location information is a marking image obtained by marking the photographed portion on an image of the photographed subject, and therefore, the marking image is displayed on the back-side monitor 122.
Next, on the basis of operation input from the camera operation unit 136, the presence or absence of an instruction for starting photographing is determined (step S13). An instruction for starting photographing is given by, for example, half-push of the release button. The user (photographing person) gives an instruction for starting photographing at a position from which the photographed portion can be photographed.
In a case where an instruction for starting photographing is given, a live preview is started (step S14). An image captured by the image sensor 112 through the imaging lens 110 is displayed on the back-side monitor 122 in real time.
In a case where no instruction for starting photographing is given, the presence or absence of an instruction for termination is determined (step S15). For example, in a case of switching to another mode (for example, a usual photographing mode or a playback mode) or in a case of turning off the power, it is determined that an instruction for termination is given. In this case, the photographing guide mode ends.
Simultaneously with the start of the live preview, the reference image is detected from the captured image in real time (step S16). On the basis of the result of detection, it is determined whether the reference image is detected (step S17).
In a case where the reference image is detected, it is determined whether the reference image fits in the angle of view (step S18). That is, it is determined whether the entire reference image is detected.
In a case where the reference image fits in the angle of view, a blue detection frame FB indicating the region of the reference image is superimposed on the live preview image and displayed (step S19).
On the other hand, in a case where the reference image does not fit in the angle of view, that is, only part of the reference image is detected, a red detection frame FR indicating the region of the reference image is superimposed on the live preview image and displayed (step S20).
Thereafter, the presence or absence of an instruction for actual photographing is determined (step S21). That is, the presence or absence of an instruction for taking in an image for recording is determined. An instruction for actual photographing is given by pressing the release button.
In a case where an instruction for actual photographing is given, a process for actual photographing is performed, and an acquired image is recorded (step S22). That is, an image for recording is taken in and recorded to the memory card 120. Thereafter, the photographing guide mode ends.
On the other hand, in a case where no instruction for actual photographing is given, the presence or absence of an instruction for termination is determined (step S23). In a case where it is determined that an instruction for termination is given, the photographing guide mode ends. On the other hand, in a case where it is determined that no instruction for termination is given, the flow returns to step S16, and detection of the reference image is performed again.
Accordingly, with the photographing system of this embodiment, a guide for photographing is appropriately provided, and therefore, a specified photographed portion can be easily photographed.
As described above, a photographed portion may be specified by touching a region that is set as the photographed portion so as to surround the region on the screen on which an image of the photographed subject is displayed. That is, a region that is set as a photographed portion is surrounded with a frame to specify the photographed portion.
In a case where a region that is set as a photographed portion is surrounded with a frame to specify the photographed portion, as illustrated in
It is preferable to make the position and size of the frame W indicating a specified region be adjustable. For example, it is preferable to make the position of the frame W be adjustable by touching and swiping the region of the frame W and to make the size of the frame W be enlarged or reduced by performing a pinch-in or pinch-out operation in the region of the frame W.
Note that in the above-described embodiment, a configuration is employed in which a photographed portion is specified on the touch panel 16; however, means for specifying a photographed portion is not limited to this. In a case where the computer that forms the photographed-portion setting apparatus 10 includes an input device, such as a mouse, a joystick, or a keyboard, a configuration can be employed in which such an input device is used to specify a photographed portion.
In the above-described embodiment, an example case has been described where a portion omitted from photographing is specified as a photographed portion; however, a portion that is set as a photographed portion is not limited to this. For example, in a panoramic composite image, a region having an image quality less than or equal to a specified criterion needs to be re-photographed, and therefore, such a region can be specified as a photographed portion.
In the above-described embodiment, an example case where only one photographed portion is set has been described; however, a plurality of photographed portions can be set.
In the above-described embodiment, a configuration is employed in which a user specifies a portion that is set as a photographed portion on the screen on which an image of the photographed subject is displayed to set the photographed portion; however, a configuration can be employed in which a photographed portion is automatically set.
The photographed-portion setting apparatus 10 of this example automatically detects a portion that needs to be re-photographed from an image of the photographed subject and automatically sets a photographed portion. Specifically, the photographed-portion setting apparatus 10 of this example is configured as an apparatus that automatically detects from a panoramic composite image, a region whose image is partially missing (a region whose image is incomplete) and a region having an image quality less than or equal to a specified criterion and sets the regions as photographed portions.
As illustrated in
The re-photographed portion detection unit 30G analyzes an image of a photographed subject and detects a portion that needs to be re-photographed. In this example, the re-photographed portion detection unit 30G detects from a panoramic composite image, a region whose image is partially missing and a region having an image quality less than or equal to a specified criterion as portions that need to be re-photographed. The result of detection by the re-photographed portion detection unit 30G is fed to the photographed-portion setting unit 30C.
The photographed-portion setting unit 30C sets a photographed portion on the basis of the result of detection by the re-photographed portion detection unit 30G. For example, the photographed-portion setting unit 30C surrounds the detected region with a rectangular frame to set the photographed portion. Information about the set photographed portion is fed to the location information generation unit 30D, the reference image generation unit 30E, and the display control unit 30B.
The display control unit 30B displays the information indicating the set photographed portion on the display 15. For example, the display control unit 30B superimposes on the image of the photographed subject that is being displayed on the display 15 and displays a frame indicating the photographed portion.
Accordingly, a configuration can be employed in which a photographed portion is automatically set. In such a case where a portion that needs to be re-photographed is automatically detected and set as a photographed portion, it is preferable to allow the user to correct the automatically set photographed portion. For example, in the above-described example, a frame that indicates a photographed portion is superimposed on the image of the photographed subject that is being displayed on the display 15 and displayed, and therefore, it is preferable to allow correction of the position, size, and so on of the frame as necessary. Further, it is preferable to allow addition, deletion, and so on of a frame as necessary. A correction is made by using the touch panel 16.
In the above-described embodiment, as information indicating the location of a photographed portion, an image obtained by marking the photographed portion on an image of the photographed subject is generated; however, a form of information indicating the location of a photographed portion is not limited to this. Any information may be used as long as a photographed portion can be identified.
In a case where a photographed subject is divided into a plurality of zones and each zone is assigned a lot number, the location of a photographed portion can be indicated by using information of a lot number.
In the example illustrated in
In photographing for the purpose of panoramic composition, it is a common practice to divide a photographed subject into regions in a grid from and assign a lot number to each divided region. In such a case, information of lot numbers used in photographing can be used as is.
As illustrated in
In the example illustrated in
In a case where a photographed subject is a structure such as a bridge, GPS information of a portion specified as a photographed portion (information about the position on the globe) may be used as location information.
For example, in a case of photographing by using a camera including a GPS function, GPS information is recorded to tag information of the image (for example, Exif (Exchangeable image file format) tag information). Therefore, in a case where an image of a photographed subject is a panoramic composite image obtained by performing panoramic composition for a plurality of images and each of the images before composition has GPS information, the GPS information of each image can be used to acquire GPS information of a photographed portion. For example, GPS information of an image of a portion specified as a photographed portion or GPS information of an image of its vicinity is acquired, and GPS information of the portion specified as a photographed portion can be acquired accordingly. Alternatively, from GPS information of an image of a vicinity of a portion specified as a photographed portion, GPS information of the photographed portion is calculated, and GPS information of the photographed portion can be acquired accordingly.
In a case of a structure such as a bridge, GPS information can be acquired by using map data, data of a design drawing, and so on.
Note that GPS information at least includes information of the latitude and the longitude and may additionally include information of the altitude.
In a case where an image obtained by marking a photographed portion on an image of the photographed subject is generated as information indicating the location of the photographed portion, the image may be reduced so as to have a specified size and the generated reduced image may be used as a marking image. Accordingly, a load of, for example, transmitting data of the image to the digital camera 100 can be reduced. It is preferable to set the reduced size by taking into consideration the resolution and so on of a display unit included in the photographing apparatus.
In a case where lot-number information is used as information indicating the location of a photographed portion, the digital camera 100 acquires and displays on the back-side monitor 122 the lot-number information. In this case, the digital camera 100 may simultaneously acquire an image of the photographed subject and displays the image on the back-side monitor 122 together with the lot-number information.
In a case where GPS information is used as information indicating the location of a photographed portion, the digital camera 100 acquires and displays on the back-side monitor 122 the GPS information.
In a case where the digital camera 100 includes a GPS function, it is preferable to acquire GPS information of the current location and display the GPS information on the back-side monitor 122 together with GPS information of a photographed portion.
In the case where the digital camera 100 includes a GPS function, GPS information of the current location is acquired and compared with GPS information of a photographed portion. In a case where the pieces of GPS information match or the current location is within a certain area centered on the photographed portion, sound or display may be used to give a notification that the pieces of GPS information match or the current location is within a certain area centered on the photographed portion.
In the above-described embodiment, the photographing apparatus is formed of a digital camera; however, the configuration of the photographing apparatus is not limited to this. The photographing apparatus can be formed of a portable electronic device including a photographing function, such as a smartphone, a tablet computer, or a laptop computer.
The configuration of the photographing apparatus of the photographing system of this embodiment is different from that in the photographing system of the above-described first embodiment. The photographing system of this embodiment performs photographing with an unattended aircraft equipped with a digital camera.
As illustrated in
The unattended aircraft 212 is a drone and has an airframe 216 including a plurality of propellers 218 that are rotated to thereby fly in the air. The unattended aircraft 212 is an example of an unattended vehicle.
The digital camera 210 is an example of a photographing unit. The digital camera 210 has a pan and tilt function and can change the photographing direction. The pan and tilt function is implemented by a pan and tilt mechanism not illustrated. The pan and tilt mechanism is an example of a photographing direction change unit. The digital camera 210 is remotely operated by the controller 214 and panned and tilted.
The controller 214 wirelessly communicates with the unattended aircraft 212 and remotely operates the unattended aircraft 212 and the digital camera 210. The controller 214 is an example of a remote operation unit.
The controller 214 includes an operation unit 220, a controller display 222, a wireless communication unit 224, and a controller microcomputer 226.
The operation unit 220 includes various operating members for operating the unattended aircraft 212 and the digital camera 210 and a signal processing circuit for the operating members and outputs a signal based on an operation of each operating member to the controller microcomputer 226. The operating members for operating the unattended aircraft 212 include, for example, an operating member for instructing the unattended aircraft 212 to climb or descend and an operating member for instructing the unattended aircraft 212 to circle. The operating members for operating the digital camera 210 include operating members of a typical camera (a release button, a zoom button, and so on) and a pan and tilt operating member.
The controller display 222 is formed of, for example, an LCD. The controller display 222 is an example of a second display unit. Display on the controller display 222 is controlled by the controller microcomputer 226. The controller microcomputer 226 controls display on the controller display 222 via a display driver 222A.
The wireless communication unit 224 uses an antenna 224A to wirelessly communicate with the unattended aircraft 212 and the photographed-portion setting apparatus 10.
The controller microcomputer 226 is formed of a microcomputer that includes a CPU, a ROM, and a RAM. The controller microcomputer 226 generates a control signal corresponding to an operation of the operation unit 220 and transmits the control signal to the unattended aircraft 212 via the wireless communication unit 224. The controller microcomputer 226 acquires an image captured by the digital camera 210, information about the flight state of the unattended aircraft 212, and so on from the unattended aircraft 212 via the wireless communication unit 224 and displays the received image and information on the controller display 222. Further, the controller microcomputer 226 acquires location information of a photographed portion and a reference image from the photographed-portion setting apparatus 10 to provide a guide for photographing the photographed portion.
As illustrated in
The photographing information acquisition unit 226A acquires location information and a reference image from the photographed-portion setting apparatus 10. The photographing information acquisition unit 226A communicates with the photographed-portion setting apparatus 10 via the wireless communication unit 224 and acquires location information and a reference image from the photographed-portion setting apparatus 10. The acquired location information is fed to the first guide unit 226B, and the acquired reference image is fed to the reference image detection unit 226C.
The first guide unit 226B provides a guide for photographing a photographed portion on the basis of the location information. As in the above-described first embodiment, it is assumed that location information is formed of a marking image (an image obtained by marking the photographed portion on an image of the photographed subject). The first guide unit 226B displays the marking image on the controller display 222 to provide a guide for photographing the photographed portion. The photographing person views the marking image displayed on the controller display 222 to grasp the approximate position of the photographed portion.
The reference image detection unit 226C detects in real time, a region that matches the reference image from an image captured by the digital camera 210.
The second guide unit 226D provides a guide for photographing the photographed portion on the basis of the result of detection by the reference image detection unit 226C. In this embodiment, the image captured by the digital camera 210 is displayed on the controller display 222 in real time as a live preview image and in a case where the reference image is detected, information indicating the region of the reference image is superimposed on the live preview image and displayed to thereby provide a guide for photographing the photographed portion. More specifically, a frame (detection frame) that indicates the region of the reference image is superimposed on the live preview image and displayed to thereby provide a guide for photographing the photographed portion.
Processes performed by the photographed-portion setting apparatus 10 are the same as those performed in the photographing system of the above-described first embodiment. Therefore, a photographing method using the unattended aircraft 212 is described below.
First, location information and a reference image are acquired. The controller 214 wirelessly communicates with the photographed-portion setting apparatus 10 via the wireless communication unit 224 and acquires location information and a reference image from the photographed-portion setting apparatus 10.
After the location information and the reference image have been acquired, first, the location information is displayed on the controller display 222. As described above, the location information is a marking image obtained by marking the photographed portion on an image of the photographed subject, and therefore, the marking image is displayed on the controller display 222. The user (photographing person) views the marking image displayed on the controller display 222 to grasp where the specified photographed portion is.
The user (photographing person) operates the controller 214 on the basis of the grasped information to move the unattended aircraft 212 to a position from which the specified photographed portion can be photographed. During flight, the unattended aircraft 212 transmits an image captured by the digital camera 210 to the controller 214 in real time. The controller 214 displays the image transmitted from the unattended aircraft 212 on the controller display 222 in real time. That is, the image is displayed as a live preview image.
The controller 214 detects in real time, the reference image from the image transmitted from the unattended aircraft 212. In a case where the reference image is detected, a frame (detection frame) that indicates the region of the reference image is superimposed on the live preview image and displayed. At this time, in a case where the reference image fits in the angle of view, a blue frame is displayed. On the other hand, in a case where the reference image does not fit in the angle of view, that is, in a case where only part of the reference image is detected, a red frame (detection frame) is displayed. The user (photographing person) views this display on the controller display 222 and can confirm the position of the specified photographed portion. In a case where the frame does not fit in the angle of view, the user (photographing person) adjusts the position and/or the photographing direction on the basis of the display on the controller display 222. When the frame fits in the angle of view, the user (photographing person) takes in an image for recording and terminates photographing.
Accordingly, also in the photographing system of this embodiment, a guide is appropriately provided for photographing, and a specified photographed portion can be easily photographed.
The photographing system of this embodiment is configured as a system that automatically photographs a specified photographed portion using an unattended aircraft. The photographing system of this embodiment is the same as the photographing system of the above-described second embodiment except for automatic photographing. Accordingly, only a configuration for implementing automatic photographing is described below.
An automatic photographing function is implemented by the controller 214. That is, when the controller microcomputer 226 of the controller 214 executes a specific program, the controller 214 functions as an automatic photographing control device.
As illustrated in
The photographing information acquisition unit 226A acquires location information and a reference image from the photographed-portion setting apparatus 10. The photographing information acquisition unit 226A communicates with the photographed-portion setting apparatus 10 via the wireless communication unit 224 and acquires location information and a reference image from the photographed-portion setting apparatus 10. In this embodiment, the photographing information acquisition unit 226A acquires GPS information as location information of a photographed portion. The acquired location information is fed to the first guide information generation unit 226E, and the acquired reference image is fed to the reference image detection unit 226C.
The first guide information generation unit 226E generates information for guiding the unattended aircraft 212 to a position from which the photographed portion can be photographed, on the basis of the location information. Specifically, the first guide information generation unit 226E generates a flight route up to a position from which the photographed portion can be photographed as guide information. The first guide information generation unit 226E is an example of a first guide unit.
The movement control unit 226F controls movement (flight) of the unattended aircraft 212 in accordance with the guide information generated by the first guide information generation unit 226E. Specifically, the movement control unit 226F generates a signal for controlling flight of the unattended aircraft 212 in accordance with the guide information (information about the flight route) generated by the first guide information generation unit 226E to move the unattended aircraft 212 to a position from which the photographed portion can be photographed.
The reference image detection unit 226C detects in real time, a region that matches the reference image from an image captured by the digital camera 210.
The second guide information generation unit 226G generates information for guiding the unattended aircraft 212 to a position from which the photographed portion is photographed, on the basis of the result of detection by the reference image detection unit 226C. Specifically, the second guide information generation unit 226G calculates the movement direction and the amount of movement of the unattended aircraft 212 necessary for the reference image to fit in the angle of view on the basis of the result of detection of the reference image and positional information of the current position of the unattended aircraft 212 and generates guide information. The second guide information generation unit 226G is an example of a second guide unit.
The automatic adjustment unit 226H automatically adjusts the position of the unattended aircraft 212 in accordance with the guide information generated by the second guide information generation unit 226G. Specifically, the automatic adjustment unit 226H generates a signal for controlling flight of the unattended aircraft 212 in accordance with the guide information generated by the second guide information generation unit 226G and moves the unattended aircraft 212 to a position at which the photographed portion fits in the angle of view.
Processes performed by the photographed-portion setting apparatus 10 are substantially the same as those in the photographing system of the above-described first embodiment. Therefore, a photographing method using the unattended aircraft 212 is described below.
First, location information and a reference image are acquired. The controller 214 wirelessly communicates with the photographed-portion setting apparatus 10 via the wireless communication unit 224 and acquires location information and a reference image from the photographed-portion setting apparatus 10. Note that as described above, in the photographing system of this embodiment, GPS information is acquired as location information. Therefore, the photographed-portion setting apparatus 10 generates GPS information as location information and transmits the GPS information to the digital camera 210.
After the location information and the reference image have been acquired, first, a flight route up to a position from which the photographed portion can be photographed is set. The flight route is set on the basis of the location information.
After the flight route has been set, the unattended aircraft 212 flies in accordance with the set flight route and moves to the position from which the photographed portion can be photographed. During flight, the unattended aircraft 212 transmits an image captured by the digital camera 210 to the controller 214 in real time. The controller 214 displays the image transmitted from the unattended aircraft 212 on the controller display 222 in real time. That is, the image is displayed as a live preview image.
The controller 214 detects in real time, the reference image from the image transmitted from the unattended aircraft 212. In a case where the reference image is detected, a frame (detection frame) that indicates the region of the reference image is superimposed on the live preview image and displayed. Simultaneously, the controller 214 automatically adjusts the position of the unattended aircraft 212 and/or the photographing direction on the basis of the result of detection of the reference image such that the reference image fits in the angle of view. For example, the controller 214 automatically adjusts the position of the unattended aircraft 212 and/or the photographing direction such that the center of the reference image is positioned at the center of the angle of view and the outer edge of the reference image fits in the angle of view. When the reference image fits in the angle of view, the controller 214 starts actual photographing. After actual photographing, the unattended aircraft 212 returns to the flight start point.
Accordingly, with the photographing system of this embodiment, when only a photographed portion is specified, the specified photographed portion can be automatically photographed.
In this embodiment, a configuration is employed in which movement to a position from which photographing can be performed and adjustments of the photographing position and the photographing direction after movement are all performed automatically; however, a configuration can be employed in which only one of the movement or the adjustments is automatically performed. For example, a configuration can be employed in which movement to a position from which photographing can be performed is automatically performed and adjustments of the photographing position and the photographing direction after movement are manually performed. Alternatively, a configuration can be employed in which movement to a position from which photographing can be performed is manually performed and adjustments of the photographing position and the photographing direction after movement are automatically performed.
In the above-described embodiment, to check an image that is being captured, a live preview image is displayed on the controller display 222; however, a live preview image need not be displayed. Therefore, in the photographing system of this embodiment, the controller 214 need not include a display.
In the above-described embodiments, a configuration is employed in which the photographed-portion setting apparatus acquires an image obtained by performing panoramic composition as an image of a photographed subject; however, an image of a photographed subject is not limited to this. As an image of a photographed subject, an image thereof captured as a single image may be acquired.
In the above-described embodiments, the photographed-portion setting apparatus is formed of a tablet computer; however, the configuration of the photographed-portion setting apparatus is not limited to this. The photographed-portion setting apparatus may be formed of, for example, a smartphone or a personal computer.
Further, in the above-described embodiments, a configuration is employed in which an image of a photographed subject is acquired from an external device. In a case where an image of a photographed subject is an image obtained by performing panoramic composition, the photographed-portion setting apparatus may include a panoramic composition function. In this case, the photographed-portion setting apparatus acquires a group of images to be subjected to panoramic composition from an external device and internally performs panoramic composition for the acquired group of images. That is, a panoramic composition unit included in the photographed-portion setting apparatus performs a composition process to internally generate an image (panoramic composite image) of the photographed subject.
In the above-described embodiments, an example case where an unattended aircraft is used as an unattended vehicle to perform photographing has been described; however, an unattended vehicle that can be used in the present invention is not limited to this. The present invention is also applicable to a case where photographing is performed by using, for example, an unattended land vehicle that travels on land without human attendance, an unattended boat that sails on the water without human attendance, an unattended submarine that travels under the water without human attendance, or an unattended spacecraft that flies in space without human attendance. Further, an unattended vehicle of the present invention includes the concept of a robot.
Hardware for implementing the photographed-portion setting apparatus can be formed of various processors. The various processors include a CPU (central processing unit), which is a general-purpose processor executing a program to function as various processing units, a programmable logic device (PLD), such as an FPGA (field-programmable gate array), which is a processor having a circuit configuration that is changeable after manufacture, and a dedicated electric circuit, such as an ASIC (application-specific integrated circuit), which is a processor having a circuit configuration specifically designed to perform specific processing. One processing unit that constitutes the photographed-portion setting apparatus may be configured as one of the various processors described above or two or more processors of the same type or different types. For example, one processing unit may be configured as a plurality of FPGAs or a combination of a CPU and an FPGA. Further, a plurality of processing units may be configured as one processor. As the first example of configuring a plurality of processing units as one processor, a form is possible where one processor is formed of a combination of one or more CPUs and software, and the processor functions as the plurality of processing units, a representative example of which is a computer, such as a client or a server. As the second example thereof, a form is possible where a processor is used in which the functions of the entire system including the plurality of processing units are implemented as one IC (integrated circuit) chip, a representative example of which is a system on chip (SoC). As described above, regarding the hardware configuration, the various processing units are configured by using one or more of the various processors described above. Further, the hardware configuration of the various processors is more specifically an electric circuit (circuitry) in which circuit elements, such as semiconductor elements, are combined.
Hardware for implementing the functions of the camera microcomputer in the photographing apparatus can be similarly formed of various processors.
1 bridge
2 main girder
3 cross girder
4 sway bracing
5 lateral bracing
6 floor slab
7 panel
10 photographed-portion setting apparatus
11 CPU
15 display
16 touch panel
17 GPS receiving unit
18 camera unit
19 microphone unit
20 speaker unit
21 communication unit
21A antenna
22 short-range wireless communication unit
22A antenna
23 sensor unit
24 media drive
25 memory card
30A photographed-subject image acquisition unit
30B display control unit
30C photographed-portion setting unit
30D location information generation unit
30E reference image generation unit
30F photographing information transmission processing unit
30G re-photographed portion detection unit
100 digital camera
110 imaging lens
110A lens driving unit
112 image sensor
112A sensor driving unit
114 analog signal processing unit
116 digital signal processing unit
118 media drive
120 memory card
122 back-side monitor
122A monitor driver
124 short-range wireless communication unit
124A antenna
126 communication unit
126A antenna
128 flash
128A flash light emission control unit
130 microphone
132 speaker
134 audio signal processing unit
136 camera operation unit
140 camera microcomputer
140A photographing information acquisition unit
140B first guide unit
140C reference image detection unit
140D second guide unit
200 photographing apparatus
210 digital camera
212 unattended aircraft
214 controller
216 airframe
218 propeller
220 operation unit
222 controller display
222A display driver
224 wireless communication unit
224A antenna
226 controller microcomputer
226A photographing information acquisition unit
226B first guide unit
226C reference image detection unit
226D second guide unit
226E first guide information generation unit
226F movement control unit
226G second guide information generation unit
226H automatic adjustment unit
C arrow pointing set photographed portion
FB detection frame
FR detection frame
IM marking image
P portion omitted from photographing
PI panoramic composite image
RF reference image
W frame indicating specified region
S1 to S6 procedure of process performed by photographed-portion setting apparatus
S11 to S23 procedure of process performed by digital camera
Number | Date | Country | Kind |
---|---|---|---|
2019-020702 | Feb 2019 | JP | national |
The present application is a Continuation of PCT International Application No. PCT/JP2020/002949 filed on Jan. 28, 2020 claiming priority under 35 U.S.C § 119(a) to Japanese Patent Application No. 2019-020702 filed on Feb. 7, 2019. Each of the above applications is hereby expressly incorporated by reference, in its entirety, into the present application.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/JP2020/002949 | Jan 2020 | US |
Child | 17383520 | US |