This patent application is based on and claims priority pursuant to 35 U.S.C. § 119(a) to Japanese Patent Application No. 2021-123525, filed on Jul. 28, 2021, in the Japan Patent Office, the entire disclosure of which is hereby incorporated by reference herein.
Embodiments of the present disclosure relate to an image-capturing device and an image-capturing system.
As an image-capturing system that captures images of views surrounding the center of the system (or an image-capturing device), a hand-held full-spherical image capturing device incorporating multiple wide-angle lenses such as fish-eye lenses, super wide-angle lenses to capture a full-spherical image is known.
An image-capturing device includes: an imager extending along a first axis, the imager configured to capture an image; and a grip coupled to the imager, extending along a second axis tilted to the first axis, and having an elongated shape to have a periphery thereof gripped with hand of an operator of the image-capturing device.
An image-capturing system includes an image-capturing device including: an imager extending along a first axis, the imager including: a phototransmitter configured to emit light to an object; and a photosensor configured to receive light reflected from the object so as to capture an image; and a grip coupled to the imager, extending along a second axis tilted to the first axis, and having an elongated shape to have a periphery thereof gripped with hand of an operator of the image-capturing device; and an image processing device including processing circuitry configured to: receive data of the image from the image-capturing device; and output information on a distance to the object based on time of light emission of the phototransmitter and time of light reception of the photosensor using the received data.
A more complete appreciation of the disclosure and many of the attendant advantages and features thereof can be readily obtained and understood from the following detailed description with reference to the accompanying drawings, wherein:
The accompanying drawings are intended to depict embodiments of the present invention and should not be interpreted to limit the scope thereof. The accompanying drawings are not to be considered as drawn to scale unless explicitly noted. Also, identical or similar reference numerals designate identical or similar components throughout the several views.
In describing embodiments illustrated in the drawings, specific terminology is employed for the sake of clarity. However, the disclosure of this specification is not intended to be limited to the specific terminology so selected and it is to be understood that each specific element includes all technical equivalents that have a similar function, operate in a similar manner, and achieve a similar result.
Referring now to the drawings, embodiments of the present disclosure are described below. As used herein, the singular forms “a,” “an,” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise.
Embodiments of the present invention achieves facilitation of capturing of an omnidirectional image.
Hereinafter, embodiments will be described with reference to the accompanying drawings. In order to facilitate understanding of the description, the same components in the drawings are denoted by the same reference numerals as much as possible, and redundant description is omitted.
The image-capturing device 1 includes an imager 10 and a grip 20. The imager 10 is formed to extend along a predetermined center axis C1 (a first axis). The imager 10 captures an omnidirectional image whose center is the center axis C1. The “omnidirectional image whose center is the center axis” refers to, for example, a full-spherical image in which the diameter of a sphere representing the captured range of the full-spherical image (the image-capturing area R of the image-capturing device 1) is the center axis C1 (see
The grip 20 is a portion that is gripped by the operator during the use of the image-capturing device 1. The grip 20 is coupled to the imager 10. The grip 20 has an elongated shape to allow the operator to grip its periphery with the hand.
The imager 10 is provided with various elements to acquire information used to generate an omnidirectional image. Among various types of methods for generating an omnidirectional image, the present embodiment uses a Time Of Flight (TOF) that involves a distance image composed of a group of points representing distances between the image-capturing device 1 and surrounding objects.
The imager 10 includes multiple phototransmitters 12 and multiple photosensors 13 to obtain multiple distance images, which are combined to generate one distance image.
The phototransmitters 12 include light source units 12A and 12a as semiconductor lasers, and optical systems each including an optical element such as a fisheye lens. Such phototransmitters 12 each diverge a light beam from a corresponding one of the light source units 12A and 12a to emit a light beam diverging with a wide angle.
The photosensors 13 include TOF distance image sensors 13A and 13a (hereinafter also referred to as TOF sensors 13A and 13a) and image-forming optical systems each including an optical element such as a fish-eye lens. Light beams reflected from an object to be captured surrounding the image-capturing device 1 back to the image-capturing device 1 after emitted from the multiple phototransmitters 12 are condensed on the light-receiving areas of the TOF distance image sensors 13A and 13a.
In the present embodiment, the imager 10 further includes multiple red-green-blue (RGB) imagers 11 to obtain multiple RGB images, which are combined together in the same manner as the distance images to generate one RGB image.
The RGB imagers 11 includes an optical element such as a fish-eye lens, and solid-state image sensors 11A and 11a such as complementary metal oxide semiconductor (CMOS).
Such an image-capturing device 1 according to the present embodiment creates a digital twin with distance information and RGB information in a space surrounding the image-capturing device 1 by combining the distance image and RGB image obtained in a wide range around the image-capturing device 1.
The RGB imagers 11, the phototransmitters 12, and the photosensors 13 of the imager 10 are typically arranged to face radially outward from the center axis C1.
The multiple RGB imagers 11 is typically arranged substantially uniformly in the circumferential direction around the center axis C1. The multiple phototransmitters 12 and the multiple photosensors 13 are arranged in the same manner as the RGB imagers 11.
Any of the RGB imagers 11, the phototransmitters 12, and the photosensors 13 may be arranged along the extending direction of the center axis C1 (e.g., at an upper position in the vertical direction in
In response to receiving an ON signal (a shooting-start signal) from the shooting switch 15, the control unit 141 outputs synchronization signals to the image sensors 11a and 11A, light source units 12a and 12A, and the TOF sensors 13a and 13A to control the entire operation of the processing circuit 14. The control unit 141 outputs a signal instructing emission of an ultrashort pulse to each of the light source units 12a and 12A, and outputs a signal instructing generation of a TOF image to each of the TOF sensors 13a and 13A at the same timing. Further, the control unit 141 outputs a signal instructing each of the image sensors 11a and 11A to capture an image. The image sensors 11a and 11A each capture an image during the emission of the light source units 12a and 12A or in a period immediately before or after that emission of the light source units 12a and 12A.
The RGB image data acquisition unit 142 acquires RGB image data captured by the image sensors 11a and 11A in accordance with the instructions to capture an image from the control unit 14, and outputs full-spherical RGB image data. The monochrome processing unit 143 performs processing for converting data type in preparation for a process of matching that data with the TOF image data acquired by the TOF sensors 13a and 13A. In this example, the monochrome processing unit 143 converts the full-spherical RGB image data into a full-spherical monochrome image.
The TOF image data acquisition unit 144 acquires the TOF image data generated by the TOF sensors 13a and 13A in accordance with the instruction to generate TOF image data, output from the control unit 141, and outputs TOF image data of the full sphere.
Notably, in the present embodiment, the TOF image data acquisition unit 144 is also referred to as a range sensor (ranging unit) that outputs information on the distance to the object to be captured, the information obtained based on the time at which each phototransmitter emits light and the time at which each photosensor receives light.
The resolution enhancement unit 145 enhances (increases) the resolution of the TOF image data of the full sphere as a monochrome image. Specifically, the resolution enhancement unit 145 replaces the value of the distance associated with each pixel of the TOF image data of the full sphere with the value (gray scale value) of the monochrome image of the full sphere, thus to use the values of the monochrome image of the full sphere. Further, the resolution enhancement unit 145 increases the resolution of the monochrome image of the full sphere up to the resolution of the full-spherical RGB image data acquired from the image sensors 11a and 11A. Such enhancement of the resolution (the conversion to high resolution) is performed by performing, for example, regular up-conversion processing. Another conversion of the resolution may be super-resolution processing involving acquiring multiple frames of the TOF image data of the full sphere, which have been continuously generated, for example; and adding the distance for each point between adjacent frames of the multiple frames.
The matching processing unit 146 extracts the amount of features on textured portions of the full-spherical monochrome image having the resolutions increased from those of the full-spherical TOF image data and the monochrome image of the full-spherical RGB image data, and performs matching processing with the extracted amount of features. For example, the matching processing unit 146 extracts an edge from each monochrome image and performs matching processing based on pieces of information on the extracted edges. Alternatively, the matching processing may involve performing featurization of changes in texture (e.g., scale invariant feature transform (SIFT)). Herein, the matching processing refers to searching for pixels corresponding to each other.
Specific examples of the matching processing include block matching. The block matching involves calculating the similarities between pixel values of a block of M×M (M is a positive integer) pixels around a pixel to be referred in one image and pixel values of a block of M×M pixels around a pixel serving as the centerpiece of search in another image, and determines, as the pixels corresponding to each other, pixels the similarity between which is the highest.
There are various methods for calculating such similarities. For example, an equation indicating a normalized correlation coefficient (NCC) may be used. With increasing NCC, the similarity is higher. The NCC becomes 1 when the pixel values of the blocks exactly match with each other.
Further, since the distance data of the textureless region is obtained from the TOF image data of the full sphere, the matching process may be weighted according to the region. For example, in the calculation of expression indicating the NCC, calculation for weighting a portion (i.e., the textureless region) of an image other than an edge may be performed.
Alternatively, selective correlation coefficient (SCC) may be used instead of the NCC.
The reprojection processing unit 147 reprojects the TOF image data of the full sphere indicating the distance to each position within the range to be measured to two-dimensional coordinates (a screen coordinate system) of the imager 11. The reprojection of the reprojection processing unit 147 refers to obtaining coordinates of the three-dimensional points calculated by the TOF sensors 13a and 13A in the images of the image sensors 11a and 11A. The TOF image data of the full sphere indicates the positions of three-dimensional points in a coordinate system with a distance-information acquisition unit (i.e., the photosensor 13), particularly a wide-angle lens at the center. This means that the three-dimensional points indicated by the TOF image data of the full sphere is reprojected onto the coordinate system having the imager 11 (mainly the fish-eye lens) as the center.
For example, the reprojection processing unit 147 performs a process of translating the coordinates of the three-dimensional point of the full-spherical TOF image data to the coordinates of the three-dimensional point for which the imager 11 is at the center and then transforming the coordinates of the three-dimensional point of the full-spherical TOF image data into a two-dimensional coordinate system (the screen coordinate system) indicated by the full-spherical RGB image data. Thus, the coordinates of the three-dimensional points of the TOF image data of the full sphere and the coordinates of the two-dimensional image information of the full sphere captured by the imager 11 are associated with each other. The reprojection processing unit 147 associates the coordinates of the three-dimensional points of the TOF image data of the full sphere with the coordinates of the two-dimensional image information of the full sphere captured by the imager 11.
The disparity calculation unit 149 calculates the disparity at each position from the differences in distance between the pixels corresponding to each other, which have been obtained from the matching process. The disparity matching processing allows a reduction in processing time and acquisition of distance information in more detail and with higher resolution by searching the pixels surrounding the reprojected coordinates (position) using the coordinates reprojected and transformed by the reprojection processing unit 147.
The semantic segmentation unit 148 performs semantic segmentation processing to generate segmentation data. In some embodiments, the disparity matching processing uses the segmentation data generated by the semantic segmentation unit 148. The use of the segmentation data enables acquisition of further more detailed and higher-resolution distance information.
In some other embodiments, the disparity matching processing may be performed only on an edge portion and a portion having a large amount of features, and a propagation process may be performed on the other portion by using, for example, an RGB image feature of the full sphere or a stochastic method using the TOF image data of the full sphere.
The semantic segmentation unit 148 uses deep learning to assign a segmentation label indicating an object to an input image within a measurement range (the range to be measured). Assigning a segmentation label to an input image enables each pixel of the TOF image data of the full sphere to be bound to any of multiple distance regions divided among the distance values, thus allowing a higher reliability of the calculation.
The three-dimensional reconstruction processing unit 150 acquires the full-spherical RGB image data from the RGB image data acquisition unit 142, reconstructs three-dimensional data of the full sphere based on the distance information output from the disparity calculation unit 149, and outputs a high-density three-dimensional point group of the full sphere in which color information is added to each three-dimensional point. The three-dimensional reconstruction processing unit 150 is an example of a three-dimensional information determiner that determines three-dimensional information.
The determination unit 160 acquires the full-spherical RGB image data from the RGB image data acquisition unit 142, acquires from the reprojection processing unit 147 the TOF image data of the full sphere converted into the two-dimensional coordinate system indicated by the full-spherical RGB image data, and determines the presence or absence of reflection of a particular object in the captured image based on the acquired data, outputting the determination result (determination as to the presence or absence of a particular object in the captured image) to the display control unit 170.
The display control unit 170 acquires the full-spherical RGB image data from the RGB image data acquisition unit 142 and causes the display unit 30 to display two-dimensional image information based on the acquired full-spherical RGB image data. In addition, the display control unit 170 causes the display unit 30 to display a display image including information indicating the determination result acquired from the determination unit 160 and two-dimensional image information.
The display control unit 170 is an example of an output unit that outputs two-dimensional image information captured by the imager 11 separately from three-dimensional information. The display unit 30 is an example of a destination device to which the two-dimensional image information is output.
In some embodiments, the display control unit 170 acquires three-dimensional data of the full sphere from the three-dimensional reconstruction processing unit 150 and causes the display unit 30 to display the three-dimensional information. Specifically, the display control unit 170 selects data to be displayed by the display unit 30, between the three-dimensional information and the two-dimensional image information according to a prescribed condition. The prescribed condition is set by the user in advance. The display control unit 170 outputs two-dimensional image information separately from three-dimensional information.
The transmission-reception unit 180 communicates with the external device 300 by wire or wireless. Specifically, the transmission-reception unit 180 transmits (outputs), via the network 400, the three-dimensional data of the full sphere output from the three-dimensional reconstruction processing unit 150 and the two-dimensional image information of the full sphere output from the RGB image data acquisition unit 142 to the external device 300 that performs three-dimensional reconstruction processing.
In the present embodiment, the two-dimensional image information captured by the imager 11 refers to original two-dimensional image information for creating two-dimensional image data for display or refers to the two-dimensional image data for display. In one example, the image-capturing device 1 creates two-dimensional image data from the original two-dimensional image information. In another example, the image-capturing device 1 outputs the original two-dimensional image information to the external device 300, and the external device 300 creates two-dimensional image data for display from the received original two-dimensional image information.
The transmission-reception unit 180 is an example of an output unit that outputs three-dimensional information, and the external device 300 is an example of a destination device to which three-dimensional information is output.
In some examples, the transmission-reception unit 180 transmits only the three-dimensional data of the full sphere without transmitting the two-dimensional image information of the full sphere. Further, the transmission-reception unit 180 may be configured by a portable storage medium such as a secure digital (SD) card or an interface circuit to communicate with a personal computer (PC).
In the present embodiment, the image-capturing device 1 includes: the TOF distance image acquisition means (unit) including the phototransmitters 12 and the photosensors 13; and RGB image acquisition means (unit) including the RGB imagers 11. In some embodiments, the image-capturing device 1 includes one of the TOF distance image acquisition means and the RGB image acquisition means. Further, the distance image acquisition means (unit) is not limited to the TOF distance image acquisition means, and may be, for example, a stereo camera. The RGB image acquisition means is not limited to the CMOS, and may be another type of image sensor. Further, the RGB image acquisition means may acquire a gray scale image instead of the RGB image.
As illustrated in
In the present embodiment, an image-capturing area R of the imager 10 refers to a portion of the full-spherical area, whose center is the imager 10, excluding the non-image-capturing area N. In other words, the non-image-capturing area N is referred to also as a blind spot of a full-spherical image. Hereinafter, the non-image-capturing area N is sometimes referred to as a blind spot region (non-image-capturing area N).
As illustrated in
In particular, the image-capturing device 1 according to the present according to the present embodiment features the configuration of the grip 20 described above. As described above, the grip 20 having an elongated shape to allow the operator to grip its periphery with the hand is coupled to the imager 10.
Further, the grip 20 is coupled to a portion of the imager 10, the portion being proximate to the non-image-capturing area N (i.e., the blind spot region) along the center axis C1. In other words, the portion of the imager 10 along the center axis C1 is the lower side of the imager 10 when the image-capturing device 1 is in use. This arrangement allows the operator to more likely be within the non-image-capturing area N of the imager 10 as illustrated in
Further, in the present embodiment, as illustrated in
The tilt of the grip 20 will be further described. As illustrated in
In this case, the direction in which the longitudinal direction (the center axis C2) of the grip 20 is tilted relative to the center axis C1 of the imager 10 is a direction in which the ulnar flexion angle A of the operator's wrist (hand grasping the grip 20) (see
Herein, an angle between the direction (the upward direction in
Such a configuration of the image-capturing device 1 in which the longitudinal direction (i.e., the center axis C2) of the grip 20 is tilted relative to the center axis C1 of the imager 10 facilitates adjustment of the center axis C1 of the imager 10 in the vertical direction as well as capturing of an omnidirectional image while the operator holds the image-capturing device 1 by hand to capture an image.
This configuration further facilitates adjustment of the non-image-capturing area N (i.e., the blind spot) of the imager 10 to be within a desired range (e.g., the lower side of the imager 10 in the vertical direction), and thus enables a smaller blind spot in a captured omnidirectional image and prevents the operator from being unintentionally within the image-capturing area R to be reflected in a captured image.
In some cases, the ulnar flexion angle (or the radial flexion (bending) angle) of the wrist is to be increased while the operator holds the image-capturing device 1 with the center axis C1 of the image-capturing area parallel to the vertical direction, which is intentionally oriented by the operator during use. This causes the operator difficulty in moving the wrist sufficiently or comfortably within the movable range of the wrist. This might cause the operator to apply unnatural force to the image-capturing device 1 and produce camera shake to more likely cause image blurring. However, the present embodiment enables a posture of the ulnar flexion (or radial flexion) of the wrist, which allows the operator to move the wrist without difficulty (or comfortably) within the movable range of the wrist when the operator adjusts the center axis C1 of the imager 10 in the vertical direction, because of the tilting of the longitudinal direction of the grip 20 in the direction in which the ulnar flexion angle A or the radial flexion angle of the wrist (the operator's hand gripping the grip 20) is reduced. This prevents the occurrence of image blurring of a captured image due to camera shake.
Such a reduction in the ulnar flexion angle of the operator's wrist (hand) gripping the grip 20 reduces the load on the operator during the operation of the image-capturing device 1 and further facilitates capturing of an omnidirectional image.
As illustrated in
As illustrated in
The third finger support 23, the fourth finger support 24, and the fifth finger support 25, which are referred to as a support are provided along the extending direction (i.e., the center axis C2) of the grip 20.
The first finger support 21 is disposed so as to face the third finger support 23, the fourth finger support 24, and the fifth finger support 25, which are disposed across the central axis C2 coincident with the extending direction (the longitudinal direction) of the grip 20, from the first finger support 21. The longitudinal direction of the grip 20 is tilted relative to the center axis C1 of the imager 10, in a direction from the first finger support 21, the third finger support 23, the fourth finger support 24, to the fifth finger support 25.
Such an arrangement of the first to fifth finger supports 21 to 25 facilitates the operator's posture of gripping the grip 20 as illustrated in
The first finger support 21 and the third finger support 23 are proximate to the imager 10 in the extending direction (i.e., along the center axis C2) of the grip 20. In other words, the first finger support 21 and the third finger support 23 are closer to the imager 10 than the fourth finger support 24 and the fifth finger support 25 in the extending direction, or the longitudinal direction of the grip 20. The second finger support 22 is between the imager 10 and the third finger support 23.
Referring to
With an omnidirectional image captured by the image-capturing device 100 having a non-image-capturing area N (i.e., the blind spot region) at a lower part of the image-capturing device 100, the configuration of
Capturing an image by a simple linear columnar image-capturing system will be considered. As an example, a case where the bending angle B (see
In consideration of this, the imager 10 of the image-capturing device 1 fails to be perpendicular to the ground unless the ulnar flexion angle A of the wrist satisfies the following equation (1):
Bending Angle of Shoulder Joint B (150 degrees)−Angle E Between hand and Pillar D Gripped by Hand (10 to 20 degrees)−Ulnar Flexion Angle of Wrist A=Image-Capturing System angle (90 degrees) (1)
The image-capturing system angle on the right side of the above equation (1) is the angle between the center axis C1 of the imager 10 and the horizontal direction. In order to satisfy the condition of the above equation, the ulnar flexion angle A of the wrist is to be about 40 to 50 degrees.
There is a finding that bending the wrist with an ulnar flexion angle of about 40 to 50 degrees may be a heavy load for a human.
For this reason, the image-capturing device 100 according to the comparative example held by a person having a range of motion equal to or less than the average may have the center axis C1 of the imager 10 tilted as illustrated in
The image-capturing device 100 according to the comparative example, i.e., the mage-capturing system of a simple straight pillar likely causes any of the following issues: The operator fails to bend the wrist to a desired angle (i.e., until the image-capturing device 100 turns vertical to the ground), and the center axis C1 of the imager 10 of the image-capturing device 100 tilts; and camera shake is likely to occur because of the operator's capturing an image while forcedly bending the wrist.
In the image-capturing device 1 according to an embodiment, however, a tilt angle θ is provided between the center axis C1 of the imager 10 (i.e., the center axis C1 of the image-capturing area passing through the center of the image-capturing area R) and the extending direction (the center axis C2) of the grip 20. This configuration enables the operator to hold the image-capturing device 1 above the head with a wrist posture that allows the operator to move the wrist comfortably as appropriate within the movable range (the ulnar flexion range) of the wrist, so as to orient the imager 10 of the image-capturing device 1 vertical to the ground.
The following describes a method of setting the tilt angle θ of the grip 20 will be described with further reference to
The tilt angle θ of the grip 20 is set as follows to allows users of more ages and genders to use the image-capturing device 1: For the ulnar flexion angle A of the wrist in the image-capturing posture, an angle of 15 degrees)(°) that allows bending of the wrist comfortably without applying a force is a reference value of the ulnar flexion of the wrist. This reference value is determined in consideration of holding the grip with an allowance of an angle of 10 degrees (°) for a further movement of the wrist within the movable range, assuming that some users fail to bend the wrist to an angle of greater than 25 degrees with reference to the minimum value of 60 to 69 years old males in
Using the angles A, B, and E as defined above, the tilt angle θ is calculated to range from 125 to 175 degrees (°) as illustrated in
Through the above-described setting method, the inventors of the present invention have conceived of the tilt angle θ of the grip 20 ranging from 125 to 175 degrees, which are preferable in the present embodiment.
Notably, with the grip 20 having a tilt angle θ of greater than 175 degrees, the ulnar flexion angle of the wrist of the operator is to be increased to adjust the center axis C1 of the imager 10 in the vertical direction. This means that an ulnar flexion angle sufficient to adjust the center axis C1 of the imager 10 in the vertical direction might be difficult to obtain irrespective of the differences in the movable range of the ulnar flexion movement between age and sex groups as illustrated in
Next, the advantageous effects of the present embodiment will be described with reference to
As illustrated in
The image-capturing device 100 according to a comparative example as illustrated in
As illustrated in
By contrast, in the comparative example at which the tilt angle θ is 180 degrees, the angle F of the imager 10 of the image-capturing device 100 ranges from 95 to 145 degrees, and the angle G with respect to the vertical direction ranges from 5 to 55 degrees)(°. In other words, the imager 10 is tilted relative to the vertical direction at a tilt angle of 55 degrees at maximum. When the present embodiment is compared with the comparative example, the present embodiment allows a reduction in the tilt of the imager 10 by 30 degrees in absolute value.
The difference G between the angle F of the imager 10 and the image-capturing device 1 and the vertical direction 1 in the worst of the assumed postures is the smallest at the tilt angle θ of 150 degrees, but is constantly smaller than the case of 0=180 degrees within the range of 125 to 175 degrees described above.
Next, modification 1 will be described with reference to
As illustrated in
A case is considered in which a flat surface is provided at a lower portion of the image-capturing device 1A to form the bottom portion 26, and the image-capturing device 1A is placed on a horizontal surface such as the ground or a desk. In the image-capturing device 1 of
Such an image-capturing device 1A according to modification 1 that can stand by itself on the horizontal surface such as the ground or a desk enables capturing of images via wireless control and will be widely available in applications other than gripping operation.
In order to make the projection point I of the center of gravity H of the image-capturing device 1A fall within the range of the bottom portion 26 as in modification 1, for example, as illustrated in
Further, the virtual plane P of the bottom portion 26 is preferably orthogonal to the center axis C1 of the imager 10. In other words, the normal direction of the bottom portion 26 is preferably coincident with the center axis C1. This arrangement that allows the projection point I of the center of gravity H to be within the range of the bottom portion 26 enables the image-capturing device 1A to reliably stand by itself. Further, the image-capturing device 1A according to modification 1 in which the center axis C1 of the imager 10 is vertical to the bottom portion 26 enables minimization of the blind spot (non-image-capturing area) on the plane on which the image-capturing device 1A is placed.
As long as the image-capturing device 1A can stand by itself, the normal direction of the bottom portion 26 may not be coincide with the center axis C1 of the imager 10.
The following describes the advantageous effects of modification 1 with reference to
As illustrated in
Similarly to the case in which the tilt angle θ is 150 degrees as illustrated in
Next, modification 2 will be described with reference to
As illustrated in
This configuration enables a tilt angle θ that allows the user (operator) to most comfortably hold the grip 20 to be set according to the characteristics of the user. For example, with an increase in the tilt angle θ to reduce the difference between the center axis C1 of the imager 10 and the center axis C2 (the extending direction) of the grip 20 as illustrated in
Next, modification 3 will be described with reference to
As illustrated in
For example, as illustrated in
As illustrated in
Another embodiment will be described with reference to
The image-capturing system 200 in
The image-capturing device 1D in
The processing circuit 4 in the image-capturing device 1 illustrated in
Similarly to
Similarly to
Unlike
The display device 500 in
The transmission-reception unit 180 receives the full-spherical RGB image data and the full-spherical TOF image data transmitted from the image-capturing device 1D.
The RGB image data acquisition unit 542 acquires the full-spherical RGB image data from the transmission-reception unit 180, and the TOF image data acquisition unit 544 acquires the full-spherical TOF image data from the transmission-reception unit 180. The remaining configurations of the RGB image data acquisition unit 542 and the TOF image data acquisition unit 544 are similar to those of the RGB image data acquisition unit 142 and the TOF image data acquisition unit 144 in
The monochrome processing unit 543, the TOF image data acquisition unit 544, the resolution enhancement unit 545, the matching processing unit 546, the reprojection processing unit 547, the semantic segmentation unit 548, the disparity calculation unit 549, the three-dimensional reconstruction processing unit 550, and the determination unit 560 are configured similarly to the monochrome processing unit 143, the TOF image data acquisition unit 144, the resolution enhancement unit 145, the matching processing unit 146, the reprojection processing unit 147, the semantic segmentation unit 148, the disparity calculation unit 149, the three-dimensional reconstruction processing unit 150, and the determination unit 160 in
The display control unit 530 acquires the full-spherical RGB image data from the RGB image data acquisition unit 542 and causes the display unit 520 to display a two-dimensional image based on the acquired full-spherical RGB image data. Alternatively, the display control unit 530 acquires the full-spherical three-dimensional data from the three-dimensional reconstruction processing unit 550 and causes the display unit 520 to display three-dimensional image.
The display control unit 530 causes the display unit 520 to display a display image including information indicating the determination result acquired from the determination unit 160 and a two-dimensional image or a three-dimensional image.
As described above, the display device 500 includes the transmission-reception unit 510 (an example of a receiver) that receives an output of the imager 11 that captures an image of an object to be captured and an output of the distance-information acquisition unit (the photosensor 13) that receives light reflected from the object after emitting light to the object; the determination unit 560 that determines the presence or absence of a particular object based on the output from the distance-information acquisition unit (the photosensor 13) and the output from the imager 11, which are received by the transmission-reception unit 510; and the display control unit 530 that causes a display unit to display a display image, which differs according to the presence or absence of the particular object based on the determination result of the determination unit 560.
Examples of the particular object includes a near object, a high-reflection object, a far object, a low-reflection object, and an image-blurring region.
The display device 500 includes the display control unit 530 that causes the display unit 520 to display a display image including identification information for identifying a particular object and a three-dimensional image 3G determined by the three-dimensional reconstruction processing unit 550 based on the determination result of the determination unit 560 that determines whether the particular object is present based on both an output of the imager 11 that captures an image of the object and an output of the distance-information acquisition unit (the photosensor 13) that receives light reflected from the object after emitting light to the object.
Notably, the display device 500 is an example of an information processing apparatus (external information processing apparatus) outside the image-capturing device 1D. The display device 500 is, for example, a personal computer (PC), a server, a smartphone, or a tablet PC.
Unlike the image-capturing device 1 according to an embodiment, the image-capturing system 200 includes the TOF image data acquisition unit 544 as a range sensor (ranging unit) in the information processing apparatus outside the image-capturing device 1D. The external information processor also serves to generate a three-dimensional image. The image-capturing device 1D merely serves to capture an RGB image or a TOF image. As described above, the image-capturing device 1D of the image-capturing system 200 merely acquires image information. This simplifies the structure of the image-capturing device 1D and achieves weight reduction and miniaturization. Since the image-capturing device 1D is to be held by the operator during use, the weight reduction and miniaturization of the image-capturing device 1D increases usability.
The present embodiment has been described above with reference to specific examples. However, the present disclosure is not limited to these specific examples. These specific examples to which appropriate design modifications are added by those skilled in the art are also included in the scope of the present disclosure as long as the features of the present disclosure are included. Each element included in each specific example described above and the arrangement, condition, shape, and the like thereof are not limited to those illustrated, and can be appropriately changed. The elements included in the specific examples described above may be combined as appropriate as long as there is no technical contradiction.
The above-described embodiments are illustrative and do not limit the present invention. Thus, numerous additional modifications and variations are possible in light of the above teachings. For example, elements and/or features of different illustrative embodiments may be combined with each other and/or substituted for each other within the scope of the present invention.
Any one of the above-described operations may be performed in various other ways, for example, in an order different from the one described above.
Each of the functions of the described embodiments may be implemented by one or more processing circuits or circuitry. Processing circuitry includes a programmed processor, as a processor includes circuitry. A processing circuit also includes devices such as an application specific integrated circuit (ASIC), a digital signal processor (DSP), a field programmable gate array (FPGA), and conventional circuit components arranged to perform the recited functions.
Number | Date | Country | Kind |
---|---|---|---|
2021-123525 | Jul 2021 | JP | national |