The present disclosure relates to a technique for identifying an image in an environment requiring privacy protection, particularly in home or indoor.
For example, Patent Literature 1 discloses an image identification system where an identifier receives a computational image captured by a light-field camera or the like to identify an object included in the computational image using a trained identification model.
The computational image is difficult to be visually recognized by a person due to blurring that is intentionally created due to an influence such as superimposition of a plurality of images each having different viewpoints, or a subject image that is less likely to be focused due to non-use of a lens. Thus, the computational image is preferably used to construct an image identification system in an environment requiring privacy protection, such as home or indoor.
On the other hand, in a case where the imaging pattern of the computational image is known, the computational image may be restored to an unblurred image. Therefore, a user who is a subject has an issue that the user cannot be sure whether privacy of the captured computational image is protected. This issue is referred to as a psychological load on a user.
In Patent Literature 1, no measure is taken against this issue, and thus a psychological load on the user is desired to be reduced by implementing effective technical measures.
Patent Literature 1: WO 2019/054092 A
The present disclosure has been made to solve the above issue, and an object of the present disclosure is to provide a technique that enables a reduction of a psychological load on the user while protecting the privacy of a subject.
An image identification system according to one aspect of the present disclosure includes a first camera that includes a mask and an image sensor, the mask having a changeable mask pattern having a plurality of pinholes, and captures a computational image that is an image with blurring, an image identification unit that identifies the computational image using an identification model that uses the computational image captured by the first camera as input data and an identification result as output data, a mask identification unit that, after the mask pattern is changed, identifies the mask pattern that has been changed, and an identification model change unit that changes the identification model in accordance with the mask pattern identified by the mask identification unit.
According to the present disclosure, a psychological load on a user can be reduced while privacy of a subject is being protected.
In home, indoor, or the like, various recognition techniques are important, such as human activity recognition of a person in an environment or person recognition of a device operator. In recent years, a technique called deep learning has attracted attention for object identification. The deep learning is machine learning using a neural network having a multilayer structure, and enables achieving more accurate identification performance as compared with a conventional method by using a large amount of learning data. In such object identification, image information is particularly effective. Various methods have been proposed for greatly improving conventional object identification capability by using a camera as an input device and performing deep learning using image information as an input.
Unfortunately, disposing a camera in home or the like causes a problem in that privacy is violated when a captured image leaks to the outside due to hacking or the like. Thus, a measure is required to protect privacy of a subject even when a captured image leaks to the outside.
Computational images captured by a light-field camera or the like are difficult to be visually recognized by a person due to blurring that is intentionally created due to an influence such as superimposition of a plurality of images each having a different viewpoint, or a subject image that is less likely to be focused due to non-use of a lens. Thus, the computational image is preferably used to construct an image identification system in an environment requiring privacy protection, such as home or indoor.
The image identification system disclosed in Patent Literature 1 is configured such that a target area is imaged by a light-field camera or the like, and a computational image acquired by the imaging is input to an identifier. This configuration allows the identifier to identify an object included in the computational image using a trained identification model. When the target area is imaged by a light-field camera or the like that captures a computational image as described above, privacy of a subject can be protected even when the captured image leaks to the outside due to the computational image that is difficult to be visually recognized by a person.
As another viewpoint of privacy protection, it is also important to reduce a psychological load on a user who is to be imaged by an image recognition device. Capturing a blurred image enables appeal for protecting privacy of a subject. However, when computational imaging information is set in a region unrelated to the user, such as a factory of a manufacturer, a psychological load on the user may increase due to a suspicion that the manufacturer can restore a normal image from a blurred image.
To solve the above problems, the following technique is disclosed.
(1) An image identification system according to one aspect of the present disclosure includes a first camera that includes a mask and an image sensor, the mask having a changeable mask pattern having a plurality of pinholes, and captures a computational image that is an image with blurring, an image identification unit that identifies the computational image using an identification model that uses the computational image captured by the first camera as input data and an identification result as output data, a mask identification unit that, after the mask pattern is changed, identifies the mask pattern that has been changed, and an identification model change unit that changes the identification model in accordance with the mask pattern identified by the mask identification unit.
According to this configuration, even if the mask pattern of the mask in which the plurality of pinholes are formed is changed, the identification model is changed in accordance with the changed mask pattern, thus preventing degradation of identification accuracy in image identification using the identification model. Furthermore, since the degree of blurring of the computational image captured by the first camera can be changed by changing the mask pattern of the mask, the psychological load on the user can be reduced while privacy of a subject is being protected.
(2) The image identification system according to (1) may further include a mask change unit that changes the mask pattern of the mask.
According to this configuration, the mask pattern of the mask is mechanically changed, thereby enhancing the accuracy of changing the mask pattern.
(3) In the image identification system according to (1) or (2), the first camera may include a multi-pinhole camera, the mask may include a multi-pinhole mask in which a plurality of masks are overlaid, the plurality of masks may respectively have mask patterns different from each other, and the mask pattern of the multi-pinhole mask may be changed when at least one of the plurality of masks is removed.
According to this configuration, the mask pattern can be easily changed by changing the number of the plurality of masks constituting the multi-pinhole mask.
(4) In the image identification system according to (1) or (2), the first camera may include a multi-pinhole camera, the mask may include a multi-pinhole mask in which a plurality of masks are overlaid, the plurality of masks may respectively have mask patterns different from each other, and at least one of the plurality of pinholes formed in one of the plurality of masks may be shielded by another one of the plurality of masks.
According to this configuration, when one of the plurality of masks is removed, the number of the pinholes of the multi-pinhole mask is changed, thereby easily changing the mask pattern.
(5) In the image identification system according to (1) or (2), the first camera may include a multi-pinhole camera, the mask may include a multi-pinhole mask in which a plurality of masks are overlaid, the plurality of masks may respectively have mask patterns different from each other, and at least one of the plurality of pinholes formed in one of the plurality of masks may be present on a position identical to a position of at least one of the plurality of pinholes formed in another one of the plurality of masks.
According to this configuration, the pinholes can be reliably formed in the multi-pinhole mask in which the plurality of masks are overlaid.
(6) In the image identification system according to (5), the one pinhole included in the one mask may have a size different from a size of the other pinhole included in the other mask at the position identical to the position of the one pinhole.
According to this configuration, when the one mask or the other mask is removed, the size of the pinhole is changed, thereby easily changing the mask pattern.
(7) In the image identification system according to any one of (1) to (6), the mask identification unit may acquire mask information about the mask pattern from the computational image captured by the first camera, and identify the mask pattern of the mask based on the acquired mask information.
According to this configuration, the changed mask pattern can be identified from the computational image captured by the first camera.
(8) The image identification system according to (7) may further include a light emitting unit, in which the mask identification unit may acquire a point spread function as the mask information from an image including the light emitting unit, the image being captured by the first camera.
According to this configuration, the changed mask pattern can be identified by using the point spread function.
(9) The image identification system according to any one of (1) to (6) may further include a mask identification information acquisition unit that acquires mask identification information for identifying the mask, in which the mask identification unit may identify the mask pattern of the mask based on the acquired mask identification information.
According to this configuration, the changed mask pattern can be identified from mask ID information provided in the mask.
(10) In the image identification system according to any one of (1) to (6), the mask may include a first mask and a second mask overlaid on the first mask, the system may further include a marker position specifying unit that specifies positions of a plurality of markers formed on each of the first mask and the second mask to detect a rotation angle of the first mask with respect to the second mask, and the mask identification unit may detect the rotation angle of the first mask with respect to the second mask based on the specified positions of the plurality of markers and identify the mask pattern of the mask based on the detected rotation angle.
According to this configuration, the changed mask pattern can be identified from the rotation angle of the first mask with respect to the second mask, the rotation angle being detected based on the positions of the markers provided on the mask.
(11) The image identification system according to any one of (1) to (10) may further include a storage unit that stores a plurality of mask patterns and a plurality of identification models in association with each other, in which the identification model change unit may specify an identification model associated with the mask pattern identified by the mask identification unit from among the plurality of identification models stored in the storage unit, and change the current identification model to the specified identification model.
According to this configuration, since the plurality of mask patterns and the plurality of identification models are stored in advance in the storage unit in association with each other, the identification model associated with the changed mask pattern can be easily specified, and the identification model can be easily changed.
(12) The image identification system according to any one of (1) to (10) may further include a learning unit that acquires a first learning image captured by a second camera that captures an image without blurring or an image with blurring less than blurring by the first camera and a correct answer label given to the first learning image, generates a second learning image with blurring based on the mask pattern identified by the mask identification unit and the first learning image, and performs machine learning using the second learning image and the correct answer label to create an identification model for identifying the computational image captured by the first camera, in which the identification model change unit may change the current identification model to the identification model created by the learning unit.
According to this configuration, since a new identification model corresponding to the changed mask pattern can be created, it is possible to prevent degradation of identification accuracy in image identification using the identification model.
Further, the present disclosure can be implemented not only as the image identification system having the characteristic configuration as described above, but also as an image identification method or the like for executing characteristic processing corresponding to the characteristic configuration of the image identification system. In addition, the characteristic processing included in such an image identification method can be implemented as a computer program to be executed by a computer. Therefore, even other aspects below can achieve an effect as in the above image identification system.
(13) An image identification method according to another aspect of the present disclosure is an image identification method in an image identification system, the method including acquiring a computational image that is an image with blurring captured by a first camera that includes a mask and an image sensor, the mask having a changeable mask pattern having a plurality of pinholes, identifying the computational image using an identification model that uses the computational image captured by the first camera as input data and an identification result as output data, after the mask pattern is changed, identifying the mask pattern that has been changed, and changing the identification model in accordance with the identified mask pattern.
(14) An image identification program according to another aspect of the present disclosure is a program that causes a computer to perform operations including acquiring a computational image that is an image with blurring captured by a first camera that includes a mask and an image sensor, the mask having a changeable mask pattern having a plurality of pinholes, identifying the computational image using an identification model that uses the computational image captured by the first camera as input data and an identification result as output data, after the mask pattern is changed, identifying the mask pattern that has been changed, and changing the identification model in accordance with the identified mask pattern.
(15) A computer-readable non-temporary recording medium according to another aspect of the present disclosure includes an image identification program recorded therein, the image identification program causing a computer to perform operations including acquiring a computational image that is an image with blurring captured by a first camera that includes a mask and an image sensor, the mask having a changeable mask pattern having a plurality of pinholes, identifying the computational image using an identification model that uses the computational image captured by the first camera as input data and an identification result as output data, after the mask pattern is changed, identifying the mask pattern that has been changed, and changing the identification model in accordance with the identified mask pattern.
Hereinafter, embodiments of the present disclosure will be described with reference to the drawings. Note that elements denoted by the same reference numerals in different drawings represent the same or corresponding elements. Further, each of the embodiments to be described below shows one specific example of the present disclosure. Numerical values, shapes, constituents, steps, order of steps, and the like shown in the embodiments below are merely one example, and are not intended to limit the present disclosure. Furthermore, a constituent that is not described in an independent claim representing the highest concept among constituents in the embodiments below is described as an arbitrary constituent. Furthermore, in all the embodiments, respective contents can be combined.
The image identification system 1 of the first embodiment includes a first camera 101, a storage unit 102 that stores an identification model, an image identification unit 103, a mask identification unit 105, an identification model change unit 106, an output unit 107, and a light emitting unit 108. The mask identification unit 105, the identification model change unit 106, and the image identification unit 103 include a processor such as a central processing unit (CPU) and a memory such as a semiconductor memory. The storage unit 102 is a hard disk drive (HDD), a solid state drive (SSD), a semiconductor memory, or the like. The output unit 107 is a display device, a speaker, or the like.
The image identification system 1 may include, for example, an image identification device including the storage unit 102, the image identification unit 103, the mask identification unit 105, and the identification model change unit 106, and may be configured so that the first camera 101, the output unit 107, and the light emitting unit 108 are connected to the image identification device. In addition, the image identification device may be a server. Further, the image identification system 1 may include, for example, an image identification device including the first camera 101, the storage unit 102, the image identification unit 103, the mask identification unit 105, the identification model change unit 106, and the output unit 107, and may be configured so that the light emitting unit 108 is connected to the image identification device.
Unlike a normal camera that captures a normal image without blurring, the first camera 101 captures a computational image that is an image with blurring (step S101). Although a subject in the computational image cannot be recognized by a person who views the computational image itself due to intentionally created blurring, an image can be generated from the computational image by performing image processing on the captured computational image, the image being able to be recognized by the person or identified by the image identification unit 103.
The first camera 101 includes a mask having a changeable mask pattern in which a plurality of pinholes are formed, and an image sensor. Further, the first camera 101 is either a coded aperture camera including a mask having a mask pattern having a transmittance different for each region, or a multi-pinhole camera in which a mask having a mask pattern having a plurality of pinholes is disposed on a light receiving surface of the image sensor. The multi-pinhole camera is disclosed in conventional literature (e.g. M. Salman Asif, Ali Ayremlou, Ashok Veeraraghavan, Richard Baraniuk, and Aswin Sankaranarayanan, “FlatCam: Replacing Lenses with Masks and Computation”, International Conference on Computer Vision Workshop (ICCVW), 2015, pp. 663-666). The mask pattern of the mask is configured to be changeable.
The multi-pinhole camera 301 is one example of the first camera 101. The multi-pinhole camera 301 illustrated in
The pinhole image of the subject differs depending on a position and a size of each pinhole 301aa, so that the image sensor 301b acquires a superimposed image (multiple image) in a state where a plurality of pinhole images are superimposed while being slightly shifted. The positions of the plurality of pinholes 301aa affect the position of the subject projected on the image sensor 301b, and the sizes of the pinholes 301aa affect the degree of blurring of the pinhole image.
Using the multi-pinhole mask 301a enables acquiring a plurality of pinhole images each having a different position and a different degree of blurring while superimposing the images. That is, the multi-pinhole camera 301 can acquire a computational image in which multiple images and blurring are intentionally created. Thus, the image to be captured becomes a superimposed image, and the multi-pinhole camera 301 can acquire a blurred image in which privacy of the subject is protected. Further, when the pinholes are changed in number, position, and size, images each having a different blurring pattern can be acquired.
Next, the image identification unit 103 performs identification processing on the computational image captured by the first camera 101 (step S102). Using the identification model that is a learning result of the learning device, the image identification unit 103 identifies, as to the computational image captured by the first camera 101, category information about a subject, such as a person (including an action or expression of the person), an automobile, a bicycle, or a traffic light included in the computational image, and position information about the subject. Machine learning such as deep learning using a multilayer neural network may be used for learning for creating an identification model.
The image identification unit 103 identifies the computational image using an identification model that uses the computational image captured by the first camera 101 as input data and the identification result as output data.
Next, the output unit 107 outputs the identification result from the image identification unit 103 (step S103). The output unit 107 may include an interface unit to present the identification result to the user with an image, text, voice, or the like. In addition, the output unit 107 may include a device control unit, and a control method may be changed in accordance with the identification result.
Further,
When the mask pattern of the first camera 101 is changed, the mask identification unit 105 identifies the mask pattern of the mask of the first camera 101 (step S202). Note that the processing in step S202 may be performed periodically, for example, every 1 minute. After the mask pattern is changed, the mask identification unit 105 identifies the changed mask pattern.
Further, when the user changes the mask pattern, the input unit of the image identification system may receive an input of a mask pattern identification instruction from the user. Then, when the input of the mask pattern identification instruction is received, the mask identification unit 105 may identify the changed mask pattern. In this case, the image identification system 1 may include the input unit instead of the light emitting unit 108.
The mask identification unit 105 may determine whether the point spread function in the image captured by the first camera 101 has been changed. Then, in a case where the point spread function is changed, the mask identification unit 105 may determine that the mask pattern has been changed and identify the changed mask pattern.
Next, the identification model change unit 106 changes the identification model in accordance with the mask pattern identified by the mask identification unit 105 (step S203). The identification model change unit 106 changes the identification model in accordance with the mask pattern identified by the mask identification unit 105.
Here, the storage unit 102 stores a plurality of mask patterns and a plurality of identification models in advance in association with each other. The identification model change unit 106 specifies an identification model associated with the mask pattern identified by the mask identification unit 105 from among the plurality of identification models stored in the storage unit 102. The identification model change unit 106 changes the current identification model used by the image identification unit 103 to the specified identification model, and stores the changed identification model in the storage unit 102.
Note that the storage unit 102 may store a plurality of pieces of mask information and the plurality of identification models in advance in association with each other. The mask information is, for example, a point spread function. The identification model change unit 106 may specify an identification model associated with the mask information acquired by the mask identification unit 105 from among the plurality of identification models stored in the storage unit 102.
Further, in a case where the mask information is the point spread function, the identification model change unit 106 may specify an identification model associated with a point spread function most similar to the point spread function acquired by the mask identification unit 105 from among the plurality of identification models stored in the storage unit 102. The identification model change unit 106 determines a correlation (similarity) between the acquired point spread function and a point spread function stored in advance to be able to specify the point spread function most similar to the acquired point spread function.
In a case where the first camera 101 is the multi-pinhole camera 301, the multi-pinhole mask 301a corresponding to the mask of the first camera 101 has a configuration in which a plurality of masks are overlaid. For example, the user changes the mask pattern of the multi-pinhole mask 301a by removing at least one of the plurality of masks disposed in an overlaid manner.
In the first embodiment, the first camera 101 is the multi-pinhole camera 301, the mask is the multi-pinhole mask 301a in which a plurality of masks are overlaid, and the plurality of masks respectively have different mask patterns. When at least one of the plurality of masks is removed, the mask pattern of the multi-pinhole mask 301a is changed.
The nine pinholes 401a to 409a illustrated in
These masks 400a to 400c are fixed so that the pinhole 401a, the pinhole 401b, and the pinhole 401c are aligned with each other, the pinhole 402a and the pinhole 402b are aligned with each other, the pinhole 403a and the pinhole 403b are aligned with each other, the pinhole 404a and the pinhole 404b are aligned with each other, and the pinhole 405a and the pinhole 405b are aligned with each other. At this time, the pinholes 402a to 405a of the mask 400a and the pinholes 402b to 405b of the mask 400b are fixed so as to be aligned with a shielding portion where the pinhole 401c is not formed in the mask 400c. In addition, in the mask 400a, among the eight pinholes 402a to 409a excluding the central pinhole 401a, the four pinholes 406a to 409a at positions different from the pinholes 402b to 405b are fixed so as to be aligned with the shielding portion where the pinholes 401b to 405b are not formed in the mask 400b.
That is, at least one of the plurality of pinholes formed in one of the plurality of masks is at the position identical to the position of as at least one of the plurality of pinholes formed in another one of the plurality of masks. In addition, at least one of the plurality of pinholes formed in one of the plurality of masks is shielded by another one of the plurality of masks.
In the multi-pinhole mask 301a where the three masks 400a, 400b, and 400c are overlaid as described above, one pinhole constituted by the pinholes 401a, 401b, and 401c is formed. Therefore, it can be said that the multi-pinhole mask 301a where the three masks 400a, 400b, and 400c are overlaid has a mask pattern where one pinhole is formed. In a case where each of the three masks 400a, 400b, and 400c has a sufficiently small thickness, the multi-pinhole mask 301a where the three masks 400a, 400b, and 400c are overlaid can be said to be substantially equal to the mask where one pinhole is formed, that is, the mask 400c illustrated in
When imaging is performed by the multi-pinhole camera 301 including the multi-pinhole mask 301a where the three masks 400a, 400b, and 400c are overlaid as described above, among light emitted from a subject included in an imaging range, light that has passed through the pinholes 401a, 401b, and 401c constituting the above-described one pinhole is received on a light receiving surface of the image sensor 301b.
In the multi-pinhole mask 301a where the three masks 400a, 400b, and 400c are overlaid, when the mask 400c is removed by the user, the multi-pinhole mask 301a has a configuration where the two masks 400a and 400b are overlaid.
In the multi-pinhole mask 301a where the two masks 400a and 400b are overlaid, five pinholes are formed. The five pinholes include (1) a pinhole constituted by the pinholes 401a and 401b, (2) a pinhole constituted by the pinholes 402a and 402b, (3) a pinhole constituted by the pinholes 403a and 403b, (4) a pinhole constituted by the pinholes 404a and 404b, and (5) a pinhole constituted by the pinholes 405a and 405b. Therefore, it can be said that the multi-pinhole mask 301a where the two masks 400a and 400b are overlaid has a mask pattern where five pinholes are formed. Further, in a case where each of the two masks 400a and 400b has a sufficiently small thickness, the multi-pinhole mask 301a where the two masks 400a and 400b are overlaid can be said to be substantially equal to the mask where five pinholes are formed, that is, the mask 400b illustrated in
When imaging is performed by the multi-pinhole camera 301 having the multi-pinhole mask 301a where the two masks 400a and 400b are overlaid, among light emitted from a subject included in an imaging range, (1) light passing through the pinholes 401a and 401b constituting the multi-pinhole, (2) light passing through the pinholes 402a and 402b constituting the multi-pinhole, (3) light passing through the pinholes 403a and 403b constituting the multi-pinhole, (4) light passing through the pinholes 404a and 404b constituting the multi-pinhole, and (5) light passing through the pinholes 405a and 405b constituting the multi-pinhole are received on the light receiving surface of the image sensor 301b.
As described above, when the mask 400c is removed from the three masks 400a, 400b, and 400c, the mask pattern of the multi-pinhole mask 301a is changed from the mask pattern where one pinhole is formed to the mask pattern where five pinholes are formed.
In a case where the mask 400b is removed by the user from the multi-pinhole mask 301a where the two masks 400a and 400b are overlaid, the multi-pinhole mask 301a is configured only by the mask 400a. The multi-pinhole mask 301a constituted by one mask 400a becomes a mask where nine pinholes are formed, that is, the mask 400a illustrated in
As described above, when the mask 400b is removed from the two masks 400a and 400b, the mask pattern of the multi-pinhole mask 301a is changed from the mask pattern where the five pinholes are formed to the mask pattern where the nine pinholes are formed.
As described above, in order to remove the mask 400b and the mask 400c, the mask 400a, the mask 400b, and the mask 400c may be fixed with a detachable adhesive such as a seal, and the mask pattern may be changed by the user manually removing each of the mask 400b and the mask 400c from the mask 400a.
Certainly, the user may change the mask pattern by adding a mask instead of removing some masks. In a state where the mask 400a illustrated in
Further, the user fixes the mask 400c to the mask 400b so that the pinhole 401a, the pinhole 401b, and the pinhole 401c are aligned with each other in the state where the two masks 400a and 400b are overlaid. As a result, the multi-pinhole mask 301a has a configuration where the three masks 400a, 400b, and 400c are overlaid, and the multi-pinhole mask 301a can be said to be substantially equal to the mask where one pinhole is formed as described above, that is, the mask 400c illustrated in
Certainly, the user may change the mask pattern by replacing one mask instead of removing one of the plurality of masks. In a state where the mask 400a illustrated in
Furthermore, in the first embodiment, one pinhole included in one of the plurality of masks has a size equal to a size of another pinhole included in another mask at the same position as the above-described one pinhole among the plurality of masks, but the present disclosure is not particularly limited thereto. For example, the one pinhole included in one of the plurality of masks may have a size different from the size of the other pinhole included in the other mask at the same position as of the one pinhole described above among the plurality of masks. For example, even if five pinholes including the pinhole 401b to the pinhole 405b are formed in the mask 400b, and five pinholes of the mask 400a, that is, the pinhole 401a to the pinhole 405a are aligned with the pinhole 401b to the pinhole 405b of the mask 400b, at least one pinhole among the five pinholes of the mask 400a may have a size greater than the size of the pinhole at the same position as the pinhole described above in the mask 400b. As a result, the mask pattern of the multi-pinhole mask where the mask 400a and the mask 400b are overlaid can be made different from the mask pattern of the multi-pinhole mask including only the mask 400a.
Certainly, the user may change the mask pattern by rotating one of the plurality of masks instead of removing one of the plurality of masks.
The multi-pinhole mask 301a has a configuration where the mask 400d illustrated in
When the mask 400d is rotated 30 degrees clockwise from the state of
On the other hand, when the mask 400d is rotated 60 degrees counterclockwise from the state of
Certainly, the mask 400d may be fixed, and the mask 400e may be rotatable. Further, both the mask 400d and the mask 400e may be rotatable.
Thereafter, a configuration of the multi-pinhole camera 301 where a user can arbitrarily rotate a mask in the first modification of the first embodiment will be described with reference to
The multi-pinhole camera 301 includes the multi-pinhole mask 301a where the mask 400d that is rotatable and the mask 400e that does not rotate with respect to a housing 501 are overlaid. A grip portion 502 operable by a user is connected to the rotatable mask 400d. Further, the user can fix or rotate the mask 400d with respect to the housing 501 by gripping and operating the grip portion 502. In such a mechanism, a screw may be provided in the grip portion 502, the mask 400d may be fixed by tightening the screw, and the mask 400d may be rotatable by loosening the screw.
As illustrated in
Furthermore, in the multi-pinhole camera 301 where the mask can be arbitrarily rotated by the user, the plurality of pinholes of the multi-pinhole mask 301a may be disposed so as not to be rotationally symmetric with respect to a vertical line passing through a rotation axis of the mask 400d as illustrated in
As a matter of course, the multi-pinhole camera 301, in which the user can arbitrarily rotate the mask, may be configured without the grip portion 502. Another configuration of the multi-pinhole camera 301 where a user can arbitrarily rotate the mask in a second modification of the first embodiment will be described with reference to
The multi-pinhole camera 301 includes the multi-pinhole mask 301a including the mask 400d that is rotatable and the mask 400e that does not rotate. The rotatable mask 400d is fixed to a first lens barrel 511. Further, the image sensor 301b and the non-rotating mask 400d are installed in a second lens barrel 512 different from the first lens barrel 511. The first lens barrel 511 and the second lens barrel 512 are rotatable with a thread configuration. That is, the second lens barrel 512 is provided outside the first lens barrel 511, and a male thread is cut outside the first lens barrel 511 serving as a joint portion, and a female thread is cut inside the second lens barrel 512. Further, the first lens barrel 511 includes the male thread to which a fixture 513 is first attached, and then the second lens barrel 512 is attached. As with the second lens barrel 512, a female thread is cut also in the fixture 513. With this configuration described above, when the first lens barrel 511 is screwed into the second lens barrel 512, a screwing depth is changed in accordance with a position of screwing the fixture 513 into the first lens barrel 511, and thus a rotation angle of the rotatable mask 400d can be changed.
As illustrated in
Further, in a third modification of the first embodiment, the multi-pinhole mask may be achieved by a user making an arbitrary hole in the mask.
Next, an image identification system in a fourth modification of the first embodiment will be described.
The image identification system 1B includes the first camera 101, the storage unit 102 that stores an identification model, the image identification unit 103, a mask change unit 104, the mask identification unit 105, the identification model change unit 106, the output unit 107, and an input unit 109.
The image identification system 1B may include, for example, an image identification device including the storage unit 102, the image identification unit 103, the mask change unit 104, the mask identification unit 105, and the identification model change unit 106, and may be configured so that the first camera 101, the output unit 107, and the input unit 109 are connected to the image identification device. In addition, the image identification device may be a server. Further, the image identification system 1B may be, for example, an image identification device including the first camera 101, the storage unit 102, the image identification unit 103, the mask change unit 104, the mask identification unit 105, the identification model change unit 106, the output unit 107, and the input unit 109. Note that the image identification device included in the image identification system 1B may be connected to the light emitting unit 108 illustrated in
Further, in the fourth modification of the first embodiment, the multi-pinhole mask 301a includes a spatial light modulator 520 capable of arbitrarily changing the position and size of a hole. The mask change unit 104 changes the position and size of the hole of the multi-pinhole mask 301a. Such a multi-pinhole mask 301a can be achieved by the spatial light modulator such as liquid crystal on silicon (LCOS) or a spatial light phase modulator.
The mask change unit 104 changes the mask pattern of a mask.
The input unit 109 receives an input for changing the mask pattern of the mask from the user. A user inputs the positions and sizes of the plurality of pinholes of the multi-pinhole mask 301a to the input unit 109. For example, the input unit 109 may receive selection of one mask pattern from among a plurality of mask patterns from the user. The user may select one mask pattern from the plurality of mask patterns.
The mask change unit 104 instructs the first camera 101 to change the mask pattern of the multi-pinhole mask 301a in accordance with the positions and sizes of the plurality of pinholes received by the input unit 109. The spatial light modulator 520 included in the first camera 101 changes the positions and sizes of the plurality of pinholes in response to the instruction from the mask change unit 104.
First, the mask change unit 104 changes the mask pattern of the first camera 101 (step S201). The mask change unit 104 notifies the mask identification unit 105 of the changed mask pattern, that is, the positions and sizes of the plurality of pinholes.
Next, the mask identification unit 105 identifies the mask pattern of the mask of the first camera 101 (step S202). The mask identification unit 105 identifies the mask pattern of the mask based on the changed mask pattern notified from the mask change unit 104.
Furthermore, the mask change unit 104 may change the mask pattern by applying an external force to the multi-pinhole mask 301a to deform the mask. The multi-pinhole camera 301 configured so that the mask is deformed by application of an external force in a fifth modification of the first embodiment will be described with reference to
The multi-pinhole mask 301ac has a configuration where the first mask 301a1, the second mask 301a2, and the third mask 301a3 are overlaid. The multi-pinhole camera 301, which is one example of the first camera 101, has a drive unit that independently applies an external force to each of the first mask 301a1, the second mask 301a2, and the third mask 301a3. Here, each mask has a shape in which a fan shape and a circular ring are combined. As a matter of course, this configuration is one example, and the shape is not limited to the fan shape, and the number of the plurality of masks constituting the multi-pinhole mask 301ac is not limited to 3. Each mask has one or a plurality of pinholes. Further, the mask may have no pinhole.
The first mask 301a1 has two pinholes 301aa1 and 301aa2. The second mask 301a2 has one pinhole 301aa3. The third mask 301a3 has two pinholes 301aa4 and 301aa5. The drive unit moves these three masks with application of an external force to create various mask patterns.
The mask change unit 104 instructs the first camera 101 to change the mask pattern of the multi-pinhole mask 301ac in accordance with the number and positions of the plurality of pinholes received by the input unit 109. The drive unit included in the first camera 101 applies an external force to the plurality of masks in response to the instruction from the mask change unit 104 to change the number and positions of the plurality of pinholes of the multi-pinhole mask 301ac.
As described above, the multi-pinhole mask 301ac can be changed in number and positions of pinholes by application of an external force. For example, when an external force is applied to the multi-pinhole mask 301ac illustrated in
Certainly, the mask change unit 104 may change not only the number or positions of the plurality of pinholes of the multi-pinhole mask but also the sizes of the plurality of pinholes.
The multi-pinhole mask 301ad has a plurality of pinholes and is made of a material having elasticity. The multi-pinhole mask 301ad has, for example, three pinholes. The multi-pinhole camera 301, which is one example of the first camera 101, has a plurality of independently controllable drive units 521 to 524. As a matter of course, the number of drive units does not need to be four. The mask change unit 104 can change the positions or sizes of the plurality of pinholes in the multi-pinhole mask 301ad by moving each drive unit.
The mask change unit 104 instructs the first camera 101 to change the mask pattern of the multi-pinhole mask 301ad in accordance with the positions and sizes of the plurality of pinholes received by the input unit 109. Each of the drive units included in the first camera 101 applies an external force to the multi-pinhole mask 301ad in response to the instruction from the mask change unit 104 to change the positions and sizes of the plurality of pinholes of the multi-pinhole mask 301ad.
Note that the number and positions of the plurality of pinholes of the multi-pinhole mask 301ad may be changed by moving the multi-pinhole mask 301ad parallel with the drive direction. For example, the number of pinholes may increase as the pinholes outside the imaging range move into the imaging range, and the number of pinholes may decrease as the pinholes within the imaging range move out of the imaging range.
On the other hand,
Note that in the first embodiment, the user overlays the plurality of masks of the multi-pinhole mask, but the present disclosure is not particularly limited thereto. The multi-pinhole camera 301 may include a drive unit that inserts at least one of the plurality of masks into the front of the image sensor 301b or removes at least one of the plurality of masks from the front of the image sensor 301b. In this case, the mask change unit 104 may control the drive unit in response to an instruction from the input unit 109 to change the number of masks constituting the multi-pinhole mask. For example, in a case where the multi-pinhole mask 301a includes the three masks 400a to 400c illustrated in
Further, in the first embodiment, the user rotates a rotatable mask among the plurality of masks of the multi-pinhole mask, but the present disclosure is not particularly limited thereto. The multi-pinhole camera 301 may include a drive unit that rotates at least one of the plurality of masks. In this case, the mask change unit 104 may cause the drive unit in response to an instruction from the input unit 109 to rotate at least one of the plurality of masks constituting the multi-pinhole mask. For example, in a case where the multi-pinhole mask 301a includes the rotatable mask 400d illustrated in
The mask identification unit 105 may identify the mask pattern of the mask by acquiring a point spread function (PSF) of the multi-pinhole mask 301a from the image captured by the first camera 101. The PSF is a transfer function of the first camera 101 such as a multi-pinhole camera or a coded aperture camera, and is expressed by the following formula (1).
y=k*x (1)
In Formula (1), y represents a blurred computational image captured by the first camera 101, k represents a PSF, and x represents a normal image obtained by a normal camera imaging without blurring imaging a scene captured by the first camera 101. Further, * represent a convolution operator.
The mask identification unit 105 acquires mask information about the mask pattern from the computational image captured by the first camera 101, and identifies the mask pattern of the mask based on the acquired mask information. The mask identification unit 105 acquires a point spread function as the mask information from the image including the light emitting unit 108, the image being captured by the first camera 101.
First, the mask identification unit 105 instructs the light emitting unit 108 existing in the imaging environment to light up (step S301). The light emitting unit 108 is a light source that can be regarded as a point light source existing in the environment, and is a light emitting diode (LED) mounted on an electric apparatus or an LED for illumination, for example. In addition, the light emitting unit 108 may emit only a part of light of a monitor such as an LED monitor.
Next, the light emitting unit 108 performs lighting in response to a lighting instruction from the mask identification unit 105 (step S302).
Next, the mask identification unit 105 instructs the first camera 101 to perform imaging (step S303). As a result, the first camera 101 can image the lit light emitting unit 108.
Next, in response to the imaging instruction from the mask identification unit 105, the first camera 101 captures a computational image during lighting-up (step S304). The computational image obtained by imaging the lit light emitting unit 108 is input from the first camera 101 to the mask identification unit 105. The mask identification unit 105 acquires the computational image obtained by imaging the lit light emitting unit 108 from the first camera 101. The mask identification unit 105 temporarily stores the acquired computational image.
Next, the mask identification unit 105 instructs the light emitting unit 108 as a light source to light up (step S305).
Next, the light emitting unit 108 lights out in response to the lighting-out instruction from the mask identification unit 105 (step S306).
Next, the mask identification unit 105 instructs the first camera 101 to perform imaging (step S307). As a result, the first camera 101 can image the lit-out light emitting unit 108.
Next, in response to the lighting-out instruction from the mask identification unit 105, the first camera 101 captures a computational image during the lighting-out (step S308). The computational image obtained by imaging the lit-out light emitting unit 108 is input from the first camera 101 to the mask identification unit 105. The mask identification unit 105 acquires the computational image obtained by imaging the lit-out light emitting unit 108 from the first camera 101. The mask identification unit 105 temporarily stores the acquired computational image.
Next, the mask identification unit 105 calculates a difference image between the computational image during the lighting-up of the light emitting unit 108 and the computational image during the lighting-out of the light emitting unit 108 (step S309).
The image during the lighting-up of the light emitting unit 108 and the image during the lighting-out of the light emitting unit 108 are desirably captured in a time difference as little as possible. Calculating the difference image between the image during lighting-up of the light emitting unit 108 and the image during lighting-out of the light emitting unit 108 as described above enables acquiring a PSF being an image of only the light emitting unit 108 in a lighting-on state without being affected by a subject in the environment.
Next, processing of the mask identification unit 105 that identifies the mask pattern using the PSF will be described with reference to
Further, the light emitting unit 108 may be a light of a smartphone or a mobile phone of a user. In this case, lighting up or off the light as the light emitting unit 108 may be implemented by the user operating a smartphone or a mobile phone.
Furthermore, the mask identification unit 105 may identify the mask pattern by acquiring a light transport matrix (LTM), which is a transport function used in the light-field camera, instead of acquiring the point spread function of the multi-pinhole mask 301a from the image captured by the first camera 101. In this case, the mask identification unit 105 may use the plurality of light emitting units 108 dispersedly disposed in the environment, acquire the PSF at the plurality of positions, and set the PSF as the LTM.
Further, in a case where the mask identification unit 105 identifies the mask pattern by acquiring the LTM of the multi-pinhole mask 301a from the image captured by the first camera 101, the PSFs at a plurality of positions may be acquired using the plurality of light emitting units 108, and the PSFs may be set as the LTM.
First, the mask identification unit 105 calculates a plurality of PSFs respectively corresponding to the plurality of light emitting units 108 (step S331). As described above, this may be calculated using the difference image between a computational image during the lighting-up of the plurality of light emitting units 108 and a computational image during the lighting-out of the plurality of light emitting units 108. This processing enables acquiring the PSFs at a plurality of positions on the image.
Next, the mask identification unit 105 acquires the LTM by interpolating the calculated plurality of PSFs (step S332). The mask identification unit 105 calculates the PSFs in all the pixels of the image by performing the interpolation processing on the calculated plurality of PSFs, and sets the PSFs as the LTM. Such interpolation processing may use general image processing such as morphing.
Furthermore, in a case where the mask identification unit 105 identifies the mask pattern using the LTM, the image identification system may include one light emitting unit 108 instead of the plurality of light emitting units 108. The one light emitting unit 108 may be used at a plurality of positions. This processing may be implemented by using a light of a smartphone or a mobile phone as the light emitting unit 108, and lighting up and out the one light emitting unit 108 while a user changes its location, for example. Alternatively, an LED mounted on a drone or a vacuum cleaner robot may be used as the light emitting unit 108.
Further, the mask identification unit 105 may determine the state of the image quality of the mask information acquired based on the computational image captured by the first camera 101, and switch contents of the processing in accordance with the determination result. The mask information is a PSF.
First, the mask identification unit 105 calculates the difference image between the computational image during the lighting-up of the light emitting unit 108 and the computational image during the lighting-out using the method similar to steps S301 to S309 in
Next, the mask identification unit 105 checks the image quality of the calculated difference image and determines whether the image quality of the calculated difference image is sufficient (step S310). Here, in a case where a determination is made that the image quality of the difference image is not sufficient, that is, the image quality of the difference image is insufficient (NO in step S310), the processing returns to step S301, and the mask identification unit 105 instructs the light emitting unit 108 to light up. The mask identification unit 105 instructs the light emitting unit 108 to light up, instructs the first camera 101 to image the light emitting unit 108 in the lighting-up state, instructs the light emitting unit 108 to light off, and instructs the first camera 101 to image the light emitting unit 108 in the lighting-out state. In step S309, the mask identification unit 105 calculates a new difference image between the computational image during the lighting-up of the light emitting unit 108 and the computational image during the lighting-out of the light emitting unit 108, and then performs the processing in step S310 again.
On the other hand, in a case where the determination is made that the image quality of the difference image is sufficient (YES in step S310), the mask identification unit 105 stores the calculated difference image as a PSF (step S311).
The determination may be made whether the image quality of the PSF is sufficient using, for example, a difference image. The PSF needs to show nothing except for the point light source. Therefore, the difference image between the lighting-up state and the lighting-out state is used. Unfortunately, in a case where there is a change in scene such as a large movement of a person or a dramatic change in brightness in the environment between the imaging during the lighting-up and the imaging during the lighting-out, the change appears in the difference image, and thus an accurate PSF cannot be acquired. Therefore, the mask identification unit 105 may count the number of pixels having luminance of a certain value or more in the difference image, and may determine that the image quality of the difference image is insufficient in a case where the counted number of pixels is a threshold or more. Further, the mask identification unit 105 may determine that the image quality of the difference image is sufficient in a case where the counted number of pixels is smaller than the threshold.
Note that, in step S310, the mask identification unit 105 may determine whether the image quality of the calculated difference image is an allowable value or more. The mask identification unit 105 may count the number of pixels having luminance of the certain value or more in the difference image, and may determine that the image quality of the difference image is smaller than the allowable value in a case where the counted number of pixels is the threshold or more. Furthermore, in a case where the counted number of pixels is smaller than the threshold, the mask identification unit 105 may determine that the image quality of the difference image is the allowable value or more.
Further, in a case where the setting of the first camera 101 is inappropriate, the image quality of the difference image may be degraded. For example, in a case where the exposure time of the first camera 101 is too short or the gain of signal amplification for adjusting the brightness of image data is too small, the luminance of the light emitting unit 108 is buried in noise, and the image quality of the difference image (PSF) may be deteriorated. Conversely, in a case where the exposure time of the first camera 101 is too long or the gain is too great, the luminance value of a high luminance region in the image exceeds an upper limit value of the sensing range and is saturated, and thus the image quality of the difference image (PSF) may be deteriorated.
Therefore, the mask identification unit 105 may check the maximum luminance value of the image during the lighting-up of the light emitting unit 108, and may determine that the image quality of the difference image (PSF) is insufficient in a case where the maximum luminance value is equal to or greater than the upper limit value or equal to or smaller than the lower limit value. In a case where the mask identification unit 105 determines the image quality of the difference image (PSF) based on whether the maximum luminance value is equal to or smaller than the lower limit value in the image during the lighting-up of the light emitting unit 108, a determination can be made whether the luminance of the light emitting unit 108 is buried in noise. Further, in a case where the mask identification unit 105 determines the image quality of the difference image (PSF) based on whether the maximum luminance value is equal to or greater than the upper limit value in the image during the lighting-up of the light emitting unit 108, a determination can be made whether the luminance of the light emitting unit 108 exceeds the sensing range and is saturated. In a case where the luminance of the light emitting unit 108 is buried in noise or the luminance of the light emitting unit 108 is saturated, the mask identification unit 105 may change the setting of the first camera 101 so that the maximum luminance value falls within a predetermined range between the upper limit value and the lower limit value.
Note that, the mask identification unit 105 may check the maximum luminance value of the image during the lighting-out of the light emitting unit 108, and may determine that the image quality of the difference image (PSF) is insufficient in a case where the maximum luminance value is equal to or greater than the upper limit value or equal to or smaller than the lower limit value.
The processing in step S304 in
Next, the mask identification unit 105 acquires a computational image captured by the first camera 101 during the lighting-up of the light emitting unit 108 (step S321).
Next, the mask identification unit 105 determines whether the maximum luminance value of the acquired computational image is equal to or greater than an upper limit value Th1 to determine whether the computational image is saturated (step S322). In a case where a determination is made that the maximum luminance value is equal to or greater than the upper limit value Th1, that is, a determination is made that the acquired computational image is saturated (YES in step S322), the mask identification unit 105 instructs the first camera 101 to again perform imaging with a shorter exposure time (step S323). The processing then returns to step S304.
On the other hand, in a case where a determination is made that the maximum luminance value is smaller than the upper limit value Th1 (NO in step S322), the mask identification unit 105 determines whether the maximum luminance value of the acquired computational image is equal to or smaller than a lower limit value Th2 to determine whether the luminance of the light emitting unit 108 is buried in noise (step S324). Note that the lower limit value Th2 is smaller than the upper limit value Th1. In a case where a determination is made that the maximum luminance value is equal to or smaller than the lower limit value Th2, that is, a determination is made that the luminance of the light emitting unit 108 is buried in noise (YES in step S324), the mask identification unit 105 instructs the first camera 101 to again perform imaging with a longer exposure time (step S325). The processing then returns to step S304.
On the other hand, in a case where a determination is made that the maximum luminance value is greater than the lower limit value Th2 (NO in step S324), the mask identification unit 105 determines that the image quality of the acquired computational image is sufficiently high with the current exposure time. In this case, the processing proceeds to step S305, and the mask identification unit 105 instructs the light emitting unit 108 to light out. Then, the mask identification unit 105 further instructs the first camera 101 to perform imaging with the exposure time equal to in the imaging during the lighting-up of the light emitting unit 108, thereby acquiring a computational image during the lighting-out of the light emitting unit 108.
Note that as for the computational image acquired when the light emitting unit 108 lights up, the mask identification unit 105 may control the exposure time of the first camera 101 so that the maximum luminance value falls within the predetermined range between the upper limit value Th1 and the lower limit value Th2, as described above.
As a matter of course, the mask identification unit 105 may change the setting other than the exposure time of the first camera 101. For example, the mask identification unit 105 may change the gain of the first camera 101.
In a case where a determination is made in step S322 that the maximum luminance value is equal to or greater than the upper limit value Th1, that is, a determination is made that the acquired computational image is saturated (YES in step S322), the mask identification unit 105 instructs the first camera 101 to again perform imaging with a smaller gain (step S327). The processing then returns to step S304.
In a case where a determination is made in step S324 that the maximum luminance value is equal to or smaller than the lower limit value Th2, that is, a determination is made that the luminance of the light emitting unit 108 is buried in noise (YES in step S324), the mask identification unit 105 instructs the first camera 101 to again perform imaging with a greater gain (step S328). The processing then returns to step S304.
On the other hand, in a case where a determination is made that the maximum luminance value is greater than the lower limit value Th2 (NO in step S324), the mask identification unit 105 determines that the image quality of the acquired computational image is sufficiently high with the current gain. In this case, the processing proceeds to step S305, and the mask identification unit 105 instructs the light emitting unit 108 to light out. Then, the mask identification unit 105 further instructs the first camera 101 to perform imaging with the exposure time and gain equal to in the imaging during the lighting-up of the light emitting unit 108, thereby acquiring a computational image during the lighting-out of the light emitting unit 108.
Note that as for the computational image acquired when the light emitting unit 108 lights out, the mask identification unit 105 may control the gain of the first camera 101 so that the maximum luminance value falls within the predetermined range between the upper limit value Th1 and the lower limit value Th2, as described above.
Further, the mask identification unit 105 may control the luminance of the light emitting unit 108 instead of the exposure time or gain of the first camera 101. That is, in a case where a determination is made that the luminance of the light emitting unit 108 is saturated, the mask identification unit 105 controls the light emitting unit 108 so that the luminance decreases. Conversely, in a case where a determination is made that the luminance of the light emitting unit 108 is buried in noise, the mask identification unit 105 controls the light emitting unit 108 so that the luminance increases. Increasing the luminance of the light emitting unit 108 increases a luminance difference between the light emitting unit 108 and the noise.
In a case where a determination is made that the image quality of the difference image is insufficient, the mask identification unit 105 may select another light emitting unit present in the target area and instruct the another light emitting unit to light up and light off. This configuration is effective for a light source having directivity because image quality inevitably deteriorates depending on a positional relationship between the first camera 101 and the light emitting unit 108.
In the first embodiment, the mask identification unit 105 acquires mask information about the mask pattern from the computational image captured by the first camera 101, and identifies the mask pattern of the mask based on the acquired mask information. Conversely, in a second embodiment, the mask identification unit may identify a mask pattern by reading mask ID information described in a mask. This is implemented, for example, by reading mask ID information embedded around a mask. The mask ID information is, for example, a barcode, a two-dimensional code, color information, or the like.
The image identification system 1C of the second embodiment includes a first camera 101, a storage unit 102, an image identification unit 103, a mask identification unit 105, an identification model change unit 106, an output unit 107, an ID information acquisition unit 111, and a second camera 121. The mask identification unit 105, the identification model change unit 106, the image identification unit 103, and the ID information acquisition unit 111 include a processor such as a CPU and a memory such as a semiconductor memory.
In addition, in a case where the ID information acquisition unit 111 transmits and receives information to and from second camera 121 via a wired or wireless network, the ID information acquisition unit 111 includes a reception unit. The reception unit includes a reception circuit as hardware. Further, in a case where the mask identification unit 105 transmits and receives information to and from the second camera 121 via a wired or wireless network, the mask identification unit 105 includes a transmission unit. The transmission unit includes a transmission circuit as hardware. On the other hand, the second camera 121 includes a transmission circuit and a reception circuit as hardware.
The image identification system 1C may include, for example, an image identification device including the storage unit 102, the image identification unit 103, the mask identification unit 105, the identification model change unit 106, and the ID information acquisition unit 111, and may be configured so that the first camera 101, the output unit 107, and the second camera 121 are be connected to the image identification device. In addition, the image identification device may be a server. Further, the image identification system 1C may include, for example, an image identification device including the first camera 101, the storage unit 102, the image identification unit 103, the mask identification unit 105, the identification model change unit 106, the output unit 107, and the ID information acquisition unit 111, and may be configured so that the second camera 121 is connected to the image identification device.
The Second camera 121 is a camera different from first camera 101, and captures an image without blurring. The second camera 121 is disposed on the subject side captured by the first camera 101. The first camera 101 is a multi-pinhole camera 301. The second camera 121 images a multi-pinhole mask 301a included in the multi-pinhole camera 301 on the subject side. The mask identification unit 105 instructs the second camera 121 to capture an image. The second camera 121 acquires an image obtained by imaging the multi-pinhole mask 301a on the subject side in response to the instruction from the mask identification unit 105.
The ID information acquisition unit 111 acquires mask ID information (mask identification information) for identifying a mask from an image captured by the second camera 21.
The mask identification unit 105 identifies a mask pattern of the mask based on the mask ID information acquired by the ID information acquisition unit 111.
Processing of the mask identification unit 105 for acquiring mask ID information to identify a mask pattern will be described with reference to
The multi-pinhole mask 301a is configured by overlaying the mask 400a illustrated in
The mask ID information 601 to 603 is represented by, for example, a two-dimensional code. The ID information acquisition unit 111 acquires the mask ID information 601 to 603 about the masks 400a to 400c by reading the two-dimensional codes described respectively in the masks 400a to 400c included in a captured image. The two-dimensional codes respectively representing the mask ID information 601 to 603 may be printed respectively on the masks 400a to 400c, or may be affixed with a seal or the like.
Further, the mask identification unit 105 stores mask information corresponding one-to-one to the mask ID information 601, 602, and 603 in advance. The mask information corresponding to the mask ID information 601 is, for example, information about a PSF of the multi-pinhole mask 301a including only the mask 400a. In a case where the multi-pinhole mask 301a includes the mask 400a, the multi-pinhole mask 301a has nine pinholes as described above. Therefore, the mask information corresponding to the mask ID information 601 is the PSF of the mask 400a having nine pinholes.
The mask information corresponding to the mask ID information 602 is, for example, information about the PSF of the multi-pinhole mask 301a where the two masks 400a and 400b are overlaid. In a case where the multi-pinhole mask 301a includes the two masks 400a and 400b, the multi-pinhole mask 301a has five pinholes. Note that, in a case where the thicknesses of the masks 400a and 400b are sufficiently small, the mask information corresponding to the mask ID information 602 is the PSF of the mask 400b having five pinholes.
The mask information corresponding to the mask ID information 603 is, for example, information about the PSF of the multi-pinhole mask 301a where the three masks 400a, 400b, and 400c are overlaid. In a case where the multi-pinhole mask 301a includes the three masks 400a, 400b, and 400c, the multi-pinhole mask 301a has one pinhole. Note that, in a case where the thicknesses of the masks 400a, 400b, and 400c are sufficiently small, the mask information corresponding to the mask ID information 603 is the PSF of the mask 400c having one pinhole.
As described above, these masks are fixed with a detachable adhesive. A user manually removes each mask to change the mask pattern of the multi-pinhole mask 301a. Note that the mask change unit 104 may change the mask pattern of the multi-pinhole mask 301a by controlling the drive unit to replace each mask.
In a case where the second camera 121 images the multi-pinhole mask 301a where the three masks 400a, 400b, and 400c are overlaid, as illustrated in
In addition, in a case where the mask 400c is removed from the multi-pinhole mask 301a where the three masks 400a, 400b, and 400c are overlaid, the multi-pinhole mask 301a is configured by the two overlaid masks 400a and 400b. In a case where the second camera 121 images the multi-pinhole mask 301a where the two masks 400a and 400b are overlaid, as illustrated in
In addition, in a case where the mask 400b is removed from the multi-pinhole mask 301a where the two masks 400a and 400b are overlaid, the multi-pinhole mask 301a is configured only by the mask 400a. In a case where the second camera 121 images the multi-pinhole mask 301a having the mask 400a, as illustrated in
In order for the ID information acquisition unit 111 to acquire such mask ID information, the second camera 121 needs to acquire an image obtained by imaging the multi-pinhole mask 301a on the subject side.
The second camera 121 is another camera different from the first camera 101, and is disposed on a subject side where the multi-pinhole camera 301 corresponding to the first camera 101 performs imaging. For example, second camera 121 captures an image without blurring. For example, the mask identification unit 105 can acquire an image obtained by imaging the multi-pinhole mask 301a on the subject side by instructing the second camera 121 to perform imaging. The captured image includes the mask ID information 600.
In the modification of the second embodiment, the mirror 122 is disposed instead of the second camera 121. In this case, the image identification system 1C does not include the second camera 121. The first camera 101 images the mirror 122 in which the multi-pinhole mask 301a having the mask ID information 600 described on the surface thereof is reflected. The ID information acquisition unit 111 acquires an image captured by the multi-pinhole camera 301 corresponding to the first camera 101. The ID information acquisition unit 111 acquires mask ID information from a computational image captured by the multi-pinhole camera 301.
The mirror 122 is disposed on the subject side where the multi-pinhole camera 301 corresponding to the first camera 101 performs imaging. The multi-pinhole mask 301a in which the mask ID information 600 is described is reflected in the mirror 122. When the mask identification unit 105 instructs the multi-pinhole camera 301 to perform imaging, the light emitted from the mask constituting the multi-pinhole mask 301a is reflected by the mirror 122 and then reaches the light receiving surface of the image sensor 301b. Therefore, the mask identification unit 105 can acquire an image obtained by imaging the multi-pinhole mask 301a on the subject side by instructing the multi-pinhole camera 301 to perform imaging.
Note that, since the image captured by the multi-pinhole camera 301 is a computational image with blurring, the ID information acquisition unit 111 needs to acquire the mask ID information from the computational image. Therefore, for example, the storage unit 102 may store in advance a mask ID identification model for identifying the mask ID information about the multi-pinhole mask 301a from the computational image captured by the multi-pinhole camera 301. The ID information acquisition unit 111 may identify the mask ID information from the image of the multi-pinhole mask 301a, the image being captured by the multi-pinhole camera 301, using the mask ID identification model. The ID information acquisition unit 111 may input the image captured by the multi-pinhole camera 301 to the mask ID identification model and acquire the mask ID information output as the identification result from the mask ID identification model. The mask ID identification model is created by learning processing performed by a learning device.
In a case where the mask ID identification model is constructed by a multilayer neural network, the learning device may create the mask ID identification model by performing machine learning with deep learning using, as teacher data, the image of the multi-pinhole mask 301a including the mask ID information, the image being captured by the normal camera, and annotation information corresponding to the mask ID information given to the image.
Note that the second camera 121 and the mirror 122 are necessary in a case where the ID information acquisition unit 111 reads the mask ID information 600, but are unnecessary in a case where the image identification unit 103 identifies an image captured by the first camera 101. For example, it is desirable to configure so that when the ID information acquisition unit 111 reads the mask ID information 600, the second camera 121 or the mirror 122 is installed on the subject side where the multi-pinhole camera 301 performs imaging, and the installed second camera 121 or mirror 122 is removed after the reading of the mask ID information 600 is completed.
In a third embodiment, in a case where the user changes the mask pattern by rotating one of the plurality of masks instead of removing one of the plurality of masks, the mask identification unit 105 may recognize the rotation angle of the mask by reading marker information described on the mask and identify the mask pattern. In the third embodiment, the multi-pinhole mask 301a is configured so that two masks 400d and 400e described with reference to
The image identification system 1D of the third embodiment includes a first camera 101, a storage unit 102, an image identification unit 103, a mask identification unit 105, an identification model change unit 106, an output unit 107, a marker position specifying unit 112, and a second camera 121. The mask identification unit 105, the identification model change unit 106, the image identification unit 103, and the marker position specifying unit 112 include a processor such as a CPU and a memory such as a semiconductor memory.
In addition, in a case where the marker position specifying unit 112 transmits and receives information to and from second camera 121 via a wired or wireless network, the marker position specifying unit 112 includes a reception unit. The reception unit includes a reception circuit as hardware. Further, in a case where the mask identification unit 105 transmits and receives information to and from the second camera 121 via a wired or wireless network, the mask identification unit 105 includes a transmission unit. The transmission unit includes a transmission circuit as hardware. On the other hand, the second camera 121 includes a transmission circuit and a reception circuit as hardware.
The image identification system 1D may include, for example, an image identification device including the storage unit 102, the image identification unit 103, the mask identification unit 105, the identification model change unit 106, and the marker position specifying unit 112, and may be configured so that the first camera 101, the output unit 107, and the second camera 121 are connected to the image identification device. In addition, the image identification device may be a server. Further, the image identification system 1D may include, for example, an image identification device including the first camera 101, the storage unit 102, the image identification unit 103, the mask identification unit 105, the identification model change unit 106, the output unit 107, and the marker position specifying unit 112, and may be configured so that the second camera 121 is connected to the image identification device.
The multi-pinhole mask 301a includes the mask 400d and the mask 400e overlaid on the mask 400d. The mask 400d is one example of a first mask, and the mask 400e is one example of a second mask.
The marker position specifying unit 112 specifies the positions of a plurality of markers formed on each of the mask 400d and the mask 400e in order to detect the rotation angle of the mask 400d with respect to the mask 400e.
The mask identification unit 105 detects the rotation angle of the mask 400d with respect to the mask 400e based on the positions of the plurality of markers specified by the marker position specifying unit 112, and identifies the mask pattern of the mask based on the detected rotation angle.
Processing for recognizing a rotation angle of one mask by reading identification information described in a plurality of masks constituting the multi-pinhole mask 301a and identifying a mask pattern based on the recognized rotation angle will be described with reference to
As illustrated in
The marker 611 is formed at a predetermined position on the outer peripheral portion of the mask 400d. For example, the marker 611 is a number “1”. The marker 612 is formed at a predetermined position on the outer peripheral portion of the mask 400e, and the marker 613 is formed at a position facing the marker 612 on the outer peripheral portion of the mask 400e. A straight line connecting the marker 612 and the marker 613 passes through the center point of the mask 400e. For example, the marker 612 and the marker 613 are numbers “2”. Note that the markers 611, 612, and 613 are not limited to a number, and may be a bar code, a two-dimensional code, color information, or the like, and may be any information that can identify each marker. Further, the markers 611, 612, and 613 may be printed on the masks 400d and 400e, or may be attached with a seal or the like.
The multi-pinhole mask 301a illustrated in
The marker position specifying unit 112 specifies the positions of the markers 611, 612, and 613 from the image obtained by imaging the multi-pinhole mask 301a. For example, the marker position specifying unit 112 specifies the coordinate positions of the markers 611, 612, and 613 in the image.
The mask identification unit 105 detects the rotation angle of the mask 400d with respect to the mask 400e based on the positions of the markers 611, 612, and 613 specified by the marker position specifying unit 112, and identifies the mask pattern of the mask based on the detected rotation angle.
The mask identification unit 105 stores mask information corresponding to the rotation angle of the mask 400d with respect to the mask 400e in advance. The mask information corresponding to the rotation angle is, for example, information about the PSF that changes in accordance with the rotation angle of the mask 400d with respect to the mask 400e. As the mask 400d rotates with respect to the mask 400e, the positions and the number of the plurality of pinholes formed in the multi-pinhole mask 301a change.
In a case where the second camera 121 images the multi-pinhole mask 301a where the two masks 400d and 400e are overlaid, as illustrated in
The mask identification unit 105 acquires mask information corresponding to the rotation angle of the mask 400d with respect to the mask 400e from the memory. The mask information corresponding to the rotation angle of the mask 400d with respect to the mask 400e is a PSF corresponding to the mask pattern in accordance with the rotation angle.
The identification model change unit 106 changes the identification model in accordance with the mask pattern (mask information) identified by the mask identification unit 105.
In this way, since the image identification unit 103 can identify the computational image using the identification model optimal for the mask pattern of the multi-pinhole mask 301a of the first camera 101, high identification accuracy can be maintained even if the mask is changed.
In the first to third embodiments, the storage unit 102 stores a plurality of mask patterns and a plurality of identification models in advance in association with each other, and the identification model change unit 106 selects an identification model associated with the mask pattern identified by the mask identification unit 105 from the plurality of identification models stored in the storage unit 102.
On the other hand, the image identification system according to the fourth embodiment may further include a learning device. The learning device may learn the identification model in accordance with the mask pattern identified by the mask identification unit 105.
The image identification system 1A of the fourth embodiment includes a first camera 101, a storage unit 102, an image identification unit 103, a mask identification unit 105A, an identification model change unit 106A, an output unit 107, a light emitting unit 108, and the learning device 2. The mask identification unit 105A, the identification model change unit 106A, and the image identification unit 103 include a processor such as a CPU and a memory such as a semiconductor memory.
The image identification system 1A may include, for example, an image identification device including the storage unit 102, the image identification unit 103, the mask identification unit 105A, and the identification model change unit 106A, and may be configured so that the first camera 101, the output unit 107, the light emitting unit 108, and the learning device 2 are connected to the image identification device. In addition, the image identification device and the learning device 2 may be a server. Further, the image identification system 1A may include, for example, an image identification device including the first camera 101, the storage unit 102, the image identification unit 103, the mask identification unit 105A, the identification model change unit 106A, and the output unit 107, and may be configured so that the light emitting unit 108 and the learning device 2 are connected to the image identification device. In addition, the image identification device may include the learning device 2.
The learning device 2 acquires a first learning image captured by a camera (second camera) that captures an image with no blurring or an image with a blurring less than of an image captured by the first camera 101 and a correct answer label given to the first learning image. The learning device 2 generates a blurred second learning image based on the mask pattern identified by the mask identification unit 105A and the first learning image. The learning device 2 creates an identification model for identifying a computational image captured by the first camera 101 by performing machine learning using the second learning image and the correct answer label.
The identification model change unit 106A changes the current identification model to the identification model created by the learning device 2.
The learning device 2 includes a learning database 201, a mask information acquisition unit 202, a database correction unit 203, and a learning unit 204. The learning device 2 trains the identification model to be used for identification of the image identification unit 103 by using the mask information identified by the mask identification unit 105. As described above, a PSF may be used as the mask information. The learning database 201 is a storage unit such as an HDD, an SSD, or a semiconductor memory.
The mask information acquisition unit 202, the database correction unit 203, and the learning unit 204 include a processor such as a CPU as hardware. In addition, in a case where the mask information acquisition unit 202 transmits and receives information to and from the mask identification unit 105A via a wired or wireless network, the mask information acquisition unit 202 includes a reception unit. The reception unit includes a reception circuit as hardware. On the other hand, the mask identification unit 105A includes a transmission circuit and a processor such as a CPU as hardware.
Further, in a case where the learning unit 204 transmits and receives information to and from the identification model change unit 106A via the network, the learning unit 204 includes a transmission unit. The transmission unit includes a transmission circuit as hardware. The identification model change unit 106A includes a reception circuit and a processor such as a CPU as hardware.
First, the mask information acquisition unit 202 acquires the mask information identified by the mask identification unit 105A from the mask identification unit 105A (step S401). In a case where the mask identification unit 105A includes a transmission unit and the mask information acquisition unit 202 includes a reception unit, the mask information may be transmitted and received in a wired or wireless manner. In a case where the mask information acquisition unit 202 includes an interface, the mask information may be given by a user.
Subsequently, the database correction unit 203 corrects the learning database 201 using the mask information acquired by the mask information acquisition unit 202 (step S402). For example, when the image identification unit 103 identifies a behavior of a person in an environment, the learning database 201 holds an image without blurring captured by a normal camera, and annotation information (correct answer label) that is given to each image and that indicates a position at which the person has performed what kind of behavior in the image. When a normal camera is used, annotation information may be assigned to an image captured by the camera. However, when a computational image is acquired by a multi-pinhole camera or a light-field camera, it is difficult to assign annotation information to the image because a person cannot find what the image shows even when looking at the image. Additionally, even when learning processing is performed on an image captured by a normal camera significantly different from the first camera 101, the image identification unit 103 does not increase in identification accuracy.
Therefore, a database in which annotation information is added in advance to an image captured by a normal camera is held as the learning database 201. The database correction unit 203 uses the mask information acquired from the mask identification unit 105A to transform an image captured by a normal camera into an image to be captured by the first camera 101, thereby creating a learning data set corresponding to the mask pattern of the first camera 101 and correcting the learning database 201. The identification model is created by performing the learning processing using the created learning data set. Then, the image identification unit 103 performs identification using the created identification model, thereby improving identification accuracy. Therefore, the database correction unit 203 calculates, from a captured image z that is obtained from the normal camera and is prepared in advance in the learning database 201, a corrected image y based on the following formula (2) using a PSF that is mask information acquired by the mask information acquisition unit 202.
y=k*z (2)
In the formula (2), k represents the PSF that is the mask information acquired by the mask information acquisition unit 202, and * represents a convolution operator.
The database correction unit 203 stores the corrected image and the annotation information in the learning database 201.
Next, the learning unit 204 performs the learning processing using the corrected learning database 201 (step S403). The learning unit 204 performs the learning processing by using the corrected image calculated by the database correction unit 203 and the annotation information held in advance in the learning database 201, and creates a trained identification model. For example, in a case where the identification model is constructed by a multilayer neural network, the learning unit 204 performs machine learning with Deep Learning using, as teacher data, the corrected image and the annotation information added to the image captured by the normal camera used when the corrected image is calculated. As a prediction error correction algorithm, a back propagation method or the like may be used.
The trained identification model created by the learning unit 204 is output to the identification model change unit 106A. The identification model change unit 106A stores the received trained identification model in the storage unit 102. For example, the identification model change unit 106A replaces the identification model stored in the storage unit 102 with the received trained identification model, and stores the replaced trained identification model as a new identification model in the storage unit 102. In this way, the identification model change unit 106A changes the identification model in accordance with the mask pattern identified by the mask identification unit 105A.
The image identification unit 103 reads the new identification model from the storage unit 102, and identifies the computational image captured by the first camera 101 using the read new identification model.
Note that, in the fourth embodiment, a configuration where the learning device 2 is independent from the identification model change unit 106A has been described, but the present disclosure is not limited thereto. The identification model change unit 106A may include the learning device 2.
The corrected image is an image that matches the PSF corresponding to the mask pattern of the mask of the first camera 101. Therefore, in the learning using the corrected image, a new identification model more suitable for the mask pattern of the mask of the first camera 101 is created by the learning processing of the learning device 2. That is, even if the mask pattern of the mask of the first camera 101 is changed, the new identification model more suitable for the changed mask pattern of the mask is created by the learning processing of the learning device 2. Therefore, even if the mask pattern of the mask of the first camera 101 is changed, the image identification unit 103 performs the identification processing using the created new identification model, thus implementing the highly accurate identification processing.
As described above, the image identification system 1A of the fourth embodiment can improve the identification accuracy by training the identification model in accordance with the mask pattern of the mask of the first camera 101. The image identification system 1A of the fourth embodiment is particularly effective for sensing in an environment where privacy needs to be considered. This is because the first camera 101 is different from a normal camera, and it is difficult for a person to recognize a subject even if the person sees the captured image itself.
Note that in each of the above embodiments, each constituent may include dedicated hardware or may be implemented by execution of a software program suitable for each constituent. Each constituent may be implemented by a program execution unit, such as a CPU or a processor, reading and executing a software program recorded in a recording medium such as a hard disk or a semiconductor memory.
Some or all functions of the devices according to the embodiments of the present disclosure are implemented as large scale integration (LSI), which is typically an integrated circuit. These functions may be individually integrated into one chip, or may be integrated into one chip so as to include some or all functions. Circuit integration is not limited to LSI, and may be implemented by a dedicated circuit or a general-purpose processor. A field programmable gate array (FPGA), which can be programmed after manufacturing of LSI, or a reconfigurable processor in which connection and setting of circuit cells inside LSI can be reconfigured may be used.
Some or all functions of the devices according to the embodiments of the present disclosure may be implemented by a processor such as a CPU executing a program.
The numerical figures used above are all illustrated to specifically describe the present disclosure, and the present disclosure is not limited to the illustrated numerical figures.
The order in which each step shown in the above flowcharts is executed is for specifically describing the present disclosure, and may be any order other than the above order as long as a similar effect is obtained. Some of the above steps may be executed simultaneously (in parallel) with other steps.
Since the technique according to the present disclosure can reduce the psychological load on a user while protecting the privacy of a subject, the technique according to the present disclosure is useful as a technology for identifying an image in an environment requiring privacy protection, in particular, in the home or indoors.
Number | Date | Country | Kind |
---|---|---|---|
2021-212421 | Dec 2021 | JP | national |
Number | Date | Country | |
---|---|---|---|
Parent | PCT/JP2022/046813 | Dec 2022 | WO |
Child | 18750039 | US |