The present disclosure relates to an image search device, an image search method, and a storage medium storing an image search program.
There is an image search device that searches for a gallery image in which the same subject as a subject appearing in a query image appears from among a plurality of gallery images (see, for example, Patent Literature 1).
The image search device includes a feature table that holds feature amounts extracted from each of the plurality of gallery images. Further, the image search device includes an image feature extracting unit that extracts a feature amount from a query image, and a feature collating unit that searches for a gallery image in which the same subject as a subject appearing in the query image appears from among the plurality of gallery images by collating each feature amount held in the feature table with the feature amount extracted by the image feature extracting unit.
Patent Literature 1: JP 2015-2547 A
In a case where the camera that has captured the query image and the camera that has captured the gallery image are different types of cameras, a type of a feature amount that can be extracted from the query image and a type of a feature amount that can be extracted from the gallery image may be different. Thus, in the image search device disclosed in Patent Literature 1, in a case where the camera that has captured the query image and the camera that has captured the gallery image are different types of cameras, the image feature extracting unit cannot always extract the same type of feature amount as the feature amount extracted from the gallery image from the query image. Therefore, in the image search device, in a case where the image feature extracting unit cannot extract the same type of feature amount from the query image, there is a problem that the feature collating unit cannot search for the gallery image in which the same subject as the subject appearing in the query image appears.
The present disclosure has been made to solve the above-described problems, and an object of the present disclosure is to obtain an image search device, an image search method, and an image search program capable of searching for a gallery image in which the same subject as a subject appearing in a query image appears, even if a camera that has captured the query image and a camera that has captured the gallery image are different types of cameras.
An image search device according to the present disclosure includes a processor; and a memory storing a program, upon executed by the processor, to perform a process: to acquire image data indicating each of a plurality of images captured by a camera and identification data indicating a type of the camera that has captured each of the images; to specify a type common in types of feature amounts extractable from each of the images on a basis of a type of a camera indicated by each piece of the identification data acquired; to extract a feature amount with respect to the type specified from the image indicated by each piece of the image data acquired; and
According to the present disclosure, even if a camera that has captured a query image and a camera that has captured a gallery image are different types of cameras, it is possible to search for a gallery image in which the same subject as a subject appearing in the query image appears.
Hereinafter, in order to describe the present disclosure in more detail, embodiments for carrying out the present disclosure will be described with reference to the accompanying drawings.
In
The camera 1-n (n=1, . . . , N) captures a person appearing at the installation place as a subject, and outputs image data indicating an image in which the subject appears to the image search device 2.
Further, the camera 1-n outputs identification data indicating the type of the camera to the image search device 2.
Here, the camera 1-n captures a person appearing at the installation place as a subject. However, this is merely an example, and the camera 1-n may capture, for example, a robot appearing at the installation place or an animal appearing at the installation place as a subject.
The image search device 2 includes a data acquiring unit 11, a data holding unit 12, a common type specifying unit 13, a feature amount extracting unit 14, a query image selecting unit 15, and an image search unit 16.
The image search device 2 is a device that searches for a gallery image in which the same subject as a subject appearing in a query image appears.
The man-machine interface unit 3 includes a man-machine interface such as a touch panel.
The man-machine interface unit 3 receives the selection of the query image and outputs a selection signal indicating the selected query image to the image search device 2.
Further, the man-machine interface unit 3 displays the gallery image retrieved by the image search device 2.
The data acquiring unit 11 is implemented by, for example, a data acquiring circuit 21 illustrated in
The data acquiring unit 11 acquires image data indicating an image captured by the camera 1-n (n=1, . . . , N) and identification data indicating the type of the camera 1-n.
The data acquiring unit 11 outputs each of the image data and the identification data to the data holding unit 12.
The data holding unit 12 is implemented by, for example, a data holding circuit 22 illustrated in
The data holding unit 12 holds each of the image data and the identification data output from the data acquiring unit 11.
Further, the data holding unit 12 holds a feature amount extracted by the feature amount extracting unit 14.
The common type specifying unit 13 is implemented by, for example, a common type specifying circuit 23 illustrated in
The common type specifying unit 13 acquires each piece of the identification data acquired by the data acquiring unit 11 from the data holding unit 12.
The common type specifying unit 13 specifies a type common in types of feature amounts extractable from each of the images on the basis of the type of the camera 1-n (n=1, . . . , N) indicated by each piece of the identification data.
The common type specifying unit 13 outputs a specification result of the common type to the feature amount extracting unit 14.
The feature amount extracting unit 14 is implemented by, for example, a feature amount extracting circuit 24 illustrated in
The feature amount extracting unit 14 acquires each piece of the image data acquired by the data acquiring unit 11 from the data holding unit 12.
The feature amount extracting unit 14 extracts a feature amount with respect to the type specified by the common type specifying unit 13 from an image indicated by each piece of the image data.
The feature amount extracting unit 14 outputs each feature amount to the data holding unit 12.
The query image selecting unit 15 is implemented by, for example, the query image selecting circuit 25 illustrated in
The query image selecting unit 15 displays an image indicated by each piece of image data held in the data holding unit 12 on the touch panel of the man-machine interface unit 3.
The query image selecting unit 15 acquires a selection signal indicating a query image whose selection has been received by the man-machine interface unit 3.
The query image selecting unit 15 outputs the selection signal indicating the query image to the image search unit 16.
The image search unit 16 is implemented by, for example, an image search circuit 26 illustrated in
The image search unit 16 acquires a feature amount extracted from the query image indicated by the selection signal output from the query image selecting unit 15 as a feature amount extracted from a query image from among a plurality of feature amounts held in the data holding unit 12.
The image search unit 16 compares the feature amount extracted from the query image indicated by the selection signal with the feature amount of each gallery image held in the data holding unit 12. The gallery image is an image other than the query image among a plurality of images held in the data holding unit 12.
The image search unit 16 searches for a gallery image in which the same subject as a subject appearing in the query image appears on the basis of a comparison result of the feature amounts.
For example, the image search unit 16 outputs the retrieved gallery image to the man-machine interface unit 3.
In
The data holding circuit 22 corresponds to, for example, a nonvolatile or volatile semiconductor memory such as a random access memory (RAM), a read only memory (ROM), a flash memory, an erasable programmable read only memory (EPROM), or an electrically erasable programmable read only memory (EEPROM), a magnetic disk, a flexible disk, an optical disk, a compact disk, a mini disk, or a digital versatile disc (DVD).
Further, each of the data acquiring circuit 21, the common type specifying circuit 23, the feature amount extracting circuit 24, the query image selecting circuit 25, and the image search circuit 26 corresponds to, for example, a single circuit, a composite circuit, a programmed processor, a parallel-programmed processor, an application specific integrated circuit (ASIC), a field-programmable gate array (FPGA), or a combination thereof.
The components of the image search device 2 are not limited to those implemented by dedicated hardware, and the image search device 2 may be implemented by software, firmware, or a combination of software and firmware.
Software or firmware is stored in a memory of the computer as a program. The computer means hardware that executes a program, and corresponds to, for example, a central processing unit (CPU), a central processing device, a processing device, an arithmetic device, a microprocessor, a microcomputer, a processor, or a digital signal processor (DSP).
In a case where the image search device 2 is implemented by software, firmware, or the like, the data holding unit 12 is configured on a memory 41 of the computer. An image search program for causing a computer to execute each processing procedure in the data acquiring unit 11, the common type specifying unit 13, the feature amount extracting unit 14, the query image selecting unit 15, and the image search unit 16 is stored in the memory 41. Then, the processor 42 of the computer executes the image search program stored in the memory 41.
Furthermore,
Next, the operation of the image search device 2 illustrated in
When a person appears at the installation place of the camera 1-n (n=1, . . . , N), the camera 1-n captures the person as a subject.
The camera 1-n outputs image data indicating an image in which the subject appears to the image search device 2.
The N cameras 1-1 to 1-N may be different types of cameras from each other, or may be the same type of cameras.
Examples of the camera 1-n include a visible camera, a depth camera, an infrared camera, and light detection and ranging (LiDAR).
Further, the camera 1-n outputs identification data indicating the type of the camera to the image search device 2.
The identification data is data indicating that the camera 1-n is a visible camera in a case where the camera 1-n is, for example, a visible camera, and indicating that the camera 1-n is a depth camera in a case where the camera 1-n is, for example, a depth camera. Further, the identification data is data indicating that the camera 1-n is an infrared camera in a case where the camera 1-n is, for example, an infrared camera, and indicating that the camera 1-n is a LiDAR in a case where the camera 1-n is, for example, a LiDAR.
When each of the image data and the identification data is output from the camera 1-n (n=1, . . . , N), the data acquiring unit 11 acquires each of the image data and the identification data (step ST1 in
The data acquiring unit 11 outputs each of the image data and the identification data to the data holding unit 12.
The data holding unit 12 holds each of the image data and the identification data output from the data acquiring unit 11.
The common type specifying unit 13 acquires each piece of the identification data acquired by the data acquiring unit 11 from the data holding unit 12.
The common type specifying unit 13 specifies a type common in types of feature amounts extractable from each of the images on the basis of the type of the camera 1-n (n=1, . . . , N) indicated by each piece of the identification data (step ST2 in
The example of
Further, it is indicated that a type of a feature amount that can be extracted from an image captured by the infrared camera is silhouette, texture, or temperature.
It is indicated that a type of a feature amount that can be extracted from an image captured by LiDAR is silhouette or texture.
Therefore, for example, when N=2, if the camera 1-1 is a visible camera and the camera 1-2 is a depth camera, the common type specifying unit 13 specifies that a type common in types of feature amounts extractable from each of the images captured by the respective cameras is “color”, “silhouette”, or “texture”.
For example, when N=3, if the camera 1-1 is a visible camera, the camera 1-2 is a depth camera, and the camera 1-2 is an infrared camera, the common type specifying unit 13 specifies that a type common in types of feature amounts extractable from each of the images captured by the respective cameras is “silhouette” or “texture”.
The common type specifying unit 13 outputs a specification result of the common type to the feature amount extracting unit 14.
The feature amount extracting unit 14 acquires all the image data acquired by the data acquiring unit 11 from the data holding unit 12.
The feature amount extracting unit 14 acquires a specification result of the common type from the common type specifying unit 13.
The feature amount extracting unit 14 selects a feature amount extracting method capable of extracting a feature amount with respect to a type indicated by the specification result from among a plurality of feature amount extracting methods (step ST3 in
In the example of
In a case where the common type is “silhouette” or “texture”, CNN (1), CNN (2), or Histogram of Oriented Gradients (HOG) can be used as the feature amount extracting method.
In a case where the common type is “temperature”, the CNN (2) can be used as the feature amount extracting method.
For example, in a case where the common type is “color”, the CNN (1) is a learning model that learns a feature amount with respect to “color” if image data indicating a color image is given as input data and a feature amount with respect to “color” extracted from the color image is given as training data at the time of learning. In a case where image data indicating a color image is given as input data at the time of inference, the CNN (1) outputs a feature amount with respect to “color” corresponding to the image data.
For example, in a case where the common type is “temperature”, the CNN (2) is a learning model that learns a feature amount with respect to “temperature” if image data indicating a grayscale infrared image is given as input data and a feature amount extracted from the infrared image is given as training data at the time of learning. In a case where image data indicating an infrared image is given as input data at the time of inference, the CNN (2) outputs a feature amount with respect to “temperature” corresponding to the image data.
For example, in a case where the common type is “color”, the feature amount extracting unit 14 selects the CNN (1) from the CNN (1), the CNN (2), and the HOG.
For example, in a case where the common type is “silhouette” or “texture”, the feature amount extracting unit 14 selects one of the CNN (1), the CNN (2), or the HOG from the CNN (1), the CNN (2), and the HOG.
For example, in a case where the common type is “temperature”, the feature amount extracting unit 14 selects the CNN (2) from the CNN (1), the CNN (2), and the HOG.
The feature amount extracting unit 14 extracts a feature amount with respect to the type specified by the common type specifying unit 13 from an image indicated by each piece of the image data using the selected feature amount extracting method (step ST4 in
The feature amount extracting unit 14 outputs each feature amount to the data holding unit 12.
The data holding unit 12 holds each feature amount output from the feature amount extracting unit 14.
In a case where the user desires to search for a gallery image in which the same subject as a subject appearing in a query image appears, the user operates the man-machine interface unit 3 to request for execution of image search processing.
The man-machine interface unit 3 receives the request for execution of the image search processing, and outputs the request for execution of the image search processing to the query image selecting unit 15.
When the request for execution of the image search processing is output from the man-machine interface unit 3, the query image selecting unit 15 outputs all the image data held in the data holding unit 12 to the man-machine interface unit 3.
The man-machine interface unit 3 acquires all image data from the query image selecting unit 15, and displays images indicated by respective pieces of the image data on the display.
The user operates the man-machine interface unit 3 to select a query image from among all the images displayed on the display.
The man-machine interface unit 3 receives the selection of the query image and outputs a selection signal indicating the selected query image to the query image selecting unit 15.
The query image selecting unit 15 outputs the selection signal indicating the query image to the image search unit 16.
The image search unit 16 acquires the selection signal from the query image selecting unit 15.
The image search unit 16 acquires a feature amount extracted from the query image indicated by the selection signal from among a plurality of feature amounts held in the data holding unit 12.
Further, the image search unit 16 acquires a feature amount extracted from each of gallery images which are images other than the query image indicated by the selection signal from among the plurality of feature amounts held in the data holding unit 12.
The image search unit 16 compares the feature amount of the query image with the feature amount of each of the gallery images.
The image search unit 16 searches for a gallery image in which the same subject as a subject appearing in the query image appears on the basis of a comparison result of the feature amounts (step ST5 in
Hereinafter, the search processing for the gallery image by the image search unit 16 will be specifically described.
Here, for convenience of description, it is assumed that the number of gallery images held in the data holding unit 12 is M (M is an integer equal to or more than 1), and a feature amount extracted from a gallery image Gm (m=1, . . . , M) is Fgm. Further, it is assumed that the feature amount extracted from a query image Q is Fq.
As a result of comparison between the feature amount Fq extracted from the query image Q and the feature amount Fgm extracted from the gallery image Gm, the image search unit 16 calculates a similarity Sq,gm between the feature amount Fq extracted from the query image Q and the feature amount Fgm extracted from the gallery image Gm.
Examples of a method of calculating the similarity Sq,gm include a method of calculating a Euclidean distance between the feature amount Fq extracted from the query image Q and the feature amount Fgm extracted from the gallery image Gm, and a method of calculating a cosine similarity between the feature amount Fq extracted from the query image Q and the feature amount Fgm extracted from the gallery image Gm.
The image search unit 16 searches for a gallery image Gj in which the similarity Sq,gm with the feature amount Fq is larger than a threshold Th among M gallery images G1 to GM as a gallery image in which the same subject as a subject appearing in the query image Q appears. The threshold Th may be stored in an internal memory of the image search unit 16 or may be given from the outside of the image search device 2. j=1, . . . , J, and J is an integer equal to or more than 0 and equal to or less than M.
As image data indicating a gallery image in which the same subject as the subject appearing in the query image Q appears, the image search unit 16 outputs image data indicating the gallery image Gj in which the similarity Sq,gm with the feature amount Fq is larger than the threshold Th to a monitoring device (not illustrated) or the like.
Further, the image search unit 16 outputs, to the man-machine interface unit 3, the image data indicating the gallery image Gj in which the similarity Sq,gm with the feature amount Fq is larger than the threshold Th.
The man-machine interface unit 3 displays the gallery image Gj on the display. By displaying the gallery image Gj on the display, it is possible to track a person who is the subject.
In the above-described first embodiment, the image search device 2 is configured to include the data acquiring unit 11 to acquire image data indicating each of a plurality of images captured by the camera 1-n (n=1, . . . , N) and identification data indicating a type of the camera that has captured each of the images, and the common type specifying unit 13 to specify a type common in types of feature amounts extractable from each of the images on the basis of a type of a camera indicated by each piece of the identification data acquired by the data acquiring unit 11. Further, the image search device 2 includes the feature amount extracting unit 14 to extract a feature amount with respect to the type specified by the common type specifying unit 13 from the image indicated by each piece of the image data acquired by the data acquiring unit 11, and the image search unit 16 to compare a feature amount extracted from a query image that is any one image included in the plurality of images among a plurality of feature amounts extracted by the feature amount extracting unit 14 with a feature amount extracted from each of one or more gallery images that are images other than the query image included in the plurality of images, and search for a gallery image in which the same subject as a subject appearing in the query image appears on the basis of a comparison result of the feature amounts. Therefore, even if the camera that has captured the query image and the camera that has captured the gallery image are different types of cameras, the image search device 2 can search for the gallery image in which the same subject as the subject appearing in the query image appears.
In the image search device 2 illustrated in
The data acquiring unit 11 can identify the type of the camera on the basis of, for example, an extension of image data or a data array of the image data. In this case, the data acquiring unit 11 outputs data indicating a type identification result to the data holding unit 12 as identification data.
Examples of the extension of the image data indicating an image captured by the visible camera include “.crw” or “.arw”. Examples of the extension of the image data indicating an image captured by the depth camera include “.heif” or “.heic”. Examples of the extension of the image data indicating an image captured by the infrared camera include “.iri” or “.six”. Examples of the extension of the image data indicating an image captured by LiDAR include “.obj” and “.dxf”.
In a second embodiment, an image search device 2 including a feature amount compressing unit 17 that compresses each feature amount extracted by the feature amount extracting unit 14 will be described.
The image search device 2 illustrated in
The feature amount compressing unit 17 is implemented by, for example, a feature amount compressing circuit 27 illustrated in
The feature amount compressing unit 17 compresses each of feature amounts extracted by the feature amount extracting unit 14, and outputs each of the feature amounts after compression to the data holding unit 12.
The data holding unit 12 holds feature amounts after compression by the feature amount compressing unit 17 instead of holding feature amounts extracted by the feature amount extracting unit 14.
The image search unit 18 is implemented by, for example, an image search circuit 28 illustrated in
The image search unit 18 acquires a feature amount after compression related to a query image indicated by a selection signal output from the query image selecting unit 15 from among a plurality of feature amounts after compression held in the data holding unit 12.
The image search unit 18 compares the feature amount after compression related to the query image indicated by the selection signal with a feature amount after compression related to each of gallery images held in the data holding unit 12.
The image search unit 18 searches for a gallery image in which the same subject as a subject appearing in the query image appears on the basis of a comparison result of the feature amounts after compression.
For example, the image search unit 18 outputs the retrieved gallery image to the man-machine interface unit 3.
In
The data holding circuit 22 corresponds to, for example, a nonvolatile or volatile semiconductor memory such as RAM, ROM, a flash memory, EPROM, or EEPROM, a magnetic disk, a flexible disk, an optical disk, a compact disk, a mini disk, or DVD.
Further, each of the data acquiring circuit 21, the common type specifying circuit 23, the feature amount extracting circuit 24, the feature amount compressing circuit 27, the query image selecting circuit 25, and the image search circuit 28 corresponds to, for example, a single circuit, a composite circuit, a programmed processor, a parallel-programmed processor, ASIC, FPGA, or a combination thereof.
The components of the image search device 2 are not limited to those implemented by dedicated hardware, and the image search device 2 may be implemented by software, firmware, or a combination of software and firmware.
In a case where the image search device 2 is implemented by software, firmware, or the like, the data holding unit 12 is configured on the memory 41 illustrated in
Further,
Next, the operation of the image search device 2 illustrated in
The feature amount compressing unit 17 acquires each feature amount extracted by the feature amount extracting unit 14.
The feature amount compressing unit 17 compresses each feature amount by a compression method such as sparse coding.
The feature amount compressing unit 17 outputs each feature amount after compression to the data holding unit 12.
The data holding unit 12 holds each feature amount after compression by the feature amount compressing unit 17.
The image search unit 18 acquires a selection signal output from the man-machine interface unit 3.
The image search unit 18 acquires a feature amount after compression related to the query image indicated by the selection signal from among a plurality of feature amounts after compression held in the data holding unit 12.
Further, the image search unit 18 acquires a feature amount after compression related to each of gallery images which are images other than the query image indicated by the selection signal from among the plurality of feature amounts after compression held in the data holding unit 12.
The image search unit 18 compares the feature amount after compression related to the query image with the feature amount after compression related to each of the gallery images.
The image search unit 18 searches for a gallery image in which the same subject as a subject appearing in the query image appears on the basis of a comparison result of the feature amounts after compression.
Hereinafter, the search processing of the gallery image by the image search unit 18 will be specifically described.
Here, for convenience of description, it is assumed that the number of gallery images held in the data holding unit 12 is M (M is an integer equal to or more than 1), and a feature amount after compression related to a gallery image Gm (m=1, . . . , M) is CFgm. Further, it is assumed that the feature amount after compression related to the query image Q is CFq.
As a result of comparison between the feature amount CFq after compression related to the query image Q and the feature amount CFgm after compression related to the gallery image Gm, the image search unit 18 calculates a similarity CSq,gm between the feature amount CFq after compression related to the query image Q and the feature amount CFgm after compression related to the gallery image Gm.
Examples of a method of calculating the similarity CSq,gm include a method of calculating a Euclidean distance between the feature amount CFq after compression related to the query image Q and the feature amount CFgm after compression related to the gallery image Gm, and a method of calculating a cosine similarity between the feature amount CFq after compression related to the query image Q and the feature amount CFgm after compression related to the gallery image Gm.
The image search unit 18 searches for the gallery image Gj in which the similarity CSq,gm with the feature amount CFq is larger than the threshold The among M gallery images G1 to GM as a gallery image in which the same subject as a subject appearing in the query image Q appears. The threshold The may be stored in an internal memory of the image search unit 18 or may be given from the outside of the image search device 2. j=1, . . . , J, and J is an integer equal to or more than 0 and equal to or less than M.
As image data indicating a gallery image in which the same subject as the subject appearing in the query image Q appears, the image search unit 18 outputs image data indicating the gallery image Gj in which the similarity CSq,gm with the feature amount CFq is larger than the threshold Thc to a monitoring device (not illustrated) or the like.
Further, the image search unit 18 outputs, to the man-machine interface unit 3, the image data indicating the gallery image Gj in which the similarity CSq,gm with the feature amount CFq is larger than the threshold Thc.
The man-machine interface unit 3 displays the gallery image Gj on the display.
In the above-described second embodiment, the image search device 2 illustrated in
In a third embodiment, an image search device 2 including a data acquiring unit 19 that associates authentication information acquired from an authentication device 4 with an image captured by a camera 1-n will be described.
For example, when an employee identification (ID) card is brought close by a person who appears at a place where the camera 1-n is installed, the authentication device 4 acquires authentication information of the person from the ID card. The person is a subject appearing in an image captured by the camera 1-n.
In the image search device 2 illustrated in
The authentication device 4 outputs the authentication information to the image search device 2.
The image search device 2 illustrated in
In the image search device 2 illustrated in
The data acquiring unit 19 is implemented by, for example, a data acquiring circuit 29 illustrated in
Similarly to the data acquiring unit 11 illustrated in
The data acquiring unit 19 acquires the authentication information from the authentication device 4, and associates the authentication information with the image captured by the camera 1-n.
The data acquiring unit 19 outputs each of the image data and the identification data to the data holding unit 12.
The image search unit 20 is implemented by, for example, an image search circuit 30 illustrated in
Similarly to the image search unit 16 illustrated in
Similarly to the image search unit 16 illustrated in
Similarly to the image search unit 16 illustrated in
Unlike the image search unit 16 illustrated in
For example, the image search unit 20 outputs the retrieved gallery image to the man-machine interface unit 3.
In
The data holding circuit 22 corresponds to, for example, a nonvolatile or volatile semiconductor memory such as RAM, ROM, a flash memory, EPROM, or EEPROM, a magnetic disk, a flexible disk, an optical disk, a compact disk, a mini disk, or DVD.
Further, each of the data acquiring circuit 29, the common type specifying circuit 23, the feature amount extracting circuit 24, the query image selecting circuit 25, and the image search circuit 30 corresponds to, for example, a single circuit, a composite circuit, a programmed processor, a parallel-programmed processor, ASIC, FPGA, or a combination thereof.
The components of the image search device 2 are not limited to those implemented by dedicated hardware, and the image search device 2 may be implemented by software, firmware, or a combination of software and firmware.
In a case where the image search device 2 is implemented by software, firmware, or the like, the data holding unit 12 is configured on the memory 41 of the computer. An image search program for causing a computer to execute processing procedures in the data acquiring unit 19, the common type specifying unit 13, the feature amount extracting unit 14, the query image selecting unit 15, and the image search unit 20 is stored in the memory 41 illustrated in
Further,
Next, the operation of the image search device 2 illustrated in
The authentication device 4 may be installed at a place where each of the N cameras 1-1 to 1-N is installed.
For example, when N cameras 1-1 to 1-N are installed in a company, the authentication device 4 may be installed at an entrance of the company where the camera 1-1 is installed, and the authentication device 4 may be installed at an exit of the company where the camera 1-N is installed. It is assumed that the authentication device 4 is not installed at a place where each of the cameras 1-2 to 1-(N-1) is installed.
In such a case, when the ID card is brought close by the person appearing at the entrance of the company, the authentication device 4 installed at the entrance of the company acquires authentication information IDk of the person from the ID card, and outputs the authentication information IDk to the image search device 2. k is an integer equal to or more than 1. The authentication information IDk is a unique number or the like different for each person.
When the ID card is brought close by the person appearing at the exit of the company, the authentication device 4 installed at the exit of the company acquires the authentication information IDk of the person from the ID card, and outputs the authentication information IDk to the image search device 2.
Similarly to the data acquiring unit 11 illustrated in
In a case where the authentication device 4 is installed at the entrance of the company where the camera 1-1 is installed, the data acquiring unit 19 acquires the authentication information IDk from the authentication device 4 installed at the entrance of the company, and associates the authentication information IDk with an image captured by the camera 1-1.
That is, the data acquiring unit 19 adds the authentication information IDk to the image data indicating the image captured by the camera 1-1.
In a case where the authentication device 4 is installed at the exit of the company where the camera 1-N is installed, the data acquiring unit 19 acquires the authentication information IDk from the authentication device 4 installed at the exit of the company, and associates the authentication information IDk with an image captured by the camera 1-N.
That is, the data acquiring unit 19 adds the authentication information IDk to the image data indicating the image captured by the camera 1-N.
The data acquiring unit 19 outputs the image data indicating the image captured by the camera 1-n (n=1, . . . , N) and the identification data indicating the type of the camera to the data holding unit 12.
Here, the authentication information IDk is added to the image data indicating the image captured by the camera 1-1 and the image data indicating the image captured by the camera 1-N, and the authentication information IDk is not added to the image data indicating the image captured by the camera 1-n (n=2, . . . , N−1).
The data holding unit 12 holds each of the image data and the identification data output from the data acquiring unit 11.
Similarly to the image search unit 16 illustrated in
Similarly to the image search unit 16 illustrated in
Further, similarly to the image search unit 16 illustrated in
Similarly to the image search unit 16 illustrated in
Similarly to the image search unit 16 illustrated in
For example, when the query image is an image captured by the camera 1-1, a subject appearing in the gallery images captured by the cameras 1-2 and 1-N may be the same as the subject appearing in the query image. In such a case, only the gallery image captured by the camera 1-2 may be retrieved as the gallery image in which the same subject as the subject appearing in the query image appears. For example, in a case where the subject wears a coat when captured by the cameras 1-1 and 1-2, but the subject does not wear the coat when captured by the camera 1-N, only the gallery image captured by the camera 1-2 may be retrieved as the gallery image in which the same subject as the subject appearing in the query image appears.
In a case where the authentication information IDk is added to the image data indicating the query image, the image search unit 20 searches for image data in which the same authentication information as the authentication information IDk is added among the image data indicated by one or more gallery images held in the data holding unit 12.
In a case where the authentication information IDk related to the same person is acquired by the authentication device 4 installed at the entrance of the company where the camera 1-1 is installed and the authentication device 4 installed at the exit of the company where the camera 1-N is installed, the gallery image captured by the camera 1-N is retrieved in addition to the gallery image captured by the camera 1-2.
For example, the image search unit 20 outputs the retrieved gallery image to the man-machine interface unit 3.
The man-machine interface unit 3 displays the retrieved gallery images on the display. By displaying the retrieved gallery images on the display, it is possible to track a person who is the subject.
In the above-described third embodiment, the image search device 2 illustrated in
In a fourth embodiment, an image search device 2 will be described in which a data acquiring unit 11′ extracts a region in which a subject appears from an image indicated by each piece of image data and outputs image data of the extracted region.
The image search device 2 includes a data acquiring unit 11′, a data holding unit 12, a common type specifying unit 13, a feature amount extracting unit 14′, a query image selecting unit 15, and an image search unit 16.
The data acquiring unit 11′ is implemented by, for example, a data acquiring circuit 21′ illustrated in
The data acquiring unit 11′ acquires image data indicating an image captured by the camera 1-n (n=1, . . . , N) and identification data indicating the type of the camera.
The data acquiring unit 11′ extracts a region in which the subject appears from the image indicated by each piece of image data.
The data acquiring unit 11′ outputs each of the image data and the identification data of the extracted region to the data holding unit 12.
The feature amount extracting unit 14′ is implemented by, for example, a feature amount extracting circuit 24′ illustrated in
The feature amount extracting unit 14′ acquires each piece of image data acquired by the data acquiring unit 11′ from the data holding unit 12.
The feature amount extracting unit 14′ extracts a feature amount with respect to the type specified by the common type specifying unit 13 from the image of the region indicated by each piece of image data.
The feature amount extracting unit 14′ outputs each feature amount to the data holding unit 12.
In the image search device 2 illustrated in
In
Each of the data acquiring circuit 21′, the common type specifying circuit 23, the feature amount extracting circuit 24′, the query image selecting circuit 25, and the image search circuit 26 corresponds to, for example, a single circuit, a composite circuit, a programmed processor, a parallel-programmed processor, ASIC, FPGA, or a combination thereof.
The components of the image search device 2 are not limited to those implemented by dedicated hardware, and the image search device 2 may be implemented by software, firmware, or a combination of software and firmware.
In a case where the image search device 2 is implemented by software, firmware, or the like, the data holding unit 12 is configured on the memory 41 of the computer. An image search program for causing a computer to execute processing procedures in the data acquiring unit 11′, the common type specifying unit 13, the feature amount extracting unit 14′, the query image selecting unit 15, and the image search unit 16 is stored in the memory 41 illustrated in
Furthermore,
Next, the operation of the image search device 2 illustrated in
When each of the image data and the identification data is output from the camera 1-n (n=1, . . . , N), the data acquiring unit 11′ acquires each of the image data and the identification data.
The data acquiring unit 11′ extracts a region in which the subject appears from the image indicated by each piece of image data.
The shape of the region in which the subject appears is, for example, a rectangle. As a method of extracting a rectangular region, for example, there is a method of extracting a rectangular region using a pre-learned model such as a single shot multibox detector (SSD) in addition to a method of searching a maximum inclusion rectangle from a foreground-extracted region such as a background difference.
By the data acquiring unit 11′ extracting the region in which the subject appears, the feature amount extracted by the feature amount extracting unit 14′ is substantially the feature amount of the subject.
The data acquiring unit 11′ outputs each of the image data and the identification data of the extracted region to the data holding unit 12.
The data holding unit 12 holds each of the image data and the identification data of the region output from the data acquiring unit 11′.
In the image search device 2 illustrated in
The feature amount extracting unit 14′ acquires all the image data acquired by the data acquiring unit 11′ from the data holding unit 12.
The feature amount extracting unit 14′ acquires a specification result of the common type from the common type specifying unit 13.
The feature amount extracting unit 14′ selects a feature amount extracting method capable of extracting a common type of feature amount from a plurality of feature amount extracting methods, similarly to the feature amount extracting unit 14 illustrated in
The feature amount extracting unit 14′ extracts a feature amount from an image indicated by each piece of image data using the selected feature amount extracting method.
The feature amount extracting unit 14′ outputs each feature amount to the data holding unit 12.
The data holding unit 12 holds each feature amount output from the feature amount extracting unit 14′.
The feature amount extracted by the feature amount extracting unit 14′ is generally the feature amount of the subject. On the other hand, the feature amount extracted by the feature amount extracting unit 14 illustrated in
In the above-described fourth embodiment, the image search device 2 illustrated in
Note that, in the present disclosure, free combinations of the embodiments, modifications of any components of the embodiments, or omissions of any components in the embodiments are possible.
The present disclosure is suitable for an image search device, an image search method, and an image search program.
1-1 to 1-N: camera, 2: image search device, 3: man-machine interface unit, 4: authentication device, 11, 11′: data acquiring unit, 12: data holding unit, 13: common type specifying unit, 14, 14′: feature amount extracting unit, 15: query image selecting unit, 16: image search unit, 17: feature amount compressing unit, 18: image search unit, 19: data acquiring unit, 20: image search unit, 21, 21′: data acquiring circuit, 22: data holding circuit, 23: common type specifying circuit, 24, 24′: feature amount extracting circuit, 25: query image selecting circuit, 26: image search circuit, 27: feature amount compressing circuit, 28: image search circuit, 29: data acquiring circuit, 30: image search circuit, 41: memory, 42: processor
This application is a Continuation of PCT International Application No. PCT/JP2022/005773, filed on Feb. 15, 2022, which is hereby expressly incorporated by reference into the present application.
| Number | Date | Country | |
|---|---|---|---|
| Parent | PCT/JP2022/005773 | Feb 2022 | WO |
| Child | 18789070 | US |