The present disclosure relates to a signal processing method and a signal processing apparatus.
Compressed sensing is a technique for generating a larger number of data than observed data by assuming that a data distribution of an observation target is sparse in a certain space (e.g., a frequency space). The compressed sensing is applicable to an imaging system that generates an image including more information from a small number of observed data. In a case where the compressed sensing is applied to an imaging system, an optical filter having a function of coding an image of light in terms of space and wavelength can be used, for example. Such an imaging system can acquire a compressed image by imaging a subject through the optical filter and generate a reconstructed image including more information than the compressed image by computation. This can obtain various effects such as increase of resolution and the number of wavelengths of an image, shortening of an imaging time, and higher sensitivity.
U.S. Pat. No. 9,599,511 (hereinafter referred to as Patent Literature 1) discloses an example of applying a compressed sensing technique to a hyperspectral camera that acquires images of wavelength bands each having a narrow bandwidth. According to the technique disclosed in Patent Literature 1, it is possible to generate a high-resolution and multiwavelength hyperspectral image.
Japanese Unexamined Patent Application Publication No. 2017-208641 (hereinafter referred to as Patent Literature 2) discloses a super-resolution method for generating a high-resolution image from a small number of observation information by using a compressed sensing technique.
U.S. Patent Application Publication No. 2019/0340497 (hereinafter referred to as Patent Literature 3) discloses a method for generating an image of higher resolution than an acquired image by applying convolutional neural network (CNN) to the acquired image.
One non-limiting and exemplary embodiment provides a technique of increasing efficiency and performance of processing for generating a reconstructed image including more information from a compressed image.
In one general aspect, the techniques disclosed here feature a signal processing method executed by using a computer, the method including acquiring a compressed image including compressed information of a subject; acquiring a first parameter group and a reconstruction matrix used in first reconstruction processing for generating a reconstruction target image from the compressed image; generating the reconstruction target image on the basis of the compressed image, the first parameter group, and the reconstruction matrix; acquiring a second parameter group used in second reconstruction processing for generating a reconstructed image from the compressed image; generating the reconstructed image on the basis of the compressed image, the second parameter group, and the reconstruction matrix; and correcting the second parameter group on the basis of the reconstruction target image and the reconstructed image.
According to the technique of the present disclosure, it is possible to increase efficiency and performance of processing for generating a reconstructed image including more information from a compressed image.
It should be noted that general or specific aspects of the present disclosure may be implemented as a system, an apparatus, a method, an integrated circuit, a computer program, a computer-readable storage medium such as a storage disc, or any selective combination thereof. Examples of the computer-readable storage medium may include a volatile storage medium and a non-volatile storage medium such as a compact disc-read only memory (CD-ROM). The apparatus may include one or more apparatuses. In a case where the apparatus includes two or more apparatuses, the two or more apparatuses may be disposed in one piece of equipment or may be separately disposed in two or more separate pieces of equipment. In the specification and claims, the “apparatus” can mean not only a single apparatus, but also a system including apparatuses.
Additional benefits and advantages of the disclosed embodiments will become apparent from the specification and drawings. The benefits and/or advantages may be individually obtained by the various embodiments and features of the specification and drawings, which need not all be provided in order to obtain one or more of such benefits and/or advantages.
An embodiment described below illustrates a general or specific example. Numerical values, shapes, materials, constituent elements, the way in which the constituent elements are disposed and connected, steps, the order of steps, layout of a display screen, and the like in the embodiment below are examples and do not limit the technique of the present disclosure. Among constituent elements in the embodiment below, constituent elements that are not described in independent claims indicating highest concepts are described as optional constituent elements. Each drawing is a schematic view and is not necessarily strict illustration. In each drawing, substantially identical or similar constituent elements are given identical reference signs. Repeated description is sometimes omitted or simplified.
In the present disclosure, all or a part of any of circuit, unit, device, part or portion, or any of functional blocks in the block diagrams may be, for example, implemented as one or more of electronic circuits including a semiconductor device, a semiconductor integrated circuit (IC), or a large scale integration (LSI). The LSI or IC can be integrated into one chip, or also can be a combination of plural chips. For example, functional blocks other than a memory may be integrated into one chip. The name used here is LSI or IC, but it may also be called system LSI, very large scale integration (VLSI), or ultra large scale integration (ULSI) depending on the degree of integration. A Field Programmable Gate Array (FPGA) that can be programmed after manufacturing an LSI or a reconfigurable logic device that allows reconfiguration of the connection or setup of circuit cells inside the LSI can be used for the same purpose.
Further, it is also possible that all or a part of the functions or operations of the circuit, unit, device, part or portion are implemented by executing software. In such a case, the software is recorded on one or more non-transitory recording media such as a ROM, an optical disk or a hard disk drive, and when the software is executed by a processor, the software causes the processor together with peripheral devices to execute the functions specified in the software. A system or apparatus may include such one or more non-transitory recording media on which the software is recorded and a processor together with necessary hardware devices such as an interface.
In the present disclosure, data or a signal representing an image is sometimes referred to simply as an “image”.
Various algorithms can be applied to processing for generating a reconstructed image including more information from a compressed image including less information. For example, various algorithms based on the compressed sensing technique or various algorithms based on machine learning such as deep learning can be used. The individual algorithms have respective unique characteristics. For example, one algorithm enables high-accuracy reconstruction, but requires a larger computation amount and a longer reconstruction processing time. On the other hand, another algorithm enables reconstruction in a short time, but is inferior in reconstruction accuracy.
As an example, a system that inspects a product carried by a carrier device such as a belt conveyor on the basis of a hyperspectral image is discussed here. In such a system, an imaging device including an optical filter such as the one disclosed in Patent Literature 1 can be used. Such an imaging device generates a compressed image by sequentially imaging a product through an optical filter that codes an image of light in terms of wavelength. A reconstructed image (e.g., a hyperspectral image) by applying computation using an algorithm such as compressed sensing or machine learning to the generated compressed image. It is possible to perform inspection as to whether or not a product has an abnormality, whether or not a foreign substance is contained in a product, or the like on the basis of the generated reconstructed image. Such a system requires real-time processing. It is therefore necessary to perform reconstruction processing in a short time by using a high-speed algorithm. However, in such an algorithm, it is typically necessary to set many parameters to appropriate values for accurate reconstruction, and a method for efficiently optimizing the parameters is needed.
The present disclosure is based on the above discussion, and provides a technique for efficiently optimizing one or more parameters in an algorithm for reconstruction processing actually used in a scene such as inspection.
The “compressed image” is an image of a relatively small information amount acquired by imaging. The compressed image can be, for example, image data in which information on wavelength bands is compressed as a single piece of image information but is not limited to this. The compressed image may be image data for generating a Magnetic Resonance Imaging (MRI) image. Alternatively, the compressed image may be image data for generating a high-resolution image.
The “reconstruction target image” is data of an image that is a target of reconstruction. The reconstruction target image is generated from the compressed image by the first reconstruction processing using a first algorithm. As the first algorithm, an algorithm that requires a large computation amount and has high reconstruction performance can be employed, for example. For example, an algorithm that performs reconstruction processing on the basis of compressed sensing can be employed as the first algorithm. The first parameter group is set before the first reconstruction processing is performed. The first parameter group includes one or more parameters. The first parameter group may include parameters or may include a single parameter. A parameter included in the first parameter group is sometimes referred to as a “first parameter”. The first parameter group may be set by a user or may be automatically set by a system. A set value of the first parameter group can be stored in a storage medium such as a memory. In the following description, the reconstruction target image is sometimes referred simply as a “target image”.
The “reconstructed image” is an image generated for a purpose such as inspection or analysis. The reconstructed image is generated from the compressed image by the second reconstruction processing using the second algorithm. As the second algorithm, an algorithm of a smaller computation load than the first algorithm can be employed. For example, an algorithm of a higher speed than the first algorithm or an algorithm that consumes less memory than the first algorithm can be employed as the second algorithm. The second parameter group is set before the second reconstruction processing is performed. The second parameter group includes one or more parameters. The second parameter group may include parameters or may include a single parameter. A parameter included in the second parameter group is sometimes referred to as a “second parameter”. The second parameter group may include a larger number of parameters than the first parameter group. For example, the number of parameters of the second parameter group may be two times as large as the number of parameters of the first parameter group or larger, may be five times as large as the number of parameters of the first parameter group or larger, or may be ten times as large as the number of parameters of the first parameter group or larger. The second parameter group may include, for example, 10 or more parameters, 30 or more parameters, or 50 or more parameters.
The “reconstruction matrix” is matrix data used in the first reconstruction processing and the second reconstruction processing. The reconstruction matrix can be, for example, stored in a storage medium such as a memory in a form such as a table. Therefore, the reconstruction matrix is sometimes referred to as a “reconstruction table”. The reconstruction matrix can be, for example, a matrix reflecting characteristics of an optical filter used in imaging based on compressed sensing.
The correction of the second parameter group in step S16 can include correcting the one or more parameters included in the second parameter group so that the reconstructed image approaches the reconstruction target image. For example, the correction of the second parameter group can include finding an error evaluation value concerning an error between the reconstruction target image and the reconstructed image and correcting the one or more parameters included in the second parameter group so that the error evaluation value is minimized. This makes it possible to tune the second parameter group so that the reconstructed image approaches the reconstruction target image.
The final second parameter group may be decided by repeating the process in step S15 and the process in step S16 plural times. That is, the signal processing method may include correcting the second parameter group on the basis of the reconstruction target image and the reconstructed image and deciding the final second parameter group by repeating generating a reconstructed image by using the corrected second parameter group plural times. This makes it possible to optimize the second parameter group, for example, so that the reconstructed image almost matches the reconstruction target image.
The second reconstruction processing may include processing based on a trained model trained through machine learning such as deep learning. Such processing is high-speed processing and can generate a reconstructed image in a short time. The algorithm based on machine learning is required to optimize a large number of parameters. In the present embodiment, it is possible to efficiently optimize the parameters on the basis of a high-accuracy reconstruction target image.
The first reconstruction processing need not include the processing based on a trained model trained through machine learning. The first reconstruction processing can include, for example, iterative operation for minimizing or maximizing an evaluation function based on the compressed image and the reconstruction matrix. An algorithm that performs such iterative operation can perform high-accuracy reconstruction, but requires a large computation amount and cannot generate a reconstructed image in a short time. Therefore, the first algorithm that performs the first reconstruction processing is not used in an actual environment such as inspection and is used to generate a reconstruction target image that is referred to for correction of the second parameter group in the second algorithm used in an actual environment. It is possible to improve reconstruction performance of the second reconstruction processing by correcting the second parameter group by using a high-accuracy reconstruction target image generated by the first reconstruction processing.
The compressed image may be an image in which spectral information of a subject is coded. In other words, the compressed image may be an image obtained by compressing information on wavelength bands of a subject as a single monochromatic image. In this case, the reconstruction target image and the reconstructed image may each include information on images corresponding to the wavelength bands. Therefore, images of the wavelength bands (e.g., hyperspectral image) can be generated from the compressed image.
The above method may further include displaying, on a display device, a graphical user interface (GUI) for allowing a user to enter the first parameter group. The user can thus set the first parameter group on the GUI.
The above method may further include repeating the correcting the second parameter group on the basis of the reconstruction target image and the reconstructed image and the generating the reconstructed image by using the corrected second parameter group a predetermined number of times unless an end condition is satisfied, calculating an error evaluation value concerning an error between the reconstructed image after the predetermined number of times of the repetition and the reconstruction target image, and displaying, on the display device, a GUI that prompts the user to perform at least one of re-entry of the first parameter group, change of the predetermined number of times, or change of the end condition in a case where the error evaluation value is larger than a threshold value. This allows the user to change a condition for generation of the reconstruction target image in a case where the error evaluation value concerning the error between the reconstructed image and the reconstruction target image does not become equal to or less than the threshold value.
The calculating the error evaluation value may include extracting a first region in the reconstruction target image, extracting a second region corresponding to the first region in the reconstructed image, and deciding the error evaluation value on the basis of a difference between the first region and the second region. The first region and the second region can be, for example, decided on the basis of a region designated by the user. This makes it possible to correct the second parameter group so that an error in the regions extracted from the reconstructed image and the reconstruction target image becomes small.
A signal processing apparatus according to another embodiment of the present disclosure includes one or more processors and a memory in which a computer program to be executed by the one or more processors is stored. The computer program causes the one or more processors to execute the signal processing method described above. That is, the computer program causes the one or more processors to execute (a) acquiring a compressed image including compressed information of a subject, (b) acquiring a first parameter group and a reconstruction matrix used in first reconstruction processing for generating a reconstruction target image from the compressed image, (c) generating the reconstruction target image on the basis of the compressed image, the first parameter group, and the reconstruction matrix, (d) acquiring a second parameter group used in second reconstruction processing for generating a reconstructed image from the compressed image, (e) generating the reconstructed image on the basis of the compressed image, the second parameter group, and the reconstruction matrix, and (f) correcting the second parameter group on the basis of the reconstruction target image and the reconstructed image.
According to the above configuration, the second parameter for generating a reconstructed image can be properly corrected on the basis of the reconstruction target image generated by the first reconstruction processing and the reconstructed image generated by the second reconstruction processing.
Next, an example of a configuration of an imaging system that can be used in an exemplary embodiment of the present disclosure is described.
In
The filter array 110 according to the present embodiment is an array of light-transmitting filters that are arranged in rows and columns. The filters include kinds of filters that are different from each other in spectral transmittance, that is, wavelength dependence of light transmittance. The filter array 110 outputs incident light after modulating an intensity of the incident light for each wavelength. This process using the filter array 110 is referred to as “coding”, and the filter array 110 is sometimes called a “coding element” or a “coding mask”.
In the example illustrated in
The optical system 140 includes at least one lens. Although the optical system 140 is illustrated as a single lens in
The filter array 110 may be disposed away from the image sensor 160.
The image sensor 160 is a monochromatic photodetector that has photodetection elements (hereinafter also referred to as “pixels”) that are arranged within a two-dimensional plane. The image sensor 160 can be, for example, a charge-coupled device (CCD), a complementary metal oxide semiconductor (CMOS) sensor, or an infrared array sensor. Each of the photodetection elements includes, for example, a photodiode. The image sensor 160 need not necessarily be a monochromatic sensor. For example, a color-type sensor including red (R)/green (G)/blue (B) filters, R/G/B/infrared (IR) filters, or R/G/B/transparent (W) filters may be used. Use of a color-type sensor can increase an amount of information concerning a wavelength and can improve accuracy of reconstruction of the hyperspectral image 20. A wavelength range to be acquired may be any wavelength range, and is not limited to a visible wavelength range and may be a wavelength range such as an ultraviolet wavelength range, a near-infrared wavelength range, a mid-infrared wavelength range, or a far-infrared wavelength range.
The processing apparatus 200 can be a computer including one or more processors and one or more storage media such as a memory. The processing apparatus 200 generates data of the reconstructed images 20W1, 20W2, . . . , and 20WN on the basis of the compressed image 10 acquired by the image sensor 160. The processing apparatus 200 may be incorporated into the imaging device 100.
In the example illustrated in
In the example illustrated in
In the example illustrated in
A certain cell among all cells, for example, a half of all the cells may be replaced with a transparent region. Such a transparent region allows transmission of light of all of the wavelength bands W1 to WN included in the target wavelength region W at equally high transmittance, for example, transmittance of 80% or more. In such a configuration, transparent regions can be, for example, disposed in a checkerboard pattern. That is, a region in which light transmittance varies depending on a wavelength and a transparent region can be alternately arranged in two alignment directions of the regions of the filter array 110.
Data indicative of such a spatial distribution of spectral transmittance of the filter array 110 can be acquired in advance on the basis of design data or actual calibration and stored in a storage medium included in the processing apparatus 200. This data is used for arithmetic processing which will be described later.
The filter array 110 can be, for example, constituted by a multi-layer film, an organic material, a diffraction grating structure, or a microstructure containing a metal. In a case where a multi-layer film is used, for example, a dielectric multi-layer film or a multi-layer film including a metal layer can be used. In this case, the filter array 110 is formed so that at least one of a thickness, a material, and a laminating order of each multi-layer film varies from one cell to another. This can realize spectral characteristics that vary from one cell to another. Use of a multi-layer film can realize sharp rising and falling in spectral transmittance. A configuration using an organic material can be realized by varying contained pigment or dye from one cell to another or laminating different kinds of materials. A configuration using a diffraction grating structure can be realized by providing a diffraction structure having a diffraction pitch or depth that varies from one cell to another. In a case where a microstructure containing a metal is used, the filter array 110 can be produced by utilizing dispersion of light based on a plasmon effect.
Next, an example of signal processing performed by the processing apparatus 200 is described. The processing apparatus 200 generates the multiwavelength hyperspectral image 20 on the basis of the compressed image 10 output from the image sensor 160 and spatial distribution characteristics of transmittance for each wavelength of the filter array 110. The multiwavelength means, for example, a larger number of wavelength regions than RGB (red, green, and blue) three wavelength regions acquired by a general color camera. The number of wavelength regions can be, for example, 4 to approximately 100. The number of wavelength regions is referred to as “the number of bands”. The number of bands may be larger than 100 depending on intended use.
Data to be obtained is data of the hyperspectral image 20, which is expressed as f. The data f is data including image data f1 of an image corresponding to the wavelength band W1, the image data f2 of an image corresponding to the wavelength band W2, . . . , and image data fN of an image corresponding to the wavelength band WN where N is the number of bands. It is assumed here that a lateral direction of the image is an x direction and a longitudinal direction of the image is a y direction, as illustrated in
In the formula (1), g is a one-dimensional vector of n×m rows and 1 column based on the image data g, which is two-dimensional data of n×m pixels.
In the formula (1), f1 is a one-dimensional vector of n×m rows and 1 column based on the image data f1, which is two-dimensional data of n×m pixels, f2 is a one-dimensional vector of n×m rows and 1 column based on the image data f2, which is two-dimensional data of n×m pixels, . . . , and fN is a one-dimensional vector of n×m rows and 1 column based on the image data fN, which is two-dimensional data of n×m pixels.
In the formula (1), f is a one-dimensional vector of n×m×N rows and 1 column. A matrix H represents conversion of performing coding and intensity modulation of components f1, f2, . . . , and fN of f by using different pieces of coding information (also referred to as “mask information”) for the respective wavelength bands and adding results thus obtained. Accordingly, H is a matrix of n×m rows and n×m×N columns. This matrix H is sometimes referred to as a “reconstruction matrix”.
It seems that when g and the matrix H are given, f can be calculated by solving an inverse problem of the formula (1). However, since the number of elements n×m×N of the data f to be obtained is larger than the number of elements n×m of the acquired data g, this problem is an ill-posed problem and cannot be solved. In view of this, the processing apparatus 200 finds a solution by using a method of compressed sensing while utilizing redundancy of the images included in the data f. Specifically, the data f to be obtained is estimated by solving the following formula (2).
In the formula (2), f′ represents the estimated data f. The first term in the parentheses in the above formula represents a difference amount between an estimation result Hf and the acquired data g, that is, a residual term. Although a sum of squares is a residual term in this formula, an absolute value, a square-root of sum of squares, or the like may be a residual term. The second term in the parentheses is a regularization term or a stabilization term. The formula (2) means that f that minimizes a sum of the first term and the second term is found. The function in the parentheses in the formula (2) is called an evaluation function. The processing apparatus 200 can calculate, as the final solution f′, f that minimizes the evaluation function by convergence of solutions by recursive iterative operation.
The first term in the parentheses in the formula (2) means operation of finding a sum of squares of a difference between the acquired data g and Hf obtained by converting f in the estimation process by the matrix H. Φ(f) in the second term is a constraint condition in regularization of f and is a function reflecting sparse information of the estimated data. This function brings an effect of smoothing or stabilizing the estimated data. The regularization term can be, for example, expressed by discrete cosine transform (DCT), wavelet transform, Fourier transform, total variation (TV), or the like of f. For example, in a case where total variation is used, stable estimated data with suppressed influence of noise of the observed data g can be acquired. Sparsity of the target 70 in the space of the regularization term varies depending on texture of the target 70. A regularization term that makes the texture of the target 70 more sparse in the space of the regularization term may be selected. Alternatively, regularization terms may be included in calculation. τ is a weight coefficient. As the weight coefficient τ becomes larger, an amount of reduction of redundant data becomes larger, and a compression rate increases. As the weight coefficient τ becomes smaller, convergence to a solution becomes weaker. The weight coefficient τ is set to such a proper value that f converges to a certain extent and is not excessively compressed.
Note that in the configurations of
Through the above processing, the hyperspectral image 20 can be generated from the compressed image 10 acquired by the image sensor 160.
For example, a system that inspects or analyzes a target carried by a carrier device on the basis of a hyperspectral image can be constructed by using the imaging system described above. In such a system, it is required to perform reconstruction processing for generating a reconstructed image from a compressed image in a short time in order to realize real-time inspection or analysis. However, an existing algorithm based on compressed sensing requires a large calculation amount and sometimes cannot complete the processing in a required short time although the algorithm has high reconstruction performance. On the other hand, there is an existing algorithm that enables high-speed processing, but such an algorithm has a large number of parameters to be set, and optimization of the parameters is difficult or complicated.
Processing for generating a reconstructed image from a compressed image can be performed not only by using an algorithm using compressed sensing based on the above formula (2), but also by using an algorithm using machine learning. Examples of the algorithm using machine learning include an algorithm that generates a reconstructed image by applying a trained model trained by deep learning to a compressed image. A period required for reconstruction processing can be shortened by using such an algorithm. However, optimization of parameters is needed for high-accuracy reconstruction. In view of this, a method for optimizing parameters in a machine learning algorithm on the basis of a reconstruction target image generated by using an algorithm based on compressed sensing can be used. This makes it possible to efficiently optimize the parameters.
The processing apparatus 200 includes one or more processors 210 such as a CPU or a GPU and one or more memories 250. The memory 250 stores therein a computer program to be executed by the processor 210 and various kinds of data generated by the processor 210. The computer program causes the processor 210 to perform the signal processing method illustrated in
In the example illustrated in
The reconstruction target image generation unit 212 generates a reconstruction target image by performing first reconstruction processing based on the compressed image generated by the imaging device 100 and the first parameter group and the reconstruction matrix stored in the memory 250. A first algorithm is used in the first reconstruction processing. The generated reconstruction target image is output to the display device 520 and the parameter optimization unit 216.
The image reconstruction unit 214 generates a reconstructed image by performing second reconstruction processing based on the compressed image, the second parameter group, and the reconstruction matrix. A second algorithm is used in the second reconstruction processing. The generated reconstructed image is output to the display device 520 and the parameter optimization unit 216.
The parameter optimization unit 216 corrects the second parameter group on the basis of the reconstruction target image and the reconstructed image and outputs the corrected second parameter group to the image reconstruction unit 214. The parameter optimization unit 216 corrects the second parameter group, for example, so that a difference between the reconstructed image and the reconstruction target image becomes small. The processing for correcting the second parameter group on the basis of the reconstruction target image and the reconstructed image and the processing for generating the reconstructed image by using the corrected second parameter group are repeated a predetermined number of times, and a final value of the second parameter group is decided. The second parameter group is thus optimized.
In a case where a hyperspectral image is generated from the compressed image, the reconstruction target image and the reconstructed image each include images corresponding to wavelength bands. The wavelength bands can include, for example, four or more bands having a relatively narrow band width such as a band of wavelengths 400 nm to 410 nm and a band of wavelengths 410 nm to 420 nm. The correcting the second parameter group so that the reconstructed image approaches the reconstruction target image can include correcting one or more values corresponding to one or more parameters included in the second parameter group so that the images of the bands of the reconstructed image approach the images of the corresponding bands of the reconstruction target image. For example, the second parameter group can be corrected so that an image of wavelengths 400 nm to 410 nm of the reconstructed image approaches an image of wavelengths 400 nm to 410 nm of the reconstruction target image and an image of wavelengths 410 nm to 420 nm of the reconstructed image approaches an image of wavelengths 410 nm to 420 nm of the reconstruction target image.
The difference between the reconstructed image and the reconstruction target image can be evaluated on the basis of an error evaluation value. The error evaluation value can be, for example, calculated by calculating an error for each pair of images of wavelength bands that correspond on a one-to-one basis between the reconstructed image and the reconstruction target image and summing up or averaging these errors. That is, minimizing the error evaluation value can include minimizing the sum or an average of the errors concerning the respective wavelength bands.
The display device 520 displays the reconstruction target image and the reconstructed image generated in the process of optimizing the second parameter group.
The input device 510 can include a device such as a keyboard or a mouse used by a user to set various setting items such as the first parameter group.
The second reconstruction processing in the example of
In the example of
The second reconstruction processing need not necessarily be performed by a machine learning algorithm. For example, the second reconstruction processing may be performed by an algorithm based on compressed sensing.
In the first reconstruction processing illustrated in
Any compressed sensing algorithm such as Iterative shrinkage/thresholding, Two-step iterative shrinkage/thresholding, or Generalized alternating projection-total variation can be used as the first algorithm in the example of
In step S101, the processor 210 acquires a compressed image generated by the imaging device 100. The processor 210 may acquire the compressed image directly from the imaging device 100 or may acquire the compressed image via a storage medium such as the memory 250.
In step S102, the processor 210 acquires the reconstruction matrix from the memory 250. The reconstruction matrix is generated in advance and is stored in the memory 250. Step S102 may be performed before step S101 or may be performed concurrently with step S101.
In step S103, the processor 210 acquires a value of the first parameter group from the memory 250. In a case where a user sets the first parameter group by using the input device 510, the processor 210 acquires a set value of the first parameter group.
In step S104, the processor 210 generates a reconstruction target image on the basis of the compressed image, the reconstruction matrix, and the first parameter group. This processing corresponds to the first reconstruction processing and is performed by using the first algorithm. In a case where the first algorithm is an algorithm that performs iterative operation based on compressed sensing, the reconstruction target image is generated by performing the iterative operation a preset number of times.
In step S105, it is determined whether or not the reconstruction target image is a desired image. This determination can be performed on the basis of user's operation using the input device 510. For example, the processor 210 causes the generated reconstruction target image and a GUI for allowing the user to check whether or not to employ the reconstruction target image to be displayed on the display device 520, and it may be determined that the reconstruction target image is a desired image in a case where the user approves employment of the reconstruction target image.
Through the above processing, generation of a reconstruction target image by the first reconstruction processing is completed. Subsequently, a reconstructed image is generated by the second reconstruction processing.
In step S106, the processor 210 sets the second parameter group to an initial value stored in the memory 250.
In step S107, the processor 210 generates a reconstructed image on the basis of the compressed image, the reconstruction matrix, and the second parameter group. This processing corresponds to the second reconstruction processing and is performed by using the second algorithm. In a case where the second algorithm is an algorithm that performs iterative operation based on compressed sensing, a reconstructed image is generated by performing iterative operation a preset number of times.
In step S108, the processor 210 evaluates an error by comparing the reconstructed image with the reconstruction target image. For example, the processor 210 decides an error evaluation value by using an error evaluation function indicative of a difference between the reconstructed image and the reconstruction target image. For example, Mean Squared Error (MSE) can be used as the error evaluation function. The MSE is expressed by the following formula (3).
In the formula (3), n and m represents the number of pixels in a vertical direction and the number of pixels in a horizontal direction in an image, respectively, fi,j represent pixel values of i rows and j columns of a correct image, and Ii,j represent pixel values of i rows and j columns of an estimated reconstructed image. The error can be expressed not only by the MSE, but also by another error evaluation index such as Root MSE (RMSE), Peak-Signal-to-Noise-Ratio (PSNR), Mean Absolute Error (MAE), Structural Similarity (SSMI), or a spectral angle.
In the present embodiment, the reconstruction target image and the reconstructed image each includes images corresponding to wavelengths bands. For example, images such as an image corresponding to a band of wavelengths 400 nm to 410 nm and an image corresponding to a band of wavelengths 410 nm to 420 nm can be generated in the first reconstruction processing and the second reconstruction processing. In such a case, an error evaluation function such as the MSE can be calculated for each pair of corresponding bands between the reconstructed image and the reconstruction target image. The error evaluation value can be decided by summing up or averaging values of error evaluation functions calculated for the respective bands.
The reconstruction target images may include a first reconstruction target image corresponding to the wavelength band W1, a second reconstruction target image corresponding to the wavelength band W2, . . . , and an N-th reconstruction target image corresponding to the wavelength band WN.
The reconstructed images may include a first reconstructed image corresponding to the wavelength band W1, a second reconstructed image corresponding to the wavelength band W2, . . . , and an N-th reconstructed image corresponding to the wavelength band WN.
Error evaluation values concerning errors between the reconstruction target images and the reconstructed images may be decided on the basis of a first error evaluation value concerning an error between the first reconstruction target image and the first reconstructed image, a second error evaluation value concerning an error between the second reconstruction target image and the second reconstructed image, . . . , and an N-th error evaluation value concerning an error between the N-th reconstruction target image and the N-th reconstructed image.
The following may be established: (the error evaluation values concerning errors between the reconstruction target images and the reconstructed images)={(the first error evaluation value)+(the second error evaluation value)+ . . . +(N-th error evaluation value)}.
The following may be established: (the error evaluation values concerning errors between the reconstruction target images and the reconstructed images)={(the first error evaluation value)+(the second error evaluation value)+ . . . +(the N-th error evaluation value)}/N.
The first error evaluation value may be an MSE between the first reconstruction target image and the first reconstructed image, the second error evaluation value may be an MSE between the second reconstruction target image and the second reconstructed image, . . . , and the N-th error evaluation value may be an MSE between the N-th reconstruction target image and the N-th reconstructed image.
In step S109, the processor 210 updates the second parameter group so that an error between the reconstructed image and the reconstruction target image becomes small. For example, the second parameter group can be updated so that the error evaluation value becomes small by using a method such as a gradient descent method or Bayesian optimization.
In step S110, the processor 210 determines whether or not a preset loop end condition is satisfied. The loop end condition can be, for example, a condition that the optimization loop in steps S107 to S109 has repeated a predetermined number of times, a condition that the error evaluation value between the reconstructed image and the reconstruction target image has become smaller than a threshold value, or the like. In a case where the loop end condition is not satisfied, step S107 is performed again. In a case where the loop end condition is satisfied, the processing ends.
The processor 210 may cause a reconstructed image in a generation process to be displayed on the display device 520 together with the reconstruction target image while performing the processes in steps S107 to S109. This allows a user to check if optimization of the second parameter group has been successfully performed by comparing the reconstructed image and the reconstruction target image.
In the example of
In subsequent step S208, the processor 210 evaluates an error by comparing the reconstructed image with the reconstruction target image. In this process, the processor 210 gives weights to evaluation values for the respective regions so that reconstruction accuracy of the designated region improves. For example, in the example of
evaluation value=a×the evaluation value of the region A+b×the evaluation value of the region B
As a result, an error in the region B does not affect the evaluation value much, and an error in the region A markedly affects the evaluation value. The coefficient a can be set to a value larger than the coefficient b, for example, to a value 1.5, 2, or more times larger than the coefficient b. By such processing, reconstruction placing priority on reconstruction accuracy in a region where an important subject is present is performed.
In the example of
As described above, in the present modification, the processor 210 repeats the correcting the second parameter group on the basis of the reconstruction target image and the reconstructed image and the generating the reconstructed image by using the corrected second parameter group a predetermined number of times unless the end condition is satisfied. The processor 210 calculates an error evaluation value concerning an error between the reconstructed image and the reconstruction target image after the predetermined number of times of repetition. In a case where the error evaluation value is larger than a threshold value, the processor 210 causes a GUI prompting a user to perform at least one of re-entry of the first parameter group, change of the predetermined number of times, or change of the end condition to be displayed on the display device 520. This makes it possible to generate a more suitable reconstructed image by generating the reconstruction target image again by using the first parameter group having a more appropriate value, changing the predetermined number of times, or changing the end condition in a case where the difference between the reconstructed image and the reconstruction target image is large.
The processor 210 extracts a first region from the reconstruction target image, extracts a second region corresponding to the first region from the reconstructed image, and decides an error evaluation value on the basis of a difference between the first region and the second region. The first region can be, for example, designated by a user. This makes it possible to optimize the second parameter group, for example, so that a reconstructed image whose error in a region of high importance is small is generated.
Although a hyperspectral image is generated from a compressed image in the present embodiment, a range of application of the technique of the present disclosure is not limited to generation of a hyperspectral image. For example, the technique of the present disclosure is applicable to generation of a higher-resolution reconstructed image from a low-resolution compressed image, generation of an MRI image from a compressed image, generation of a three-dimensional image from a compressed image, and the like.
The following technique is disclosed by the above description of the embodiment.
A signal processing method performed by using a computer, the signal processing method including:
According to this configuration, the second parameter group can be corrected properly and efficiently. As a result, quality of a reconstructed image generated by the second reconstruction processing can be improved.
The method according to technique 1, in which
According to this configuration, it is possible to properly and efficiently correct the second parameter group including a larger number of parameters than the first parameter group.
The method according to technique 1 or 2, in which
According to this configuration, it is possible to properly correct the second parameter group and generate a reconstructed image close to a reconstruction target image.
The method according to technique 3, in which
According to this configuration, it is possible to properly correct the second parameter group and generate a reconstructed image having a small error.
The method according to any one of techniques 1 to 4, in which
According to this configuration, it is possible to generate a reconstructed image in a shorter time by using the second algorithm of a smaller computation amount than the first algorithm.
The method according to technique 5, in which
According to this configuration, it is possible to generate a reconstructed image in a shorter time by the second reconstruction processing including processing based on a trained model.
The method according to technique 6, in which
According to this configuration, it is possible to generate a more accurate reconstruction target image, and it is therefore possible to correct the second parameter group to a more appropriate value.
The method according to any one of techniques 1 to 7, in which
According to this configuration, it is possible to properly generate a reconstructed image including information on images corresponding to wavelength bands.
The method according to any one of techniques 1 to 8, further including deciding a final value of the second parameter group by repeating the correcting the second parameter group on the basis of the reconstruction target image and the reconstructed image and generating the reconstructed image by using the corrected second parameter group plural times.
According to this configuration, it is possible to correct the second parameter group to a more appropriate value, and it is therefore possible to improve accuracy of a reconstructed image.
The method according to any one of techniques 1 to 9, further including displaying, on a display device, a graphical user interface (GUI) for allowing a user to enter the first parameter group.
According to this configuration, the user can adjust the first parameter group, and it is therefore possible to generate a more appropriate reconstruction target image. As a result, it is possible to correct the second parameter group to a more appropriate value and improve accuracy of a reconstructed image.
The method according to technique 10, further including:
According to this configuration, it is possible to correct the second parameter group to a more appropriate value. As a result, it is possible to generate a more accurate reconstructed image.
The method according to technique 11, in which
According to this configuration, it is possible to reduce a reconstruction error concerning a subject included in the extracted region.
A signal processing apparatus including:
According to this configuration, it is possible to properly and efficiently correct the second parameter group. As a result, it is possible to improve quality of a reconstructed image generated by the second reconstruction processing.
A modification of the embodiment of the present disclosure may be as follows.
A method according to a first item is a method performed by a computer, and the method includes causing a computer to:
The technique of the present disclosure is useful, for example, for a camera and a measurement device that acquires a multiwavelength or high-resolution image. The technique of the present disclosure is, for example, applicable to sensing for a biological, medical, or cosmetic purpose, a food foreign substance or residual pesticide test system, a remote sensing system, and an on-vehicle sensing system.
| Number | Date | Country | Kind |
|---|---|---|---|
| 2022-134391 | Aug 2022 | JP | national |
| Number | Date | Country | |
|---|---|---|---|
| Parent | PCT/JP2023/027993 | Jul 2023 | WO |
| Child | 19038677 | US |