The present invention relates to an image processing apparatus, an image processing method, and a program, in particular, an image processing apparatus, an image processing method, and a program suitable to a case in which processed images are easily compared when a plurality of different image processings are applied on one moving image.
In a case where a plurality of different image processings are applied on a certain image and the image processing results are mutually compared and reviewed, a method of alternately displaying the plurality of processed images after the image processings on one monitor for comparison is easily considerable. Also, a method of displaying the plurality of processed images on the monitor at the same time for comparison also exists (for example, see Patent Documents 1 and 2).
However, in a case where the plurality of comparison target processed images are still images, the processed images arranged and displayed on the monitor can be thoroughly compared and reviewed, but in a case where the plurality of comparison target processed images are moving images, the images are changed in a moment, and it is therefore difficult to perform the comparison only by simply looking at the processed images arranged and displayed on the monitor.
The present invention has been made in view of the above-mentioned circumstances and is aimed at making it possible to easily compare processed images when a plurality of different image processings are applied on one moving image.
An image processing apparatus according to an aspect of the present invention includes: a plurality of image processing means configured to perform a plurality of different image processings on one input image which is an image constituting a moving image and is sequentially input; and synthesized image generation means configured to generate a synthesized image in which a plurality of processed images which are respectively processed by the plurality of image processing means are synthesized, in which the synthesized image changes in accordance with results of the plurality of image processings.
Each of the plurality of image processing means can include: image quality change processing means configured to change an image quality of the input image into image qualities different in each of the plurality of image processing means; and expansion processing means configured to perform an expansion processing while using a predetermined position of the image after the change processing which is subjected to the change processing by the image quality change processing means as a reference, and control means configured to decide the predetermined position on the basis of change processing results by the image quality change processing means of the plurality of image processing means can be further provided.
The control means can decide a position as the predetermined position where a difference is large when the plurality of images after the change processings are mutually compared.
The synthesized image can be composed of a main image which is the processed image instructed by a user and a plurality of sub images which are the other processed images, and the synthesized image generation means can change an arrangement of the processed image which is set as the main image and the processed images set as the sub images on the basis of an instruction of the user.
The synthesized image generation means can perform a highlight display of the sub image selected by the user.
The plurality of image processing means can perform the plurality of different image processings by using a class classification adaptive processing.
Each of the plurality of image processing means can include: tracking processing means configured to track the predetermined position of the input image in tracking systems different in each of the plurality of image processing means; and expansion processing means configured to perform an expansion processing while using a tracking position which is a result of the tracking processing by the tracking processing means as a reference.
Control means configured to supply the tracking position selected by a user among the plurality of tracking positions as the predetermined position to the tracking processing means can be further provided.
The synthesized image can be composed of a main image which is the processed image instructed by a user and a plurality of sub images which are the other processed images, and the synthesized image generation means can change an arrangement of the sub images in accordance with a ratio of a horizontal component and a vertical component of a tracking difference vector representing a difference between the tracking position of the main image and the tracking position of the sub image.
Detection means configured to detect a time code representing which scene of the moving image the input image is on the basis of characteristic amounts of the plurality of input images and the same number of decision means configured to decide the predetermined positions as the number of the image processing means can be further provided, in which the decision means can store the predetermined position in the input image while corresponding to the time code and decide the different predetermined positions by each of the plurality of decision means corresponding to the detected time code, and each of the plurality of image processing means can include expansion processing means configured to perform an expansion processing while using the predetermined position decided by the decision means as a reference.
The plurality of expansion processing means can include the expansion processing means configured to perform a high image quality expansion processing and the expansion processing means configured to perform a low image quality expansion processing, and control means can be further provided which is configured to control a supply of an expanded image selected by a user among a plurality of expanded images subjected to the expansion processing by each of the plurality of expansion processing means to the expansion processing means at the predetermined position decided by the decision means so as to be processed by the expansion processing means configured to perform the high image quality expansion processing.
The synthesized image can be composed of a main image which is the processed image instructed by the user and a plurality of sub images which are the other processed images, and the synthesized image generation means can change an arrangement of the processed images so as to set the expanded image processed by the expansion processing means configured to perform the high image quality expansion processing as the main image.
The synthesized image generation means can calculate correlation values between the expanded image of the main image and the expanded images of the sub images and change an arrangement of the plurality of sub images in a descending order.
An image processing method according to an aspect of the present invention includes: performing a plurality of different image processings on one input image which is an image constituting a moving image and is sequentially input; and generating a synthesized image in which a plurality of processed images obtained as a result of being subjected to the image processings are synthesized, in which the synthesized image changes in accordance with results of the plurality of image processings.
A program according to an aspect of the present invention causes a computer to execute a processing including: performing a plurality of different image processings on one input image which is an image constituting a moving image and is sequentially input; and generating a synthesized image in which a plurality of processed images obtained as a result of being subjected to the image processings are synthesized, in which the synthesized image changes in accordance with results of the plurality of image processings.
According to an aspect of the present invention, the plurality of different image processings are performed on one input image which is the image constituting the moving image and is sequentially input, and the synthesized image is generated in which the plurality of processed images obtained as the result of being subjected to the image processings are synthesized.
According to an aspect of the present invention, it is possible to readily compare the processed images when the plurality of different image processings are applied on one moving image.
200 IMAGE PROCESSING APPARATUS, 213-1 TO 213-3 IMAGE PROCESSING UNIT, 214 IMAGE SYNTHESIS UNIT, 217 CONTROL UNIT, 241A TO 241C IMAGE QUALITY CHANGE PROCESSING UNIT, 242A TO 242C EXPANSION PROCESSING UNIT, 262 IMAGE COMPARISON UNIT, 341A TO 341C TRACKING PROCESSING UNIT, 242A′ TO 242C′ EXPANSION PROCESSING UNIT, 471 SYNCHRONIZATION CHARACTERISTIC AMOUNT EXTRACTION UNIT, 472A TO 472C SEQUENCE REPRODUCTION UNIT, 242A″ TO 242C″ EXPANSION PROCESSING UNIT, 473 SWITCHER UNIT
Embodiments of an image processing apparatus (system) to which the present invention is applied will be described, but beforehand, a class classification adaptive processing utilized for a signal processing performed by the image processing apparatus will be described. It should be noted that the class classification adaptive processing is an example of a processing utilized for the signal processing performed by the image processing apparatus, and the signal processing performed by the image processing apparatus may also be one not utilizing the class classification adaptive processing.
Also, herein, the class classification adaptive processing will be described while using an image conversion processing for converting first image data (image signal) into the second image data (image signal) as an example.
The image conversion processing for converting the first image data into the second image data becomes various signal processings depending on a definition of the first and second image data.
That is, for example, when the first image data is set as the image data with a low spatial resolution and also the second image data is set as the image data with a high spatial resolution, the image conversion processing can be mentioned as a spatial resolution creation (improvement) processing for improving the spatial resolution.
Also, for example, when the first image data is set as the image data with a low S/N (Signal/Noise) and also the second image data is set as the image data with a high S/N, the image conversion processing can be mentioned as a noise removal processing for removing the noise.
Furthermore, for example, when the first image data is set as the image data with a predetermined pixel number (size) and also the second image data is set as the image data with a pixel number more or fewer than the first image data, the image conversion processing can be mentioned as a resize processing for performing the resize (expansion or reduction) of the image.
Also, for example, when the first image data is set as the image data with a low temporal resolution and also the second image data is set as the image data with a high temporal resolution, the image conversion processing can be mentioned as a temporal resolution creation (improvement) processing for improving the temporal resolution.
Furthermore, for example, when the first image data is set as decoded image data obtained by decoding image data coded in units of block such through an MPEG (Moving Picture Experts Group) coding and also when the second image data is set as the image data before the coding, the image conversion processing can be mentioned as a distortion removal processing for removing various distortions such as a block distortion generated by the MPEG coding and decoding.
It should be noted that in the spatial resolution creation processing, when the first image data which is the image data with the low spatial resolution is converted into the second image data which is the image data with the high spatial resolution, the second image data can be set as the image data with the same pixel number as the first image data and can also be the image data with a pixel number larger than the first image data. In a case where the second image data is set as the image data with the pixel number larger than the first image data, the spatial resolution creation processing is a processing for improving the spatial resolution and also a resize processing for expanding the image size (pixel number).
As described above, depending on the image conversion processing, various signal processings can be realized depending on how to define the first and second image data.
In the class classification adaptive processing serving as the above-described image conversion processing, through a computation using a tap coefficient of a class obtained by performing a class classification on (a pixel value of) an attention pixel which attracts an attention among the second image data into one of a class among a plurality of classes and (a pixel value of) a pixel of the first image data selected with respect to the attention pixel, (the pixel value of) the attention pixel is obtained.
That is,
In the image conversion apparatus 1, image data supplied thereto is supplied as the first image data to tap selection unit 12 and 13.
An attention pixel selection unit 11 sequentially sets a pixel constituting the second image data as an attention pixel and supplies information representing the attention pixel to a necessary block.
The tap selection unit 12 selects some of (pixel values of) pixels constituting the first image data used for predicting (the pixel value of) the attention pixel as a prediction tap.
Specifically, the tap selection unit 12 selects a plurality of pixels of the first image data at positions spatially or temporarily close to a position in the temporal space of the attention pixel as the prediction tap.
The tap selection unit 13 selects some of pixels constituting the first image data used for performing the class classification for classifying the attention pixel into one of some classes as a class tap. That is, similarly as the tap selection unit 12 selects the prediction tap, the tap selection unit 13 selects the class tap.
It should be noted that the prediction tap and the class tap may have the same tap structure or may also have different tap structures.
The prediction tap obtained in the tap selection unit 12 is supplied to a prediction computation unit 16, and the class tap obtained in the tap selection unit 13 is supplied to a class classification unit 14.
On the basis of the class tap from the tap selection unit 13, the class classification unit 14 performs the class classification of the attention pixel and supplies a class code corresponding to the class obtained as the result to a coefficient output unit 15.
Herein, as a method of performing the class classification, for example, ADRC (Adaptive Dynamic Range Coding) or the like can be adopted.
In the method using the ADRC, (the pixel values of) the pixels constituting the class tap are subjected to an ADRC processing, and the class of the attention pixel is decided while following an ADRC code obtained as the result.
It should be noted that in a K-bit ADRC, for example, the maximum value MAX and a minimum value MIN of the pixel value for the pixel constituting the class tap are detected, and DR=MAX−MIN is set as a local dynamic range of an aggregate. On the basis of this dynamic range DR, the pixel values of the respective pixels constituting the class tap are re-quantized into K bits. That is, from the pixel values of the respective pixels constituting the class tap, the minimum value MIN is subtracted, and the subtracted value is divided by DR/2K (re-quantized). Then, a bit sequence in which the pixel values in the K bits of the respective pixels constituting the class tap obtained in the above-mentioned manner are arranged in a predetermined order is output as the ADRC code. Therefore, in a case where the class tap is subjected, for example, to a 1-bit ADRC processing, the pixel values of the respective pixels constituting the class tap is divided by an average value of the maximum value MAX and the minimum value MIN (cutting of a fractional part), and according to this, the pixel values of the respective pixels are set in 1 bit (binarized). Then, a bit sequence in which the pixel values in 1-bit are arranged in a predetermined order is output as the ADRC code.
It should be noted that it is also possible to cause the class classification unit 14 to output a pattern of the pixel values of, for example, the pixels constituting the class tap as the class code as it is. However, in this case, when the class tap is constructed by the pixel values of N pixels, and the pixel values of the respective pixels are allocated with K bits, the number in the case of the class code output by the class classification unit 14 is (2N)K patterns, which is an enormous number exponentially in proportion to the bit number K of the pixel values of the pixels.
Therefore, in the class classification unit 14, the class classification is preferably performed while the information amount of the class tap is compressed by the above-mentioned ADRC processing, a vector quantization, or the like.
The coefficient output unit 15 stores the tap coefficients for each class obtained through a learning which will be described below and further outputs the tap coefficient (the tap coefficient of the class represented by the class code supplied from the class classification unit 14) stored in an address corresponding to the class code supplied from the class classification unit 14 among the stored tap coefficients. This tap coefficient is supplied to the prediction computation unit 16.
Herein, the tap coefficient is comparable to a coefficient multiplied with input data in a so-called tap in a digital filter.
The prediction computation unit 16 obtains the prediction tap output by the tap selection unit 12 and the tap coefficient output by the coefficient output unit 15 and uses the prediction tap and the tap coefficient to perform a predetermined prediction computation for obtaining a predicted value of a true value of the attention pixel. According to this, the prediction computation unit 16 obtains and outputs (a predicted values of) the pixel value of the attention pixel, that is, the pixel value of the pixel constituting the second image data.
Next, with reference to a flow chart of
In step S11, the attention pixel selection unit 11 selects one which is not set as the attention pixel yet as the attention pixel among the pixels constructing the second image data with respect to the first image data input to the image conversion apparatus 1. For example, among the pixels constructing the second image data, in a raster scan order, one which is not set as the attention pixel yet is selected as the attention pixel.
In step S12, the tap selection unit 12 and the tap selection unit 13 respectively select the prediction tap and the class tap regarding the attention pixel from the first image data supplied thereto. Then, the prediction tap is supplied from the tap selection unit 12 to the prediction computation unit 16, and the class tap is supplied from the tap selection unit 13 to the class classification unit 14.
The class classification unit 14 receives the class tap regarding the attention pixel from the tap selection unit 13, and in step S13, performs the class classification of the attention pixel on the basis of the class tap. Furthermore, the class classification unit 14 outputs a class code representing the class of the attention pixel obtained as a result of the class classification to the coefficient output unit 15.
In step S14, the coefficient output unit 15 obtains and outputs the tap coefficient stored in an address corresponding to the class code supplied from the class classification unit 14. Furthermore, in step S14, the prediction computation unit 16 obtains the tap coefficient output by the coefficient output unit 15.
In step S15, the prediction computation unit 16 uses the prediction tap output by the tap selection unit 12 and the tap coefficient obtained from the coefficient output unit 15 to perform a predetermined prediction computation. According to this, the prediction computation unit 16 obtains and outputs the pixel value of the attention pixel.
In step S16, the attention pixel selection unit 11 determines whether or not the second image data which is not set as the attention pixel yet exists. In step S16, in a case where it is determined that the second image data which is not set as the attention pixel yet exists, the processing returns to step S11, and afterward the similar processing is repeatedly performed.
Also, in step S16, in a case where it is determined that the second image data which is not set as the attention pixel yet does not exist, the processing is ended.
Next, the prediction computation in the prediction computation unit 16 of
Now, a case is considered, for example, in which the image data with a high image quality (high image quality image data) is set as the second image data, and also the image data with a low image quality (low image quality image data) whose image quality (resolution) is decreased while the high image quality image data is subjected to filtering by an LPF (Low Pass Filter) or the like is set as the first image data to select the prediction tap from the low image quality image data, and by using the prediction tap and the tap coefficient, a pixel value of the pixel in the high image quality image data (high image quality pixel) is obtained (predicted) through a predetermined prediction computation.
As the predetermined prediction computation, for example, when a linear first-order prediction computation is adopted, the pixel value y of the high image quality pixel is obtained through the next linear primary expression.
Where, in Expression (1), xn denotes a pixel value of the pixel in the n-th low image quality image data constituting the prediction tap regarding the high image quality pixel y (hereinafter, which is appropriately referred to as low image quality pixel), and wn denotes n-th tap coefficient to be multiplied with (the pixel value of) the n-th low image quality pixel. It should be noted that in Expression (1), the prediction tap is constituted by N low image quality pixel x1, x2, . . . , xN.
Herein, the pixel value y of the high image quality pixel can also be obtained through a higher-order expression equal to or higher than second-order instead of the linear primary expression shown in Expression (1).
Now, when a true value of the pixel value of the k-th sample high image quality pixel is denoted by yk and also a predicted value of the true value yk obtained through Expression (1) is denoted by yk′, a prediction error ek thereof is represented by the following expression.
[Expression 2]
e
k
=y
k
−y
k′ (2)
Now, as the predicted value yk′ in Expression (2) is obtained while following Expression (1), when yk′ in Expression (2) is replaced while following Expression (1), the following expression is obtained.
It should be however that in Expression (3), xn, k denotes n-th the low image quality pixel constituting the prediction tap regarding the k-th sample high image quality.
A tap coefficient wn where the prediction error ek is 0 in Expression (3) (or Expression (2)) becomes an optimal one for predicting the high image quality pixel, and regarding all the high image quality pixels, it is generally difficult to obtain such tap coefficient wn.
In view of the above, if the method of least squares is adopted as a rule representing that the tap coefficient wn is the optimal one, for example, the optimal tap coefficient wn can be obtained by minimizing a total sum E for the square errors represented by the following expression.
It should be however that in Expression (4), K denotes the number of samples of sets of the high image quality pixel yk and the low image quality pixels x1, k, x2, k, . . . , xN, k constituting the prediction tap regarding the high image quality pixel yk (the number of learning material samples).
The minimum value (local minimal value) of the total sum E for the square errors in Expression (4) is obtained by wn in which, as shown in Expression (5), the total sum E is subjected to partial differentiation by the tap coefficient wn is set as 0.
In view of the above, when the above-mentioned Expression (3) is subjected to partial differentiation by the tap coefficient wn, the following expression is obtained.
From Expressions (5) and (6), the following expression is obtained.
As ek in Expression (7) is assigned to Expression (3), Expression (7) can be represented by a normal equation shown in Expression (8).
The normal equation of Expression (8) can solve the tap coefficient wn, for example, by using the discharge method (Gauss-Jordan elimination method) or the like.
As the normal equation of Expression (8) is established for each class and solved, the optimal tap coefficient (herein, the coefficient for minimizing the total sum E for the square errors) wn can be obtained for each class.
A learning material image storage unit 31 stores learning material image data used for leaning the tap coefficient wn. Herein, for the learning material image data, for example, the high image quality image data with a high resolution can be used.
A teacher data generation unit 32 reads out the learning material image data from the learning material image storage unit 31. Furthermore, the teacher data generation unit 32 generates the teacher for the learning on the tap coefficient (true value), that is, teacher data which is a pixel value of a mapping destination of mapping serving as a prediction computation through Expression (1) from the learning material image data to be supplied to a teacher data storage unit 33. Herein, the teacher data generation unit 32 supplies, for example, the high image quality image data serving as the learning material image data to the teacher data storage unit 33 as the teacher data as it is.
The teacher data storage unit 33 stores the high image quality image data serving as the teacher data supplied from the teacher data generation unit 32.
A student data generation unit 34 reads out the learning material image data from the learning material image storage unit 31. Furthermore, the student data generation unit 34 generates a student for the learning on the tap coefficient, that is, student data which is a pixel value of a conversion target by the mapping serving as a prediction computation through Expression (1) from the learning material image data to be supplied to a student data storage unit 174. Herein, the student data generation unit 34 performs, for example, a filtering on the high image quality image data serving as the learning material image data for decreasing the resolution to generate the low image quality image data and supplies this low image quality image data as the student data to the student data storage unit 35.
The student data storage unit 35 stores the student data supplied from the student data generation unit 34.
A learning unit 36 sequentially sets pixels constituting the high image quality image data serving as the teacher data stored in the teacher data storage unit 33 as an attention pixel and regarding the attention pixel, selects the low image quality pixel with the same tap structure as one selected by the tap selection unit 12 of
That is,
An attention pixel selection unit 41 sequentially selects the pixels constituting the teacher data stored in the teacher data storage unit 33 as the attention pixel and supplies information representing the attention pixel to a necessary block.
A tap selection unit 42 selects, regarding the attention pixel, the same pixel as the one selected by the tap selection unit 12 of
A tap selection unit 43 selects, regarding the attention pixel, the same pixel as the one selected by the tap selection unit 13 of
On the basis of the class tap output by the tap selection unit 43, the class classification unit 44 performs the same class classification as the class classification unit 14 of
The supplement unit 45 reads out the teacher data (pixel) which becomes the attention pixel from the teacher data storage unit 33 and performs supplement in which the attention pixel and the student data (pixel) constituting the prediction tap supplied from the tap selection unit 42 regarding the attention pixel are set as the targets for each class code supplied from the class classification unit 44.
That is, the supplement unit 45 is supplied with the teacher data yk stored in the teacher data storage unit 33, the prediction tap xn, k output by the tap selection unit 42, and the class code output by the class classification unit 44.
Then, for each class corresponding to the class code supplied from the class classification unit 44, the supplement unit 45 uses the prediction tap (student data) xn, k to perform a multiplication of mutual student data in a matrix on the left side in Expression (8) (xn, kxn′, k) and a computation comparable to the summation (Σ).
Furthermore, for each class corresponding to the class code supplied from the class classification unit 44, the supplement unit 45 all the same uses the prediction tap (student data) xn, k and the teacher data yk to perform a multiplication of the student data xn, k in the vector on the right side in Expression (8) and the teacher data yk (xn, kyk) and a computation comparable to the summation (Σ).
That is, the supplement unit 45 stores the component (Σxn, kxn′, k) in a matrix on the left side in Expression (8) and the component (Σxn, kyk) in a vector on the right side obtained regarding the teacher data set as the attention pixel in the previous time in a built-in memory thereof (not shown) and with respect to the component (Σxn, kxn′, k) in the matrix or the component (Σxn, kyk) in the vector, regarding the teacher data newly set as the attention pixel, supplements the corresponding component xn, k+1xn′, k+1 or xn, k+1yk+1 calculated by using the teacher data yk+1 and the student data xn, k+1 (performs the addition represented by the summation in Expression (8)).
Then, the supplement unit 45 sets all the teacher data stored in the teacher data storage unit 33 (
The tap coefficient calculation unit 46 obtains and outputs the optimal tap coefficient wn for the respective classes by solving the normal equation for the respective classes supplied from the supplement unit 45.
The coefficient output unit 15 in the image conversion apparatus 1 of
Herein, depending on a manner of selecting the image data serving as the student data corresponding to the first image data and the image data serving as the teacher data corresponding to the second image data, for the tap coefficient, as described above, one for performing the various image conversion processings can be obtained.
That is, as described above, by performing the learning on the tap coefficient while the high image quality image data is set as the teacher data corresponding to the second image data and also the low image quality image data in which the spatial resolution of the high image quality image data is degraded is set as the student data corresponding to the first image data, for the tap coefficient, as shown in the first from the top of
It should be noted that in this case, the pixel number of the first image data (student data) may be the same as or may be smaller than that of the second image data (teacher data).
Also, for example, by performing the learning on the tap coefficient while the high image quality image data is set as the teacher data and also the image data where noise is overlapped with respect to the high image quality image data serving as the teacher data is set as the student data, for the tap coefficient, as shown in the second from the top of
Furthermore, for example, by performing the learning on the tap coefficient while certain image data is set as the teacher data and also the image data where the pixel number of the image data serving as the teacher data is thinned out is set as the student data, for the tap coefficient, as shown in the third from the top of
It should be noted that the tap coefficient for performing the expansion processing can also be obtained by the learning on the tap coefficient while the high image quality image data is set as the teacher data and also the low image quality image data in which the spatial resolution of the high image quality image data is degraded by thinning out the pixel number is set as the student data.
Also, for example, by performing the learning on the tap coefficient while the image data with a high frame rate is set as the teacher data and also the image data in which the frames of the image data with the high frame rate serving as the teacher data are thinned out is set as the student data, for the tap coefficient, as shown in the fourth from the top of
Next, with reference to a flow chart of
First, in step S21, the teacher data generation unit 32 and the student data generation unit 34 generate the teacher data and the student data from the learning material image data stored in the learning material image storage unit 31 and to be respectively supplied to the teacher data storage unit 33 and the student data generation unit 34 and stored.
It should be noted that in the teacher data generation unit 32 and the student data generation unit 34, respectively, what kinds of student data and teacher data are generated are varied depending on the learning on the tap coefficient performed for one of the processings among the image conversion processings of the above-mentioned types.
After that, the processing proceeds to step S22, and in the learning unit 36 (
In step S24, the class classification unit 44 performs the class classification of the attention pixel on the basis of the class tap regarding the attention pixel and outputs the class code corresponding to the class obtained as the result to the supplement unit 45.
In step S25, the supplement unit 45 reads out the attention pixel from the teacher data storage unit 33 and performs the supplement of Expression (8) for each of the class codes supplied from the class classification unit 44 while the attention pixel and the student data constituting the prediction tap selected with regard to the attention pixel supplied from the tap selection unit 42 are set as the targets.
In step S26, the attention pixel selection unit 41 determines whether or not the teacher data which is not set as the attention pixel is still stored in the teacher data storage unit 33. In step S26, in a case where it is determined that the attention pixel is still stored in the teacher data storage unit 33, the processing returns to step S22, and afterward the similar processing is repeatedly performed.
Also, in step S26, in a case where it is determined that the attention pixel is not stored in the teacher data storage unit 33, the processing proceeds to step S27, and the supplement unit 45 supplies the matrix on the left side and the vector on the right side in Expression (8) for each of the classes obtained through the processing up to the present in steps S22 to S26 to the tap coefficient calculation unit 46.
Furthermore, in step S27, by solving the normal equation for each of the classes constituted by the matrix on the left side and the vector on the right side in Expression (8) supplied from the supplement unit 45 for each of the classes, the tap coefficient calculation unit 46 obtains and outputs the tap coefficient wn for each of the classes, and the processing is ended.
It should be noted that due to a state or the like where the number of the learning material image data is not sufficient, a class may be generated with which the number of the normal equations necessary for obtaining the tap coefficient cannot be obtained. However, as to such a class, the tap coefficient calculation unit 46 is configured to output, for example, a default tap coefficient.
It should be noted that in the drawing, a part corresponding to the case in
To the coefficient output unit 55, in addition to the class (class code) supplied from the class classification unit 14, for example, a parameter z input from the outside in accordance with an operation of the user is supplied. As will be described below, the coefficient output unit 55 generates the tap coefficient for each class corresponding to the parameter z and outputs the tap coefficient for the class from the class classification unit 14 among the tap coefficients for the respective classes to the prediction computation unit 16.
A coefficient generation unit 61 generates a tap coefficient for each class on the basis of coefficient seed data stored in a coefficient seed memory 62 and a parameter z stored in a parameter memory 63 to be supplied to a coefficient memory 64 and stored in an overwriting manner.
The coefficient seed memory 62 stores the coefficient seed data for each class obtained through a learning on the coefficient seed data which will be described below. Herein, the coefficient seed data is so-called data becoming a seed for generating the tap coefficient.
The parameter memory 63 stores the parameter z input from the outside in accordance with the operation of the user or the like in an overwriting manner.
The coefficient memory 64 stores the tap coefficient for each class supplied from the coefficient generation unit (the tap coefficient for each class corresponding to the parameter z). Then, the coefficient memory 64 reads out the coefficient for the class supplied from the class classification unit 14 (
In the image conversion apparatus 51 of
When the parameter z is stored in the parameter memory (when the storage content of the parameter memory 63 is updated), the coefficient generation unit 61 reads out the coefficient seed data for each class from the coefficient seed memory 62 and also reads out the parameter z from the parameter memory 63, and on the basis of the coefficient seed data and the parameter z, obtains the tap coefficient for each class. Then, the coefficient generation unit 61 supplies the tap coefficient for each class to the coefficient memory 64 to be stored in an overwriting manner.
The image conversion apparatus 51 stores the tap coefficient, and in the coefficient output unit 55 provided instead of the coefficient output unit 15 for outputting the tap coefficient, except for generating and outputting the tap coefficient corresponding to the parameter z, a processing is performed which is similar to the processing following the flow chart of
Next, a prediction computation in the prediction computation unit 16 of
As in the case according to the embodiment of
Herein, the pixel value y of the high image quality pixel can also be obtained through a higher-order expression equal to or higher than the second order instead of the linear primary expression shown in Expression (1).
According to the embodiment of
Where, in Expression (9), βm, n denotes the m-th coefficient seed data used for obtaining the n-th tap coefficient wn. It should be noted that in Expression (9), the tap coefficient wn is obtained by using M coefficient seed data β1, n, β2, n, . . . , βM, n.
Herein, from the coefficient seed data βm, n and the parameter z, the expression for obtaining the tap coefficient wn is not limited to Expression (9).
Now, a value zm−1 decided by the parameter z in Expression (9) is defined by the following expression while a new variable tm is introduced.
[Expression 10]
t
m
=z
m−1 (m=1, 2, . . . , M) (10)
As Expression (10) is assigned to Expression (9), the following expression is obtained.
According to Expression (11), the tap coefficient wn is obtained by the linear primary expression based on the coefficient seed data βm, n and the variable tm.
Incidentally, now, when a true value of the pixel value for the high image quality pixel of the k-sample is denoted by yk and also a predicted value of the true value yk obtained through Expression (1) is denoted by yk′, the prediction error ek is represented by the following expression.
[Expression 12]
e
k
=y
k
−y
k′ (12)
Now, as the predicted value yk′ in Expression (12) is obtained while following Expression (1), yk′ in Expression (12) is replaced while following Expression (1), the following expression is obtained.
Where, in Expression (13), xn, k denotes the n-th low image quality pixel constituting the prediction tap with regard to the high image quality pixel of the k-sample.
As Expression (11) is assigned to wn in Expression (13), the following expression is obtained.
Although the coefficient seed data βm, n where the prediction error ek in Expression (14) is 0 becomes an optimal one for predicting the high image quality pixel, regarding all the high image quality pixels, it is generally difficult to obtain such coefficient seed data βm, n.
In view of the above, if the method of least squares is adopted as a rule representing that the coefficient seed data βm, n is optimal, for example, the optimal coefficient seed data βm, n can be obtained by minimizing the total sum E for the square errors represented by the following expression.
Where, in Expression (15), K denotes the sample number (the number of learning material samples) of sets of the high image quality pixel yk and the low image quality pixels x1, k, x2, k, . . . , xN, k constituting the prediction tap with regard to the high image quality pixel yk.
The minimum value (local minimal value) of the total sum E for the square errors in Expression (15) is given by βm, n, as shown in Expression (16), where the total sum E subjected to a partial differentiation by the coefficient seed data βm, n is set as 0.
As Expression (13) is assigned to Expression (16), the following expression is obtained.
Now, xi, p, j, q and yi, p are defined as shown in Expressions (18) and (19).
In this case, Expression (17) can be representing by the normal equation shown in Expression (20) using Xi, p, j, q and Yi, p.
The normal equation of Expression (20) can be solved, for example, by using the discharge method (Gauss-Jordan elimination method) or the like, with regard to the coefficient seed data βm, n.
In the image conversion apparatus 51 of
It should be noted that in the drawing, a part corresponding to the case in the learning apparatus 21 of
Similarly as in the student data generation unit 34 of
It should be however that to the student data generation unit 74, in addition to the learning material image data, some of values in a range where the parameter z supplied to the parameter memory 63 of
The student data generation unit 74 performs a filtering on the high image quality image data serving as the learning material image data, for example, through an LPF with a cutoff frequency corresponding to the parameter z supplied thereto, so that the low image quality image data serving as the student data is generated.
Therefore, regarding the high image quality image data serving as the learning material image data, in the student data generation unit 74, Z+1 types of the low image quality image data serving as the student data with different resolutions are generated.
It should be noted that herein, for example, as the number of the parameter z increases, an LPF with a higher cutoff frequency is used for filtering the high image quality image data, and the low image quality image data serving as the student data is generated. Therefore, herein, as the low image quality image data corresponding to the parameter z with a larger value, the spatial resolution is higher.
Also, according to the present embodiment, to simplify the description, in the student data generation unit 74, the low image quality image data is generated in which the spatial resolutions in both the directions of the horizontal direction and the vertical direction of the high image quality image data are decreased by an amount corresponding to the parameter z.
The learning unit 76 uses the teacher data stored in the teacher data storage unit 33, the student data stored in the student data storage unit 35, and the parameter z supplied from the parameter generation unit 81 to obtain and output the coefficient seed data for each class.
The parameter generation unit 81 generates, for example, z=0, 1, 2, . . . , Z described above as some of values in a range where the parameter z can take to be supplied to the student data generation unit 74 and the learning unit 76.
Similarly as in the tap selection unit 42 of
Similarly as in the tap selection unit 43 of
It should be however that in
The supplement unit 95 reads out the attention pixel from the teacher data storage unit 33 of
That is, the supplement unit 95 is supplied with the teacher data yk serving as the attention pixel stored in the teacher data storage unit 33, the prediction tap xi, k(xj, k) with regard to the attention pixel output by the tap selection unit 42, and the class of the attention pixel output by the class classification unit 44, and also the parameter z at the time of generating the student data constituting the prediction tap with regard to the attention pixel is supplied from the parameter generation unit 81.
Then, for each class supplied from the class classification unit 44, the supplement unit 95 uses the prediction tap (student data) xi, k(xj, k) and the parameter z to perform a multiplication of the student data for obtaining the component Xi, p, j, q defined by Expression (18) in a matrix on the left side in Expression (20) and the parameter z (xi, ktpxj, ktq) and a computation comparable to the summation (E). It should be noted that tp of Expression (18) is calculated from the parameter z while following Expression (10). tq of Expression (18) is also similar.
Furthermore, for each class corresponding to the class code supplied from the class classification unit 44, the supplement unit 95 all the same uses the prediction tap (student data) xi, k, the teacher data yk, and the parameter z to perform a multiplication of the student data xi, k for obtaining the component Yi, p defined by Expression (19) in a vector on the right side in Expression (20), the teacher data yk, and the parameter z (xi, ktpyk) and a computation comparable to the summation (E). It should be noted that tp of Expression (19) is calculated from the parameter z while following Expression (10).
That is, the supplement unit 95 stores the component Xi, p, j, q in a matrix on the left side and the component Yi, p in a vector on the right side in Expression (20) obtained regarding the teacher data set as the attention pixel in the previous time in a built-in memory thereof (not shown), and with respect to the component Xi, p, j, q in the matrix or the component Yi, p in the vector, regarding the teacher data newly set as the attention pixel, supplements the teacher data yk, the student data xi, k(xj, k), and the corresponding component xi, ktpxj, ktq or xi, ktpyk calculated by using the parameter z (performs the addition represented by the summation in the component Xi, p, j, q of Expression (18) or the component Yi, p of Expression (19)).
Then, the supplement unit 95 performs the above-mentioned supplement on while all the teacher data stored in the teacher data storage unit 33 as the attention pixels for the parameters z for all the values of 0, 1, . . . , Z, and thus, with regard to the respective classes, when the normal equation shown in Expression (20) is established, the normal equation is supplied to a coefficient seed calculation unit 96.
The coefficient seed calculation unit 96 obtains and outputs the respective coefficient seed data for each class βm, n by solving the normal equation for each of the classes supplied from the supplement unit 95.
Next, with reference to a flow chart of
First, in step S31, the teacher data generation unit 32 and the student data generation unit 74 respectively generate the teacher data and the student data from the learning material image data stored in the learning material image storage unit 31 to be output. That is, the teacher data generation unit 32 outputs the learning material image data, for example, as the teacher data as it is. Also, the student data generation unit 74 is provided with the parameter z with a value of Z+1 generated by the parameter generation unit 81. The student data generation unit 74 performs the filtering, for example, on the learning material image data through the LPF with the cutoff frequency corresponding to the parameter z with the value of Z+1 (0, 1, . . . , Z) from the parameter generation unit 81, so that regarding the teacher data for the respective frames (the learning material image data), the student data with the frames Z+1 is generated and output.
The teacher data output by the teacher data generation unit 32 is supplied to the teacher data storage unit 33 to be stored, and the student data output by the student data generation unit 74 is supplied to the student data storage unit 35 to be stored.
After that, in step S32, the parameter generation unit 81 sets the parameter z, for example, 0 as an initial value to be supplied to the tap selection units 42 and 43 of the learning unit 76 (
In step S34, regarding the attention pixel, the tap selection unit 42 selects the prediction tap from the student data stored in the student data storage unit 35 corresponding to the parameter z output by the parameter generation unit 81 (the student data generated by filtering the learning material image data corresponding to the teacher data which becomes the attention pixel through the LPF with the cutoff frequency corresponding to the parameter z) to be supplied to the supplement unit 95. Furthermore, in step S34, regarding the attention pixel, the tap selection unit 43 all the same selects the class tap from the student data stored in the student data storage unit 35 corresponding to the parameter z output by the parameter generation unit 81 to be supplied to the class classification unit 44.
Then, in step S35, the class classification unit 44 performs the class classification of the attention pixel on the basis of the class tap regarding the attention pixel and outputs the class of the attention pixel obtained as the result to the supplement unit 95.
In step S36, the supplement unit 95 reads out the attention pixel from the teacher data storage unit 33 and uses the attention pixel, the prediction tap supplied from the tap selection unit 42, and the parameter z output by the parameter generation unit 81 to calculate the component xi, ktpxj, Ktq in a matrix on the left side and the component xi, KtpyK in a vector on the right side in Expression (20). Furthermore, the supplement unit 95 performs the supplement of the attention pixel, the prediction tap, and the component xi, Ktpxj, Ktq in the matrix and the component xi, KtpyK in the vector obtained from the parameter z with respect to one corresponding to the class of the attention pixel from the class classification unit 44 among the component in the matrix and the component in the vector already obtained.
In step S37, the parameter generation unit 81 determines whether or not the parameter z output by itself is equal to Z which is a maximum value of the value that can be taken. In step S37, in a case where it is determined that the parameter z output by the parameter generation unit 81 is not equal to the maximum value Z (smaller than the maximum value Z), the processing proceeds to step S38, and the parameter generation unit 81 adds 1 to the parameter z and outputs the added value as the new parameter z to the tap selection units 42 and 43 of the learning unit 76 (
Also, in step S37, in a case where it is determined that the parameter z is equal to the maximum value Z, the processing proceeds to step S39, and the attention pixel selection unit 41 determines whether or not the teacher data which is not set as the attention pixel is still stored in the teacher data storage unit 33. In step S38, in a case where it is determined that the teacher data which is not set as the attention pixel is still stored in the teacher data storage unit 33, the processing returns to step S32, and afterward the similar processing is repeatedly performed.
Also, in step S39, in a case where it is determined that the teacher data which is not set as the attention pixel is not stored in the teacher data storage unit 33, the processing proceeds to step S40, and the supplement unit 95 supplies the matrix on the left side and the vector on the right side in Expression (20) for each of the classes obtained through the processing up to now to the coefficient seed calculation unit 96.
Furthermore, in step S40, by solving the normal equation for each of the classes constituted by the matrix on the left side and the vector on the right side in Expression (20) for each of the classes supplied from the supplement unit 95, the coefficient seed calculation unit 96 obtains and outputs the coefficient seed data βm, n for the respective classes, and the processing is ended.
It should be noted that due to a state or the like where the number of the learning material image data is not sufficient, a class may be generated with which the number of the normal equations necessary for obtaining the coefficient seed data cannot be obtained. However, as to such a class, the coefficient seed calculation unit 96 is configured to output, for example, default coefficient seed data.
Incidentally, in the learning apparatus 71 of
That is, while the high image quality image data serving as the learning material image data is set as the teacher data and also the low image quality image data in which the high image quality image data is subjected to the filtering by the LPF with the cutoff frequency corresponding to the parameter z to decrease the horizontal resolution and the vertical resolution is set as the student data, first, the learning apparatus 71 obtains the tap coefficient wn for minimizing the total sum for the square errors of the predicted value y of the teacher data predicted through the linear first-order prediction expression of Expression (1) by using the tap coefficient wn and the student data xn for each value of the parameter z (herein, z=0, 1, . . . , Z). Then, the learning apparatus 71 obtains the coefficient seed data βm, n for minimizing the total sum for the square errors of the predicted value of the tap coefficient wn serving as the teacher data predicted from the variable tm corresponding to the coefficient seed data βm, n and the parameter z which is the student data through Expression (11) while the tap coefficient wn obtained for each value of the parameter z is set as the teacher data and also the parameter z is set as the student data.
Herein, the tap coefficient wn for minimizing (diminishing) the total sum E for the square errors of the predicted value y of the teacher data predicted by the linear first-order prediction expression of Expression (1) can be obtained with regard to the respective classes for each of the values of the parameter z(z=0, 1, . . . , Z), similarly as in the case in the learning apparatus 21 of
Incidentally, the tap coefficient is obtained from the coefficient seed data βm, n and the variable tm corresponding to the parameter z as shown in Expression (11). Then, now, if this tap coefficient obtained through Expression (11) is denoted by wn′, the coefficient seed data βm, n for setting an error en as 0 which is represented by the following Expression (21) between the optimal tap coefficient wn and the tap coefficient wn′ obtained through Expression (11) becomes the optimal coefficient seed data for obtaining the optimal tap coefficient wn, but with regard to all the tap coefficients wn, it is generally difficult to obtain such coefficient seed data βm, n.
[Expression 21]
e
n
=w
n
−w
n′ (21)
It should be noted that Expression (21) can be transformed into the following expression by Expression (11).
In view of the above, all the same, if the method of least squares is adopted as a rule representing that the coefficient seed data βm, n is optimal, for example, the optimal coefficient seed data βm, n can be obtained by minimizing the total sum E for the square errors represented by the following expression.
The minimum value (local minimal value) of the total sum E for the square errors of Expression (23) is given, as shown in Expression (24), by βm, n with which the total sum E subjected to the partial differentiation by the coefficient seed data is set as 0.
Expression (22) is assigned to Expression (24), so that the following expression is obtained.
Now, Xi, j, and Yi are defined as shown in Expressions (26) and (27).
In this case, Expression (25) can be represented by the normal equation shown in Expression (28) using Xi, j and Yi.
The normal equation of Expression (28) can also solve the coefficient seed data βm, n, for example, by using the discharge method (Gauss-Jordan elimination method) or the like.
It should be noted that in the drawing, a part corresponding to the case in the learning apparatus 21 of
A supplement unit 115 is provided with the class of the attention pixel output by the class classification unit 44 and the parameter z output by the parameter generation unit 81. Then, the supplement unit 115 reads out the attention pixel from the teacher data storage unit 33 and performs the supplement while the attention pixel and the student data constituting the prediction tap regarding the attention pixel supplied from the tap selection unit 42 are set as the targets for each class supplied from the class classification unit 44 and also for each value of the parameters z output by the parameter generation unit 81.
That is, the supplement unit 115 is supplied with stored in the teacher data yk the teacher data storage unit (
Then, for each class supplied from the class classification unit 44, and also for each value of the parameters z output by the parameter generation unit 81, the supplement unit 115 uses the prediction tap (student data) xn, k to perform a multiplication of mutual student data in a matrix on the left side in Expression (8) (xn, kxn′, k) and a computation comparable to the summation (Σ).
Furthermore, for each class supplied from the class classification unit 44 and also for each value of the parameters z output by the parameter generation unit 81, the supplement unit 115 all the same uses the prediction tap (student data) xn, k and the teacher data yk to perform the multiplication (xn, kyk) of the student data xn, k and the teacher data yk in the vector on the right side in Expression (8) and a computation comparable to the summation (Σ).
That is, the supplement unit 115 stores the component (Σxn, kxn′, k) in a matrix on the left side and the component (Σxn, kyk) in a vector on the right side in Expression (8) obtained regarding the teacher data set as the attention pixel in the previous time in a built-in memory thereof (not shown) and performs the supplement of the corresponding component xn, k+1xn′, k+1 or xn, k+1yk+1 calculated by using the teacher data yk+1 and the student data xn, k+1 regarding the teacher data newly set as the attention pixel with respect to the component (Σxn, kxn′, k) in the matrix or the component (Σxn, kyk) in the vector (performs the addition represented by the summation in Expression (8)).
Then, by performing the above-mentioned supplement while all the teacher data stored in the teacher data storage unit 33 is set as the attention pixel, with regard to the respective classes, for each value of the parameters z, when the normal equation shown in Expression (8) is established, the supplement unit 115 supplies the normal equation to the tap coefficient calculation unit 46.
Therefore, similarly as in the supplement unit 45 of
By solving the normal equation supplied from the supplement unit 115 for each value of the parameters z with regard to the respective classes, the tap coefficient calculation unit 46 obtains the optimal tap coefficient wn for each value of the parameters z with regard to the respective classes to be supplied to a supplement unit 121.
The supplement unit 121 performs the supplement while (the variable tm corresponding to) the parameter z supplied from the parameter generation unit 81 (
That is, the supplement unit 121 uses the variable ti(tj) obtained from the parameter z supplied from the parameter generation unit 81 (
Herein, as the component Xi, j is decided by only the parameter z and is not relevant to the class, the calculation for the component Xi, j is completed by only once in actuality without a necessity to be performed for each class.
Furthermore, the supplement unit 121 uses the variable ti obtained from the parameter z supplied from the parameter generation unit 81 through Expression (10) and the optimal tap coefficient wn supplied from the tap coefficient calculation unit 46 to perform the multiplication (tiwn) of the variable ti corresponding to the parameter z for obtaining the component Yi defined by Expression (27) in the vector on the right side of Expression (28) and the optimal tap coefficient wn and a computation comparable to the summation (Σ) for each class.
For the respective classes, by obtaining the component Xi, j represented by Expression (26) and the component Yi represented by Expression (27), with regard to the respective classes, when the normal equation of Expression (28) is established, the supplement unit 121 supplies the normal equation to a coefficient seed calculation unit 122.
By solving the normal equation of Expression (28) for each class supplied from the supplement unit 121, the coefficient seed calculation unit 122 obtains and outputs the respective coefficient seed data for each of the classes βm, n.
The coefficient seed memory 62 in the coefficient output unit 55 of
It should be noted that in the learning on the coefficient seed data too, similarly as in the case of the learning on the tap coefficient described in
That is, in the above-mentioned case, the learning on the coefficient seed data is performed in which the learning material image data is set as the teacher data corresponding to the second image data as it is and also the low image quality image data in which the spatial resolution of the learning material image data is degraded is set as the student data corresponding to the first image data, and thus one for performing the image conversion processing serving as the spatial resolution creation processing for converting the first image data into the second image data in which the spatial resolution is improved can be obtained as the coefficient seed data.
In this case, in the image conversion apparatus 51 of
Also, for example, by performing the learning on the coefficient seed data while the high image quality image data is set as the teacher data and also with respect to the high image quality image data serving as the teacher data and the image data on which the noise at a level corresponding to the parameter z is overlapped is set as the student data, one for performing the image conversion processing serving as the noise removal processing for converting the first image data into the second image data in which the noise therein is removed (reduced) can be obtained as the coefficient seed data. In this case, in the image conversion apparatus 51 of
Also, for example, by performing the learning on the coefficient seed data while certain image data is set as the teacher data and also the image data in which the pixel number of the image data serving as the teacher data is thinned out while corresponding to the parameter z is set as the student data or the image data with a predetermined size is set as the student data and also the image data in which the pixel of the image data serving as the student data is thinned out at a thinning out date corresponding to the parameter z is set as the teacher data, one for performing the image conversion processing serving as the resize processing for converting the first image data into the second image data in which the size is expanded or reduced can be obtained as the coefficient seed data. In this case, in the image conversion apparatus 51 of
It should be noted that in the above-mentioned case, as shown in Expression (9), the tap coefficient wn is defined by β1, nz0+β2, nz1+ . . . +βM, nzM−1, and with this Expression (9), the tap coefficient wn for improving the spatial resolutions in the horizontal and vertical directions while both corresponding to the parameter z is obtained. However, as the tap coefficient wn, it is also possible to obtain one for independently improving the horizontal resolution and the vertical resolution while corresponding to the independent parameters zx and zy.
That is, instead of Expression (9), for example, the tap coefficient wn is defined by the third-order expression β1, nzx0zy0+β2, nzx1zy0+β3, nzx2zy0+β4, nzx3zy0+β5, nzx0zy1+β6, nzx0zy2+β7, nzx0zy3+β8, nzx1zy1+β9, nzx2zy1+β10, nzx1zy2, and also the variable tm defined by Expression (10) is defined by, for example, t1=zx0zy0, t2=zx1zy0, t3=zx2zy0, t4=zx3zy0, t5=zx0zy1, t6=zx0zy2, t7=zx0zy3, t8=zx1zy1, t9=zx2zy1, t10=zx1zy2 instead of Expression (10). In this case too, the tap coefficient wn can be eventually represented by Expression (11), and therefore, in the learning apparatus 71 of
Other than that, for example, in addition to the parameter zx and zy respectively corresponding to the horizontal resolution and the vertical resolution, further, by introducing a parameter zt corresponding to a resolution in the time direction, it is possible to obtain the tap coefficient wn for independently improving the horizontal resolution, the vertical resolution, and the time resolution respectively while corresponding to the independent parameters zx, zy, and zt.
Also, regarding the resize processing too, similarly to the case of the spatial resolution creation processing, in addition to the tap coefficient wn for resizing the horizontal and vertical directions at an expansion rate (or a reduction rate) both corresponding to the parameter z, it is possible to obtain the tap coefficient wn for independently resizing the horizontal and vertical directions at expansion rates respectively corresponding to the parameters zx and zy.
Furthermore, in the learning apparatus 71 of
An embodiment of the image processing apparatus which is an image processing apparatus provided with a processing unit for performing the above-mentioned class classification adaptive processing to which the present invention is applied will be described below. In other words, the image processing apparatus (image processing system) which will be described below is an apparatus provided with the above-mentioned image conversion apparatus as an image processing unit (image processing units 213-1 to 213-3 of
An image processing apparatus 200 of
To the image processing apparatus 200, a moving image is input. A plurality of images constituting the image are sequentially obtained by the image input unit 211 and supplied to the image distribution unit 212 as the input images.
The image distribution unit 212 supplies the input image supplied from the image input unit 211 to the image processing units 213-1 to 213-3 and the image synthesis unit 214.
Under a control by the control unit 217, the respective image processing units 213-1 to 213-3 execute a predetermined image processing on the input image supplied from the image distribution unit 212 at the same time (in parallel) and supply processed images which are the images after the processing to the image synthesis unit 214. Herein, the image processing performed by the image processing units 213-1 to 213-3 is, for example, a high image quality realization processing for setting an image quality into a high image quality, an expansion processing for expanding a predetermined area of the image, or the like. Also, those image processing units 213-1 to 213-3 execute mutually different processings, in which, for example, degrees of the image quality are different or the areas on which the expansion processing is performed are different. For the image processings performed by these image processing units 213-1 to 213-3, it is possible to utilize the above-mentioned class classification adaptive processing.
The image synthesis unit 214 is supplied with the input image from the image distribution unit 212 and also supplied with the processed images respectively from the image processing units 213-1 to 213-3.
Under a control by the control unit 217, the image synthesis unit 214 uses the input image and the three types of the processed images to generate a synthesized image to be supplied to the image presentation unit 215. Also, the image synthesis unit 214 supplies a main image which is an image functioning as a principal among a plurality of images used for the synthesized image to the image recording unit 216.
By displaying the synthesis image supplied from the image synthesis unit 214 on a predetermined display unit, the image presentation unit 215 presents the synthesis image to the user. The image recording unit 216 records the main image supplied from the image synthesis unit 214 on a predetermined recording medium.
The control unit 217 obtains operation information which is information where the user operates a remote commander not shown in the drawing or the like and supplies control information corresponding to the operation to the image processing units 213-1 to 213-3 and the image synthesis unit 214.
In
In
The image processing unit 213-1 is composed of an image quality change processing unit 241A and an expansion processing unit 242A, and the image processing unit 213-2 is composed of an image quality change processing unit 241B and the expansion processing unit 242B. Also, the image processing unit 213-3 is composed of an image quality change processing unit 241C and an expansion processing unit 242C.
Also, the control unit 217 is composed of a control instruction unit 261 and an image comparison unit 262.
The image quality change processing units 241A to 241C apply different image parameters to generate and output images different in the image quality. For example, the image quality change processing units 241A to 241C generate the images different in the image quality such that the levels (degrees) of the image quality are higher in the order of the image quality change processing units 241A, 241B, and 241C. The levels of the image quality are decided by the control information supplied from the control instruction unit 261.
The image quality change processing units 241A to 241C can respectively execute, for example, the above-mentioned class classification adaptive processing, and as the image parameters in this case, the parameter z for specifying the resolution and the noise removal degree, for example, and the coefficient seed data can be set. Also, a general image processing other than the class classification adaptive processing may be adopted, and parameters for changing a hue, a luminance, a γ value γ, and the like can also be set.
The image quality change processing unit 241A supplies the image after the image quality change processing (hereinafter, which is referred to as processing A image) to the image comparison unit 262, the expansion processing unit 242A, and the image synthesis unit 214. The image quality change processing unit 241B supplies the image after the image quality change processing (hereinafter, which is referred to as the processing B image) to the image comparison unit 262, the expansion processing unit 242B, and the image synthesis unit 214. The image quality change processing unit 241C supplies the image after the image quality change processing (hereinafter, which is referred to as the processing C image) to the image comparison unit 262, the expansion processing unit 242C, and the image synthesis unit 214.
The expansion processing units 242A to 242C respectively execute an expansion processing based on a position (p, q) supplied from the image comparison unit 262. That is, the expansion processing units 242A to 242C generate expanded images where a predetermined area is expanded while the position (p, q) is set as the center to be supplied to the image synthesis unit 214. Herein, an area size indicating how much of the area is expanded with respect to the position (p, q) is previously specified by the user on a setting screen or the like and decided. It should be noted that in a case where the area size decided by the user specification is not matched with the size of the respective areas of a display screen 270 which will be described below with reference to
As a system of the expansion processing executed by the expansion processing units 242A to 242C, in addition to the above-mentioned class classification adaptive processing, a linear interpolation system, a bi-linear system, or the like can be adopted.
The control instruction unit 261 supplies control information for controlling the processing contents to the image quality change processing units 241A to 241C, the expansion processing units 242A to 242C, and the image synthesis unit 214 on the basis of the operation information indicating the contents operated by the user. For example, as described above, the control instruction unit 261 supplies the parameter for deciding the levels of the image quality to the image quality change processing units 241A to 241C as the control information and supplies the parameter for deciding the area size where the expansion processing is performed to the expansion processing units 242A to 242C as the control information.
The image comparison unit 262 performs a mutual comparison computation on the processed images respectively supplied from the image quality change processing units 241A to 241C to detect the position (pixel) (p, q) where the image quality is most different among the processing A image, the processing B image, and the processing C image respectively output by the image quality change processing units 241A to 241C to be supplied to the expansion processing units 242A to 242C.
With reference to
The image comparison unit 262 is supplied from the control instruction unit 261 with BLOCKSIZE (By, By) which is the comparison area size and a processing frame number FN as control information. Also, the image comparison unit 262 is supplied from the image quality change processing units 241A to 241C with the processing A image to the processing C image. Herein, the processing A image to the processing C image are images where a corner on the upper left of the image is set as an origin and which are composed of X pixels in the horizontal direction and Y pixels in the vertical direction.
First, as shown in
Then, the image comparison unit 262 executes the following processing on the processing A image to the processing C image at a predetermined frame supplied from the image quality change processing units 241A to 241C (for example, at the first frame). That is, the image comparison unit 262 calculates a difference square sum d11(1, 1) of luminance values (pixel values) of mutual pixels at the same position between the comparison target image A (1, 1) and a comparison target image B (1, 1). Also similarly, the image comparison unit 262 calculates a difference square sum d12(1, 1) of luminance values of mutual pixels at the same position between the comparison target image B (1, 1) and a comparison target image C (1, 1) and calculates a difference square sum d13(1, 1) of mutual pixels at the same position between the comparison target image A (1, 1) and the comparison target image C (1, 1).
Next, the image comparison unit 262 obtains a total of the calculated difference square sum d11(1, 1), the difference square sum d12(1, 1), and the difference square sum d13(1, 1) to be set as a difference square sum d1(1, 1). That is, the image comparison unit 262 calculates d1(1, 1)=d11(1, 1)+d12(1, 1)+d13(1, 1).
The image comparison unit 262 executes the above-mentioned processing while the position (1, 2) to (X′, Y′) of the processing A image to the processing C image are set as the reference position. Herein, X′=X−Bx+1 and Y′=Y−By+1 are established. That is, the image comparison unit 262 obtains the difference square sums d1(1, 1) to d1(X′, Y′) regarding all the reference positions (1, 1) to (X′, Y′) with which the processing A image to the processing C image can be set without a protrusion with respect to the input image at the first frame.
Furthermore, the image comparison unit 262 repeatedly performs the processing for obtaining the difference square sums d1(1, 1) to d1(X′, Y′) on the input image of the processing frame number FN supplied from the control instruction unit 261.
As a result, as shown in
Then, the image comparison unit 262 calculates the total sum of the difference square sums d(x′, y′)=Σdk(x′, y′) with regard to all x′=1, 2, . . . , X′ and y′=1, 2, . . . , Y′ and obtains and decides the position (p, q) where the total sum d(p, q) is largest among the total sums d(1, 1) to d(X′, Y′) which are the calculation results. Therefore, (p, q) is one of (1, 1) to (X′, Y′). It should be noted that Σ denotes a sum from a time when k=1 is set until k=FN is set.
With reference to a flow chart of
First, in step S61, the image comparison unit 262 stores BLOCKSIZE (Bx, By) supplied from the control instruction unit 261 and the processing frame number FN therein.
In step S62, the image comparison unit 262 selects the processing A image to the processing C image at a predetermined frame. For example, the processing A image to the processing C image on which the image quality change processing is performed on the input image at the first frame are selected.
In step S63, the image comparison unit 262 decides a predetermined pixel of the processing A image to the processing C image as the reference position. For example, the image comparison unit 262 decides the position (1, 1) as the reference position. According to this, the comparison target images A (1, 1) to C (1, 1) are decided.
In step S64, the image comparison unit 262 calculates the difference square sum of the luminance values of the mutual pixels at the same position between the comparison target image A and the comparison target image B. For example, in the comparison target images A (1, 1) to C (1, 1) regarding the reference position (1, 1), the image comparison unit 262 calculates the difference square sum d11(1, 1) of the luminance values of the mutual pixels at the same position between the comparison target image A (1, 1) and the comparison target image B (1, 1).
In step S65, the image comparison unit 262 calculates the difference square sum of the luminance values of the mutual pixels at the same position between the comparison target image B and the comparison target image C. For example, in the comparison target images A (1, 1) to C (1, 1) regarding the reference position (1, 1), the image comparison unit 262 calculates the difference square sum d12(1, 1) of the mutual pixels at the same position between the comparison target image B (1, 1) and the comparison target image C (1, 1).
In step S66, the image comparison unit 262 calculates the difference square sum of the mutual pixels at the same position between the comparison target image A and the comparison target image C. For example, in the comparison target images A (1, 1) to C (1, 1) regarding the reference position (1, 1), the image comparison unit 262 calculates the difference square sum d13(1, 1) of the mutual pixels at the same position between the comparison target image A (1, 1) and the comparison target image C (1, 1).
In step S67, the image comparison unit 262 obtains a total of the difference square sums obtained in steps S64 to S66. For example, the image comparison unit 262 obtains a the calculated difference square sum d11(1, 1), the difference square sum d12(1, 1), and the difference square sum d13(1, 1) to be set as the difference square sum d1(1, 1).
In step S68, the image comparison unit 262 determines whether or not all the pixels with which the processing A image to the processing C image can be set without a protrusion are set as the reference position. In step S68, in a case where it is determined that all the pixels are not set as the reference position yet, the processing returns to step S63. According to this, the pixel which is not set as the reference position yet is set as the next reference position, and a subsequent processing is repeatedly performed.
On the other hand, in step S68, in a case where it is determined that all the pixels are set as the reference position, the processing proceeds to step S69, and the image comparison unit 262 determines whether or not the frame where the difference square sum is currently calculated is the last frame of the processing frame number FN.
In step S69, in a case where it is determined that the frame where the difference square sum is currently calculated is not the last frame of the processing frame number FN, the processing returns to step S62, and the subsequent processing is repeatedly performed.
On the other hand, in step S69, in a case where it is determined that the frame where the difference square sum is currently calculated is the last frame of the processing frame number FN, that is, in a case where the difference square sum is calculated regarding the input image of the processing frame number FN, in step S70, the image comparison unit 262 calculates the total sum of the difference square sums d(x′, y′)=Σdk(x′, y′) with regard to the pixel (1, 1) to (X′, Y′) to calculate the total sum of the difference square sums d(1, 1) to d(X′, Y′).
In step S71, the image comparison unit 262 obtains and decides obtains the position (p, q) where the total sum d(p, q) is largest among the calculated total sums of the difference square sums d(1, 1) to d(X′, Y′). Also, the image comparison unit 262 supplies the decided position (p, q) to the expansion processing units 242A to 242C, respectively, and the processing is ended.
In the image comparison unit 262, in the above-mentioned manner, the processing A image to the processing C image are compared, and the position (p, q) is obtained.
Next, with reference to a flow chart of
First, in step S81, the image distribution unit 212 distributes the input image supplied from the image input unit 211. That is, the image distribution unit 212 supplies the image input to the image processing units 213-1 to 213-3 and the image synthesis unit 214.
In step S82, the image quality change processing units 241A to 241C execute an image quality change processing which is a processing of changing an image quality on the image input. It should be noted that in the image quality change processing units 241A to 241C, by the control of the control instruction unit 261, the image quality change processing is performed so that the image qualities of the generated images are different from each other. The processing A image to the processing C image after the image quality change processing are supplied to the image comparison unit 262.
In step S83, the image comparison unit 262 executes the image comparison processing described with reference to
In step S84, the expansion processing units 242A to 242C execute an expansion processing of expanding a part of the input processed image. That is, the expansion processing unit 242A generates an expanded image A where a predetermined area is expanded while the position (p, q) of the processed A image supplied from the image quality change processing unit 241A is set as the reference to be supplied to the image synthesis unit 214. The expansion processing unit 242B generates an expanded image B where a predetermined area is expanded while the position (p, q) of the processed B image supplied from the image quality change processing unit 241B is set as the reference to be supplied to the image synthesis unit 214. The expansion processing unit 242C generates an expanded image C where a predetermined area is expanded while the position (p, q) of the processed C image supplied from the image quality change processing unit 241C is set as the reference to be supplied to the image synthesis unit 214.
In step S85, the image synthesis unit 214 uses the input image supplied from the image distribution unit 212 and the expanded images A to C supplied from the expansion processing units 242A to 242C and generates a synthesized image to be supplied to the image presentation unit 215. Also, the image synthesis unit 214 supplies one image selected among the image input and the expanded images A to C as a main image to the image recording unit 216.
It should be noted that which image among the image input and the expanded images A to C is set as the main image id decided by an instruction of the user supplied via the control unit 217 as will be described with reference to
In step S86, by displaying the synthesis image supplied from the image synthesis unit 214 on the predetermined display unit, the image presentation unit 215 presents the synthesis image to the user. Also, in step S86, the image recording unit 216 records the main image supplied from the image synthesis unit 214 in a predetermined recording medium, and the processing is ended.
With reference to
The image synthesis unit 214 generates a synthesized image so that expanded images can be displayed on the respective areas where the display screen 270 is divided into a main screen area 281 shown in
The display screen 270 is composed of the main screen area 281 and the sub screen areas 282-1 to 282-3 arranged on the right side thereof. The sub screen areas 282-1 to 282-3 are arranged lined up in the up and down direction, and an overall height of the three sub screen areas 282-1 to 282-3 is the same as a height of the main screen area 281.
The synthesized image shown, for example, in
In the synthesized image shown in
Then, a highlight display 291 is displayed in one of the sub screen areas 282-1 to 282-3, and the user can move the highlight display 291 to a desired position of the sub screen areas 282-1 to 282-3 by operating up and down keys (not shown) of the remote commander. That is, the image synthesis unit 214 generates an image in which the highlight display 291 is overlapped on the synthesized image to be supplied to the image presentation unit 215.
When control information indicating that the up key of the remote commander is operated is supplied to the image synthesis unit 214 in a state where the highlight display 291 is at the sub screen area 282-1, the image synthesis unit 214 moves the highlight display 291 to the sub screen area 282-3.
Also, when control information indicating that the down key of the remote commander is operated is supplied to the image synthesis unit 214 in a state where the highlight display 291 is at the sub screen area 282-3, the image synthesis unit 214 moves the highlight display 291 to the sub screen area 282-1.
Alternatively, when control information indicating that a decision key (not shown) of the remote commander is operated is supplied to the image synthesis unit 214 in a state where the highlight display 291 is at the sub screen area 282-3, the image synthesis unit 214 generates a synthesized image shown in
Also, in a case where control information indicating that the decision key (not shown) of the remote commander is operated is supplied to the image synthesis unit 214 in a state where the highlight display 291 is at the sub screen area 282-3, the image synthesis unit 214 can also generate a synthesized image shown in
Next, a GUI for the user to set parameters necessary to the first image processing will be described.
With respect to the image quality change processing performed by the image quality change processing units 241A to 241C, by using a parameter setting screen shown in
In the parameter setting screen shown in
In the zoom setting box 301, it is possible to decide a zoom rate at the time of performing the expansion processing with respect to the image input in the image quality change processing units 241A to 241C. Although not described so far, in the image quality change processing units 241A to 241C, the image input after is expanded at a predetermined zoom rate, the image quality can be changed. At about which zoom rate the image quality change processing units 241A to 241C expand the image input is specified in this zoom setting box 301.
After the user moves a cursor 305 to the zoom setting box 301 and the selection is made by the decision key, by operating the up and down keys, the zoom rate is set as a desired value. In
In the resolution change boxes 302A to 302C, it is possible to decide a resolution which is a parameter for deciding the image quality when the image quality change processing units 241A to 241C perform the image quality change processing. This resolution is, for example, the spatial resolution.
In a case where the respective parameters of the resolution change boxes 302A to 302C are changed, similarly as in the case of changing the zoom rate, the user moves the cursor 305 to the resolution change boxes 302A to 302C where the change is desired and changes the resolution (numeric value) by the up and down key. According to this, the parameter z and the coefficient seed data in a case where the class classification processing is applied.
By moving the cursor to the end box 304 and operating a decision button, the changed parameter is stored inside the control instruction unit 261 and also supplied from the control instruction unit 261 to the image quality change processing units 241A to 241C.
In the parameter setting screen shown in
In a case where BLOCKSIZE (Bx, By) is changed, the user changes a size of an area frame 311 displayed on the parameter setting screen shown in
In a case where the processing frame number FN is changed, after the user moves a cursor 312 to a frame number setting box 314, a numeric value displayed therein is changed by the up and down key. An end box 313 is a button operated at the time of instructing an end of the parameter setting similarly as in the end box 304. Upon the end of the parameter setting screen, the parameter after the change is supplied via the control instruction unit 261 to the image comparison unit 262 and stored therein.
As described above, according to the first embodiment the image processing apparatus 200 shown in
The image processing unit 213-1 is composed of a tracking processing unit 341A and an expansion processing unit 242A′, the image processing unit 213-2 is composed of a tracking processing unit 341B and an expansion processing unit 242B′. Also, the image processing unit 213-3 is composed of a tracking processing unit 341C and an expansion processing unit 242C′.
To the tracking processing units 341A to 341C, the position (hereinafter, which is referred to as user instruction point) (x, y) in the image input instructed by the user at a predetermined timing and the zoom rate z are supplied as the initial value (x, y, z) from the control unit 217.
On the basis of the initial value (x, y, z), the tracking processing units 341A to 341C execute the tracking processing for tracking the user instruction point (x, y) of the image input. It should be noted that the tracking processing units 341A to 341C execute the tracking processing in mutually different tracking processing systems. Therefore, the results of executing the tracking processing while the same user instruction point (x, y) is set as the reference are not necessarily the same. Details of the tracking processing systems performed by the respective tracking processing units 341A to 341C will be described below with reference to
The tracking processing unit 341A supplies a tracking processing result (xa, ya, z) composed of the position after the tracking (xa, ya) and the zoom rate z to the expansion processing unit 242A′ and the control unit 217. The tracking processing unit 341B supplies the tracking processing result (xb, yb, z) composed of the position after the tracking (xb, yb) and the zoom rate z to the expansion processing unit 242B′ and the control unit 217. The tracking processing unit 341C supplies the tracking processing result (xc, yc, z) composed of the position after the tracking (xc, yc) and the zoom rate z to the expansion processing unit 242C′ and the control unit 217.
It should be noted that in a case where the respective tracking processing units 341A to 341C are not particularly necessarily distinguished, those are simply referred to as tracking processing unit 341.
The expansion processing units 242A′ to 242C′ execute an expansion processing similarly as in the expansion processing units 242A to 242C according to the first embodiment and supply expanded images A′ to C′ after the expansion processing to the image synthesis unit 214. An area where the expansion processing is respectively performed by the expansion processing units 242A′ to 242C′ is an area decided by the tracking processing result (xa, ya, z), (xb, yb, z) of the image input supplied from the tracking processing units 341A to 341C or (xc, yc, z).
The control unit 217 supplies the user instruction point (x, y) instructed by the user and the zoom rate z as the initial value (x, y, z) to the tracking processing units 341A to 341C. Also, in the second image processing according to the second embodiment, the expanded images A′ to C′ after the expansion processing by the expansion processing units 242A′ to 242C′ are displayed on one screen at the same time as described with reference to
With reference to
The tracking processing unit 341A detects BLOCKSIZE (Bx, By) in which the user instruction point (x(0), y(0)) is set as the center with respect to the image input at the time t=0 as the search template. It should be noted that BLOCKSIZE (Bx, By) according to the second embodiment does not need to have the same value as BLOCKSIZE (Bx, By) according to the first embodiment. Also, according to the second embodiment, BLOCKSIZE (Bx, By) is set as a square (Bx=By) and simply described as BLOCKSIZE.
Then, when the image input at the time t=1 is supplied, the tracking processing unit 341A detects AREASIZE (Ax, Ay) while the user instruction point (x(0), y(0)) is set as the center as a search target image. It should be noted that according to the second embodiment, AREASIZE (Ax, Ay) is set as a square (Ax=Ay) and simply described as AREASIZE.
In
Next, the tracking processing unit 341A obtains a difference square sum of the luminance value d(x′, y′) with respect to the pixel (x′, y′)(x′=1, 2, . . . , X′, y′=1, 2, . . . , Y′) in AREASIZE. Herein, X′=Ax−Bx+1 and Y′=Ay−By+1 are established. It should be noted that the difference square sum d(x′, y′) according to this second embodiment is a value different from the difference square sum d(x′, y′) according to the first embodiment.
The tracking processing unit 341A obtains and decides the position (v, w) where the difference square sum d(v, w) is smallest among the difference square sums of the luminance values d(1, 1) to d(X′, Y′) regarding all the pixels (1, 1) to (X′, Y′) with which BLOCKSIZE in AREASIZE can be set without a protrusion. Therefore, (v, w) is one of (1, 1) to (X′, Y′).
Then, the tracking processing unit 341A assigns the position (v, w) to the following.
x(t+1)=v+(BLOCKSIZE−AREASIZE)/2+x(t)
y(t+1)=w+(BLOCKSIZE−AREASIZE)/2+y(t)
Thus, the tracking position (x(1), y(1)) in the image input at the time t=1 is obtained. The thus obtained tracking position (x(1), y(1)) is supplied together with the zoom rate (z) to the expansion processing unit 242A′ as the tracking processing result (xa, ya, z) at the time t=1.
First, in step S101, the tracking processing unit 341A obtains the initial value (x, y, z) supplied from the control unit 217. The obtained initial value (x, y, z) is stored inside the tracking processing unit 341A.
In step S102, the tracking processing unit 341A detects the search template from the image input at the time t=0 supplied from the image distribution unit 212. To be more specific, the tracking processing unit 341A detects BLOCKSIZE while the user instruction point (x(0), y(0)) is set as the center with respect to the image input at the time t=0 supplied from the image distribution unit 212 as the search template.
In step S103, the tracking processing unit 341A stands by until the image input at a next time is supplied from the image distribution unit 212.
In step S104, the tracking processing unit 341A detects the search target image from the image input at the next time. That is, the tracking processing unit 341A detects AREASIZE as the search target image while the user instruction point (x(0), y(0)) is set as the center with respect to the image input at the next time.
In step S105, the tracking processing unit 341A obtains the difference square sums of the luminance values d(1, 1) to d(X′, Y′) with regard to all the pixels (1, 1) to (X′, Y′) with which BLOCKSIZE in AREASIZE can be set without a protrusion. The obtained difference square sums of the luminance values d(1, 1) to d(X′, Y′) are stored inside the tracking processing unit 341A as evaluation value tables.
In step S106, the tracking processing unit 341A obtains and decides the position (v, w) where the difference square sum d(v, w) is smallest among the difference square sums of the luminance values d(1, 1) to d(X′, Y′).
In step S107, on the basis of the position (v, w), the tracking processing unit 341A obtains the tracking position of the image input at the next time. For example, in a case where the next time is the time t+1, on the basis of the position (v, w), the tracking processing unit 341A calculates as follows.
x(t+1)=v+(BLOCKSIZE−AREASIZE)/2+x(t)
y(t+1)=w+(BLOCKSIZE−AREASIZE)/2+y(t)
Thus, the tracking position (x(1), y(1)) in the image input at the time t+1 is obtained. In step S107, also, the obtained tracking position is supplied together with the zoom rate as the tracking processing result (xa, ya, z) at the next time to the expansion processing unit 242A′.
In step S108, the tracking processing unit 341A determines whether or not the next image input is supplied. In step S108, in a case where it is determined that the next image input is supplied, the processing returns to step S104, and the subsequent processing is repeatedly executed.
On the other hand, in step S108, in a case where it is determined that the next image input is not supplied, the processing is ended.
As described above, in the tracking processing by the tracking processing unit 341A, at the time t, when the user instruction point (x(t), y(t)) with respect to the image input is instructed by the user, the search template set to the image input where the user instruction point (x(t), y(t)) is specified and the search target images set to the image inputs sequentially input are compared, so that the user instruction point (x(t), y(t)) is tracked. The tracking processing result (x(t+1), y(t+1)) and the zoom rate z are supplied as the tracking processing result (xa, ya, z) to the expansion processing unit 242A′.
This tracking processing by the tracking processing unit 341A is a general method called block matching.
In contrast to this, a system of the tracking processing performed by the tracking processing unit 341B is different in that in step S108 of the tracking processing shown in
That is, in the tracking processing by the tracking processing unit 341A, the search template is not changed while the image input when the user instructs the user instruction point (x(t), y(t)) is regularly set as the reference image, but a different point is that in the tracking processing by the tracking processing unit 341B, the search template is also set by the image input at the latest time. This tracking processing by the tracking processing unit 341B has an advantage as superior to a shape change of the tracking target but on the other hand also has an aspect that the tracking target is gradually shifted.
On the other hand, in the tracking processing by the tracking processing unit 341C, a different point is that the difference square sums d(1, 1) to d(X′, Y′) calculated in step S105 in the tracking processing shown in
As described above, the tracking processing units 341A to 341C execute the tracking processing in respectively different tracking processing systems and supply the tracking processing results (xa, ya, z), (xb, yb, z), and (xc, yc, z) obtained as the results to the expansion processing units 242A′ to 242C′ on a one-to-one basis.
Next, with reference to a flow chart of
First, in step S121, the image distribution unit 212 determines whether or not the image input is supplied from the image input unit 211. In step S121, in a case where it is determined that the image input is not supplied, the processing is ended.
On the other hand, in step S121, in a case where it is determined that the image input is supplied from the image input unit 211, the processing proceeds to step S122, and the image distribution unit 212 distributes the supplied image input. That is, the image distribution unit 212 supplies the image input and the image processing units 213-1 to 213-3 and the image synthesis unit 214.
In step S123, as described with reference to
The tracking processing result (xa, ya, z) by the tracking processing unit 341A is supplied to the expansion processing unit 242A′, and the tracking processing result (xb, yb, z) by the tracking processing unit 341B is supplied to the expansion processing unit 242B′. Also, the tracking processing result (xc, yc, z) by the tracking processing unit 341C is supplied to the expansion processing unit 242C′.
In step S124, the expansion processing units 242A′ to 242C′ execute the expansion processing in parallel for expanding a part of the image input which is input from the image distribution unit 212. That is, the expansion processing unit 242A′ generates an expanded image A′ which is expanded at the zoom rate z while the position after the tracking (xa, ya) supplied from the tracking processing unit 341A is set as the center to be supplied to the image synthesis unit 214. The expansion processing unit 242B′ generates an expanded image B′ which is expanded at the zoom rate z while the position after the tracking (xb, yb) supplied from the tracking processing unit 341B is set as the center to be supplied to the image synthesis unit 214. The expansion processing unit 242C′ generates an expanded image C′ which is expanded at the zoom rate z while the position after the tracking (xc, yc) supplied from the tracking processing unit 341C is set as the center to be supplied to the image synthesis unit 214.
In step S125, the image synthesis unit 214 uses the input image supplied from the image distribution unit 212 and the expanded images A′ to C′ supplied from the expansion processing units 242A′ to 242C′ to generate a synthesized image to be supplied to the image presentation unit 215. Also, the image synthesis unit 214 supplies the main image which is an image arranged in the main screen area 281 in the synthesized image to the image recording unit 216.
In step S126, By displaying the synthesis image supplied from the image synthesis unit 214 on the predetermined display unit, the image presentation unit 215 presents the synthesis image to the user. Also, in step S126, the image recording unit 216 records the main image supplied from the image synthesis unit 214 on the predetermined recording medium.
In step S127, the control unit 217 determines whether or not one expanded image of the expanded images A′ to C′ displayed in the image presentation unit 215 is selected by the user.
In step S127, in a case where it is determined that one expanded image of the expanded images A′ to C′ is not selected by the user, the processing returns to step S121, and the subsequent processing is repeatedly executed.
On the other hand, in step S127, in a case where it is determined that one expanded image of the expanded images A′ to C′ is selected by the user, the processing proceeds to step S128, and the control unit 217 supplies the tracking processing result of the expanded image selected by the user to the tracking processing units 341A to 341C as the next initial value (x, y, z). After that, the processing returns to step S121, and the subsequent processing is repeatedly executed.
With reference to
For example, as an initial state, in the image presentation unit 215, it is supposed that the display screen 360A is displayed. On the display screen 360A, the expanded image A′ from the expansion processing unit 242A′ is displayed in the main screen area 281, and the image input is displayed in the sub screen area 282-1. Also, the expanded image B′ is displayed in the sub screen area 282-2, and the expanded image C′ is displayed in the sub screen area 282-3.
In this initial state, when the user operates a down key DN of the remote commander, the image presentation unit 215 obtains the operation via the control unit 217 to display the display screen 360B. On the display screen 360B, in addition to the display of the display screen 360A, the highlight display 291 for highlighting a predetermined sub screen area is displayed in the sub screen area 282-1.
In a state where the display screen 360B is displayed, when the user operates the down key DN of the remote commander, the image presentation unit 215 displays the display screen 360C on which the highlight display 291 is moved to the sub screen area 282-2.
Next, in a state where the display screen 360C is displayed, when the user operates a decision key RTN of the remote commander, the image presentation unit 215 displays the display screen 360D on which the expanded image A′ of the main screen area 281 and the expanded image B′ of the sub screen area 282-2 are switched the display screen 360D on which.
In a state where the display screen 360D is displayed, when the user operates an up key UP of the remote commander, the image presentation unit 215 displays the display screen 360E on which the highlight display 291 is moved to the sub screen area 282-1.
Furthermore, in a state where the display screen 360E is displayed, when the user operates the up key UP of the remote commander, as the current highlight display 291 is the sub screen area 282-1 on the top among the sub screen areas 282-1 to 282-3, the image presentation unit 215 displays the display screen 360F on which the sub screen area 282-3 is moved to the highlight display 291.
In a state where the display screen 360F is displayed, when the user further operates the up key UP of the remote commander, the image presentation unit 215 displays the display screen 360G on which the highlight display 291 is moved to the sub screen area 282-2.
Then, in a state where the display screen 360G is displayed, when the user operates the decision key RTN of the remote commander, the image presentation unit 215 displays the display screen 360H on which the expanded image B′ of the main screen area 281 and the expanded image A′ of the sub screen area 282-2 are switched.
In a state where the display screen 360H is displayed, when the user operates the down key DN of the remote commander, the image presentation unit 215 displays the display screen 3601 on which the sub screen area 282-3 is moved to the highlight display 291.
Then, in a state where the display screen 360I is displayed, when the user operates the decision key RTN of the remote commander, the image presentation unit 215 displays the display screen 360J on which the expanded image At of the main screen area 281 and the expanded image C′ of the sub screen area 282-3 are switched.
As described above, according to the second embodiment, the expanded image selected by the user among the sub screen areas 282-1 to 282-3 is displayed in the main screen area 281, and the expanded image displayed so far in the main screen area 281 is displayed in the sub screen area selected among the sub screen areas 282-1 to 282-3. That is, the expanded image in the main screen area 281 and the expanded image selected by the user among the sub screen areas 282-1 to 282-3 are switched.
Next, the shift of the tracking position in a case where, as described in
In
As described above, the tracking processing units 341A to 341C perform the tracking processing in respectively different tracking systems, when the user selects one of the expanded images A′ to C′ displayed in the sub screen areas 282-1 to 282-3, the control unit 217 supplies the tracking processing result of the selected expanded image as the next initial value (x, y, z) to the tracking processing units 341A to 341C, so that the tracking position is reset.
In the example shown in
At a time x(20) in a state in which the display screen 360C (
At a time x(40) in a state in which the display screen 360G (
Similarly, at a time x(45) after a predetermined period of time from the time x(40), in a state in which the display screen 3601 (
As described above, the tracking processing results obtained when the tracking processing units 341A to 341C perform the tracking processing in the different tracking processing systems are displayed at the same time and presented to the user, so that the user can select the tracking processing result where advantages of the respective tracking processing systems appear. The expanded image (one of the expanded images A′ to C′) selected by the user and displayed in the main screen area 281 is recorded in the image recording unit 216, so that the user can obtain still more desired processing results.
Now, in the image processing apparatus 200 of
Also, it is supposed that the tracking processing result by the tracking processing unit 341C is selected by the user. That is, on the main screen area 281 of the display screen 400 of
Herein, when the tracking position 411C displayed in the main screen area 281 is set as the reference, the tracking position 411A is shifted in position in the x direction. On the other hand, the tracking position 411B is shifted in position in the y direction.
Under such a condition, for example, a display shown in B of
In B of
On the other hand, in C of
It should be noted that dotted lines in B of
When the display screens in B of
In view of the above, the image synthesis unit 214 changes the expanded images displayed in the sub screen areas 282-1 and 282-2 in accordance with the tracking processing results so as to be easily compared with the expanded image displayed in the main screen area 281.
To be more specific, the image synthesis unit 214 obtains a difference between the tracking position of the tracking processing unit 341 displayed in the main screen area 281 and the respective tracking positions of the unselected remaining two tracking processing units 341 as a tracking difference vector and further obtains a ratio of the horizontal component and the vertical component of the obtained tracking difference vector (the vertical component/the horizontal component). Then, the image synthesis unit 214 displays the expanded image in which the tracking position of the tracking processing unit 341 corresponding to the larger one among the obtained two ratios is set as the reference in the sub screen area 282-2 and the expanded image in which the tracking position of the tracking processing unit 341 corresponding to the small one is set as the reference in the sub screen area 282-1.
By doing so, the expanded images can be displayed in the sub screen areas 282-1 and 282-2 so as to be easily compared with the expanded image displayed in the main screen area 281.
The image processing unit 213-1 is composed of an expansion processing unit 242A″, and the image processing unit 213-2 is composed of an expansion processing unit 242B″. Also, the image processing unit 213-3 is composed of an expansion processing unit 242C″.
The control unit 217 is composed of a synchronization characteristic amount extraction unit 471, sequence reproduction units 472A to 472C, a switcher unit 473, and a control instruction unit 474.
Similarly as in the expansion processing units 242A′ to 242C′ according to the second embodiment, the expansion processing units 242A″ to 242C″ execute the expansion processing to be supplied to the expanded images after the processing to the image synthesis unit 214. The area in which the expansion processing units 242A″ to 242C″ respectively perform the expansion processing is a predetermined area of the image input decided by a zoom parameter (xa″, ya″, z) , (xb″, yb″, z) , or (xc″, yc″, z) supplied from the switcher unit 473. It should be noted that hereinafter, the expanded image expanded on the basis of the zoom parameter (xa″, ya″, z) is set as the expanded image A″, the expanded image expanded on the basis of the zoom parameter (xb″, yb″, z) is set as the expanded image B″, and the expanded image expanded on the basis of the zoom parameter (xc″, yc″, z) is set as the expanded image C″.
The expansion processing units 242A″ to 242C″ execute the expansion processing in the respectively difference systems. In the image synthesis unit 214, the expanded images after the expansion processing A″ to C″ are displayed on the display screen 270 composed of the main screen area 281 and the sub screen areas 282-1 to 282-3 shown in
The expansion processing By the expansion processing unit 242A″ can be set, for example, as a processing adopting the above-mentioned class classification adaptive processing. Also, the expansion processing performed by the expansion processing units 242B″ and 242C″ is set as, for example, a processing based on the linear interpolation, and intervals for interpolation can be set different from each other in the expansion processing units 242B″ and 242C″.
The synchronization characteristic amount extraction unit 471 of the control unit 217 stores therein a synchronization characteristic amount time table 461 (
The synchronization characteristic amount extraction unit 471 calculates (extracts) the synchronization characteristic amount of the input image supplied from the image distribution unit 212 and refer to the synchronization characteristic amount time table 461, so that the time code currently corresponding to the input image supplied from the image distribution unit 212 is detected. The synchronization characteristic amount extraction unit 471 supplies the detected time code to the sequence reproduction units 472A to 472C.
The sequence reproduction units 472A to 472C respectively store a parameter table in which time codes are associated with zoom parameters. The zoom parameters stored by the sequence reproduction units 472A to 472C as the parameter tables are mutually different. Therefore, the same time code is supplied from the synchronization characteristic amount extraction unit 471 to the sequence reproduction units 472A to 472C, but different zoom parameters are supplied to the switcher unit 473 from the respective sequence reproduction units 472A to 472C.
To be more specific, while corresponding to the time code supplied from the synchronization characteristic amount extraction unit 471, the sequence reproduction unit 472A supplies the zoom parameter (xa″, ya″, z) to the switcher unit 473. While corresponding to the time code supplied from the synchronization characteristic amount extraction unit 471, the sequence reproduction unit 472B supplies the zoom parameter (xb″, yb″, z) to the switcher unit 473. While corresponding to the time code supplied from the synchronization characteristic amount extraction unit 471, the sequence reproduction unit 472C supplies the zoom parameter (xc″, yc″, z) to the switcher unit 473.
The zoom parameter (xa″, ya″, z) denotes the center position (xa″, ya″) when the expansion processing is performed and the zoom rate z. The same applies to the zoom parameter (xb″, yb″, z) and (xc″, yc″, z) too. It should be noted that the zoom rate z is common among the zoom parameters output by the sequence reproduction units 472A to 472C, but the zoom rate z can also be set as different values in the sequence reproduction units 472A to 472C.
To the switcher unit 473, selection information is supplied from the control instruction unit 474 which is information indicating that the expanded image displayed in the main screen area 281 of the display screen 270 is selected while the user operates the remote commander or the like and indicating one of the expanded images A″ to C″.
The switcher unit 473 appropriately selects the zoom parameter (xa″, ya″, z), (xb″, yb″, z), and (xc″, yc″, z) so that the expansion processing at the highest image quality is performed on the expanded image indicated by the selection information to be supplied to the expansion processing units 242A″ to 242C″ on a one-on-one basis. That is, in a case where the expanded image A″ is supplied as the selection information, the switcher unit 473 supplies the zoom parameter (xa″, ya″, z) to the expansion processing unit 242A″, in a case where the expanded image B″ is supplied as the selection information, supplies the zoom parameter (xb″, yb″, z) to the expansion processing unit 242A″, and in a case where the expanded image C″ is supplied as the selection information, supplies the zoom parameter (xc″, yc″, z) to the expansion processing unit 242A″.
The control instruction unit 474 supplies the expanded image instructed by the user by operating the remote commander or the like to be displayed in the main screen area 281 of the display screen 270 to the switcher unit 473 as the selection information. Also, the control instruction unit 474 supplies the operation information indicating the operations of the down key DN, the up key, and the decision key RTN, and the like of the remote commander to the image synthesis unit 214.
The image synthesis unit 214 generates a synthesized image in which the expanded images A″ to C″ supplied from the expansion processing units 242A″ to 242C″ and the image input supplied from the image distribution unit 212 are synthesized to be supplied to the image presentation unit 215. Herein, the image synthesis unit 214 generates the synthesized image so that the expanded image supplied from the expansion processing unit 242A″ is arranged in the main screen area 281 of the display screen 270.
Also, on the basis of the operation information from the control instruction unit 474, the image synthesis unit 214 performs the highlight display 291 in a predetermined area in the sub screen areas 282-1 to 282-3.
Next, with reference to
First, the synchronization characteristic amount extraction unit 471 calculates the synchronization characteristic amount of the input image supplied from the image distribution unit 212.
Then, the synchronization characteristic amount extraction unit 471 detects the time code having the same synchronization characteristic amount from the synchronization characteristic amount time table 461. In the time table 461 of
Next, with reference to
In the initial state, the display screen 480A is displayed. In the display screen 480A, the expanded image A″ expanded on the basis of the zoom parameter (xa″, ya″, z) instructed by the sequence reproduction unit 472A is displayed in the main screen area 281. The image input is displayed in the sub screen area 282-1. The expanded image B″ expanded on the basis of the zoom parameter (xb″, yb″, z) instructed by the sequence reproduction unit 472B is displayed in the sub screen area 282-2. The expanded image C″ expanded on the basis of the zoom parameter (xc″, yc″, z) instructed by the sequence reproduction unit 472C is displayed in the sub screen area 282-3.
That is, the initial state, as shown in
While returning back to
In a state in which the display screen 480B is displayed, when the user operates the down key DN of the remote commander, the image presentation unit 215 displays the display screen 480C in which the highlight display 291 is moved to the sub screen area 282-2.
Next, in a state in which the display screen 480C is displayed, when the user operates the decision key RTN of the remote commander, the selection information indicating that the expanded image B″ is selected is supplied from the control instruction unit 474 to the switcher unit 473.
As shown in
As a result, the display screen 480D of
Next, with reference to a flow chart of
First, in step S141, the image distribution unit 212 determines whether or not the image input is supplied from the image input unit 211. In step S141, in a case where it is determined that the image input is not supplied, the processing is ended.
On the other hand, in step S141, in a case where it is determined that the image input is supplied from the image input unit 211, the processing proceeds to step S142, and the image distribution unit 212 distributes the supplied image input. That is, the image distribution unit 212 supplies the image input to the synchronization characteristic amount extraction unit 471, the sequence reproduction units 472A to 472C, the expansion processing units 242A″ to 242C″, and the image synthesis unit 214.
In step S143, the synchronization characteristic amount extraction unit 471 calculates the synchronization characteristic amount of the input image supplied from the image distribution unit 212, and in step S144, by referring to the time table 461, the time code corresponding to the calculated synchronization characteristic amount is detected. The detected time code is supplied to the sequence reproduction units 472A to 472C.
In step S145, the sequence reproduction units 472A to 472C refer to the parameter table in which the time codes are associated with the zoom parameters and supply the zoom parameter corresponding to the supplied time code to the switcher unit 473. The sequence reproduction unit 472A supplies the zoom parameter (xa″, ya″, z) to the switcher unit 473, and the sequence reproduction unit 472B supplies the zoom parameter (xb″, yb″, z) to the switcher unit 473. The sequence reproduction unit 472C supplies the zoom parameter (xc″, yc″, z) to the switcher unit 473.
In step S146, on the basis of the selection information from the control instruction unit 474, the switcher unit 473 supplies the zoom parameter supplied from the sequence reproduction units 472A to 472C to the expansion processing units 242A″ to 242C″.
In step S147, the expansion processing units 242A″ to 242C″ respectively execute the expansion processing to supply the expanded images after the processing to the image synthesis unit 214.
In step S148, the image synthesis unit 214 uses the input image supplied from the image distribution unit 212 and the expanded images A″ to C″ subjected to the expansion processing to generate a synthesized image to be supplied to the image presentation unit 215. Herein, the image synthesis unit 214 generates the synthesized image so that the expanded image supplied from the expansion processing unit 242A″ is displayed in the main screen area 281. Also, the image synthesis unit 214 supplies the main image which is the image arranged in the main screen area 281 among the synthesized image to the image recording unit 216.
In step S149, the image presentation unit 215 displays the synthesized image supplied from the image synthesis unit 214 in a predetermined display unit to be presented to the user. Also, in step S149, the image recording unit 216 records the main image supplied from the image synthesis unit 214 on the predetermined recording medium. After the processing in step S149, the processing returns to step S141, and the subsequent processing is repeatedly executed.
As described above, in the third image processing by the image processing apparatus 200 of
The expanded images A″ to C″ subjected to the expansion processing are images where areas decided by the different zoom parameters (xa″, ya″, z), (xb″, yb″, z), and (xc″, yc″, z) are expanded, which are therefore respectively different images. Therefore, by sequentially switching the image input or the expanded images A″ to C″, the user can perform editing as if the inputs from a plurality of camera image frames are switched in the main screen area 281. Also, other unselected image input or expanded images A″ to C″ are also displayed in the sub screen areas 282-1 to 282-3, and therefore the user can perform editing while performing the comparison.
Also, the switcher unit 473 switches the zoom parameters supplied to the expansion processing units 242A″ to 242C″ in accordance with a selection of the user, and thus the main image displayed in the main screen area 281 and the main image recorded in the image recording unit 216 can be set as the expanded images expanded by the expansion processing regularly having the high image quality. According to this, it is possible to perform the expansion processing at the high image quality on the expanded image to be viewed on a large screen or to be recorded, and on the other hand it is possible to adopt an inexpensive processing unit on the expanded image which is not necessary for the recording. Thus, the overall cost can be suppressed, and also the performances which the expansion processing units 242A″ to 242C″ have can be effectively utilized. That is, it is possible to effectively distribute the resource for the processing.
Next, a modified example of the third embodiment will be described.
In the above-mentioned example, as shown in A of
To be more specific, the image synthesis unit 214 calculates, periodically or at a predetermined timing, a correlation value corr of the image displayed in the main screen area 281 for the display and the respective three images input in the sub screen areas 282-1 to 282-3 for the display through the following Expression (29).
In Expression (29), pv1(x, y) denotes a luminance value in a predetermined position (x, y) of the image displayed in the main screen area 281, pv2(x, y) denotes a luminance value in the position (x, y) corresponding to one of the comparison target images displayed in the sub screen areas 282-1 to 282-3, pv1
Then, the image synthesis unit 214 generates a synthesized image in which the display is performed from the top of the sub screen areas 282-1 to 282-3 in the descending order of the three calculated correlation values corr to be supplied to the image presentation unit 215. According to this, the user can easily find a desired image from the switching candidate images.
As described above, according to the above-mentioned first to third embodiments, a plurality of different image processings are applied on the input moving image, and the images after the processings are displayed at the same time, so that it is possible to easily perform the comparison.
It should be noted that in the above-mentioned example, the image synthesis unit 214 generates the synthesized image while corresponding to the display screen 270 (
The other synthesis methods by the image synthesis unit 214 include synthesis methods shown in A of
As shown in B of
Also, the highlight display can be set as a display other than the frame display enclosing the surrounding of the area shown in
The above-mentioned series of processings can be executed by hardware and can also be executed by software. In a case where the above-mentioned series of processings is executed by the software, a program structuring the software is installed from a recording medium into a computer incorporated in dedicated-use hardware or, for example, a general-use personal computer or the like which is capable of executing various functions by installing various programs.
In the computer, a CPU (Central Processing Unit) 601, a ROM (Read Only Memory) 602, and a RAM (Random Access Memory) 603 are mutually connected by a bus 604.
To the bus 604, furthermore, an input and output interface 605 is connected. To the input and output interface 605, an input unit 606 composed of a key board, a mouse, a microphone, or the like, an output unit 607 composed of a display, a speaker, or the like, a storage unit 608 composed of a hard disk drive, a non-volatile memory, or the like, a communication unit 609 composed of a network interface or the like, and a drive 610 for driving removable media 611 such as a magnetic disk, an optical disk, an opto-magnetic disk, or a semiconductor memory are connected.
In the computer configured as described above, the CPU 601 loads, for example, the programs stored in the storage unit 608 via the input and output interface 605 and the bus 604 onto RAM 603 to be executed, so that the above-mentioned first to third image processings are performed.
The programs executed by the computer (the CPU 601) are recorded, for example, in the removable media 611 which is package media composed of the magnetic disk (including a flexible disk), the optical disk (CD-ROM (Compact Disc-Read Only Memory), DVD (Digital Versatile Disc) or the like), the opto-magnetic disk, the semiconductor memory, or the like, or are provided via a wired or wireless transmission medium such as a local area network, the internet, or digital satellite broadcasting.
Then, the programs can be installed into the storage unit 608 via the input and output interface 605 by mounting the removable media 611 to the drive 610. Also, the programs can be received by the communication unit 609 via the wired or wireless transmission medium and installed into the storage unit 608. In addition, the programs can be previously installed into the ROM 602 or the storage unit 608.
It should be noted that the program executed by the computer may be a program where the processings are performed in a time sequence manner while following the order described in the present specification but may also be a program where the processings are performed in parallel or at a necessary timing when a call is performed or the like.
In the present specification, steps described in the flow chart of course include the processing executed in a time sequence manner while following the stated order but also the processing executed in parallel or individually instead of not necessarily being processed in the time sequence manner.
Also, in the present specification, the system represents an entire apparatus composed of a plurality of apparatuses.
Embodiments of the present invention is not limited to the above-mentioned embodiments, and various modifications can be made without departing from the gist of the present invention.
Number | Date | Country | Kind |
---|---|---|---|
2007-187879 | Jul 2007 | JP | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/JP2008/062997 | 7/18/2008 | WO | 00 | 1/12/2010 |