The field of the invention is that of signal compression, in particular of a digital image or a sequence of digital images, divided into blocks of pixels.
The invention more particularly relates to the transformation of a block of pixels, originating from a prediction or not. It finds particular application in a transform competition context.
The coding/decoding of digital images applies in particular to images from at least one video sequence comprising:
The present invention applies similarly to the encoding/decoding of 2D or 3D type images.
The invention may especially, but not exclusively, apply to video coding implemented in current AVC video coders (for “Advanced Video Coding”) and HEVC (for “High Efficiency Video Coding”) and their extensions (MVC, 3D-AVC, MV-HEVC, 3D-HEVC, post-HEVC, etc.), and the corresponding decoding.
We consider a conventional compression scheme of a digital image, according to which the image is divided into blocks of pixels. A current block to be coded, which constitutes an initial coding unit, is generally divided into a variable number of sub-blocks according to a predetermined cutting mode. In relation to
Each CU undergoes an encoding or decoding operation consisting of a sequence of operations, including in a non-exhaustive manner a prediction, a residue calculation, a transformation, a quantisation and an entropy coding. This sequence of operations is known from the prior art and presented in relation to
During a step E0, the first block CTU to be processed is selected as the current block c. For example, this is the first block (in lexicographic order). The size of this block is for example 64×64 according to the HEVC standard.
It is assumed that there are D partitions in possible CUs numbered from 1 to D, with D a non-zero integer and that the partitioning used on block c corresponds to the partitioning number d. For example, there can be 4 possible sizes of square partitioning in sub-blocks of size 4×4, 8×8, 16×16, and 32×32 according to a regular cutting mode of type “quad tree” as specified in the HEVC standard. Partitioning into rectangular sub-blocks is also possible.
In the following, we denote by current block a sub-block P from the partitioning of the CTU block c. The steps that will be described are repeated for the other sub-blocks. During a step E1, a prediction Pr of the block CU P is determined. It is a prediction block constructed by known means, typically by motion compensation (block from a previously decoded reference image), or by intra prediction (block constructed from decoded pixels belonging to the image ID). The prediction information related to Pr is coded in the bit stream TB or compressed file FC. We assume here that there are P possible prediction modes m1, m2, . . . , mp, with P a non-zero integer. For example, the prediction mode chosen for the current block x is the mode mp. Some prediction modes are associated with an Intra type prediction, others with an INTER type prediction.
During a step E2, an original residue R is formed by subtracting R=P−Pr from the prediction Pr of the current block P, at the current frame P.
In E3, we identify a transform T to apply to the residue R or to a sub-block derived from a subdivision of R.
The transformation step plays a crucial role in such a video coding scheme: indeed, this step concentrates the information before the quantisation operation. As a result, a set of pixels before coding is shown on a small number of non-zero frequency coefficients representing the same information. Thus, instead of transmitting a large number of coefficients, only a small number will be needed to faithfully reconstruct a block of pixels.
This transformation step is complex to implement both coder side and decoder side, which must implement the inverse transformation of that applied by the coder.
During a step E4, the residue R is transformed into a transformed residue block, so-called RT, by the identified transform. Alternatively, in the absence of prediction, a transformed block RT is obtained from the block c. This is for example a block-type transform or a wavelet transform, all known to those skilled in the art and in particular implemented in the JPEG/MPEG standards for the DCT/DST and JPEG2000 for the wavelet transform.
In image and video coding, orthogonal or quasi-orthogonal block transforms (4×4, 8×8, etc.) are generally used. The most used transforms are based on cosines or sines. They are usually referred to as DTT (for “Discrete Trigonometric Transforms”). The DCT is thus present in most standards for image and video. Recently the HEVC standard has also introduced the DST (for “Discrete Sine Transform”) for the coding of particular residues in the case of 4×4 blocks.
In fact, approximations of these transforms are used, the calculations being carried out on integers. In general, the bases of transforms are approximated to the nearest integer, after multiplication by a factor which conditions the precision given to the approximation (this factor is often at the power of 2 generally of 8 or 10 bits).
As an example, transforms used by the HEVC standard on 4×4 block size are presented in connection with
In E5, in a manner known in the state of the art, these coefficients are traversed in a predetermined order so as to constitute a one-dimensional vector RQ[j], where the index j varies from 0 to Nb−1, with Nb an integer equal to the number of pixels in the block x. The index j is called the frequency of the coefficient RQ[j]. Classically, these coefficients are scanned by globally increasing or decreasing order of frequency values, for example according to a predetermined path, for example diagonal or horizontal.
In E6, the transformed block RT is quantised by conventional quantisation means, for example scalar or vector, into a quantised block RQ comprising as many coefficients Nb as the block RT.
During a step E7, the information relating to the coefficients of the block RQ is coded by entropy coding, for example according to a Huffman coding or arithmetic coding technique. This information includes at least the amplitude of the coefficients and their sign. By amplitude, here is meant the absolute value of the coefficient. Conventionally, a piece of information representative of the fact that the coefficient is non-zero, can be coded for each coefficient. Then, for each non-zero coefficient, one or more pieces of information relating to the amplitude are coded. Coded amplitudes CA are obtained. The signs of non-zero coefficients are also coded. In general, they are simply coded by a bit 0 or 1, each value corresponding to a given polarity. Such coding provides efficient performances because, due to the transformation, the values of the amplitudes to be coded are mostly equal to zero.
Concerning the applied transform, in the case of the HEVC standard, the decoder is indicated by a bit, so-called “transform_skip_flag”, the inverse transform to be applied among the two alternatives DST or absence of transform. This case occurs in the case of 4×4 size blocks.
In E8, the coded data relating to the current block x are inserted in the bit stream TB.
The other sub-blocks which constitute the block R of the image I1 are processed in the same way, then the other blocks CTU of the image I1, as well as the blocks of the following images of the sequence.
The step of transforming a current block into a transformed block generates complex calculations, which the person having skills in the art has attempted to simplify and/or accelerate.
This complexity increases with the size of the processed blocks. In HEVC, the transforms can reach a size of 32×32.
We especially know from the paper of Markus Püschel entitled Algebraic Signal Processing Theory: Cooley-Tukey Type Algorithms for DCTs and DSTs, published in IEEE Transactions on Signal Processing, in April 2008, a methodology for rapid implementation of the families of DCT or DST transformations. This method relates in particular to trigonometric DCT and DST transforms of types I to VIII and it is shown that fast algorithms for very particular sizes (typically to the power of 2 or close to a power of two) can be implemented.
A disadvantage of these types of DCT or DST transforms is that all of them are not applicable to all block sizes, especially among those specified in the HEVC standard. For example, the DST as specified in HEVC is limited to 4×4 block sizes because it induces significant complexity for larger sizes.
Another disadvantage is that some transforms do not have fast algorithms.
These objectives, as well as others that will appear later, are achieved by means of a method for decoding at least one digital image, from a bit stream comprising coded data representative of said image, said image being divided into a plurality of blocks processed in a defined order, said method comprising the following steps, implemented for a block, so-called transformed current block:
Such a method is particular in that at least one of said first and second transformation substeps comprises, for a row vector respectively a column vector, said an input vector:
We consider a separable transformation implemented from two orthogonal or quasi-orthogonal transforms.
The invention is based on an entirely new and inventive approach, whereby at least one of the orthogonal transformation substeps is replaced with at least two partial transformations of the row vectors or column vectors of the block to be transformed. Each partial transformation applies to a subvector of the row vector or column vector to be processed and is the only transformation applied to its elements, since the coefficients resulting from this transformation are directly placed back into an output vector intended for the next subtransformation step. There is therefore no recombination of the values obtained with other data resulting from a transformation of other elements of the input vector.
According to the invention, the sum of the sizes of the formed subvectors is equal to the size of the input vector, and the composition of the transformed vector comprises the insertion of elements of the first and of the at least one transformed second subvector to the initial positions of the first subvector and of the at least a second subvector formed in the input vector.
Thus, all elements of the input vectors, i.e. row or column vectors, are each processed by one and only one subtransform. As proved later in the description, breaking down into partial subtransformations proposed by the invention leads to a number of operations less than or equal to that necessary to apply a single transform of size equal to that of the input vectors.
According to another aspect of the invention, the first and second subtransforms are of different types.
In a context of transform competition, the coder selects, for the current block, the transforms which provide the best rate-distortion compromise. When an association of partial subprocesses according to the invention is chosen for at least one of the first or second substeps, the resulting gain in coding complexity is added to the gain in compression.
According to another aspect of the invention, one of the at least two partial sub-transforms is an identity transform.
As a result, at least one subset of the pixels/coefficients of the vectors (rows or columns) is not transformed during the transformation substep, which has the effect of reducing the complexity of the overall transformation.
According to yet another aspect of the invention, the size of at least one of the subvectors of the input vector is odd in number.
One advantage is that it can be applied a partial subtransform of odd size, for which there is a fast algorithm. This is the case, for example, of DTT type or DCT I type transforms.
The method which has just been described in its various embodiments is advantageously implemented by a device for decoding a digital image, from a bit stream comprising coded data representative of said image, said image being divided into a plurality of blocks processed in a defined order, said device comprising the following units, capable of being implemented for a block, so-called transformed current block:
Such a device is particular in that at least one of said first and second transformation sub-units comprises, for a so-called row vector respectively a column vector, said an input vector:
Correlatively, the invention also relates to a method for coding a digital image, said image being divided into a plurality of blocks of pixels processed in a defined order, said method comprising the following steps, implemented for a current block of predetermined dimensions:
According to the invention, such a method is particular in that at least one of said first and second transformation substeps comprises, for a row vector respectively a column vector, so-called input vector:
The method for coding a digital image according to the invention performs the inverse transformation step of the one implemented by the method for decoding a digital image which has just been described with its different embodiments.
The coding method according to the invention is implemented advantageously by a device for coding a digital image, said image being divided into a plurality of blocks of pixels processed in a defined order, said device comprising the following units, capable of being implemented for a current block (x) of predetermined dimensions:
Such a device is particular in that at least one of said first and second transformation sub-units comprises, for a so-called row vector respectively a column vector, said an input vector:
The invention also relates to a signal carrying a bit stream comprising coded data representative of a digital image, said digital image being divided into blocks of processed pixels in a defined order, a transformed block being obtained by transforming the pixels of a current block, said transformation comprising a first substep intended to produce an intermediate block, that applies to the row vectors respectively column vectors of the current block, the second intended to produce a block of pixels applied to the column vectors respectively row vectors of the intermediate block, resulting from the first substep.
Such a signal is particular in that, at least one of said first and second transformation substeps comprises forming a first subvector of size K<N respectively N from adjacent elements of the row vector respectively column vector and at least one second subvector of size equal to N−K respectively M−K respectively from adjacent elements of the input vector, not included in the first sub-vector, such that the sum of the sizes of the formed subvectors is equal to the size of the input vector, the application of a partial subtransform of size K×K to said sub-vector and applying a second size subtransform (N−K)×(N−K) to said second sub-vector, and the building of a row vector respectively column vector transformed by inserting the first transformed subvector and the at least one second transformed subvector, said signal comprises an identifier representative of said partial sub-transform.
The invention further relates to a user terminal characterised in that it comprises the device for coding a digital image and the device for decoding a digital image which have just been described.
The invention also relates to a computer program comprising instructions for implementing the steps of a method for decoding a digital image as described above, when this program is executed by a processor.
The invention also relates to a computer program comprising instructions for implementing the steps of a method for coding a digital image as described above, when this program is executed by a processor.
These programs can use any programming language. They can be downloaded from a communication network and/or recorded on a computer-readable medium.
The invention finally relates to recording media, which are readable by a processor, whether or not integrated into the device for coding a digital image and to the device for decoding a digital image according to the invention, which might be removable, respectively storing a computer program implementing a coding method and a computer program implementing a decoding method, as described above.
Other features and advantages of the invention will appear more clearly on reading the following description of an embodiment of the invention, given as a simple illustrative and non-limiting example, and the appended drawings among which:
In the remainder of the description, the framework of a coding scheme of a digital image has been chosen, for example as previously described in relation with
Note however that the prediction step of a current block E1 previously described is optional and that the next step E4 of transformation can therefore be applied directly to the pixels of the current block.
Let us consider a rectangular current block of size N×M (N columns, M rows), with N and M non-zero integers and at least one transform Tl to be applied to this block. The transform Tl belongs to a set of L transforms {T0, T1, . . . Tl . . . , TL−1}, with L a non-zero integer.
In relation to
In E40, the M row vectors Ln[m] of size N are extracted from the current block x, with m an integer between 0 and M−1. In E41, the M vectors Ln[m] are transposed to be presented as column vectors of size N. In E42, they are transformed by a first transform B1 of size N×N. An intermediate block of size M×N is obtained. In E43, N column vectors of size M are extracted and then transposed into E44. The transposed vectors Col[n] with n between 0 and N−1 are transformed into E45 by a second transform Al of size M×M.
Thus A1 and B1 can operate as follows:
X
l
=A
l·(Bl·xt)t (1)
Al and Bl are transforms that can express themselves in the form of square matrices (respectively of size M×M and N×N) and which are respectively adapted to transform each constituent vector of the pixel block s. Xl is the block of coefficients resulting from the transformation of the block x by the transform Tl.
Xl can also be expressed in other equivalent forms, for example by omitting the transposition of the block x. In this case we perform the calculation:
X
l
=B
l·(Al·x)t (2)
In this case, the N column vectors are transformed at first with a size M of transformation, followed by transformation on the resulting rows by M transforms of size N.
In known manner, transforms are typically applied to row/column vectors of sizes equal to powers of 2. For example, the HEVC standard transforms row/column vectors of sizes 4, 8, 16 or 32.
It is assumed here that the first and second transforms Al, Bl are orthogonal or quasi-orthogonal transforms. Such transformations have a property that a transformation matrix multiplied by its transpose only exhibits the constant non-zero diagonal terms. In the quasi-orthogonal case, the terms outside the diagonal have negligible values with respect to the amplitude of the diagonal terms. So the inverse of an orthogonal matrix is approximated by its transpose within one factor, so that at decoding, the transpose of the matrix used in the coding can be applied.
It will be noted that the transforms Al and Bl and can also be applied through a rapid implementation by an algorithm that represents the form of a diagram composed of butterflies (that is to say, linear combinations of terms taken in pairs). This applies to all trigonometric transformations, as described in the article by Markus Püschel already cited.
One known way to appreciate the computational complexity of such a transformation step is to count the number of addition and multiplication operations required. Table 1 below gives some examples of computational complexity estimates made using this method.
There are the “general case” in which a transformation is applied in matrix form and particular cases (DCT I, DCT II) that rely on fast implementations of transformations under Article Markus Püschel cited above. The number of operations reported relates to the processing of a pixel vector.
A transformation matrix comprising N rows and N columns is now considered. In the general case, we have N multiplications and N−1 additions for each row of the matrix. As a result, we have N*N and N*(N−1) multiplications and additions in total. We then have a number of operations per transformed value of 2N−1.
It is noted that the complexity increases significantly with the size of the block to be processed. We also note that there are so-called fast transforms, which are more economical in computing resources.
In the case of quasi-orthogonal transforms, the inverse transformation being the same as applying the transformation with transposed coefficients, the computational complexity for the inverse is identical to the direct transformation.
In relation to
In a known manner, the first step E42 of subtransformation by the transform Bl can be carried out using a matrix Q of size N×N.
In the particular example in
For example, the step of transformation of the M transposed row vectors of size N LnT[m] by the particular subtransform Bl of size N×N is performed by applying a first subtransform R of size K×K to a subset of the pixels of the column vectors of the block and of a second subtransform S of size (N−K)×(N−K) to the rest of the pixels of the column vectors of this block. All the elements of a vector are thus processed only once by only one of the two subpartial transforms.
From a mathematical point of view, the transform Bl can then be expressed in the form of a matrix MBl comprising in its diagonal a submatrix MR of size K×K corresponding to the first subtransform R and a submatrix MS of size (N−K)×(N−K) corresponding to the second subtransform S, the rest of the elements of the matrix MBl is zero, as shown in
Thus, according to this embodiment of the invention, to transform a vector with the matrix MBl, it is sufficient to perform a transformation by the matrix MR, which processes the first K values of a vector LnT[m] and a transformation by the matrix MS which covers the remaining N−K pixels of this vector.
In relation with
These operations are repeated for the M transposed row vectors of the block to be transformed. The M transformed vectors are returned to their initial positions to form the intermediate block BI.
We will now show that splitting the matrix MBl, into one part MR processing K points and into one part MS processing N−K points (with K<N−K, without loss of generality), generates a reduction in complexity.
For a transformation of size N, applied in matrix form:
For a transformation of size K, applied in matrix form:
For a transformation of size N−K, applied in matrix form:
If we cumulate the computations associated with the transformations N−K and K, we count:
It will be noted that in the particular case of a type II DCT (refer to Table 1 above), Q is of size 8 and requires 40 operations in total, whereas the implementation of the transform of the partial transform of size 7 involves only 38 operations.
In the particular case of type 1 DCT of size 3 and 4 it is understood that it is notoriously more favourable to process a size 3 transform which requires 6 operations to be compared to the 12 operations required by size 4.
In a first particular example, N=8, K=4 and the complexity is evaluated if R and S are chosen as DCT II type transformations, for which a fast algorithm is available.
The computational complexity is then 12+12=24 operations for the 8 pixels constituting the vector, i.e. 3 operations per pixel. This is to be compared to a DCT II size 8 transform, which requires 5 operations per pixel.
In a second particular example, N=8, K=4 and the complexity is evaluated if R and S are chosen as transformations implemented in the form of a matrix product without any available fast algorithm.
The computational complexity is then 28+28=56 operations for the 8 pixels constituting the vector, i.e. 7 operations per pixel. This is to be compared to a size 8 transform, which requires 15 operations per pixel.
In a third particular example, N=8, K=4 and the complexity is evaluated if R is selected as a Type II DCT and S as an applied transform as a matrix product. This makes it possible to apply a suitable transformation to a part of the vectors of the block, whereas this adapted transform can for example be of the RDOT type as described in the O. Sezer article already cited.
The computational complexity is then 12+28=40 operations for the 8 pixels constituting the vector, i.e. 5 operations per pixel. This should be compared with a size 8 transform applied as a matrix product that requires 15 operations per pixel.
As a result, the invention notoriously reduces computational complexity. In the two cases presented above, 3/5=60% and 7/15=47% and 5/15=33% respectively of the complexity of a full size transformation are used.
According to a second embodiment of the invention, the first subtransform of size K×K is an identity-type transformation. It is expressed as a matrix containing only identical terms on its diagonal. The matrix MBl corresponding is shown on
By way of illustration, the identity transforms I3 of size 3×3, I2 of size 2×2 and I1 of size 1 are presented below:
In
In both cases, the application of the subtransform R=I1 respectively I2 requires no calculation in this particular case, so the application of the matrix MBl has the same complexity as the matrix MS.
Three cases are presented with R=I1 and S respectively taking the form of a matrix of size N−K=3, a DCT-II of size N−K=3 and a DCT-II of size N−K=7.
Table 2 shows that with this embodiment, the complexity ratio is therefore 15/28=53% (28 corresponds to the complexity due to a 4×4 matrix of any type) and 8/12=67% (12 corresponds to the case where there is a fast DCT II transform of size 4) and 38/40=95% of the initial complexity.
The first embodiment therefore allows a higher gain complexity in the latter mode.
We shall now describe a third embodiment of the invention which relates to a transform competition context. In connection with
We know from the article by Arrufat et al, entitled “Low Complexity Transform Competition for HEVC” published in the Proceedings of the ICASSP of March 2016, a selection method of a select group of transforms from a list of several “complete” transforms of separable type, optimised for distortion or based on discrete trigonometric transformations, so-called DTT, to which the DCT and DST transforms belong.
In a first step, according to the prior art, the compression performance of a coder with complete transformations, that is to say of sizes equal to those of the processed block, is assessed. To do this, we have a set of 16 types of trigonometric transforms of size 4, for example those for which we have a fast implementation as described in the Püschel article already cited.
For example, the list of the following transforms (8 types of DCT and 8 types of DST) is assessed: DCT I, DCT-II, DCT-III, DCT-IV, DCT-V, DCT-VI, DCT-VII, DCT-VIII, DST-I, DST-II, DST-III, DST-IV, DST-V, DST-VI, DST-VII, DST-VIII.
In total there are therefore 16×16=256 possible combinations when combining the vertical (Bl) and horizontal (Al) transforms.
To assess these different combinations of transforms, for each pair, a compactness in a distortion/parcimony plane, as described in the article by Sezer already cited, is measured on a set of residual image signals the following quantity:
xi are the residual blocks collected on a large set of various images. In this particular case, they are derived from 4×4 block size for intra prediction angular index 26 according to the HEVC standard.
are the residual blocks quantised by thresholding resulting from the transformation of xi by Tl composed of Al and Bl.
∥⋅∥0 represents the zero norm, that is the number of non-zero coefficients in .
N,M are the dimensions of the block and H is the number of blocks considered.
λ is a Lagrange multiplier that allows to set the constraint on the number of non-zero coefficients in , according to the Sezer publication cited above.
For example, the best transformations Al, Bl, i.e., those producing the 5 weakest measures j(λ) are chosen. By learning, in accordance with the publication of Arrufat, the most effective five pairs of transforms are selected to code a set of residual blocks from a vertical prediction. In this example, at the end of the learning phase, the set of the 5 best pairs of transforms is that of table 3:
The compactness measurement (equation (1)) produces a value of 29.76 for this combination of 5 pairs.
According to the third embodiment of the invention, combinations of partial subtransformations of sizes K×K and (N−K)×(N−K) are added to the initial set of subtransforms considered for learning. In the example of a 4×4 block, these are partial subtransforms of sizes 1×1 and 3×3 or 2×2 and 2×2. The subtransformations in this embodiment are of identity type. They are denoted ID1 and ID2 for the respective sizes 1 and 2.
These combinations are advantageously proposed in association with a complete transform. In other words, at least two partial subtransforms are broken down, either the transform Al or the transform Bl.
Learning is carried out from the set of transforms thus completed and, analogously to the previous case, the 5 pairs A1, B1 which contribute to the smallest compactness measurement are selected.
We obtain the following selection of transformations:
A compactness measurement equal to 29.34 is obtained for this new combination, which is equivalent to an improvement of 1.38% in compactness compared to the state-of-the-art embodiment.
The invention is therefore remarkable in that it offers, in addition to a reduction in the number of operations as shown above, in relation to the first two embodiments, an improvement in the compression of the signal.
It should be noted that in the combinations presented, at least one of the partial subtransforms is equal to the identity and the other to a DST or DCT. Of course, the invention is not limited to these examples and covers the case of combinations of at least two more complex partial subtransforms, such as for example an ID and an adapted RDOT transform or a DST VII and an adapted RDOT transform.
The coding method produces a coded data stream representative of the at least one input image. The bit stream TB produced by the coding method according to the invention is for example transmitted in the form of a signal to a decoder or a receiver comprising a decoding device via a telecommunication network.
It is assumed that the bit stream TB has been received by a decoding device implementing the decoding method according to the invention. This decoding method will now be described in relation with
In D0, the first block to be processed is initially selected as the current block C′. For example, this is the first block (in lexicographic order). This block has Nb pixels, for example 64×64 as specified in the HEVC standard for CTU blocks.
As described for the encoding method, the block current considered in the following may be the block C′ itself or can be derived from a partition of the block C′ into sub-blocks CU. It may further be a residue of a block obtained by subtracting a prediction of a sub-block CU or of the current block. We shall also designate by current block a sub-block derived from a division of a sub-block CU, residual or not, before its transformation by a transform T1.
During a step D1, the coded data relating to the current block C′ are read. The coded data includes coding parameters, such as, for example, the prediction mode used, or in a context of transform competition, a marking of an identifier of the transform applied to the current block and the values relating to the amplitudes and to the signs of the quantised residual coefficients of the current block.
When the determined prediction mode indicates that a prediction has been made by the coder, the current block is predicted at D2, according to the prediction mode determined from an already processed block. A predicted block Pr′ is obtained.
During a step D3, the data coded representative quantised values, residual of the current block (values and signs of the coefficients) are decoded, and a one-dimensional vector of values RQ′[i] is formed with i being an integer between 0 and M×N−1. It is understood that this is the inverse operation of that of entropy coding previously described in relation to the coding method.
In D4, the data of the current block RQ′[i] are dequantised. A vector R′[i] is obtained.
In D5, a reorganisation of the data of the one-dimensional residual vector in the current block is performed, in a process that is the inverse of the current block path described in step E5 of
In D6, the transform to be applied to the current block is identified. In known manner, and for example in accordance with the specifications of the HEVC standard, the decoder can access the identifier ID-TR of this transform, which has been previously associated with the prediction mode of the current block. More particularly, in a context of transform competition, the identifier of the transform can be received in the bit stream as a coding parameter, read during step D1 and then decoded. Whether in a context of transform competition or not, this transform identifier actually makes it possible to identify the two orthogonal or quasi-orthogonal transforms Al and Bl, which are successively applied to the current block.
According to the invention, at least one of the two substeps of transformations being implemented by the implementation of at least two partial subtransforms to disjoint subsets of elements of the column vectors respectively row vectors of the current block R′, the step of identification of the transform therefore comprises obtaining an ordered list of identifiers of at least three transform identifiers, including two consecutive identifiers correspond to the partial sub-transforms of one of the transformation substeps and one identifier to the transform implemented in the other substep (Al or Bl). For example, for the second row of the table 4, the identifier obtained will include Al=ID1, DST-IV and Bl=DCT-V.
In a step D7, the transforms (Al, Bl) corresponding to the identifiers are applied to the dequantised data obtained at D6. This transform corresponds to the reverse operation to that performed at the coder. For example, if at coding, the transform Bl was broken down into two partial subtransforms R and S, as shown in connection with
Since the subtransforms applied to the decoder are transposed from those applied to the coder, the complexity generated is the same and therefore has an advantage over the state of the art.
In a step D8, the block of pixels c′ of the decoded image is reconstructed from the block r′ obtained and integrated with the image ID being decoded. If the block is a residual block, it is added the prediction Pr′ of the current block obtained during step D2.
During a step D9, whether the current block is the last block to process the decoder, will be tested, given the order of travel defined above. If yes, the decoding process has finished its processing. If no, the next step is the step of selecting the next block D0 and the decoding steps D1 to D9 previously described are repeated for the next block selected.
It will be noted that the invention which has just been described can be implemented by means of software and/or hardware components. In this context, the terms “module” and “entity”, used in this document, can correspond either to a software component, or to a hardware component, or to a set of hardware and/or software components, capable of implementing the function(s) described for the module or entity concerned.
In relation to
For example, the device 100 comprises a processing unit 110, equipped with a processor μ1 and driven by a computer program Pg1 120, stored in a memory 130 and implementing the method according to the invention.
At initialisation, the code instructions of the computer program Pg1 120 are for example loaded into a RAM before being executed by the processor of the processing unit 110. The processor of the processing unit 110 implements the steps of the method described above, according to the instructions of the computer program 120.
In this embodiment of the invention, the device 100 comprises at least one transforming unit (TRANS) of the current block into a transformed block, said current block comprising M row vectors and N column vectors. The transforming unit is adapted to apply a first subtransform of the N row vectors transposed for providing an intermediate residual block formed from the transposed row vectors transformed and a second subtransformation of the M transposed row vectors of the intermediate block into transposed row vectors transformed from which the transformed block is formed. The device 100 further comprises an ENC coder of the transformed block coefficients and a bit stream INSERT constructor comprising coded data representative of the coded picture which is capable of inserting the coded coefficients. According to the invention, for at least one of said subtransformations, the transformer comprises the sub-units for forming at least a first subvector of size K<N or N respectively from adjacent elements of the input vector, for transforming the first subvector into a first transformed subvector by applying a partial subtransform of K×K size and building the transformed vector by insertion of said at least one first transformed subvector.
Advantageously, such a device 100 can be integrated into a user terminal TU. The device 100 is then arranged to cooperate at least with the next module of the terminal TU:
In relation to
For example, the device 200 comprises a processing unit 210, equipped with a processor μ2 and driven by a computer program Pg2 220, stored in a memory 230 and implementing the method according to the invention.
At initialisation, the code instructions of the computer program Pg2 220 are for example loaded into a RAM before being executed by the processor of the processing unit 210. The processor of the processing unit 210 implements the steps of the method described above, according to the instructions of the computer program 120.
In this embodiment of the invention, the device comprises a decoder (DEC) of the coefficients of the current block transformed from coded data read in the bit stream, a reverse transformer (TRANS−1) of the current block transformed into a decoded block, adapted to provide two successive reverse subtransformations, the first subtransformation producing an intermediate block and applied to column vectors, respectively row vectors of the current block, the second producing a block of pixels applicable to the row vectors, respectively the column vectors of the intermediate block, resulting from the first subtransformation.
According to the invention, the reverse transformer comprises the sub-units for forming at least one first subvector of size K<N or N respectively from adjacent elements of the input vector, for transforming the first sub vector into a first subvector transformed by application of a partial subtransform of size K×K and building the transformed vector by insertion of said at least one first transformed subvector.
The decoding device 200 further comprises an image constructor (RECONST), decoded from the decoded block.
Advantageously, such a device 200 can be integrated into a user terminal TU. The device 200 is then arranged to cooperate at least with the next module of the terminal TU:
An exemplary embodiment of the invention improves the situation discussed above with respect to the prior art.
An exemplary embodiment of the invention particularly overcomes these disadvantages of the prior art.
More precisely, an exemplary embodiment of the invention proposes a solution that makes it possible to reduce the complexity of the calculations implemented when applying a transform to a block of pixels, to the coder as well as to the decoder.
An exemplary embodiment of the proposes a gain in complexity that has no impact on the efficiency in compression.
It goes without saying that the embodiments which have been described above have been given for purely indicative and non-limiting reasons, and that many modifications can easily be made by those skilled in the art without departing from the scope. of the invention.
Number | Date | Country | Kind |
---|---|---|---|
16 53704 | Apr 2016 | FR | national |
This application is a Section 371 National Stage Application of International Application No. PCT/FR2017/050817, filed Apr. 5, 2017, the content of which is incorporated herein by reference in its entirety, and published as WO 2017/187040 on Nov. 2, 2017, not in English.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/FR2017/050817 | 4/5/2017 | WO | 00 |