Optimized video compression from a single process step

Abstract
A method of optimizing data compression in a data compression process. A single compressor is provided for performing video signal data compressor operations (122). The compressor receives a video signal that includes spatial and temporal data from a plurality of video signal frames. The compressor performs motion compensation (124, 126) on the plurality of video frames by simultaneously registering the plurality of video signal frames to produce a plurality of motion vectors. A performance metric including energy compaction parameters is then generated from the plurality of motion vectors (128). The performance metric is applied to the motion compensation step (124, 126) to optimize rate distortion performance in the motion compensation step and in subsequent data compression process steps.
Description

BACKGROUND OF THE INVENTION
1. Technical Field
This invention relates generally to digital video signal compression, and more particularly to a method of digital video signal energy compaction method that determines rate distortion performance during the initial motion compensation step, thereby achieving a high signal compression ratio and a low transmission rate.
2. Discussion
Present video applications require large amounts of data to be transmitted at high bit rates and with a minimal amount of signal distortion. For example, the uncompressed data bit rates for digital video, such as VCR-grade video (SIF), broadcast television video (CCIR-601) and high definition television (HDTV) are 4 Mbps, 15 Mbps, and 16Mbps, respectively. In an uncompressed state, these data bit rates are too high to allow such video signals to be transmitted and processed in a commercially feasible manner. Therefore, in order to process such video signals in a practical manner, such video signals must be compressed prior to being transmitted.
In response to the proliferation of video based applications and products, an industry wide need for creation of a standard video signal compression syntax arose. A group under the International Standards Organization (ISO), known informally as the Moving Pictures Experts Group (MPEG), was formed to define standards for digital video and audio compression. MPEG has created a standardized syntax by defining the content of a compressed video signal bit stream and the method of decompressing the bit stream subsequent to its transmission. The methods of compression, however, have not been defined, thus allowing individual manufacturers to develop various methods of actually compressing the data bit stream within the defined standards.
MPEG has to date defined two syntaxes widely used in the digital videos industry. A syntax known as MPEG-1 was defined to be applicable to a wide range of bit rates and sample rates. Particularly, MPEG-1 is suitable for use in CD/ROM applications and other non-interlaced video applications having transmission rates of about 1.5 Mb/s. A second syntax known as MPEG-2 was defined for representation of broadcast video, and other video signal applications having coded bit rates of between 4 and 9 Mb/s. MPEG-2 syntax is also applicable to applications such as HDTV and other applications requiring efficient coding of interlaced video.
While the above discussed MPEG-1 and MPEG-2 syntaxes exhibit adequate performance characteristics, the ongoing evolution of digital video dictates the need for further advancement in the art, as the present MPEG video syntax definitions do have associated limitations. For example, temporal redundancy, a phenomenon which enhances video compression by minimizing data bit rate transmission for non-changing video pixels, is an efficient method of maximizing video signal compression. Present MPEG-1 and 2 data compression-based methods utilize temporal compression.
However, the MPEG-1 and 2-based temporal compression is based on a frame by frame judgement basis so that the methods do not take full advantage of temporal compression. In particular, commercially standard MPEG-1 and MPEG-2 syntaxes only partially utilize temporal redundancy. In addition, present MPEG syntax requires numerous optimization variables (such as predictive frame, bi-linear frame and intraframe variables) to be calculated, transmitted and then decoded by MPEG signal decompressors. The use of these numerous variables adds both computational time and complexity to the data compression. Also, while many current MPEG syntaxes exhibit an associated bit rate compression of as high as 30:1, increasingly complex and data-intensive video applications require higher compression rates for real-time processing. Although data compression methods claiming compression ratios as high as 200:1 do exist, such methods subsample an array of pixels in a video frame sequence (i.e., throw away every other pixel) and utilize other shortcuts to achieve high compression.
Further, in the above-discussed MPEG-1 and MPEG-2 syntaxes, a data compression optimization operation known as Block Matching Algorithm (BMA) is the most computationally intensive operation in the MPEG video compression process. The BMA operation, although it is exhaustive, requires an inordinate amount of processing power. For example, LSI Logic of Milpitas, Calif. presently manufactures a ASIC chip, commercially sold under the name AMEP, which is capable of performing computations at a rate equivalent to 500 Intel Pentium 166 MHz processors. Two of these chips are required just to perform the BMA step, while typically only three additional chips are required to perform the remaining data compression calculations. As a result, the MPEG-2 committee has recently called for new alternative methods to improve upon the existing Block Matching Algorithm (BMA), both to lessen the complexity of the data compression operation and to lower the associated data transmission rate.
Data compression optimization methods which minimize data transmission rate do exist. For example, pending U.S. patent application entitled "Method and System of Three Dimensional Compression of Digital Video Systems", Ser. No. 08/621,855 by Niesen, now U.S. Pat. No. 5,933,143 assigned to TRW, Inc., which is incorporated by reference, discloses video data compression optimization methods performed after video compression transform operations by optimizing video compression based on transform coefficient transmission order. The methods are realized in the final steps of the process after motion compensation and the transform coefficients have been calculated.
While compression methods such as those identified above represent a significant advancement in data compression, there is still room for advancement in the art. In particular, there is a need for a data compression maximization method that is realized in the initial steps of the data compression process to minimize rate distortion as early in the process as possible to thereby optimize the data compression achieved in subsequent steps.
SUMMARY OF THE INVENTION
In accordance with the teachings of the present invention, a method of optimizing digital video signal compression is provided. The method of the present invention exhibits a higher compression ratio than current MPEG standards. At the same time, the method can be a less computationally intensive operation than the conventional BMA. In particular, the video signal compression method of the present invention, which is based on a first order Markov model, maximizes signal rate distortion performance early in the compression process by applying a compaction performance metric to the data compression motion compensation step. Because the motion compensation step is linked directly to rate distortion performance of all remaining data compression steps, data compression is increased. The method may be realized in an MPEG compatible format, thereby allowing the method to be used with current digital video signal applications.
In the inventive approach, a method is provided for compressing a digital video signal. In its present form, the method uses a subset of Discrete Cosine Transform (DCT) coefficients generated in the compression process to approximate the status of remaining coefficients. The method includes the step of providing a compressor for performing video signal data compression operations. A video signal is received at the compressor and includes spatial and temporal data from a plurality of video signal frames. The compressor compensates for motion in the plurality of video frames, and generates a performance metric to maximize energy compaction. The compressor applies the performance metric at the step of compensating for motion to optimize rate distortion performance in subsequent data compression process steps.





BRIEF DESCRIPTION OF THE DRAWINGS
Other objects and advantages of the invention will become apparent upon reading the following detailed description and upon reference to the drawings, in which:
FIGS. 1a and 1b are block diagrams for the signal compressor and decompressor, respectively, in which the signal compression methods of the present invention are realized;
FIG. 2 is a flow diagram of a prior art MPEG data compression process;
FIG. 3 illustrates a group of video frames processed by the motion compensation step of the MPEG-2 data compression process illustrated in FIG. 2;
FIG. 4 illustrates the motion compensation grouping of the frames of FIG. 3;
FIG. 5 illustrates the motion compensation processing sequence of the frames of FIG. 3;
FIG. 6 illustrates the group of video frames of FIG. 3 after being subjected to a 3-D discrete cosine transform;
FIG. 7 shows a three-dimensional quantization matrix containing three-dimensional coefficients after a one dimension discrete cosine transform is performed on all of the video frames of FIG. 3;
FIG. 8 is a flow diagram of a data compression process utilizing energy compaction in the motion compensation step according to a preferred embodiment of the present invention;
FIG. 9 illustrates a group of video frames processed by the motion compensation step of the data compression process illustrated in FIG. 3;
FIG. 10 illustrates the group of video frames of FIG. 9 after being transformed from the spatial domain to the frequency domain; and
FIG. 11 is a flow diagram of the methodology of the motion compensation and energy compaction steps of a first preferred embodiment of the present invention; and
FIG. 12 is a flow diagram of the methodology of motion compensation and energy compaction steps of a second preferred embodiment of the present invention.





DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
Referring generally to FIGS. 1a-1b, a system in which the preferred embodiment of the present invention is realized is shown generally at 10a and 10b. The data compression method of the present invention maximizes data compression by applying a compaction performance metric, discussed in detail below, to a motion compensation step in the data compression operation to maximize data compression by minimizing rate distortion through the remaining compression steps. As a result, complexity of the data compression operation is greatly reduced and the compression ratio is greatly increased when compared to present MPEG-based methods.
Referring in particular to FIG. 1a, the video signal is generated at a video signal source 12. It is contemplated that this video signal source could be any source of a digital video signal, or an analog video signal that may be later converted to digital through means such as an analog-to-digital video frame grabber 14. Such a video signal source may include, but is not limited to, a video cassette recorder (VCR), a CD ROM, broadcast television signal generator, a high definition television signal generator, or any type of computer network application or other real-time digital video source. The digital video signal generated from the video signal source 12 and/or the video frame grabber 14 is then input into a signal compressor, indicated generally at 16. The signal compressor includes a processor 18, a memory 20, and a power source 22. Also, an external storage medium, such as a central hard drive 24, may be operatively connected to the CPU to add further computation and memory capabilities to the compressor.
Preferably, the above described signal compressor is preferably a Sun Sparc 10 Workstation. However, it should be appreciated that other computers, such as an IBM or IBM compatible personal computer having a Intel Pentium.RTM. processor chip or any other processing unit having equivalent processing capabilities can also be utilized to implement the data compression method of the present invention. The method of the present invention may be programmed into one of the above signal compressors using ANSI C programming language. However, it should also be appreciated that any other computer language capable of implementing the present method may be used.
In addition to the processor 18, the signal compressor also includes an application specific integrated circuit (ASIC) 30 for performing discrete cosine transforms (DCT) or other transform operations in accordance with the preferred method of the present invention as will be described below. The signal compressor 18 may also include a histogram chip 32 which may be utilized for forming histograms for data transmission purposes, such as those described in pending patent application Ser. No. 08/621,855 to Niesen which has been incorporated by reference.
It should be appreciated that the energy compaction method may be realized through the processor 18 via conventional software programming techniques. Alternatively, the energy compaction method of the present invention may be realized through implementation of VLSI hardware, such as the VLSI chip shown at 34. Realization of the present invention through software, hardware, or a combination of both involves tradeoffs between overall system expense and processing speed, as is well known in the art.
Referring to FIG. 1b, a signal receiver is shown generally at 10b. The signal receiver includes a signal decompressor, shown generally at 40. As with the signal compressor, the decompressor may be a Sun Sparc 10 Workstation, an IBM personal computer with an Intel Pentium.RTM. processor, or any other processing unit having similar processing capabilities. It should be appreciated that the compressor and the decompressor may in fact be the same processor, and that the decompressor undoes what the processor does. The decompressor includes a processor 42 for performing inverse transforms to reconstruct the compressed video frames transmitted from the signal compressor 16. Specifically, an ASIC 44 performs these inverse transforms. The processor 42 also includes a memory 46 and a power supply 48, of the type well known in the art. A video monitor 50 is also connected to the processor 42 for displaying the transmitted, reconstructed video frames. The video monitor is of the type well known in the art and may be a television, a computer display, or any other commercially available pixel-based video screen.
Referring to FIG. 2, a diagram illustrating the steps of a conventional MPEG-2 based data compression process is shown at 50. At 52, a sequential group of video frames, generated by the video signal source 12 and shown generally at 54 in FIG. 3, is input into the signal compressor 16. Each of the video frames represents a temporally successive arrangement of pixels, as indicated by the time designation 56 in FIG. 3 and as is well known in the art. As shown in FIG. 2, in accordance with conventional MPEG-2 syntax, after being input into the compressor, the sequence of frames is subjected to a motion compensation analysis at step 58. The general mathematical representation of the motion compensation is as follows. The displacements .DELTA.x.sub.n and .DELTA.y.sub.n for motion compensation of a macroblock frame n (s.sub.n) with respect to a macroblock from frame O (S.sub.o) are: ##EQU1## over all possible .DELTA.x.sub.n, .DELTA.y.sub.n within a window of pixels, which is preferably 16 pixels by 16 pixels, in frame S.sub.o. The initial (I) frame is used as a reference frame S.sub.o, as shown at 60 in FIG. 4. The remaining frames S.sub.n are treated as either predicted (P) frames or bilinear interpolated (B) frames. The window of pixels is moved around each of the frames to generate data for the above equation, and, as shown at 62 in FIG. 5, motion compensation is performed on the group of frames two frames at a time in a predetermined sequence.
Referring again to FIG. 2, at 66, a two-dimensional discrete cosine transform (DCT) is performed on the data contained within the frames to transform the data from the spatial position domain to the spatial frequency domain. The two-dimensional spatial DCT of the plurality of video frames is indexed by the "time tag" index t. ##EQU2## where N is the number of frames in the group of frames, and: ##EQU3##
Also at 66, a one-dimensional discrete cosine transform is performed on the data in the video frames in the temporal dimension to effectively produce a "temporal frequency tag" for each of the spatial transform coefficients. As a result, the group of frames shown in FIG. 3 is transferred from the temporal domain to the frequency domain, as shown at 64 in FIG. 6.
After the transform process, an energy compaction calculation is performed at 67. The compaction process optimizes the correlation between transform coefficients in the higher and lower bands and places most of the coefficient energy in the lower bands for properly motion compensated blocks. The compaction process also involves a double penalty: 1. suboptimality due to the BMA used for motion compensation and 2. the requirement that residual motion error be transformed efficiently. The resulting data for each video frame is quantized and input at 68 into a two-dimensional quantization matrix as shown generally at 70 in FIG. 7. Each of the resulting matrices typically includes twelve coefficients per row, indicated by the arrow U at 72 in FIG. 6, and twelve coefficients per column, indicated by the arrow V at 74, corresponding to the group of twelve video frames.
At 80 in FIG. 2, the processor determines the probability of each coefficient being zero for the particular processed group of frames and arranges the coefficients accordingly in a zig-zag transmission block order utilized in conventional MPEG-2 data compression methods.
Subsequent to the coefficients being arranged in the above-described zig-zag order transmission block in the quantization matrices, optimal distortion and transmission rate calculations are performed on the data to be transmitted to the signal decompressor 42. At 82, the MPEG-2 syntax performs a two dimensional zero run length coding on the zig-zag ordered coefficients to optimize the zero run length for each group of twelve frames. As a result, the transmitted compressed bit stream includes an ordering with the coefficient having the least probability of being zero being first transmitted, and with the coefficients having the least probability of being zero being transmitted in increasing probability order. In the zig-zag order, the first row of the first matrix of quantized coefficients is the column from the processed group of frames with the least number of zeros. The second row of the first matrix of the quantized coefficients is the column from the processed group of frames with the second least number of zeros. The remaining rows of the first matrix are filled out with similar ordering regarding number of zero coefficients in a column. Rows of the second matrix are filled, if there are any remaining columns, with non-zero coefficients. This zero run length coding minimizes the amount of data having no associated energy, and represented by coefficients having a zero value, being transmitted.
The data transmission rate is calculated at 82 using calculations typically based on standard MPEG syntax. However, the transmission rate may be determined through methods such as the three-dimensional transmission rate method described in the pending U.S. patent application Ser. No. 08/621,855 to Niesen, which has been incorporated by reference. The entropy calculations slightly overestimate the transmission rate. The MPEG method typically calculates rate determination using Huffman coding, which is realized through the following equation: ##EQU4## where L.sub.i is the zero run length plus one and p.sub.i is the probability of a sum of .sub.I zeros.
The above-described MPEG-2 syntax utilizes energy compaction in association with the ordering of the quantized coefficients in the quantization matrices, and therefore does not take full advantage of the energy compaction. It is important that the energy compaction step is performed as early on the data compression process as possible, as the optimal performance of the ordering and coding steps is increased if signal coefficients input into the steps exhibit good energy compaction. As will now be described, the data compression method of the present invention optimizes data compression by combining the energy compaction process step with the data compression motion compensation step, thereby fully exploiting the advantages associated with the energy compaction.
Referring to FIG. 8, a preferred embodiment of the methodology used to realize the optimized data compression method of the present invention is shown generally at 90. The methodology of the present invention at 92 receives a set of digitized video frames, shown at 94 in FIG. 9, and performs motion compensation on the frames at 96. The number of frames comprising a processed group preferably is 8, but may range from 2 to N, where N is an even integer depending upon the video processing application and the memory available to support storing N frames. The motion compensation step according to the present invention is based on a first order Markov model for temporal intensity variation and therefore matches frame trends in the group of frames being processed, as opposed to the more narrowly-focused temporal minimization of the differences in two successive frames associated with conventional MPEG-2 syntax. As a result, it more accurately processes frame data typically associated with pixel data input from camera zooming or panning operations. For example, a scene having a progressive enlargement of an object due to focussing will experience a sequential increase/decrease in average luminance intensity. Such a scene change would be best handled by the proposed invention.
Energy compaction is performed on the frame data at 98 on the temporal frequency tag described above in conjunction with the motion compensation step and prior to the transformation at 100 of the data from the temporal to the frequency domain, as shown at 102 in FIG. 10. Integrating the energy compaction process into the motion compensation step prior to the transform process thereby allows energy compaction dependent steps, such as the quantizing, ordering and coding steps similar to those discussed above and shown at 104, 106 and 108, to maximize data compression, because motion compensation effects on energy compaction are optimized before the compressed video data is output at 110 and before step 100.
Realization of the methodology of the present invention will now be discussed. Referring in particular to the transform step 100 shown in FIG. 10, the majority of coefficient energy for the 3-D DCT transform resides in the temporal coefficients S(0,0,w) and in the S(u,v,0) coefficients. The S(0,0,w) coefficients can be expressed as the difference of pixel intensities from blocks of various combinations of two frames within the set of the group of pictures. The cost metric for motion compensation as discussed below is the energy compaction, as measured by an estimate of the basis restriction error (BRE) for S(0,0,w) coefficients. The S(0,0,0) coefficient is not included in the energy compaction optimization since it does not involve frame differences. Since the S(0,0,4) coefficient requires that four frame difference variables be generated just to optimize it, it is also not included in the energy compaction optimization.
The use of the coefficient energy compaction method of the present invention is an extension of the 2-dimension temporal MPEG predictive coding method that conventionally utilizes a BMA method as discussed above. The BMA minimizes the energy in the frame difference of macroblocks. For two frames, the temporal high frequency coefficient for S(0,0,1) is expressed as ##EQU5##
The energy of the S(0,0,1) coefficient is the product of N and the square of the above expression. The cost criteria for the BMA used in MPEG is the energy contribution of the high frequency coefficient, expressed by the s(0,0,1) expression, when x'=x and y'=y. Thus, the BMA acts to minimize a portion of the BRE for two frames. The motion compensation method of the present invention is extended to greater than two frames (N.sub.t >2), using the BRE estimate as the motion compensation performance metric.
The motion compensation method of the present invention is motivated by the concept that the BMA approximately minimizes the energy in the high temporal frequency subband for two frames. The minimization of high temporal frequency energy is extended to greater than two frames by minimizing the BRE in the set of S(0,0,w) coefficients, where S(0,0,w)= ##EQU6## Though the idea can be applied to any even number of frames, it is applied to a set of eight frames in the following analysis. The transform method proposed here could also be applied to a 3-D coefficient vector with u.noteq.0v.noteq.0, such as S(1,1,w). For this case, ##EQU7## The S(1, 1, W) can be segregated into pixel groupings that are differences scaled by the same constant.
The method purposed here applies to Discrete Sire and Discrete Furrier Transforms also.
In the present invention, frame differences are properly weighted and summed to form 3-D DCT coefficients as follows:
S(0,0,1)=0.98.gamma.3+0.83.gamma.4+0.55.gamma.5+0.19.gamma.6
S(0,0,2)=1.84.gamma.1-0.92.gamma.3+0.38.gamma.4-0.38.gamma.5+0.38.gamma.6
S(0,0,3)=0.83.gamma.3+0.19.gamma.4-0.98.gamma.5-0.55.gamma.6
S(0,0,5)=0.55.gamma.3-0.98.gamma.4+0.19.gamma.5+0.83.gamma.6
S(0,0,6)=0.54.gamma.1-0.92.gamma.2-0.92.gamma.3+0.54.gamma.4-0.54.gamma.5-0.92.gamma.6
S(0,0,7)=0.19.gamma.3-0.54.gamma.4+0.83.gamma.5-0.98.gamma.6
The random variables .gamma.1, .gamma.2, .gamma.3, .gamma.4, .gamma.5, and .gamma.6 are the difference of N-by-N blocks of pixels, and are defined as: ##EQU8## where x'=x+.delta.x, y'=y+.delta.y, and .delta.x and .delta.y are allowed displacements within the reference frame.
In the .gamma..sub.1 expression, s(x,y,3) is a search frame while s(x',y',0) is a reference frame in the expansion for .gamma..sub.6. Assuming that the frame difference .gamma..sub.1 is small, s(x',y',0) can be used to approximate s(x',y',3). In the expression for .gamma..sub.4 and .gamma..sub.6, the reference frames s(x',y',2) and s(x',y',3) can be approximated by s(x',y',1) and s(x',y',0), respectively. Using the suggested substitution for reference frames in .gamma..sub.4 and .gamma..sub.6, six displacement vectors are generated. Reference frames 1 and are used in their original state in subsequent video compression process steps. Using the aforementioned approach, the frames 0 and 1 can be used as reference frames for 3-D motion compensation of a group of eight frames (See FIG. 13).
The objective of the motion compensation/energy compaction method of the present invention is to position the blocks from the group of eight frames within the search range of the reference frame to minimize the BRE of the coefficient vector s.sub.00.
Thus, the basis restriction error (BRE), which essentially defines energy compaction capability, is expressed as: ##EQU9## as the mean squared error between the original block of pixels and the reconstructed block of pixels. The reconstructed block of pixels is formed after truncating a number of the transform coefficients of higher frequency. As a result, the basis restriction error is expressed as: ##EQU10## for a unitary transform as the sum of squares of the truncated coefficients, as the n highest spatial frequency coefficients and the m highest temporal frequency coefficients are truncated.
The coefficient vector is expressed as s.sub.00 =[S(0,0,1), S(0,0,2), S(0,0,3), S(0,0,5), S(0,0,6), S(0,0,7)].sup.T.
The coefficient vector o is expressed in terms of the set of frame differences in the .gamma. vector, where .gamma.=[(.gamma..sub.1, .gamma..sub.2, .gamma..sub.3, .gamma..sub.4, .gamma..sub.6 ].sup.T. ##EQU11## Then, set s.sub.00 =K.gamma.. As a result, s.sup.T.sub.00 s.sub.00 =.gamma..sup.T K.gamma.=.gamma..sup.T L .gamma., where L=K.sup.T K.
In an alternative embodiment of the present invention, a different metric for motion compensation is to use the magnitude, or inner product, of the coefficient vector s.sub.00, instead of the BRE. In this approach, the magnitude s.sub.00, s.sub.00.sup.T s.sub.00, would be used to determine the best set of displacements that minimize the inner product. In this alternative approach, the Rayleigh quotient is used to assist in minimization of the inner product of the temporal coefficient vector s.sub.00. The inner product of the vector s.sub.00 can be described in terms of the frame difference vector .gamma. and a matrix L where,
s.sub.00.sup.T s.sub.00 =.gamma..sup.T L.gamma..
The coefficient matrix can be decomposed as
L=Q.sup.T .LAMBDA.Q
where Q=[e.sub.1, e.sub.2, e.sub.3, e.sub.4, e.sub.5, e.sub.6 ].sup.T and ##EQU12##
The eigenvectors e.sub.1, and e.sub.6 correspond to the smallest eigenvalue .lambda..sub.1 and the largest eigenvalue .lambda..sub.6, respectively.
The Rayleigh quotient ##EQU13## is minimized to the value .lambda..sub.1 when x=e.sub.1, and maximized to .lambda..sub.6 when x=e.sub.6.
The Rayleigh quotient is minimized when the frame difference vector
.gamma.=.alpha.e.sub.1,
where .alpha. is a scaling constant with .alpha..noteq.0. Substituting .alpha.e.sub.1 for x and s.sub.00.sup.T s.sub.00 for x.sup.T Lx, the Rayleigh quotient becomes ##EQU14##
The coefficient vector inner product, s.sub.00.sup.T s.sub.00, is minimized to value .alpha..sub.2 .lambda..sub.1, when .gamma.=.alpha.e.sub.1 is satisfied for the minimum value of .vertline..alpha..vertline., ##EQU15## The minimization of the above is the motive for the following. Values for .alpha. and .gamma. are next determined that satisfy the above minimization equation. Where .gamma..noteq.0 and when each component of .gamma. is minimized, the magnitude from .gamma.=.alpha.e.sub.1, is taken: ##EQU16##
The following procedure is performed twice, once for .alpha.>0 and once for .alpha.<0. The value of .vertline..alpha..vertline..sub.max is used to set a maximum value for .alpha.>0 and a minimum value for .alpha.<0. If .gamma..sub.1+ is the frame difference that produces the smallest 20 value greater than .alpha.e.sub.1 (i), .gamma..sub.1 is the frame difference that produces the smallest value less than .alpha.e.sub.1 (i) and .gamma..sub.10 is a frame difference such that .gamma..sub.10 =.alpha.e.sub.1 (i), all possible combinations of .gamma..sub.1+, .gamma..sub.1-, and .gamma..sub.10 are formed into trial vectors .alpha.e.sub.1, where each trial vector is identified by its j index, to evaluate the set of frame differences that best minimize the Rayleigh quotient. Each trial vector is evaluated in the product of the Rayleigh quotient and the frame difference. The trial vector that best approximates .alpha.e.sub.1 is that .alpha.e.sub.1j which minimizes
.alpha..sup.2 e.sub.1j e.sup.T.sub.1J R(.alpha.e.sub.1j).A-inverted.e.sub.1j
The effectiveness of this alternative 3-D motion compensation energy compaction method of the present invention is determined by the minimum value achieved by the above trial vector. The energy in the residual motion error for the 3-D motion compensation method provides a standard method for comparison to other motion compensation methods.
The alternative method of the present invention acts to minimize the sum of the squares of the spatio-temporal frequency components of s.sub.00, as the BRE metric for motion compensation acts to maximize the energy compaction of s.sub.00. Such a method is synergistic with the subsequent transform process at 100 in FIG. 8, which also acts to maximize the energy compaction of the transform coefficients. The energy compaction method of the present invention reduces residual motion error in the overall data compression process, as the vector coefficient values generated through data compression method utilizing the energy compaction of the present invention are lower in energy than the corresponding vector coefficient values generated by the block matching algorithm (BMA) of conventional MPEG-2 based methods. The energy compaction method of the present invention has a performance metric that is more robust than that generated in conventional MPEG-2 schemes, and therefore has a stronger tie-in with a minimization of energy in the residual motion error. The flow diagram shown at 120 in FIG. 11 summarizes the methodology associated with motion compensation step 96 and the energy compaction step 98 of FIG. 8 in more detail. In step 122, video data is input into the data compressor. Next, the motion compensation is performed on the video data, and coefficient vectors are generated at 124, while displacement vectors are generated at 126 as discussed above in conjunction with FIG. 8. If the BRE-based energy compaction methodology of the present invention is utilized, the BRE is next generated at step 128. At step 130, the methodology determines if the BRE has been minimized. If not, the methodology returns to step 124 and repeats steps 124, 126 and 128 to minimize the BRE. If, at step 130, the methodology determines that the BRE has been minimized, the corresponding displacement vectors generated at step 124 are stored in memory at step 132.
The methodology of the alternative embodiment of the present invention is shown in the flow diagram at 130 in FIG. 12. Steps 132, 134 and 136 correspond generally to steps 122, 124 and 126 of the flow diagram in FIG. 11. However, at step 138, the methodology generates trial vectors as described above. At step 140, coefficient and trial vectors are compared. At step 142, the methodology determines whether the difference between the coefficient and trial vectors is small. If the difference is not small, the methodology returns to step 134. Steps 134, 136, 138 and 140 are repeated until the difference between the generated trial and coefficient vectors is small. Subsequently, at step 144, the displacement vectors are stored for use in the transform process at 100 in FIG. 8.
Because the motion compensation process in MPEG provides motion vectors alone as its output, the more accurate motion vectors from the present invention could be substituted in for the MPEG motion vectors to provide a lower bit rate and lower distortion modified MPEG video compression process.
From the foregoing explanation, it should be appreciated that the video signal compression methods of the present invention provide methods for compressing digital video signals in a manner that achieves low bit rate transmission with minimal detected distortion. The method of the present invention maximizes use of energy compaction by performing an energy compaction process early on in the data compression process, thereby maximizing data compression in subsequent energy compaction dependent steps. The method of the present invention utilizes a more robust performance metric that generates coefficient values having lower energy levels than the coefficient values generated by the BMA of conventional intake two methods and thereby reduces residual motion error in data compression. The data compression method of the present invention also exhibits increased compression ratios when compared to present MPEG-1 and MPEG-2 based data compression methods while at the same time minimizes data compression complexity.
Various other advantages of the present invention will become apparent to those skilled in the art after having the benefit of studying the foregoing text and drawings, taken in conjunction with the followings claims.
Claims
  • 1. A method of optimizing data compression in a data compression process, comprising the steps of:
  • providing a signal compressor for performing video signal data compression operations;
  • receiving a video signal at said signal compressor, said video signal including spatial and temporal data from a plurality of video signal frames;
  • compensating for motion in said plurality of video frames by simultaneously registering said plurality of video signal frames to produce a plurality of motion vectors;
  • generating a performance metric including energy compaction parameters; and
  • applying said performance metric to said step of compensating for motion to optimize rate distortion performance therein and in subsequent data compression process steps.
  • 2. The method of claim 1, further comprising the steps of:
  • transforming said conditioned spatial and temporal data from a location domain to a frequency domain after said step of applying said performance metric;
  • ordering said transformed spatial and temporal data coefficients into a coefficient matrix for data transmission purposes; and
  • coding said conditioned coefficient matrix for data transmission purposes;
  • said step of applying said performance metric to said step of compensating for motion thereby optimizing rate distortion performance in said transforming, ordering and coding steps.
  • 3. The method of claim 2, wherein said step of transforming comprises performing a two-dimensional discrete cosine transform on said spatial and temporal data.
  • 4. The method of claim 1, wherein said step of generating a performance metric comprises the step of minimizing temporal coefficient vector components where the temporal coefficient vector components are for higher temporal frequencies.
  • 5. The method of claim 4, wherein said step of minimizing temporal coefficient vector components is realized through application of the Rayleigh quotient.
  • 6. The method of claim 1, wherein said step of applying said performance metric comprises applying a basis restriction error to a coefficient vector generated in said step of compensating for motion to minimize high temporal frequency energy for more than two of said plurality of video signal frames.
  • 7. The method of claim 1, wherein said step of applying said performance metric comprises minimizing a coefficient vector inner product to optimize rate distortion performance.
  • 8. A method of compressing video signal data, comprising the steps of:
  • providing a signal compressor for performing video compression operations;
  • inputting video signal data including a plurality of video frames into said signal compressor;
  • compensating for motion in said plurality of video frames;
  • compacting the energy level of said motion compensated plurality of video frames during said step of compensating for motion in said plurality of video frames;
  • performing a two-dimensional transformation of said video signal data from spatial and temporal position domains to a frequency domain to obtain two-dimension spatial frequency coefficients;
  • creating a plurality of two-dimension quantization matrices each containing two quantization coefficients corresponding to said two dimension temporal and spatial frequency coefficients from one of said video frames in said video signal;
  • creating a two-dimensional ordering of said transmission block from said data in said quantization matrices with an optimized distortion rate resulting from said step of compacting the temporal frequency vector energy level; and
  • coding said data in said quantization matrices for data transmission purposes with an optimized distortion rate resulting from said step of compacting the temporal frequency vector energy level.
  • 9. A method of transmitting a digital video signal, comprising the steps of:
  • compensating for motion in digital video signal data;
  • simultaneously compacting energy in said video signal data;
  • spatially and temporally transforming said digital video signal data;
  • quantizing said spatially and temporally transformed digital video signal data to calculate optimal transmission and signal distortion rates;
  • coding said spatially transformed, quantized digital video signal data;
  • transmitting said coded, quantized spatially transformed digital video signal information with an optimal signal distortion rate resulting from said step of compacting energy.
  • 10. The method of claim 9, wherein said steps of transforming, compacting, quantizing, and coding are combined into a single data compression process step.
  • 11. The method of claim 9, wherein said step of simultaneously compacting energy comprises generating a performance metric to define energy compaction in said step of compensating for motion.
  • 12. The method of claim 11, wherein said step of generating a performance metric comprises the steps of:
  • generating a coefficient vector representing a plurality of frame differences from an input set of video frames;
  • generating a Basis Restriction Error for determining a mean squared error between an original block of pixels and a reconstructed block of pixels; and
  • minimizing the Basis Restriction Error to maximize energy compaction.
  • 13. The method of claim 11, wherein said step of generating a performance metric comprises the steps of:
  • generating a coefficient vector representing a plurality of frame differences of an input set of video frames;
  • calculating the magnitude of the coefficient vector; and
  • calculating a set of displacement vectors that minimizes an inner product of the coefficient vector through use of the magnitude of the coefficient vector to thereby maximize energy compaction.
  • 14. A data compression system, comprising:
  • a signal input for receiving a video signal containing a plurality of frames of video data;
  • a video signal compressor, comprising:
  • a processor that performs motion compensation on a group of video frames, that generates an energy compaction performance metric to identify temporal trends in said group of video frames, and that applies said performance metric to said group of video frames as said processor performs said motion compensation to generate a plurality of motion vectors;
  • a memory operatively associated with said processor that stores said group of video frames and said plurality of motion vectors during processing thereof;
  • said processor also being operative to transform said group of video signals into a group of frequency domain coefficients, order said transform coefficients, and code said ordered transform coefficients in an optimized manner subsequent to application of said performance metric to said group of video frames to generate compressed video signal data; and
  • a signal output that outputs said compressed video signal data for transmission thereof.
  • 15. An energy compaction method for an MPEG-based data compression process, comprising the steps of:
  • inputting video data representative of a group of video frames;
  • generating a plurality of displacement vectors from frame differences in said group of video frames;
  • generating a coefficient vector expressed in terms of said frame differences;
  • generating a basis restriction error estimate based on said steps of generating displacement vectors and generating a coefficient vector;
  • minimizing said basis restriction error to determine an optimal group of said displacement vectors; and
  • storing said optimal group of displacement vectors for use in subsequent data compression steps.
  • 16. A method of compacting energy in a data compression process, comprising the steps of:
  • inputting video data representative of a group of video frames;
  • generating a plurality of displacement vectors representing frame differences in said group of video frames;
  • generating a coefficient vector in response to said step of generating a plurality of displacement vectors;
  • forming a plurality of trial vectors from combinations of said frame differences in said group of video frames;
  • evaluating said coefficient vector through said trial vectors to determine a set of frame differences that best minimizes an associated Rayleigh quotient; and
  • storing a set of optimal displacement vectors in response to said step of evaluating said coefficient vector for subsequent data compression.
US Referenced Citations (12)
Number Name Date Kind
5121191 Cassereau Jun 1992
5223926 Stone et al. Jun 1993
5227878 Puri et al. Jul 1993
5293434 Feig et al. Mar 1994
5361105 Iu Nov 1994
5367629 Chu et al. Nov 1994
5424779 Odaka Jun 1995
5428693 Murakami et al. Jun 1995
5452104 Lee Sep 1995
5530661 Garbe et al. Jun 1996
5933193 Niesen Aug 1999
5936669 Niesen Aug 1999
Non-Patent Literature Citations (7)
Entry
The MPEG standard (www.crs4.it/.about.luigi/mpeg/mpeg2.html).
MPEG-2 FAQ version 3.7 (May 11, 1995); by Chad Fogg (efogg@chromatic.com).
Pennebaker & Mitchell JPEG: Still Image Data Compression Standard, Chapter 4--The Discrete Cosine Transform (DCT)--(1993).
Visual Psychophysics, Digital Pictures: Representation and Compression Chapters 3 and 4--Netravali & Haskell (1988).
IEEE Transactions on Electromagnetic Compatibility, vol. EMC-18, No. 1, Feb. 1976.
IEEE Transactions on Image Processing, vol. 3, No. 5, Sep. 1994 Entitled Three-Dimensional Subband Coding with Motion Compensation by Jens Rainer Ohm, Member, IEEE.
IEEE Transactions on Image Processing, vol. 3, No. 5, Sep. 1994 Entitled Multirate 3-D Subband Coding of Video by David Taubman and Avideh Zakhor.