a-d illustrate first and second embodiments of a one-dimensional image encoding process that provide for converting from a first relatively lower-resolution display format to a second relatively-higher-resolution display format, operating on an image of relatively lower detail;
a-d illustrates first and second embodiments of a one-dimensional image encoding process that provide for converting from a first relatively lower-resolution display format to a second relatively-higher-resolution display format, operating on an image of relatively higher detail;
a illustrates a first embodiment of a first group of pixels of a relatively-lower-resolution digitized image of a third embodiment of an encoded image in accordance with the third aspect of the image encoding process;
b illustrates a first embodiment of extended data of the third embodiment of an encoded image that can be combined with the relatively-lower-resolution digitized image illustrated in
a illustrates a first embodiment of a one-dimensional encoding process in accordance with the third aspect of the image encoding process;
b illustrates a second aspect of the first embodiment of the one-dimensional encoding process in accordance with the third aspect of the image encoding process;
a illustrates a first embodiment of a two-dimensional encoding process in accordance with the third aspect of the image encoding process
b illustrates a set of interpolation equations associated with the first embodiment of the two-dimensional encoding process illustrated in
a illustrates a first embodiment of a one-dimensional decoding process in accordance with the third aspect of the image encoding process, for decoding an image encoded in accordance with the first embodiment of the one-dimensional encoding process illustrated in
b illustrates a second embodiment of a one-dimensional decoding process in accordance with the third aspect of the image encoding process, for decoding an image encoded in accordance with the second embodiment of the one-dimensional encoding process illustrated in
c illustrates a set of equations associated with the first and second embodiments of the one-dimensional decoding processes illustrated in
a illustrates a first embodiment of a two-dimensional decoding process in accordance with the third aspect of the image encoding process, for decoding an image encoded in accordance with the first embodiment of the one-dimensional encoding process illustrated in
b illustrates a set of equations associated with the first embodiment of the two-dimensional decoding processes illustrated in
a illustrates a first portion of a second embodiment of extended data of a fifth embodiment of an encoded image that can be combined with the relatively-lower-resolution digitized image illustrated in
b illustrates a second embodiment of a first group of pixels of a relatively-lower-resolution digitized image of the fifth embodiment of an encoded image in accordance with the third aspect of the image encoding process;
c illustrates a second portion of the second embodiment of extended data of the fifth embodiment of an encoded image that can be combined with the relatively-lower-resolution digitized image illustrated in
a illustrates a third embodiment of a one-dimensional encoding process in accordance with the third aspect of the image encoding process;
b illustrates a second aspect of the third embodiment of the one-dimensional encoding process in accordance with the third aspect of the image encoding process;
a illustrates a second embodiment of a two-dimensional encoding process in accordance with the third aspect of the image encoding process;
b illustrates details of a second embodiment of a two-dimensional encoding process in accordance with the third aspect of the image encoding process illustrated in
c illustrates a set of interpolation equations associated with the second embodiment of the two-dimensional encoding process illustrated in
a illustrates a third embodiment of a one-dimensional decoding process in accordance with the third aspect of the image encoding process, for decoding an image encoded in accordance with the third embodiment of the one-dimensional encoding process illustrated in
b illustrates a fourth embodiment of a one-dimensional decoding process in accordance with the third aspect of the image encoding process, for decoding an image encoded in accordance with the fourth embodiment of the one-dimensional encoding process illustrated in
c illustrates equations associated with the third and fourth embodiments of the one-dimensional decoding processes illustrated in
a illustrates a third embodiment of a first group of pixels of a relatively-lower-resolution digitized image of a seventh embodiment of an encoded image in accordance with the third aspect of the image encoding process;
b illustrates a third embodiment of extended data that can be combined with the relatively-lower-resolution digitized image of the seventh embodiment of an encoded image illustrated in
a-d illustrate first through fourth portions of a first intermediate process for determining the set of interpolation equations associated with the third embodiment of the two-dimensional encoding process illustrated in
a illustrates a fifth embodiment of a one-dimensional encoding process in accordance with the third aspect of the image encoding process;
b illustrates a sixth embodiment of a one-dimensional encoding process in accordance with the third aspect of the image encoding process;
c illustrates a set of equations associated with the fifth and sixth embodiments of the one-dimensional encoding process illustrated in
a illustrates a fifth embodiment of a one-dimensional decoding process in accordance with the third aspect of the image encoding process, for decoding an image encoded in accordance with the fifth embodiment of the one-dimensional encoding process illustrated in
b illustrates a sixth embodiment of a one-dimensional decoding process in accordance with the third aspect of the image encoding process, for decoding an image encoded in accordance with the sixth embodiment of the one-dimensional encoding process illustrated in
a illustrates a fifth aspect of a one-dimensional decoding process as applied along a row dimension of the associated relatively-low-resolution and relatively-high-resolution images; and
b illustrates the fifth aspect of a one-dimensional decoding process as applied along a column dimension of the associated relatively-low-resolution and relatively-high-resolution images.
Referring to
For example, in accordance with a first aspect of a system for generating, transmitting and displaying encoded images 20.1, the relatively-higher-resolution digitized image 12 is initially captured by a camera 22 that incorporates an associated camera lens 24 that provides for focusing a raw image 26—of an associated scene 28 being imaged—onto either an imaging medium 30—for example, film, for example, 35 millimeter film 30′ as used for recording cinematic movies—or an imaging sensor 32—for example, and array of photodetectors, for example, a CCD imaging sensor.
The aspect ratio of an image is given by the ratio of the width to the height of the image. A wide range of aspect ratios have been used in systems for generating and displaying images, and different aspect ratios can be used for different elements of such systems. For example, the aspect ratio of 35 millimeter film 30′ is typically 1.33:1, or a ratio of 4:3, based upon a standard originally established by Thomas Edison. This same 1.33:1 aspect ratio has been used in the National Television System Committee (NTSC) standard for analog televisions. However, a wider aspect ratio is generally considered to be more aesthetically pleasing. For example, many major motion pictures are created in an image aspect ratio of approximately 2.37:1. When recording an image having an aspect ratio different from that of the imaging medium 30 or imaging sensor 32, an anamorphic lens 34 can be used in cooperation with the camera lens 24 so as to shrink or expand the raw image 26 in one direction or the other so as to use substantially all of the active portion of the imaging medium 30 or imaging sensor 32, so as to provide for better image resolution and a better associated overall signal-to-noise ratio. The original aspect ratio of a raw image 26 that was recorded anamorphically can then be recovered by using a complementary anamorphic magnification upon reconstruction thereof. A raw image 26 recorded on an imaging medium 30 can be digitized with a digitizer 36 to produce the relatively-higher-resolution digitized image 12 than would otherwise be produced directly by a corresponding imaging sensor 32.
The displays 16, 16.1 upon which the encoded image 14 is ultimately displayed exhibit a variety of formats that are generally characterized by both the associated aspect ratio and the associated resolution, the latter of which is generally expressed as the pixel dimensions of the associated two-dimensional display area, for example, W×H, where W is the number of pixels 37 in the width of the display 16, 16.1, and H is the number of pixels in the height of the display 16, 16.1. For example, one set of displays 16, 16.1 uses an aspect ratio of approximately 2.37:1—also generally referred to approximately as 21:9—with associated display resolutions, in order of increasing total number of pixels 37, of 1920×810 and 2560×1080. As another example, another set of displays 16, 16.1 uses an aspect ratio of approximately 1.78:1—also generally referred to approximately as 16:9—with associated display resolutions, in order of increasing total number of pixels 37, of 1920×1080, 2560×1440 and 3840×2160. For example, BLU-RAY DISC™ Video supports 16:9 aspect ratios with resolutions of 1920×1080, anamorphic 1920×1080, 1280×720, anamorphic 720×480, and anamorphic 720×576.
In order to display an image having a relatively higher aspect ratio without distortion on a display 16, 16.1 having a relatively lower aspect ratio, black horizontal bars are generally added at the top and bottom of the display 16, 16.1 where there is no image. For example, when a major motion picture having an aspect ratio of 2.37:1 is transferred to a high definition disc (such as a BLU-RAY DISC™ Video), black horizontal bars are added to the top and bottom of the video image so that the total area is in the format of 16:9 (or 1.78:1). This format is ideal for displaying the image in “letterbox” format on displays 16, 16.1 also having the 16:9 aspect ratio. Alternatively, when such movies are shown to the full width of a display having a 2.37:1 aspect ratio, the video image can be electronically stretched to fit the full vertical extent of the display 16, 16.1, thereby removing the black bars to show the movie in its proper format. As a second alternative, some users find the black letterbox bars so objectionable that even on a 16:9 aspect ratio display 16, 16.1 they will electronically stretch the image in both directions to overcome the black bars without otherwise distorting the aspect ratio of the image, even though the left and right sides of the image are now cut off by the width of the display 16, 16.1. However, in both cases an electronic vertical stretch of the image is simply some form of mathematical up-scaling or interpolation of each column of original pixel values into a longer (taller) column of pixel values, which otherwise does not add detail to the displayed image. However, if there is sufficient detail in the original relatively-higher-resolution digitized image 12, and if the associated display 16 is capable of displaying the additional pixels 37 associated therewith, then as described hereinbelow, the encoded image 14 provides for reconstructing the relatively-higher-resolution digitized image 12 in a form that can be rendered on that associated display 16, for example, so as to provide for vertically expanding the encoded image 14 so as to make use of the full resolution of the display 16.
Alternatively, the horizontal and vertical pixel resolution of a legacy display 16.1 may be lower than that of the relatively-higher-resolution digitized image 12, in which case, the encoded image 14 is formatted for input to the legacy display 16.1 without requiring further image processing. For example, an original image resolution of 3840×2160 pixels may be encoded into the encoded image 14, wherein the central 1920×1080 pixels—constituting an associated relatively-lower-resolution image 38—represent a lower resolution representation of the full 3840×2160 relatively-higher-resolution digitized image 12 suitable for immediate input to a legacy display 16.1 without requiring any decoding, whereas the remaining information is retained in the encoded image 14 so that the full 3840×2160 relatively-higher-resolution digitized image 12 can be reconstructed for input to a 3840×2160 display by appropriately decoding the encoded image 14. In this case, a 1920×1080 display 16, 16.1 need only directly display this relatively-lower-resolution image 38 without decoding, whereas a 3840×2160 display 16 can employ the associated decoding process to display the full resolution 3840×2160 relatively-higher-resolution digitized image 12.
The encoded image 14 is transmitted to the associated display(s) 16, 16.1 by conventional means. For example, the encoded image 14 undergoes a standard image formatting 40, for example, compression and encoding, as is conventionally used to format conventional images for transmission to associated display systems. For example, the encoded image 14 may be converted to any of a variety of formats, including but not limited to, either the MPEG-1, MPEG-2, MPEG-3 or MPEG-4 standard of the Moving Pictures Experts Group (MPEG) or the various standards of the Society of Motion Picture and Television Engineers (SMPTE). The formatted encoded image 14′ is then transmitted as an image signal 41, also more generally referred to as an image representation 41, over an associated image transmission medium 42—for example, either wirelessly, by a conductive transmission line, for example, cable or DSL, by DVD or BLU-RAY DISC™, or streamed over the internet—to an associated receiver or player 44 that extracts the encoded image 14 from the formatted encoded image 14′ in accordance with the associated compression and encoding standard, and then either inputs the encoded image 14 directly to a legacy display 16.1, or decodes the encoded image 14 with an image decoder 46—using an associated image decoding process 46′—for example, that could be embedded in either the receiver or player 44 or in the associated displays 16.
Referring to
Referring to
Because this example involves two different images that differ only in the vertical direction, it is sufficient to illustrate the principles of the one-dimensional encoding process 300 in this case by examining an arbitrary column of pixel values since the same process could be applied independently to any other column. Referring now to
In accordance with the first embodiment of the one-dimensional encoding process 300.1, the encoded image 14, 14.1 of
Referring to
The one-dimensional encoding process 300 may be performed independently along different directions—height and width—of the image, and may be performed for a plurality of times in either direction, either consecutively or interleaved with respect to the direction of encoding (i.e. H, V, V or V, H, V, etc.).
Note that the encoding process may be performed in a number of sequential stages with varying algorithms as desired, each producing a smaller, lower resolution approximation of the initial image of the previous stage while at the same time producing an increased number of pixel values surrounding the increasingly lower resolution image region. The initial image would then be reconstructed by applying the appropriate inverse operation of each stage in the reverse sequence.
The one-dimensional encoding process 300—performed in one or both directions—provides a relatively lower resolution relatively-lower-resolution image 38 within the encoded image 14 that may be displayed without additional processing while ideally masking off or otherwise not displaying the peripheral pixel values for maximum aesthetic value. The resultant relatively lower resolution relatively-lower-resolution image 38 is made as representative of the higher resolution image as possible, in accordance with known algorithms in the art of image sampling. However, more complex algorithms may exhibit diminishing improvements relative to the increase in electronic processing involved to reconstruct the image.
Consider the comparison in image fidelity shown in
Referring again to
The resulting 1920×1080 encoded image can be played on standard 1920×1080 display devices. Since the source content is first formed vertically stretched by 33% before creating the initial image and since the lower resolution region of the encoded image is vertically scaled by the inverse amount, the visual result of the lower resolution image has a final aspect ratio of 2.37:1. The simple instruction to the display 16.1 of blanking or turning off the 135 rows of pixels 37 above and below the relatively-lower-resolution image 38 effectively provides an image of 1920×810 pixels 37 of, in one set of embodiments, substantially similar quality to the 1920×810 image of 2.37:1 movies provided by conventional transmission and storage media, including BLU-RAY DISCs™. However, in this case encoded values within the 1920×810 image combined with the additional values within the remaining 270 rows can be employed using the inverse process of the encoding algorithm to restore the full 1920×1080 anamorphic image with a significant increase in vertical resolution (33.33%) for those display systems which can take advantage of this improvement in image quality. In particular, such display systems may include a 1920×1080 (1.78:1) projector fitted with an anamorphic lens to optically reformat the 1920×1080 (1.78:1) decoded image into the 2.37:1 aspect ratio. Such display systems may further include those with 2560 horizontal×1080 (2.37:1) vertical pixels 37, wherein the 1920 columns of pixels 37 of the initial image are electronically stretched to fill the 2560 horizontal columns of pixels 37 of the display, thereby horizontally stretching the image by 133.33% to render a final image again having the 2.37:1 visual aspect ratio of the original content.
Referring to
The decoding process 900 may be readily incorporated into the image processing hardware components and/or software of a variety of devices including but not limited to displays, set top content boxes, BLU-RAY DISC™ and similar media players, computer systems, gaming consoles, dedicated image processors, video distribution and storage systems and even handheld devices. Furthermore, such devices may further include a capability of receiving and/or storing one or more separate segments of an initial image, one of which includes the relatively-lower-resolution image 38, and where such segments may potentially arrive at different times and/or through different communication methods, and then recombining these segments in synchronization and with sufficient data to apply the appropriate decoding algorithm and to therefore recreate the initial high resolution image. Such segments may each arrive, for example, through individual optical discs (such as a BLU-RAY DISC™) or streaming content. In the case of the relatively-lower-resolution image 38 residing on one optical disc and the remaining encoded values of the encoded image arriving on a second optical disc, a synchronized dual optical disc player provides for reading both discs in synchronization to decode the respective encoded values and therefore to reconstruct the relatively-higher-resolution initial image in real time without the need for storage of any segment.
Referring to
Accordingly, in accordance with one example of this second aspect of a system for generating, transmitting and displaying encoded images 20.2, a relatively-higher-resolution digitized image 12 with a native aspect ratio of 16:9 is transferred to an initial image of 2560 columns×1440 rows of pixels 37 (1.78:1). The encoding process is then applied to both dimensions of the image to create a centralized, lower resolution image of 1920 columns×1080 rows of pixels 37 (1.78:1). Because this example applies the same 75% scaling in both directions, the lower resolution image retains the identical aspect ratio of the initial image. The relatively-lower-resolution image 38 (i.e. the encoded HD (EHD) content) may further be communicated and displayed without the additional encoded values (i.e. the extended data (ED) content) through conventional means of HDTV, such as using a BLU-RAY DISC™ for storage and a 1920×1080 (1.78:1) TV for display. The additional encoded values (i.e. the extended data (ED) content) that, in this case, surround the relatively-lower-resolution image 38 in the encoded higher resolution image, may be communicated by a separate means. In this case there are fewer pixel values representing the additional encoded image (i.e. the extended data (ED) content) than in the relatively-lower-resolution image 38 (i.e. the encoded HD (EHD) content). Therefore, the entire additional information can also be communicated through conventional means of HDTV, such as using a second BLU-RAY DISC™ or using streaming technology. Accordingly, a display capable of presenting a higher resolution such as 2560×1440 (1.78:1) pixels 37 can be accompanied by an electronic processing capability to receive both sets of communications to rebuild the full encoded image 14 and subsequently decode this image to create the 2560×1440 (1.78:1) initial relatively-high-resolution image.
The one-dimensional encoding process 300 provides for independently encoding and/or decoding in multiple dimensions. For example, upon receiving the relatively-lower resolution 1920×1080 (1.78:1) image 38 plus sufficient remaining encoded values (i.e. the extended data (ED) content), an alternative display solution may be to decode only in the horizontal direction, so as to produce a relatively-higher-resolution image of 2560×1080 (2.37:1) that can be shown on a display 16 of the same resolution. In this case of course, the content will be horizontally stretched. However, for some users this wider aspect ratio has value that makes up for the fact that the image is distorted.
In accordance with another aspect, the relatively-lower-resolution image 38 of the encoded image 14 is sent first and immediately followed by the remaining pixel values (i.e. the extended data (ED) content), of the encoded image. However, in this embodiment, the receiving display may either allow this process to continue, decoding the two packets of information as they come in, or the display 16 (or accompanying device) may at any time direct the sending device to exclude the additional information packet due to bandwidth limitations or due to the user directing the display to discontinue processing the information to reconstruct the initial image. In many cases, especially relying on wireless communications, the bandwidth of those communications may vary. Accordingly, this provides for at least the relatively-lower-resolution image 38 to be communicated and displayed during low bandwidth availability but then switched to the higher resolution initial image when bandwidth availability is higher and/or when the user desires a higher resolution image.
The image encoder 10 provides an encoded image which can be shown either as a low resolution representation of the initial image or, with appropriate decoding, as the higher resolution initial image itself. If raw pixel values of the extended data (ED) content outside the relatively-lower-resolution image region are displayed without decoding, those displayed values will typically appear as visual noise. Accordingly, these pixel values should be suppressed by the display 16, 16.1 or by an intermediate device. A simple method of performing such suppression is to allow the user the ability to effectively adjust the extent of black image bars on the display 16, 16.1 until only the relatively-lower-resolution image 38 is shown. With typical content such as video, the size of the relatively-lower-resolution image 38 is a constant for the entire content, so the extent of such black bars need only be adjusted at the beginning of the viewing experience. However, the size and aspect ratio of the relatively-lower-resolution image 38 may be encoded within an information frame typically placed at the beginning of the content. The display 16, 16.1 or intermediate device may therefore interpret this basic information and automatically turn off the pixels 37 outside the relatively-lower-resolution image region.
For example, an initial image of 1920 columns×1080 rows of pixels 37 (1.78:1) formed by transferring a movie with a native aspect ratio of 2.37:1 but produced with a 33.33% vertical (anamorphic) stretch, employs the full resolution of 1920×1080 (1.78:1). An encoded image 14 can now be produced by applying the encoding algorithm only in the vertical direction to produce a lower resolution image of 1920 columns×810 rows of pixels 37 (2.38:1), leaving the 135 rows of pixels 37 above and below the lower resolution image for redistributed pixel values to support later reconstruction of the higher resolution initial image. Since the entire encoded image is in the standard resolution of HDTV, it could simply be shown on a standard 1920×1080 (1.78:1) HDTV display 16, 16.1. However, if all the pixel values of the encoded image are shown then there will be 135 rows above and below the lower resolution image which will appear as noise. Therefore, the HDTV display 16, 16.1 or some intermediate device would allow the 135 rows above and below the lower resolution image to be turned off as a setting either selected by the user or automatically selected by the display 16, 16.1 or the intermediate device based on initial information provided by the content.
Note again that since the above example considers 2.37:1 content anamorphically stretched before encoding, the relatively-lower-resolution image 38 of the encoded image 14 (i.e. within the 810 central rows of 1080 total rows) will be of the proper 2.37:1 aspect ratio when displayed on the 1920×1080 (1.78:1) display 16, 16.1. The blanking process applied to the 135 rows above and below the relatively-lower-resolution image 38 effectively just creates the standard “letterboxed” appearance of a conventional HDTV showing native 2.37:1 content.
The above-described encoding process 200 and associated decoding process 900 can help establish a market for displays 16 having higher resolution than the HDTV resolution of 1920×1080 pixels 37, by enabling a single content source to support both 1920×1080 pixel imaging without encoding and also relatively-higher-resolution imaging when used in cooperation with the decoding process 900, so as to provide for developing a content base that can support the market development of higher resolution displays 16.
Although the relatively-lower-resolution image 38 has been illustrated centered in the associated encoded image 14, it should be understood that the relatively-lower-resolution image 38 could be located anywhere within the associated encoded image 14. Furthermore, there may be certain applications, such as stereo three-dimensional imaging, for which there may be a plurality of relatively-lower-resolution image 38 associated with a common encoded image 14.
Referring to
Referring to
For example, referring to
Alternatively, referring to
In accordance with the third aspect of the image encoding process, the location of the respective centers 128, 130 of the first 110 and second 124 kernel arrays correspond to the same region of the relatively-higher-resolution digitized image 12 so as to be substantially aligned with one another relative to the associated digitized image. The operation(s) of transforming each first kernel array 110 into the corresponding second kernel array 124 is or are symmetric with respect to the corresponding centers 128, 130 thereof and is or are identical for each set of corresponding first 110 and second 124 kernel arrays, which provides using a relatively simple and fast associated data processing algorithm. The operation or operations of transforming each first kernel array 110 into the corresponding second kernel array 124 is or are formulated as associated interpolations, wherein the locations of the transformed pixels in the second kernel array 124 are symmetrically shifted in location by a relatively small perturbation δ, symmetric relative to the center 130 of the second kernel array 124, with a substantially zero total net shift within each second kernel array 124, so as to diminish the influence of the fundamental spatial frequencies of the underlying sampling process that might otherwise cause associated noticeable sampling artifacts in the resulting associated image, and so that each second kernel array 124 appears in the resulting relatively-lower-resolution image 38 to be aligned with the corresponding center 128 of the corresponding first kernel array 110 of the relatively-higher-resolution digitized image 12, so as to provide for a relatively better macroscopic representation of the original image. Accordingly, each first 110 and second 124 kernel array may be macroscopically considered as an effective larger pixel of the underlying digitized image, so that a best down-sampled representation of each second kernel array 124—independent of the others—provides for a corresponding best aggregate representation of the entire relatively relatively-lower-resolution image 38.
In accordance with the third aspect of the image encoding process, the value of each down-sampled pixel 37′ of the second kernel array 124 is generated from a linear spatial interpolation of corresponding associated relatively adjacent pixels 37 of the first kernel array 110, but with the associated interpolation coefficients modified relative to corresponding values for the actual relative spatial locations of the associated original 37 and down-sampled 37′ pixels. More particularly, the associated interpolation coefficients are calculated assuming that the associated down-sampled pixel 37′ is shifted in space by a relatively small perturbation δ relative to a corresponding nominal geometric location. Accordingly, the resulting interpolation process introduces an intentional error in either the value of the down-sampled pixel 37′ at the corresponding nominal location, or in the location of the down-sampled pixel 37′ for a corresponding value. Furthermore, the direction of the spatial shifts is varied for different down-sampled pixels 37′ of the second kernel array 124 so that there is substantially zero net total spatial shift of all geometric locations within each second kernel array 124. This variation in spatial shifts of the down-sampled pixels 37′ within the second kernel array 124 provides for effectively diminishing the influence of the associated fundamental spatial frequencies of the regularly spaced array of geometric down sampled pixel locations, wherein sampling artifacts would otherwise typically result from the interaction of these fundamental sampling spatial frequencies with spatial frequencies of the original image. Accordingly, the above-described spatial shifts effectively provide for sacrificing an increase in location sampling error for the benefit of a decrease in more noticeable sampling artifacts.
For example, for original pixels 37, designated by values A1 and A2 that geometrically span one down-sampled pixel 37′ designated by value B1, with the down-sampled pixel B1 separated from pixels A1 and A2 by distances d1 and d2, respectively, then the value B1 of the down-sampled pixel 37′ is given by the following linear interpolation of values A1 and A2:
and perturbation δ is the effective shift of the down-sampled pixel B1 from its nominal uniformly-spaced location of the corresponding sampling point in the corresponding encoded image 14. If the value of the perturbation δ is set to zero, then the value of the down-sampled pixel B1 is given by a conventional spatial linear interpolation of values A1 and A2 at the nominal location of the down-sampled pixel B1. The modified interpolation method of equations (1)-(3) may be interpreted as either a) resulting from an effective offset of the sampling location from the corresponding geometric location of a corresponding down-sampled pixel 37′ in a regularly spaced down-sampled image pixel array, b) resulting from a conventional linear interpolation applied to the geometric location of an irregularly spaced, down-sampled image pixel array, or c) resulting from a conventional linear interpolation applied to a regularly-spaced down-sampled image pixel array using associated coefficients that are modified with some type of bias that effectively alters their values.
Furthermore, when applied to the interpolation of different pairs of pixels 37 in the first kernel arrays 110, the amount and direction of the perturbation δ is symmetric with respect to the centers 128, 130 of the first 110 and second 124 kernel arrays so that within the second kernel array 124, the sum of all resulting effective shifts in the locations of the down-sampled pixels 37′ is substantially zero. The particular values of the associated interpolation coefficients α, β may be determined empirically so that the resulting encoded image 14 provides the best subjective appearance, wherein the resulting set of associated interpolation coefficients α, β will have an underlying associated corresponding value for the associated perturbation δ.
Generally, the down-sampling ratio R is the ratio of the number of pixels 37 along a single dimension 104, 106 in the first kernel array 110 to the corresponding number of down-sampled pixels 37′ along the same dimension 104, 106 in the second kernel array 124. The down-sampling ratio R—or product of plural down-sampling ratios R—will depend upon the configurations of the original relatively-higher-resolution digitized image 12 in relation to the relatively-lower-resolution display format 18 of the associated displays 16, 16.1 to be used to display the resulting corresponding relatively-lower-resolution image 38 of the associated second two-dimensional array 108 of down-sampled pixels 37′. Furthermore, each down-sampling ratio R has a corresponding particular set of best interpolation coefficients α, β, each of which may be empirically determined for best subjective appearance, and which may be different for different down-sampling ratios R.
For example, standard display resolutions are frequently used for video and computer displays 16, 16.1. For example, the video resolution of Quad Full High Definition (“QFHD”) is 3840×2160 pixels 37. In one embodiment, a resolution of some high resolution computer monitors is 2560×1600 pixels 37, 37′. Accordingly, a down-sampling ratio R of 3-to-2 applied to a QFHD image will produce a relatively-lower-resolution image 38 with 2560×1440 pixels 37′ that will fill the width of such a computer display 16, 16.1. A second down-sampling ratio R of 4-to-3 operating on this intermediate image will result in a standard Full High Definition (“HD”) relatively-lower-resolution image 38″ of 1920×1080 pixels 37, 37′. Equivalently, the 2560×1440 intermediate image contains ⅔ the resolution of the QFHD image in each dimension and the 1920×1080 HD image contains ¾ the resolution of the intermediate image in each dimension. Therefore, relative to the QFHD original image, the HD image has a resolution equal to the product these ratios, or ½ the resolution in each dimension. The 4-to-3 down-sampling ratio R applied in one dimension is also useful with anamorphic imaging for video applications. It should be understood that the 3-to-2 and 4-to-3 down-sampling ratios R described herein are examples that are used for purposes of illustration, and that these particular examples should not otherwise be considered to be limiting.
For example, referring to
Pixels A1 and A3 are located on the corresponding first 134.1 and second 134.2 edges of the first kernel array 110, 110.1, pixel A2 is located at the center 128 of the first kernel array 110, 110.1, and the down-sampled pixels B1, B2 are located on the corresponding first 136.1 and second 136.2 edges of the second kernel array 124, 124.1, wherein down-sampled pixel B1 is interpolated between corresponding pixels A1 and A2, and down-sampled pixel B2 is interpolated between corresponding pixels A3 and A2. Values of α=2 and β=1 for the interpolation coefficients α, β appeared to provide for best subjective appearance of the resulting relatively-lower-resolution image 38, and to provide for relatively fast associated data processing, the latter of which is described more fully hereinbelow.
Referring to
α=γ·(d2+δ)=2 (6)
β=γ·(d1−δ)=1 (7)
so that for associated distances d1 and d2 are each having a value of ½, d1=d2=½, γ=3 and δ=⅙. Accordingly, for best subjective appearance, the values of the down-sampled pixels B1, B2 are interpolated as if each was shifted from its respective nominal location by a distance of ⅙ away from the center 128 of the first kernel array 110, 110.1 to corresponding shifted locations B1*, B2*.
Alternatively, referring to
For values of α=2 and β=1 for the interpolation coefficients α, β, the values of the down-sampled pixels B1, B2 are given by:
Equations (8) and (9) can be implemented by a relatively fast data processing algorithm by using binary shift operations to perform the associated multiplications and divisions, which is provided for by use of appropriate values for the associated interpolation coefficients α, β, assuming that the associated relatively-higher-resolution digitized image 12 and relatively-lower-resolution image 38 each comprise digital values for the associated pixels 37 and down-sampled pixels 37′.
For example, a multiplication of a binary value X by an nth power of 2, i.e. 2n, is equivalent to left shifting the binary value by n bits, which is represented herein by X<<n. Similarly, a division of a binary value Y by an nth power of 2, i.e. 2n, is equivalent to right shifting the binary value by n bits, which is represented herein by Y>>n.
Division by a value of (2n−1) can be approximated using the following formulae based upon the geometric series:
or, for x=2n−1,
Accordingly,
so that, for division of X by 3=22−1, equation (11.1) with n=2 becomes:
According, using binary shift operations for a relatively fast algorithmic implementation for a one-dimensional interpolation of pixels 37: A1, A2, A3 in accordance with a down-sampling ratio R of 3-to-2 so as to generate the corresponding down-sampled pixels 37′: B1, B2 using values of α=2 and β=1 for the interpolation coefficients α, β for best subjective quality of the resulting relatively-lower-resolution image 38, equations (8) and (9) are then approximated by:
where p is equal to half the number of bits in the digital representation of the values of the pixels 37, 37′.
Referring to
The first 132.1 and second 132.2 one-dimensional encoding processes can be performed sequentially, in either order, to provide for transforming each 3-by-3 first kernel array 110, 110.3 into a corresponding 2-by-2 second kernel array 124, 124.3.
Alternatively, referring to
Referring to
Similarly, referring to
If the first 132.1 and second 132.2 one-dimensional encoding processes had been performed sequentially, in a given order, to provide for transforming each 3-by-3 first kernel array 110, 110.3 into a corresponding 2-by-2 second kernel array 124, 124.3, then each second kernel array 124, 124.3 would then be decoded—in combination with the second group 116, 116.1 of the third plurality of pixels 118, 118.1—so as to form the corresponding first kernel array 110, 110.3 by associated second 140.2 and first 140.1 one-dimensional decoding processes performed in the reverse order to the corresponding encoding processes 132.1, 132.2, so that the last one-dimensional encoding process 132.2, 132.1 to have been performed is decoded first, and the first one-dimensional encoding process 132.1, 132.2 to have been performed is decoded last.
For values of α=2 and β=1 for the interpolation coefficients α, β used in the associated first 132.1 or second 132.2 one-dimensional encoding processes, and for pixel 37: A2 having been stored as the residual datum in the second group 116, 116.1 of the third plurality of pixels 118, 118.1, then the resulting equations of the corresponding first 140.1 or second 140.2 one-dimensional decoding processes become:
which can be implemented as follows using binary shift operations for associated multiplications and divisions:
A1=(B1<<1+B1−A2)>>1, (16)
and
A3=(B2<<1+B2−A2)>>1. (17)
Referring to
Referring to
Alternatively, referring to
Accordingly, the encoded images 14.6, 14.7 illustrated in
Alternatively, the 4-to-3 encoding process 144 could operate directly on the relatively-higher-resolution digitized image 12 of the first two-dimensional array 102, either alone, or followed by the 3-to-2 encoding process 126.
For example, referring to
wherein equations (18) and (20) are based on equation (1), and provide for respective interpolations that incorporate symmetric shifts (i.e. perturbations δ) with respect to the respective centers 128, 130 of the first 110, 110.4 and second 124, 124.4 kernel arrays.
Pixels A1 and A4 are located on the corresponding first 134.1 and second 134.2 edges of the first kernel array 110, 110.4, pixels A2 and A3 straddle the center 128 of the first kernel array 110, 110.1, the down-sampled pixels B1, B3 are located on the corresponding first 136.1 and second 136.2 edges of the second kernel array 124, 124.4, and down-sampled pixel B2 is located at the center 130 of the second kernel array 124, 124.4 wherein down-sampled pixel B1 is interpolated between corresponding pixels A1 and A2, down-sampled pixel B3 is interpolated between corresponding pixels A4 and A3, and down-sampled pixel B2 is the average of pixels A2 and A3.
Values of α=3 and β=1 for the interpolation coefficients α, β appeared to provide for best subjective appearance of the resulting relatively-lower-resolution image 38 in combination with relatively fast associated data processing using the binary shifting techniques for multiplications and divisions, as described more fully hereinbelow.
Referring to
α=γ·(d2+δ)=3 (21)
β=γ·(d1−δ)=1 (22)
so that for associated distances d1 and d2 each having a value of ½, d1=d2=½, γ=4 and δ=¼. Accordingly, for best subjective appearance, the values of the down-sampled pixels B1, B3 are interpolated as if each was shifted from its respective nominal location by a distance of ¼ away from the center 128 of the first kernel array 110, 110.4 to corresponding shifted locations B1*, B3*.
Alternatively, referring to
Interpolation equations (18)-(20) can be implemented in a relatively fast data processing algorithm by using binary shift operations to perform the associated multiplications and divisions, which is provided for by use of appropriate values for the associated interpolation coefficients α, β—assuming that the associated relatively-higher-resolution digitized image 12 and relatively-lower-resolution image 38 each comprise digital values for the associated pixels 37 and down-sampled pixels 37′—for example, by choosing values of the interpolation coefficients α, β for which a) the sum (α+β) is the minimum sum that provides for best subjective image quality, b), α and β are non-negative integers, and c) the sum (α+β)=2m, where m is a non-negative integer. Furthermore, the associated data processing operations are facilitated by choosing a value for the associated perturbation δ—that provides for the associated effective shift in sampling location—so that the resulting associated interpolation coefficients α, β become simple fractions that support implementation of the associated interpolation equations with relatively faster binary operations instead of relatively slower, albeit more general, floating-point operations. Although a shift adjustment to facilitate computational efficiency may create an additional error in geometric location of a given pixel value, it is presumed that such an error would not be as visually noticeable as would be the associated reduction in sampling artifacts. However, it should be understood that the interpolation equations can generally be implemented with any kind of operations, for example, either floating-point operations, integer operations, binary operations, or a combination thereof.
The following combinations of interpolation coefficients α, β are examples of values for which (α+β)=2m:
The selection of the particular interpolation coefficients α, β that provide for best resulting image quality can be subjective. Typically the best realistic representation of an original image is a compromise between a high clarity but artifact-laden image at one extreme and a perceptibly soft but artifact-free image at the other. Equations (18)-(20) were applied to a variety of digital test images using the above values for the interpolation coefficients α, β in order to determine which set of values provided for the best subjective image quality. For example, the first set of interpolation coefficients α, β, with α=1 and β=0, the resulting down-sampled image exhibited relatively high clarity but noticeable residual jaggedness in curved lines. For the second set of interpolation coefficients α, β, with α=3 and β=1, the resulting down-sampled image exhibited substantially reduced artifacts with only a slight reduction in clarity. Furthermore, the results using this second set of interpolation coefficients α, β (α=3, β=1) was visually similar to that of far more complicated interpolation methods. Experiments with additional sets of values for the interpolation coefficients α, β yielded at best only very relatively minor improvements in visual quality that were not considered sufficiently significant to justify the presumed increase in reconstruction error.
Using values of α=3 and β=1 for the interpolation coefficients α, β, the resulting equations for the values of the down-sampled pixels B1, B2, B3 are given by:
Equations (19), (23) and (24) can be implemented in a relatively fast data processing algorithm by using binary shift operations to perform the associated multiplications and divisions, which is provided for by use of appropriate values for the associated interpolation coefficients α, β, assuming that the associated relatively-higher-resolution digitized image 12 and relatively-lower-resolution image 38 each comprise digital values for the associated pixels 37 and down-sampled pixels 37′. According, the relatively fast algorithmic implementation of equations (19), (23) and (24) to provide for a one-dimensional interpolation of pixels 37: A1, A2, A3, A4 in accordance with a down-sampling ratio R of 4-to-3 so as to generate the corresponding down-sampled pixels 37′: B1, B2, B3 using values of α=3 and β=1 for the interpolation coefficients α, β for best subjective quality of the resulting relatively-lower-resolution image 38 is approximated by:
B1=(A1+A1<<1+A2)>>2 (25)
B2=(A1+A3)>>1 (26)
B3=(A4+A4<<1+A3)>>2 (27)
Referring to
The third 132.3 and fourth 132.4 one-dimensional encoding processes can be performed sequentially, in either order, to provide for transforming each 4-by-4 first kernel array 110, 110.6 into a corresponding 3-by-3 second kernel array 124, 124.6.
Alternatively, referring to
Referring to
Similarly, referring to
If the third 132.3 and fourth 132.4 one-dimensional encoding processes had been performed sequentially, in a given order, to provide for transforming each 4-by-4 first kernel array 110, 110.6 into a corresponding 3-by-3 second kernel array 124, 124.6, then each second kernel array 124, 124.6 would then be decoded—in combination with the second group 116, 116.2 of the third plurality of pixels 118, 118.2—so as to form the corresponding first kernel array 110, 110.6 by associated fourth 140.4 and third 140.3 one-dimensional decoding processes performed in the reverse order to the corresponding encoding processes 132.3, 132.4, so that last one-dimensional encoding process 132.4, 132.3 to have been performed is decoded first, and the first one-dimensional encoding process 132.3, 132.4 to have been performed is decoded last.
For values of α=3 and β=1 for the interpolation coefficients α, β using in the associated third 132.3 or fourth 132.4 one-dimensional encoding processes, and for pixel 37: A1 having been stored as the residual datum in the second group 116, 116.2 of the third plurality of pixels 118, 118.2, then the resulting equations of the corresponding third 140.3 or fourth 140.4 one-dimensional decoding processes become:
which can be implemented as follows using binary shift operations for associated multiplications and divisions:
where, as for equations (12) and (13) hereinabove, p is equal to half the number of bits in the digital representation of the values of the pixels 37, 37′.
The choice of storing pixel 37: A1 instead of pixel 37: A2 as the residual datum in the second group 116, 116.2 of the third plurality of pixels 118, 118.2 provides for improved accuracy, because otherwise, if pixel 37: A2 had been saved as the residual datum, then pixel 37: A1 would have been given by:
for which the result may have been fractional with an associated truncation error.
Referring to
First, the values of central pixels 37.2: A22, A32, A33, A23 are determined from the central down-sampled pixel 37.2′: B22 in combination with stored pixels 37: γ5, γ6, γ7 as defined in
Then, using equations derived from those illustrated in
Edge pixels 37.3: A12 and A13 are determined from corresponding edge down-sampled pixel 37.3′: B12 in combination with stored pixel 37: γ1 as defined in
Edge pixels 37.3: A21 and A31 are determined from corresponding edge down-sampled pixel 37.3′: B21 in combination with stored pixel 37: γ2 as defined in
Edge pixels 37.3: A42 and A43 are determined from corresponding edge down-sampled pixel 37.3′: B32 in combination with stored pixel 37: γ3 as defined in
Edge pixels 37.3: A24 and A34 are determined from corresponding edge down-sampled pixel 37.3′: B23 in combination with stored pixel 37: γ4 as defined in
Finally, using equations derived from those illustrated in
Referring to
Alternatively, referring to
Referring to
In accordance with the third embodiment of the two-dimensional encoding process 138.3, the corresponding second group 116, 116.3 of the third plurality of pixels 118, 118.3 includes a combination of original pixel values (C12-C25) and calculated pixel values (γ1-γ7) as illustrated in
Accordingly, generally, the above-described one- 132 and two- 138 dimensional encoding processes provide for quickly down-sampling a first plurality of pixels 100, 37 containing a relatively-higher-resolution digitized image 12 so as to generate both a first group 112 of a second plurality of pixels 114 containing a relatively-lower-resolution image 38 and a second group 116 of a third plurality of pixels 118, wherein the first group 112 of the second plurality of pixels 114 alone provides for displaying a relatively high quality lower-resolution representation of the original image, and in combination with the second group 116 of the third plurality of pixels 118 using a corresponding one- 140 or two-142 decoding process provides for relatively quickly reconstructing the original relatively-higher-resolution digitized image 12 substantially without loss of associated visual detail. The above-described one- 132 and two- 138 dimensional encoding processes provide for relatively fast operation by using linear interpolation implemented with integer arithmetic operations during both the encoding and decoding processes. In one set of embodiments, associated multiplications and divisions are implemented using binary shift operations where possible to provide for faster associated data processing.
The sampling shifts, i.e. perturbations δ, that are symmetric with respect to the centers 128, 130 of each associated kernel array 110, 124 constituting the digitized image provide for a net zero shift within each kernel array 110, 124, and provide for the associated linear interpolation process to be applied with a given down-sampling ratio R. Providing there are a sufficient number of pixels 37 in the original relatively-higher-resolution digitized image 12, a plurality of associated encoding processes 132, 138 using various associated down-sampling ratios R may be applied in sequence so as to effectively produce a new down-sampling ratio R equal in value to the product of all the individual down-sampling ratios R. Furthermore, if an original relatively-higher-resolution digitized image 12 does not have a sufficient number of pixels 37 for an integral number of associated first kernel arrays 110, then the relatively-higher-resolution digitized image 12 can be padded with zero-valued pixels 37 as necessary to accommodate an integral number of associated first kernel arrays 110, followed by appropriate truncation of the resulting zero values in the second kernel arrays 124 of the down-sampled relatively-lower-resolution image 38. It should be understood that the method of creating associated optimized encoding algorithms for specific resolutions of original and down-sampled images, as well as the algorithms themselves, may therefore be find application beyond the particular examples illustrated herein.
Referring again to
Processes for mitigating such distortion are described hereinbelow in respect of an example of an abstract encoded image 14, 14.10 and an associated process illustrated in
The encoded HD (EHD) content and extended data (ED) content may be incorporated together within a common associated encoded image 14, 14.10—i.e. comprising a third two-dimensional array 111—in a variety of formats. For example, in accordance with one set of embodiments of a first aspect of formatting the encoded HD (EHD) content and extended data (ED) content within the associated encoded image 14, 14.10, the encoded HD (EHD) content is incorporated as a single contiguous two-dimensional array of pixels 37—i.e. a fifth two-dimensional array 122—within the encoded image 14, 14.10, and the extended data (ED) content is stored in a complementary region or regions of pixels 37 of the encoded image 14, 14.10. Accordingly, in respect of the above example illustrated in
Referring to
The above-described errors in the reconstruction of the relatively-higher-resolution digitized image 12—caused by the process of standard image formatting 40 of the encoded image 14 formatted in accordance with the first aspect—can also be mitigated by appropriate configuration of the associated encoding 126, 132, 138, 144, 146 or decoding 140 processes so as to reduce the sensitivity of resulting values of reconstructed pixels 37 of the relatively-higher-resolution digitized image 12 to errors in the associated formatted encoded image 14′, and in the encoded image 14 reconstructed therefrom, caused by the process of standard image formatting 40 and by the associated inverse process.
For example, the following examples illustrate the affect of particular formulations of the equations associated with a 4-to-3 encoding process 144 and associated decoding process 140 on the sensitivity of resulting values of reconstructed pixels 37 of the relatively-higher-resolution digitized image 12 to errors in the associated extended data (ED) content caused by the associated process of standard image formatting 40, e.g. compression, or by the associated inverse process.
From equations (19), (23) and (24), the values Bi of the second kernel array 124 are given from the corresponding values Ai of the first kernel array 110, as follows for a one-dimensional encoding process 132:
In accordance with a first aspect of a one-dimensional decoding process 140′, either A1 or A4 is stored—for example, A1—in the extended data (ED) content, and the resulting corresponding reconstructed values Ai of the first kernel array 110 are given as a function of the value of A1 in combination with the values B, of the second kernel array 124 by:
Accordingly, if the process of standard image formatting 40 in generating the formatted encoded image 14′ from the encoded image 14, or its inverse in generating the encoded image 14 from the formatted encoded image 14′, causes an error ε in the value of A1, then reconstructed value
Accordingly, by substituting the value
and
Accordingly, the maximum error in the reconstructed pixels 37 is less than or equal to three times the error ε in the values A1 of the pixels 37 of each group of four pixels 37 that are stored in the extended data (ED) content.
In accordance with a second aspect of a one-dimensional decoding process 140″, either A2 or A3 is stored—for example, A2—in the extended data (ED) content, and the resulting corresponding reconstructed values Ai of the first kernel array 110 are given as a function of the value of A2 in combination with the values Bi of the second kernel array 124 by:
Accordingly, if the process of standard image formatting 40 in generating the formatted encoded image 14′ from the encoded image 14, or its inverse in generating the encoded image 14 from the formatted encoded image 14′, causes an error ε in the value of A2, then reconstructed value
Accordingly, by substituting the value
Accordingly, the maximum error in the reconstructed pixels 37 is less than or equal to the error ε in the values A2 of the pixels 37 of each group of four pixels 37 that are stored in the extended data (ED) content.
Generally, the extended data (ED) content is not limited to storing values of original pixels 37 of the relatively-higher-resolution digitized image 12, but instead, may contain values derived the original pixels 37 of the relatively-higher-resolution digitized image 12, for example, from a combination of values of original pixels 37 sufficient in combination with the encoded HD (EHD) content to algebraically reconstruct the original pixels 37 of the relatively-higher-resolution digitized image 12. For example, the encoded HD (EHD) content may be based upon a difference of two original pixel values having the least influence on the down-sampled and reconstructed images. For example, in accordance with a third aspect of a one-dimensional decoding process 140′″, the difference between A2 and A3 of two relatively central pixels 37, for example,
C=A2−A3 (39)
is stored in the extended data (ED) content, and the resulting corresponding reconstructed values Ai of the first kernel array 110 are given as a function of the value C in combination with the values Bi of the second kernel array 124 by:
(found by simultaneously solving A2−A3=C and A2+A3=2B2 from equations (39) and (29), respectively)
Accordingly, if the process of standard image formatting 40 in generating the formatted encoded image 14′ from the encoded image 14, or its inverse in generating the encoded image 14 from the formatted encoded image 14′, causes an error ε in the value of C, then reconstructed value
Accordingly, by substituting the value
Accordingly, the maximum error in the reconstructed pixels 37 is less than or equal to half the error ε in the values of C of each group of four pixels 37 that are stored in the extended data (ED) content.
Accordingly, in reconstructing the values of four original pixels 37 from corresponding values of three down-sampled pixels 37′ as described hereinabove, the value stored in the extended data (ED) content is adapted to have relatively minor affect on the corresponding associated pixels 37 of the relatively-higher-resolution digitized image 12 that are reconstructed from the encoded HD (EHD) content in combination with the extended data (ED) content. Accordingly, any error or noise in a value stored in the extended data (ED) content that has a lesser contribution to corresponding reconstructed values will result in correspondingly lower error or noise in the resulting reconstructed relatively-higher-resolution digitized image 12.
Although the above first through thirds aspects of the one-dimensional decoding process 140′, 140″, 140′″ have been illustrated with respect to a 4-to-3 encoding process 144, it should be understood that a similar analysis can be applied with other encoding processes, for example, the 3-to-2 encoding process 126 or the 6-to-3 encoding process 146 that have been described hereinabove. Generally, for any combination of encoding and decoding processes, the error or noise in the reconstruction of the relatively-higher-resolution digitized image 12 caused by the process of standard image formatting 40 can be reduced by adapting the decoding process, and possibly also the associated encoding process, to minimize the following cost function:
It should be understood that the algebraic manipulations are not necessarily limited to those of the first through thirds aspects of the one-dimensional decoding process 140′, 140″, 140′″ described hereinabove.
Furthermore, the associated encoding processes may also be considered for manipulation to minimize or reduce the contribution of a particular stored original data value to the error or noise in the resulting reconstruction of the relatively-higher-resolution digitized image 12. However, such alteration may result in a tradeoff in the fidelity of the associated relatively-lower-resolution image 38.
For decoding processes for which the value stored in the extended data (ED) content is an algebraic manipulation of values of the original pixels 37 of the relatively-higher-resolution digitized image 12, it is beneficial that the resulting value that is stored in the extended data (ED) content be of similar magnitude to the other pixels 37 of the encoded image 14 so as to mitigate against the above-described distortion that might result from the process of standard image formatting 40 in generating the formatted encoded image 14′ from the encoded image 14, or its inverse in generating the encoded image 14 from the formatted encoded image 14′. For example, because all the values of pixels 37 in the relatively-higher-resolution digitized image 12 are non-negative, then the above-described distortion can be mitigated to at least some extent if the resulting values that are stored in the extended data (ED) content are also non-negative. For example, in cooperation with the third aspect of the one-dimensional decoding process 140′″, this can be achieved by storing the non-negative value C′ in the extended data (ED) content, instead of just storing the value C, wherein C′ is given by:
wherein γ is the maximum value of a pixel 37 (for example, 255 for an 8-bit representation of a single color of a pixel value).
Accordingly, by appropriate formulation of the associated decoding process, the reduction in susceptibility to errors ε caused by the process of standard image formatting 40 in generating the formatted encoded image 14′ from the encoded image 14, or its inverse in generating the encoded image 14 from the formatted encoded image 14′, provides for a substantial improvement in the resulting reconstructed relatively-higher-resolution digitized image 12 that is sufficient to achieve acceptable results when using the first aspect of formatting the encoded HD (EHD) content and extended data (ED) content within the associated encoded image 14, 14.10 because of substantially less sensitivity to compression errors and associated noise. This is because the associated values stored in the extended data (ED) content are differences in values of adjacent pixels 37 in the original relatively-higher-resolution digitized image 12, so that the variation in intensity or color between adjacent corresponding pixel values of the extended data (ED) content is relatively small and therefore relatively less sensitive to associated errors from image compression during the process of standard image formatting 40. Furthermore, any error resulting from image compression and decompression during the process of standard image formatting 40 in generating the formatted encoded image 14′ from the encoded image 14, or its inverse in generating the encoded image 14 from the formatted encoded image 14′, would be affecting the corresponding difference between neighboring pixels 37 in the reconstructed relatively-higher-resolution digitized image 12, rather then directly affecting the values of the pixels 37, which substantially reduces the visibility of such an error.
Referring to
In accordance with a fourth aspect of an associated decoding process, the encoded HD (EHD) content may be decoded directly to provide an approximation of the relatively-higher-resolution digitized image 12 without necessitating the associated extended data (ED) content, whereby the data of the extended data (ED) content is instead estimated from the associated data of the encoded HD (EHD) content, so as to provide for displaying the resulting approximation of the relatively-higher-resolution digitized image 12 on a display 16 having a relatively higher resolution than the resolution of the relatively-lower-resolution image 38 provided by the encoded HD (EHD) content alone.
For example, equations (8) and (9) of the one-dimensional 3-to-2 encoding process 126,—for example, as illustrated in
wherein
Equations (14) and (15) of the corresponding associated one-dimensional decoding process 140.1, 140.2 illustrated in
wherein Ē is the associated transformation matrix that provides for implementing the associated one-dimensional decoding process 140.1, 140.2.
Substituting equation (45) into equation (46), if the residual pixel 37 value (A2) is saved and used, the associated one-dimensional decoding process 140.1, 140.2 provides for reconstructing all pixels exactly, as follows:
However, for purposes of decoding in accordance with the fourth aspect of the associated decoding process, the value (A2) corresponding to the residual pixel 37 of the 3-to-2 encoding process 126 may be approximated from the values of the second plurality of pixels 114 (B1, B2), for example, by an approximation function ƒ(B1, B2), so as to provide for reconstructing an approximation of the relatively-higher-resolution digitized image 12 for display on a display 16 having a relatively higher resolution than the resolution of the relatively-lower-resolution image 38 provided by the encoded HD (EHD) content of the second plurality of pixels 114 (B1, B2), alone, without requiring the corresponding original second group 116 of the third plurality of pixels 118 (A2). Accordingly, using this approximation ƒ(B1, B2) for the residual pixel 37 value (A2), the corresponding approximation of the corresponding pixels 37 (A1, A2, A3) of the relatively-higher-resolution digitized image 12 for the one-dimensional decoding process 140.1, 140.2 is then given by:
wherein
For example, in one embodiment, the value (A2) corresponding to the residual pixel 37 is approximated by the average of surrounding values of the second plurality of pixels 114 (B1, B2), or ƒ(B1, B2)=(B1+B2)/2, so that equation (48) becomes:
From equation (45), the vector
wherein
The approximation of the pixels 37 (A1′, A2′, A3′) of the relatively-higher-resolution digitized image 12 is then given as follows by substituting equation (50) into equation (49):
wherein the matrix
The difference between the approximate (A1′, A2′, A3′) and actual (A1, A2, A3) values of the pixels 37 of the relatively-higher-resolution digitized image 12 is then given by:
and the associated sum of squares of the differences between the approximate (A1′, A2′, A3′) and actual (A1, A2, A3) values is then given by an error measure Q as follows:
Q=Trace(ĀT·[
which, for a given set of pixel 37 values (A1, A2, A3), provides a measure of the quality or fidelity of the approximation, and which can be used to select amongst possible approximation functions ƒ(B1, B2) so as to provide improving the quality or fidelity of the approximation, if possible.
As a second example of the fourth aspect of the associated decoding process, equations (23), (19) and (24) of the one-dimensional 4-to-3 encoding process 144, —for example, as illustrated in
wherein
Equations (34.1), (29) and (30.2) of the corresponding associated one-dimensional decoding process 140.3, 140.4 illustrated in
wherein
Substituting equation (54) into equation (55), if the residual pixel 37 value (A2) is saved and used, the associated one-dimensional decoding process 140.3, 140.4 provides for reconstructing all pixels exactly, as follows:
However, for purposes of decoding in accordance with the fourth aspect of the associated decoding process, the value (A2) corresponding to the residual pixel 37 of the 3-to-2 encoding process 126 may be approximated from the values of the second plurality of pixels 114 (B1, B2), for example, by an approximation function ƒ(B1, B2, B3), so as to provide for reconstructing an approximation of the relatively-higher-resolution digitized image 12 for display on a display 16 having a relatively higher resolution than the resolution of the relatively-lower-resolution image 38 provided by the encoded HD (EHD) content of the second plurality of pixels 114 (B1, B2, B3), alone, without requiring the corresponding original second group 116 of the third plurality of pixels 118 (A2). Accordingly, using this approximation for the residual pixel 37 value (A2), the corresponding approximation of the corresponding pixels 37 (A1, A2, A3, A4) of the relatively-higher-resolution digitized image 12 for the one-dimensional decoding process 140.1, 140.2 is then given by:
wherein
For example, in one embodiment, the value (A2) corresponding to the residual pixel 37 is approximated by the average of surrounding values of the second plurality of pixels 114 (B1, B2), or ƒ(B1, B2)=(B1+B2)/2, so that equation (57) becomes:
From equation (54), the vector
wherein
The approximation of the pixels 37 (A1′, A2′, A3′, A4′) of the relatively-higher-resolution digitized image 12 is then given as follows by substituting equation (59) into equation (58):
wherein the matrix
The difference between the approximate (A1′, A2′, A3′, A4′) and actual (A1, A2, A3, A4) values of the pixels 37 of the relatively-higher-resolution digitized image 12 is then given by:
and the associated sum of squares of the differences between the approximate (A1′, A2′, A3′, A4′) and actual (A1, A2, A3, A4) pixel 37 values is then given by an error measure Q as follows:
Q=Trace(ĀT·[
which, for a given set of pixel 37 values (A1, A2, A3, A4), provides a measure of the quality or fidelity of the approximation, and which can be used to select amongst possible approximation functions ƒ(Bh B2, B3) so as to provide improving the quality or fidelity of the approximation, if possible.
Similarly, the fourth aspect of the decoding process may be applied to the two-dimensional decoding processes 142.1, 142.2, wherein, for example, the associated approximation functions ƒ( ) may generally be dependent upon values from either adjacent rows, adjacent columns, or both adjacent rows and adjacent columns of the associated first group 112 of the second plurality of pixels 114. For example, for the second kernel array 124, 124.3 illustrated in
The above-described encoding and decoding processes can be adapted for relatively fast computation, for example, using integer arithmetic operations instead of floating point operations, or using relatively fast binary shift operations for multiplications or divisions by powers of two. With division operations, a non-integer result may need to be approximated by an integer result, which is most efficiently accomplished by simply truncating the resulting value to exclude any remainder. However, truncation is not as accurate as rounding when the value being truncated is closer in magnitude to the next higher integer value. A more accurate approach would be to round the initial value to the closest integer. However, rounding is inherently a more complex process and computationally expensive than truncation, and therefore less desirable. The affect of a resulting quotient truncation error may be mitigated in a division operation by a priori increasing the value of the dividend by an amount sufficient for the result to be relatively close in a value to a corresponding result that would have been achieved by rounding. For example, when dividing an integer by a value of four, there are only four possible remainders, as follows: 0.0, 0.25, 0.50 and 0.75. A truncation of this remainder results in an associated truncation error. Assuming that each remainder has an equal likelihood, then the corresponding average truncation error would be 0.375. The average truncation error can be reduced by adding a value of half the divisor to the dividend. For example, for a divisor having a value of four, the addition of a value of two to the dividend results in possible net truncation errors of 0.0, 0.25, 0.50 and 0.25, resulting in an average truncation error of 0.25. Similarly, for a divisor having a value of three, the addition of a value of one to the dividend will similarly reduce the associated average truncation error.
Furthermore, if an arithmetic operation would result in a pixel value having a value that is either less than zero or greater than the maximum pixel value—for example, 255 for a pixel color represented by 8 bits, —then such a value would be replaced with a corresponding clipped value, for example, replaced with zero if the value is below zero or replaced with 255 if the value is greater than 255.
If the cumulative effects of truncation and other computational errors still ultimately impact the fidelity of resultant images, additional modifications similar to those above may be contemplated to improve the empirical result.
Referring to
B1=(4A1+A2)/5, (63)
B2=(A2+A3)/2, (64)
B3=(4A4+A3)/5, (65)
and
γ=(A3−A2+μ)/2. (66)
or generally in respect of equations (63) and (65):
B1=(αA1+βA2)/(α+β), (63.1)
B3=(αA4+βA3)/(α+β), (65.1)
and
wherein μ is the maximum value of a pixel 37 (for example, 255 if a single color of a pixel value is represented by 8 bits). Alternatively, referring to
Alternatively, equations (63)-(66) may be expressed in vector-matrix form as:
The second kernel arrays 124, 124.5, 124.6 and the corresponding associated extra data pixels γ are respectively stored in a corresponding first 112, 112.2 and second 116, 116.2 groups of the second 114, 114.2 and third 118, 118.2 pluralities of pixels, as described more fully hereinabove.
Accordingly, an original relatively-higher-resolution digitized image 12 having 4N pixels 37 in one dimension is treated as having N cells, i.e. first kernel arrays 110, 110.4, 110.5, of four pixels 37 each, wherein N is a positive integer, and a down-sampled relatively-lower-resolution image 38 having 3N pixels in the same dimension is treated as having N cells, i.e. second kernel arrays 124, 124.5, 124.6, of three pixels each. The sequential position of any original image cell 110, 110.5, 110.6 is in correspondence with the same sequential position of the down-sampled image cell 124, 124.5, 124.6, wherein the four pixel 37 values of a given cell A (110, 110.5, 110.6) of the original image are identified sequentially as A1, A2, A3 and A4 and the three down-sampled pixel 37′ values of cell B (124, 124.5, 124.6) of the down sampled relatively-lower-resolution image 38 are identified sequentially as B1, B2 and B3 Accordingly, A1 is on one side, or the “edge”, of the cell A (110, 110.5, 110.6) corresponding to the edge pixel B1 on the same side. Similarly, A4 and B3 are the corresponding edge pixels 37, 37′ on the opposite side of their respective cells A, B.
Referring to
A2=B2−γ+μ/2, (68)
A1=(5B1−A2)/4, (69)
A3=2B2−A2, (70)
and
A4=(5B3−A3)/4. (71)
or, in respect of equations (69) and (71), generally:
A1=((α+β)B1−βA2)/α, (69.1)
A4=((α+β)B3−βA3)/α. (71.1)
Alternatively, equations (68)-(71) may be expressed in vector-matrix form as a function of the down-sampled pixels 37′: B1, B2, B3 and the extra data pixel γ as:
The formulation of equations (68)-(71) or equation (72) beneficially provides for a greater—i.e. more heavily weighted—dependence of the reconstructed pixels 37: A1, A2, A3, A4, upon the associated down-sampled pixel values 37′: B1, B2, B3 than upon the value of the associated extra data pixel γ, thereby providing for increased fidelity of the reconstructed relatively-higher-resolution digitized image 12 if the extra data pixel γ is either erroneous or must be estimated.
Furthermore, equations (68)-(71) or equation (72) may be beneficially reformulated so as to provide for the evaluation thereof using relatively fast binary operations. The reformulated equations may be adapted with additional values so as to minimize the effects of associated truncation errors, and optimized so as to reduce or minimize associated aggregate global errors as opposed to minimizing the error of each associated operation within the associated algorithm. Furthermore, the reformulated equations may be adapted to reduce or minimize the relative computational complexity of the associated reconstruction algorithm, for example, so as to reduce or minimize the number of operations during reconstruction. For example, equations (63)-(66) and (68)-(71) may be rewritten as follows to take advantage of binary operations while simultaneously minimizing aggregate truncation error and minimizing the computational complexity of the associated reconstruction process:
B1=(A1<<2+A2+2)/5 (73)
B2=(A2+A3)>>1 (74)
B3=(A4<<2+A3+2)/5 (75)
γ=(A3−A2+μ)>>1 (76)
A2=B2−γ+half_μ (77)
A1=(B1<<2+B1−A2)>>2 (78)
A3=B2<<1−A2 (79)
A4=(B3<<2+B3−A3)>>2 (80)
wherein half_μ is the integer value of μ/2, the symbol “<<” indicates that the binary form of the preceding value is shifted left by the number of bits indicated to the right of the symbol, and the symbol “>>” indicates a similar operation but shifting to the right by the indicated number of bits.
For example, X>>2 means shifting the binary representation of X twice to the right, which excluding the affects of truncation, is equivalent to dividing X by 4. It should be understood that individually, equations (73)-(80) may result in corresponding associated individual truncation errors. Accordingly, in one embodiment, each associated computed value is set to the nearest extreme of an acceptable value range if that computed value is beyond that range. For example, if A2 is found to be less than zero and the acceptable range is zero to some positive integer, then A2 would be set to zero.
Accordingly, equations (73)-(76) provide for implementing the associated one-dimensional encoding process 132.5, 132.6, and equations (77)-(80) provide for implementing the corresponding associated one-dimensional decoding process 140.5, 140.6, using relatively simple binary operations that can be evaluated relatively quickly.
For example, in one set of embodiments, where possible, the computational complexity of reconstruction process—for example, as embodied by the one-dimensional decoding process 140.5, 140.6 for a one-dimensional reconstruction process—is minimized, even at the expense of greater relative complexity of the associated encoding process—for example, as embodied by the one-dimensional encoding process 132.5, 132.6 for a one-dimensional encoding process—when generating image content for mass distribution, for example, the creation of an optical video disc, for example, a BLU-RAY DISC™. In such applications, the associated down-sample process, i.e. the image encoding process, does not necessarily need to be fast or may be performed by relatively fast processing equipment. However, the resulting image-encoded products used for playback of such content generally require real time reconstruction processing using relatively simpler or less powerful computational resources that benefit from reduced computational complexity in any algorithms implemented thereby, for example, as would be associated with the decoding or reconstruction operations in playback components such as BLU-RAY players.
It has been observed that the relatively-lower-resolution image 38 down-sampled in accordance with equations (73)-(76) resulted in a relatively minor increase in associated aliasing artifacts relative to the original relatively-higher-resolution digitized image 12. However these artifacts were relatively inconsequential for image test patterns and virtually unnoticeable in actual imagery.
The above-described fourth aspect of the associated decoding process provides for decoding the associated encoded HD (EHD) content directly without necessitating the associated extended data (ED) content. For example, equation (57) provides for estimating four sequential pixels 37: A1, A2, A3, A4 responsive to the values of associated down-sampled pixels 37′: B1, B2, B3 without necessitating a value for a corresponding extra data pixel γ. However, in accordance with a fifth aspect—which is governed by the same transformation equations as the fourth aspect, —the associated decoding process can be used to up-sample an original relatively-lower-resolution image 38 so as to form a corresponding approximated relatively-higher-resolution digitized image 12 suitable for display on a corresponding relatively-higher-resolution display 16.
In accordance with another example, the fourth and fifth aspects of the associated decoding process may be adapted so as to depend upon an estimate of the difference, or equivalently, the slope, between two pixels 37, 37′, so as to provide for further reducing or minimizing associated reconstruction errors.
More particularly, in accordance with one embodiment, substituting equation (66) in equation (68), or equation (76) in equation (77), respectively, gives:
A2=B2−(A3−A2)/2, (81.1)
or
A2=B2−(A3−A2)>>1. (81.2)
Accordingly, the only extra data needed to reconstruct A2 is the difference between A3 and A2 or, equivalently, the slope between those two pixel 37 values. Furthermore, the value of A2 is primarily responsive to the value of B2, a known value, because in calculating the value of A2, the value of B2 is weighted by unity, whereas the corresponding slope between A3 and A2 is weighted by half. Accordingly, if the slope between A3 and A2 can be estimated, equations (81.1 and 81.2) would be expected to provide a more accurate estimate of A2 than that provided by estimating A2 as a weighted average of B1 and B2, for example, as in equation (57).
For example, in accordance with one embodiment, the slope between A3 and A2 is estimated as a weighted average of known slopes closest to those points. More particularly, the slope determined by values B3 and B2 is at a location that is relatively close to one side of A3 in sampling space and the slope determined by values B2 and B1 is at a location that is relatively close to the opposite side of A2 in sampling space. Accordingly, in one embodiment, the slope between A3 and A2 is estimated by the average value of the above two slopes, i.e. the slope determined by the values B3 and B2 averaged with the slope determined by the values B2 and B1, which mathematically resolves to half the difference between B3 and B1, i.e. (B3−B1)/2. Furthermore, because the distance between pixel locations B3 and B1 is not the same as the distance between pixel locations A2 and A3, a multiplier α may be adapted to scale this distance to provide a more accurate estimate of the slope between A3 and A2, for example, so as to provide for the following estimate of A2.
A2=B2−α(B3−B1)>>2 (82)
The value of multiplier α can be determined either directly or empirically. For example, the multiplier α can be empirically determined so as to provide for the best subjective visual fidelity in a given reconstructed image, or a given set of reconstructed images, while also providing for equation (82) to be implemented using relatively simple binary operations. For example, in one embodiment, the value of multiplier α was empirically determined to be 5/4, thereby providing for equation (82) to be implemented with the following relatively simple binary operations:
A2=B2−((B3−B1)<<2+B3−B1)>>4 (83)
As an alternative to the use of a single empirical result, the value of multiplier α may be represented by a range between empirically determined reasonable extremes so as to provide for a particular user may select from within that range the value of the multiplier α that they might consider to provide the subjectively best image reconstruction. For example, a relatively lesser value of 5/32 for the multiplier α produces a somewhat softer but less digital appearance, whereas a relatively greater value of 11/32 for the multiplier α produces a clearer—but more digital and therefore less natural—appearance. Equation (82) may therefore be rewritten to provide a narrow range of multiplier α while simultaneously providing for relatively simple binary operations, as follows:
A2=B2−[(B3−B1)<<3+D1*((B3−B1)<<1)+D2*(B3−B1)]>>5 (84)
wherein parameters D1 and D2 may adopt any of values −1, 0 or 1, depending on user preference, which effectively provides for seven different values of multiplier α in equation (82).
Accordingly, equations (77)-(84) provide for implementing the corresponding associated one-dimensional decoding process 140.5, 140.6, using relatively simple binary operations that can be evaluated relatively quickly, using a relatively minimal amount of estimation, for example, whereby one estimated data value (extra data pixel γ) is used to determine four relatively high resolution pixel 37 values (A1, A2, A3, A4), and whereby the one estimated data value (extra data pixel γ) has a relatively low influence on the estimated pixel 37 values (A1, A2, A3, A4), so that the resulting representation of the relatively-higher-resolution digitized image 12, although possibly imperfect, is of relatively high fidelity, and may be calculated relatively quickly relative to conventional scaling approaches.
Referring to
A2=B2−[(B3−B1)<<3+D1*((B3−B1)<<1)+D2*(B3−B1)]>>5 (84)
A1=(B1<<2+B1−A2)>>2 (78)
A3=B2<<1−A2 (79)
A4=(B3<<2+B3−A3)>>2 (80)
It should be understood that each pixel 37, 37′ will generally comprise a vector of independent image property values, referred to herein as a pixel vector, for example associated values for associated red, green and blue subpixels for color rendering, and that the above described algebraic operations (e.g. multiplication, division, binary shifting, addition, multiplication) on a given pixel would be separately performed on each element of the associated pixel vector or associated subpixels.
It should also be understood that the first group 112 of the second plurality of pixels 114 could be used without necessitating separate storage and/or transmission of the second group 116 of the third plurality of pixels 118, so as to provide for an alternative system and method for displaying a relatively-lower resolution image 38, for example, an alternative method of down-sampling a relatively-higher-resolution digitized image 12.
While specific embodiments have been described in detail in the foregoing detailed description and illustrated in the accompanying drawings, those with ordinary skill in the art will appreciate that various modifications and alternatives to those details could be developed in light of the overall teachings of the disclosure. It should be understood, that any reference herein to the term “or” is intended to mean an “inclusive or” or what is also known as a “logical OR”, wherein when used as a logic statement, the expression “A or B” is true if either A or B is true, or if both A and B are true, and when used as a list of elements, the expression
“A, B or C” is intended to include all combinations of the elements recited in the expression, for example, any of the elements selected from the group consisting of A, B, C, (A, B), (A, C), (B, C), and (A, B, C); and so on if additional elements are listed. Furthermore, it should also be understood that the indefinite articles “a” or “an”, and the corresponding associated definite articles “the’ or “said”, are each intended to mean one or more unless otherwise stated, implied, or physically impossible. Yet further, it should be understood that the expressions “at least one of A and B, etc.”, “at least one of A or B, etc.”, “selected from A and B, etc.” and “selected from A or B, etc.” are each intended to mean either any recited element individually or any combination of two or more elements, for example, any of the elements from the group consisting of “A”, “B”, and “A AND B together”, etc. Yet further, it should be understood that the expressions “one of A and B, etc.” and “one of A or B, etc.” are each intended to mean any of the recited elements individually alone, for example, either A alone or B alone, etc., but not A AND B together. Furthermore, it should also be understood that unless indicated otherwise or unless physically impossible, that the above-described embodiments and aspects can be used in combination with one another and are not mutually exclusive. Accordingly, the particular arrangements disclosed are meant to be illustrative only and not limiting as to the scope of the invention, which is to be given the full breadth of the appended claims, and any and all equivalents thereof.
The instant application is a division of International Application No. PCT/US2012/054532 filed on 10 Sep. 2012, which claims the benefit of the following prior U.S. provisional applications: U.S. Provisional Application Ser. No. 61/532,589 filed on 9 Sep. 2011, U.S. Provisional Application Ser. No. 61/577,638 filed on 19 Dec. 2011, U.S. Provisional Application Ser. No. 61/590,053 filed on 24 Jan. 2012, U.S. Provisional Application Ser. No. 61/601,080 filed on 21 Feb. 2012, and U.S. Provisional Application Ser. No. 61/658,903 filed on 12 Jun. 2012. Each of the above-identified applications is incorporated by reference herein in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
5444491 | Lim | Aug 1995 | A |
5477397 | Naimpally et al. | Dec 1995 | A |
5589993 | Naimpally | Dec 1996 | A |
5621660 | Chaddha et al. | Apr 1997 | A |
5768535 | Chaddha et al. | Jun 1998 | A |
6025882 | Geshwind | Feb 2000 | A |
6184935 | Iaquinto et al. | Feb 2001 | B1 |
6249549 | Kim | Jun 2001 | B1 |
6510177 | De Bonet et al. | Jan 2003 | B1 |
6539120 | Sita et al. | Mar 2003 | B1 |
6683980 | Meier et al. | Jan 2004 | B1 |
6788347 | Kim et al. | Sep 2004 | B1 |
6791578 | Ubillos | Sep 2004 | B1 |
6925249 | Meier et al. | Aug 2005 | B1 |
6990150 | Fang | Jan 2006 | B2 |
6999129 | Shirahama et al. | Feb 2006 | B2 |
7075559 | Ubillos | Jul 2006 | B2 |
7272567 | Fejzo | Sep 2007 | B2 |
7295608 | Reynolds et al. | Nov 2007 | B2 |
7305174 | Meier et al. | Dec 2007 | B2 |
7386049 | Garrido et al. | Jun 2008 | B2 |
7392195 | Fejzo | Jun 2008 | B2 |
7599434 | Reynolds et al. | Oct 2009 | B2 |
7656950 | Garrido et al. | Feb 2010 | B2 |
7667707 | Margulis | Feb 2010 | B1 |
7813428 | Yoon et al. | Oct 2010 | B2 |
7930184 | Fejzo | Apr 2011 | B2 |
7956930 | Sullivan | Jun 2011 | B2 |
7961963 | Sun | Jun 2011 | B2 |
20040017852 | Garrido et al. | Jan 2004 | A1 |
20050031321 | Ubillos | Feb 2005 | A1 |
20060039472 | Barbarien et al. | Feb 2006 | A1 |
20060210185 | Sun | Sep 2006 | A1 |
20080008252 | Xu et al. | Jan 2008 | A1 |
20080211901 | Civanlar et al. | Sep 2008 | A1 |
20080232452 | Sullivan et al. | Sep 2008 | A1 |
20100046612 | Sun et al. | Feb 2010 | A1 |
20110194024 | Stefanoski et al. | Aug 2011 | A1 |
20120075526 | DeHaan et al. | Mar 2012 | A1 |
Number | Date | Country |
---|---|---|
1643771 | Apr 2006 | EP |
2101313 | Sep 2009 | EP |
1643771 | Dec 2009 | EP |
2262230 | Dec 2010 | EP |
2009052772 | Jul 2008 | WO |
Entry |
---|
Schwarz, Heiko, Marpe, Detlev and Wiegand, Thomas, “Overview of the Scalable Video Coding Extension of the H.264/AVC Standard”, IEE Transactions on Circuits and Systems for Video Technology, vol. 17, No. 9, Sep. 9, 2007, pp. 1103-1120. |
Unknown author(s), “Scalable Video Coding ”, Wikipedia, May 4, 2011, downloaded from www.wikipedia.org on Sep. 29, 2011, 5 pages. |
Unknown author(s), “H.264/MPEG-4 AVC ”, Wikipedia, Sep. 22, 2011, downloaded from www.wikipedia.org on Sep. 29, 2011, 5 pages. |
Unknown author(s), “PALplus”, Wikipedia, Jun. 10, 2012, downloaded from www.wikipedia.org on Aug. 21, 2012, 5 pages. |
PCT-ISA/US, International Search Report and Written Opinion of International Searching Authority, Search History, Nov. 30, 2012, 19 pages. |
PCT-IPEA/US, International Preliminary Examination Report on Patentability, Sep. 27, 2013, 37 pages. |
Muresan, D. Darian and Parks, Thomas W., “Adaptively Quadratic (Aqua) Image Interpolation)”, IEEE Transactions on Image Processing, vol. 13, No. 5, May 2004, pp. 690-698. |
DTS, Inc., “DTS-HD High Definition Audio”, 2006, downloaded from www.dts.com on Sep. 11, 2011, 23 pages. |
PCT-ISA/US, International Search Report and Written Opinion of International Searching Authority & Search History in International Application No. PCT/US12/54532, Nov. 30, 2012, 19 pages. |
PCT-IPEA/US, International Preliminary Examination Report on Patentability in International Application No. PCT/US12/54532, Sep. 27, 2013, 37 pages. |
Number | Date | Country | |
---|---|---|---|
20140184636 A1 | Jul 2014 | US |
Number | Date | Country | |
---|---|---|---|
61532589 | Sep 2011 | US | |
61577638 | Dec 2011 | US | |
61590053 | Jan 2012 | US | |
61601080 | Feb 2012 | US | |
61658903 | Jun 2012 | US |
Number | Date | Country | |
---|---|---|---|
Parent | PCT/US2012/054532 | Sep 2012 | US |
Child | 14201746 | US |