Video systems have been developed, in part, to transmit video and multimedia data over networks. In some instances, the networks may be wired or wireless and may include television broadcast, satellite, cellular, and WiFi networks. Important to a video experience is the quality of video received for viewing by a user. In the event parts of the video received at a receiving end of a video transmission are lost in the transmitting of the data (e.g., a lossy network), then the user's video viewing experience may be compromised due to video degradation.
A number of techniques have been proposed to compensate for lost video data by making adjustments at the receiving end of a video transmission link. In many respects, such previous techniques operate by considering an average packet loss rate for the data transmitted over a video transmission channel and based on the average packet loss rate for the transmitted video, compensating for that average packet loss rate.
While techniques considering an average packet loss rate for a video transmission channel may or may not address some broad video quality issues to an extent, such methods and systems are not typically adaptive or responsive to a variety transmitted types of video. Improving the effectiveness and efficiency of video transmissions and other data based on information specific to the transmitted data is therefore seen as important.
Aspects of the present disclosure herein are illustrated by way of example and not by way of limitation in the accompanying figures. For purposes related to simplicity and clarity of illustration rather than limitation, aspects illustrated in the figures are not necessarily drawn to scale. Further, where considered appropriate, reference labels have been repeated among the figures to indicate corresponding or analogous elements.
The following description describes a method or system that may support processes and operation to improve effectiveness and an efficiency of a video transmission by providing a mechanism to estimate a packet loss impact due to the loss of each individual packet of a video bitstream. The disclosure herein provides numerous specific details such regarding a system for implementing the processes and operations. However, it will be appreciated by one skilled in the art(s) related hereto that embodiments of the present disclosure may be practiced without such specific details. Thus, in some instances aspects such as control mechanisms and full software instruction sequences have not been shown in detail in order not to obscure other aspects of the present disclosure. Those of ordinary skill in the art will be able to implement appropriate functionality without undue experimentation given the included descriptions herein.
References in the present disclosure to “one embodiment”, “some embodiments”, “an embodiment”, “an example embodiment”, “an instance”, “some instances” indicate that the embodiment described may include a particular feature, structure, or characteristic, but that every embodiment may not necessarily include the particular feature, structure, or characteristic. Moreover, such phrases are not necessarily referring to the same embodiment. Further, when a particular feature, structure, or characteristic is described in connection with an embodiment, it is submitted that it is within the knowledge of one skilled in the art to affect such feature, structure, or characteristic in connection with other embodiments whether or not explicitly described.
Some embodiments herein may be implemented in hardware, firmware, software, or any combinations thereof. Embodiments may also be implemented as executable instructions stored on a machine-readable medium that may be read and executed by one or more processors. A machine-readable storage medium may include any tangible non-transitory mechanism for storing information in a form readable by a machine (e.g., a computing device). In some aspects, a machine-readable storage medium may include read only memory (ROM); random access memory (RAM); magnetic disk storage media; optical storage media; flash memory devices; and electrical and optical forms of signals. While firmware, software, routines, and instructions may be described herein as performing certain actions, it should be appreciated that such descriptions are merely for convenience and that such actions are in fact result from computing devices, processors, controllers, and other devices executing the firmware, software, routines, and instructions.
At operation 105, a plurality of blocks of pixel data comprising a encoded video bitstream are generated. The plurality of blocks of pixel data may be generated by a video encoder. The generated blocks of pixel data may be configured as packets of data, in accordance with one or more video transmission protocols. The size and structure of the packet may be based, in part, on the particular video transmission protocol used. The plurality of blocks of pixel data may typically be transmitted over a network via a transmission channel to a video receiver. The video receiver may include a video decoder to decode the encoded video bitstream.
A video sequence encoded by a video encoder may have side information generated for the video sequence by the video encoder. At operation 110, side information associated with the plurality of blocks of pixel data is generated. In some aspects, the side information may, in at least some part, be descriptive of the video sequence with which it is associated. In some aspects, the side information may be used to enhance or process the encoded video bitstream, including transmitting and processing the video based, at least in part, on data in the side information. In some aspects, processing of the video may include processing operations performed by a decoder at a receiver end of a transmission link. In some embodiments, the side information may include or form part of a header of the associated video, include metadata, or other data structures.
In general, the side information herein may include any additional information other than the encoded video bitstream. In some aspects, the side information may be encapsulated with the video (e.g., embedded in a header of transmitted video) or provided at a platform level by sharing the side information with other functional or operational blocks on the platform. In some aspects, the side information may include metadata descriptive of and associated with the encoded video bitstream.
The flow diagram of
At operation 205, video is received. The video may be received by video encoder 300 including functional components to implement the further operations of process 200. It is noted that an error concealment block may usually be implemented at a decoder to conceal the macroblocks (MBs) lost during a transmission of a video sequence encoded bitstream. In some embodiments herein, an error concealment module may be included at the encoder to estimate the concealed pixel values in each MB during the encoding process, as expressed at operation 205. In particular, video encoder 300 may include an error concealment module 310 to calculate how a packet of video will be concealed if that individual video packet were to be lost during a transmission of the video bitstream.
As an example, if frame copy concealment is used as the error concealment method to conceal lost pixels, then the lost pixel is concealed by copying from the collocated pixel in the previous frame. That is, the concealment pixel value fin (pixel i in frame n)=fin-1, where fin-1 is the compressed pixel value for pixel i in frame n−1 (i.e. the previous frame) and can be obtained from a frame buffer.
At operation 210, a distortion calculation is determined. In some embodiments, based on the estimated concealed pixel value and the compressed pixel value of operation 205, a value is calculated of the distortion caused by the packet loss at a pixel level. That is, a calculation is made to determine what distortion will be introduced if the (current) video packet is lost. For example, distortion calculation module 315 may calculate the distortion between the concealed video packet and the correctly received video packet.
As an example, different distortion metrics may be applied here to calculate the distortion. In some embodiments, a mean square error (MSE) is applied. In such embodiments, a distortion for pixel i in frame n may be expressed by diγn=(fin−fin)2. Other distortion metrics may also be used, including MS-SSIM (Multi-scale Structural SIMilarity) and VQM (Video Quality Metric) as well as other reference and non-reference metrics.
At operation 215, an estimate of how the loss of the current video packet will impact future video frame quality is determined. Referring to
In some embodiments, the frame type and the frame index of a frame n are also known at the encoder. As such, the individual packet loss impact to the group of pictures (GOP) of the video can be estimated based on the above information and an error propagation pattern. For example, for an I (i.e., “intra-coded picture”) or P (i.e., “predicted picture”) frame, the errors in one frame usually propagates to all of the frames in the GOP while for a B (i.e., “Bi-predictive picture”) frame, packet loss only causes distortion in the current frame. In the present disclosure, a model is built to estimate the packet loss impact to the GOP. The model assumes that errors in one frame will propagate constantly to the following frames in a GOP and a packet loss in a P frame causes significant distortion to the adjacent B frame. Thus, the packet loss impact of packet p to the GOP can be estimated by:
where ρtype is the frame type of packet p, ρindex is the frame index in the GOP of packet p, kp is the number of frames affected by the lost packet p defined as:
and Doffset is a constant to capture the distortion to the adjacent B frame:
In some embodiments, the particular algorithms used to determine the packet loss impact estimation herein may be varied to correspond to, for example, other error concealment techniques used by error concealment module 315. However, the general framework disclosed herein is applicable across a variety of use cases.
It is noted that the packet loss impact is calculated or estimated for each packet. Therefore, the side information output by packet loss impact estimation module 320 provides an indication of the importance of each individual video packet based on the packet loss visibility.
Returning to process 200, the process may proceed to operation 220. At operation 220, the individual packet loss impact determined at operation 215 may be reported, stored, or used in the transmission of the video bitstream. The individual packet loss impact may be used in some instances to dynamically adjust the transmission of the video, in an effort to minimize or eliminate the degradation of the video that would result from lost video packets. Thus, a video transmission method may be implemented that prioritizes video packets determined in operations 205-215 to significantly impact the quality of the video at a receiver end.
Applicant has realized the effectiveness of the methods disclosed herein by determining an estimation based on the methods and evaluating the performance of the estimation for three video sequences, including the publically available “Parkrun”, “Stockholm”, and “Shields” video sequences under different bitrates. The video sequences used had a video resolution of 1280×720 and a frame rate of 30 fps. A Vanguard H.264 Codec was used to encode the video in IBPBP coding structure with a GOP size of 30. PSNR (peak signal-to-noise ratio) and MS-SSIM (Multi-Scale Structural SIMilarity) were used as video quality metrics in the estimation method. Comparing the actual packet loss impact to the estimated packet loss impact for a first GOP of Stockholm encoded video at 2.277 Mbps, the estimation method herein accurately matches with the actual packet loss impact, where the mean PSNR and MS-SSIM estimation error are 0.14 and 0.0008, respectively.
Tables 1a and 1b below include mean estimation errors (i.e., mean absolute difference between actual and estimated results) for the Stockholm, Shields, and Parkrun video sequences under different bitrates. Table 1 a uses PSNR as the video quality metric and table 1b uses MS-SSIM as the video quality metric. As demonstrated in the tables 1a and 1 b, the estimation determined by the methods and processes disclosed herein compare very favorably to the actual observed losses.
In some aspects, the packet loss impact estimation determined by a method or process herein may be used to improve or enhance video transmission of a network. In some embodiments, an output of, for example, video encoder 300 at 320 can be used for a video adaptation scheme to improve an end-user's video quality experience.
As an example to demonstrate the effectiveness of methods and systems of the present disclosure, a use case is presented that incorporates the packet loss impact estimation herein with an adaptive channel selection. In the present example, it is assumed that three channels having a packet loss rate of 0%, 1% and 5% are available for video transmission. In a baseline scheme, each of the three available channels is randomly selected for transmission of one third of the video packets. According to an adaptive channel selection scheme using the packet loss impact estimation herein (e.g., a process that may include operation 220 of
In this use case, a Stockholm (300 frames) sequence is encoded and transmitted through the three channels using the baseline scheme and the adaptive channel selection scheme. The average PSNR over 20 runs for the baseline scheme and proposed scheme are 30.19 dB and 31.30 dB, respectively. Thus, it is seen that the packet loss impact estimation disclosed herein can be incorporated with a channel adaption method to improve the end-to-end video quality.
All systems and processes discussed herein may be embodied in program code stored on one or more computer-readable media. Such media may include, for example, a floppy disk, a CD-ROM, a DVD-ROM, one or more types of “discs”, magnetic tape, a memory card, a flash drive, a solid state drive, and solid state Random Access Memory (RAM), Read Only Memory (ROM) storage units, and other non-transitory media. Furthermore, the systems and apparatuses disclosed or referenced herein (e.g., system 200) may comprise hardware, software, and firmware, including general purpose, dedicated, and distributed computing devices, processors, processing cores, and microprocessors. In some aspects, the processes and methods disclosed herein may be delivered and provided as a service. Embodiments are therefore not limited to any specific combination of hardware and software.
Embodiments have been described herein solely for the purpose of illustration. Persons skilled in the art will recognize from this description that embodiments are not limited to those described, but may be practiced with modifications and alterations limited only by the spirit and scope of the appended claims.
Number | Name | Date | Kind |
---|---|---|---|
5872807 | Booth et al. | Feb 1999 | A |
6175259 | Mann et al. | Jan 2001 | B1 |
6606005 | Chang | Aug 2003 | B2 |
6826233 | Oosawa | Nov 2004 | B1 |
7042258 | Booth et al. | May 2006 | B2 |
20030099298 | Rose | May 2003 | A1 |
20050169312 | Cakareski et al. | Aug 2005 | A1 |
20070030894 | Tian et al. | Feb 2007 | A1 |
20070160137 | Guo et al. | Jul 2007 | A1 |
20080089414 | Wang | Apr 2008 | A1 |
20080107154 | Hsiao-Chyi | May 2008 | A1 |
20090103635 | Pahalawatta | Apr 2009 | A1 |
20110080952 | Vadapalli et al. | Apr 2011 | A1 |
Number | Date | Country |
---|---|---|
201028015 | Jul 2010 | TW |
201105112 | Feb 2011 | TW |
2007084475 | Jul 2007 | WO |
2012001377 | Feb 2012 | WO |
Entry |
---|
Liao et al., “Rate-distortion based mode selection for video coding over wireless networks with burst losses,” in Packet Video Workshop, 2009. PV 2009. 17th International, vol., No., pp. 1-10, May 11-12, 2009, hereinafter “Liao”. |
“PCT Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration”, May 29, 2013, for International Application No. PCT/US2013/024864, 10pgs. |
“Notification Concerning Transmittal of International Preliminary Report on Patentability, Chapter I of the Patent Cooperation Treaty”, Oct. 9, 2014, for International Application No. PCT/US2013/031536, 6 pgs. |
“English Translation of Search Report of R.O.C.”, for Taiwan Patent Application No. 102104861, completed Mar. 21, 2016, 1 page. |
“Korean Office Action: The Notice of Preliminary Rejection”, dated Oct. 30, 2015, for Korean Application No. 2014-7027169, 5pgs. |
“English Translation of Korean Office Action: The Notice of Preliminary Rejection”, dated Oct. 30, 2015, for Korean Application No. 2014-7027169, 4pgs. |
Number | Date | Country | |
---|---|---|---|
20130259136 A1 | Oct 2013 | US |