This application claims the benefit, under 35 U.S.C. §365 of International Application PCT/IB2008/000223, filed Jan. 31, 2008, which was published in accordance with PCT Article 21(2) on Aug. 6, 2009, in English. Further, this application is related to International Application PCT/IB2008/000224, filed Jan. 31, 2008, entitled, “METHOD AND SYSTEM FOR LOOK DATA DEFINITION AND TRANSMISSION, which was published in accordance with PCT Article 21(2) on Aug. 6, 2009, in English.
The present principles generally relate to multimedia interfaces and, more particularly, to a method and system for look data definition and transmission over a high definition multimedia interface (HDMI).
Currently, when delivering a video content product either for home use or for professional use, there is one singular color decision made for that video delivery product, which is typically representative of the video content creator's intent. However, different usage practices of the content may occur so that the content's color decision may have to be altered. For instance, such different usage practices may involve different display types such as a front projection display, a direct view display, or a portable display, each requiring some change to the color decision to provide an optimal display of such video content.
A method and system in accordance with various embodiments of the present principles address the deficiencies of the prior art by providing look data definition and transmission over a high definition multimedia interface (HDMI).
According to an aspect of the present principles, there is provided a method. The method includes generating metadata for video content. The metadata is for altering the video content before display thereof by accounting for variations between different display devices and variations between different creative intents by a content creator. The method further includes preparing the video content and the metadata for transmission over a high definition multimedia interface.
According to another aspect of the present principles, there is provided a system. The system includes a metadata generator and a metadata transmission preparation device. The metadata generator is for generating metadata for video content. The metadata for altering the video content before display thereof by accounting for variations between different display devices and variations between different creative intents by a content creator. The metadata transmission preparation device is for preparing the video content and the metadata for transmission over a high definition multimedia interface.
The teachings of the present principles can be readily understood by considering the following detailed description in conjunction with the accompanying drawings, in which:
It should be understood that the drawings are for purposes of illustrating the concepts of the invention and are not necessarily the only possible configuration for illustrating the invention. To facilitate understanding, identical reference numerals have been used, where possible, to designate identical elements that are common to the figures.
The present principles advantageously provide a method and system for look data definition and transmission over a high definition multimedia interface (HDMI). Although the present principles will be described primarily within the context of a transmission system relating to a source device and a display device, the specific embodiments of the present invention should not be treated as limiting the scope of the invention.
The functions of the various elements shown in the figures can be provided through the use of dedicated hardware as well as hardware capable of executing software in association with appropriate software. When provided by a processor, the functions can be provided by a single dedicated processor, by a single shared processor, or by a plurality of individual processors, some of which can be shared. Moreover, explicit use of the term “processor” or “controller” should not be construed to refer exclusively to hardware capable of executing software, and can implicitly include, without limitation, digital signal processor (“DSP”) hardware, read-only memory (“ROM”) for storing software, random access memory (“RAM”), and non-volatile storage. Moreover, all statements herein reciting principles, aspects, and embodiments of the invention, as well as specific examples thereof, are intended to encompass both structural and functional equivalents thereof. Additionally, it is intended that such equivalents include both currently known equivalents as well as equivalents developed in the future (i.e., any elements developed that perform the same function, regardless of structure).
Thus, for example, it will be appreciated by those skilled in the art that the block diagrams presented herein represent conceptual views of illustrative system components and/or circuitry embodying the principles of the invention. Similarly, it will be appreciated that any flow charts, flow diagrams, state transition diagrams, pseudocode, and the like represent various processes which may be substantially represented in computer readable media and so executed by a computer or processor, whether or not such computer or processor is explicitly shown
Reference in the specification to “one embodiment” or “an embodiment” of the present principles means that a particular feature, structure, characteristic, and so forth described in connection with the embodiment is included in at least one embodiment of the present principles. Thus, the appearances of the phrase “in one embodiment” or “in an embodiment” appearing in various places throughout the specification are not necessarily all referring to the same embodiment.
Moreover, as used herein, with respect to the transmission and receipt of metadata, the phrase “in-band” refers to the transmitting and/or receiving of such metadata together with the color corrected picture content to be displayed by a consumer device. In contrast, the phrase “out-of-band” refers to the transmitting and/or receiving of the metadata separately with respect to the color corrected picture content to be displayed by a consumer device.
Further, as used herein, the term “scene” refers to a range of picture frames in a motion picture, usually originating from a single “shot”, meaning a sequence of continuous filming between scene changes.
Also, as used herein, the phrase “Look Data Management” refers to the editing of look data, the transmission, and the application.
Additionally, as used herein, the phrase “compact disc player” refers to any of a standard definition compact disc player, a BLU-RAY digital video disc player, a high definition digital video disc player, and so forth.
Moreover, as used herein, “unused gamut profile” refers to a gamut profile that is currently not used in version 1.3 (or any preceding version) of the HDMI Standard.
Further, as used herein, the phrase “look data”, and term “metadata” as it relates to such look data, refers to data such as, for example, integer, non-integer values, and/or Boolean values, used for and/or otherwise relating to color manipulation, spatial filtering, motion behavior, film grain, noise, editorial, and tone mapping. Such look data and/or metadata may be used to control, turn on or turn off relating mechanisms for implementing the preceding, and to modify the functionality of such. Furthermore, look data and/or metadata may include a specification of a mapping table.
For example, in an embodiment directed to color manipulation, a color mapping table could be realized by means of a 1-D LUT (one-dimensional Look Up Table), a 3-D LUT (three-dimensional Look Up Table), and/or 3×3 LUTs. As an example, in the case of a 3-D LUT, such LUT is used to receive three input values, each value representing one color component, Red, Green, or Blue, and producing a predefined triplet of output values, e.g., Red, Green, and Blue, for each individual Red, Green, and Blue input triplet. In this case, the metadata from a content source to a content consumption device (e.g., a display device) would then include a LUT specification.
Another embodiment may involve the specification of a mapping function such as, for example, circuitry and/or so forth for performing a “GOG” (Gain, Offset, Gamma), which is defined as follows:
Vout=Gain*(Offset+Vin)^Gamma, for each color component.
In such a case, the look data and/or metadata would include nine (9) values, one set of Gain, Offset, and Gamma for each of the three color components.
Look data, as used herein, is used to influence these mechanisms; there can be several sets of look data, in order to implement transmission/storage of not only one, but several looks.
Of course, the present principles are not limited to the preceding embodiments and, given the teachings of the present principles provided herein, other embodiments involving other implementations of look data and/or metadata are readily contemplated by one of ordinary skill in this and related arts, while maintaining the spirit of the present principles. Look data is further described herein at least with respect to
For example,
The display device 130 (and/or a device(s) disposed between the transmission medium 120 and the display device 130 and connected to these devices) can include a receiver 161, a storage device 162, and/or a metadata applier 162 for respectively receiving, storing, and applying the metadata.
For example,
At step 404, the look data is prepared for transmission, which can involve, but is not limited to, generating one or more Look Data Elementary Messages for the look data (previously generated at step 402), generating one or more look data packets that respectively include one or more Look Data Elementary Messages, storing look data on a disk, and the like. The method then proceeds to step 406.
At step 406, the look data and the video content are transmitted to a display device using HDMI. Such transmission can involve, for example, but is not limited to, using HDMI color metadata, CEA/HDMI vendor specific information frames, HDMI/CEC (consumer electronic control) protocol, and/or the like. With respect to using HDMI color metadata for the transmission, such use can involve using a gamut Boundary description (GBD) metadata container. With respect to using CEA/HDMI vendor specific information frames for the transmission, such use can involve applying GBD flow control to vendor specific information frames. With respect to using the HDMI CEC protocol for the transmission, such use can involve adding a network abstraction layer on top of CEC, enabling Quality of Service (QoS), and timing CEC to video. The method then proceeds to step 408.
At step 408, the video content is received, stored, and modified in accordance with the look data and the modified video content is displayed on the display device. The method 400 can then be exited.
It is to be appreciated that the preceding order and use of received, stored, and modified can vary depending on the actual implementation. For example, storage can correspond to the metadata being provided on a storage medium and/or can correspond to temporally storing the same on the content rendition side for subsequent processing.
In one embodiment of the present invention, the principles of the present invention are used to create content for High Definition-Digital Video Discs (HD DVDs) and/or BLU-RAY discs by encoding the content in accordance with the International Organization for Standardization/International Electrotechnical Commission (ISO/IEC) Moving Picture Experts Group-4 (MPEG-4) Part 10 Advanced Video Coding (AVC) standard/International Telecommunication Union, Telecommunication Sector (ITU-T) H.264 recommendation (hereinafter the “MPEG-4 AVC standard”), storing the content on a disc, then controlling signal processing units in a display to alter the video data for display. In such an application, look data is stored on the disc. The look data is then transmitted to the display using a high definition multimedia interface (HDMI). Various exemplary methods for using HDMI to transmit the look data are described herein. Of course, it is to be appreciated that the present principles are not limited to solely the described embodiments and, given the teachings of the present principles provided herein, one of ordinary skill in this and related arts will contemplate these and various other embodiments and variations thereof, while maintaining the spirit of the present principles.
It is to be appreciated that, in various embodiments, the present principles can be used in a professional or semiprofessional environment including, but not limited to, processing “Digital Dailies” in motion picture production.
Scene Bound Data
In one embodiment, the look data 500 can be shared among scenes 515 by not updating the look data 500 on scene changes if it is determined that the look data 500 is equal between scenes 515. Thus, the look data 500 stays valid until the look data 500 is invalidated or updated. Such an invalidation can include a disabling of the application of the LDEM metadata, by asserting a “FALSE” to the “Data Valid” tag in a “Look Data Elementary Message”. The alternative is to send a new LDEM with the same Tag ID.
Look Data Packet
In one embodiment of the present invention, for Look Data Packet transmission, a “KLV” (Key, Length, Value) metadata concept is implemented, however, other known Look Data Packet transmission concepts can be implemented. That is, while one or more embodiments are described herein with respect to the KLV metadata concept, it is to be appreciated that the present principles are not limited to solely implementing the KLV metadata concept and, thus, other approaches to implementing the Look Data Packets can also be implemented in accordance with various embodiments of the present invention, while maintaining the spirit of the present principles.
The KLV concept is useful for the transmission devices to understand when a packet transmission is concluded without having to parse the content. This is illustrated in
More specifically and referring to
Further, each packet can include a value field 630 for carrying the payload portion of the packet. In one embodiment, the word size of the payload contents can be determined by a metadata tag. In one embodiment of the present invention, the payload can include, for example, individual “Look Data Elementary Messages”, where another layer of KLV can be used or, alternatively, only KV (Key and Value).
Look Data Elementary Messages
1. Color Manipulation
In one embodiment of the present invention, color manipulation can be defined in a Look Data Elementary Message. That is, color manipulation can be implemented, for example, by one or more 3D-LUT's, one or more 1D-LUT's, and/or one or more 3×3 LUT's. For example, an exemplary definition of such Look Data Elementary Messages is provided in
More specifically,
The length definition section in the Value section 820 of
Word=RED<<20+GREEN<<10+BLUE.
Word=LUT[0]<<20+LUT[1]<<10+LUT[2].
where A1 and B1 is RED or CIE_X, A2 and B2 is GREEN or CIE_Y, and A3 and B3 is BLUE or CIE_Z and the sequence of order is C1-C2-C3. In the Look Data Elementary Message 1200 of
Word=C1<<20+C2<<10+C3.
2. Spatial Filter
In an embodiment of the present invention, spatial filtering control can be specified in a Look Data Elementary Message. For example, the spatial response or frequency response can be altered using spatial domain filtering. One exemplary method of changing the spatial frequency response is to use a bank of finite impulse response (FIR) filters, each tuned to one particular center frequency.
In one embodiment, the frequency response of a picture is manipulated by changing the filter coefficients (C0 . . . CN), in order to enhance or attenuate a frequency detail. For example,
For example,
3. Motion Behavior
In one embodiment, motion behavior control can be specified in a Look Data Elementary Message, utilizing a message that contains information for allowing the display to align the motion behavior to a desired motion behavior. This information carries the specification of the desired motion behavior, and additionally can carry helper data from a content preprocessing unit that simplifies processing in the display. For example,
4. Film Grain
In an embodiment, film grain control can be specified in a Look Data Elementary Message. In one embodiment of the present invention, the film grain message can be taken from the MPEG-4 AVC Standard, payload type=19.
5. Noise
In an embodiment, noise control can be specified in a Look Data Elementary Message. That is, it is possible to add a determined level of White Noise, same to all color channels, or one particular level/behavior per channel within the Look Data Elementary Message for noise. Moreover, in an embodiment, noise can be removed from one or more color channels. In one embodiment, the noise characteristic can be changed by modifying the frequency response in the same manner as the spatial response, as described above.
6. Editorial
In an embodiment, the editorial of one or more scenes can be specified in a Look Data Elementary Message. For example, it is possible to cut out one or more segments of a scene or groups of scenes in accordance with a Look Data Elementary Message of the present invention. As such, the cut scene can be displayed at a later time with an update of the Editorial data. Thus, in an embodiment, a “cut list” of IN and OUT time codes within a particular scene can be transmitted. In one embodiment, the first frame of a scene would have the time code 00:00:00:00 (HH:MM:SS:FF).
7. Tone Mapping
In one embodiment, tone mapping is specified in a Look Data Elementary Message. Tone mapping can be used, for example, when converting a high dynamic range image to a low dynamic range image. As an example, a typical application could be the conversion from a 10 bit encoded image to an 8 bit or 7 bit image. It is to be appreciated that the present principles are not limited to any particular tone mapping algorithm and, thus, any approach to tone mapping can be used in accordance with the present invention, while maintaining the spirit of the present principles. As one example, tone mapping can be specified in a supplemental enhancement information (SEI) message in the MPEG-4 AVC Standard. For example,
Look Data Transmission
In HDMI, there are different methods for transmitting look data. Some exemplary methods for transmitting look data include, but are not limited to, the use of the “Gamut Metadata Package” for data other than gamut metadata, the use of a “Vendor Specific Info Frame”, and the use of consumer electronic control (CEC) vendor specific commands.
1. HDMI Color Metadata
Noting that the current version of the HDMI Specification is version 1.3A, there has been a new method for transferring colorimetric metadata via HDMI since version 1.3 of the HDMI Specification. In one embodiment of the present invention, instead of transmitting only colorimetric metadata, the transmission possibility is used to transmit the “Look Data Packet”. Therefore, the use of a Gamut Profile is proposed, which is not used by the current HDMI specification, version 1.3A, for example GBD_profile=7. HDMI specification version 1.3 allows for up to 800 HDMI packets in one single transmission, but future versions of the specification may provide a different total number of packets. The time for this, however, can last up to 10 video fields, but again, this may change with future versions of the interface specification. With 28 bytes per HDMI packet, this would sum up to 21.8 Kbytes.
Hence, in embodiments of the present invention relating to such look data transmission, it should be ensured that the Look Data Packet is not larger than the maximum size of the HDMI gamut metadata packet. In addition, as Look Data Packets may have to be adapted from scene to scene, where a scene defined to be a range of video fields that share Look Data Packet Data, the scene preceding such instance of an update should be no shorter than the time it takes for “Look Data Packet” (LDP) transmission of the current scene.
For using the HDMI colorimetric metadata according to the specification version 1.3, the length of the packet is to be calculated, and GBD_Length_H (High Byte), and GBD_Length_L (Low. Byte) is to be filled into the first two bytes of the Gamut Metadata Packet, as shown in
In one embodiment, an optional checksum ban be performed of the whole packet including the GBD Header and the Look Data Packet, plus any fill data if applicable.
As depicted, the data can be divided into individual HDMI interface packets for transmission, 22 bytes for the first GBD packet, and 28 bytes for all remaining packets. If the last packet cannot be filled completely with “Look Data Packet” data, then it has to be filled with the aforementioned “fill data” which, in one embodiment of the present invention, can include one or more “0's”. For data flow, the HDMI GBD data flow mechanism is used, with “Next_Field”, “Affected_Gamut_Seq_Num”, “Current_Gamut_Seq_Num”, and “Packet_Seq” (see
Alternatively, a communication method, as described below with respect to “HDMI CEC protocol” can be used, however, the GBD method can be preferential since it features an inbuilt frame synchronization method.
2. CEA/HDMI Vendor Specific Info Frame
Instead of implementing the HDMI GBD metadata as described above, in alternate embodiments of the present invention a “vendor specific Info frame” can be used in accordance with the present invention. Vendor specific info frames are described, for example, in chapter 6.1 of the CEA-861-D Specification. The HDMI Specification permits the use of CEA-861-D info frames, as described in chapter 5.3.5 thereof. In fact, info frame packets are 28 bytes in length. The only difference compared to Gamut Metadata packets is that the packet size is limited to one Info Frame only. In one embodiment of the present invention, it is proposed to use the GBD metadata flow control for a vendor specific info frame as well. That is, in one embodiment, the following modification is used: due to the above mentioned restriction of vendor specific info frames to one packet only, the length field is 5 bits in size only. This means that the length info and cyclic redundancy code (CRC) info should be placed in the “GBD-like” header, the Look Data Packet header (see
Accordingly,
=0 (0b00) Intermediate packet in sequence
=1 (0b01) First packet in sequence
=2 (0b10) Last packet in sequence
=3 (0b11) Only packet in sequence.
3. HDMI CEC Protocol
Consumer electronic control (CEC) is a bidirectional control bus in HDMI. It is a shared media to be used by several audio/visual (NV) devices that are connected to this bus.
It is very slow in nature, with a raw data transmission speed in the range of one hundred to two hundred bits per second. A single vendor specific CEC message has a maximum raw size of 16×10 bits according to HDMI Specification, Version 1.3a, and a maximum of 11×8 bits of raw payload. Considering protocol and data flow overhead of 100%-200%, the transmission of one CEC message will take several seconds. This means that if the same amount of data is transmitted as is the maximum possible with the earlier two methods, namely 21.8 kBytes, it would take several minutes to transmit. This is only true if no other device uses the bus during this time, in which case the transmission time would be further increased.
Therefore, it is advisable that the look data packet is limited in size. Certain Look Data Elementary Messages are impractical to use in a day-to-day use (especially LUT download, see
Nevertheless, considering the payload size of a CEC frame, it is almost unavoidable that the look data packets will be longer than one CEC frame. Due to the fact that CEC has been designed for simple applications, an abstraction layer can be implemented on top of the CEC to make the communication more robust in accordance with an embodiment of the present invention.
More specifically and with respect to the International Electrotechnical Commission Open System Interconnect (ISO/OSI) reference model, the CEC functionality has a physical layer, a data link and parts of the network layer implemented. The quality of service (QoS) is not provided by the network layer. Thus, in an embodiment of the present invention, the QoS is addressed in a layer that is to be implemented on top of the CEC protocol.
Subsequently, the CRC and the look data packet are split into frames of appropriate size for communication by a packet to frames block 3020. In an embodiment of the present invention, a packet size of 88 bytes can be used as an example. In such an embodiment, the first CEC message carries 8 bit of CRC data, and the following messages can then carry 8 bits more payload data since the CRC data needs to be communicated only once per Look Data Packet.
Alternatively and as depicted in
In one embodiment of the present invention, on a reception side, the exact opposite is done. That is, the frames are reassembled to become a look data packet, the checksum is calculated, and then the Look Data Elementary Messages are disassembled from the look data packet to feed the application on the reception side, which typically is a display that now changes the look according to the intent specified by the content creator.
It should be noted that the CRC block 3015 is part of the network layer 2900. In the case of a CRC error, either of the following two exemplary approaches can be performed, although it should be noted that the principles of the present invention are not limited to solely the following described approaches. In a first approach, the look data packet can be discarded. In a second approach, a re-transmission request can be issued. In the case of the first approach, the previously transmitted packet would stay valid.
Referring back to
As opposed to some of the previously mentioned methods, CEC does not have a common time base with the video signal. Therefore, in an embodiment of the present invention to synchronize a look data packet, a validation signal at the end of a LDP transmission is used, in order to time the loading of the parameters and data of the transmitted LDP into the video processing blocks of the sink device. As such, a look data packet gets transmitted over CEC and stays invalid until a special “Validate” CEC command is transmitted.
However, the validation cannot be exactly timed. As such in one embodiment of the present invention, one possibility is to estimate the uncertainty of application time, and ensure that the change in video processing is not disturbed. Scene change blanking can be used. The “Validate” signal can be as short as 1 byte, but with CEC bits overhead it will add up to a minimum of 60 bits plus start bit, as shown in
Therefore, the transmit time can be calculated by: CEC Starttime+20×CEC nominal data bit period. HDMI Specification Version 1.3a suggests a nominal CEC Start Time of 4.5 milliseconds, and a nominal CEC Data Period time of 2.4 milliseconds. This results in 4.5 milliseconds+60×2.4 milliseconds=148.5 milliseconds. Therefore, the “Validate” signal will have a field delay of 9 for 60 Hz, or 8 for 50 Hz, or 4 for 24 Hz. However, this may change with newer versions of the CEC specification in the HDMI specification.
As such, the “Validate” command has to be requested from a CEC generator at least 9 fields prior to application in case of a field frequency of 60 Hz. Since the HDMI sink control is over both video and CEC, in order to avoid the wrong LDP data being applied to a given picture content, it is proposed that the picture content gets blanked or otherwise made not vulnerable to LDP changes during the transition phase. The transition phase is determined to be the time duration between the transmission time assuming the fastest possible CEC transition speed by means of the HDMI specification, and the slowest CEC transmission, plus possible processing delay in the sink.
In order to overcome the synchronization problem of scene based look data packet changes, the following exemplary method is proposed, in accordance with an embodiment of the present invention. That is, to synchronize CEC with video, a physical level CEC action is performed. This can be accomplished, for example, by an apparatus depicted in
More specifically,
At step 3420, a “Validate” signal transmission is requested from the CEC generator in the source device in block 3420. The method 3400 then proceeds to step 3430.
At step 3430, the CEC generator starts transmitting the CEC “Validate” command, starting at the next VSYNC event. The method 3400 then proceeds to step 3440.
At step 3440, the sink receives the “Validate” signal. The method 3400 then proceeds to step 3450.
At step 3450, the sink waits until the transmission is finished and with the next following VSYNC signal event in the sink device, it validates the LDP data. The method 3400 then proceeds to step 3460.
At step 3460, the sink device applies the LDP data content to the video processing blocks. The method 3400 can then be exited.
As mentioned above, the transmission time of a “Validate” signal will be approximately 52.5 milliseconds. Therefore, the “Validate” signal will have a field delay of 4 for 60 Hz, of 3 for 50 Hz and for 24 Hz. Therefore, it is necessary to have the LDP transmission finished and the “Validate” signal transmission initiated approximately 4 frames prior to its application. There will then be no uncertainty of LDP application.
In an alternate embodiment of the present invention, another transmission method can include using novel, future networking possibilities for LDP transmission. For example, HDMI may adopt a new networking channel on a top layer/level in the future, or replace some of the existing HDMI-specific data transmission possibilities. This new transmission method may be based on known networking technology, and may be asynchronous to video. This networking technology can be used to transmit the LDP packets the same way, and to use the methods of CEC transmission, such as video synchronization, and packet control described in conjunction with the various embodiments of the present invention described herein.
Having described preferred embodiments for a method and system for look data definition and transmission over HDMI (which are intended to be illustrative and not limiting), it is noted that modifications and variations can be made by persons skilled in the art in light of the above teachings. It is therefore to be understood that changes may be made in the particular embodiments of the invention disclosed which are within the scope and spirit of the invention as outlined by the appended claims. While the forgoing is directed to various embodiments of the present invention, other and further embodiments of the invention may be devised without departing from the basic scope thereof.
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/IB2008/000223 | 1/31/2008 | WO | 00 | 11/13/2010 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2009/095732 | 8/6/2009 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
5373327 | McGee et al. | Dec 1994 | A |
5838318 | Porter et al. | Nov 1998 | A |
6043909 | Holub | Mar 2000 | A |
6067070 | Suzuki et al. | May 2000 | A |
7053927 | Jones et al. | May 2006 | B2 |
7092742 | Rytivaara et al. | Aug 2006 | B2 |
7583355 | Bogdanowicz et al. | Sep 2009 | B2 |
7822270 | Van Hoof et al. | Oct 2010 | B2 |
20030053683 | Newman et al. | Mar 2003 | A1 |
20040006575 | Visharam et al. | Jan 2004 | A1 |
20040113933 | Guler | Jun 2004 | A1 |
20050146735 | Ternasky et al. | Jul 2005 | A1 |
20050280853 | Newman et al. | Dec 2005 | A1 |
20060110031 | Bala et al. | May 2006 | A1 |
20060280360 | Holub | Dec 2006 | A1 |
20060291569 | Kabuto et al. | Dec 2006 | A1 |
20070022464 | Dean | Jan 2007 | A1 |
20070268411 | Rehm et al. | Nov 2007 | A1 |
20070291179 | Sterling et al. | Dec 2007 | A1 |
20080172708 | Perry et al. | Jul 2008 | A1 |
20080266459 | Butterworth | Oct 2008 | A1 |
20080288995 | Diab et al. | Nov 2008 | A1 |
20090147021 | Glen | Jun 2009 | A1 |
20090161017 | Glen | Jun 2009 | A1 |
20090178097 | Kim et al. | Jul 2009 | A1 |
20090278984 | Suzuki et al. | Nov 2009 | A1 |
20100033627 | Hayashi et al. | Feb 2010 | A1 |
20110013833 | Hoof et al. | Jan 2011 | A1 |
Number | Date | Country |
---|---|---|
1838083 | Sep 2007 | EP |
1838083 | Sep 2007 | EP |
WO 2006039357 | Apr 2006 | WO |
WO2006039357 | Apr 2006 | WO |
2007111843 | Oct 2007 | WO |
WO 2007132877 | Nov 2007 | WO |
Entry |
---|
Vetter, “Data Encoding Protocol Using Key-Length-Value”, Proposed SMPTE Standard 336M, MPEG00/5926, Jan. 2000, 30 pages. |
“High-Definition Multimedia Interface”, Specificatioin Version 1.3, Internet Citation, URL: http://www.hdmi.org/download/HDMI—Spec—1.3—GMI.pdf, Jun. 22, 2006. |
Vetter, “SMPTE Key-Length-Value (KLV) Protocol for Data Encoding”, Video Standards and Drafts, ISO/IEC JTC1/SC29/WG11, MPEG99/M5038, Melbourn, Australia, Sep. 1999. |
Wilkinson, “The SMPTE Data Coding Protocol and Dictionaries”, SMPTE Journal, vol. 109, No. 7, Scarsdale, NY, Jul. 1, 2000, pp. 579-586. |
(International Search Report Dated: Sep. 18, 2008). |
Number | Date | Country | |
---|---|---|---|
20110064373 A1 | Mar 2011 | US |