This application is based upon and claims the benefit of priority from the prior Japanese Patent Application No. 2004-106709 filed on Mar. 31, 2004 the entire contents of which are incorporated herein by reference.
1) Field of the Invention
The present invention relates to the broad realm of Internet technology and deals more specifically with hypermedia. Still more specifically, the invention relates to metadata descriptive of information to be exhibited in association with each object (e.g. figure or any other subject of interest) appearing in a motion video program, as well as to systems, methods, and computer programs for delivery of the metadata, as from a metadata distributor on the server side to a video player on the client side, for reconstruction of the received metadata, and for reproduction of the video program in synchronism with the reconstructed metadata.
2) Description of the Related Art
Hypermedia has found widespread use in recent years on the Internet as a refinement of hypertext. Written in Hypertext Markup Language (HTML), hypertext enables the user to access additional web content relating to the textural or still-pictorial matter on the display screen via website hyperlinks. Hypermedia may be envisaged as an adaptation of hypertext to motion video programs. It defines hyperlinks from the figures and other objects in a video program to associated information, either textural or pictorial, that is explanatory or illustrative of the objects. The video viewer may click on a desired object (e.g. one of the individual players in the case of a soccer game program) on the display screen. Thereupon some predefined information on that object (e.g. career of the soccer player that has been clicked) will appear on the display screen. Japanese Patent No. 3,266,278 is hereby cited as dealing with hypermedia.
Video-oriented hypermedia requires use of what is herein termed “object region data” indicative of the time-space zone occupied by each object and immediate vicinity thereof on the display screen. Object region data is obtainable by any such familiar methods as the serial image masking method of binary or greater values, the arbitrary shape coding of MPEG4 by the Moving Picture Experts Group, or the delineation of the loci of the salient points of images. (Reference may be had to ISO/IEC 14496 for more details on MPEG4.) A realization of video-oriented hypermedia demands additional data including action script indicative of a particular action to be taken, in response to the clicking of an object, to show information on that object. (The word “information” will be hereinafter spelled “information” for simplicity.) The term “metadata” as used in the context of video programs refers to all such data indicative of object regions, object properties, and action to be taken.
Video-oriented hypermedia finds application to video CDs and DVDs (digital versatile disks), on which there are prerecorded both video data and metadata. It also lends itself to use on the Internet or other computer network, from which the individual users may download both video data and metadata in streaming mode.
One of the problems encountered in the practice of hypermedia is that not all the video CDs and DVDs on the market or in users' possession today are of hypermedia format; in fact, there are incalculable numbers of non-hypermedia video disks, having video programs recorded thereon but no metadata. A solution to this problem is to make the streaming delivery, over the Internet or like computer network of client/server architecture, of the metadata that has been newly prepared to complement the preexisting video programs on the CDs and DVDs of the kind in question. The viewers are then enabled to derive the full benefits of hypermedia from their inherently non-hypermedia video disks.
As heretofore practiced, however, the streaming distribution of metadata for video programs has had some difficulties left unresolved. The metadata must be distributed in small units in anticipation of data drop-outs and in support of random access to any desired part of the metadata. The individual metadata units must be so configured as to independently perform all the purposes for which they are intended, in the face of drop-outs from other units. These requirements have resulted in redundancy, such that unnecessarily large amounts of metadata have had to be sent over the Internet for the purposes set forth above.
It is an object of the present invention to at least solve the problems in the conventional technology.
Metadata for an object in video according to one aspect of the present invention includes first and second pieces of the metadata each including object property information of the object, property identification information for the object property information, and object region information indicating a position of the object in an image. If the second piece is the same as the first piece in a property of the object property information, the property identification information of the second piece is the same as the property identification information of the first piece and the second piece does not include the property in the object property information of the second piece.
An apparatus for distributing metadata for an object in video according to another aspect of the present invention includes a storage unit storing the metadata including first and second pieces of the metadata, and a metadata distributor sending the first and second pieces to a video player via a network. The first and second pieces of the metadata each include object property information of the object, property identification information for the object property information, and object region information indicating a position of the object in an image. If the second piece is the same as the first piece in a property of the object property information, the property identification information of the second piece is the same as the property identification information of the first piece and the second piece does not include the property in the object property information of the second piece.
An apparatus for playing video with metadata for an object in the video according to still another aspect of the present invention includes a metadata receiver receiving the metadata from a metadata distributor via a network, a metadata interpreter, and a player. The metadata includes first and second pieces of the metadata each including object property information of the object, property identification information for the object property information, and object region information indicating a position of the object in an image. The metadata interpreter obtains, if the property identification information of the second piece is the same as the property identification information of the first piece and the second piece does not include a property of the object property information of the first piece in the object property information of the second piece, the property from the object property information of the first piece, for the object property information of the second piece. The player plays the video with the first piece and the second piece including the property obtained.
A system according to still another aspect of the present invention includes the metadata distributing device and the video playing device according to the invention, which are connected via a network.
A method according to still another aspect of the present invention is of distributing metadata for an object in video. The metadata includes first and second pieces each including object property information of the object, property identification information for the object property information, and object region information indicating a position of the object in an image. The method includes sending the first and second pieces to a video player via a network. If the second piece is the same as the first piece in a property of the object property information, the property identification information of the second piece is the same as the property identification information of the first piece and the second piece does not include the property in the object property information of the second piece.
A method according to still another aspect of the present invention is of playing video with metadata for an object in the video. The method includes receiving the metadata from a metadata distributor via a network. The metadata includes first and second pieces each including object property information of the object, property identification information for the object property information, and object region information indicating a position of the object in an image. The method also includes obtaining, if the property identification information of the second piece is the same as the property identification information of the first piece and the second piece does not include a property of the object property information of the first piece in the object property information of the second piece, the property from the object property information of the first piece, for the object property information of the second piece; and playing the video with the first piece and the second piece including the property obtained.
The computer program product according to still another aspect of the present invention causes a computer to perform the method according to the present invention.
The other objects, features, and advantages of the present invention are specifically set forth in or will become apparent from the following detailed description of the invention when read in conjunction with the accompanying drawings.
Exemplary embodiments of metadata, a metadata distributor, a video player, a system including the metadata distributor and the video player, and methods of distributing metadata and playing video relating to the present invention will be explained in detail below with reference to the accompanying drawings.
Shown in
For delivering metadata on demand from each client, the metadata distributor 101 comprises a hard disk drive (HDD) 113, a session controller 114, and a transmitter 115. The HDD 113 includes hard disks storing the metadata to be delivered to the clients, although other storage media could be employed as well, examples being semiconductor memories and magnetic tape. The metadata stored here is associated with, or complementary to, the inherently non-hypermedia video programs to be reproduced by the video player 100, as will be detailed presently.
The session controller 114 is designed for exchange of control data with the video player 100 via the network 102 for each session of metadata delivery. The transmitter 115 sends the metadata, retrieved from the HDD 113 via the session controller 114, to the video player 100 via the network 102. The transmitter 115 is equipped to schedule metadata delivery so that the metadata requested by the client may be sent in proper timing.
Inputting the streaming metadata from the metadata distributor 101, the video player 100 reproduces video data and processes the metadata in synchronism with the video reproduction. Thus the video player 100 gains hypermedia functions in addition to those normally associated with simple video reproduction. In terms of hardware the video player 100 comprises a player engine 119, an interface handler 107, an audio/video renderer 112, a network manager 120, a media decoder 121, a script interpreter 111, an input device 117, and a display 118.
The player engine 119 includes known means for reproduction of a video program 103 stored, either digitally or otherwise, on any such storage medium 104 as DVD, video CD, hard disk, video tape, or semiconductor memory. The storage medium 104 also stores metadata according to the invention.
The player engine 119 is shown as additionally comprising a controller 105 and an audio/video decoder 106. The controller 105 controls data read out from the storage medium 104. More specifically, in response to user instructions entered on the input device 117 and delivered through the interface handler 107, the controller 105 dictates the start and stop of video retrieval from the storage medium 104 as well as random access to any desired point thereon.
The video program to be reproduced may be recorded compressed on the storage medium 104. Then the audio/video decoder 106 will reconvert the incoming video program into the original form. The video program on the storage medium 104 may be accompanied by audio data, as is usually the case. In that case the audio/video decoder 106 will separate the video and the audio data and decode them independently.
The interface handler 107 controls interfaces between all such component modules of the video player 100 as the player engine 119, network manager 120, metadata decoder 110, script interpreter 111, and audio/video renderer 112.
Additionally, the interface handler 107 is coupled to the input device 117 to receive therefrom each input event and deliver the same to any appropriate module. The audio/video renderer 112 processes the decoded video and audio data from the audio/video decoder 106 for rendering on the display 118. Possibly, video signals may be supplied not only from the audio/video decoder 106 but from the metadata decoder 110 too. In that case the audio/video renderer 112 synthesizes the video signals from both audio/video decoder 106 and metadata decoder 110 and delivers the resulting composite video signals to the display 118. The renderer 112 coacts with the controller 105 and decoder 106 for actual video reproduction as hypermedia.
The network manager 120 comprises a session controller 108 and a receiver 109. The session controller 108 exchanges control data with the metadata distributor 101 via the network 102. The control data sent from video player 100 to metadata distributor 101 includes requests for media or metadata, for session start or session end, for metadata delivery start, and for pause in metadata delivery. The control data sent from metadata distributor 101 to video player 100, on the other hand, includes status data such as OKs and errors.
In response to a metadata request from the video player 100, the metadata distributor 101 sends the desired metadata to the video player in streaming mode via the network 102. The receiver 109 of the network manager 120 inputs the streaming metadata and sequentially transfers the same to the metadata decoder 110 of the media decoder 110.
The metadata decoder 110 analyzes the incoming metadata as taught by the instant invention. Referring to the timestamps of the video data being processed by the audio/video decoder 106, the metadata decoder 110 decodes the required metadata in synchronism with the video data. Further, in response to the object marking information (yet to be detailed) which is included in the metadata, the metadata decoder 110 creates video data needed for marking the object regions, as by masking, and delivers such data to the audio/video renderer 112.
Additionally, the metadata decoder 110 creates object property tables listing correlations between object property information and identification data. For those object metadata units from which object property information has been omitted either in part or in whole, the metadata decoder 110 relies on the object property tables for recovering the missing object property information from the object metadata units bearing the same IDs, as will become better understood as the description progresses.
The metadata decoder 110 has still another function that manifests itself when the viewer specifies some object by clicking or like input event. Ascertaining the object thus specified, the metadata decoder 110 obtains an action command such as that for exhibition of predefined additional information pertaining to the specified object. The obtained action command is sent to the script interpreter 111 via the interface handler 107. A yet further function of the metadata decoder 110 is to erase from the memory any metadata that has become obsolete or unnecessary. The metadata decoder 110 relies for the last mentioned function upon the timestamps included in the metadata and those of the video data being played.
The script interpreter 111 interprets and implements the command specified by the action script (yet to be detailed) included in the object metadata according to the invention. The script to be executed when the viewer specifies an object is sent from the metadata decoder 110 to the script interpreter 111 via the interface handler 107.
The input device 117 may take the form of one or more of such known devices as a mouse, a touchpad, and a keyboard. The display 118 may be of either plasma, light-emitting diode, liquid crystal, or electroluminescent construction.
High reliability is a prerequisite of communication protocol for use in data delivery between video player 100 and metadata distributor 101 via the network 102. An example of protocol meeting this requirement is Real-Time Streaming Protocol/Transmission Control Protocol/Internet Protocol (RTSP/TCP/IP). Real-time Transport Protocol/User Datagram Protocol/Internet Protocol (RTP/UDP/IP) is preferred for metadata delivery from distributor 101 to video player 100 by reason of swiftness of data transmission. The required two-way data transfer between video player 100 and metadata distributor 101 may take place via either the same network or separate ones.
In practice both video player 100 and metadata distributor 101 may be of the familiar computer hardware comprising a central processor unit (CPU), memories including a read-only memory (ROM) and a random access memory (RAM), a HD drive, external storage means such as a CD or DVD drive, a display, and input means such as a keyboard and a mouse. Alternatively, the video player 100 may be a commercial device dedicated solely for reproduction of DVD or other video disks.
The video reproduction program for use on the video player 100 and the metadata distribution program for use on the metadata distributor 101 may both be supplied in the form of installable or executable files. Such files may be recorded on computer-readable storage media such as CD-ROMs, flexible magnetic disks, CD-Rs, or DVDs.
A further, and far more large-scale, application of the invention is a computer network as typified by the Internet. The noted video reproduction program and metadata distribution program may both be installed on a computer connected to the network, thereby permitting the individual users to download the programs. Or the video reproduction and metadata distribution programs may be presented or distributed over the network.
The video reproduction program and metadata distribution program according to the invention are both of modular configuration set forth with reference to
As the viewer points this object region with a cursor 201 and clicks, information relating to that object (e.g. the career of, and other personal information on, the clicked soccer player) will be exhibited as at 203. The object information thus exhibited has been encoded in the metadata stored on the HDD 113,
The metadata for use in the practice of this particular embodiment of the invention includes both object region data indicative of the zones of the objects appearing in each video program recorded on the storage medium 104 of the video player 100, and object property data indicative of how each object is marked or indicated on the display screen by the video player 100 and of the action to be taken by the video player when that object, or object region, is clicked by the viewer.
Reference may be had to
Object regions may be converted into object region data by some such known method as shape coding by MPEG-4 or the time-space zone descriptors of MPEG-7. The MPEG-4 shape coding method and MPEG-7 time-space descriptor method are alike in lessening data amounts by taking advantage of the time correlations of object regions. Consequently, these methods inherently possess the shortcoming that data may become undecodable, or that, in the event of a data dropout at a certain moment in time, not only that data but the neighboring data may also become undecodable. These shortcomings will present no serious problem at all in the illustrated embodiment of the invention, in which each lasting object region is divided into a series of separately encoded segments. The encoding of the successive objective zone segments expedites random access to any such segments and mitigates the effects of partial data dropouts.
As has been set forth with reference to
The object property information on the section 402 indicates how the associated object is marked off (as by contouring, labeling or highlighting on the display screen, and the action to be taken by the video player 200 when the object is clicked. More specifically, as shown in
The hierarchy information on the subsection 500 of the object property information section 402 represents the hierarchical rank of each object in cases where a plurality of object regions overlap one upon another on the display screen. One of such overlapping objects must be chosen as the one that has been specified by the view, it being undesirable that information concerning all the overlapping objects be displayed at one time. An example of hierarchy information is a series of numbers from zero to 255. The smaller the hierarchy number, the more is the object positioned in the foreground of the scene being shown on the display screen. Thus does the hierarchy information makes it possible to choose, for example, the foremost object of the scene when the viewer clicks on the overlapping parts of two or more object regions.
The action script on the subsection 501 of the object property information section 402 specifies one of the possible actions to be taken by the video player 100 upon clicking of an object Such possible actions may include an exhibition of an HTML file, a jump to another video program or to a different scene of the program now in progress, or an execution of an application program. Alternatively, the action script may simply indicate the name of a script file on which is specified the actual action to be conducted. The script file indicated by the script may be cross-referenced by the script interpreter 111,
The object marking information on the subsection 502 of the object property information section 402 determines how the objects, or object regions, are marked or visually indicated on the display screen. For example, an object region may be either highlighted by a difference in brightness, bound off by an outline or contour, or labeled by a name tag attached in the neighborhood of the object The object marking information specifies one of these marking methods and additionally contains the parameters for such marking.
With reference back to
The object ID on the ID subsection 600 uniquely identifies each object appearing in the video program. However, different object IDs may be assigned to one figure or other subject if such one figure or other subject is to be dealt with as different objects.
The media ID on the ID subsection 601 specifies the video program for which the object metadata is intended. Some video programs are so made that, like the multiangle features (yet to be described) of some DV Ds, different images are presented depending upon which of the two or more different languages to be spoken or which of two or more different camera angles are chosen by the viewer. When the video program on the storage medium 104,
The action ID on the ID subsection 602 is identification information for the action script on the object property information subsection 501. The marking information ID on the ID subsection 603 is identification information for the object marking information on the object property information subsection 502.
Notwithstanding the showing of
It is among the features of the present invention that the ID section 403,
All such object metadata units will contain the same action script if, as has been assumed above, the video player 100 is to take the same action in response to the viewer clicking of the same object. In that case, according to the teachings of the present invention, the same object ID may be used for the whole series of object metadata units for the same object, and the object property information under consideration may be omitted from all but at least one of the series of object metadata units. The amount of metadata delivered from metadata distributor 101 to video player 100 will be significantly reduced in this manner.
For each series of object metadata units bearing the same ID there must be at least one where the object property information section 402 is left intact. Inputting the other object metadata units from which object property information has been omitted, the video player 100 will then be capable of regaining the action script and other object property information by referring to the object metadata unit from which the object property information has not been omitted. Object marking information may be regained likewise by the video player 100.
Unlike the showing of
According to a further feature of the invention, by use of the same ID, only changes in the object property information included in the preceding object metadata unit may be delivered from metadata distributor 101 to video player 100. For most efficient delivery of successive object metadata units that differ only slightly in object property information, the same ID may be used for such units, and only the differences in object properties may be sent to the video player.
Receiving such object metadata units, the video player may store the object property information for each ID. The stored object property information may be renewed upon receipt of object property information that requires such renewal. The video player 100 may then perform the required operations as dictated by the object information stored for each ID.
It is also desirable for most efficient metadata delivery that the object metadata units be arranged on the HDD 113,
Referring now more specifically to
The horizontal axes in
Out of this object metadata stream 1206, the video player 100 needs only those sets of object metadata units which meet the viewer settings of camera angle and language. For example, if Camera Angle II is chosen, the object metadata set 1203 is needed during the time period of
Possibly, however, the viewer may frequently change the camera angle and/or language settings. This possibility is a justification of thorough delivery of the complete object metadata stream 1206 from metadata distributor 101 to video player 100. The video player 100 may then be so constructed as to select the required set or sets of object metadata units for each camera angle and language setting.
Flowcharted in
If the view starts the reproduction of a video program from its very beginning, the object metadata stream may be delivered from metadata distributor 101 to video player 100 from its beginning. However, as each video program is random-accessed, so must be the object metadata stream for that program. The object metadata must then be delivered from any required point intermediate the beginning and end of the complete object metadata stream. The Access Point Table is preformed and introduced into the HDD 113,
A closer study of
With reference back to
The timestamps of random-accessed points on the video program may not be listed on the Access Point Table. In such cases an access point corresponding to a timestamp close to the unlisted timestamp may first be ascertained, and where the object metadata starts to be delivered may be determined by referring to the object metadata timestamps in the neighborhood of that access point. Alternatively, the metadata distributor may be programmed to search the Access Point Table for a timestamp earlier than the random-accessed point on the video program and to start metadata delivery at that earlier access point.
The block S1306 of the On-Demand Metadata Delivery Program is followed by another logical node S1307. which asks if all the required object metadata has been sent. The answer “no” to this query directs the program back to the block S1301, so that the steps of S1301-S1306 are cyclically repeated until all the necessary object metadata is delivered to the video player 100.
As has been mentioned in connection with the blocks 1305 and 1306 of the On-Demand Metadata Delivery Program, the object metadata is sent from metadata distributor 101 to video player 100 in the form of a packet or packets.
One series of object metadata may possibly be too large to be accommodated in one packet, in which case the maximum possible proportion of the metadata series may be loaded in one packet, and the balance in the next.
When the user of the video player 100 manipulates the input device 117 for playback of some chosen video program, this video player requests the metadata distributor 101 for information concerning the object metadata stream to be distributed therefrom, by the DESCRIBE method of RTSP (Step S1701). It is understood that the video player 100 has been informed of the IP address of the metadata distributor 101 for the delivery of the metadata corresponding to the video program to be played.
The metadata distributor 101 responds at Step S1702 to the information requests from the video player 100 supplies the information on object metadata in the form of a Session Description Protocol (SDP) file. The SDP file may include all such information as the session protocol version, session ownership, session name, connection information, session time information, metadata name, metadata properties, etc.
Then at Step S1703 the video player 100 requests the metadata distributor 101 for the setup of a session by the SETUP method of RTSP. In response to this request the metadata distributor 101 makes ready the streaming delivery of metadata and sends a session ID to the video player 100 at Step 1704. Then at Step S1705 the video player 100 requests the metadata distributor 101 for transmission of the object metadata by the PLAY method of RTSP. This request includes information indicative of the timestamp of the beginning of the video program. The metadata distributor 101 sends a confirmation message in response to the object metadata transmission request (Step S1706). Then, ascertaining where the object metadata stream starts to be delivered, the metadata distributor 101 actually commences the delivery of the object metadata in packet form to the video player 100 by RTP (Step S1707).
At Step 1708 is shown the video player 100 as sending a session termination request by the RTSP TEARDOWN method, as when the video reproduction by the video player 100 has come to an end or when object metadata delivery from the metadata distributor 101 is to be discontinued. The metadata distributor 101 responds to the session termination request by stopping data delivery, and so ending the session, and proceeds to send a confirmation message to the video player 100 (Step S1709). Thereupon the session ID that has been in use is invalidated.
Receiving the metadata stream from the metadata distributor 101 as above, the video player 100 operates as dictated by the Metadata Decode Program which is preinstalled thereon and which is flowcharted in
Then comes a logical node S1804 where the metadata decoder 110 refers to the media ID on the ID subsection 601,
The controller 105,
Next comes another logical node S1808 which asks whether, for each incoming object metadata unit, the object marking information is omitted or not. The answer “yes” to this query leads to the block S1809, at which the required object marking information is ascertained by referring to one of the Object Property Tables. If the answer to the node S1808 is “no”, on the other hand, then the object marking information can be obtained directly from that object metadata unit.
Still another logical node S1810 is next encountered which asks the metadata decoder 110 whether the ascertained object marking information specifies some marking of the object or not. If it does, the next block S1811 dictates the production of the required marking image such as the masking or outlining of the object region. The object marking image is then sent to the audio/video renderer 112,
The answer to the node S1810 may be “no”; that is, no marking method may be specified for the object region. No marking image need be formed in that case, so that the routine jumps to the final block S1812, bypassing the block S1811.
According to this subroutine the metadata decoder 110 obtains still-untabulated object metadata from the unshown buffer at the block S1901 and gets the IDs of such object metadata at the next block S1901. Then, at a logical node S1903, the metadata decoder 110 determines whether the successive IDs have already been registered on the preexisting Object Property Tables. If they are not, the metadata decoder 110 proceeds to enter the associated object properties and their IDs on the Object Property Tables at the block S1904. The object properties now under consideration are action script and object marking information, and the IDs are those associated with these properties. There are thus created the aforementioned Action Script Table, which lists correlations between action IDs and action scripts, and the Object Marking Information Table which lists correlations between object marking information IDs and object marking information.
Following the tabulation of the object properties and associated IDs at the block S1904, it is checked at another logical node S1905 whether the unshown buffer contains still-unprocessed object metadata or not. The cycle of the steps S1901-S1905 is repeated for all the object metadata buffered.
The flowchart given in
At the next block S2103 the metadata decoder 110 responds by determining the object that has been specified by the viewer, on the bases of the incoming timestamp and coordinate position. Such determination will be easily accomplished since the metadata decoder 110 decodes the object metadata in step with the progress of video reproduction and knows the object regions at the timestamp when the object region is clicked. If two or more object regions overlap in the clicked position, it is assumed that the view has clicked upon the object that is positioned most in the foreground of the scene. Toward this end the metadata decoder 110 may refer to the hierarchy information on the subsection 500,
The object that has been presumably selected by the viewer having been specified as above, the metadata decoder 110 proceeds to check, at another logical node S2104, if the action script 501,
Upon receipt of the action script 501 the script interpreter 111 interprets and executes the same at the final block S2107. Thus, for example, the video player may either exhibit an HTML file specified or display some video program, resulting in realization of hypermedia. The HTML files or video programs to be specified by the action script 501 may be either prerecorded on the video player 100 or delivered over the Internet or like computer network.
Telop exhibition is also possible in synchronism with the reproduction of a video program as a kind of object marking according to the object marking information.
The display time information 2400 represents the beginning and ending moments of telop display, or the beginning moment and length of telop display. This display time information will be unnecessary if the beginning and end of telop display agree with the beginning and end of the associated object region appearing on the display screen.
The style information 2401 specifies the font, and its size, of the text to be exhibited. The highlight information 2402 indicates those of the telop characters or words which are to be highlighted. The karaoke mode information 2403 dictates the timing of color change from the beginning toward the end of the characters or words exhibited; for example, it may specify three seconds for color change from the first to the tenth characters, two seconds from the eleventh to the twenty-fourth characters, and so forth. The scroll information 2404 indicates the speed and direction of text scroll through the telop zone. The blink information 2405 specifies the words or characters to be blinked as well as the frequency of the blinking. The text length 2406 indicates the number of words or characters to be exhibited on the telop zone. The text 2407 indicates the characters or words to be actually exhibited on the telop zone by the number specified by the text length 2406.
Use of the object metadata including the object marking information of
Let it be assumed that, as drawn in
A further example of object marking is the outlining of text on the display screen in the course of video reproduction by the video player 100, as pictured in
In order to determine the size and position of the outline to be exhibited, the coordinate positions of some preselected points of each outline may be set forth as part of the object region data. In
Thus the present invention succeeds in the provision of hypermedia with the delivery of metadata, from which object property information is omitted wherever possible, from metadata distributor 101 to video player 100. In delivering metadata from metadata distributor 101 to video player 100, for each object metadata unit that is to bear the same action script as the other units, the same action ID is given to the object metadata unit as that of the other units, and the action script is omitted. Similarly, for each object metadata that is to bear the same object marking information as the other units, the same marking information ID 603,
According to the further features of the invention, upon receipt of the metadata, there are created in the video player 100 both Action Script Table, which lists correlations between action IDs and action scripts, and Object Marking Information Table which lists correlations between object marking information IDs and object marking information. For all the object metadata units from which action script or object marking information has been omitted, the missing action script or object marking information is retrieved from these object property tables. Accurate display of object information is therefore possible from the received metadata despite the omission of some object property information therefrom during its delivery from the object distributor.
In the first preferred embodiment set forth hereinbefore with reference to
Both video player and metadata distributor can be the same in terms of hardware as those shown at 100 and 101 in
At the block S3107 of the modified Metadata Decode Program the metadata decoder 110 reconstructs the decoded object metadata, filling up the omissions by the subroutine given in
The answer to the node S3109 may be “no”; that is, no marking of the object region may be specified. No marking image need be formed in this case, so that the routine jumps to the final block S3111, bypassing the block S3110.
Reference is now invited to the flowchart of
The answer “yes” to the node S3202 directs the subroutine to the block S3203, where the previously reconstructed object metadata of the same ID is obtained. That is, if action script has been omitted, the previously reconstructed object metadata of the same action ID is obtained, and if object marking information has been omitted, the previously reconstructed object metadata of the same making information ID 603,
Following the block S3204, another logical node S3205 requires, in short, that the steps S3201-S3204 be performed for all the object metadata that need reconstruction. The required object property information will thus be added to all the object metadata.
At the next block S3304 the metadata decoder 110,
Thus, as in the first describe embodiment of the invention, video-oriented hypermedia is realized with a substantial reduction in the amount of metadata delivered to each client video player. Those object metadata units from which the object property information has been omitted can be accurately reconstructed by the video player for display of any associated information upon clicking of each object. Upon receipt of a packet or packets of metadata including incomplete units, which lack in some object property information such as action script or object marking information, the video player reconstructs the received metadata altogether. This feature of the alternate embodiment is conducive to the swift, accurate exhibition of desired additional information in response to the clicking of an object.
A further reduction of the amount of metadata distributed to each client is accomplished by omitting object region data, in addition to the omission of part or all of the object property information as in the foregoing two embodiments, wherever such omission is possible. Omission of object region data is possible, however, only in cases where the object region represented is immobile. This third embodiment also presupposes use of the client/server architecture such as that of
A typical example of object region permitting omission of object region data according to the teachings of this embodiment is a fixed telop zone seen at 3401 in
In
It will be observed that no change in the position of the telop zone occurs in the plane defined by the axes X and Y, the telop zone itself being assumed to be fixed on the display screen. Consequently, only if the fixed position of the telop zone is fully described in the first object region data unit 3501, the succeeding units 3502-3505 will be unnecessary, these being capable of restoration later in the video player by replication of the first unit.
Thus, in this embodiment of the invention, the object region data indicative of the fixed telop zone is stored only in the first of the series of object metadata units and omitted from the rest Such omission of object region data is possible not only with the fixed telop zone but further with any other object whose position on the display screen does not change with time. Additional curtailment is thus accomplished in the amounts of data that must be transmitted from metadata distributor 101 to video player 100.
Receiving from the metadata distributor 101 the metadata stream from which object region data has been omitted wherever possible as above, the video player 100 is to operate according to the Metadata Decode Program flowcharted in both
A block S3708,
Next comes a logical node S3709 which asks whether, for each incoming object metadata unit, object region data is omitted or not. The answer “yes” to this query directs the program to the block S3710, at which the Object region Table is referred to ascertain the required object region data from the object ID in question. If the answer to the node S3709 is “no”, on the other hand, then the object region data is obtained directly from that object metadata unit, and the block S3710 is bypassed. The subsequent steps of operation, from logical node S3711 to block S3715, are akin to the corresponding steps S1808-S1812 of the Metadata Decode Program of
Following the tabulation of the object region data and associated IDs at the block S1904, it is checked at another logical node S3805 whether the unshown buffer contains still-unprocessed object metadata or not. The cycle of the steps S3801-S3804 is repeated for all the object metadata which is buffered.
The teachings of this third embodiment of the invention are applicable to each series of object metadata units bearing the same object region data, that is, the object region data representative of a fixed telop or like object region set forth with reference to
The preparation of the Object region Table is, however, not an essential feature of the present invention; instead, the object metadata units from which object region data has been omitted may first be reconstructed as in the embodiment of
Contrary to the teachings of all the foregoing embodiments of the invention, it is not an absolute requirement that metadata be stored on the metadata distributor 101 and delivered to the video player 100. A possible alteration of this metadata delivery system, apparent to the specialists of the Internet and allied arts, is to store metadata on the video player 100 itself, for use whenever an associated video program is played.
This embodiment of the invention represents a total departure from the client/server architecture of
The volume/file structure information area is assigned to the universal disk format (UDF) bridge structure. The UDF bridge volume is recognized according to Part Two of ISO/IEC 13346. The space for such volume recognition consists of a consecutive sector, starting from the first logic sector of the volume space. The first hexadecimal logic sector is reserved for use of the system specified in ISO 9660 by International Organization for Standardization. Such volume/file structure information area is needed to assure compatibility with the conventional DVD video standards.
The DVD video area has prerecorded thereon management information known as Video Manager (VMG) and one or more video title sets (VTS#1-VTS#n). The VMG covers all the VTS existing in the DVD video area and includes Video Manager Information (VMGI), optional video object set for VMG menu (VMGM_VOBS), and VMG backup. Each VTS includes VTS information (VTSI), optional video object set for VTS menu (VTSM_VOBS), video object set for titles (e.g. motion pictures) in a VTS (VTSTT_VOBS), and VTSI backup. The DVD video area of such content is also needed for compatibility with the conventional DVD video standards.
The menu for selective playback of the prerecorded titles (VTS#1-VTS#n) and similar features are prescribed by the provider (manufacturer of the DVD in this case) in the VMG The menu for selective playback of chapters in each title (e.g. VTS#1) and the procedure of cell playback are also predefined by the provider in the VTSI. The DVD user has therefore been conventionally able to enjoy the recordings on the disk according to the menu of the VMG/VTSI which has been prepared by the provider or the program chain information (PGCI) included in the VTSI. However, according to the conventional DVD video standards, it has been impossible for the viewer to play back the motion picture, musical performance, or like content of the VTS in a manner different from what the VMG/VTSI dictates.
The Enhanced DVD of
The ENAV contents include sound data, still-picture data, font text data, motion video data, animation data, etc., in addition to ENAV document for controlling the playback of all such data. The ENAV document describes, in markup or script language, the methods of playback (marking method, procedure of playback, procedure of playback switching, selection of the object for playback, etc.) of the ENAV contents and/or DVD video contents. The markup language in use may be either Hypertext Markup Language/Extensible Hypertext Markup Language (HTML/XHTML) or Synchronized Multimedia Integration Language (SMIL), and the script language may be either European Computer Manufacturers Association (ECMA) script or JavaScript. Various language combinations are possible.
The present invention suggests prerecording of object metadata on the Enhanced DVD as part of the ENAV contents. As in all the foregoing embodiments of the invention, the object metadata is designed to enable the viewer access some predefined information on each object appearing in the DVD video program, by clicking on that object on the display screen. The object metadata may be prerecorded on the disk just as it is on the storage medium 113 of the metadata distributor 101 in the first three embodiments of the invention. Also as in the foregoing embodiments, the object property information included in the object metadata may be omitted wherever possible in order to save the disk surface area required for such metadata recording and to make most of the storage capacity of the disk.
Additional advantages and modifications will readily occur to those skilled in the art. Therefore, the invention in its broader aspects is not limited to the specific details and representative embodiments shown and described herein. Accordingly, various modifications may be made without departing from the spirit or scope of the general inventive concept as defined by the appended claims and their equivalents.
Number | Date | Country | Kind |
---|---|---|---|
2004-106709 | Mar 2004 | JP | national |