The present invention relates to a data processor and processing method for writing stream data of a moving picture stream on a storage medium such as an optical disk.
Various types of data streams have been standardized to compress and encode video data at low bit rates. A system stream compliant with the MPEG-2 System standard (ISO/IEC 13818-1) is known as one such data stream. There are three types of system streams, namely, a program stream (PS), a transport stream (TS) and a PES stream.
Recently, another data stream, complying with the MPEG-4 system standard (ISO/IEC 14496-1), has been being defined. In a format compliant with the MPEG-4 system standard, video streams, including MPEG-2 or MPEG-4 video streams, and various types of audio streams are multiplexed together, thereby generating moving picture stream data. Furthermore, according to a format compliant with the MPEG-4 system standard, auxiliary information is defined. The auxiliary information and a moving picture stream are defined as a single file (which is called an “MP4 file”). The data structure of an MP4 file is based on, and an extension of, a QuickTime® file format of Apple Corporation. It should be noted that as for a system stream compliant with the MPEG-2 System standard, no data structure storing the auxiliary information (such as access information, special playback information and recording date) is defined. This is because the auxiliary information is included in a system stream according to the MPEG-2 System standard.
Video data and audio data would often be recorded on a magnetic tape in the past. Recently, however, optical disks such as DVD-RAMs and MOs have attracted much attention as storage media that will soon replace magnetic tapes.
In the moving picture stream 22 of the MP4 file shown in
It should be noted that a data stream with a structure including a moving picture stream and auxiliary information as in the data structure of the MP4 file shown in
It should be noted that the UDF standard corresponds to the installing terms of the ISO/IEC 13346 standard. By connecting a DVD-RAM drive to a computer such as a PC by way of a 1394 interface and a serial bus protocol 2 (SBP-2), the PC can also treat a file that was written in a UDF compliant format as a single file.
By using allocation descriptors, the file entry manages the continuous data areas (CDAs) a, b, c and the data area d where the data is stored. More specifically, if the writing control section 341 finds a defective logical block while writing the MP4 file on the continuous data area a, then the writing control section 341 will skip that defective logical block and continue to write the file from the beginning of the continuous data area b. Next, if the writing control section 341 finds a non-writable PC file storage area while writing the MP4 file on the continuous data area b, then the writing control section 341 will resume writing the file from the beginning of the continuous data area c. On having written the file, the writing control section 341 will write the auxiliary information 21 on the data area d. As a result, the file VR_MOVIE.VRO is made up of the continuous data areas d, a, b and c.
As shown in
In playing back an MP4 file, the data processor 350 retrieves and receives a moving picture stream by way of the pickup 330 and a reading section 321 and gets the stream decoded by an MPEG-2 decoding section 311, thereby generating a video signal and an audio signal, which are eventually output through a video signal output section 310 and an audio signal output section 312, respectively. Reading the data from the DVD-RAM disk and outputting the read data to the MPEG-2 decoding section 311 are carried out concurrently. In this case, the data read rate is set higher than the data output rate and is controlled such that the data to be played back does not run short. Accordingly, if the data is continuously read and output, then extra data can be obtained by the difference between the data read rate and the data output rate. By using that extra data as the data to be output while data reading is discontinued by pickup's jump, continuous playback is realized.
Specifically, supposing the rate of reading the data from the DVD-RAM disk 331 is 11 Mbps, the maximum rate of outputting the data to the MPEG-2 decoding section 311 is 8 Mbps and the longest time it takes to move the pickup is 3 seconds, data of 24 megabits, which corresponds with the amount of data to be output to the MPEG-2 decoding section 311 while the pickup is moving, is needed as the extra output data. To secure this amount of data, the data needs to be read for eight seconds on end. That is to say, the continuous reading needs to last for the amount of time that is obtained by dividing 24 megabits by the difference between the data read rate of 11 Mbps and the data output rate of 8 Mbps.
Accordingly, while the continuous reading is carried out for eight seconds, data of 88 megabits, which should be output in eleven seconds, is read out. Thus, if a continuous data area with a size corresponding to at least eleven seconds is secured, then continuous data playback can be guaranteed.
It should be noted that several defective logical blocks may be included within the continuous data area. In that case, however, the continuous data area needs to have a size corresponding to an amount of time that is slightly longer than eleven seconds with the expected read time, which it will take to read those defective logical blocks during the playback operation, taken into account.
In performing the process of deleting a stored MP4 file, the writing control section 341 performs predetermined deletion processing by controlling the writing section 320 and reading section 321. In the MP4 file, the auxiliary information includes presentation timings (i.e., time stamps) of all frames. Accordingly, in partially deleting an intermediate portion of a moving picture stream, only the time stamps in the auxiliary information need to be deleted. It should be noted that in an MPEG-2 system stream, the moving picture stream should be analyzed to ensure continuity even at the partially deleted portion. This is because the time stamps are dispersed over the stream.
The MP4 file format is characterized by storing the video frames or audio frames of a video/audio stream as a single set without dividing each frame. At the same time, the MP4 file format defines access information, which enables random access to any arbitrary frames, for the first time ever for any international standard defined so far. The access information is defined on a frame-by-frame basis, and may include the frame size, frame period, and address information for a frame. More specifically, the access information is stored for every unit (e.g., every display period of 1/30 second for a video frame and every 1,536 samples for an audio frame (in AC-3 audio, for example)). Accordingly, if the presentation timing of a video frame needs to be changed, only the access information thereof should be changed, and the video/audio stream does not always have to be changed. The access information of that type has a data size of about 1 megabyte per hour.
As to the data size of the access information, according to Non-Patent Document No. 1, the access information compliant with the DVD Video recording standard needs to have a data size of 70 kilobytes per hour. The data size of the access information as defined by the DVD Video recording standard is less than one-tenth of that of the access information included in the auxiliary information of an MP4 file.
Also, the optical disk drive disclosed in Patent Document No. 1 not only writes video frames on a GOP basis, not on a frame basis, but also writes each audio frame continuously for a period of time corresponding to one GOP. The optical disk drive defines the access information on a GOP basis, too, thereby cutting down the required data size of the access information.
Furthermore, the MP4 file describes a moving picture stream in accordance with the MPEG-2 Video standard but is not compatible with a system stream as defined by the MPEG-2 System standard. Thus, the MP4 file cannot be edited by utilizing the moving picture editing capability of any application used extensively today on PCs, for example. This is because the editing capability of a lot of applications is targeted on a moving picture stream compliant with the MPEG-2 System standard. Furthermore, the MP4 file standard defines no decoder model to ensure playback compatibility for a moving picture stream portion. Then, absolutely no piece of software or hardware compliant with the MPEG-2 System standard, which has circulated very widely today, can be used at all.
Meanwhile, a play list function for picking and combining together preferred playback ranges of a moving picture file to make a single piece of work has been realized. This play list function is normally carried out as a virtual editing process without directly editing any moving picture file recorded. A play list is made up of MP4 files by newly generating a Movie Atom. In making a play list up of MP4 files, if multiple playback ranges have the same stream attribute, then the same Sample Description Entry is used and the redundancy of Sample Description Entry can be reduced. However, when a seamless play list that guarantees seamless playback is described by making use of this feature, it is difficult to describe the stream attribute information on a playback range basis.
An object of the present invention is to provide a data structure, of which the access information has a small data size and which can be used even in an application designed for a conventional format, and also provide a data processor that can perform processing based on such a data structure.
Another object of the present invention is to realize an editing process of combining video and audio seamlessly in a format compatible with a conventional stream that should have audio gaps. A more specific object thereof is to get such editing done on video and audio that is described as an MP4 stream. An additional object thereof is to combine audio naturally at every connection point.
Yet another object of the present invention is to realize an editing process that combines a plurality of contents together such that the user can specify his or her desired audio connection form (e.g., whether or not the audio should fade).
A data processor according to the present invention includes: a writing section for arranging a plurality of moving picture streams, each including video and audio to play back synchronously with each other, and writing the streams as at least one data file on a storage medium; and a writing control section for locating a mute interval between two moving picture streams that are going to be played back continuously. The writing control section provides additional audio data representing audio to be reproduced in the mute interval located, and the writing section stores the provided additional audio data on the storage medium such that the additional audio data is associated with the data file.
The writing control section may further use audio data, which is stored in a predetermined terminal range of one of the two continuously played moving picture streams that is going to be played earlier than the other, and may provide the additional audio data including the same audio as that stored in the predetermined terminal range.
Alternatively, the writing control section may further use audio data, which is stored in a predetermined terminal range of one of the two continuously played moving picture streams that is going to be played later than the other, and may provide the additional audio data including the same audio as that stored in the predetermined terminal range.
The writing section may write the provided additional audio data just before where the mute interval is stored, thereby associating the additional audio data with the data file.
The writing section may write the arranged moving picture streams as a single data file on the storage medium.
Alternatively, the writing section may write the arranged moving picture streams as multiple data files on the storage medium.
The writing section may write the provided additional audio data just before where one of the two continuously played moving picture stream data files, which is going to be played later than the other, is stored, thereby associating the additional audio data with the data file.
The writing section may write information about the arrangement of the moving picture streams as at least one data file on the storage medium.
The mute interval may be shorter than the time length of a single audio decoding unit.
A video stream in each said moving picture stream may be an MPEG-2 video stream, and the same MPEG-2 video stream buffer conditions may have to be met by the two continuously played moving picture streams.
The writing section may further write information for controlling an audio level before and after the mute interval on the storage medium.
The writing section may write the moving picture streams in a physically continuous data area on the storage medium on the basis of either a predetermined playback duration or a predetermined data size, and may also write the additional audio data just before the continuous data area.
A data processing method according to the present invention includes the steps of: writing an arrangement of a plurality of moving picture streams, each including video and audio to play back synchronously with each other, as at least one data file on a storage medium; and controlling writing by locating a mute interval between two moving picture streams that are going to be played back continuously. The step of controlling writing includes providing additional audio data representing audio to be reproduced in the mute interval located, and the step of writing includes associating the provided additional audio data with the data file and storing the additional audio data on the storage medium.
The step of controlling writing may include further using audio data, which is stored in a predetermined terminal range of one of the two continuously played moving picture streams that is going to be played earlier than the other, and providing the additional audio data including the same audio as that stored in the predetermined terminal range.
Alternatively, the step of controlling writing may include further using audio data, which is stored in a predetermined terminal range of one of the two continuously played moving picture streams that is going to be played later than the other, and providing the additional audio data including the same audio as that stored in the predetermined terminal range.
The step of writing may include writing the provided additional audio data just before where the mute interval is stored, thereby associating the additional audio data with the data file.
The step of writing may include writing the arranged moving picture streams as a single data file on the storage medium.
Alternatively, the step of writing may include writing the arranged moving picture streams as multiple data files on the storage medium.
The step of writing may include writing the provided additional audio data just before where one of the two continuously played moving picture stream data files, which is going to be played later than the other, is stored, thereby associating the additional audio data with the data file.
The step of writing may include writing information about the arrangement of the moving picture streams as at least one data file on the storage medium.
Another data processor according to the present invention includes: a reading section for reading at least one data file, including a plurality of moving picture streams with video and audio to be played back synchronously with each other, and additional audio data associated with the at least one data file from a storage medium; a reading control section for controlling reading by generating a control signal based on time information that is added to the moving picture streams to play back the video and audio synchronously with each other; and a decoding section for decoding the moving picture streams in response to the control signal, thereby outputting signals representing the video and audio. When two moving picture streams are played back continuously by using the data processor, the reading control section outputs a control signal instructing that audio represented by the additional audio data be output after one of two moving picture streams has been played back and before the other moving picture stream is played back.
Another data processing method according to the present invention includes the steps of: reading at least one data file, including a plurality of moving picture streams with video and audio to be played back synchronously with each other, and additional audio data associated with the at least one data file from a storage medium; generating a control signal based on time information that is added to the moving picture streams to play back the video and audio synchronously with each other; and decoding the moving picture streams in response to the control signal, thereby outputting signals representing the video and audio. When two moving picture streams are played back continuously, the step of generating the control signal includes outputting a control signal instructing that audio represented by the additional audio data be output after one of two moving picture streams has been played back and before the other moving picture stream is played back.
A computer program according to the present invention makes a computer function as a data processor that performs the following processing steps when read and executed by the computer. Specifically, by executing the computer program, the data processor performs the processing steps of: acquiring a plurality of moving picture streams, each including video and audio to play back synchronously with each other, and writing the streams as at least one data file on a storage medium; and controlling writing by locating a mute interval between two moving picture streams that are going to be played back continuously. The step of controlling writing includes providing additional audio data representing audio to be reproduced in the mute interval located, and the step of writing includes associating the provided additional audio data with the data file and storing the additional audio data on the storage medium.
The computer program may be stored on a storage medium.
Another data processor according to the present invention writes a plurality of encoded data, compliant with the MPEG-2 System standard, as a single data file such that audio data of a predetermined length is associated with the data file.
Still another data processor according to the present invention reads a data file, including a plurality of encoded data compliant with the MPEG-2 System standard, and audio data associated with the data file. In reading the encoded data, the data processor reads the audio data associated with the data file in a mute interval of the encoded data.
Hereinafter, preferred embodiments of the present invention will be described with reference to the accompanying drawings.
The portable videocorder 10-1 receives a broadcast program via its attached antenna and compresses the moving pictures of the broadcast program, thereby generating an MP4 stream. The camcorder 10-2 records not only video but also its accompanying audio, thereby generating another MP4 stream. In an MP4 stream, the video and audio data are encoded by a predetermined compression coding method and are stored in accordance with the data structure described herein. The portable videocorder 10-1 and camcorder 10-2 either store the generated MP-4 streams on a storage medium 131 such as a DVD-RAM or output the streams through a digital interface such as an IEEE 1394 or USB port. It should be noted that the portable videocorder 10-1 and camcorder 10-2 needs to have even smaller sizes. Thus, the storage medium 131 does not have to be an optical disk with a diameter of 8 cm but may be an optical disk with a smaller diameter, for example.
The PC 10-3 receives the MP4 streams by way of either the storage medium or a transmission medium. If the respective appliances are connected together through a digital interface, then the PC 10-3 can receive the MP4 streams from the respective appliances by controlling the camcorder 10-2 and so on as external storage devices.
If the PC 10-3 has application software or hardware that can cope with the MP4 stream processing of the present invention, then the PC 10-3 can play back the MP4 streams just as defined by the MP4 file standard. On the other hand, if the PC 10-3 cannot cope with the MP4 stream processing of the present invention, then the PC 10-3 can play back the moving picture streams in accordance with the MPEG-2 System standard. It should be noted that the PC 10-3 can also perform editing processing such as partial deletion on the MP4 streams. In the following description, the portable videocorder 10-1, camcorder 10-2 and PC 10-3 shown in
First, the MP4 stream writing function of the data processor 10 will be described. The data processor 10 includes a video signal input section 100, an MPEG2-PS compressing section 101, an audio signal input section 102, an auxiliary information generating section 103, a writing section 120, an optical pickup 130 and a writing control section 141 as respective components regarding this function.
The video signal input section 100 is implemented as a video signal input terminal and receives a video signal representing video data. The audio signal input section 102 is implemented as an audio signal input terminal and receives an audio signal representing audio data. For example, the video signal input section 100 and audio signal input section 102 of the portable videocorder 10-1 (see
The MPEG2-PS compressing section (which will be simply referred to herein as a “compressing section”) 101 receives the video and audio signals, thereby generating an MPEG-2 program stream (which will be referred to herein as an “MPEG2-PS”) compliant with the MPEG-2 System standard. The MPEG2-PS generated may be decoded by itself in accordance with the MPEG-2 System standard. The MPEG2-PS will be described in further detail later.
The auxiliary information generating section 103 generates auxiliary information for the MP4 stream. The auxiliary information includes reference information and attribute information. The reference information is used to identify the MPEG2-PS that has been generated by the compressing section 101 and may include the file name of the MPEG2-PS being written and its storage location on the DVD-RAM disk 131. On the other hand, the attribute information describes the attributes of a sample unit of the MPEG2-PS. As used herein, the “sample” refers to the minimum management unit in a sample description atom (to be described later) as in the auxiliary information defined by the MP4 file standard. The attribute information includes data size, playback time and so on for each sample. One sample may be a data unit that can be accessed at random, for example. In other words, the attribute information is needed to read the sample. Among other things, the sample description atom (to be described later) is sometimes called “access information”.
Specific examples of the attribute information include the address of the data storage location, a time stamp representing playback timing, an encoding bit rate, and information about codec. The attribute information is provided for each of the video data and the audio data in every sample. Except for the field description to be mentioned explicitly soon, the attribute information complies with the contents of auxiliary information for a conventional MP4 stream 20.
As will be described later, one sample according to the present invention is a single video object unit (VOBU) in the MPEG2-PS. It should be noted that “VOBU” refers to the video object unit as defined by the DVD Video recording standard. The auxiliary information will be described in further detail later.
In accordance with the instruction given by the writing control section 141, the writing section 120 controls the pickup 130, thereby writing data at a particular location (i.e., address) on the DVD-RAM disk 131. More specifically, the writing section 120 writes the MPEG2-PS, generated by the compressing section 101, and the auxiliary information, generated by the auxiliary information generating section 103, on the DVD-RAM disk 131 as respectively different files.
The data processor 10 further includes a continuous data area detecting section (which will be simply referred to herein as a “detecting section”) 140 and a logical block management section (which will be simply referred to herein as a “management section”) 143 that operate during the data write operation. In accordance with the instruction given by the writing control section 141, the continuous data area detecting section 140 checks the availability of sectors, which are managed by the logical block management section 143, thereby detecting a physically continuous unused area available. The writing control section 141 instructs the writing section 120 to write the data on that unused area. A specific data writing method may be similar to that already described with reference to
Hereinafter, the data structure of the MP4 stream will be described with reference to
The auxiliary information 13 includes reference information dref to make reference to the MPEG2-PS 14 and further includes attribute information that describes the attributes of each video object unit (VOBU) of the MPEG2-PS 14. Since the attribute information describes the attributes of each VOBU, the data processor 10 can find a VOBU at any arbitrary location in the MPEG2-PS 14 on a VOBU basis and can perform playback and editing processing thereon.
The MPEG2-PS 14 is a moving picture stream, which is compliant with the MPEG-2 System standard and which is made up of video and audio packs that are interleaved together. Each video pack includes a pack header and encoded video data, while each audio pack includes a pack header and encoded audio data. In the MPEG2-PS 14, the data is managed on a video object unit (VOBU) basis, where a VOB includes moving picture data, each unit of which has a length corresponding to a video playback duration of 0.4 to 1 second. The moving picture data includes a plurality of video packs and a plurality of audio packs. By reference to the information described in the auxiliary information 13, the data processor 10 can locate and read any arbitrary VOBU. It should be noted that each VOBU includes at least one GOP.
The MP4 stream 12 of the present invention is partly characterized in that the MPEG2-PS 14 can be decoded not only by reference to the attribute information 13, which complies with the data structure of an MP4 stream as defined by the MPEG-4 system standard, but also in accordance with the MPEG-2 System standard. The auxiliary information file and the PS file are stored separately, and therefore, the data processor 10 can analyze and process them independently of each other. For example, an MP4 stream player, which can carry out the data processing of the present invention, can adjust the playback duration of the MP4 stream 12 according to the attribute information 13, sense the encoding method of the MPEG2-PS 14 and decode it by its associated decoding method. On the other hand, a conventional apparatus that can decode an MPEG2-PS may decode it in accordance with the MPEG-2 System standard. Thus, even any currently popular version of software or hardware, which is compliant with only the MPEG-2 System standard, can also play back a moving picture stream included in the MP4 stream.
Optionally, not only the sample description atom on the VOBU basis but also another sample description atom, which uses a number of frames of the audio data of the MPEG2-PS 14, corresponding to a predetermined amount of time, as a management unit, may be provided as shown in
Next, the data structure of each video object unit (VOBU) of the MPEG2-PS 14 will be described.
Each of the video and audio packs included in the VOBU of the MPEG2-PS 14 is composed of the data included in its associated elementary stream (Video) or (Audio) so as to have a data size of 2 kilobytes. As described above, each pack is provided with a pack header.
It should be noted that if another elementary stream (not shown) is provided for auxiliary video data such as subtitle data, then each VOBU of the MPEG2-PS 14 further includes packs of that auxiliary video data.
Next, the data structure of the auxiliary information 13 in the MP4 stream 12 will be described with reference to FIGS. 15 and 16.
According to the present invention, the attributes of a sample unit are described by using “Data Reference Atom (drf)” 15 and “Sample Table Atom (stbl)” 16, in particular. As described above, one sample corresponds to one video object unit (VOBU) of the MPEG2-PS. “Sample Table Atom” 16 includes the six low-order atoms shown in
In “Sample Description Atom” 17 shown in
The entry 18 includes that sample's “Presentation Start Time” and “Presentation End Time”, which store the timing information of the first video frame and the timing information of the last video frame, respectively. The entry 18 further includes the attribute information (“Video ES Attribute”) of the video stream within the sample and the attribute information (“Audio ES Attribute”) of the audio stream within the same sample. As shown in
The entry 18 further includes a discontinuity point start flag and seamless information. These pieces of information are described if there are a number of PS streams in a single MP4 stream 12 as will be described later. For example, a discontinuity point start flag of “0” indicates that the previous moving picture stream and the current moving picture stream are a completely continuous program stream. On the other hand, a discontinuity point start flag of “1” shows that those moving picture streams are discontinuous program streams. And if those streams are discontinuous, the seamless information may be described in order to play back a moving picture or audio without a break even at a discontinuity point of the moving picture or audio. The seamless information includes audio discontinuity information and SCR discontinuity information during the playback. The audio discontinuity information includes the presence or absence of a mute interval (i.e., the audio gap shown in
By providing the discontinuity point start flag, the switching point of Sample Description Entries and the continuity switching point of moving picture streams can be defined independently of each other. As shown in
The data structures of the auxiliary information 13 and MPEG2-PS 14 of the MP4 stream 12 shown in
Hereinafter, it will be described with reference to
Then, in Step 213, the writing section 120 determines the file name and storage location of the MPEG2-PS to be written on the DVD-RAM disk 131. Next, in Step 214, the auxiliary information generating section 103 acquires the file name and storage location of the PS file and specifies the contents to be described as the reference information (i.e., Data Reference Atom shown in
Subsequently, in Step 215, the auxiliary information generating section 103 acquires data representing the playback duration, data size and so on for each of the VOBUs defined in the MPEG2-PS 14 and specifies the contents to be described as the attribute information (i.e., Sample Table Atom shown in
Thereafter, in Step 216, the auxiliary information generating section 103 generates the auxiliary information based on the reference information (i.e., Data Reference Atom) and the attribute information (i.e., Sample Table Atom).
Next, in Step 217, the writing section 120 outputs the auxiliary information 13 and MPEG2-PS 14 as the MP4 stream 12 and writes them on the DVD-RAM disk 131 as an auxiliary information file and a PS file, respectively. By performing this procedure, the MP4 stream is generated and written on the DVD-RAM disk 131.
Hereinafter, the MP4 stream reading function of the data processor 10 will be described with reference to
First, in accordance with an instruction given by the reading control section 142, the reading section 121 controls the pickup 130 so as to read the MP4 file from the DVD-RAM disk 131 and acquire the auxiliary information 13. The reading section 121 outputs the acquired auxiliary information 13 to the reading control section 142. Also, in response to a control signal supplied from the reading control section 142 to be described later, the reading section 121 reads the PS file from the DVD-RAM disk 131. The control signal is a signal designating the PS file to read (“MOV001.MPG”).
The reading control section 142 receives the auxiliary information 13 from the reading section 121 and analyzes its data structure, thereby acquiring the reference information 15 (see
The MPEG2-PS decoding section 111 receives the MPEG2-PS 14 and the auxiliary information 13 and decodes the MPEG2-PS 14 into video data and audio data in accordance with the attribute information contained in the auxiliary information 13. More specifically, the MPEG2-PS decoding section 111 reads the data format (“data_format”), the video stream attribute information (“video ES attribute”) and the audio stream attribute information (“audio ES attribute”) of Sample Description Atom 17 (see
The video signal output section 110 is implemented as a video signal output terminal to output the decoded video data as a video signal, while the audio signal output section 112 is implemented as an audio signal output terminal to output the decoded audio data as an audio signal.
The MP4 stream reading process by the data processor 10 begins by reading a file with an extension “MP4” (i.e., “MOV001.MP4”) as in the conventional process of reading an MP4 stream file. More specifically, this process may be carried out in the following manner. First, the reading section 121 reads the auxiliary information file (“MOV001.MP4”). Next, the reading control section 142 analyzes the auxiliary information 13, thereby extracting the reference information (i.e., Data Reference Atom). Then, in accordance with the reference information extracted, the reading control section 142 outputs a control signal instructing that the PS file, making up the same MP4 stream, be read. In this preferred embodiment, the control signal output by the reading control section 142 instructs that the PS file (“MOV001.MPG”) be read.
Next, in response to the control signal, the reading section 121 reads the designated PS file. Thereafter, the MPEG2-PS decoding section 111 receives the MPEG2-PS 14 and auxiliary information 13 contained in the data file read and analyzes the auxiliary information 13, thereby extracting the attribute information. Then, by reference to Sample Description Atom 17 (see
It should be noted that any conventional player or playback software that can read a stream compliant with the MPEG-2 System standard can read the MPEG2-PS 14 just by reading the PS file. In that case, however, the player does not have to be able to read the MP4 stream 12. Since the MP4 stream 12 is made up of the auxiliary information 13 and MPEG2-PS 14 as two separate files, the PS file in which the MPEG2-PS 14 is stored can be easily identified by the extension, for example, and read.
In
Hereinafter, a modified example of the PS file to make up the MP4 stream 12 will be described.
In the auxiliary information file, reference information (dref, see
In the auxiliary information file, reference information (dref, see
A procedure of reading two PS files, storing mutually discontinuous MPEG2-PS's, by providing two pieces of reference information for the files has been described with reference to
Suppose the data of PS #1 is processed first, and then that of PS #3 is processed. The DTS video frame on the second row and the PTS video frame on the third row represent time stamps of a video frame. As can be seen from these time stamps, PS files #1 and #3 can be played back without discontinuing the video. As to an audio frame, however, there is a mute interval, in which no data is present for a certain period of time, after PS #1 has been played and before PS #3 starts being played. With such an interval left, seamless playback could not be achieved.
Thus, PS #2 is newly provided and a PS file, including an audio frame for the purpose of seamless connection, is provided such that the auxiliary information file can make reference to that file. This audio frame includes audio data to fill the mute interval. For example, the audio data that was written synchronously with the end of the moving picture of PS #1 is copied. As can be seen from the audio frame row shown in
In
In the moving picture streams PS #1 and PS #3, the video stream thereof preferably satisfies the VBV buffer condition of the MPEG-2 Video standard continuously before and after its connection point. This is because if the buffer condition is satisfied, no underflow will occur in the video buffer in the MPEG2-PS decoding section, and therefore, the reading control section 142 and MPEG2-PS decoding section 111 can easily play back the video seamlessly.
By performing these processing steps, even a number of discontinuous PS files can be read and decoded continuously with no time gap left.
In the example shown in
In the preferred embodiments described above, the moving picture stream is supposed to be an MPEG-2 program stream. Alternatively, the moving picture stream may also be an MPEG-2 transport stream (which will be referred to herein as an “MPEG2-TS”) as defined by the MPEG-2 System standard.
As in the MP4 stream 12 shown in
A time stamp is added to the MPEG2-TS 14. More specifically, in this MPEG2-TS 14, a time stamp of 4 bytes to be referred to at the time of transmission is additionally provided before a transport packet (which will be referred to herein as a “TS packet”) of 188 bytes. Accordingly, a TS packet containing video (V_TSP) and a TS packet containing audio (A_TSP) are each made up of 192 bytes. It should be noted that the time stamp may be provided behind the TS packet.
In the MP4 stream 12 shown in
Alternatively, one frame may be handled as one sample, and a plurality of frames may be treated as one chunk.
Even if the data structure shown in
Next, the file structure of another data format, to which the data processing described above is also applicable, will be described with reference to
In the MTF 32, if two adjacent MPEG2-PS's 14 are one continuous program stream, then a single piece of reference information may be provided for the continuous range, thereby making up one MP4 stream. On the other hand, if two adjacent MPEG2-PS's 14 are a discontinuous program stream, then the data address of the discontinuity point may be included in the attribute information as shown in
It has been described how to handle an MPEG-2 system stream by extending the MP4 file format that was standardized in 2001. Alternatively, according to the present invention, the MPEG-2 system stream may also be handled even by extending the QuickTime file format or the ISO Base Media file format, too. This is because most of the specifications of the MP4 file format and the ISO Base Media file format are defined based on, and have the same contents as, the QuickTime file format.
Hereinafter, audio processing for realizing a seamless playback will be described. First, a conventional seamless playback will be described with reference to
It should be noted that if the moving picture file has a longer playback duration, then there will be multiple sets of post recording areas and moving picture stream areas of a predetermined length. If these sets are written continuously on a DVD-RAM disk, then the moving picture file is stored so as to be interleaved with those post recording areas. This format is adopted to make the data stored in any of those post recording areas easily accessible in a short time even while the moving picture file is being accessed.
Also, the video streams in the moving picture file are supposed to satisfy the VBV buffer conditions as defined by the MPEG-2 Video standard continuously before and after the connection point between PS #1 and PS #3 (as well as the connection conditions for realizing a seamless playback at a connection point between two streams as defined by the DVD-VR standard).
To avoid such audio discontinuity, a fade-out, fade-in measure may be taken before and after the audio gap. That is to say, by applying the fade-out and fade-in in 10 ms each before and after the audio gap of a seamless playback, noise to be caused by the sudden audio stoppage can be eliminated and the audio can be heard more natural. However, if the fade-out and fade-in were applied every time the audio gap is produced, then stabilized audio level could not be provided depending on the type of the audio material in question, and good audiovisual state could not be maintained anymore. That is why it is sometimes necessary to eliminate the mute range caused by the audio gap during the playback.
For that purpose, the following measure is taken in this preferred embodiment.
To make such a data arrangement, the writing control section 141 realizes a seamlessly reproducible data structure, which allows an audio gap, for the data that is located around the connection point between the moving picture streams PS #1 and PS #3 that should be connected seamlessly together. At this point in time, it is known whether or not there is any no-data interval (i.e., a mute interval) corresponding to one audio frame or less (i.e., whether or not there is an audio gap), and an audio frame including the audio data to be lost in that audio gap interval and the length of the audio gap interval are also known. The audio gap is produced in almost all cases. Next, the writing control section 141 transmits the data of the audio that should be reproduced in that audio gap interval to the writing section 120 and makes the writing section 120 store it as an audio file and associate it with the moving picture file. As used herein, “to associate” means providing a post recording data area just before the moving picture file is stored and storing additional audio data in that data area. It also means associating that moving picture file and a file storing the audio data with a moving picture track and an audio track, respectively, in the auxiliary information (Movie Atom). This audio data may be audio frame data compliant with the AC3 format, for example.
As a result, the moving picture data file shown in
On the other hand, portion (b) of
The mute interval caused by the audio gap can be eliminated by any of these playback processes. In any of the examples shown in portions (a) and (b) of
It should be noted that if two moving picture streams are connected together using the play list shown in
The bridge file is a moving picture file MOVE0002.MPG including a connecting portion of PS #1 and PS #3. The video streams in the two moving picture streams are supposed to satisfy the VBV buffer conditions of the MPEG-2 Video standard before and after this connecting portion. That is to say, the data structure is supposed to be the same as that shown in
Each of these moving picture files has a predetermined duration (of 10 seconds to 20 seconds, for example) as in
In Edit List Table, track_duration specifies the video duration of the playback range and media_time specifies the location of the playback range in the moving picture file. As the location of this playback range, the location of the top video of the playback range is represented as a time offset value, which is defined by regarding the top of the moving picture file as time zero. Media_time=−1 means a pause interval in which nothing is played back during track_duration. As media_rate, 1.0, meaning 1× playback, is set. The reading section reads Edit List Atom from both the PS track and the audio track alike, thereby carrying out a playback control based on it.
In this manner, a play list for realizing seamless playback of video and audio can be provided such that its form is compatible with a conventional stream that is supposed to have an audio gap. That is to say, either a seamless playback using the audio gap or a seamless playback using the overlapping audio frame may be selected arbitrarily. Consequently, even an apparatus that cope with only the conventional audio gap can perform the seamless playback at the stream connection point at least by the conventional method.
In addition, the connection point can be controlled finely according to the contents of the audio.
Besides, Sample Description Entry, which cuts down the redundancy of an MP4 file play list and which can provide detailed description required for a seamless play list, is realized.
According to the present invention, the seamless playback of video and audio is realized by recording the overlapping portion of the audio. Alternatively, the video and audio may be played back pseudo-seamlessly by skipping the playback of a video frame without using that overlapping portion.
Also, in the preferred embodiment described above, the overlapping portion of the audio is recorded in the post recording area. Alternatively, the overlapping portion may also be stored in Movie Data Atom in the play list file. In AC3, for example, a single frame may have a data size of several kilobytes. Optionally, instead of the STC continuity flag shown in
In the present invention, only the audio frame including the audio overlapping portion is recorded in the post recording area separately from the portion of PS #3. Alternatively, both the extended portion shown in
The audio control information is included in the seamless information of a PS track. Optionally, the audio control information may also be included in the seamless information of an audio track as well. Even so, the fade-out and fade-in are also controlled in a similar manner just before and right after the connection point.
A method of playing back an audio frame continuously before and after a connection point without applying fade-out or fade-in to the connection point was mentioned. This is a technique effectively applicable to AC-3, MPEG Audio Layer 2 and other compression methods.
In the preferred embodiments of the present invention described above, the MPEG2-PS 14 shown in
In the preferred embodiments of the present invention described above, the overlapping audio is supposed to be recorded in the post recording area. Alternatively, the overlapping audio may also be recorded elsewhere but its storage location is preferably as physically close to the moving picture file as possible.
The audio file is supposed to be made up of AC-3 audio frames. Optionally, the audio file may also be stored in either an MPEG-2 program stream or in an MPEG-2 transport stream.
In the data processor 10 shown in
The data processor 10 performs the processing of generating, writing and reading a data stream according to a computer program. For example, the processing of generating and writing the data stream may be carried out by executing a computer program that is described based on the flowchart shown in
The file system is supposed to be compliant with UDF but may also be compliant with FAT, NTFS or any other standard. The video is supposed to be an MPEG-2 video stream but may also be an MPEG-4 AVC, for example. Also, the audio is supposed to be compliant with AC-3 but may also be compliant with LPCM, MPEG-Audio or any other appropriate standard. Furthermore, the moving picture stream is supposed to have a data structure of an MPEG-2 program stream, for example, but may also be any other type of data stream as long as the video and audio are multiplexed together.
According to the present invention, while the data structure of auxiliary information is adapted to the up-to-date standard so as to comply with the ISO standard, a data structure for a data stream, having a format equivalent to a conventional one, and a data processor, operating on such a data structure, are provided as well. Since the data stream is compatible with the conventional format, any existent application can use the data stream. Consequently, every piece of existent software and hardware can be used effectively. In addition, the present invention also provides a data processor that can play back not just video but also audio without a break at all when two moving picture streams are combined together by editing. Furthermore, since the data processor is still compatible with the conventional data stream, compatibility with existent playback equipment is guaranteed, too.
Number | Date | Country | Kind |
---|---|---|---|
2003-289769 | Aug 2003 | JP | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/JP04/11678 | 8/6/2004 | WO | 2/3/2006 |