In systems for editing and producing video and audio, it is often useful to carry various kinds of data that are adjunct to the audio and video signal data. Some of these data are associated with a video or audio sequence as a whole, or associated with a segment, and are not synchronous. Examples include descriptive metadata, such as information about the people involved in the production, date and location, as well as structural metadata, such as image format/type (e.g., HD, SD) and bit rate. However an increasing range of information in the form of data streams that are time-synchronized with the audio and video content is also being generated. This information is distinct from the audio and video portions of the signal, and represents additional information pertaining to the production and/or subject matter of the material.
In existing media editing and production systems, time-synchronous adjunct data are usually embedded within the audio or video signal. In digital video broadcasts, the data is often placed in the ancillary portion of the signal. In file-based media, such as those generated by cameras, disk recorders, video servers, and other video or audio acquisition and playback devices, the data is present in various forms within the media files, often as embedded portions of the video frames.
Such schemes for handling time-synchronous adjunct data streams limit the ability of editors to edit or otherwise manipulate the data streams independently of the picture and audio tracks.
In general, the methods, systems, and computer program products described herein enable time-synchronous data to be represented as distinct entities that are independently editable from the program picture and sound. They further enable the time-synchronous data to be acquired and converted into a flexible format, extracted into indexed asset management systems for analysis, aggregation, and search, and to be accurately created, edited, and played back in sync with picture and sound.
In general, in one aspect, a method of editing a time-based media composition using a media editing software application, wherein the composition includes at least one media track and at least one data track, includes: displaying a timeline representing a current state of the media composition, wherein the timeline includes the at least one media track and the at least one data track, wherein the data track references at least one data stream comprising a stream of data packets, wherein each of the data packets occupies a unique time address corresponding to a media element on the media track; enabling a user of the media editing software application to perform a first editing operation to the time-based media composition, wherein the first editing operation is applied to both the at least one media track and the at least one data track, thereby maintaining temporal synchrony between the at least one media track and the at least one data track; and enabling the user of the media editing software application to perform a second editing operation to the media composition, wherein the second editing operation is applied to only one of the at least one media track and the at least one data track, thereby altering a temporal relationship between at least a portion of the at least one media track and the at least one data track.
Various embodiments include one or more of the following features. At least one of the first and second editing operations includes moving at least one boundary of a segment of the composition. At least one of the first and second editing operations includes altering a temporal location within the composition of a segment of the composition. The time-based media composition is a video composition. The media element is a video frame. The time-based media composition is an audio composition. One of the data tracks references text data. The text data is one of the set consisting of closed captions, open captions, subtitles, and teletext. One of the data tracks references film metadata and/or camera data. The method further enables a user to combine data referenced by at least one data track and at least one media track into a combination media stream, and to output the combination media stream. The at least one data track includes a plurality of data streams, and the method further enables a user to select and output a subset of the plurality of data streams. Further enabling a user to convert data referenced by the at least one data track into a form that is searchable by a media asset management system, and storing the converted data in a database.
In general, in another aspect, a computer program product includes: a computer-readable medium with computer program instructions encoded thereon, wherein the computer program instructions, when processed by a computer, instruct the computer to perform a method of editing a time-based media composition, wherein the composition includes at least one media track and at least one data track, the method comprising: displaying a timeline representing a current state of the media composition, wherein the timeline includes the at least one media track and the at least one data track, wherein the data track references at least one data stream comprising a stream of data packets, wherein each of the data packets occupies a unique time address corresponding to a media element on the media track; enabling a user of the computer to perform an editing operation to the time-based media composition, wherein the editing operation is applied to both the at least one media track and the at least one data track, thereby maintaining temporal synchrony between the at least one media track and the at least one data track; and enabling the user of the computer to perform an editing operation to the media composition, wherein the editing operation is applied to only one of the at least one media track and the at least one data track, thereby altering a temporal relationship between at least a portion of the at least one media track and the at least one data track.
In general, in a further aspect, a system for editing a time-based media composition, the system including: a memory for storing computer-readable instructions; and a processor connected to the memory, wherein the processor, when executing the computer-readable instructions, causes the media composition system to perform a method of editing a media composition, wherein the composition includes at least one media track and at least one data track, the method comprising: displaying a timeline representing a current state of the media composition, wherein the timeline includes the at least one media track and the at least one data track, wherein the data track references at least one data stream comprising a stream of data packets, wherein each of the data packets occupies a unique time address corresponding to a media element on the media track; enabling a user to perform an editing operation to the time-based media composition, wherein the editing operation is applied to both the at least one media track and the at least one data track, thereby maintaining temporal synchrony between the at least one media track and the at least one data track; and enabling the user to perform an editing operation to the media composition, wherein the editing operation is applied to only one of the at least one media track and the at least one data track, thereby altering a temporal relationship between at least a portion of the at least one media track and the at least one data track.
In general, in still a further aspect a method of editing a video composition using a video editing software application, wherein the video composition includes at least one video track and at least two data tracks, the method including: displaying a timeline representing a current state of the video composition, wherein the timeline includes the at least one video track and the at least two data tracks, wherein each data track references at least one data stream comprising a stream of data packets, wherein each of the data packets occupies a unique time address corresponding to a frame time on the video track, and wherein each of the packets within a given stream are characterized by a common data type, and wherein each of a first data track and a second data track includes a data stream of a first data type; and enabling a user of the video editing software application to select one of the data stream of the first type from the first data track and the data stream of the first type from the second data track, wherein the selected data stream is selectively included in at least one of playback or export of the video composition.
Time-synchronous data pertaining to video and audio programs include descriptive, structural, and enhancement data. Types of time-synchronous data include, but are not limited to: closed and open captions; subtitles; teletext; film metadata, such as edgecodes, keycodes, and pull-down flags; timecodes; camera data, such as position, zoom, iris, and gain; location data, such as GPS coordinates; program description; 3D stereo parameters; aspect ratio conformance, such as Advanced Format Descriptor; and interactive television programming. Audio-specific examples include: gain, pan, and mixing parameters; MIDI data; and audio timecodes. A system for facilitating the editing of video programs with enhancement data is described in U.S. Pat. No. 7,930,624, entitled Editing Time-Based Media with Enhanced Content, which is incorporated herein by reference.
Time-synchronous data are organized into a hierarchy of data structures including clips, tracks, streams, and packets, as shown in
Data packets are organized into streams, each stream including only a single type of packet, i.e., corresponding to a single synchronous data type. Within each stream, each packet occupies a unique time address, which corresponds to a time interval of a video frame. Thus, there can be no more than one packet at a given frame address, but it is not necessary for each frame address in the range of frames in the content to be assigned a packet in the stream, as shown, for example in the sparse stream illustrated in
One or more streams are organized into tracks, in which each stream has the same time length, and each track includes no more than one stream of a given type (ID).
Multiple data tracks are assembled into clips, with each track having the same duration and starting time address. Video and/or audio tracks are included in the clips, and the clip structure ensures that data streams are synchronized with each other and with audio and video content.
Multiple clips are further arranged in a desired order into a sequence. In a sequence, any portion of a data track or data stream (accurate to a video frame time) can be selected and inserted. This enables the user of a digital media production system to alter the order and timing of the data in clips to create a new order and timing.
The digital media production system processes the sequence by following a chain of references beginning with the reference to a temporal span of a data track of the sequence, which points to the clip encapsulating the data, which in turn points to the data packets themselves. Such a chain of references is illustrated in
With the described data structure, the digital media processing system is able to perform a number of functions relating to the synchronous data. A fundamental new capability is playback and/or storing of streams of arbitrary data in absolute synchronization with video and audio. The synchronization may be accurate down to the level of individual video frame times. Note, the synchronization extends through each temporal location of the sequence, and is not limited to start/end point synchronization. The structure also permits the data streams to be played or processed from multiple disk files, including files that may be stored in different locations. For example, during the editing of a television program, it is often necessary to use footage from archives that have their own streams of closed captions encoded into files. The described system is able to incorporate these files into the production process with transcoding, re-wrapping, combining, or otherwise converting the stream files, thus speeding up the process of editing data-oriented material in a deadline-driven production environment. In another example, it may be necessary to replace a small portion of the data in a sequence to correct a mistake, such as a misspelled caption or incorrect station branding. The described system may be used to prepare a small stream file containing only the corrected data with a reference to the new file in the original sequence in place of the erroneous data. This approach avoids the need to fully recreate the original data track containing the mistake, which results in a more efficient and faster workflow.
The system is able to combine, filter, and convert the streams of data in real-time or as a non-real-time process.
The system also permits a user to filter data according to selected rules and parameters to determine what media and data flows to processes downstream of the media editing system. In the example illustrated in
The data may be converted into a form that allows searching or analysis. For example, data may be converted into text and read into a media asset management system, which can then be searched. Searchable data may include closed caption text, program description data, positional data such as GPS coordinates, and interactive program content. For example, a documentary or reality film maker may use GPS location data to search and sort available media based on the locations at which they were shot.
The data structure described herein permits a digital media production system that includes an interactive non-linear video or audio editing system to also edit and play back data tracks with the same degree of flexibility hitherto available only for media tracks.
An example of a graphical user interface for a non-linear editing system in which an editor is able to manipulate multiple data tracks is illustrated in
The various components of the system described herein may be implemented as a computer program using a general-purpose computer system. Such a computer system typically includes a main unit connected to both an output device that displays information to a user and an input device that receives input from a user. The main unit generally includes a processor connected to a memory system via an interconnection mechanism. The input device and output device also are connected to the processor and memory system via the interconnection mechanism.
One or more output devices may be connected to the computer system. Example output devices include, but are not limited to, liquid crystal displays (LCD), plasma displays, cathode ray tubes, video projection systems and other video output devices, printers, devices for communicating over a low or high bandwidth network, including network interface devices, cable modems, and storage devices such as disk or tape. One or more input devices may be connected to the computer system. Example input devices include, but are not limited to, a keyboard, keypad, track ball, mouse, pen and tablet, communication device, touch-screen, and data input devices. The invention is not limited to the particular input or output devices used in combination with the computer system or to those described herein.
The computer system may be a general purpose computer system which is programmable using a computer programming language, a scripting language or even assembly language. The computer system may also be specially programmed, special purpose hardware. In a general-purpose computer system, the processor is typically a commercially available processor. The general-purpose computer also typically has an operating system, which controls the execution of other computer programs and provides scheduling, debugging, input/output control, accounting, compilation, storage assignment, data management and memory management, and communication control and related services. The computer system may be connected to a local network and/or to a wide area network, such as the Internet. The connected network may transfer to and from the computer system program instructions for execution on the computer, media data, metadata, review and approval information for a media composition, media annotations, and other data.
A memory system typically includes a computer readable medium. The medium may be volatile or nonvolatile, writeable or nonwriteable, and/or rewriteable or not rewriteable. A memory system typically stores data in binary form. Such data may define an application program to be executed by the microprocessor, or information stored on the disk to be processed by the application program. The invention is not limited to a particular memory system. Time-based media may be stored on and input from magnetic or optical discs, which may include an array of local or network attached discs.
A system such as described herein may be implemented in software or hardware or firmware, or a combination of the three. The various elements of the system, either individually or in combination may be implemented as one or more computer program products in which computer program instructions are stored on a non-transitory computer readable medium for execution by a computer, or transferred to a computer system via a connected local area or wide are network. Various steps of a process may be performed by a computer executing such computer program instructions. The computer system may be a multiprocessor computer system or may include multiple computers connected over a computer network. The components described herein may be separate modules of a computer program, or may be separate computer programs, which may be operable on separate computers. The data produced by these components may be stored in a memory system or transmitted between computer systems.
Having now described an example embodiment, it should be apparent to those skilled in the art that the foregoing is merely illustrative and not limiting, having been presented by way of example only. Numerous modifications and other embodiments are within the scope of one of ordinary skill in the art and are contemplated as falling within the scope of the invention.
Number | Name | Date | Kind |
---|---|---|---|
4538188 | Barker et al. | Aug 1985 | A |
4685003 | Westland | Aug 1987 | A |
4746994 | Ettlinger | May 1988 | A |
5012334 | Etra | Apr 1991 | A |
5045940 | Peters et al. | Sep 1991 | A |
5097351 | Kramer | Mar 1992 | A |
5196933 | Henot | Mar 1993 | A |
5214528 | Akanabe et al. | May 1993 | A |
5237648 | Mills et al. | Aug 1993 | A |
5267351 | Reber et al. | Nov 1993 | A |
5274758 | Beitel et al. | Dec 1993 | A |
5307456 | MacKay | Apr 1994 | A |
5317732 | Gerlach, Jr. et al. | May 1994 | A |
5355450 | Garmon et al. | Oct 1994 | A |
5390138 | Milne et al. | Feb 1995 | A |
5404316 | Klingler et al. | Apr 1995 | A |
5428731 | Powers, III | Jun 1995 | A |
5442744 | Piech et al. | Aug 1995 | A |
5467288 | Fasciano et al. | Nov 1995 | A |
5488433 | Washino et al. | Jan 1996 | A |
5489947 | Cooper | Feb 1996 | A |
5493568 | Sampat et al. | Feb 1996 | A |
5513306 | Mills et al. | Apr 1996 | A |
5515490 | Buchanan et al. | May 1996 | A |
5534942 | Beyers et al. | Jul 1996 | A |
5537141 | Harper et al. | Jul 1996 | A |
5537157 | Washino et al. | Jul 1996 | A |
5539869 | Spoto et al. | Jul 1996 | A |
5541662 | Adams et al. | Jul 1996 | A |
5561457 | Cragun et al. | Oct 1996 | A |
5568275 | Norton et al. | Oct 1996 | A |
5577190 | Peters | Nov 1996 | A |
5584006 | Reber et al. | Dec 1996 | A |
5585858 | Harper et al. | Dec 1996 | A |
5592602 | Edmunds et al. | Jan 1997 | A |
5613057 | Caravel | Mar 1997 | A |
5617146 | Duffield et al. | Apr 1997 | A |
5619636 | Sweat et al. | Apr 1997 | A |
5623308 | Civanlar et al. | Apr 1997 | A |
5652714 | Peterson et al. | Jul 1997 | A |
5659790 | Kim et al. | Aug 1997 | A |
5659792 | Walmsley | Aug 1997 | A |
5659793 | Escobar et al. | Aug 1997 | A |
5664216 | Blumenau | Sep 1997 | A |
5680619 | Gudmundson et al. | Oct 1997 | A |
5682326 | Klingler et al. | Oct 1997 | A |
5684963 | Clement | Nov 1997 | A |
5712953 | Langs | Jan 1998 | A |
5724605 | Wissner | Mar 1998 | A |
5752029 | Wissner | May 1998 | A |
5754851 | Wissner | May 1998 | A |
5760767 | Shore et al. | Jun 1998 | A |
5764275 | Lappington et al. | Jun 1998 | A |
5767846 | Nakamura et al. | Jun 1998 | A |
5781435 | Holroyd et al. | Jul 1998 | A |
5801685 | Miller et al. | Sep 1998 | A |
5822019 | Takeuchi et al. | Oct 1998 | A |
5826102 | Escobar et al. | Oct 1998 | A |
5852435 | Vigneaux et al. | Dec 1998 | A |
5860073 | Ferrel et al. | Jan 1999 | A |
5861881 | Freeman et al. | Jan 1999 | A |
5889514 | Boezeman et al. | Mar 1999 | A |
5892506 | Hermanson | Apr 1999 | A |
5892507 | Moorby et al. | Apr 1999 | A |
5905841 | Peters et al. | May 1999 | A |
5907366 | Farmer et al. | May 1999 | A |
5910825 | Takeuchi | Jun 1999 | A |
5926613 | Schaffer | Jul 1999 | A |
5930445 | Peters et al. | Jul 1999 | A |
5946445 | Peters et al. | Aug 1999 | A |
5969716 | Davis et al. | Oct 1999 | A |
5977962 | Chapman et al. | Nov 1999 | A |
5978648 | George et al. | Nov 1999 | A |
5982445 | Eyer et al. | Nov 1999 | A |
5995951 | Ferguson | Nov 1999 | A |
5999173 | Ubillos | Dec 1999 | A |
6016362 | Kato et al. | Jan 2000 | A |
6037932 | Feinleib | Mar 2000 | A |
6038573 | Parks | Mar 2000 | A |
6058236 | Peters et al. | May 2000 | A |
6081262 | Gill et al. | Jun 2000 | A |
6091407 | Boetje et al. | Jul 2000 | A |
6092122 | Liu et al. | Jul 2000 | A |
6118444 | Garmon et al. | Sep 2000 | A |
6195497 | Nagasaka et al. | Feb 2001 | B1 |
6199082 | Ferrel et al. | Mar 2001 | B1 |
6201924 | Crane et al. | Mar 2001 | B1 |
6212527 | Gustman | Apr 2001 | B1 |
6230173 | Ferrel et al. | May 2001 | B1 |
6236395 | Sezan et al. | May 2001 | B1 |
6243087 | Davis et al. | Jun 2001 | B1 |
6249280 | Garmon et al. | Jun 2001 | B1 |
6262723 | Matsuzawa et al. | Jul 2001 | B1 |
6262724 | Crowe et al. | Jul 2001 | B1 |
6330004 | Matsuzawa et al. | Dec 2001 | B1 |
6353461 | Shore et al. | Mar 2002 | B1 |
6400378 | Snook | Jun 2002 | B1 |
6404978 | Abe | Jun 2002 | B1 |
6411725 | Rhoads | Jun 2002 | B1 |
6426778 | Valdez, Jr. | Jul 2002 | B1 |
6430355 | Nagasawa | Aug 2002 | B1 |
6476828 | Burkett et al. | Nov 2002 | B1 |
6484199 | Eyal | Nov 2002 | B2 |
6489969 | Garmon et al. | Dec 2002 | B1 |
6544294 | Greenfield et al. | Apr 2003 | B1 |
6546405 | Gupta et al. | Apr 2003 | B2 |
6553142 | Peters | Apr 2003 | B2 |
6564263 | Bergman et al. | May 2003 | B1 |
6597375 | Yawitz | Jul 2003 | B1 |
6618547 | Peters et al. | Sep 2003 | B1 |
6654933 | Abbott et al. | Nov 2003 | B1 |
6715126 | Chang et al. | Mar 2004 | B1 |
7725812 | Balkus et al. | May 2010 | B1 |
7930624 | Phillips et al. | Apr 2011 | B2 |
20020188628 | Cooper et al. | Dec 2002 | A1 |
20030018609 | Phillips et al. | Jan 2003 | A1 |
20040268224 | Balkus et al. | Dec 2004 | A1 |
20050120127 | Bradley et al. | Jun 2005 | A1 |
20080298768 | Isobe et al. | Dec 2008 | A1 |
20090282080 | Schlossberg et al. | Nov 2009 | A1 |
20110064146 | Chen et al. | Mar 2011 | A1 |
20110154290 | Kelly | Jun 2011 | A1 |
20110191661 | Phillips et al. | Aug 2011 | A1 |
Number | Date | Country |
---|---|---|
0403118 | Dec 1990 | EP |
0469850 | Feb 1992 | EP |
0526064 | Feb 1993 | EP |
0564247 | Oct 1993 | EP |
0592250 | Apr 1994 | EP |
0596823 | May 1994 | EP |
0613145 | Aug 1994 | EP |
0689133 | Oct 1995 | EP |
0706124 | Apr 1996 | EP |
2336025 | Oct 1999 | GB |
WO8807719 | Oct 1988 | WO |
WO9321636 | Oct 1993 | WO |
WO9403897 | Feb 1994 | WO |
WO9429868 | Dec 1994 | WO |
WO9626600 | Aug 1996 | WO |
WO9631829 | Oct 1996 | WO |
WO9636007 | Nov 1996 | WO |
WO9712342 | Apr 1997 | WO |
WO9737497 | Oct 1997 | WO |
WO9804984 | Feb 1998 | WO |
WO9952045 | Oct 1999 | WO |
WO0073875 | Dec 2000 | WO |
Entry |
---|
Bulterman, D.C.A. et al., “GRiNS: A Graphical Interface for creating and playing SMIL Documents”, Computer Networks & ISDN Systems, Sep. 30, 1998, North Holland Publishing, Amsterdam, NL, vol. 30, NR. 1-7, pp. 519-529. |
Candan, K. Selcuk, et al., “CHIMP: A Framework for Supporting Distributed Multimedia Document Authoring and Presentation”, ACM Multimedia 96, Boston, MA, pp. 329-340. |
Fahmi, H. et al., “Distributed framework for real-time multimedia object communication”, Object-Oriented Real-Time Distributed Computing 2000, ISORC 2000, Proceedings, Third IEEE International Symposium, Newport, CA, USA, Mar. 15-17, 2000, Los Alamitos, CA, USA, IEEE Computer Society, pp. 252-259. |
Mines, Robert F. et al., “DAVE: A Plug and Play Model for Distributed Multimedia Application Development”, ACM Multimedia 94, San Francisco, CA, pp. 59-66. |
Nicol, John R., et al., “How the Internet Helps Build Collaborative Multimedia Applications”, Communications of the ACM, Jan. 1999, vol. 42, No. 1, pp. 79-85. |
Sauer, S. et al., “Extending UML for modeling of multimedia applications”, Visual Languages, 1999, Proceedings 1999 IEEE Symposium, Tokyo, JP, Sep. 13-16, 1999, Los Alamitos, CA, USA, IEEE Computer Society, pp. 80-87. |
AAF Specification Version 1.0 DR4, 2000, pp. i-181. |
Ackermann, Phillipp, “Design and Implementation of an Object-oriented Media Composition Framework”, Int'l, Computer Music Conference, 1994. |
Agnew, J.N. et al., “The Right Tool for the Job: A Quest for the Perfect Authoring Package”, Proceedings of SIGDOC '92: 10th Annual ACM Conference on Systems Documentation, Oct. 1992, Ottawa, Ontario, Canada, pp. 253-258. |
Authorware Users Guide, Macromedia Inc., 1992, pp. 8-9, 12-13, 15, 17, 20, 25, 28-31, 36-37, 43, 46-47, 115, 178-179, 191, 232, 242, 301-302. |
Baecker, R et al., “A Multimedia System for Authoring Motion Pictures”, ACM Multimedia 1996, pp. 31-42. |
Blakowski, Gerold, et al., “Tools for Specifying and Executing Synchronized Multimedia Presentations”, Proc. 2nd Int'l. Workshop on OS for Digital Audio Video, Nov. 18-19, 1991, pp. 271-279. |
The Brughetti Corporation, Product Information for Pure Image Creator, SLICE Transition Maker and AIR Playout Manager, 1994, (11 pages). |
Buchanan, M. Cecelia et al., “Automatic Temporal Layout Mechanisms”, Proc. ACM Multimedia 1993, pp. 341-350. |
Buchanan, M. Cecelia et al., “Specifying Temporal Behavior in Hypermedia Documents”, ECHT '92, pp. 262-271. |
Buford, John F., “Integrating Object-Oriented Scripting Languages with HyTime”, from Proceedings of ICMCS 1994 IEEE, pp. 1-10. |
Bulterman, Dick C.A. et al., “Multimedia Authoring Tools: State of the Art and Research Challenges”, LNCS 1000, 1995, pp. 1-17. |
Chiu, Patrick et al., A Dynamic Grouping Technique for Ink and Audio Notes, USIT-98, San Francisco, CA 1998 ACM, pp. 195-202. |
Chiueh, Tzi-cker, et al., Zodiac: A History-Based Interactive Video Authoring System; ACM Multimedia '98, Bristol, UK, 1998 ACM, pp. 435-443. |
Chua, Tat-Seng et al., “A Video Retrieval and Sequencing System”, ACM Transactions on Information Systems, vol. 13, No. 4, Oct. 1995, pp. 373-407. |
“The Coming Revolution”, Communications of the ACM, Jul. 1989, vol. 32, No. 7, pp. 794-901. |
Csinger, Andrew et al., “AI Meets Authoring: User Models for Intelligent Multimedia, Artificial Intelligence review, special issue on user modeling”, 8, pp. 447-468, 1995. |
Drapeau, George D. et al., “MAEstro—A Distributed Multimedia Authoring Environment”, USENIX, Summer 1991, pp. 315-328. |
Drapeau, George D., “Synchronization in the MAEstro Multimedia Authoring Environment”, ACM Multimedia 1993, pp. 331-340. |
Davenport, Gloriana et al., “Cinematic Primitives for Multimedia”, 1991 IEEE, Jul. 7, 1991, pp. 67-74. |
The Electronic Scrapbook: Towards an Intelligent Home-Video Editing System, by Amy Susan Bruckman, Sep. 1991, video Mosaic: Laying Out Time in a Physical Space. |
Eun, Seong Bae et al., “Specification of Multimedia Composition and A Visual Programming Environment”, Proceeding of the first ACM Int'I. Conference on Multimedia, pp. 167-173, 1993. |
Gephard, C., Der Widersenstigen Zahmung: Computer & Video Radio Fernsehen Electronik, vol. 44, No. 1, Jan. 1, 1995, Berlin, DE. |
Gibbs, Simon, “Composition Multimedia and Active Objects”, in Proc. OOPSLA 1991, pp. 1-16. |
Gibbs, Simon, “Data Modeling of time-Based Media”, SIGMOD'94, 1994 ACM, pp. 91-102. |
Gibbs, Simon, “Video Nodes and Video Webs: Uses of Video in Hypermedia”, Proc. ACM Conf. on Hypertext, 1992, p. 3. |
Girgensohn, Andreas et al., “A Semi-automatic Approach to Home Video Editing”, UIST '00, San Diego, CA , 2000 ACM, CHI Letters vol. 2, 2, pp. 81-89. |
Gruenette R., “Authorware and IconAuthor”, CD-ROM Professional, vol. 8, No. 10, Oct. 1995, pp. 81-89. |
Hardman, H., “New Authoring Tools for Windows”, BYTE, vol. 18, No. 9, Aug. 1993, pp. 153-156. |
Hardman, Lynda et al., “Authoring Interactive Multimedia: Problems and Prospects”, CWI Quarterly, vol. 7(1) 1994, pp. 47-66. |
Hardman, Lynda et al., “Structured Multimedia Authoring”, ACM Multimedia 1993, pp. 283-289. |
Herman, I. et al., “MADE: A Multimedia Application Development Environment”, CWI Quarterly, vol. 7(1) 1994, pp. 27-46. |
Hirata, Kyoji, et al., “Content-Oriented Integration in Hypermedia Systems”, Hypertext '86, 7th ACM Conf. On Hypertext, Conf. 7, Mar. 1996, pp. 11-21. |
Hirzalla, Nael, et al., “A temporal Model for Interactive Multimedia Scenarios”, in IEEE Multimedia, vol. 2, No. 3, Fall 1995, pp. 24-31. |
Hudson, Scott E. et al., “The Walk-Through Approach to Authoring Multimedia Documents”, Multimedia '94, 1994 ACM, pp. 173-180. |
Hung, Yung-Chen, “A Hypervideo System Generator”, Software Practice and Experience, vol. 17(11), Nov. 1997, pp. 1263-1281. |
Koefel, J.F. et al., “Improving Visual Programming Languages for Multimedia Authoring”, Educational Multimedia and Hypermedia Annula, 1993, Proceeding of Ed-Media 93-World on Educational Multimedia and Hypermedia, Jun. 1993, Orlando, FL, pp. 286-293. |
Kurlander, David et al., “A Visual Language for Browsing, Undoing, and Redoing Graphical Interface Commands”, in Visual Languages and Visual Programming, 1990, pp. 257-275. |
Mackay, W.E. et al., “Video Mosaic: Laying Out Time in a Physical Space”, Proceedings of Multimedia '94, San Francisco, CA, ACM, 8 pages. |
Mackay, W.E. et al., “Virtual Video Editing in Interactive Multimedia Applications”, Communications of the ACM, Jul. 1989, vol. 32, No. 7, pp. 802-810. |
Macromind Mediamaker, Macromind Inc., 1990, pp. 129, 165 and 166. |
Matthews, James et al., “VideoScheme: A Programmable Video Editing System for Automation and Media Recognition”, ACM Multimedia 1993. |
Ohanian, Tom, “The Avid 1 Media Composer”, Int'l. Broadcast Engineer, Jan. 1991, pp. 22-23. |
Open Media Framework, OMF Interchange Specification, Avid Technology, Inc., Version 2.0, Nov. 29, 1995, pp. i-245. |
Pazandak, Paul et al., “A Multimedia Temporal Specification Model and Language”, Technical Report 94-33, University of Minnesota, 1994. |
Ryan, Mark et al., “A Development Process for Large Multimedia Titles”, 1994 ACM, pp. 126-138. |
Sack, Warren et al., “IDIC: Assembling Video Sequences from Story Plans and Content Annotations”, In: Proceedings of IEEE International Conf. on Multimedia Computing and Systems in Boston, MA, IEEE Computer Science Society Press, pp. 30-36, 1994. |
Schleicher, Stephen, “How to Create Web Markers in Premiere 6”, Digital Video Editing, Digital Media Online, 2001. |
“Schneller Zugrigg: Video Machine Von Fast”, Radio Fernshen Elektonik, vol. 44, No. 11, Nov. 1, 1995, Berlin, DE. |
Siocchi, A. et al., “The Integrator: A Prototype for Flexible Development of Interactive Digital Applications”, Interactive Multimedia, vol. 2, No. 3, 1993, pp. 5-26. |
Tonomura, Yoshinobo, “Video Handling Based on Structured Information for Hypermedia Systems”, Proceedings of the Int'l Conference of Multimedia Information Systems, Jan. 1991, pp. 333-334. |
Ueda, Hirotada et al., “Automatic Scene Separation and Tree Structure GUI for Video Editing”, ACM Multimedia 96, Boston, MA, 1996 ACM, pp. 405-406. |
Ueda, Hirotada et al., “Automatic Structure Visualization for Video Editing”, Interchi '93, 1993 ACM, Apr. 24-29, 1993, pp. 137-141; 545-546. |
Uedo, Hirotada et al., Impact: An Interactive Natural-Motion-Picture Dedicated Multimedia Authoring System, 1991 ACM, pp. 343-350. |
Ueda, Hirotoda et al., “Impact: Interactive Motion Picture Authoring System for Creative Talent”, 1993 ACM, p. 525. |
Van der Sluis, Ielka et al., “Enriching Textual Documents with Time-codes from Video Fragments”, In: Proceedings RIAO 2000 Content-Based Multimedia Information Access, Paris, Apr. 2000, pp, 431-440. |
Wahl, Thomas et al,, “Representing Time in Multimedia Systems”, Proc. IEEE Int'l Conf. On Multimedia 1994, pp. 538-543. |
Wahl, Thomas et al., “Tiempo: Temporal Modeling and Authoring of Interactive Multimedia”, Proc. Int'l Conf. of Multimedia, 1995, pp. 274-277. |
Weiss, Ron et al., “Content-Based Access to Algebraic Video”, Proceedings Int'l Conf. on Multimedia Computing and Systems, Boston, MA, May 1994, pp. 140-151. |
Zhang et al., “Video parsing, retrieval and browsing: an integrated and content-based solution”, Proceedings of the third ACM Int'l Conference on Multimedia, published 1995, p. 15-25, renumbered in printout copy as p. 1-18. |
Number | Date | Country | |
---|---|---|---|
20120301112 A1 | Nov 2012 | US |