The invention relates to the transfer, distribution, and play of streamed information or content in a network environment. More particularly, the invention relates to the creation of streamed and loopable content in a network environment.
The Internet comprises a web of computers and networks, which are widely spread throughout the world. The Internet currently comprises millions of network connections, and is used by millions of people, such as for business, education, entertainment, and/or basic communication.
Digital content, such as sound recordings, e.g. songs, are often transferred across the Internet. In addition to the basic transfer of song files, numerous network enabled radio stations have been introduced, which provide content to listeners at computers across the Internet. Network enabled radio has significantly increased the magnitude and variety of content to recipients, as compared to conventional over-the-air radio broadcasts
While there are numerous Internet radio stations currently in operation, there are many technological shortcomings in the delivery of digital content to listeners. For example, buffering between songs, i.e. tracks, and even during tracks, is a common occurrence, which commonly diminishes the quality of a broadcast for listeners. As well, a short or long duration failure across the network, e.g. a blackout, results in the cessation of a music presentation, further diminishing the user experience.
As well, current content delivery systems do not offer sufficient flexibility and/or scalability for future network architectures and increased market demands. As the number of Internet radio stations increases to meet consumer demand, and as the number and variety of content recipients, i.e. listeners, increases, it will be necessary to provide substantial improvements in content delivery architectures.
Pull vs. Push Mod Is for Content Delivery. In content delivery systems which operate on a push distribution model, a source complex makes an outbound connection to a distribution point, and pushes data to the distribution point at a rate determined by the source complex. However, in a content delivery system which operates on a push distribution model, broadcasters are required to be aware of the network architecture. Therefore, every time a distribution point is added, the broadcast configuration is required to change, to make an outbound connection to the new distribution point. As well, the implementation of fail over and/or load balancing logic typically requires that a push system frequently reconfigure both the distribution points and the broadcaster hosts.
A pull model typically requires less buffering logic than a push model for the broadcaster, because the broadcaster just sends data obliviously, i.e. the distribution point is required to receive the data and feed a local buffer appropriately. In a content delivery system which operates on a pull distribution model, a distribution point initiates the connection with a broadcaster, and requests a desired stream identifier.
Several structures and methods have been described for the distribution of content in a network environment.
N. Dwek, Multimedia Content Delivery System and Method, U.S. Pat. No. 6,248,946, describes “A system and method for delivering multimedia content to computers over a computer network, such as the Internet, includes a novel media player which may be downloaded onto a user's personal computer. The media player includes a user interface which allows a listener to search an online database of media selections and build a custom playlist of exactly the music selections desired by the listener. The multimedia content delivery system delivers advertisements which remain visible on a user's computer display screen at all times when the application is open, for example, while music selections are being delivered to the user. The advertisements are displayed in a window which always remains on a topmost level of windows on the user's computer display screen, even if the user is executing one or more other programs with the computer.”
M. DeLorenzo, Multi-Room Entertainment System with In-Room Media Player, U.S. Pat. No.6,438,450, describes “A plurality of media data, including audio data or audio/video data, are stored in a central database. A plurality of in-room, user interface systems access the media data through a central server. The central server presents to the in-room system a selection menu through which at least one of the media data may be selected. Upon selection of a media data by the user interface, the central server accesses the selected media data from the central database and transmits it to the in-room system. The media data may be transmitted by downloading the data to an intermediate system, playing the media data at the intermediate system and outputting the played media data to the in-room system through a communications line. The media data may also be transmitted by streaming the media data to the in-room system through a communications line. The central server may present to the in-room system any of a number of additional menus including a purchase menu through which the selected media data may be purchased, an activation menu through which communication between the in-room system and the central server may be established for a period of time, a radio menu through which any of a plurality of programmed media-data channels may be accessed and a mood menu through which the brightness of the image displayed on the in-room system video monitor may be affected.”
Other structures and methods have been described for the distribution of content in a network environment, such as: Streaming Information Providing Method, European STREAM SOURCING CONTENT DELIVERY SYSTEM patent application Ser. No. 1187 423; O. Hodson, C. Perkins, and V. Hardman, Skew Detection and Compensation for Internet Audio Applications; 2000 IEEE International Conference on Multimedia and Expo; 2000; C. Aurrecoechea, A. Campbell, and Linda Hauw, A Survey of Qos Architectures, Center for Telecommunication Research, Columbia University; S. Cen, C. Pu, R. Staehli, and J. Walpole, A Distributed Real-Time MPEG Video Audio Player, Oregon Graduate Institute of Science and Technology; N. Manouselis, P. Karampiperis, I. Vardiambasis, and A. Maras, Digital Audio Broadcasting Systems under a Qos Perspective, Telecommunications Laboratory, Technical University of Crete; Helix Universal Gateway Configuration Guide, RealNetworks Technical Blueprint Series; Jul. 21, 2002; Helix Universal Server from RealNetworks Helix Universal Gateway Helix Universal Server, www.realnetworks.com; Media Delivery and Windows Media Services 9 Series.
Other systems describe various details of audio distribution, streaming, and/or the transfer of content in a network environment, such as G. France and S. Lee, Method for Streaming Transmission of Compressed Music, U.S. Pat. No. 5,734,119; D. Marks, Group Communications Multiplexing System, U.S. Pat. No. 5,956,491; M. Abecassis, Integration of Music From a Personal Library with Real-Time Information, U.S. Pat. No. 6,192,340; J. Logan, D. Goessling, and C. Call, Audio Program Player Including a Dynamic Program Selection Controller, U.S. Pat. No. 6,199,076; E. Sitnik, Multichannel Audio Distribution System Having Portable Receivers, U.S. Pat. No. 6,300,880; M. Bowman-Amuah, Method For Providing Communication Services Over a Computer Network System, U.S. Pat. No. 6,332,163; H. Ando, S. Ito, H. Takahashi, H. Unno, and H. Sogabe, Information Recording Device and A Method of Recording Information by Setting the Recording Area Based on Contiguous Data Area, U.S. Pat. No. 6,530,037; P. Hunt and M. Bright, Method and Apparatus for Intelligent and Automatic Preference Detection of Media Content, U.S. Patent Application Publication No. U.S. Pat. No. 2002 0078056; G. Beyda and K. Balasubramanian, Hybrid Network Based Advertising System and Method, U.S. Patent Application Publication No. U.S. Pat. No. 2002 0082914; System and Method for Delivering Plural Advertisement Information on a Data Network, International Publication No. WO 02/063414; Method for Recording and/or Reproducing Data on/from Recording/Recorded Medium, Reproducing Apparatus, Recording Medium, Method for Recognizing Recording/Recorded Medium, and Method for Recording and/or Reproducing Data for Apparatus Using Recording/Recorded Medium, European Patent Application No. EP 1 178 487; Method and System for Securely Distributing Computer Software Products, European Patent Application No. EP 1 229 476; Information Transmission System, Information Transmission Method, Computer Program Storage Medium Where Information Transmission Program is Stored, European Patent Application No. EP 1 244 021; Digital Content Publication, European Patent Application No. EP 1 267 247; File and Content Management, European Patent Application No. EP 1 286 351; S. Takao; Y. Kiyoshi; W. Kazuhiro; E. Kohei. Packet Synchronization Recovery Circuit; Section: E, Section No. 1225, Vol. 16, No. 294, Pg. 120; Jun. 29, 1992; R. Sion, A. Elmagarmid, S. Prabhakar, and A. Rezgui, Challenges in Designing a Qos Aware Media Repository, Purdue University; Z. Chen, S. Tan, R. Campbell, and Y. Li, Real Time Video and Audio in the World Wide Web, University of Illinois at Urbana-Champaign; Content Networking with the Helix Platform, RealNetworks White Paper Series; Jul. 21, 2002; C. Hess, Media Streaming Protocol: An Adaptive Protocol for the Delivery of Audio and Video over the Internet, University of Illinois at Urbana-Champaign, 1998; R. Koster, Design of a Multimedia Player with Advanced Qos Control, Oregon Graduate Institute of Science and Technology, Jan. 1997; C. Poellabauer and K. Schwan, Coordinated CPU and Event Scheduling for Distributed Multimedia Applications, College of Computing Georgia Institute of Technology, R. West, Computing Science Department Boston University; and Windows Media.
While some content delivery technologies describe the delivery of streamed content across a network, existing systems do not adequately provide a seamless delivery to a large number of recipients, nor do such technologies provide a “fail safe” seamless playback of content upon failure across the network.
It would be advantageous to provide a system and an associated method which provides a seamless delivery of songs to a large number of recipients, which provides a “fail safe” seamless playback of content upon failure across the network. The development of such a content delivery system would constitute a major technological advance.
It would also be advantageous to provide a system and an associated method which provides delivery of content as well as metadata to multiple distribution points, and has the capability of broadcasting content indefinitely, even if a database or content store fails. The development of such a content delivery system would constitute a major technological advance.
The stream sourcing content delivery system goes to a database and builds a physical stream, based on a schedule. The stream source content delivery system works at a station ID (SID), finds the order of the delivery of content for the station based upon the schedule, and downloads a plurality of music files, e.g. 6 hours of music, to its hard drive to enable play back. The system then concatenates the files, to create a stream, and awaits the request of one or more stream recipients. Some preferred system embodiments further comprise a fail-safe mode, whereby a loop of music is generated from the downloaded stream, and is delivered to one or more users when further access to content is interrupted, such that recipients experience an uninterrupted delivery of a plurality of files, e.g. songs. A stream source content delivery system provides flexibility and scalability for large number of stations, e.g. up to 100 stations, and/or listeners.
A stream source content delivery system 12 fetches 22 songs 21, e.g. 21a-21n, from a content store 20 and stores them on a local disk 25 to be played by the user U. The system 12 loads these songs 21 into a memory 27, and streams the songs 21. It is sometimes the case that access to the content store 20 is lost. If a connection is lost between the stream sourcing content delivery server 12 and the content store 20 or local disk 25, the system 12 preferably goes into a looping behavior 200 (
The stream sourcing content delivery system 12 fetches songs 21, e.g. 21a-21n, from a content store 20 and stores them on a local disk 25, to be played for a user U at a client terminal or computer 32, e.g. 32a in
It is sometimes the case that access to the content store 20 is lost. The stream sourcing content delivery system 12 loads these songs 21 into memory 27 and streams them to listeners 32. If a connection is lost between the stream sourcing content delivery system 12 and the content store 20 or local disk 25, the system 12 goes into a looping behavior 200 (
Some preferred embodiments of the stream sourcing content delivery system 12 provide, on a per file basis, an adjustable bit rate at which a stream 28 is sent 190,192 (
Some preferred embodiments of the stream sourcing content delivery system 12 also preferably provide the insertion of metadata 210 (
Some system embodiments 12 act as a component of a streaming architecture, such as for a Radio@ streaming architecture, available through Radio@AOL. The stream sourcing content delivery system 12 delivers a formatted stream 28, to a distribution point 26, based on a content store 26 and a scheduling database 14.
From the high level view, the stream sourcing content delivery system 12 fetches playlists 18 from a database 15 for each station 30, e.g. 30a, that the system 12 serves. The system 12 analyzes the playlists 18, locally caches 24 the content 21a-21n for each station 30, and sets up a listen thread, i.e. stream 28. A distribution point 26 can then connect to the stream sourcing content delivery system 12, and relay the data stream 28, typically corresponding to a relay protocol.
The stream sourcing content delivery system 12 shown in
The stream sourcing content delivery system 12 typically logs the status of the system operation and hardware, such that operations personnel can properly manage the stream sourcing system hardware 12.
Some preferred embodiments of stream sourcing content delivery system 12 comprehensively control the content 20, the source complex, the distribution point 26, the transport, and the clients 32a-32j, to provide integrated flexibility and control over how content 21a-21n is delivered to users U, to ensure both that the user experience is maximized, and that system improvements and upgrades are readily implemented.
In addition to improving the backend architecture of content delivery, the stream sourcing content delivery system 12 improves user experience. For example, some preferred embodiments of the stream sourcing content delivery system 12 do not require buffering between tracks. Users do not have to wait to buffer between songs 21 on the same stations 30. Instead, there is typically a short buffering period when a user tunes into a station 30. While the user listens to a station 30, the user does not experience any buffering, unless the user has an abnormally bad network condition.
As well, some embodiments of the stream sourcing content delivery system 12 provide reduced network congestion, through the use of matched data transmission, e.g. 14 kbps codec, and through the minimization of data overhead, which improves the delivery to data to a client 32, i.e. it is less likely that a client 32 is not able to receive the necessary data for a given time slice. The stream sourcing content delivery system 12 reduces a need to rebuffer in the middle of a song 21 due to network congestion, which further provides an improved user experience.
The distribution point 26 shown in
However, the downside of a push model architecture is that broadcasters 34 must be aware of the network architecture, such as the number and locations of distribution points 26. Therefore, each time a distribution point 26 is added, the broadcast configuration is required to change, to make outbound connections to the new distribution point 26. Furthermore, a push model architecture which includes fail over and/or load balancing becomes even more complex, and requires frequent reconfiguration of both the distribution points 28 and the broadcaster hosts 12.
The stream sourcing content delivery system 12a shown in
In the “pull” model architecture 40, it is the responsibility of operations personnel to craft the network architecture as needed, whereby the majority of the network architecture is controlled by the distribution points 26.
As seen in
The stream sourcing content delivery system 12a shown in
Song Selection Models—“Plan Ahead” vs. “Just In Time” Song Selection. The stream sourcing content delivery system 12 can be configured for a wide variety of song selection models, such as “Just in Time” song selection, or “Plan Ahead” song selection.
A “Just-in-Time” song selection model requires that the song selection process verify the existence of the file 21 on disk just before it is ready to be played. In some “Just-in-Time” song selection model embodiments, tracks are typically scheduled three tracks in advance. Some embodiments of the stream sourcing content delivery system 12 comprise Just-in-Time song selection, such as to decrease the chance of violating DMCA rules, and/or to maximize the chance that content 21 is available on the system disk 25.
Since song verification and access can be an intensive and time-sensitive process, which can be disrupted with the failure of multiple parts of the system, some system embodiments 12 preferably comprise a “Plan Ahead” song selection model, in which song tracks 21 for each station 30 are scheduled far in advance, and in which the local content cache 24 is populated with an extended playlist 18 of songs 21. A “Plan Ahead” song selection model gives the broadcaster 34 an opportunity to plan ahead and pre-fetch the tracks 21 that the system 12 needs for the foreseeable future. A “Plan Ahead” song selection model also allows the caching of content 21 on the system 12, so that in the event of a failure of the database 14 and content store 20, the system 12 has sufficient content 21 to loop 200 (
System Performance and Scalability. The operating system of the stream sourcing content delivery system 12 manages the retrieval of schedules playlists 18 and content 21, the production of content streams 28, and the loop 200 of content as needed. Therefore, the system input and output (IO) hardware, comprising the network 11 and disk 25, is not the limiting factor in the performance of the system 12, since the performance of the stream sourcing content delivery system 12 is not limited by the overhead of the process. Therefore, the stream sourcing content delivery system 12 is readily scaled to meet the needs of a large number of streams per host, e.g. as much as 150 or more streams per host system 12, and/or as many as or more than 500 listeners or relays per host system 12.
While the stream sourcing content delivery system 12, is readily adapted for a wide variety of operating environments, current system embodiments typically comprise the following features:
Database Management.
The system 12 periodically wakes up 52, e.g. such as every five minutes. Upon waking 52, the system 12 logs 54 into the database 14, such as within a user/password/database format, e.g. via a PRO*C daemon.
The stream sourcing content delivery system 12 then performs a query 56 of how many total streams 28 that the system 12 is required to source, in order to allocate memory for the stream structures 28. The system 12 queries 56 the database 14 for the current number of station identities (SIDs) 30, and determines 58 if there are more results. Once the number of streams 28 has been determined, the stream sourcing content delivery system 12 allocates the appropriate space, and continues.
If there are no more results 60, the process 50 is finished for that period, and begins 62 another sleep period, and then returns 64 to wake up 52. If there are 66 more results, a determination 68 is made whether the result is a new SID 30, at step 68. If the SID determination 68 is negative 70, i.e. the results are not a new SID 30, the new playlist items are fetched 72, not including what has been previously fetched. If the SID determination 68 is positive 76, i.e. the results correspond a new SID 30, the SID configuration for the new SID 30 is retrieved 78, and the playlist 18 is fetched 80, typically starting at the current time and extending for a time period, e.g. 5 minutes. The system 12 advances 74 to the next result in the result set, and returns to the result step 58, to repeatedly process any other results, before sleeping 62.
As seen in
Retrieval of Stream Configurations. The stream sourcing content delivery system 12 then retrieves the details for each stream 28 it will source. The system 12 compares the result set to the list of streams 28 it currently has, and adds any new streams 28 to the list.
Retrieval of Playlists. For each stream configuration received in the previous step, the database thread queries the database 14, and retrieves the corresponding playlist 18 for the stream 28. The system 12 marks each playlist item 21 as “Not Cached”.
Playlist Management.
The periodic playlist management process 90 is therefore repeated for each station ID 30. The periodic playlist management process 90 guarantees that the stream sourcing content delivery system 12 has the maximum schedule for each station 30, such that the system has the greatest chance to fetch the content 21 and to prepare the content stream 28.
Content Management. The stream sourcing content delivery system 12 preferably caches content as far in the future as possible. The stream sourcing content delivery system 12 uses two types of cache management, disk cache management, and memory cache management. As well, the stream sourcing content delivery system 12 typically manages the removal of the content.
Caching On Disk.
If the determination is positive 128, e.g. the content is already cached on disk 25 from another station 30, the system 12 touches 130 the file on the disk 25, i.e. the system finds the content on disk 25, and marks the playlist item as cached. The system 12 then advances 132 to the next playlist item 21, and returns 134 to repeat the process, at step 112.
If the cache determination is negative 116, e.g. the content is not already cached on disk 25 from another station 30, the system fetches 118 the content 21 from the content store 20, touches 120 the file on the disk 25, and marks 122 the playlist item as cached on the disk 25. The system 12 then advances 124 to the next playlist item 21, and returns 126 to repeat the process, at step 112.
The stream sourcing content delivery system 12 periodically analyzes the items 21 in the playlist 18 for each stream 18. If the system 12 sees an item in the playlist 18 that hasn't been marked as cached, the system 12 attempts to cache the content 21. Before the system 12 caches the content 21, the system 12 checks to see if the content 21 is already on disk 25, i.e. the content 21 may already be cached on disk from another station 30. If the system 12 finds the content 21 on disk 25, the system 12 marks the playlist entry as cached. Otherwise, the system 12 fetches the content 21, such as by using a corresponding URL from the database 14 to fetch the content 21 via HTTP. The system 12 then marks the content 21 cached on disk 25.
Caching in Memory.
As seen in
At step 154, a determination is made whether the second track to be played 21 is cached in memory 27. If the determination 154 is positive 162, the system 12 proceeds to analyze 154 the third track 21. If the determination 154 is negative 156, the system 12 reads 158 the file off the disk 25 and caches to memory 27, while incrementing the counter.
At step 164, a determination is made whether the third track to be played 21 is cached in memory 27. If the determination 164 is positive 162, the system 12 sleeps 174 until the next track 21 finishes streaming. If the determination 154 is negative 166, the system 12 reads 168 the file off the disk 25 and caches to memory 27, while incrementing the counter.
Removing Content from Disk. The removal of content 21 from disk 25 is left to operations to manage. Since the stream sourcing content delivery system 12 does a system “touch” 120,130 (
Removing Content from Memory. When the stream sourcing content delivery system 12 finishes streaming a file, the system 12 frees the memory 27 for the track 21. Content 21 is typically cached in memory 27 uniquely for each stream 28. Therefore, there can be an overlap of content 21 between streams 28 in the stream sourcing content delivery system 12.
Stream Management.
As seen in
Starting the Stream. When a stream 28 first starts, unless the system 12 is in a failure condition, the stream sourcing content delivery system 12 attempts to start playing the next track 21 at the scheduled start time for the track 21. This ensures that multiple stream sourcing content delivery instances 12 are synchronized, both with other stream sourcing content delivery instances 12, and with and database 14.
Stream Format. Data is read from the cached files in memory 27, and is preferably encapsulated. The data is then fed into the circular buffer 186 for each stream 18.
Metadata Insertion. Metadata 210 for each track 21 is inserted just before the track data 21 is fed into the buffer 186. Metadata 210 is stored along with scheduled tracks 21 in the database 21. The metadata 210 is preferably formatted within the database 14. The stream sourcing content delivery system 12 retrieves metadata 210, along with the playlist item information 21. At the time that the track 21 will play, the metadata 210 is encapsulated, using the metaclass and metatype from the stream configuration, and the metadata message 210 is added the buffer 186. In some system embodiments 12, “0x901” is used for cached metadata 210.
Relay Functionality. The stream sourcing content delivery system 12 exposes a listen thread, which is responsible for listening for inbound connections. When a connection is established, a relay negotiation occurs, such as in compliance with a relay protocol. Upon a successful negotiation, the file descriptor for the non-blocking socket is added to the listener send list.
Time Management. Some embodiments of the stream sourcing content system 12 require that a client 32 be able to display a time elapsed per song 21. While song-lengths 204, e.g. 204a (
In an exemplary embodiment of the stream sourcing content delivery system 12 which displays time-remaining information 214, the system 12 reads the length 204 of a track 21 as one of the data fields in the playlist fetch. As a song 21 is ready to be streamed, the stream sourcing content delivery system 12 looks at the corresponding time, and creates the following cached metadata message:
After every N seconds, e.g. N=2 seconds, until the end of the track 21, the stream sourcing content delivery system 12 sends the 0x5000 message. However, instead of t=0 in the payload, the stream sourcing content delivery system 12 estimates the amount of time that has elapsed 212 (
Payload=len=<track length in seconds>;t=<time elapsed>. (1)
The frequency of the repeated pass-thru metadata 210 is preferably configurable.
In some preferred system embodiments, the display of elapsed time 212 comprises the following features:
Database is Down on Startup. Most embodiments of the stream sourcing content delivery system 12 keep a time snapshot, e.g. 5 minutes, of station information 30, playlists 18, and metadata 210. After each time period database sequence, e.g. after every 5 minutes, some embodiments of the stream sourcing content delivery system 12 writes a file to disk called FAIL0.bin. FAIL0.bin which contains station information 30, playlists 18, and metadata 210 for all streams 18.
Database and Content Store are Down at Startup. If FAIL0.bin doesn't exist and the database 14 is unavailable, the stream sourcing content delivery system exits.
Database and Content Store Fail for a Short Time. On a database sequence failure, the stream sourcing content delivery system 12 increases the frequency of polling the database 14 to every 30 seconds. For HTTP content grabs, the following applies:
Database & Content Store Fail for an Extended Time. If the database 14 and content store 20 fail for an extended period, the stream sourcing content delivery system 12 typically continues to advance through the playlist 18. If the system 12 reaches the second-to-last playlist item 21, the system 12 goes into a “looping mode” 200, and a log entry is preferably made, to note the required looping operation 200. The first track 21 in the playlist 18 is then cached into memory 27. After each track 21 finishes streaming, the stream sourcing content delivery system 12 checks for new items in the playlist 18, to stop the looping operation 200 if possible, i.e. to resume normal streaming of content 21.
Periodic Synchronization. Some embodiments of the stream sourcing content delivery system 12 comprise a periodic synchronization, such as to compensate for any time drift between playlists 18.
For example, in some system embodiments 12, whereby multiple stream sourcing content delivery processes may drift in their playlists 18 by small amounts over time, e.g. the course of a day, a synchronization may preferably be periodically performed, e.g. daily, to minimize the overall drift.
For example, in a system embodiment 12 which comprises a daily synchronization, the synchronization process is preferably performed at a time which minimizes the disruption of content playback for users, such as at late night or early morning.
For example, in an exemplary daily synchronization methodology, on embodiment of the stream sourcing content delivery system 12 stops and then begins streaming the next track that is scheduled for 5:00AM for a station 30. While such a synchronization could cause a cut in the song 21 that is being listening to, the process ensures that the system servers are synchronization, and would affect only a small group of users.
System Configuration. Some embodiments of the stream sourcing content delivery system 12 allow for the configuration of the following parameters:
Logging.
System Advantages. The stream sourcing content delivery system 12 and associated methods provide significant advantages over existing content delivery and broadcast systems, and provides improvements to the scheduling, caching, and/or playing of content, e.g. songs.
The stream sourcing content delivery system 12 delivers content 21 and metadata 210 to multiple distribution points 26, and is able to broadcast content indefinitely if the database 12 or content store 20 fails. If a connection is lost between the stream sourcing content delivery server 12 and the content store 20, the system 12 goes into a looping behavior 200, whereby the user's experience is uninterrupted. The looping behavior is avoids content blackouts for the user, i.e. the loop 200 is of sufficient length that it is typically not noticeable, and is preferably DMCA compliant.
The stream sourcing content delivery system 12 is also readily scaled to the number of broadcast streams 28a-28k, and allows operations to easily manage the source complex. The stream sourcing content delivery system 12 is readily expanded and scaled for a large number of stations 32, distribution points 26, clients, relays, and/or listeners. A plurality of systems 12 can readily be operated together, and may further comprise load balancing between systems 12.
As well, datastreams within the stream sourcing content delivery system 12 preferably comprise metadata associated with the steam and/or songs, e.g. to create song boundaries, as well as controlled buffering and synchronization.
Preferred embodiments of the stream sourcing content delivery system 12 sends content, on a per file basis, at a bit rate which matches the actual bit rate of reception and use, which avoids either over run or under run of data transfer.
Although the stream sourcing content delivery system and methods of use are described herein in connection with the delivery of music, i.e. songs, the apparatus and techniques can be implemented for a wide variety of electronic content, such as a wide variety of audio content, e.g. songs, dialog, discussion, video content, multimedia content, or any combination thereof, as desired.
Although the stream sourcing content delivery system and methods of use are described herein in connection with personal computers, mobile devices, and other microprocessor-based devices, such as portable digital assistants or network enabled cell phones, the apparatus and techniques can be implemented for a wide variety of electronic devices and systems, or any combination thereof, as desired.
As well, while the stream sourcing content delivery system and methods of use are described herein in connection with interaction between a user terminals and one or more radio station sites across a network such as the Internet, the stream sourcing content delivery system and methods of use can be implemented for a wide variety of electronic devices and networks or any combination thereof, as desired.
Accordingly, although the invention has been described in detail with reference to a particular preferred embodiment, persons possessing ordinary skill in the art to which this invention pertains will appreciate that various modifications and enhancements may be made without departing from the spirit and scope of the claims that follow.
The present application claims priority to U.S. Provisional Application No. 60/433,734, entitled “MUSIC NET,” filed Dec. 13, 2002.
Number | Name | Date | Kind |
---|---|---|---|
5168481 | Culbertson et al. | Dec 1992 | A |
5325238 | Stebbings et al. | Jun 1994 | A |
5410343 | Coddington et al. | Apr 1995 | A |
5517672 | Reussner et al. | May 1996 | A |
5528513 | Vaitzbilt et al. | Jun 1996 | A |
5585866 | Miller et al. | Dec 1996 | A |
5616876 | Cluts | Apr 1997 | A |
5644715 | Baugher | Jul 1997 | A |
5671195 | Lee | Sep 1997 | A |
5715314 | Payne et al. | Feb 1998 | A |
5734119 | France et al. | Mar 1998 | A |
5761417 | Henley et al. | Jun 1998 | A |
5774672 | Funahashi et al. | Jun 1998 | A |
5784597 | Chiu et al. | Jul 1998 | A |
5787482 | Chen et al. | Jul 1998 | A |
5790174 | Richard, III et al. | Aug 1998 | A |
5792971 | Timis et al. | Aug 1998 | A |
5802502 | Gell et al. | Sep 1998 | A |
5819160 | Foldare et al. | Oct 1998 | A |
5892900 | Ginter et al. | Apr 1999 | A |
5907827 | Fang et al. | May 1999 | A |
5910987 | Ginter et al. | Jun 1999 | A |
5913039 | Nakamura | Jun 1999 | A |
5915019 | Ginter et al. | Jun 1999 | A |
5917912 | Ginter et al. | Jun 1999 | A |
5920861 | Hall et al. | Jul 1999 | A |
5930765 | Martin | Jul 1999 | A |
5943422 | Van Wie et al. | Aug 1999 | A |
5944778 | Takeuchi et al. | Aug 1999 | A |
5949876 | Ginter et al. | Sep 1999 | A |
5956321 | Yao et al. | Sep 1999 | A |
5956491 | Marks | Sep 1999 | A |
5959945 | Kleiman | Sep 1999 | A |
5963914 | Skinner et al. | Oct 1999 | A |
5982891 | Ginter et al. | Nov 1999 | A |
5991867 | Fosmark | Nov 1999 | A |
5996015 | Day et al. | Nov 1999 | A |
6029257 | Palmer | Feb 2000 | A |
6031797 | Van Ryzin et al. | Feb 2000 | A |
6041354 | Biliris et al. | Mar 2000 | A |
6044398 | Marullo et al. | Mar 2000 | A |
6061722 | Lipa et al. | May 2000 | A |
6067562 | Goldman | May 2000 | A |
6088722 | Herz | Jul 2000 | A |
6112023 | Dave et al. | Aug 2000 | A |
6112181 | Shear et al. | Aug 2000 | A |
6138119 | Hall et al. | Oct 2000 | A |
6157721 | Shear et al. | Dec 2000 | A |
6157940 | Marullo et al. | Dec 2000 | A |
6160812 | Bauman et al. | Dec 2000 | A |
6163683 | Dunn et al. | Dec 2000 | A |
6173325 | Kukreja | Jan 2001 | B1 |
6185683 | Ginter et al. | Feb 2001 | B1 |
6185701 | Marullo et al. | Feb 2001 | B1 |
6192340 | Abecassis | Feb 2001 | B1 |
6195701 | Kaiserworth et al. | Feb 2001 | B1 |
6199076 | Logan et al. | Mar 2001 | B1 |
6222530 | Sequiera | Apr 2001 | B1 |
6226672 | DeMartin et al. | May 2001 | B1 |
6237786 | Ginter et al. | May 2001 | B1 |
6240185 | Van Wie et al. | May 2001 | B1 |
6243328 | Fenner et al. | Jun 2001 | B1 |
6243725 | Hempleman et al. | Jun 2001 | B1 |
6247061 | Douceir | Jun 2001 | B1 |
6248946 | Dwek | Jun 2001 | B1 |
6253193 | Ginter et al. | Jun 2001 | B1 |
6262569 | Carr et al. | Jul 2001 | B1 |
6263313 | Milsted et al. | Jul 2001 | B1 |
6263362 | Donoho et al. | Jul 2001 | B1 |
6266788 | Othmer et al. | Jul 2001 | B1 |
6300880 | Sitnik | Oct 2001 | B1 |
6314576 | Asamizuya et al. | Nov 2001 | B1 |
6332163 | Bowman-Amuah | Dec 2001 | B1 |
6356936 | Donoho et al. | Mar 2002 | B1 |
6363488 | Ginter et al. | Mar 2002 | B1 |
6366914 | Stern | Apr 2002 | B1 |
6389402 | Ginter et al. | May 2002 | B1 |
6421651 | Tedesco et al. | Jul 2002 | B1 |
6427140 | Ginter et al. | Jul 2002 | B1 |
6430537 | Tedesco et al. | Aug 2002 | B1 |
6434621 | Pezzillo et al. | Aug 2002 | B1 |
6434628 | Bowman-Amuah | Aug 2002 | B1 |
6438450 | DiLorenzo | Aug 2002 | B1 |
6438630 | DeMoney | Aug 2002 | B1 |
6441832 | Tao et al. | Aug 2002 | B1 |
6446080 | Van Ryzin et al. | Sep 2002 | B1 |
6446125 | Huang et al. | Sep 2002 | B1 |
6446126 | Huang et al. | Sep 2002 | B1 |
6449367 | Van Wie et al. | Sep 2002 | B2 |
6453316 | Kairbe et al. | Sep 2002 | B1 |
6477541 | Korst et al. | Nov 2002 | B1 |
6477707 | King et al. | Nov 2002 | B1 |
6492469 | Willis et al. | Dec 2002 | B2 |
6496744 | Cook | Dec 2002 | B1 |
6502194 | Berman et al. | Dec 2002 | B1 |
6505160 | Levy et al. | Jan 2003 | B1 |
6519648 | Eyal | Feb 2003 | B1 |
6526411 | Ward | Feb 2003 | B1 |
6529586 | Elvins et al. | Mar 2003 | B1 |
6536037 | Guheen et al. | Mar 2003 | B1 |
6542445 | Ijichi et al. | Apr 2003 | B2 |
6546397 | Rempell | Apr 2003 | B1 |
6550057 | Bowman-Amuah | Apr 2003 | B1 |
6601041 | Brown et al. | Jul 2003 | B1 |
6609097 | Costello et al. | Aug 2003 | B2 |
6618424 | Yamada et al. | Sep 2003 | B1 |
6618484 | Van Wie et al. | Sep 2003 | B1 |
6658568 | Ginter et al. | Dec 2003 | B1 |
6668325 | Collberg et al. | Dec 2003 | B1 |
6725275 | Eyal | Apr 2004 | B2 |
6772340 | Peinado et al. | Aug 2004 | B1 |
6772435 | Thexton et al. | Aug 2004 | B1 |
6910220 | Hickey et al. | Jun 2005 | B2 |
6950623 | Brown et al. | Sep 2005 | B2 |
7020710 | Weber et al. | Mar 2006 | B2 |
7020893 | Connelly | Mar 2006 | B2 |
7024485 | Dunning et al. | Apr 2006 | B2 |
7043479 | Ireton | May 2006 | B2 |
7136906 | Giacalone, Jr. | Nov 2006 | B2 |
7185352 | Halford et al. | Feb 2007 | B2 |
7222354 | Ching et al. | May 2007 | B1 |
7412532 | Gondhalekar et al. | Aug 2008 | B2 |
7493289 | Verosub et al. | Feb 2009 | B2 |
20010003828 | Peterson et al. | Jun 2001 | A1 |
20010030660 | Zainoulline | Oct 2001 | A1 |
20010044851 | Rothman et al. | Nov 2001 | A1 |
20020032907 | Daneils | Mar 2002 | A1 |
20020059237 | Kumagai et al. | May 2002 | A1 |
20020059624 | Machida et al. | May 2002 | A1 |
20020068525 | Brown et al. | Jun 2002 | A1 |
20020078056 | Hunt et al. | Jun 2002 | A1 |
20020082914 | Beyda et al. | Jun 2002 | A1 |
20020091761 | Lambert | Jul 2002 | A1 |
20020095510 | Sie et al. | Jul 2002 | A1 |
20020104099 | Novak | Aug 2002 | A1 |
20020107968 | Horn et al. | Aug 2002 | A1 |
20020108395 | Fujita et al. | Aug 2002 | A1 |
20020152876 | Hughes et al. | Oct 2002 | A1 |
20020152878 | Akashi | Oct 2002 | A1 |
20020158895 | Murase et al. | Oct 2002 | A1 |
20020198846 | Lao | Dec 2002 | A1 |
20030014436 | Spencer et al. | Jan 2003 | A1 |
20030018797 | Dunning et al. | Jan 2003 | A1 |
20030023973 | Monson et al. | Jan 2003 | A1 |
20030023975 | Schrader et al. | Jan 2003 | A1 |
20030028659 | Mesarina et al. | Feb 2003 | A1 |
20030028893 | Addington | Feb 2003 | A1 |
20030048418 | Hose et al. | Mar 2003 | A1 |
20030069768 | Hoffman et al. | Apr 2003 | A1 |
20030069854 | Hsu et al. | Apr 2003 | A1 |
20030121050 | Kalva et al. | Jun 2003 | A1 |
20030126275 | Mungavan et al. | Jul 2003 | A1 |
20030135605 | Pendakur | Jul 2003 | A1 |
20030195974 | Ronning et al. | Oct 2003 | A1 |
20030236906 | Klemets et al. | Dec 2003 | A1 |
20040064507 | Sakata | Apr 2004 | A1 |
20040138948 | Loomis | Jul 2004 | A1 |
20040177115 | Hollander et al. | Sep 2004 | A1 |
20040222047 | DiFranza et al. | Nov 2004 | A1 |
20050056494 | Amo et al. | Mar 2005 | A1 |
20050114529 | Lambert | May 2005 | A1 |
20050114757 | Sahota et al. | May 2005 | A1 |
20050159104 | Valley et al. | Jul 2005 | A1 |
20060155400 | Loomis | Jul 2006 | A1 |
20090164794 | Verosub et al. | Jun 2009 | A1 |
20090175591 | Gondhalekar et al. | Jul 2009 | A1 |
Number | Date | Country |
---|---|---|
1113605 | Jul 1991 | EP |
0831608 | Mar 1998 | EP |
0875846 | Nov 1998 | EP |
0984584 | Mar 2000 | EP |
0986046 | Mar 2000 | EP |
1 178 487 | Feb 2002 | EP |
1178487 | Feb 2002 | EP |
1178487 | Feb 2002 | EP |
1 187 423 | Mar 2002 | EP |
1187423 | Mar 2002 | EP |
1187423 | Mar 2002 | EP |
1 229 476 | Aug 2002 | EP |
1229476 | Aug 2002 | EP |
1229476 | Aug 2002 | EP |
1 244 021 | Sep 2002 | EP |
1244021 | Sep 2002 | EP |
1244021 | Sep 2002 | EP |
1 267 247 | Dec 2002 | EP |
1267247 | Dec 2002 | EP |
1267247 | Dec 2002 | EP |
1 286 351 | Feb 2003 | EP |
1286351 | Feb 2003 | EP |
1286351 | Feb 2003 | EP |
1187485 | Apr 2003 | EP |
2002108395 | Apr 2002 | JP |
2002318587 | Oct 2002 | JP |
2003068968 | Mar 2003 | JP |
2003069768 | Mar 2003 | JP |
497055 | Aug 2002 | TW |
WO 0110496 | Feb 2001 | WO |
WO 02063414 | Aug 2002 | WO |
WO-02063414 | Aug 2002 | WO |
Number | Date | Country | |
---|---|---|---|
20040186733 A1 | Sep 2004 | US |
Number | Date | Country | |
---|---|---|---|
60433734 | Dec 2002 | US |