Handling long-tail content in a content delivery network (CDN)

Information

  • Patent Grant
  • 10218806
  • Patent Number
    10,218,806
  • Date Filed
    Monday, July 10, 2017
    7 years ago
  • Date Issued
    Tuesday, February 26, 2019
    5 years ago
Abstract
A content delivery network has at least a first tier of servers. A content delivery method includes, at a first server in the first tier of servers, obtaining a request from a client for a resource. If the resource is available at the first server or at a peer of the first server, then the resource is served to the client from the first server. Otherwise, it is determined whether the resource is popular, and if the resource is determined to be popular, then the first server obtains the resource and the first server serves the resource to the client. If the resource is determined not to be popular, the client is directed to a second server, not in the first tier of servers, and the second server serves the resource to the client. The second server may be in a second tier of servers or it may be an origin server.
Description
FIELD OF THE INVENTION

This invention relates to content delivery, to content delivery networks (CDNs), and to frameworks and systems using CDNs.


DETAILED DESCRIPTION
Glossary

As used herein, unless stated otherwise, the following terms or abbreviations have the following meanings:

    • 1. IP means Internet Protocol.
    • 2. “IP address” means an address used in the Internet Protocol to identify electronic devices such as servers and the like.
    • 3. HTTP means Hypertext Transfer Protocol.
    • 4. URL means Uniform Resource Locator.
    • 5. DNS means Domain Name System.





BRIEF DESCRIPTION OF THE DRAWINGS

The present invention can be better understood with reference to the following drawings. The elements of the drawings are not necessarily to scale, emphasis instead being placed upon clearly illustrating the principles of the present invention. Furthermore, like referenced numerals designate corresponding parts throughout the several views.



FIG. 1 depicts a general hierarchical, multi-tiered content delivery network (CDN);



FIG. 2 shows the logical organization of servers into groups or clusters in a CDN;



FIG. 3 depicts a content delivery framework (CDF) using a content delivery network (CDN);



FIG. 4 depicts the operation of a two-level content delivery network in a content delivery framework;



FIG. 5 is a flowchart showing operation of the popularity services of the CDN of FIG. 4;



FIG. 6 is an exemplary data structure for maintaining the popularity data for a particular popularity server; and



FIG. 7 is a flowchart showing processing steps of a multi-tiered content delivery network.





OVERVIEW

The Internet and the so-called World Wide Web (the “WWW”) have become ubiquitous. Thousands or even tens of thousands of so-called content providers (publishers) now use the Internet (and, particularly, the WWW) to provide all sorts of content to tens or even hundreds of thousands of clients all over the world.


In order to offload the job of serving some or all of their content, many content providers now subscribe to so-called content delivery networks (CDNs). Using a CDN, some (or all) of a content provider's content can be served to clients from the CDN (i.e., from one or more servers in the CDN) instead of from the content provider's server(s). In a caching CDN, content that is served may also be cached on some or all of the CDN servers, either before being served or in response to specific requests for that content.


The term content as used herein means any kind of data, in any form, regardless of its representation and regardless of what it represents. Content may include, without limitation, static and/or dynamic images, text, audio content, including streamed audio, video content, including streamed video, web pages, computer programs, documents, files, and the like. Some content may be embedded in other content, e.g., using markup languages such as HTML and XML. Content includes content which is created or formed or composed specifically in response to a particular request. The term “resource” is sometimes used herein to refer to content.


Certain publishers can have large content libraries in which only a small proportion of the content (the so-called “short head”) is popular enough to benefit from serving through a caching CDN, while the majority of the content (the so-called “long tail”) is accessed only occasionally and not generally worth caching (or even serving from an edge server). This situation would be typical for a content publisher with a very large music or video library. Some music content—the popular content—may be regularly requested, whereas other music—the not popular (also referred to as unpopular) content—may be seldom if ever requested.


Content can become popular (by various measures of popularity) or fade into relative obscurity dynamically, so a content library cannot easily be explicitly partitioned. Instead, the CDN tracks popularity of certain content, and selectively migrates content toward the edge (i.e., toward the tier 1 servers) as that content becomes popular.


A CDN may have one or more tiers of servers, organized hierarchically. FIG. 1 depicts a content delivery network 100 that includes multiple tiers of servers. Specifically, the CDN 100 of FIG. 1 shows j tiers of servers, denoted Tier 1, Tier 2, Tier 3, . . . , Tier j. Each tier of servers may comprise a number of servers organized into server groups (sometimes referred to as server clusters). The Tier 1 servers are also referred to as edge servers, and Tier 1 is sometimes also referred to as the “edge” or the “edge of the CDN.” The Tier 2 servers (when present in a CDN) are also referred to as parent servers.


Furthermore, content stored in any of the intermediate tier servers (e.g., tier 2, tier 3, . . . tier j in FIG. 1) may be logically partitioned across two or more servers in the respective tier group. For example, a first portion (or subset) of a content provider's library may be stored on a first parent server in Group 1, a second portion may be stored on a second server in Group 1, and the last third may be stored on a third server in Group 1. Various partitioning methodologies are described in further detail below. In operation, the partitioning of content libraries across servers in an intermediate tier of a CDN provides a type of preemptive load-balancing such that certain parent servers are only responsible for handling a pre-defined subset of a content provider's library. In other words, redirected requests from the client to a parent/intermediate tier are distributed within the tier in accordance with the particular partitioning methodology being implemented.


For example, in the CDN 100 of FIG. 1, Tier 1 has n groups of servers (denoted “Edge Server Group 1”, “Edge Server Group 2”, . . . , “Edge Server Group n”); tier 2 (the parent servers' tier) has m server groups (the i-th group being denoted “Parent Server Group i”); and tier 3 has k server groups, and so on. Preferably each tier has the same number of server groups.



FIG. 2 shows the logical organization/grouping of servers in a CDN of FIG. 1. In the exemplary CDN of FIG. 2, each tier of servers has the same number (n) of server groups. Those of skill in the art will know and understand, upon reading this description, that each server group may have the same or a different number of servers. Additionally, the number of servers in a server group may vary dynamically. For example, additional servers may be added to a server group to deal with increased load on the group.


The servers in a server group may be homogenous or heterogeneous, and each server in a server group may comprise a cluster of physical servers sharing the same name and/or network address. An example of such a cluster is described in and co-owned patent application No. 61/064,339 (titled “Load-balancing cluster”, filed Feb. 28, 2008), the entire contents of which are incorporated herein by reference for all purposes.


Servers in the same tier and the same group are referred to as peers or peer servers.


A typical CDN has only one or two tiers of servers. A CDN with only one tier will have only edge servers, whereas a CDN with two tiers will have edge servers and parent servers. (At a minimum, a CDN should have at least one tier of servers—the edge servers.)


The grouping of servers in a tier may be based, e.g., on their physical or geographical location. For example, a particular CDN may have six groups—four groups of servers in the United States, group 1 for the West Coast, group 2 for the mid-west, Group 3 for the northeast and Group 4 for the south east; and one group each for Europe and Asia.


A particular CDN server is preferably in only one server group.


In general, some or all of the servers in each tier can exchange data with some or all of the servers in each other tier. Thus, some or all of the parent servers can exchange information with some or all of the edge servers. For the sake of simplicity, in the drawings, each tier of servers is shown as being operationally connectable to each other tier. In some CDNs, however, it may be preferable that the servers in a particular tier can only exchange information with other servers in the same group (i.e., with peer servers) and/or with other servers in the same group in a different tier. For example, in some CDNs, the edge servers in edge server group k, can exchange information with each other and with all servers in parent server group k, and so on.


A content provider's/customer's server (or servers) are also referred to as origin servers. A content provider's origin servers may be owned and/or operated by that content provider or they may be servers provided and/or operated by a third party such as a hosting provider. The hosting provider for a particular content provider may also provide CDN services to that content provider.


A CDN may also include a CDN origin/content cache tier which may be used to cache content from the CDN's subscribers (i.e., from the CDN subscribers' respective origin servers). Those of skill in the art will know and understand, upon reading this description, that a CDN can support one or more subscribers, i.e., that a CDN can function as a shared infrastructure supporting numerous subscribers. The CDN origin tier may also consist of a number of servers, and these servers may also be organized (physically and logically) into a number of regions and/or groups. The server(s) in the CDN origin tier obtain content from the subscribers' origin servers, either on an as needed basis (a pull) or in advance (via a push).


A popularity service 102 (described in greater detail below) is associated with one or more of the server groups in one or more tiers. In a presently preferred exemplary embodiment, some of the parent server groups have popularity services 102 associated therewith. Although shown as a separate component of the group, the popularity service 102 may be integrated into one of the servers in the group. In some cases, the popularity service may have its own server, distinct from any of the CDN servers. The terms “popularity service” and “popularity server” are used interchangeable herein.


In operation, when a client requests content that is to be served using a content delivery framework, the client may be served that content from a server in the CDN or, in some cases, from the subscriber/customer's origin server.


A client may be directed to a CDN and/or to a server in the CDN in any manner using any kind of server selector system 104. As understood by those of skill in the art, the server selector system 104 generally operates to direct a client's requests for content to an appropriate server in order for that content to be served to the requesting client. An appropriate server may be one which is close to the client (by some measure of cost) and/or one which is not too heavily loaded. All sorts of conditions may be applied to the term “appropriate”, and all sorts of information and tests, both static and dynamic, may be used to determine an appropriate server. The server selector system 106 may, e.g., include or operate fully or partially in Domain Name Service (DNS) servers, standalone devices, or a combination thereof. For example, the server selector system 106 may comprise a single level DNS server that selects an appropriate server based at least in part on some combination of the location of the requesting client and the load on some or all of the CDN servers. Those of skill in the art will know and understand, upon reading this description, that a client's location in a network such as the Internet may sometimes only be roughly determined, and the term “location of the client” is generally taken to be a network location corresponding to the client's network service provider.


Although shown as a component in the drawings, the server selector 106 may comprise numerous components. For example, some or all of the server selection may be based on anycast routing, and the server selector 106 may then include routers and associated tables.


In a presently preferred embodiment the server selector 106 is an intelligent traffic manager (ITM)/adaptive traffic controller (ATC) such as described in co-pending U.S. patent application Ser. No. 10/259,497, filed Sep. 30, 2002, and titled “Configurable Adaptive Global Traffic Control And Management,” (published as US 2003-0065762 A1); and in U.S. patent application Ser. No. 11/976,648, filed Oct. 26, 2007, titled “Policy-based content delivery network selection,” (collectively the “ITM applications”), the entire contents of each of which have been incorporated herein by reference for all purposes. In some embodiments the server selector 106 may include a “best” or “optimal” server selector such as disclosed in U.S. Pat. No. 6,185,598 titled, “Optimized Network Resource Location,” the entire contents of which are incorporated herein by reference for all purposes. The '598 patent refers to CDN servers as so-called repeater servers, and describes a so-called “Best Repeater Selector (BRS) mechanism”.



FIG. 3 shows a content delivery framework 300 with a two-level hierarchical CDN consisting of a tier of edge servers (Tier 1) and a tier of parent servers (Tier 2). Some or all of the edge servers can communicate with some or all of the parent servers. The edge servers are divided into n edge server groups, and the parent servers are divided into m parent server groups. In a presently preferred embodiment, the value of m is equal to the value of n, i.e., in this presently preferred embodiment there are the same number of edge server groups as there are parent server groups. A CDN origin/content cache tier stores subscriber content which it obtains from the various subscribers' origin servers. At least one of the parent server groups (in the drawing, Group 1) has a popularity service 102-1 associated therewith. Preferably more than one parent server group has an associated popularity service, and more preferably each parent server group has an associated popularity service.


Although shown in a parent tier, the popularity service may located anywhere in the system, including in the edge tier.


The popularity service may be used by certain, though not necessarily all, content. When only certain content uses the popularity service, content should be designated in order to use the popularity service.


Some or all of the edge servers in a group may use a popularity service to manage the long-tail content of various subscribers. Each edge server that uses a popularity service is referred to as being bound to that popularity service. An edge server that is bound to a popularity service is sometimes referred to herein as a “longtail coserver.”



FIGS. 4 and 5 show the operation of the popularity services of the CDN of FIG. 3.


When a client 106 requests content (e.g., using an HTTP GET request), that request is directed (e.g., by server selector 104-1) to an edge server in order for the content to be served to the client. For certain designated content a popularity check is interposed into the fill side of the caching operation. FIG. 4 illustrates the flow of messages and data in the content delivery framework 300, and FIG. 5 is a flowchart show operation of the popularity services of the CDN of FIG. 4. For the sake of this particular explanation, assume that the client's request has been directed to edge server 108. (Those of skill in the art will know and understand, upon reading this description, that the client's initial request may be directed to any tier in the CDN hierarchy, including, e.g., to the parent tier.) This server is selected using the server selection mechanisms 104-1 associated with the CDN, e.g., using one or more DNS servers and selecting an edge server based on such factors as the location of the requesting client, the load on the network, network traffic conditions, CDN policies, subscriber policies, and the like.


A client 104 requests content from an edge server 108 (at 500 in FIG. 5). The request from the client 104 arrives at edge server 108 (S1 in FIG. 4). The edge server 108 checks to see if object is present (locally or on a peer) and fresh (at 502). If so, the edge server 108 serves the object to the client 104 from the cache (S2, 504), obtaining the object from a peer, if necessary.


In some embodiments, the system may distinguish between on-net and off-net peers and same-switch peers. An on-net peer is a peer on the same backbone network; an off-net peer is a peer located on a different backbone network; and a same-switch peer is a peer directly connected to the same switch as the agent performing the check. In some embodiments, the edge server 108 may only look for the object on some of its peers (e.g., only on same-switch peers) (at 502).


If the object is not available on the edge server 108 or on a peer, the edge server 108 ascertains whether this object is served based on its popularity (i.e., whether this object has been designated so that the object's popularity will be used to determine where it will be served from) (at 506). If so, then the request is sent to the popularity service 102 associated with the edge server 108, in this case, to the popularity server for the this group (S3a).


The determination as to whether this object is designated to be served from a different location, depending on its popularity (at 506), may be made based, at least in part, on the name (hostname) used to request the object.


It is preferably to allow for a mix of edge servers, some performing popularity checking (as described above), while others do not. For those that are not running the popularity service, the name (hostname) used to request an object will resolve to a parent server (that may or may not provide popularity services). If the parent server does not provide popularity services, then the content will be obtained by the edge server from that parent server, and the content will be served to the client. On the other hand, if that parent server does provide popularity services, it can determine whether or not the edge server is a Longtail coserver based, e.g., on the IP (Internet Protocol) address of the edge server. For no-coservers, the parent server can handle the request without any popularity processing.


A request for content may be an initial request for an object or it may be a request for another part of an object, the initial part having already been served to the client. If the request is for the first part of the object (at 508), e.g., the request includes a request for the first byte of the resource (i.e., it is not a range request that starts after the beginning of the file), the popularity service 102 determines (as described below) if the object is currently popular. First, the popularity count for the current period is incremented (at 510). Based on its determination, the popularity service 102 returns one of three possible responses to the edge server 108 (S3b):

    • 1. If the object has not reached a first/minimal level of popularity (at 512): the popularity service sends the edge server an instruction (e.g., HTTP 302) to redirect the client's request to the origin server (or to the CDN origin cache) (if origin redirects are enabled) (at 514).
    • 2. If the object's popularity has exceeded the first/minimal level of popularity but has not yet exceeded a second, mid-tier threshold (at 516): the popularity service sends the edge server an instruction (e.g., HTTP 302) to redirect the client's request to a parent server (if mid-tier redirects are enabled) (at 518). Note that the redirect processing may further involve a determination of which particular parent server is responsible for storing the requested content. Such a determination would be based on how the distinct subsets of a content provider's library are partitioned across parent servers in the same intermediate tier. Further note that it is also possible for subsets of a content provider's library to be partitioned across servers residing in different intermediate tiers of a content delivery network.
    • 3. If the object's popularity has exceeded the mid-tier threshold (i.e., the object is popular): The popularity service sends the edge server an instruction to serve the content itself (at 520). In a presently preferred implementation, the popularity service sends the edge server a redirect (HTTP 302) with a “follow me” flag set, to the origin server or, if there is one, to the parent tier.


If the edge server 108 receives a redirect from the popularity service 102 without the “follow me” flag set (cases 1 and 2 above), it simply forwards the redirect to the client 104 (S4a, 522, 524). If the edge server 108 receives a “follow me” redirect, it obtains and caches the resource (at 526) and serves it to the client (at 528).


If the popularity service 102 is unreachable, unresponsive, or returns a status code indicating an error (other than HTTP 404), the object is served out of the edge's cache server (and an alert condition is raised).


Once content has been cached at an edge server, the edge server will send notifications (e.g., in the form of revalidations) to the popularity service every time it gets another request for that content. E.g., with reference to the flowchart of FIG. 5, if the edge server 108 determines (at 502) that it has the requested content (or can obtain it from a peer), then, in addition to serving the content (at 504), it also instructs the popularity server to increment the object's popularity count for the current period (at 530). This process keeps the popularity servers up to date on the relative popularity of content being served in their region.


In presently preferred embodiments, the server selection mechanism 104 does not rendezvous clients to parent servers/caches. In other words, in these embodiments, client requests are always initially directed by the server selection mechanism to an edge server. In these cases, when a request for a resource arrives at a parent server/cache, that request should preferably be served (and filled if necessary) unconditionally (since any request from a client is assumed to be the result of a redirect served by an edge server).


In an embodiment where the server selector 104 can direct client requests directly to parent servers (or to any tier other than the edge tier), a server obtaining a client request may choose to redirect that request, e.g., based on popularity. However, those of skill in the art will know and understand, upon reading this description, that it is advisable to track the redirection of a request to avoid circular and/or infinite redirection. One way to avoid such a problem is to limit the number of levels of redirection (i.e., to limit the number of redirects to follow). In a presently preferred implementation, if no final server is selected after following, e.g., thirty two redirects, an error is issued. In some embodiments, if no final sever is selected after a predefined number of redirects, then the last server reached may be used to serve the content. One way to prevent looping is to use different server names (aliases) or IP addresses when redirecting requests so that a server receiving a request can tell whether or not it is a redirect. Those of skill in the art will know and understand, upon reading this description, that information can be transferred between servers using, e.g., HTTP headers or the like.


Those of skill in the art will know and understand, upon reading this description, that in a multi-tier CDN, the popularity service may be located at any tier, or there may be popularity services at more than one tier.


The middle (Parent) tier is optional.


Step (4a) may reply with content (if popular), or with a redirect to a parent or origin server (if not), in which the client will make another request (5a or 5b) to that tier to obtain the content.


If the request is an HTTP GET request or the like, it is forwarded to the popularity service. HTTP POST requests should always be forwarded directly to the origin, since that is where they will need to be processed, and the response to a POST request should not be cached. It may sometimes be preferable to direct GET requests to a different origin server than POST requests.


While the invention has been described with reference to the HTTP protocol, those of skill in the art will know and understand, upon reading this description, that different and/or other protocols may be used and are contemplated by the inventors. HTTP is described in various documents, e.g., Hypertext Transfer Protocol—HTTP/1.1, RFC 2616, Network Working Group, the entire contents of which are incorporated herein by reference.


Those of skill in the art will know and understand, upon reading this description, that different thresholds may be established for each tier in the CDN. Further, those of skill in the art will know and understand, upon reading this description, that each content item may have its own thresholds associated therewith. In this manner, the system can check all content for popularity, with the default thresholds being zero. In this manner, every request will automatically cause the popularity to exceed the threshold and will cause the content to be cached.


By positioning Popularity Servers regionally (paired with parent cache servers), popularity and cache tiers can be managed independently, on a regional basis. Content that is popular in one region/group may not be popular in another region/group (especially if each region/group corresponds to a geographic and/or political region).


We consider it desirable that rendezvous to popularity servers prioritize so-called “regional” proximity, so that clients within the same region will tend to cast their popularity “votes” within that region and get consistent treatment of popular resources. However, if there are multiple parent cache servers available, there will generally be no attempt to rendezvous particular clients to particular parents.


Defining & Measuring Popularity


In preferred embodiments, popularity of an object/resource is measured based on the number of times that object/resource is requested in various time periods. FIG. 6 is an exemplary data structure for maintaining the popularity data for a particular popularity server. The data structure 600 in FIG. 6 is a so-called tally hash structure.


In preferred embodiments, some or all edge servers are associated with (or bound to) popularity servers. An edge server that is bound to a popularity server is sometimes referred to as a bound Longtail coserver. Each popularity server in the system allocates a tally hash structure 800 per bound Longtail coserver. A configuration provides the number of resource (hash) slots to allocate. For a presently preferred implementation, the number of hash slots is on the order of 100 million slots per coserver. Each slot is divided into a number of time buckets, preferably 16 time buckets, each bucket being represented by, e.g., a 4-bit unsigned integer. Those of skill in the art will know and understand, upon reading this description, that the selection of the size of the value in each time bucket depends on policy decisions about bounds for the popularity thresholds, and for keeping very popular resources at the edge. The size, however, is heavily influenced by a need for compactness. One 8-byte word can store all time buckets for one resource slot, and therefore, 800 MB would be required per property, and five to eight such properties could be managed per popularity server without paging.


Each time bucket represents a time period, preferably a number of seconds.


The mapping of requests/content to slots is based on some function of the object name and perhaps other information associated with the request for the object. Preferably the mapping of objects to slots is based on a hash or message digest function (such as MAD or the like) over the object name (and preferably including some parts of the query string). Each slot may therefore represent one or more resources. Each time a query/request arrives at a popularity server for an object, the hash is computed and the slot in the table 800 (for the appropriate co-server) is determined, and the counts in that slot are used. In event of a hash collision, it is therefore possible that one bucket will be receiving and representing counts for more than one object. Since this result is generally undesirable (since it could result in cache fills and edge caching of unpopular objects), the number of buckets should be chosen to be as large as practical.


Those of skill in the art will know and understand, upon reading this description, that different and/or other data structures may be used to implement the popularity counting. For example, since in most cases the total number of resources is expected to far exceed the number of popular resources, a balanced b-tree may be preferable to a hash table. In addition, it is possible to reduce the size of the hash slot by using only some part of the hash. However, reducing the number of bytes of the hash used can result in more name collisions.


Although described above with respect to popularity, those of skill in the art will know and understand, upon reading this description, that other factors may be used along with (or instead of) popularity to determine whether or not to redirect requests. A rule base may be used to augment and/or override the popularity measures for certain resources. The rules in the rule base may be static or dynamic and may be set by the CDN administrator and/or the subscriber. For example, a subscriber may not want to pay for certain content to be served from the edge, regardless of its popularity, and may set a rule accordingly (this particular result could also be achieved by setting the thresholds for that particular content to prevent it from ever being cached at the edge).


Occasional log mining could be used to look for hash collisions in actual subscriber content libraries, and the hash function and bucket sizes could be tuned as needed.


At each time bucket boundary, the popularity service will logically “rotate” the buckets and zero out the oldest tally data for each object.


Whenever a coserver's enrollment in the popularity service changes (added or dropped, or perhaps hints changed), the data structures are to be updated.


The popularity of a given object may be determined as a weighted sum of its popularity over successive time periods. More recent time periods may be given higher weights.


In order to determine which content is to be managed by the popularity service, the CDN operator and/or the subscriber may specify:

    • The tiers at which the content will be managed—edge, intermediate (i.e., parent), or origin (subscriber's or storage tier). In order to be meaningful, at least one of intermediate and origin service should be enabled.
    • Content that is to be managed based on its popularity, rather than simply always being served from a cache.


There are several reasons why a publisher/subscriber may not want the so-called “long-tail” content served from a caching CDN service, for example:

    • If cache fills are done from the subscriber's origin server and the subscriber pays for cache fill bandwidth, unnecessary fills for unpopular resources increase bandwidth costs without providing any value.
    • Serving an unpopular resource through a CDN cache adds latency (due to the cache fill), and risks forcing actually popular resources out of cache, potentially causing thrashing. The result is lower efficiency and the risk of degraded service.


For related or similar reasons, a CDN provider generally also does not want to serve long-tail content from an edge cache:

    • If cache fills are done from origin storage, unnecessary cache fills consume bandwidth both from the storage and cache systems, increasing (doubling) the internal cost to serve. This lowers efficiency, requiring relatively more bandwidth internally to serve the same content to the outside.
    • The second argument above also applies from the CDN's perspective: a CDN operator wants to minimize latency as well as the risk of thrashing in order to satisfy all of its subscribers.


      Names, Addresses & Configuration Data


As is well known, each server in a network may be addressed by means of one or more network address (e.g., Internet Protocol or IP addresses). Each server in a network may also be known by one or more names (so-called hostnames—fully qualified domain names). Hostnames may be mapped to one or more IP addresses. A hostname may correspond to (and thus resolve to) more than one server.


A system such as ITM (described in the ITM patent applications mentioned above), allows a kind of hostname (called a supername) to refer to multiple servers, and resolves the supername to a nearby server.


Preferably the server selection mechanism is ITM, and each popularity server will have supername that resolves to reach nearby popularity server.


When a popularity server shares or is co-located with a parent server, the parent server may use the name by which it was addressed to determine whether to direct a request to the popularity service. That is, parent cache servers that provide popularity service may recognize requests that use one of the aliases reserved for popularity requests, and call into the popularity service to make the fill/no fill decision and return a redirect as described above.


As noted earlier, if the server selector mechanism does not send initial requests to non-edge servers, then all parent cache servers must recognize requests that have been redirected and serve the requested resource, filling it from the origin (or another tier), if necessary.


A servers hostnames are also referred to as its aliases. Each Longtail coserver preferably has at least two aliases (three if a parent cache/server tier is used): the published supername, the hostname used for popularity service requests, and (if used) the hostname used for parent cache redirects.


Popularity servers will preferably be reached via an ITM supername, and ITM will monitor for the service's availability across the set of servers. Popularity servers should be reached using real IP addresses, and not virtual IPs, and will not necessarily be redundant within a cluster. Redundancy can be provided by having multiple servers per supername. However, preferably there will be no attempt to synchronize the popularity tallies on popularity servers, with the expected desirable effect of managing popularity separately on a “regional” basis, the granularity being determined by the number and distribution of popularity servers deployed. Should a popularity server fail, this could cause a discontinuity in popularity responses as a new server becomes active for a given edge location, but this may be mitigated (for very popular resources) by periodic background refreshes.


Information about the resources and caching strategy include the following:

    • the expected total number of resources associated with this coserver.
    • the number of buckets used to store hit counts for each resource.
    • the number of seconds that each bucket represents. Every time this interval goes by, the count in the oldest bucket is thrown out and a new bucket is started with a count of zero.
    • when the sum of all buckets for a given resource reaches this number on
    • any popularity server, the parent caches (if any) that use that server will start to cache the resource.
    • when the sum of all buckets for a given resource reaches this number on any popularity server, the edge caches that use that server will start to cache the resource.
    • the hash algorithm to apply to the resource names (optional). If not specified, a default algorithm (e.g. MD5) will be used.
    • the maximum number of this coserver's resources that should be in cache at an edge at any given time. This value may be ignored in certain embodiments.
    • the maximum number of this coserver's resources that should be in cache at a parent at any given time. This value may be ignored in certain embodiments.


Preferably resources should not be pre-expired on parent cache servers, as that will cause unnecessary requests to the origin server or queries to peer caches.


Those of skill in the art will know and understand, upon reading this description, that a decision to serve at a given tier that is based only on popularity counts, will not take into account capacity to serve at that tier—so this scheme could overload an origin server or parent tier if they do not have sufficient capacity. Further, if popularity is measured in terms of absolute thresholds on numbers of requests, and if the library is sufficiently large, this could cause cache thrashing at the parent or edge tiers.


Authentication with the origin cache or server, if needed, should be done by the server that receives the initial request from the client. During processing of a redirected request.


Various documents, including patents and patent applications, have been incorporated by reference into this application. In case of any conflict between an incorporated document and the present application, the present application, including any definitions herein, will control.


Thus is provided a feature that allows a CDN to be responsive to increasing or decreasing “popularity” of content by shaping where in the CDN content is positioned and served from.



FIG. 7 is a flowchart 700 of processing steps associated with a multi-tiered content delivery network.


In step 705, a first server in a first tier of servers obtains a request from a client for a resource. In one embodiment, the resource is available as part of a content provider's library (e.g., the resource may be a digital movie requested from an on-line streaming movie provider).


In step 710, if the resource is not available at the first server or at a peer of the first server, and additionally if it is determined that the resource is popular (using processing and functionality previously described), then the first server obtains the resource and serves it to the client. For example, the first server may obtain the resource from an origin server and/or a parent server (in any intermediate tier between the first server and the origin server).


In step 715, if the resource is determined not to be popular (again, using processing and functionality previously described), the client is directed to a second server in a second tier of servers (e.g., any intermediate tier) distinct from the first tier of servers. In this example embodiment, the second server comprises a first portion (or subset) of the content provider's library. For the purposes of this example, the first portion comprises at least the resource that was requested by the client. Furthermore, the second tier includes at least one other server that comprises a second portion of the content provider's library. According to one example embodiment, the first portion is distinct from the second portion (i.e., the portions/subsets of the library do not overlap, or overlap only minimally).


In another example embodiment, the portions of the content provider's library are logically partitioned across servers in the second tier. For example, the portions/subsets of a content provider's library can be partitioned alphabetically across servers in the second tier based on a naming convention associated with each distinct resource in the content provider's library (e.g., digital movies having titles beginning with A-F are stored on server A in a second/intermediate tier, digital movies having titles beginning with G-R are stored on server B in the second/intermediate tier, and digital movies having titles beginning with S-Z are stored on server C in the second/intermediate tier).


Assume, for example, that a content provider's library comprises multiple content-types (e.g., movies, video games, music, software patches, etc.). Per one example embodiment, it would be advantageous to store a proportionate amount of each content type on each participating intermediate tier server (e.g., 60% movies, 30% video games, 10% software downloads on each participating server in the tier). In such a scenario, if the demand for a particular content-type increases during a given time period (and/or demand for a different content-type declines), then such fluctuations in demand will generally affect each participating server proportionately.


Noted that resources may be partitioned across multiple servers by naming convention (e.g., alphabetically) as well as having proportionally distributed content per server (e.g., by content-type). With this in mind, it should be further noted that logical partitioning of resources/content among intermediate servers is not limited to just alphabetic naming convention segmentation and/or content-type proportionality. It is contemplated that the partitioning may be based on various metrics such as, but not limited to, numeric hash values (e.g., MD5) associated with content/resource URLs, the size or file type of a resource, other identifiers associated with other various naming conventions (e.g., numerical, alphanumerical, proprietary, encrypted/mangled, etc.), the relative size and traffic behavior of various resources in a content provider's library or the content provider's library as a whole, and the like.


In step 720, the step of directing the client to a second server in the second tier is performed in response to a determination of which second server actually stores the first portion. In one example, the server that stores the first portion (or, for that matter, any portion that comprises the requested resource) is identified using a hash function (similar to the processing and functionality previously described).


In step 725, the second server that stores the first portion serves the resource to the client.


The foregoing description has been presented for purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise forms disclosed. Obvious modifications or variations are possible in light of the above teachings. The embodiment or embodiments discussed were chosen and described to provide the best illustration of the principles of the invention and its practical application to thereby enable one of ordinary skill in the art to utilize the invention in various embodiments and with various modifications as are suited to the particular use contemplated. All such modifications and variations are within the scope of the invention as determined by the appended claims when interpreted in accordance with the breadth to which they are fairly and legally entitled.

Claims
  • 1. A method of content delivery in a content delivery network comprising at least a first tier of servers, the method comprising: (A) at a first server in the first tier of servers, obtaining a request from a client for a resource, wherein the resource is available as part of a content provider's library;(B) determining whether the resource is available at the first server or at a peer of the first server;(C) based on and as a result of said determining in (B), if it is determined that the resource is not available at the first server or at a peer of the first server, determining if the resource is popular, wherein the resource is part of an object, and wherein determining whether said resource is popular is based, at least in part, on whether the resource is an initial part of the object;(D) based on and as a result of said determining in (C), if the resource is determined to be popular in (C), then the first server obtaining the resource and the first server serving the resource to the client, otherwise,(E) based on and as a result of said determining in (C), if the resource is determined in (C) not to be popular, directing the client to a second server in a second tier of servers distinct from the first tier of servers,wherein distinct portions of the content provider's library are logically partitioned across servers in the second tier of servers,wherein the second server comprises a first portion of the content provider's library, the first portion comprising at least the resource, and wherein at least one other server in the second tier of servers comprises a second portion of the content provider's library, said second portion of the content provider's library being distinct from said first portion of the content provider's library, andwherein the second tier of servers is any intermediate tier of servers between the first tier of servers and an origin server that stores resources associated with the content provider's library, and(F) the second server serving the resource to the client.
  • 2. The method of claim 1, wherein the resource is available as part of the content provider's library, and wherein distinct portions of the content provider's library are logically partitioned across servers in the second tier of servers.
  • 3. The method of claim 1, wherein determining whether said resource is popular in (C) is based, at least in part, on whether the resource is for a part of the object other than the initial part of the object.
  • 4. The method of claim 3, wherein the initial part of the object has already been served to the client.
  • 5. The method of claim 1, wherein the determining in (C) whether said resource is popular is determined, at least in part, using information obtained from a popularity service running on one or more popularity servers.
  • 6. The method of claim 5, wherein the one or more popularity servers maintain information on relative popularity of content being served.
  • 7. The method of claim 6, wherein the one or more popularity servers maintain information on relative popularity of content being served in a region comprising the first server.
  • 8. The method of claim 6, wherein only certain content uses the popularity service, and wherein content is designated to be served based on popularity in order to use the popularity service.
  • 9. The method of claim 1, wherein whether or not said resource is to be served based on popularity of the resource is based at least in part on whether or not said resource is designated to be served based on popularity of the resource.
  • 10. The method of claim 1 wherein determining whether said resource is popular is based, at least in part, on a dynamic measure of popularity of the resource.
  • 11. The method of claim 1, wherein determining whether said resource is popular in (C) comprises: determining a relative popularity said resource in a region comprising the first server.
  • 12. A machine-readable non-transitory medium storing one or more programs for execution by one or more processing units of a data processing system, the one or more programs including instructions for: (A) obtaining, at a first server in a first tier of servers in a content delivery network (CDN), a request from a client for a resource, wherein the resource is available as part of a content provider's library;(B) determining whether the resource is available at the first server or at a peer of the first server;(C) based on and as a result of said determining in (B), if it is determined that the resource is not available at the first server or at a peer of the first server, then determining if the resource is popular, wherein the resource is part of an object, and wherein determining whether said resource is popular is based, at least in part, on whether the resource is an initial part of the object;(D) based on and as a result of said determining in (C), if the resource is determined to be popular in (C), then the first server (i) obtaining the resource, and (ii) serving the resource to the client, otherwise,(E) based on and as a result of said determining in (C), if the resource is determined in (C) not to be popular, directing the client to a second server in a second tier of servers distinct from the first tier of servers,wherein the second tier of servers is any intermediate tier between the first tier of servers and an origin server that stores resources associated with the content provider's library, andwherein portions of the content provider's library are logically partitioned across servers in the second tier of servers, and wherein the second server comprises a first portion of the content provider's library, the first portion comprising at least the resource, and wherein at least one other server in the second tier of servers comprises a second portion of the content provider's library, said second portion of the content provider's library being distinct from said first portion of the content provider's library; and(F) the second server serving the resource to the client.
  • 13. The machine-readable non-transitory medium of claim, 12, wherein the first server determines whether said resource is popular based on information obtained from a popularity service running on one or more popularity servers.
  • 14. The machine-readable non-transitory medium of claim 13, wherein the one or more popularity servers maintain information on relative popularity of content being served in a region comprising the first server.
  • 15. The machine-readable non-transitory medium of claim 13, wherein only certain content uses the popularity service, and wherein content is designated to be served based on popularity in order to use the popularity service.
  • 16. The machine-readable non-transitory medium of claim 12, wherein whether or not said resource is to be served based on popularity of the resource is based at least in part on whether or not said resource is designated to be served based on popularity of the resource.
  • 17. The machine-readable non-transitory medium of claim 12, wherein determining whether said resource is popular comprises: determining a relative popularity said resource in a region comprising the first server.
  • 18. A computer program product comprising non-transitory computer-readable media and having computer readable instructions stored on said non-transitory computer-readable media, the computer readable instructions including instructions for implementing a computer-implemented method, said method operable in a content delivery network (CDN), said CDN comprising a plurality of computers, each computer comprising hardware including memory and at least one processor, said method comprising: (A) at a first server in the first tier of servers in the CDN, obtaining a request from a client for a resource, wherein the resource is available as part of a content provider's library;(B) determining whether the resource is available at the first server or at a peer of the first server;(C) based on and as a result of said determining in (B), if it is determined that the resource is not available at the first server or at a peer of the first server, determining if the resource is popular, wherein the resource is part of an object, and wherein determining whether said resource is popular is based, at least in part, on whether the resource is an initial part of the object;(D) based on and as a result of said determining in (C), if the resource is determined to be popular in (C), then the first server obtaining the resource and the first server serving the resource to the client, otherwise,(E) based on and as a result of said determining in (C), if the resource is determined in (C) not to be popular, directing the client to a second server in a second tier of servers distinct from the first tier of servers,wherein the second tier of servers is any intermediate tier of servers between the first tier of servers and an origin server that stores resources associated with the content provider's library, andwherein the second server comprises a first portion of the content provider's library, the first portion comprising at least the resource, and wherein at least one other server in the second tier of servers comprises a second portion of the content provider's library, said second portion of the content provider's library being distinct from said first portion of the content provider's library; and(F) the second server serving the resource to the client.
  • 19. The computer program product of claim 18, wherein whether or not said resource is to be served based on popularity of the resource is based at least in part on whether or not said resource is designated to be served based on popularity of the resource.
  • 20. The computer program product of claim 18, wherein determining whether said resource is popular comprises: determining a relative popularity said resource in a region comprising the first server.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation of and claims the benefit of priority from U.S. patent application Ser. No. 12/880,324, entitled “HANDLING LONG-TAIL CONTENT IN A CONTENT DELIVERY NETWORK (CDN),” filed Sep. 13, 2010, the entire contents of which are fully incorporated by reference herein for all purposes. Application Ser. No. 12/880,324 is a continuation-in-part (CIP) of and claims the benefit of priority from U.S. patent application Ser. No. 12/408,681, entitled “HANDLING LONG-TAIL CONTENT IN A CONTENT DELIVERY NETWORK (CDN),” filed Mar. 21, 2009 (now U.S. Pat. No. 8,930,538 issued Jan. 6, 2015), the entire contents of which are fully incorporated by reference herein for all purposes. Application Ser. No. 12/408,681 claims priority as a non-provisional application from U.S. Provisional Application No. 61/042,412, entitled “HANDLING LONG-TAIL CONTENT IN A CONTENT DELIVERY NETWORK (CDN),” filed Apr. 4, 2008, the entire contents of which are fully incorporated by reference herein for all purposes. This application is also related to the following co-owned and co-pending patent applications, the contents of each of which are fully incorporated herein by reference for all purposes: application Ser. No. 10/073,938 filed Feb. 14, 2002, application Ser. No. 11/715,316 filed Mar. 8, 2007, application Ser. No. 11/978,656 filed Oct. 30, 2007, application Ser. No. 11/980,672 filed Oct. 31, 2007, application Ser. No. 10/259,497 filed Sep. 30, 2002, application Ser. No. 11/932,162 filed Oct. 31, 2007, application Ser. No. 11/976,648 filed Oct. 26, 2007, and application Ser. No. 12/390,560 filed Feb. 23, 2009.

US Referenced Citations (460)
Number Name Date Kind
4495570 Kitajima et al. Jan 1985 A
4591983 Bennett et al. May 1986 A
4594704 Ollivier Jun 1986 A
4726017 Krum et al. Feb 1988 A
4803641 Hardy et al. Feb 1989 A
4839798 Eguchi et al. Jun 1989 A
4847784 Clancey Jul 1989 A
4920432 Eggers et al. Apr 1990 A
4922417 Churm et al. May 1990 A
4943932 Lark et al. Jul 1990 A
4949187 Cohen Aug 1990 A
4949248 Caro Aug 1990 A
5029232 Nall Jul 1991 A
5130792 Tindell et al. Jul 1992 A
5132992 Yurt et al. Jul 1992 A
5136716 Harvey et al. Aug 1992 A
5172413 Bradley et al. Dec 1992 A
5191573 Hair Mar 1993 A
5253275 Yurt et al. Oct 1993 A
5253341 Rozmanith et al. Oct 1993 A
5287499 Nemes Feb 1994 A
5287537 Newmark et al. Feb 1994 A
5291554 Morales Mar 1994 A
5341477 Pitkin et al. Aug 1994 A
5371532 Gelman et al. Dec 1994 A
5410343 Coddington et al. Apr 1995 A
5414455 Hooper et al. May 1995 A
5442389 Blahut et al. Aug 1995 A
5442390 Hooper et al. Aug 1995 A
5442749 Northcutt et al. Aug 1995 A
5471622 Eadline Nov 1995 A
5475615 Lin Dec 1995 A
5508732 Bottomley et al. Apr 1996 A
5515511 Nguyen et al. May 1996 A
5519435 Anderson May 1996 A
5528281 Grady et al. Jun 1996 A
5539621 Kikinis Jul 1996 A
5542087 Neimat et al. Jul 1996 A
5544313 Shachnai et al. Aug 1996 A
5544327 Dan et al. Aug 1996 A
5550577 Verbiest et al. Aug 1996 A
5550863 Yurt et al. Aug 1996 A
5550982 Long et al. Aug 1996 A
5557317 Nishio et al. Sep 1996 A
5568181 Greenwood et al. Oct 1996 A
5572643 Judson Nov 1996 A
5590288 Castor et al. Dec 1996 A
5592611 Midgely et al. Jan 1997 A
5594910 Filepp et al. Jan 1997 A
5603026 Demers et al. Feb 1997 A
5614940 Cobbley et al. Mar 1997 A
5619648 Canale et al. Apr 1997 A
5623656 Lyons Apr 1997 A
5625781 Cline et al. Apr 1997 A
5627829 Gleeson et al. May 1997 A
5630067 Kindell et al. May 1997 A
5633999 Clowes et al. May 1997 A
5634006 Baugher et al. May 1997 A
5638443 Stefik et al. Jun 1997 A
5644714 Kikinis Jul 1997 A
5646676 Dewkett et al. Jul 1997 A
5649186 Ferguson Jul 1997 A
5659729 Nielsen Aug 1997 A
5666362 Chen et al. Sep 1997 A
5671279 Elgamal Sep 1997 A
5675734 Hair Oct 1997 A
5682512 Tetrick Oct 1997 A
5699513 Feigen et al. Dec 1997 A
5712979 Graber et al. Jan 1998 A
5715453 Stewart Feb 1998 A
5721914 DeVries Feb 1998 A
5734719 Tsevdos et al. Mar 1998 A
5734831 Sanders Mar 1998 A
5740423 Logan et al. Apr 1998 A
5742762 Scholl et al. Apr 1998 A
5751961 Smyk May 1998 A
5761507 Govett Jun 1998 A
5761663 Lagarde et al. Jun 1998 A
5764906 Edelstein et al. Jun 1998 A
5774660 Brendel et al. Jun 1998 A
5774668 Choquier et al. Jun 1998 A
5777988 Cisneros Jul 1998 A
5777989 McGarvey Jul 1998 A
5778187 Monteiro et al. Jul 1998 A
5781909 Logan et al. Jul 1998 A
5784058 LaStrange et al. Jul 1998 A
5796952 Davis et al. Aug 1998 A
5799141 Galipeau et al. Aug 1998 A
5802106 Packer Sep 1998 A
5802291 Balick et al. Sep 1998 A
5812769 Graber et al. Sep 1998 A
5815662 Ong Sep 1998 A
5815664 Asano Sep 1998 A
5819092 Ferguson et al. Oct 1998 A
5828847 Gehr et al. Oct 1998 A
5832069 Waters et al. Nov 1998 A
5832506 Kuzma Nov 1998 A
5832514 Norin et al. Nov 1998 A
5835718 Blewett Nov 1998 A
5845090 Collins et al. Dec 1998 A
5845303 Templeman Dec 1998 A
5856974 Gervais et al. Jan 1999 A
5862339 Bonnaure et al. Jan 1999 A
5867706 Martin et al. Feb 1999 A
5867799 Lang et al. Feb 1999 A
5870546 Kirsch Feb 1999 A
5870559 Leshem et al. Feb 1999 A
5878212 Civanlar et al. Mar 1999 A
5884038 Kapoor Mar 1999 A
5890171 Blumer et al. Mar 1999 A
5893116 Simmonds et al. Apr 1999 A
5894554 Lowery et al. Apr 1999 A
5896533 Ramos et al. Apr 1999 A
5898456 Wahl Apr 1999 A
5903723 Beck et al. May 1999 A
5907704 Gudmundson et al. May 1999 A
5913028 Wang et al. Jun 1999 A
5913033 Grout Jun 1999 A
5918010 Appleman et al. Jun 1999 A
5919247 Van Hoff et al. Jul 1999 A
5920701 Miller et al. Jul 1999 A
5931904 Banga Aug 1999 A
5933832 Suzuoka et al. Aug 1999 A
5933835 Adams et al. Aug 1999 A
5935207 Logue et al. Aug 1999 A
5940831 Takano Aug 1999 A
5944780 Chase Aug 1999 A
5951694 Choquier et al. Sep 1999 A
5956489 San Andres et al. Sep 1999 A
5956716 Kenner et al. Sep 1999 A
5958008 Pogrebisky et al. Sep 1999 A
5961596 Takubo et al. Oct 1999 A
5966440 Hair Oct 1999 A
5968121 Logan et al. Oct 1999 A
5973696 Agranat et al. Oct 1999 A
5978791 Farber et al. Nov 1999 A
5983214 Lang et al. Nov 1999 A
5983227 Nazem et al. Nov 1999 A
5987606 Cirasole et al. Nov 1999 A
5991809 Kriegsman et al. Nov 1999 A
5996025 Day Nov 1999 A
6002720 Yurt et al. Dec 1999 A
6003030 Kenner et al. Dec 1999 A
6012090 Chung et al. Jan 2000 A
6014686 Elnozahy et al. Jan 2000 A
6014698 Griffiths Jan 2000 A
6016509 Dedrick Jan 2000 A
6016512 Huitema Jan 2000 A
6018516 Packer Jan 2000 A
6021426 Douglis Feb 2000 A
6023470 Lee et al. Feb 2000 A
6029175 Chow et al. Feb 2000 A
6029176 Cannon Feb 2000 A
6035332 Ingrassia et al. Mar 2000 A
6038216 Packer Mar 2000 A
6038310 Hollywood et al. Mar 2000 A
6041324 Earl et al. Mar 2000 A
6044405 Driscoll et al. Mar 2000 A
6046980 Packer Apr 2000 A
6049831 Gardell et al. Apr 2000 A
6052718 Gifford Apr 2000 A
6052730 Felciano et al. Apr 2000 A
6065051 Steele et al. May 2000 A
6065062 Periasamy et al. May 2000 A
6070191 Narendran et al. May 2000 A
6078943 Yu Jun 2000 A
6081829 Sidana Jun 2000 A
6081835 Antcliff Jun 2000 A
6081840 Zhao Jun 2000 A
6085193 Malkin Jul 2000 A
6092112 Fukushige Jul 2000 A
6092204 Baker et al. Jul 2000 A
6094680 Hokanson Jul 2000 A
6094706 Factor et al. Jul 2000 A
6098078 Gehani Aug 2000 A
6098096 Tsirigotis et al. Aug 2000 A
6105028 Sullivan et al. Aug 2000 A
6108673 Brandt et al. Aug 2000 A
6108703 Leighton et al. Aug 2000 A
6112231 DeSimone et al. Aug 2000 A
6112239 Kenner et al. Aug 2000 A
6112240 Pogue et al. Aug 2000 A
6115357 Packer et al. Sep 2000 A
6115752 Chauhan Sep 2000 A
6119143 Dias Sep 2000 A
6125388 Reisman Sep 2000 A
6125394 Rabinovich Sep 2000 A
6128601 Van Horne et al. Oct 2000 A
6128623 Mattis et al. Oct 2000 A
6128660 Grimm et al. Oct 2000 A
6130890 Leinwand et al. Oct 2000 A
6131095 Low et al. Oct 2000 A
6134583 Herriot Oct 2000 A
6144375 Jain et al. Nov 2000 A
6144702 Yurt et al. Nov 2000 A
6151624 Teare et al. Nov 2000 A
6154744 Kenner et al. Nov 2000 A
6154753 McFarland Nov 2000 A
6157648 Volt et al. Dec 2000 A
6161137 Johnson et al. Dec 2000 A
6163779 Mantha et al. Dec 2000 A
6173311 Hassett et al. Jan 2001 B1
6173322 Hu Jan 2001 B1
6178160 Bolton et al. Jan 2001 B1
6181867 Kenner et al. Jan 2001 B1
6185598 Farber et al. Feb 2001 B1
6185619 Joffe et al. Feb 2001 B1
6189030 Kirsch et al. Feb 2001 B1
6189039 Harvey Feb 2001 B1
6195680 Goldszmidt et al. Feb 2001 B1
6205120 Packer et al. Mar 2001 B1
6212565 Gupta Apr 2001 B1
6226618 Downs May 2001 B1
6226642 Beranek et al. May 2001 B1
6230196 Guenthner et al. May 2001 B1
6233623 Jeffords et al. May 2001 B1
6240462 Agraharam et al. May 2001 B1
6249810 Kiraly Jun 2001 B1
6256675 Rabinovich Jul 2001 B1
6263313 Milsted Jul 2001 B1
6266335 Bhaskaran Jul 2001 B1
6266699 Sevcik Jul 2001 B1
6269394 Kenner et al. Jul 2001 B1
6272566 Craft Aug 2001 B1
6275470 Ricciulli Aug 2001 B1
6282569 Wallis et al. Aug 2001 B1
6286045 Griffith et al. Sep 2001 B1
6298041 Packer Oct 2001 B1
6314565 Kenner et al. Nov 2001 B1
6317787 Boyd et al. Nov 2001 B1
6324580 Jindal et al. Nov 2001 B1
6324582 Sridhar et al. Nov 2001 B1
6330602 Law et al. Dec 2001 B1
6332157 Mighdoll et al. Dec 2001 B1
6338044 Cook et al. Jan 2002 B1
6343298 Savchenko et al. Jan 2002 B1
6345294 O'Toole et al. Feb 2002 B1
6347085 Kelly et al. Feb 2002 B2
6351775 Yu Feb 2002 B1
6351776 O'Brien et al. Feb 2002 B1
6360256 Lim Mar 2002 B1
6370571 Medin Apr 2002 B1
6370580 Kriegsman Apr 2002 B2
6389462 Cohen et al. May 2002 B1
6398245 Gruse Jun 2002 B1
6405252 Gupta et al. Jun 2002 B1
6405257 Gersht et al. Jun 2002 B1
6412000 Riddle et al. Jun 2002 B1
6415280 Farber et al. Jul 2002 B1
6415368 Glance et al. Jul 2002 B1
6418421 Hurtado Jul 2002 B1
6421714 Rai et al. Jul 2002 B1
6421726 Kenner et al. Jul 2002 B1
6427170 Sitaraman et al. Jul 2002 B1
6430618 Karger et al. Aug 2002 B1
6442588 Clark et al. Aug 2002 B1
6453319 Mattis et al. Sep 2002 B1
6456630 Packer et al. Sep 2002 B1
6460082 Lumelsky Oct 2002 B1
6460085 Toporek et al. Oct 2002 B1
6463454 Lumelsky Oct 2002 B1
6463508 Wolf Oct 2002 B1
6466949 Yang et al. Oct 2002 B2
6470389 Chung et al. Oct 2002 B1
6473405 Ricciulli Oct 2002 B2
6480893 Kriegsman Nov 2002 B2
6484143 Swildens et al. Nov 2002 B1
6484204 Rabinovich Nov 2002 B1
6487555 Bharat et al. Nov 2002 B1
6490580 Dey et al. Dec 2002 B1
6493707 Dey et al. Dec 2002 B1
6496856 Kenner et al. Dec 2002 B1
6502125 Kenner et al. Dec 2002 B1
6502205 Yanai et al. Dec 2002 B1
6502215 Raad et al. Dec 2002 B2
6505248 Casper et al. Jan 2003 B1
6529477 Toporek et al. Mar 2003 B1
6553413 Leighton et al. Apr 2003 B1
6553420 Karger Apr 2003 B1
6557054 Reisman Apr 2003 B2
6564251 Katariya et al. May 2003 B2
6577595 Counterman Jun 2003 B1
6581090 Lindbo et al. Jun 2003 B1
6584083 Toporek et al. Jun 2003 B1
6587837 Spagna Jul 2003 B1
6591288 Edwards et al. Jul 2003 B1
6591299 Riddle et al. Jul 2003 B2
6594260 Aviani, Jr. et al. Jul 2003 B1
6598121 Challenger Jul 2003 B2
6611862 Reisman Aug 2003 B2
6625643 Colby et al. Sep 2003 B1
6651141 Adrangi Nov 2003 B2
6654344 Toporek et al. Nov 2003 B1
6654807 Farber et al. Nov 2003 B2
6658464 Reisman Dec 2003 B2
6665706 Kenner et al. Dec 2003 B2
6678659 Van Kommer Jan 2004 B1
6691148 Zinky et al. Feb 2004 B1
6694358 Swildens et al. Feb 2004 B1
6699418 Okada et al. Mar 2004 B2
6708137 Carley Mar 2004 B2
6718328 Norris Apr 2004 B1
6732237 Jacobs et al. May 2004 B1
6741563 Packer May 2004 B2
6741990 Nair et al. May 2004 B2
6751673 Shaw Jun 2004 B2
6754699 Swildens et al. Jun 2004 B2
6754706 Swildens et al. Jun 2004 B1
6763377 Belknap Jul 2004 B1
6763388 Tsimelzon Jul 2004 B1
6778502 Ricciulli Aug 2004 B2
6785704 McCanne Aug 2004 B1
6799221 Kenner et al. Sep 2004 B1
6801576 Haldeman et al. Oct 2004 B1
6834306 Tsimelzon Dec 2004 B1
6839758 Sorensen Jan 2005 B2
6842604 Cook et al. Jan 2005 B1
6859791 Spagna Feb 2005 B1
6859840 Singal et al. Feb 2005 B2
6870851 Leinwand et al. Mar 2005 B1
6874032 Gersht et al. Mar 2005 B2
6901604 Kiraly May 2005 B1
6915307 Mattis et al. Jul 2005 B1
6915329 Kriegsman Jul 2005 B2
6928442 Farber et al. Aug 2005 B2
6934255 Toporek et al. Aug 2005 B1
6950623 Brown et al. Sep 2005 B2
6963910 Belknap et al. Nov 2005 B1
6963980 Mattsson Nov 2005 B1
6963981 Bailey et al. Nov 2005 B1
6970432 Hankins et al. Nov 2005 B1
6973490 Robertson et al. Dec 2005 B1
6976090 Ben-Shaul et al. Dec 2005 B2
6981050 Tobias et al. Dec 2005 B1
6981180 Bailey et al. Dec 2005 B1
6996616 Leighton Feb 2006 B1
6999988 Buddhikot et al. Feb 2006 B2
7003572 Lownsbrough et al. Feb 2006 B1
7007089 Freedman Feb 2006 B2
7010578 Lewin et al. Mar 2006 B1
7012900 Riddle Mar 2006 B1
7013342 Riddle Mar 2006 B2
7024466 Outten et al. Apr 2006 B2
7032072 Quinn et al. Apr 2006 B1
7039633 Dey et al. May 2006 B1
7047300 Oehrke et al. May 2006 B1
7054902 Toporek et al. May 2006 B2
7054935 Farber et al. May 2006 B2
7058706 Iyer et al. Jun 2006 B1
7069177 Carley Jun 2006 B2
7096266 Lewin et al. Aug 2006 B2
7103564 Ehnebuske Sep 2006 B1
7103645 Leighton et al. Sep 2006 B2
7110984 Spagna Sep 2006 B1
7117259 Rohwer Oct 2006 B1
7185052 Day Feb 2007 B2
7188085 Pelletier Mar 2007 B2
7206748 Gruse Apr 2007 B1
7219153 Day May 2007 B1
7254645 Nishi Aug 2007 B2
7274658 Bornstein et al. Sep 2007 B2
7359955 Menon et al. Apr 2008 B2
7373644 Aborn May 2008 B2
7562153 Biliris et al. Jul 2009 B2
7577754 Garcia-Luna-Aceves et al. Aug 2009 B2
8930538 Fullagar et al. Jan 2015 B2
20010027491 Terretta Oct 2001 A1
20010029525 Lahr Oct 2001 A1
20010049732 Raciborski et al. Dec 2001 A1
20010051980 Raciborski et al. Dec 2001 A1
20010056500 Farber et al. Dec 2001 A1
20020010798 Ben-Shaul et al. Jan 2002 A1
20020016831 Peled et al. Feb 2002 A1
20020016835 Gamerman Feb 2002 A1
20020018449 Ricciulli Feb 2002 A1
20020023164 Lahr Feb 2002 A1
20020023165 Lahr Feb 2002 A1
20020040404 Lahr Apr 2002 A1
20020042817 Lahr Apr 2002 A1
20020046273 Lahr et al. Apr 2002 A1
20020046405 Lahr Apr 2002 A1
20020049857 Farber et al. Apr 2002 A1
20020059592 Kiraly May 2002 A1
20020066038 Mattsson May 2002 A1
20020073199 Levine et al. Jun 2002 A1
20020078233 Biliris et al. Jun 2002 A1
20020082999 Lee et al. Jun 2002 A1
20020083124 Knox et al. Jun 2002 A1
20020087797 Adrangi Jul 2002 A1
20020089470 Raman et al. Jul 2002 A1
20020092026 Janniello et al. Jul 2002 A1
20020099850 Farber et al. Jul 2002 A1
20020116444 Chaudhri et al. Aug 2002 A1
20020124080 Leighton Sep 2002 A1
20020129134 Leighton et al. Sep 2002 A1
20020131645 Hamilton Sep 2002 A1
20020133537 Lau Sep 2002 A1
20020143798 Lisiecki et al. Oct 2002 A1
20020143888 Lisiecki et al. Oct 2002 A1
20020147774 Lisiecki et al. Oct 2002 A1
20020152318 Menon et al. Oct 2002 A1
20020163882 Bornstein et al. Nov 2002 A1
20020174168 Beukema et al. Nov 2002 A1
20020184357 Traversat et al. Dec 2002 A1
20020198953 O'Rourke et al. Dec 2002 A1
20020199016 Freedman Dec 2002 A1
20030009444 Eidler et al. Jan 2003 A1
20030018966 Cook et al. Jan 2003 A1
20030028623 Hennessey et al. Feb 2003 A1
20030028626 Hennessey et al. Feb 2003 A1
20030028777 Hennessey et al. Feb 2003 A1
20030031176 Sim Feb 2003 A1
20030055972 Fuller et al. Mar 2003 A1
20030061263 Riddle Mar 2003 A1
20030061280 Bulson et al. Mar 2003 A1
20030065703 Aborn Apr 2003 A1
20030065762 Stolorz et al. Apr 2003 A1
20030078888 Lee et al. Apr 2003 A1
20030078889 Lee et al. Apr 2003 A1
20030079027 Slocombe et al. Apr 2003 A1
20030095660 Lee et al. May 2003 A1
20030105604 Ash et al. Jun 2003 A1
20030112792 Cranor et al. Jun 2003 A1
20030140111 Pace et al. Jul 2003 A1
20030149581 Chaudhri et al. Aug 2003 A1
20030154239 Davis et al. Aug 2003 A1
20030158923 Burkhart Aug 2003 A1
20030158928 Knox et al. Aug 2003 A1
20040010588 Slater Jan 2004 A1
20040022194 Ricciulli Feb 2004 A1
20040064832 Tsukidate et al. Apr 2004 A1
20040139097 Farber et al. Jul 2004 A1
20040162871 Pabla et al. Aug 2004 A1
20040177148 Tsimelzon, Jr. Sep 2004 A1
20040205162 Parikh Oct 2004 A1
20040221019 Swildens Nov 2004 A1
20050010653 McCanne Jan 2005 A1
20050021863 Jungck Jan 2005 A1
20050033858 Swildens et al. Feb 2005 A1
20050038851 Kriegsman Feb 2005 A1
20050060493 Krissell et al. Mar 2005 A1
20050100027 Leinwand et al. May 2005 A1
20050114296 Farber et al. May 2005 A1
20050240966 Hindle et al. Oct 2005 A1
20050262104 Robertson et al. Nov 2005 A1
20060015574 Seed et al. Jan 2006 A1
20060112176 Liu et al. May 2006 A1
20060143293 Freedman Jun 2006 A1
20070055764 Dilley et al. Mar 2007 A1
20080065724 Seed et al. Mar 2008 A1
20080071859 Seed et al. Mar 2008 A1
20080147866 Stolorz et al. Jun 2008 A1
20080215718 Stolorz et al. Sep 2008 A1
20090254661 Fullagar et al. Oct 2009 A1
20100169772 Stallings et al. Jul 2010 A1
20100257258 Liu et al. Oct 2010 A1
20100312861 Kolhi et al. Dec 2010 A1
20100332595 Fullagar et al. Dec 2010 A1
20110099290 Swildens et al. Apr 2011 A1
20110145386 Stolorz et al. Jun 2011 A1
Foreign Referenced Citations (52)
Number Date Country
763380 Jun 2003 AU
2202572 Oct 1998 CA
2288488 Jun 2000 CA
2335661 Mar 2001 CA
2335662 Mar 2001 CA
2467998 Apr 2006 CA
ZL99810853.7 Aug 1999 CN
1264476 Aug 2000 CN
0649121 Apr 1995 EP
0651554 May 1995 EP
0801487 Oct 1997 EP
0817444 Jan 1998 EP
0824236 Feb 1998 EP
0865180 Sep 1998 EP
1063831 Dec 2000 EP
1104555 Jun 2001 EP
2281793 Mar 1995 GB
2353877 Apr 2004 GB
140935 Mar 2006 IL
05162529 Jun 1993 JP
07066829 Mar 1995 JP
08328583 Dec 1996 JP
10027148 Jan 1998 JP
10093552 Apr 1998 JP
2000207270 Jul 2000 JP
2001007844 Jan 2001 JP
2001290787 Oct 2001 JP
2001312434 Nov 2001 JP
2002522995 Jul 2002 JP
3566626 Jun 2004 JP
2005124165 May 2005 JP
3762649 Jan 2006 JP
20010023599 Mar 2001 KR
NI1176482 Aug 2003 TW
WO-1996042041 Dec 1996 WO
WO-1997011429 Mar 1997 WO
WO-1997029423 Aug 1997 WO
WO-199742582 Nov 1997 WO
WO-1998004985 Feb 1998 WO
WO-1998006033 Feb 1998 WO
WO-199859486 Dec 1998 WO
WO-199917227 Apr 1999 WO
WO-199948246 Sep 1999 WO
WO-199953422 Oct 1999 WO
WO-2000029990 May 2000 WO
WO-2000052594 Sep 2000 WO
WO-2000060861 Oct 2000 WO
WO-200193533 Dec 2001 WO
WO-2003088065 Oct 2003 WO
WO-2009108593 Sep 2009 WO
WO-2009123868 Oct 2009 WO
WO-2012037039 Mar 2012 WO
Non-Patent Literature Citations (111)
Entry
Canada Examination Report, dated Aug. 16, 2013, Application No. 2,720,353, filed Mar. 21, 2009; 3 pgs.
Canadian Examination Report, dated Jun. 13, 2017, Application No. 2,811,210, filed Sep. 12, 2011; 3 pgs.
Canadian Examination Report, dated Sep. 8, 2014, Application No. 2,720,353, filed Mar. 21, 2009; 4 pgs.
Chinese Examination Report, dated Jun. 13, 2012, Application No. 200980120043.0, 6 pgs.
European Examination Report, dated Mar. 30, 2016, Application No. 09727342.9, filed Mar. 21, 2009; 4 pgs.
Exporting Web Server Final Report, http://www.cs.technion.ac.il/Labs/Lccn/projects/spring97/project4/final report.html (downloaded Jul. 7, 2007). Spring 1997.
Extended European Search Report dated Dec. 6, 2011, App. No. 10178695.2, 7 pgs.
Extended European Search Report, dated Nov. 9, 2015, Application No. 11825742.7, filed Sep. 12, 2011; 6 pgs.
“IBM Technical Disclosure Bulletin; Local Area Network Server Replacement Procedure”, vol. 38, No. 1 (Jan. 1995) , 235-236.
International Preliminary Report on Patentability, dated Mar. 19, 2013, Int'l Appl. No. PCT/US11/051252, Int'l Filing Date Sep. 12, 2011, 6 pgs.
International Search Report, dated Jan. 23, 2012, Intl Appl. No. PCT/US11/51252, Intl Filing Date Sep. 12, 2011, 3 pgs.
JP 10(1998)-27148 machine translation prepared by Japanese Patent Office (JPO), published Jan. 27, 1998, Title: “Server System for Internet,” Applicant: Hitachi , 23 pgs.
Patent Abstracts of Japan “Server System for Internet”, Pub. No. 10-027148, pub. date Jan. 27, 1998, Applicant: Hitachi, computer translation , 12 pgs.
Patent Abstracts of Japan, “Electronic Mail Multiplexing System and Communication Control Method in the System,” (Appln. No. JP19930162529), (Jun. 30, 1993) (Pub. No. JP 7066829).
Patent Abstracts of Japan, “Method and Device for Repeating and Converting Information”, (Appln. No. JP19960328583) (Pub. No. JP10171727) Jun. 26, 1998.
Supplemental European Search Report, dated Nov. 5, 2012, Application No. 09727342.9, 7 pgs.
Written Opinion, dated Jan. 23, 2012, Intl Appl No. PCT/US11/51252, Intl Filing Date Sep. 12, 2011, 5 pgs.
Adler, R. M. “Distributed Coordination Models for Client/Server Computing”, Computer 28, Apr. 4, 1995 , 14-22.
Aggarwal, A. et al., “Performance of Dynamic Replication Schemes for an Internet Hosting Service”, Technical Report, AT&T Labs Oct. 1998.
Almeroth, K. et al., “Scalable Delivery of Web Pages Using Cyclic Best-Effort (UDP) Multicast”, IEEE INFOCOM San Francisco, California, USA Jun. 1998 , pp. 1-24.
Andresen, D. et al., “Multiprocessor scheduling with client resources to improve the response time of WWW applications”, ACM Press, NY, Proc. 11th Inti Conf. on Supercomputing (Austria, ICS '97) Jul. 1997 , 92-99.
Andresen, et al., “SWEB: Towards a Scalable World Wide Web Server on Multicomputers”, Proc. IPPS Apr. 15, 1996 , 850-856.
Awerbuch, et al., “Competitive Distributed File Allocation”, in Proc. of the 25th Ann. ACM Symp. on Theory of Computing May 1993 , pp. 164-173.
Awerbuch, B. et al., “Distributed Paging for General Networks”, In Proc. of the 7th ACM-SIAM Symposium on Discrete Algorithms, Jan. 1996 , pp. 574-583.
Baentsch, M et al., “Enhancing the Web's Infrastructure: From Caching to Replication”, IEEE Internet Computing, Mar. 1997 , 1(2):18-27.
Baentsch, M. et al., “Introducing Application-Level Replication and Naming into Today's Web”, Computer Networks and ISDN Systems vol. 28, Nos. 7-11 May 1996 , pp. 921-930.
Bartal, Y. et al., “Competitive Algorithms for Distributed Data Management”, 24th Annual ACM STOC, May 1992, Victoria, B.C. Canada.
Basturk, E. et al., “Using network layer anycast for load distribution in the Internet”, Tech. Rep., IBM TJ. Watson Research Center Jul. 1997 , 21 pgs.
Berners-Lee, T. et al., RFC 1738—Uniform Resource Locators Dec. 1994 , pp. 1-25.
Berners-Lee, T. et al., “Hypertext Markup Language—2.0”, Internet Engineering Task Force (IETF) Request for Comments (RFC) 1866 Nov. 1995 , pp. 1-77.
Berners-Lee, T. et al., “Hypertext Transfer Protocol—HTTP/1.0”, Internet Engineering Task Force (IETF) Request for Comments (RFC) 1945 May 1996 , pp. 1-60.
Berners-Lee, T. et al., “Hypertext Transfer Protocol—HTTP/1.0”, Internet Engineering Task Force (IETF), Internet Draft, draft-ietf-http-v10-spec-00.ps. Mar. 1995 , pp. 1-59.
Berra, P. B. et al., “Architecture for Distributed Database Systems”, Computer Communications vol. 13, No. 4 May 1, 1990 , pp. 217-231.
Bestavros, A. Demand-Based Document Dissemination to Reduce Traffic and Balance Load in Distributed Information Systems, In Proc. IEEE Symp. on Parallel and Distributed Processing, San Antonio, TX Oct. 1995 , 8 pgs.
Bestavros, et al., “Server-Initiated Document Dissemination for the WWW”, IEEE Data Engineering Bulletin Sep. 1996 , 19(3): 3-11.
Bestavros, A. Speculative Data Dissermination and Service to Reduce Server Load Network Traffic and Service Time in Distributed Information Systems, In Proc. ICDE '96: The 1996 Int'l Conf. on Data Enqineering (Mar. 1996) , 4 Pages.
Bhattacharjee, S. et al., “Application-Layer Anycasting”, in Proc. IEEE INFOCOM '97 Apr. 1997 , 9 pgs.
Braun, H. et al., “Web traffic characterization: an assessment of the impact of caching documents from NCSA's web server”, Comput. Netw. ISDN Syst. 28, Dec. 1-2, 1995 , 37-51.
Brisco, T. DNS Support for Load Balancing, IETF RFC 1794 Apr. 1995 , pp. 1-7.
Carter, et al., “Dynamic server selection using bandwidth probing in wide-area networks”, Tech. Rep. BU-CS-96-007, Compo Sci. Dept., Boston University Mar. 1996 , 1-20.
Carter, et al., “Server selection using dynamic path characterization in Wide-Area Networks”, IEEE INFOCOM '97 1997 , pp. 1014-1021.
Carter, J. L. et al., “Universal Classes of Hash Functions”, Journal of Computer and System Sciences vol. 18, No. 2 Apr. 1979 , 106-112.
Cate, V. Alex: a global file system, in Proc. Usenix Conf. on File Systems May 1992 , pp. 1-11.
Chankhunthod, A. et al., “A Hierarchical Internet Object Cache”, Proc. of the 1996 USENIX Technical Conf. Jan. 1996 , 153-163.
Cohen, J. et al., “Cache Array Routing Protocol v1.1”, http://tools.ietf.org/id/draft-vinod-carp-v1-01.txt (Last-Modified: Wed, Oct. 1, 1997) Sep. 29, 1997 , 8 pages.
Colajanni, et al., “Adaptive TTL schemes for load balancing of distributed Web servers”, SIGMETRICS Perform. Eval. Rev. 25,Sep. 2, 1997 , 36-42.
Colajanni, M. et al., “Scheduling Algorithms for Distributed Web Servers”, Intl Conf. on Distributed Computing Systems (ICDCS), Baltimore, Maryland, USA May 1997 , pp. 169-176.
Crovella, et al., “Dynamic server selection in the Internet”, 3rd IEEE Workshop on the Arch. and Implementation of High Performance Computer Sys. '95 Aug. 1995 , pp. 158-162.
Danzig, P. B. et al., “An analysis of wide-area name server traffic: a study of the Internet Domain Name System”, Conf. Proc. Communications Architectures & Protocols; D. Oran, Ed. SIGCOMM '92; ACM Press, New York, NY Aug. 1992 , 281-292.
De Bra, P.M.E. et al., “Information Retrieval in the World Wide Web: Making Client-Based Searching Feasible”, Computer Networks and ISDN System, NL, North Holland Publishing, Amsterdam, vol. 27, No. 2, ISSN: 0169-7552 Nov. 1, 1994 , 183-192.
Deering, S. E. et al., “Multicast routing in datagram internetworks and extended LANs”, ACM Trans.Comput. Syst. 8,May 2, 1990 , 85-110.
Doi, K. Super Proxy Script—How to make distributed proxy servers by URL hashing, Sharp Corp., http://naragw.sharp.co.jp/sps/; download Jul. 7, 2007. dates unknown (1996-2000).
Eriksson, H. Mbone: The Multicast Backbone, Communications of the ACM vol. 37, No. 8 Aug. 1994 , pp. 54-60.
Feeley, M. et al., “Implementing Global Memory Management in a Workstation Cluster”, In Proc. 15th ACM Symp. on Operating Systems Principles Dec. 1995 , 201-212.
Fielding, R. et al., “Hypertext Transfer Protocol—HTTP/1.1”, Internet Engineering Task Force (IETF) Internet Draft, draft-ietf-http-v11-spec-00.txt Nov. 1995 , pp. 1-81.
Fielding, R. et al., “Hypertext Transfer Protocol—HTTP/1.1”, Internet Engineering Task Force (IETF) Request for Comments (RFC) 2068 Jan. 1997 , pp. 1-162.
Floyd, S. et al., “A Reliable Multicast Framework for Light-Weight Sessions and Application Level Framing”, In Proc. of ACM SIGCOMM '95, Aug. 1995 , 342-356.
Fox, A. et al., “A Framework for Separating Server Scalability and Availability from Internet Application Functionality”, PhD thesis, University of California, Berkeley, 1998 , 163 pgs.
Fox, A. et al., “Cluster-based scalable network services”, Proc. 16th ACM Symp. on Operating Systems Principles (Saint Malo, France, Oct. 5-8, 1997), W. M. Waite, Ed. SOSP '97. ACM Press, New York, NY , 78-91.
Gadde, S. et al., “Reduce, reuse, recycle: An approach to building large internet caches”, in Workshop on Hot Topics in Operating Systems (HotOS) Apr. 1997 , pp. 93-98.
Goldszmidt, M. et al., “Load Distribution for Scalable Web Servers: Summer Olympics 1996—A Case Study”, In Proc. 8th IFIPIIEEE Int'l Workshop on Distributed Systems: Operations and Management , Sydney, Australia. Oct. 1997 , 10 pgs.
Grigni, M. et al., “Tight Bounds on Minimum Broadcasts Networks”, SIAM J. Disc. Math. 4 (May 1991) , 207-222.
Guyton, et al., “Locating nearby copies of replicated Internet servers”, Proc. ACM SIGCOMM '95; pp. 288-298 Oct. 1995.
Gwertzman, J. et al., “The Case for Geographical Push-Caching”, Proc. Workshop on Hot OS '95 (May 4, 1995) , 51-55.
Gwertzman, J. et al., “World-Wide Web Cache Consistency”, Proc. 1996 USENIX Tech. Conf., pp. 141-151, San Diego, CA Jan. 1996.
Jeffery, C. et al., “Proxy sharing proxy servers”, In Proc. IEEE etaCOM Conf. May 1996 , pp. 116-119.
Karger, D. et al., “Consistent Hashing and Random Trees: Distributed Caching Protocols for Relieving Hot Spots on the World Wide Web”, In Proc. 29th Annual ACM Symp. on Theory of Computing May 1997 , 654-663.
Kim, Y. J. et al., “Clustered multi-media NOD: Popularity-based article prefetching and placement”, Sixteenth IEEE Symp. on Mass Storage Systems, San Diego, CA (Mar. 15-18, 1999 , pp. 194-202.
Korkea-Aho, M. Scalability in Distributed Multimedia Systems, Technical report TKO-B128, Helsinki University of Technology.
Kostadinova, R. Peer-to-Peer Video Streaming, [online; retrieved on Jan. 25, 2013]; Retrieved from the Internet <URL: http://www.ee.kth.se/php/modules/publications/reports/2008/XR-EE-LCN_2008_004.pdf>, especially section 5.4.1 2008 , 1-53.
Krishnamurthy, B. et al., “Study of piggyback cache validation for proxy caches in the World Wide Web”, in: Symp. on Internet Technology and Systems, USENIX Association Dec. 1997.
Kroeger, T. M. et al., “Exploring the Bounds of Web Latency Reduction from Caching and Prefetching”, Proc. Usenix Symp. Internet Technologies and Systems, Usenix Dec. 1997 , pp. 13-22.
Kwan, et al., “NCSA's World Wide Web Server: Design and Performance”, IEEE Nov. 1995 , pp. 68-74.
Little, T.D. C. et al., “Selection and Dissemination of Digital Video via the Virtual Video Browser”, Multimedia Tools and Applications vol. 1, No. 2 (Netherlands) Jun. 1995 , pp. 149-172.
Luotonen, et al., “World-Wide Web Proxies”, CERN modified May 24, 1994 Apr. 1994 , 1-8.
Malkin, G. RIP Version 2 Carrying Additional Information, Internet Engineering Task Force (IETF) Request for Comments (RFC) 1388 Jan. 1993 , pp. 1-7.
Malpani, R. et al., “Making World Wide Web Caching Servers Cooperate”, In Proc. 4th Int'l. World Wide Web Conf. (downloaded from http://www.w3.org/ConferencesIWWW4/Papers/59/ on Jul. 7, 2007) Dec. 1995 , 10 pages.
Mockapetris, P. Domain Names—Concepts and Facilities, IETF RFC 1034 Nov. 1987 , 55 pgs.
Mockapetris, P. Domain Names—Implementation and Specification, IETF RFC 1035 Nov. 1987 , 55 pgs.
Mourad, et al., “Scalable Web Server Architectures”, iscc, 2nd IEEE Symposium on Computers and Communications (ISCC '97) Jul. 1997 , pp. 12-16.
Moy, J. OSPF Version 2, Internet Engineering Task Force (IETF) Request for Comments (RFC) 1583 Mar. 1994 , 132 pgs.
Niki, Hitoshi Storage Network, Computopia vol. 36, No. 415, Japan, Computer Age Co., Ltd. (translation) Apr. 1, 2001 , p. 52.
Oguchi, et al., “A Study of Caching Proxy Mechanisms Realized on Wide Area Distributed Networks”, High Performance Distributed Computing, 5th Int'l Symposium Aug. 1996 , pp. 443-449.
Palmer, M. et al., “Fido: A Cache that Learns to Fetch”, In Proc. the 17th Int'l Conf. on Very Large Data Bases Sep. 1991 , 255-264.
Panigrahy, R. Relieving Hot Spots on the World Wide Web, Master's thesis, MIT EECS Jun. 1997 , pp. 1-66.
Parris, C. et al., “A Dynamic Connection Management Scheme for Guaranteed Performance Services in Packet-Switching Integrated Services Networks”, UC Berkeley Computer Science Division Tech. Report TR-93-005 1993 , 37 pgs.
Parris, C. et al., “The Dynamic Management of Guaranteed Performance Connections in Packet Switched Integrated Service Networks”, UC Berkeley Computer Science Division and International Computer Science Institute Tech. Report CSD-94-859 1994.
Partridge, C. et al., “Host Anycasting Service”, Internet Engineering Task Force (IETF) Request for Comments (RFC) 1546 Nov. 1993 , pp. 1-9.
Petri, S. et al., “Load Balancing and Fault Tolerance in Workstation Clusters. Migrating Groups of Communicating Processes”, Operating Systems Review, vol. 29, No. 4 Oct. 1995 , 25-36.
Plaxton, G. C. et al., “Fast Fault-Tolerant Concurrent Access to Shared Objects”, In Proc. 37th IEEE Symp. of Foundations of Computer Science Oct. 1996 , pp. 570-579.
Povey, D. et al., “A distributed internet cache”, in Proc. of the 20th Australasian Computer Science Conf. Feb. 1997.
Rabin, M. O. Efficient dispersal of information for security, load balancing, and fault tolerance, J.ACM 36, Apr. 2, 1989 , pp. 335-348.
Rabinovich, M. et al., “Dynamic Replication on the Internet Work Project No. 3116-17-7006”, AT&T Labs Research Mar. 5, 1998.
Rabinovich, M. et al., “RaDaR: A Scalable Architecture for a Global Web Hosting Service”, WWW May 8, 1999.
Reisslein, M. et al., “Interactive video streaming with proxy servers”, in Proc. of First International Workshop on Intelligent Multimedia Computing and Networking (IMMCN) 2000 , pp. 1-16.
Rodriguez, P. et al., “Improving the WWW: Caching or Multicast?”, Computer Networks and ISDN Systems North Holland Publishing, Amsterdam, NL, vol. 30, No. 22-23, ISSN: 0169-7552 Nov. 1998 , pp. 2223-2243.
Ross, K. W. Hash-Routing for Collections of Shared Web Caches, IEEE Network Magazine 11, 7:37-44 Nov.-Dec. 1997 , pp. 1-21.
Sen, S. et al., “Proxy Prefix Caching for Multimedia Streams”, IEEE Jun. 1999 , 10 pgs.
Smith, Neil What can Archives offer the World Wide Web?, Technical Report 11, University of Kent, Computing Laboratory, University of Kent, Canterbury, UK Mar. 1994 , 1-12.
Takahashi, Takao How to customize Apache Q&A—Web server using Apache, Al Publishing, Apr. 30, 1999, First Edition , pp. 63-64 (translation).
Thaler, D. G. et al., “Using name-based mappings to increase hit rates”, IEEE/ACM Trans. Netw. 6,Feb. 1, 1998 , 1-14.
Topolcic, C. Experimental Internet Stream Protocol, Version 2 (ST-II), Internet Engineering Task Force (IETF) Request for Comments (RFC) 1190 Oct. 1990 , pp. 1-148.
Touch, J. et al., “The X-Bone”, Third Global Internet Mini-Conference at Globecom '98, Sydney, Australia (pp. 44-52 of mini-conference) Nov. 1998 , pp. 59-68.
Traina, P. BGP-4 Protocol Analysis, Internet Engineering Task Force (IETF) Request for Comments (RFC) 1774 Mar. 1995 , pp. 1-10.
Vin, H. Multimedia Broadcasting Over the Internet: Part 1, IEEE Multimedia IEEE Computer Society, US, vol. 5, NR. 4 Oct. 1998 , pp. 78-82.
Vitter, J. S. et al., “Optimal Prefetching via Data Compression”, Proc. 32nd Annual IEEE Symposium on Foundations of Computer Science Oct. 1991 , 21 pages.
Wessels, D. Configuring Hierarchical Squid Caches, Aug. 19 1997 , 62 pgs.
Wessels, Duane Intelligent Caching for World-Wide Web Objects, Masters Thesis, University of Colorado (also presented at INET '95 in Jun. 1995) Jan. 1995 , 1-85.
Wessels, D. et al., “Internet Cache Protocol (ICP), Version 2”, Internet Engineering Task Force (IETF) Request for Comments (RFC) 2186 Sep. 1997 , pp. 1-9.
Wolfson, O. et al., “An Adaptive Data Replication Algorith”, ACM Transactions on Database Systems (TODS) vol. 22(4) Jun. 1997 , pp. 255-314.
Xiao, Li et al., “Exploiting neglected data locality in browsers”, in Proc. 10th Int'l World Wide Web Conference, (WWW10), Hong Kong (extended abstract) May 1-5, 2001 , 2 pgs.
Related Publications (1)
Number Date Country
20170339241 A1 Nov 2017 US
Provisional Applications (1)
Number Date Country
61042412 Apr 2008 US
Continuations (1)
Number Date Country
Parent 12880324 Sep 2010 US
Child 15645584 US
Continuation in Parts (1)
Number Date Country
Parent 12408681 Mar 2009 US
Child 12880324 US