This disclosure relates generally to a technical field of computer networking and, in one example embodiment, to a system and method of application acceleration as a service.
Content acceleration services may be limited to public applications and unsecure data in a public network. As a result, an organization may need to purchase, deploy, and/or maintain expensive infrastructure (e.g., compression appliances, decompression equipment, etc.) at each head office (e.g., the head office 102) and at each branch office (e.g., the branch offices 104A-N) to accelerate private applications and secure data.
For example, the organization may need to purchase expensive and proprietary hardware and/or software solutions that perform functions such as load balancing, compression, de-compression, and/or packet routing. In addition, the organization may need purchase expensive Multiprotocol Label Switching (MPLS) services from bandwidth providers. Such investments may be cost prohibitive for many organizations (e.g., small and medium size businesses). Furthermore, maintaining expensive hardware and software at each head office and each branch office can be complicated, unsightly, and unreliable.
Disclosed are systems and methods to provide application acceleration as a service. A collaborative document is a document that is edited simultaneously by one or more people. A computed document is a document that is generated at run-time for one user. A static document is a document that is prepared a-priori. In one aspect, a system includes a head office to serve an enterprise application that includes a collaborative document. The system also includes a branch office to request the collaborative document from the head office. In addition, the system also includes a set of Point of Presence (POP) locations between the head office and the branch office to communicate the collaborative document on behalf of the head office from a closest POP location to the head office to a closest POP location to the branch office and then onward to the branch office.
The collaborative document may be accessed and/or simultaneously modified by a number of different users at the branch office on a frequent basis. The enterprise application may include a computed document and/or a static document. The set of POP locations between the head office and the branch office may communicate the computed document and/or the static document on behalf of the head office from the closest POP location to the head office to the closest POP location to the branch office and then onward to the branch office.
The enterprise application may be an internal application of a business entity. The head office of the business entity and the branch office of the business entity may securely access the enterprise application through a private network using one or more of public addresses of source and destination routers, pools of addresses represented by a firewall, using a Multiprotocol Label Switching (MPLS) label, and/or using a Virtual Local Area Network (VLAN) tag. The enterprise application may be optionally executed at any of the set of POP locations.
In addition, the system may include an optional Customer Premise Equipment (CPE) device, an optional branch router and an optional head-office router, coupled with the head office and/or the branch office may perform a protocol independent Advanced Redundancy Removal (ARR) function to avoid sending previously sent patterns in a transport stream and/or a packet stream. The system may include the optional CPE device, the optional branch router and the optional head-office router, coupled with the head office and/or the branch office may also generate a secure transport data sent over secure tunnels of the collaborative document, the computed document, and/or the static document. The system may also include the optional CPE device, the optional branch router and the optional head-office router, coupled with the head office and/or the branch office to communicate the secured transport data between a client device in the branch office and the head office, with optional intervening firewalls, through an Internet Protocol Security (IPsec) tunnel, a Generic Routing Encapsulation (GRE) tunnel, VLAN, and/or MPLS labels using IP headers. The CPE may also perform protocol dependent split proxies on the transport stream and/or the packet stream. In addition, the CPE may perform Transmission Control Protocol (TCP) proxies, with varying policies for any of a TCP windows, buffering and/or security.
The optional CPE device may resolve a first bandwidth limitation in a first link between the closest POP location to the branch office and the branch office and to reduce communication time of the collaborative document, the computed document, and the static document in the link between the closest POP location to the branch office and the branch office. In addition, the optional CPE device may resolve a second bandwidth limitation in a second link between the closest POP location to the head office and the head office and to reduce communication time of the collaborative document, the computed document, and the static document in the link between the closest POP location to the head office and the head office. The system of POPs and optional CPEs may perform protocol dependent proxy function (e.g., singly or split across POPs and optional CPEs) to resolve bandwidth limitation or communication time reduction by simplifying the protocol or anticipating requests on behalf of the branch office users. A combination of protocol dependent (e.g., implemented through single and split protocol proxies) and protocol independent functions (e.g., implemented through ARR, TCP proxy) to solve bandwidth reduction and/or communication time reduction may be defined as the application acceleration function. When the aforementioned functions are delivered as a service, the service is called application acceleration as a service.
In addition, the system may include an optional storage of the set of POP locations and the optional CPE device to reduce an amount of data flow when the ARR function is performed. The optional storage may be a flash device, a solid state device and/or a hard drive. In addition, the system may include a public server to generate a computed document and a static document. The branch office may request the computed document and the static document from the head office through the transport stream (e.g., TCP) and/or the packet stream (e.g., IP). The set of POP locations may route a transport stream and a packet stream on behalf of the public server from a closest POP location to the public server to the closest POP location to the branch office to the branch office. There may be one or more head offices, multiple ones of the public server, and multiple ones of the branch office. The head office and/or the branch office may communicate with each other through a private network and/or a public network. The public server and/or the branch office may communicate with each other through the public network and/or the private network.
The computed document may be generated based on a recent interaction between the public server and the branch office during a secure session of the public network. The set of POP locations may be shared by one or more of licensed entities of an application acceleration service. The licensed entities may have one or more head offices and one or more branch offices. Each of the licensed entities may leverage both a shared software and/or a shared hardware infrastructure of an application acceleration service provider. An alternate path may be used to route a transport stream and/or a packet stream that includes the collaborative document, the computed document, and/or the static document between the head office and the branch office when routing through the closest POP location to the branch office is undesirable due to a network congestion, a service unavailability, and/or a segment policy.
An external service may be accessed by the branch office without communicating through the head office when a request to access the external service is communicated from the branch office directly to a particular POP location closest to the external service. A compression between each of the set of POP locations may be variable based on a path segment policy between each of the set of POP locations, the CPE, and/or the head office.
In another aspect, a system includes a business entity that has one or more head offices and one or more branch offices. A set of Point of Presence (POP) locations that are geographically proximate to the one or more head offices and the one or more branch offices may perform an application acceleration function for business entities through a placement of a collaborative document, a computed document, and/or a static document of an enterprise application at a closest POP location to a requesting entity. An optional Customer Premise Equipment (CPE) device at either of the head office and the requesting entity to perform an Advanced Redundancy Removal (ARR) function on the collaborative document, the enterprise application, and/or the computed document. In addition, the CPE may perform TCP proxies with varying policies for TCP windows, buffering and/or security. The CPE may optionally perform protocol dependent split proxies on a transport stream and/or a packet stream.
The requesting entity may be one or more of the branch office of the business entity, a secondary head office to a particular head office of the business entity in which the collaborative document, the computed document, and/or the static document is hosted, and a consumer client-side device. The application acceleration function may be offered as a service to the business entity. A public server may generate the computed document and the static document. A Content Delivery Network (CDN) may optionally host the static document. The requesting entity may request the computed document and/or the static document from the head office through the transport stream and the packet stream. The set of POP locations may route the transport stream and/or the packet stream on behalf of the public server and/or the CDN from the closest POP location to the requesting entity to the requesting entity.
In yet another aspect, a method includes serving an enterprise application comprised of a collaborative document from a head office of a business entity to a branch office. The method also includes processing a request from the branch office for the collaborative document using a processor. In addition, the method includes communicating the collaborative document through a transport stream and/or a packet stream on behalf of the head office from a closest Point of Presence (POP) location to the head office to a closest POP location to the branch office and then onward to the branch office.
The communication of the collaborative document through a transport stream and/or a packet stream on behalf of the head office from a closest Point of Presence (POP) location to the head office to a closest POP location to the branch office and then onward to the branch office may eliminate a requirement to purchase and maintain expensive hardware and software devices at the head office and the branch office to compress and decompress data, and may eliminate a need for the head office to subscribe to a Multiprotocol Label Switching (MPLS) network. The method also includes accelerating the collaborative document from the head office to the branch office.
In addition, the method may include routing the transport stream and the packet stream comprising of the collaborative document, computed document and/or the static document through an alternate path between the head office and the branch office when routing through the closest POP location to the branch office is undesirable due to a network congestion, a service unavailability, and/or a segment policy.
In an other aspect, a method includes providing an application acceleration service to a business entity having one or more head offices and one or more branch offices. The method includes placing a set of Point of Presence (POP) locations in a geographically proximate manner to the head office and the branch office. The method also includes performing an application acceleration function for business entities through a placement of at a collaborative document, a computed document, and/or a static document at a closest POP location to a requesting entity using a processor. In addition, the method includes performing an Advanced Redundancy Removal (ARR) function on the collaborative document, the computed document, and/or the static document using the processor. The ARR function may optionally perform protocol dependent split proxies on a transport stream and/or a packet stream. The method also includes performing TCP proxies with varying policies for one or more of TCP windows, buffering and security.
The methods and systems disclosed herein may be implemented in any means for achieving various aspects, and may be executed in a form of a machine-readable medium embodying a set of instructions that, when executed by a machine, cause the machine to perform any of the operations disclosed herein. Other features will be apparent from the accompanying drawings and from the detailed description that follows.
Example embodiments are illustrated by way of example and not limitation in the figures of accompanying drawings, in which like references indicate similar elements and in which:
Other features of the present embodiments will be apparent from accompanying Drawings and from the Detailed Description that follows.
Example embodiments, as described below, may be used to provide application acceleration as a service. It will be appreciated that the various embodiments discussed herein need not necessarily belong to the same group of exemplary embodiments, and may be grouped into various other embodiments not explicitly disclosed herein. In the following description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the various embodiments.
Each of the POP locations 106A-N may be access points to the Internet. For example, each of the POP location 106A-N may be physical locations that house servers, routers, ATM switches and digital/analog call aggregators. Each POP location 106A-N may be either part of the facilities of a telecommunications provider that the Internet service provider (ISP) rents and/or a location separate from the telecommunications provider. ISPs may have multiple POP locations, sometimes numbering in the thousands. The POPs 106A-N may also be located at Internet exchange points and co-location centers.
A business entity may include the head office 102 (or more head offices) and multiple branch offices 104A-N. The branch offices 104A-N may be spread across different geographies (e.g., regions, countries). The head office 102 and the branch offices 104A-N may be communicatively coupled through the WAN 114A-N. The WAN 114A-N between the head office 102 and the branch offices 104A-N may be enabled through a variety of networking protocols (e.g., Ethernet, Fractional T1/E1, T1/E1, Digital Signal 3 (DS3), Integrated Services Digital Network (ISDN), broadband (e.g., Digital Subscriber Line (DSL), cable, etc.), satellite). In one or more embodiments, the WAN 114A-N may be leased lines or Internet (e.g., egress/ingress only). In one embodiment, the head office 102 (or more head offices), the branch offices 104A-N, the public server 126 may communicate with each other through a private network, and/or the public network. The core network 112 may include the private network and the public network. In one or more embodiments, the core network 112 may use WAN 114A-N/Internet to communicate with the POPs 106A-N, the external services (e.g., such as the service providers 306-308 of
The head office 102 may serve the enterprise application 120, comprised of the collaborative document 118. The enterprise application 120 may be an internal application of the business entity (e.g., that includes one or more head offices 102 and one or more associated branch offices 104A-N). The head office 102 and the branch offices 104A-N may securely share (e.g., access, modify, etc.) the enterprise applications 120 (e.g., Enterprise Resource Planning (ERP), Customer Relationship Management (CRM), e-mail and ftp, voice and video, remote file systems, centralized backup, etc.) over the WAN 114A-N, through a private network using any of public addresses of source and destination routers, pools of addresses represented by a firewall, using a Multiprotocol Label Switching (MPLS) label, and using a Virtual Local Area Network (VLAN) tag. The collaborative document 118 (e.g. Microsoft® Word documents, Microsoft® Excel documents) may be documents that are accessed and simultaneously modified by a number of different users at the branch office 104A-N on a frequent basis through the core network 112 (e.g., through the private network in the core network 112). For example, a collaborative document may be a large structured (e.g. spreadsheet) file and/or an unstructured (e.g. word processing) file simultaneously and frequently accessed and modified by users across head office and branch offices of the same organization (e.g., same business, institution, etc.). In one embodiment, the branch office 104A-N (e.g., also known as requesting entity) may request for the collaborative documents 118, the computed document 122-124 and/or the static document 127-129 service from the head office 102. The branch office 104A-N may include several computing devices that access/modify the collaborative documents 118 in the computing devices managed by processors.
The requesting entities will be described. Particularly, requesting entities (e.g., devices at branch offices 104A-N) may be desktops and/or laptops running client applications like Windows Explorer, Microsoft® Word®, Internet Explorer®, etc. and open client connections to servers at head offices such as the head office 102. Data communication (e.g., such as the communication of the collaborative document, the static document and/or the computed document) between the head office 102 and the branch offices 104A-N may be accelerated using application acceleration services disclosed herein, according to one embodiment. In one or more embodiments, the POPs 106A-N and the optional CPEs 106A-N may perform protocol dependent proxy function (e.g., singly or split across POPs and optional CPEs) to resolve bandwidth limitation or communication time reduction by simplifying the protocol or anticipating requests on behalf of the branch office users. A combination of protocol dependent and protocol independent functions to solve bandwidth reduction and/or communication time reduction may be defined as the application acceleration function. When the aforementioned functions are delivered as a service, the service may be called application acceleration as a service
The serving entities will be described. Particularly, serving entities (e.g., the head office 102) may include servers that host and run the enterprise applications over the WAN 114A-N. The servers may include the file servers, the mail servers, the web servers, public servers, etc. The head office 102 may also include other devices like storage devices, networking devices, etc. The servers and other devices in the head office 102 may be communicatively coupled with other servers and devices in the head office 102 through the LAN 116A. The enterprise application 120 may communicate the collaborative document 118, the computed documents 112-124, and other the static documents 127-129 to the branch offices 104A-N through a transport stream (e.g., TCP) and/or a packet stream (e.g., IP). The transport stream and/or the packet stream may be routed through the POP locations 106A-N. Furthermore, the transport stream and/or the packet stream may be routed in the secure tunnels to destinations via the POPS 106A-N. In one or more embodiments, the public server 126 may generate and host the computed document 124 and/or the static document. The computed document 124 (e.g. HTML and XML) may be generated based on a recent interaction between the public server 126 and the branch office during a secure session (e.g., HTTPS) of the public network 130. In addition, the computed document 124 may be the document that may be generated based on response to a public page (e.g., response page). In one or more embodiments, the computed document 124 may be custom created for a particular user. For example, a computed document may be a confirmation page of a commerce website that is custom created for a user immediately after a recent purchase during a secure session. In one or more embodiments, the CDN 130 may be used to optionally host the static documents to reduce the amount of data communicated between the head office 102 and the branch offices 104A-N. The CDN 130 may be a system of computers networked together across the core network 112 that may cooperate transparently to distribute content for the purposes of improving performance and scalability. The CDN 130 may not host the computed document 124 as hosting becomes inefficient.
In one embodiment, the computed documents 122 may also be generated in the head office 102 and hosted by the public server 126. The static document 127 may be a copy of a content data that may be frequently accessed by the branch offices 104A-N. For example, the static document 127 may be a web page that does not change very often such as a content data of the webpage, landing page of a website etc. provided by the head office to all the branch offices 104A-N. In an alternate embodiment, the enterprise application 120 may be executed directly from any of the POP locations 106A-N rather than from the head office 102.
Similarly, devices at the branch offices 104A-N may be communicatively coupled with other devices in the branch offices 104A-N through the internal local network 116B-N respectively. The router 108A-B may be a networking device that performs a task of routing and forwarding information to destinations. The router 108A-N may communicate data and/or information between the WAN 114A-N and the LAN 116A-N of the head office 102/the branch office 104A-N. The POP 106A-N may be a pool of servers providing WAN optimization and application acceleration. The POPs 106A-N may be communicatively coupled to each other directly or indirectly through the core network 112. Both the core network 112 and WAN 114A-N, may use leased lines and/or Internet. The core network 112 that carries the transport streams and the packet streams may also be compressed.
The private network (e.g., of the core network 112) may be a network that uses private Internet Protocol (IP) addresses based on specified standard (e.g., RFC 1918, RFC 4193, etc.). The POP locations 106A-N may route the transport streams and/or the packet streams that includes the collaborative document 118, and/or the computed document 122-124 on behalf of the head office 102 from a closest POP location to the head office 102 (e.g., the POP 106A as illustrated) to a closest POP location 106B-N to the branch office 104A-N and then onward to the branch office 104A-N. Furthermore, the POP 106A may route the static document 127 on behalf of the public server 126 to the branch office 104A-N through the transport stream and/or packet stream. The private network may use Network Address Translation (NAT) gateway, or a proxy server for connecting to the public network in the core network 112.
The optional CPE 110A-N (e.g., Aryaka™ CPE) may be a device installed at the branch office 104 and/or the head office 102 for performing WAN Advanced Redundancy Removal™ (ARR). It should be noted that Aryaka™ and Advanced Redundancy Removal™ are pending U.S. federal trademarks of Aryaka, Inc. and all rights are reserved to these names.
The optional CPE 110A-N may be configured to perform secure transport of data and communicate the secured data (e.g., collaborative document 118 and the enterprise application 120) between client devices in the branch office 104A-N and the head office(s) 102, with optional intervening firewalls, through Internet Protocol Security (IPsec) tunnel, a Generic Routing Encapsulation (GRE) tunnel, VLANs, and MPLS labels using IP headers. In addition to the optional CPE, an optional branch router, and an optional head-office router (not shown in figure) may be user to perform the ARR, generation of secure transport data and communication of secure transport data over secure channels. Use of the optional CPE 110A-N may enable resolving bandwidth limitation in first/last mile.
The first mile may be a link between the closest POP location 106B-N to the branch office 104B-N and the branch office 104B-N. The last mile (e.g., also referred as second mile) may be a link between the closest POP location 106A to the head office 102 and the head office 102. The optional CPE 110A-N may reduce communication time of the collaborative document 118, the computed document 122-124, and/or the static document 127 in the link between the closest POP location 106B-N to the branch office 104B-N and the branch office 104B-N by resolving bandwidth limitation in the first mile. The optional CPE 110A-N may reduce communication time of the collaborative document 118 and the enterprise application 120 in the link between the closest POP location 106A to the head office 102 and the head office 102 by resolving bandwidth limitation in the last mile.
The use of the optional CPE 110A-N may enable faster data communication in the branch office 104A-N or the head office 102 if the communication line has a low bandwidth. However, if the branch office 104A-N and/or the head office 102 have sufficient bandwidth for data communication, the use of the optional CPE 110A-N may not be required. The POP 106A-N and the optional CPE 110A-N may have storage capabilities for performing Advanced Redundancy Removal for communicating data. The storage in the optional CPE 110A-N may be used for Advanced Redundancy Removal of data to reduce the amount of data flow. The storage in the optional CPE 110A-N may be a flash memory device, according to one embodiment. In alternate embodiments, the optional CPE 110A-N may be coupled or have internally within other types of non-volatile storage devices that includes hard drives, flash drives, solid state devices, etc. Protocol proxies (CIFS, MAPI, HTTP/HTTPS, FTP, PRINT, RTMP, RTP, Oracle, etc.) may be implemented within the POP 106A-N and/or the CPE 110A-N.
Usage of the POP 206A-B may eliminate the requirement of having intelligent synchronized WAN optimization equipments for solving latency and bandwidth at the head office 102 and the branch office 104 ends, according to one embodiment. In addition, the use of the Multiprotocol Label Switching (MPLS) may be eliminated at the core network 112 as the POPs 106A-B speeds up the data communication with no loss in packets and/or delay, according to one embodiment. According to one embodiment, the modified architecture may now be spread across the network with control over data from end to end. As a result, applications such as basic office documents (e.g., spreadsheets, word processing files, etc.), web applications (e.g., detail locators, catalogs, store locators, etc.), Internet applications, etc. may be accelerated through the acceleration as service, according to one embodiment. Large enterprise applications may also be accelerated using the POPs 106A-N. Acceleration of data may be possible as the POPs 106A-N are intelligently designed to analyze the destination of the data packet and to communicate the data packet to the destination without compromising and/or modifying client's private networks.
The POP 106A-N may have within it, a pool of servers providing application acceleration. The POP 106A-N may include the application proxies 202, the edge engines 204, the switching engines 206 and the switches 208A-B. The application proxy 202 may implement and extend a number of protocols such as CIFS, HTTP, MAPI, SMTP, etc. The edge engines 204 may perform WAN data Advanced Redundancy Removal, transparent object caching, IPSEC/SSL security, POP stream shaping, POP-POP data encoding, etc. The switching engines 206 may perform POP-POP routing, QoS, packet classification, stream shaping and load balancing. The switches 208A-B may enable communication between the application proxies 202, the edge engines 204 and the switching engines 206. The application proxies 202, the edge engines 204 and the switch 208A may function as the service server 240. In one or more embodiments, the function as the service server 240 may run on one machine, or one process shared across customers or unique per customer. The service servers 240 may provide the QoS as packets are delivered based on priority order using the application proxies 202 and edge engines 204 based on the type of the data, application of the data, secure data, etc. The switch 208B and the switching engines 206 may manage the network switching 245. The network switching 245 may be function performed by the switching engine 206 to forward the data packets through the network.
The POP 106A-N may also have the optional storage device 210 for performing ARR for transportation of data. In one or more embodiments, the storage 210 may be a shared storage. The ARR may be a class of techniques used to remove duplicate information between senders and receivers by capturing histories of data streams and holding these histories beyond the life of connections. The POPs 106A-N may be shared among different clients and different branches. In addition, the engines in the POP 106A-N may be shared by different clients. The POPs 106A-N may be centrally controlled through a control station. Also, the POPs 106A-N may provide a capability of being controlled from distributed locations.
The POP 106A-N may form a part of a core region (e.g., core region 660 as illustrated in
In an embodiment, a segment may be a communication link between the POP and other POPs. For example, the segment may be an Internet or private point-point leased line. Policies may be assigned per segment. The POPs 106A-N may be communicatively coupled to each other through transport network. Since, the POPs 106A-N are communicatively coupled to each other directly/indirectly, there may be multiple segments. Therefore, the architecture in the system may be called as multi-segment architecture. Also, communication link between each of nodes may also be called as segment. The multi-segment architecture may be based on layer 3 (network layer)/layer 4 (transport layer). Subsequently, this disclosure calls the layer-3 segmentation bypass mode packets and layer-4 segmentation intercept mode packets (using TCP proxies). The multi-segmented architecture may enable each segment to have different queuing and routing policies based on cost and congestion.
In addition, the system as described in the disclosure may include Transmission Control Protocol (TCP) proxies (e.g., at layer 4) at each segment. ACKs of TCP proxies are acknowledged by immediately following segment which may significantly reduce congestion issues and packet loss. Each segment may be configured with different TCP policies (e.g., windows scaling, Selective ACKs (SACK), Additive Increase/Multiplicative Decrease (AIMD), etc) based on speed of link, congestion, peering points and customer preference. In addition, the TCP policies may be controlled per protocol, per client. Furthermore, the TCP policies may be changed at runtime based on traffic conditions at each segment.
In one embodiment, the segments may be formed through an Internet or private point-point leased lines, pre-provisioned, etc. The POP-POP multi-segmented architecture may be lossless. The lossless ability may be achieved using a combination of layer 4 and a proprietary buffer reservation system (e.g., storage at the POPs 106A-N and optional CPE 110A-N). Furthermore, each segment may implement link utilization algorithms (e.g., interleaving algorithm, piggybacking algorithm, connection multiplexing algorithm) using transport protocols (layer 4) like standard TCP, High Speed TCP, Explicit Control Protocol (XCP) and/or even Stream Control Transmission Protocol (SCTP).
For example, the branch office 104B may require communication with the external services 306 (e.g., SAAS applications like Salesforce.com® and/or Web filtering like Websense®). Instead of sending the traffic to the POP 106A of the head office 102 for services, the POP 106C can direct the traffic to the POP 106B of the external services 306 directly. As a result, the time for the data communication of data may significantly reduce. In another embodiment, when the head office 102, wants to communicate with external services 306 (e.g. Amazon web services). The communication may be channeled directly to the external services 306 by the POP 106A-N instead of directing through the Internet. Utilizing the POPs 106 A-N instead of directing through the Internet would substantially improve latency and reduce bandwidth costs.
In an embodiment, the system may provide high availability (e.g., connectivity end to end). In an event of the POP failure (e.g., due to a network congestion, a service unavailability, a segment policy, etc.), the system may switch the coupling to a different POP. In case, when there is an intermediate POP failure, an alternate route may be determined and the data is re-routed. Data re-routing may be well understood even in a label switched network. However, in a multi-segmented network with encryption and compression, the final destination may not be known unless the payload is inspected (and the payload is encrypted and compressed). The system may provide an out-of-band protocol that may communicate the final destination address used for controlling the re-routing of the compressed/encrypted payload in the event of POP failure. In an example embodiment, the head office 102 in
In another embodiment, the set of POP locations 106A-N may be shared by more than one licensed entity of the application acceleration service. For example, the external service providers 306-308 may not be a part of the business entity. However, the service providers 306-308 may be clients of the business entity or the service providers to the business entity. The service providers 306-308 discussed herein may also be the licensed entities of the application acceleration service. Therefore, the service providers 306-308 may be entitled to use the POP locations 106A-N closest to them. As a result, the licensed entities may leverage both a shared software and a shared hardware infrastructure of an application acceleration service provider. It may be noted that the licensed entities may also have head offices and branch offices. The embodiments described herein may not be limited to hub-spoke configurations (e.g., the head office 102 serving as a hub and the branch offices 104A-N configured as a spokes). It should be noted that the embodiments described herein may also support hub-hub (e.g., the head office 102 requesting for services from an external service provider) and spoke-spoke configurations (e.g., services among the branch offices 104A-N).
The Advanced Redundancy Removal for the head office 102—the POP 106A segment may be only optional. The segment between the POP 106A and the POP 106C through a private network may have a bandwidth of 1000 Mbps 404. The compression technique that can be implemented for the segment between the POP 106A and the POP 106C with bandwidth of 1000 Mbps 404 may include adaptive Huffman coding, LZH, GZIP, PAQ, etc. Furthermore, the link between the POP 106C and the branch office 104B may have a bandwidth of 1.5 Mbps 406, and the compression model that can be implemented may be a WAN Advanced Redundancy Removal model. The Advanced Redundancy Removal resource consumption as a function of rate of compression is illustrated in
Particularly, the table in
The IPsec 602 may be a protocol suite for securing Internet Protocol (IP) communications by authenticating and/or encrypting each IP packet of a data stream. The IPsec 602 may be used to protect data flows between a pair of hosts (e.g. computer users or servers), between a pair of security gateways (e.g. routers or firewalls), and/or between a security gateway and a host.
The Insertion Header analyzer 606 may be components that read per customer packet headers (e.g., encapsulated in IPSEC over Generic Routing Encapsulation GRE) and determine the instances to redirect. The Incoming Redirector 608 may perform the actual work of redirecting the traffic to per Customer instances 609. One per Customer instance may be the Filter 610 that tracks L3 and L4 state and determines if the traffic needs to be bypassed or intercepted. Bypassed traffic may be directed to the core network via per Customer instance Core Bypasser 640. Intercepted traffic may be sent to Service Servers (e.g., the service servers 240 as illustrated in
In
In one embodiment, each customer may have thousands of client connections, originating from hundreds of desktops/laptops from each of the branch office networks 116B-N, that need to be rediscovered as client connections 618 within per customer instance TCP proxies 636. As such, there may be thousands of client connections 618 per branch. Four other proxies are illustrated in
The various embodiments described herein eliminate the burden on customers (e.g., organizations) to purchase, deploy, and maintain expensive hardware and/or software at each head office (e.g. the head office 102) and branch office (e.g., the branch offices 104A). By deploying an intelligent network of POPs 106A-N of
Furthermore, within each of the POPs 106A-N, requests may be removed, coalesced, and/or anticipated to mitigate latency, according to various embodiments. The hardware and/or software in the POPs 106A-N may be application protocol aware by reducing the state diagram of common protocols to alternate forms that remove requests, according to various embodiments. Request elimination may be completed in the cloud (e.g., in the POPs 106A-N) rather than at the head office 102 and/or the branch office 104A-N. Therefore, the various embodiments described herein eliminate significant capital outlays.
In addition, according to the various embodiments, a customer (e.g., head office 102 and the branch offices 104A-N) does not need to purchase end-end networking links (e.g., MPLS) separately according to the various embodiments disclosed herein. Rather, networking services (resources, links) may be tiered (e.g., based on speed of each segment as described in
Furthermore, capacity and load can better be managed across the large number of customers that share these networking services through the various embodiments disclosed herein. The core network 112 may provide priorities for packets between service levels (gold, silver, bronze) or class of service (voice, video, data), according to one embodiment. These priorities will be used for packet queuing, transmission or dropping. As a result, according to one embodiment, jitter, delay and packet loss for WAN connections may be mitigated through the various embodiments described in
Furthermore, the POPs 106A-N may have additional resources that may be cost prohibitive to smaller customers and may be able to offer such resources as a service to smaller customers. For example, the POPs 106A-N may have packet shaper devices that balance throughput and latency to avoid packet loss (e.g., by prioritizing traffic, delaying lower priority traffic, etc.) therefore providing bandwidth throttling and/or rate limiting functions.
The various embodiments described herein may help to alleviate problems of resource strain caused on individual customers due to protocol independent de-duplication. For example, according to the various embodiments described herein, the POPs 106A-N and the core network 112 may provide support for a variety of protocols (CIFS, MAPI, HTTP/HTTPS, FTP, PRINT, RTMP, RTP, Oracle) for bandwidth reduction and/or communication time reduction, without consuming resources of individual customers. In one or more embodiments, the aforementioned protocol support may be spread across the POPs 106A-N. The protocol support that is spread across the POPs 106A-N and the CPE 110A-N with/without connection termination may be called as split proxies as, the proxy functionality may be split across the POPs 106A-N and CPEs 110A-N. As such, through the various embodiments described in
In addition, the various embodiments described herein eliminate the need for symmetric hardware and software resources at the head office 102 and the branch offices 104A-N. Instead, symmetry may be provided through the cloud through the POPs 106A-N. Therefore, when the additional head offices 102 and/or the branch offices 104A-N are brought on board, there may be no need to purchase, deploy and configure expensive hardware and software. This may eliminate configuration challenges, delays, downtime issues, and/or design complexities caused by installation, maintenance and service of resources when new head offices and/or branch offices are brought on board.
Any equipment at the head-end or branch-end such as the optional CPE 110A-N is non-compulsory (required only if the egress/ingress bandwidth is a bottleneck). Even when used, the optional CPE 110A-N on the head-end or branch-end may only perform WAN de-duplication which may be an important design choice of performing WAN de-duplication at the link with the lowest bandwidth. By making the function of the optional CPEs 110A-N minimal, by appropriately sizing the capability of the CPE to handle the capacity of the branch (or head office) link, the entry cost of providing the service can be dramatically reduced, according to the various embodiments described herein.
In addition, the various embodiments described herein provide a distributed architecture in dealing with application acceleration according to one embodiment. Instead of stuffing all the bandwidth improvement and latency mitigation technologies in single boxes, the problem is distributed across the POPs 106A-N and the core network 112. Therefore, functions that were once required at each of the nodes (e.g., at the head office 102 and the branch offices 104A-N) such as routing, QoS, packet classification, stream shaping, load balancing, WAN data de-duplication, transparent object caching, IPSEC/SSL security, stream shaping, and data encoding can be performed at the POPs 106A-N and in the core network 112 rather than individually by customers, according to the various embodiments disclosed herein. Furthermore, application proxy nodes performing CIFS, HTTP, MAPI, SMTP, etc. can be now shared for a large number of customers rather than individually managed by customers (e.g., eliminating resource constraints at the customer premises).
In addition, it will be appreciated that the application as a service methods and system disclosed herein may enable customers to purchase services when they need them entirely themselves. In other words, the various technologies disclosed herein may be available through a ‘self service’ model in which customers can elect the acceleration services they require entirely online. Also, it will be appreciated that the application as a service methods and system disclosed herein can be billed on a pay per use basis. In other words, customers may be charged only for the services that they actually use. It will be also appreciated that the application as a service methods and system disclosed herein provide elastic scalability, in which the customers are enabled to scale their application acceleration service requirements based on their business needs quickly, efficiently, and on an as-needed basis. The application as a service methods and system disclosed herein do not require network administrators to perform complicated and tedious hardware management functions to maintain and deploy acceleration hardware at each node, thereby providing ease of management.
The application acceleration services as described in the aforementioned embodiments improves on content acceleration services by extending functionality to accelerating the service of the enterprise application and secure data acceleration as a service. In contrast, solely content acceleration services are limited to public applications and unsecure data in a public network. In other words, it should be noted that ‘application acceleration’ as described herein refers to both acceleration of services of the enterprise applications and secure data in a private network (e.g., Intranet) and/or acceleration of applications and data of a public network (e.g., Internet). Private Networks have private IP addresses and data in such networks can be routed through the public network using tunnels (e.g., IPSEC tunnel, GRE tunnel, and other similar mechanisms) with public IP address headers. Public Networks have public IP addresses and data in such networks can be routed using a core routing protocol such as the Border Gateway Protocol (BGP). Application acceleration as described herein can operate in both public network and/or private network scenarios.
A few examples that distinguish the application acceleration from a content acceleration are provided. The application acceleration may apply to enterprise applications (e.g., accessible only to employees of a company when communicating through a corporate network having a firewall) to provide business logic support functionality to an enterprise, such as to a commercial organization, which aims to improve the enterprise's productivity and/or efficiency. Furthermore, the application acceleration may provide acceleration of secure data generated from internal applications that are frequently modified (e.g., a Microsoft® Word file stored on a corporate server that is simultaneously accessed and modified by employees of a company from geographically dispersed offices). Such functions are not possible using solely content acceleration services. Unlike content acceleration, the various embodiments disclosed herein for application acceleration may be applicable to the computed documents 122-124, the static documents 127-129 and/or the collaborative documents 118.
In operation 708, the collaborative document 118 may be accelerated (e.g., using multiple POPs 106A-N) from the head office 102 to the branch office 104A-N. In operation 710, simultaneous access to the collaborative document 118 may be granted to different client devices of the branch office 104A-N. In operation 712, a simultaneously modification of the collaborative document 118 may be processed by the different client devices on a frequent basis from a closest POP location to the branch office 104A-N to a closest POP location to the head office 102. In operation 714, the ARR function may be performed to avoid sending previously sent patterns in the transport stream and/or the packet stream. In operation 716, an amount of data flow may be reduced when the ARR function is performed through an optional storage comprising a flash device, a solid state device and a hard drive
In operation 724, a second bandwidth limitation may be resolved in a second link between the closest POP location to the head office 102 and the head office 102 and communication time of the collaborative document 118 may be reduced in the link between the closest POP location to the head office 102 and the head office 102. In operation 726, the computed document 124 and the static document 127 may be generated through the public server 126. In operation 728, the transport stream and/or the packet stream that includes the computed document 124, the collaborative document 118 and/or the static document 127 may be routed on behalf of the public server 126 from the closest POP location to the branch office 104A-N to the branch office 104A-N.
The functions 901 in the application acceleration as a service may be implemented through protocol independent acceleration 905 and the protocol dependent acceleration 909. In one or more embodiments, the protocols that implement the protocol independent acceleration include the ARR, the TCP proxy, etc. In one or more embodiments, the protocols that implement the protocol dependent acceleration include the single and/or split protocol proxies.
The services 902 provided by the application acceleration as a service 950 may include network of POPs 908, the transport 910, elastic capacity 912, pay per use 928, self service 929 and no hardware management 930. In one or more embodiments, the network of POPs may be installed in various locations of the world. In some embodiments, the POPs 106A-N may be located at Internet exchange points and co-location centers. The POPs 106A-N may be a group of servers that communicate the collaborative document between the office locations (e.g., that include head office 102 and multiple branch offices 104A-N). The POPs 106A-N may support different configurations. For example, the configurations may include, but not limited to hub/hub, hub/spoke, and spoke/spoke. In one or more embodiments, the POPs 106A-N may be connected to each other directly or indirectly. However, it should be noted that the POPS 106A-N may be well networked to provide application acceleration through finding alternative paths to reduce time of data communication. The transport 910 services may include secure transport of data and communication of the secured data (e.g., collaborative document 118 and the enterprise application 120) between the offices based on any of the aforementioned configurations through the Internet Protocol Security (IPsec) tunnel, the Generic Routing Encapsulation (GRE) tunnel, VLANs, and MPLS labels using IP headers with optional intervening firewalls. The services described herein may be provided to customers on the pay per use 928 basis based on service levels (gold, silver, bronze) or class of service (voice, video, data). In addition, the customers may be charged only for the services that they actually use.
The services 902 disclosed herein also provides elastic scalability 912, whereby allowing the customers to scale their application acceleration service requirements based on their business needs quickly, efficiently, and/or on an as-needed basis. The self service 929 may enable customers to avail services when they need them entirely themselves (e.g., self service model). The services 902 provides ease of no hardware management 930, thereby eliminating a requirement of administrators to perform complicated and tedious hardware management functions to maintain and deploy acceleration hardware at each node.
The application acceleration as a service may also support multi tenancy 903 in which it provides an option to share POPs, network and software 931, performance and fault isolation 932, and quality of service by traffic/customer. The multi-tenancy 903 may enable sharing of the POPs between the customers, thereby enabling sharing of the network and software between the customers with high security (e.g., as illustrated in
The application acceleration as a service 950 may be performed using the POP locations 106A-N by communicating the secure data through the core networks 112. The application acceleration services disclosed herein do not require tags to be changes in documents (collaborative, computed, and static) to be accelerated. The application acceleration services disclosed herein can apply transparent caching as well as tag changes to point to POP locations closest to a requesting entity. For computed and static documents, the various embodiments disclosed herein may be differentiated from other methods in that multisegment policies (e.g., compression, window sizing, and securities) can be applied to scenarios in which computed and static documents are accelerated. The various routing technologies disclosed herein may apply a GRS tunnel which uses IP layer 3 to emulate MPLS like functionalities. As a result, the various embodiments disclosed herein may serve as WAN optimize replacement services in addition to application acceleration services.
The policies 914 in the multisegment 904 may include assigning the policies between the segments (e.g., see
The routing services 906 may include configurations that the system supports to enable communication of the secure data between end nodes (e.g., offices). The configuration may include hub/hub, hub/spoke and spoke/spoke. The hub/hub configuration 922 enables communication between the service providing entities. For example, the head office communicating with another head office located at different location geographically or communication in multi tenancy configuration. In one or more embodiments, the entities in the hub/hub configuration 922 may be serving entities. An example embodiment of the hub/hub configuration 922 is illustrated in
Although the present embodiments have been described with reference to specific example embodiments, it will be evident that various modifications and changes may be made to these embodiments without departing from the broader spirit and scope of the various embodiments. For example, the various devices and modules described herein may be enabled and operated using hardware circuitry (e.g., CMOS based logic circuitry), firmware, software or any combination of hardware, firmware, and software (e.g., embodied in a machine readable medium). For example, the various electrical structure and methods may be embodied using transistors, logic gates, and electrical circuits (e.g., application specific integrated (ASIC) circuitry and/or in Digital Signal Processor (DSP) circuitry).
In addition, it will be appreciated that the various operations, processes, and methods disclosed herein may be embodied in a machine-readable medium and/or a machine accessible medium compatible with a data processing system (e.g., a computer system), and may be performed in any order (e.g., including using means for achieving the various operations). Accordingly, the specification and drawings are to be regarded in an illustrative rather than a restrictive sense.
This application is a continuation application of U.S. patent application Ser. No. 13/919,018 titled APPLICATION ACCELERATION AS A SERVICE SYSTEM AND METHOD filed on Jun. 17, 2013, which itself is a continuation application of U.S. patent application Ser. No. 12/550,409 titled, APPLICATION ACCELERATION AS A SERVICE SYSTEM AND METHOD filed Aug. 31, 2009, which is issued on Jul. 16, 2013 having U.S. Pat. No. 8,489,685, which claims priority under 35 USC §119(e) to U.S. Provisional Application No. 61/226,407 titled ENTERPRISE APPLICATION AND SECURE DATA ACCELERATION AS A SERVICE SYSTEM AND METHOD filed on Jul. 17, 2009, the entirety of each of which is herein incorporated by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
5263157 | Janis | Nov 1993 | A |
5586260 | Hu | Dec 1996 | A |
5764887 | Kells et al. | Jun 1998 | A |
5765153 | Benantar et al. | Jun 1998 | A |
5864665 | Tran | Jan 1999 | A |
5915087 | Hammond et al. | Jun 1999 | A |
5968176 | Nessett et al. | Oct 1999 | A |
5983350 | Minear et al. | Nov 1999 | A |
6088451 | He et al. | Jul 2000 | A |
6112228 | Earl et al. | Aug 2000 | A |
6131163 | Wiegel | Oct 2000 | A |
6158007 | Moreh et al. | Dec 2000 | A |
6158010 | Moriconi et al. | Dec 2000 | A |
6202157 | Brownlie et al. | Mar 2001 | B1 |
6209077 | Robertson et al. | Mar 2001 | B1 |
6233612 | Fruchtman et al. | May 2001 | B1 |
6292904 | Broomhall et al. | Sep 2001 | B1 |
6460141 | Olden | Oct 2002 | B1 |
6484261 | Wiegel | Nov 2002 | B1 |
6499107 | Gleichauf et al. | Dec 2002 | B1 |
6513061 | Ebata et al. | Jan 2003 | B1 |
6542993 | Erfani | Apr 2003 | B1 |
6594763 | Madoukh | Jul 2003 | B1 |
6606708 | Devine et al. | Aug 2003 | B1 |
6704874 | Porras et al. | Mar 2004 | B1 |
6715084 | Aaron et al. | Mar 2004 | B2 |
6745245 | Carpenter | Jun 2004 | B1 |
6748540 | Canestaro et al. | Jun 2004 | B1 |
6782527 | Kouznetsov et al. | Aug 2004 | B1 |
6789202 | Ko et al. | Sep 2004 | B1 |
6826615 | Barrall et al. | Nov 2004 | B2 |
6876654 | Hegde | Apr 2005 | B1 |
6880005 | Bell et al. | Apr 2005 | B1 |
6920558 | Sames et al. | Jul 2005 | B2 |
6980989 | Silverman et al. | Dec 2005 | B2 |
7003117 | Kacker et al. | Feb 2006 | B2 |
7093283 | Chen et al. | Aug 2006 | B1 |
7093297 | Moore | Aug 2006 | B2 |
7117195 | Chantrain et al. | Oct 2006 | B2 |
7127492 | Calo et al. | Oct 2006 | B1 |
7134137 | Joshi et al. | Nov 2006 | B2 |
7140035 | Karch | Nov 2006 | B1 |
7146640 | Goodman et al. | Dec 2006 | B2 |
7150038 | Samar | Dec 2006 | B1 |
7178164 | Bonnes | Feb 2007 | B1 |
7185364 | Knouse et al. | Feb 2007 | B2 |
7210167 | Brezak et al. | Apr 2007 | B2 |
7231661 | Villavicencio et al. | Jun 2007 | B1 |
7237125 | Raley et al. | Jun 2007 | B2 |
7246230 | Stanko | Jul 2007 | B2 |
7272613 | Sim et al. | Sep 2007 | B2 |
7308716 | Danford et al. | Dec 2007 | B2 |
7318237 | Moriconi et al. | Jan 2008 | B2 |
7346775 | Gasparinl et al. | Mar 2008 | B2 |
7392546 | Patrick | Jun 2008 | B2 |
7444414 | Foster et al. | Oct 2008 | B2 |
7451488 | Cooper et al. | Nov 2008 | B2 |
7467410 | Graham et al. | Dec 2008 | B2 |
7483904 | Owen et al. | Jan 2009 | B2 |
7591000 | Griffin et al. | Sep 2009 | B2 |
7594260 | Porras et al. | Sep 2009 | B2 |
7640574 | Kim et al. | Dec 2009 | B1 |
7653930 | Griffin et al. | Jan 2010 | B2 |
7657938 | Palmer, Jr. et al. | Feb 2010 | B2 |
7673323 | Moriconi | Mar 2010 | B1 |
7676834 | Camaisa et al. | Mar 2010 | B2 |
7746496 | Henry et al. | Jun 2010 | B2 |
7765298 | Villavicencio | Jul 2010 | B2 |
7779466 | Judge et al. | Aug 2010 | B2 |
7793290 | Gebhart et al. | Sep 2010 | B2 |
7797752 | Vaidya et al. | Sep 2010 | B1 |
7810147 | Duane et al. | Oct 2010 | B2 |
7814536 | Martherus et al. | Oct 2010 | B2 |
7861288 | Tsuruoka et al. | Dec 2010 | B2 |
7861289 | Meyer et al. | Dec 2010 | B2 |
7861290 | Meyer et al. | Dec 2010 | B2 |
7865943 | Hayler et al. | Jan 2011 | B2 |
7886352 | Pandrangi et al. | Feb 2011 | B2 |
7890655 | Chandrasekaran | Feb 2011 | B2 |
7899849 | Chaudhry et al. | Mar 2011 | B2 |
7904953 | Meyer et al. | Mar 2011 | B2 |
7921456 | Chua | Apr 2011 | B2 |
7934254 | Graham | Apr 2011 | B2 |
7992189 | Griffin et al. | Aug 2011 | B2 |
8006098 | Aupperle et al. | Aug 2011 | B2 |
8010991 | Sarukkai et al. | Aug 2011 | B2 |
8020206 | Hubbard et al. | Sep 2011 | B2 |
8024771 | Malkin et al. | Sep 2011 | B2 |
8042149 | Judge | Oct 2011 | B2 |
8042181 | Judge | Oct 2011 | B2 |
8060932 | Mohammed et al. | Nov 2011 | B2 |
8079070 | Camaisa et al. | Dec 2011 | B2 |
8104075 | Spector | Jan 2012 | B2 |
8141147 | Sinclair et al. | Mar 2012 | B2 |
8160975 | Tang et al. | Apr 2012 | B2 |
8180897 | Barrall et al. | May 2012 | B2 |
8191106 | Choyi et al. | May 2012 | B2 |
8201218 | Malakapalli et al. | Jun 2012 | B2 |
8271661 | Harris et al. | Sep 2012 | B2 |
8272060 | Milliken et al. | Sep 2012 | B2 |
8296178 | Hudis et al. | Oct 2012 | B2 |
8302189 | De Los Reyes et al. | Oct 2012 | B2 |
8347373 | Crawford | Jan 2013 | B2 |
8356074 | Ehrlich et al. | Jan 2013 | B1 |
8364812 | Loach | Jan 2013 | B2 |
8397283 | Hayler et al. | Mar 2013 | B2 |
8402530 | Anderson et al. | Mar 2013 | B2 |
8429708 | Tandon | Apr 2013 | B1 |
8474034 | Zhang et al. | Jun 2013 | B2 |
8479279 | Li et al. | Jul 2013 | B2 |
8489685 | Bharadhwaj et al. | Jul 2013 | B2 |
8509237 | Akella et al. | Aug 2013 | B2 |
8539544 | Garimella et al. | Sep 2013 | B2 |
8549617 | Wei et al. | Oct 2013 | B2 |
8561167 | Alperovitch et al. | Oct 2013 | B2 |
8583793 | De Icaza | Nov 2013 | B2 |
8590004 | Comay et al. | Nov 2013 | B2 |
8595791 | Chen et al. | Nov 2013 | B1 |
8595820 | Vaidya | Nov 2013 | B1 |
8612622 | Swanson et al. | Dec 2013 | B2 |
8621075 | Luna | Dec 2013 | B2 |
8621638 | Judge et al. | Dec 2013 | B2 |
8627467 | Levy et al. | Jan 2014 | B2 |
8635690 | Alperovitch et al. | Jan 2014 | B2 |
8650608 | Ono et al. | Feb 2014 | B2 |
8661505 | Kougiouris et al. | Feb 2014 | B2 |
8677453 | Chang et al. | Mar 2014 | B2 |
8713682 | Shulman et al. | Apr 2014 | B2 |
8713694 | Harada et al. | Apr 2014 | B2 |
8719447 | Gupta et al. | May 2014 | B2 |
8819663 | Goel et al. | Aug 2014 | B2 |
8819814 | Leach et al. | Aug 2014 | B1 |
8843153 | Luna et al. | Sep 2014 | B2 |
8844040 | Harris et al. | Sep 2014 | B2 |
8849899 | L'Heureux et al. | Sep 2014 | B1 |
8849902 | Fiatal | Sep 2014 | B2 |
8850598 | Pelley et al. | Sep 2014 | B2 |
8869234 | Radkowski et al. | Oct 2014 | B2 |
8874761 | Backholm | Oct 2014 | B2 |
8893215 | Burghart | Nov 2014 | B2 |
8898340 | Vange et al. | Nov 2014 | B2 |
8903918 | Bharadhwaj et al. | Dec 2014 | B2 |
8910255 | Nice et al. | Dec 2014 | B2 |
8938773 | Hubbard | Jan 2015 | B2 |
8966576 | Giambiagi et al. | Feb 2015 | B2 |
8972493 | Cobb et al. | Mar 2015 | B2 |
8977750 | Ferris | Mar 2015 | B2 |
8978140 | Hubbard et al. | Mar 2015 | B2 |
8984056 | Ehrlich et al. | Mar 2015 | B2 |
9003524 | Baddour et al. | Apr 2015 | B2 |
9015348 | Hofmann et al. | Apr 2015 | B2 |
9021021 | Backholm et al. | Apr 2015 | B2 |
9021573 | Natarajan et al. | Apr 2015 | B2 |
9047322 | Ou et al. | Jun 2015 | B2 |
9075788 | Roth et al. | Jul 2015 | B1 |
20010049793 | Sugimoto | Dec 2001 | A1 |
20020023145 | Orr et al. | Feb 2002 | A1 |
20020108059 | Canion et al. | Aug 2002 | A1 |
20030046576 | High et al. | Mar 2003 | A1 |
20030177390 | Radhakrishnan | Sep 2003 | A1 |
20040083305 | Wang et al. | Apr 2004 | A1 |
20040107360 | Herrmann et al. | Jun 2004 | A1 |
20040111643 | Farmer | Jun 2004 | A1 |
20040158705 | Burdett et al. | Aug 2004 | A1 |
20040236963 | Danford et al. | Nov 2004 | A1 |
20040268149 | Aaron | Dec 2004 | A1 |
20050021764 | Barrall et al. | Jan 2005 | A1 |
20050138412 | Griffin et al. | Jun 2005 | A1 |
20050182958 | Pham et al. | Aug 2005 | A1 |
20050229236 | Devgan et al. | Oct 2005 | A1 |
20050251851 | Patrick et al. | Nov 2005 | A1 |
20050251852 | Patrick et al. | Nov 2005 | A1 |
20050257245 | Patrick et al. | Nov 2005 | A1 |
20050273850 | Freund | Dec 2005 | A1 |
20060005237 | Kobata et al. | Jan 2006 | A1 |
20060048210 | Hildre et al. | Mar 2006 | A1 |
20060085850 | Mayfield et al. | Apr 2006 | A1 |
20060095953 | Frank | May 2006 | A1 |
20060212925 | Shull et al. | Sep 2006 | A1 |
20060236363 | Heard et al. | Oct 2006 | A1 |
20060253605 | Sundarrajan et al. | Nov 2006 | A1 |
20060277590 | Limont et al. | Dec 2006 | A1 |
20060294297 | Gupta | Dec 2006 | A1 |
20070002862 | Kanekar et al. | Jan 2007 | A1 |
20070118881 | Mitchell et al. | May 2007 | A1 |
20070150441 | Morris | Jun 2007 | A1 |
20070150934 | Fiszman et al. | Jun 2007 | A1 |
20070157287 | Lim | Jul 2007 | A1 |
20070169168 | Lim | Jul 2007 | A1 |
20070244987 | Pedersen et al. | Oct 2007 | A1 |
20070245409 | Harris et al. | Oct 2007 | A1 |
20080028445 | Dubuc et al. | Jan 2008 | A1 |
20080060051 | Lim | Mar 2008 | A1 |
20080071859 | Seed et al. | Mar 2008 | A1 |
20080184336 | Sarukkai et al. | Jul 2008 | A1 |
20090077086 | Malkin et al. | Mar 2009 | A1 |
20090083831 | Kanai | Mar 2009 | A1 |
20090164560 | Fiatal | Jun 2009 | A1 |
20090178107 | Karjoth et al. | Jul 2009 | A1 |
20090178108 | Hudis et al. | Jul 2009 | A1 |
20090178109 | Nice et al. | Jul 2009 | A1 |
20090178131 | Hudis et al. | Jul 2009 | A1 |
20090290508 | Puthenpura et al. | Nov 2009 | A1 |
20090292824 | Marashi et al. | Nov 2009 | A1 |
20090300600 | Sugai | Dec 2009 | A1 |
20090300707 | Garimella et al. | Dec 2009 | A1 |
20100023582 | Pedersen et al. | Jan 2010 | A1 |
20100057995 | Cao et al. | Mar 2010 | A1 |
20100100949 | Sonwane et al. | Apr 2010 | A1 |
20100325692 | Rissanen | Dec 2010 | A1 |
20110016180 | Bharadhwaj et al. | Jan 2011 | A1 |
20110043652 | King et al. | Feb 2011 | A1 |
20110047620 | Mahaffey et al. | Feb 2011 | A1 |
20110276446 | Gupta et al. | Nov 2011 | A1 |
20110296523 | Schechter et al. | Dec 2011 | A1 |
20120011271 | Zhao et al. | Jan 2012 | A1 |
20120030749 | Anderson et al. | Feb 2012 | A1 |
20120089700 | Safruti et al. | Apr 2012 | A1 |
20120185937 | Levy et al. | Jul 2012 | A1 |
20120290642 | Shaughnessy et al. | Nov 2012 | A1 |
20130055344 | Rissanen et al. | Feb 2013 | A1 |
20130103786 | Miglore | Apr 2013 | A1 |
20130145421 | Kougiouris et al. | Jun 2013 | A1 |
20130174215 | Mattsson | Jul 2013 | A1 |
20130219483 | Cho et al. | Aug 2013 | A1 |
20130297700 | Hayton et al. | Nov 2013 | A1 |
20130339940 | Goel et al. | Dec 2013 | A1 |
20140032759 | Barton et al. | Jan 2014 | A1 |
20140181889 | Black | Jun 2014 | A1 |
20140237543 | Cheeniyil et al. | Aug 2014 | A1 |
20150033288 | Lim | Jan 2015 | A1 |
20150046971 | Huh et al. | Feb 2015 | A1 |
Number | Date | Country |
---|---|---|
0025214 | May 2000 | WO |
2011008419 | Jan 2011 | WO |
2012056099 | May 2012 | WO |
2014079340 | May 2014 | WO |
Entry |
---|
“Akamais Application Acceleration Services: Transform the Internet into a Business-Ready Application Delivery Platform”; Akamai Technologies, Inc.; Oct. 2008 (pp. 12). |
“A Web Service Architecture for Enforcing Access Control Policies”, Science Direct, Electronic Notes in Theoretical Computer Science 142 (2006) 47-62 by Claudio Agostino Ardagna et al. (pp. 16). |
Number | Date | Country | |
---|---|---|---|
20150058921 A1 | Feb 2015 | US |
Number | Date | Country | |
---|---|---|---|
61226407 | Jul 2009 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13919018 | Jun 2013 | US |
Child | 14529132 | US | |
Parent | 12550409 | Aug 2009 | US |
Child | 13919018 | US |