Method and device for performing integrated caching in a data communication network

Information

  • Patent Grant
  • 8739274
  • Patent Number
    8,739,274
  • Date Filed
    Wednesday, June 29, 2005
    19 years ago
  • Date Issued
    Tuesday, May 27, 2014
    10 years ago
Abstract
A device that implements a method for performing integrated caching in a data communication network. The device is configured to receive a packet from a client over the data communication network, wherein the packet includes a request for an object. At the operating system/kernel level of the device, one or more of decryption processing of the packet, authentication and/or authorization of the client, and decompression of the request occurs prior to and integrated with caching operations. The caching operations include determining if the object resides within a cache, serving the request from the cache in response to a determination that the object is stored within the cache, and sending the request to a server in response to a determination that the object is not stored within the cache.
Description
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a non-provisional of and claims priority to provisional U.S. patent application No. 60/583,784, filed Jun. 30, 2004 and entitled “Method and Device for Performing Integrated Caching in a Data Communication Network”, the entirety of which is incorporated by reference herein.


BACKGROUND OF THE INVENTION

1. Field of the Invention


The present invention relates generally to data communication networks. In particular, the present invention relates to a method and device for performing caching of data in a data communication network.


2. Background


The growth rate of network traffic continues to accelerate, albeit not quite at the exponential rates of the late 1990's. This expanding pace puts strains on the infrastructure that carries that traffic. Various solutions have arisen to permit network operators to handle this increasing problem, including the development of caching technology. With caching, content may be reused and served to clients without burdening server infrastructure. Additionally, caches can permit content to be placed closer to the end user thereby improving response time while also reducing server infrastructure burden.


Caching works by monitoring object requests made by one or more clients to a server on a network and saving the result of such request or requests in the storage of an appliance on the network. Subsequent requests for that same object are intercepted by the cache which delivers the object from its storage rather than passing the request on to the origin server. In this manner, caching provides the dual benefit of lowering response times and lowering the load on the server infrastructure.


However, as the Internet has expanded, multiple modes of processing data have developed to provide various performance, security and other benefits to network traffic. These modes of processing, however, have limited the effectiveness and performance capabilities of caches. Caches traditionally are not able to handle or process the variety of forms that data can take in the ever more complex and processed network environments and therefore the function and usefulness of such caches are commensurately limited.


Traditionally, caches have also been limited by virtue of being unable to serve to users that must be authenticated and/or authorized by the network prior to the delivery of content to such user. Authentication and authorization features have not been part of cache functionality.


Additional limitations of caches have included the inability to process and cache Hypertext Transfer Protocol over Secure Sockets Layer (HTTPS) traffic. Traditional caches have no way of decrypting, comprehending or processing encrypted traffic and therefore are unable to cache any HTTPS object or to provide end-to-end encryption for cached objects.


Furthermore, the ability of the cache to work with other network services and technologies including load balancing technology and other types of acceleration and packet processing has been less then optimal. The problem arises because of processing duplications and other inefficiencies introduced by traditional modes of integration that typically require multiple memory copies and frequent context switching to effect the integration.


The traditional approach to integrating a cache with other types of network processing usually involve placing multiple products in a line so that each function is carried out at a separate and independent node. This approach introduces considerable additional processing and a number of other inefficiencies as described below. Even if the technologies are combined within the same product, the combination is usually implemented in such a way that the benefits of caching when integrated with any other processing are diminished in the process of integration. The limitation of such integration arises as a result of the excess processing, copying and context switching involved in making these technologies work together.


The excess copying and switching results from the fact that each request or object must first be copied into a new form in the processor memory space by the application. The copying into memory space is performed in order to enable the additional processing (such as encryption/decryption, compression, connection offload or any other type of processing) to be carried out. The request or object must then be copied back to the packet level prior to being processed by the cache itself. Each additional type of processing generally requires packet copying to processor buffers, processing by the application and then copying back out for service by the cache. This approach to process integration therefore introduces multiple additional memory copies putting additional and unnecessary burdens upon the network infrastructure. An embodiment of the present invention minimizes this switching of contexts and duplicative processing as will be described below.


SUMMARY OF THE INVENTION

In order to achieve maximum performance, a device in accordance with an embodiment of the present invention efficiently integrates caching with the processing of one or more of (1) Secure Sockets Layer (SSL) encryption/decryption; (2) compression; (3) packet processing; (4) load balancing; (5) Transmission Control Protocol (TCP) buffering; (6) Authentication, Authorization and Auditing (AAA), (7) defense against distributed denial of service attacks (8), defense against HTTP/S based attacks such as Nimda, Code-Red, MyDoom etc., (9) offload of network processing, and (10) other processing, such as techniques to protect networks from network based viruses or attacks.


In particular, a device in accordance with an embodiment of the present invention integrates the foregoing technologies at the operating system (OS) or kernel level of the device rather than in the user/application level as a process or module. By integrating the cache features at this level of the computational structure and hierarchy, an embodiment of the present invention can much more efficiently synchronize caching processes with other processing of the connections to be carried out by or at the appliance including SSL decryption and end point processing, as well as the various others mentioned herein or otherwise carried out at the appliance. Such other processes can also include, in addition to each of the processes described above, connection processing, global server load balancing (GSLB), Hypertext Transfer Protocol (HTTP) compression, TCP compression, defenses against TCP and HTTP distributed denial of service attacks, HTTP virus defense scanning, and the like.


By integrating the variety of network processing and the caching at the OS or kernel level, an embodiment of the invention greatly increases the efficiency of such integration. Working in the kernel space, an embodiment of the invention enables the cache to relate to a relevant object or request as a data structure where the cache maintains equal status to such data structure as each of the other applications. By carrying out tight integration in the kernel/OS space, the cache is also able to see the semantics/data structure of an HTTP or HTTPS request and to make more efficient caching decisions as a result of having access to such data structures. An embodiment of the present invention thereby avoids the problems of context switching and duplication associated with integrating in the more obvious, traditional approach of running a process in the user space.


Further features and advantages of the invention, as well as the structure and operation of various embodiments of the invention, are described in detail below with reference to the accompanying drawings. It is noted that the invention is not limited to the specific embodiments described herein. Such embodiments are presented herein for illustrative purposes only. Additional embodiments will be apparent to persons skilled in the relevant art(s) based on the teachings contained herein.





BRIEF DESCRIPTION OF THE DRAWINGS/FIGURES

The accompanying drawings, which are incorporated herein and form part of the specification, illustrate the present invention and, together with the description, further serve to explain the principles of the invention and to enable a person skilled in the relevant art(s) to make and use the invention.



FIG. 1 illustrates an example data communication network environment in which an embodiment of the present invention may be implemented.



FIG. 2 illustrates an example software architecture of an appliance that performs integrated caching in accordance with an embodiment of the present invention.



FIG. 3 illustrates a flowchart of a method for performing integrated caching in a device in accordance with an embodiment of the present invention.



FIG. 4 illustrates an example computer system-based implementation for performing caching functions in accordance with embodiments of the present invention.





The features and advantages of the present invention will become more apparent from the detailed description set forth below when taken in conjunction with the drawings, in which like reference characters identify corresponding elements throughout. In the drawings, like reference numbers generally indicate identical, functionally similar, and/or structurally similar elements. The drawing in which an element first appears is indicated by the leftmost digit(s) in the corresponding reference number.


DETAILED DESCRIPTION OF THE INVENTION

A. Example Data Communication Network Environment



FIG. 1 illustrates an example data communication network environment 100 in which an embodiment of the present invention may be implemented. As shown in FIG. 1, example data communication network environment 100 includes a plurality of clients 102a-102n, an appliance 104, and a plurality of servers 106a-106n. Each of clients 102a-102n are communicatively coupled to appliance 104 via a public data communication network 108, while appliance 104 is communicatively coupled to servers 106a-106n via a private data communication network 110. In an embodiment, public data communication network 108 comprises the Internet and private data communication network 110 comprises an enterprise network, although the invention is not so limited.


In accordance with an embodiment of the present invention, appliance 104 includes a cache management logic and also includes or has access to a storage medium which it utilizes to implement a cache. Using these features, appliance 104 monitors object requests made by clients 102a-102n to any of servers 106a-106n. Objects returned from servers 106a-106n in response to these object requests are stored in the cache by appliance 104. Subsequent requests for the same object from any of clients 102a-102n are intercepted by appliance 104, which attempts to deliver the object from the cache rather than passing the request on to servers 106a-106n. This provides the dual benefit of reducing both the time required respond to requests from clients 102a-102n and the load on the infrastructure supporting servers 106a-106n.


As will be described in more detail herein, in an embodiment of the present invention, appliance 104 integrates this caching functionality at the operating system (OS)/kernel level of its software architecture with one or more other processing tasks, including but not limited to decryption, decompression, or authentication and/or authorization.


Example network environment 100 is presented by way of example only and is not intended to be limiting. Based on the teachings provided herein, persons skilled in the relevant art(s) will readily appreciate that the present invention may be implemented in any network environment in which object requests and responses are transferred between network nodes.


B. Integrated Caching in a Data Communication Network in Accordance with an Embodiment of the Present Invention



FIG. 2 illustrates an example software architecture 200 of appliance 104. Software architecture 200 is provided by way of illustration only and is not intended to be limiting. As shown in FIG. 2, example software architecture 200 consists of a hardware layer 206 and a software layer divided into a user space 202 and a kernel space 204.


Hardware layer 206 provides the structures upon which programs and services within kernel space 204 and user space 202 are executed. Hardware layer 206 also provides the structures which allow programs and services within kernel space 204 and user space 202 to communicate data both internally and externally with respect to appliance 104. As shown in FIG. 2, hardware layer 206 includes at least a central processing unit (CPU) 262 for executing software programs and services, a memory 264 for storing software and data, network ports 265 for transmitting and receiving data over a network, and an SSL processor 260 for performing functions related to Secure Sockets Layer processing of data transmitted and received over the network.


The operating system of appliance 104 segregates the available system memory into kernel space 204 and user space 204. In example software architecture 200, the operating system is depicted as a Unix operating system although the invention is not so limited. Kernel space 204 is strictly reserved for running Unix kernel 230, device drivers and any kernel extensions. In accordance with an embodiment of the present invention, kernel space 204 also includes a number of network services or processes working in conjunction with an integrated cache 232, the benefits of which are described in detail elsewhere herein.


As shown in FIG. 2, kernel space 204 includes at least a high-speed layer 2-7 integrated packet engine 240, an SSL engine 234, a policy engine 236 and multi-protocol compression logic 238. High speed layer 2-7 integrated packet engine 266 is responsible for managing the kernel-level processing of packets received and transmitted by appliance 104 via network ports 266. To this end, high speed layer 2-7 integrated packet engine 266 works in conjunction with SSL engine 234, integrated cache 232, policy engine 236 and multi-protocol compression logic 238. In particular, SSL engine is configured to perform SSL processing of packets, policy engine 236 is configured to perform functions related to traffic management such as request-level content switching and request-level cache redirection, and multi-protocol compression logic 238 is configured to perform functions related to compression and decompression of data. Integrated cache 232 is configured to perform an integrated caching function as described elsewhere herein.


In contrast to kernel space 204, user space 202 is the memory area used by all user mode applications. A user mode application cannot access kernel space 204 directly and must use service calls in order to access kernel services. As shown in FIG. 2, user space 202 of appliance 104 includes at least a graphical user interface (GUI) 210, a command line interface (CLI) 212, shell services 214, health monitoring programs 216, and daemon services 218. GUI 210 and CLI 212 provide a means by which a system administrator can interact with and control the operation of appliance 104. Health monitoring programs 216 are used to ensure that network systems are functioning properly and that users are receiving requested content over a network. Daemon services 218 are programs that run continuously and handle periodic service requests received by appliance 104. Each daemon program forwards the requests to other programs (or processes) as appropriate.



FIG. 3 illustrates a flowchart 300 of a sequence of events that may occur in an appliance that provides integrated caching functionality in accordance with an embodiment of the present invention. However, the invention is not limited to the description provided by the flowchart 300. Rather, it will be apparent to persons skilled in the relevant art(s) from the teachings provided herein that other functional flows are within the scope and spirit of the present invention. These other functional flows could involve different processing, different sequencing and other variations on the integration of caching that is the subject of the present invention.


The method of flowchart 300 can be implemented in one or more device(s) that are communicatively coupled to a data communication network. For example, the method of flowchart 300 can be implemented in an appliance such as appliance 104 described above in reference to FIG. 1, having a software architecture 200 as described above in reference to FIG. 2. The method of flowchart 300 will be described with continued reference to this exemplary embodiment, although the invention is not so limited.


As shown in FIG. 3, the method of flowchart 300 begins at step 302, in which appliance 104 receives an encrypted packet from one of clients 102a-102n. In an embodiment, appliance 104 is configured to act as a proxy SSL endpoint for servers 106a-106n, decrypting encrypted packets received from clients 102a-102n, and then sending them on for further processing as necessary and ultimately on to an appropriate resource based on address information within the encrypted packets. The appropriate resource may be, for example, any of servers 106a-106n or the cache managed by appliance 104. At step 304, appliance 104 performs decryption processing on the packet.


At step 306, appliance 104, which is configured in accordance with an embodiment of the present invention to carry out AAA policies for access control, authenticates and/or authorizes the client from which the encrypted packet was received.


At step 308, appliance 104, which is configured in accordance with an embodiment of the present invention to perform certain types of packet processing, carries out packet processing on the decrypted packets to reduce the connection overhead processing requirements generated by the applicable network protocols.


At step 310, appliance 104, which is configured in accordance with an embodiment of the present invention to compress and decompress content, decompresses a request associated with the packet. In an embodiment, the request comprises a web object request, although the invention is not so limited.


At step 312, appliance 104 is then able to activate the cache functionality, which receives a clear and/or authorized and/or decompressed and/or packet-processed request for an object. Because of the prior processing described in reference to steps 302, 304, 306, 308 and 310, the cache management logic can make a decision as to whether the object has been cached or is cacheable based on a clear/authorized/decompressed/packet processed request and is therefore able to process a much wider array of requests then traditional caches and to carry out the caching more efficiently than under traditional approaches. Furthermore, because the cache management logic is working in the kernel space along with the other processes, it relates to the relevant object as a data structure with equal status in relation to such data structure as each of the other applications and therefore the integration is carried out in an extremely efficient manner.


As shown at step 314, if the object is not already in the cache memory, appliance 104 sends a request on to one or more of servers 106a-106n. Before the request is sent, however, several additional processing steps may occur. For example, at step 316, appliance 104 optionally performs connection processing to ensure efficient transit of the request to the server(s) and at step 318, appliance 104 optionally makes a load balancing decision to ensure that the request is sent to the most appropriate server(s). Also, in an embodiment, the request is encrypted before it is sent to the server(s) via a back-end encryption process, thereby providing end-to-end network security. At step 320, the request is transmitted to the server(s).


At step 322, appliance 104 receives a response back from one of servers 106a-106n. If back-end encryption is supported as discussed above, appliance 104 decrypts the response from the server.


At step 324, appliance 104 compresses an object associated with the response from the server. In an embodiment, the object comprises a web object, although the invention is not so limited.


At step 326, the cache management logic in appliance 104 stores the object in the cache in compressed form. The cache management logic is able to store compressed objects in this fashion due to the processing abilities of the present invention. Once the object is stored in the cache, future client requests for the object can be served from the cache without performance of steps 316, 318, 320, 322, 324 and 326 as described above. This is indicated by the line directly connecting decision step 314 to step 328 in flowchart 300.


At step 328, after the object has been received from a server or retrieved from the cache, appliance 104 performs packet processing on the connection to more efficiently service the original client request. At step 330, the response object is then re-encrypted and delivered back to the client.


Each of the processing steps described above occurs at the kernel/OS level of appliance 104. By implementing the cache in the middle of, and integrated with, other processing steps in the kernel/OS space, an embodiment of the present invention is able to bring out additional functionality and improve performance of the cache.


Such integration permits a cache implementation in accordance with an embodiment of the invention to perform additional functions that are traditionally beyond the functional abilities of a cache. For example, an embodiment of the present invention permits the cache to work with encrypted and/or compressed objects.


Another example of additional functionality that may be achieved by an embodiment of the present invention involves the caching of end-to-end encrypted HTTPS traffic. Typically, caches only store unencrypted HTTP responses from servers. Certain caches may in some cases support SSL encrypted HTTPS delivery from the cache to the clients but, in any case, traditional caches are not able to cache responses that have been encrypted by the server and so are unable to support end-to-end (i.e. server to client) encryption. Typically, when a response is encrypted by the server in the form of HTTPS, the cache is not able to decrypt such a response and is therefore unable to store the response in its cache memory. For this reason, traditional caches fail to provide any benefit in the face of end-to-end encrypted traffic. In an embodiment of the present invention, the integrated caching appliance serves as a two-way termination point for the SSL encrypted HTTPS traffic.


For example, in a embodiment of the present invention, the integrated caching appliance acts as a termination point both to encrypted traffic between the server and the appliance, and between the appliance and the clients. In this manner, the appliance is able to decrypt and cache HTTPS-encrypted responses received from servers and when serving such responses to a client, re-encrypt such response and securely deliver it to the requesting client, thereby enabling end-to-end encryption and thus increasing the applicability of caching to a wider variety of web traffic.


In an embodiment of the present invention, the appliance can also serve as an endpoint in an SSL virtual private network (SSL VPN). In particular, the appliance can act as a proxy SSL endpoint for any resource in a private data communication network, decrypting encrypted packets received from a client and then sending them on to the appropriate destination server resource based on address information within the encrypted packets. As described in commonly-owned co-pending U.S. patent application Ser. No. 11/039,946 entitled “System and Method for Establishing a Virtual Private Network,” filed on Jan. 24, 2005, the entirety of which is incorporated by reference herein, a data communication session established between client and a gateway may be encrypted with the gateway serving as an encryption endpoint as described in the preceding paragraphs of the present application. As described in the referenced application, the client may use Secure Sockets Layer (SSL), IPSec, or some other encryption method to establish the encrypted data communication session by which an interception mechanism on the client directs traffic to the gateway while making the client browser think it is communicating directly with the destination servers or destination networks. In such an embodiment, the encrypted data communication session can be terminated at the gateway, which also includes an integrated cache as described herein. In this way caching functionality can be integrated into the SSL VPN functionality.


The gateway can also perform any applicable AAA policies to the request and consequently, the gateway will serve cached objects only to appropriately authenticated clients, as well as permitting requests only for users authorized to access a particular cached object. This is possible because the cache is integrated in such a way that the access control policies of the gateway are enforced before the cache sees any particular request. Thus, cached objects get the benefit of access control without the cache itself needing to perform the authentication and authorization.


Through the integration of the cache with such other functions, the cache itself becomes more efficient and more effective at handling the variety of data that passes across today's networks. An embodiment of the present invention also is able to improve the efficiency of the overall network performance by introducing the benefits of cache functionality to a broader array of web traffic.


Some other unique results of the mode of integration described above in accordance with an embodiment of the present invention are as follows. One result is the ability to cache pre-compressed data and serve it to compression-aware clients. Another result is the ability to cache access-controlled data. Yet another result is the ability to work with external caches to provide scalability of the cache. Because the cache is integrated with redirection and traffic management capabilities at the gateway, external caches can be deployed to provide a second-tier of caching thereby extending the capacity (and the benefits) of caching significantly. Through an embodiment of the present invention, this capacity is created without the cache module itself having to explicitly perform cache redirection policies.


In terms of performance, by integrating the cache as described above, the processors of the cache are freed from performing the variety of connection processing tasks that caches, acting as a nodes on a network, are traditionally required to perform, and are thus able to perform its caching functions at their highest performance levels. Indeed, by enabling the caching of compressed data, the cache is able to function even more efficiently and allow users to realize even higher performance.


As previously noted in this application, the efficiency arises as a result of the way the cache is integrated with the other network services and technologies including load balancing technology, encryption, AAA, compression and other types of acceleration and packet processing. As a result, processing duplications and other inefficiencies introduced by traditional modes of integration are avoided. These inefficiencies, caused by unnecessary copying and context switching, arise because each object received by the device must be copied to a message and then into a processor memory prior to processing by the relevant application. The request must then be copied back to the object or packet level for processing by the cache introducing additional memory copies. In contrast, an embodiment of the present invention carries out the integration at the OS or kernel level, thereby enabling the cache to operate on the object as a data structure where the cache has equal status as the other applications and/or processes in relating to and processing such data structure and where the need for such additional memory copies is obviated as all processes are working with the same data structures. The result is a more efficient integration.


C. Caching with Proactive Validation in a Data Communication Network in Accordance with an Embodiment of the Present Invention


Because web objects can change over time, each potentially cacheable object is said to have a useful life, or “freshness”. The concept of freshness refers to the fact that the application server that originally generated the content also determines the period of time that such object can be served by a cache that may store such object. Caches must be able to determine whether or not the copy of an object stored in its memory is still “fresh,” or whether the cache needs to retrieve a new copy of the object from the origin server. An embodiment of the present invention implements a novel approach to assuring object freshness.


Many conventional cache implementations try to keep the cached content fresh by fetching the content from the origin on a pre-determined schedule. The fetching of content from the origin occurs at times established by the cache administrator typically based on one or both of the following approaches: either at (i) regular specified intervals or (ii) when the content is about to expire.


There are two problems typically associated with the above commonly-employed approaches. First, unnecessary processing loads are imposed upon the origin server because that server is required to provide content to the cache requesting the refreshment (whether such refresh occurs at specified intervals or as the content is about to expire) without regard to whether such content will ultimately be served to clients. Second the cache incurs additional processor load based on the extra processing overhead generated because the cache needs to keep track of the elements that must be refreshed and the time at which they have to be refreshed.


A cache in accordance with an embodiment of the present invention solves the above problems using a novel pre-fetching approach. The pre-fetching of the content is not performed in accordance with a predefined schedule or just prior to expiration of the content. Instead, an embodiment of the present invention performs pre-fetching only when both of the following conditions have been met: (1) a client has made a request for the specified content and (2) that content is ‘about to expire’.


This approach addresses both problems described above. Pro-active revalidation is more likely to generate a request for refreshing of content from the origin server only where such content is being actively accessed. This minimizes the amount of ‘unnecessary’ load on the origin server. As discussed above, where the cache requests refreshment of objects that are not ultimately served to clients (or only rarely get served depending on the sensitivity of the cache), the cache is inefficiently utilizing both its own resources as well as the resources of the origin server. An embodiment of the present invention avoids the inefficient use of the cache and server resources by requesting only that content that is being actively accessed. The approach also, for the same reason, reduces the bandwidth used for pre-fetching and therefore makes more efficient use of network resources than traditional approaches.


Furthermore, an embodiment of the present invention uses the expiry information included in the cached object itself to determine whether to request refreshment of the object from the origin server. Such expiry information is typically included in the headers of the relevant object. This embodiment thus avoids the inefficiencies of storing any additional information for fetching unlike many traditional approaches which require the cache to keep a table tracking the schedule for refreshment.


Using a ‘demand-based’ pre-fetching technique also enhances benefits that are inherent to pre-fetching. This technique reduces the number of cache misses for frequently accessed objects since such objects are very likely to undergo pro-active revalidation, just before they expire. This technique can also prevent the surge of traffic to an origin server that can occur when a large response that is in great demand expires. In the traditional approach, all of the requests for such content miss the cache and get sent to the origin server because the cache content has expired. By contrast, in an embodiment of the present invention, the content of the cache memory will generally be refreshed just prior to expiration and therefore the situation where cache misses occur while the cache is refreshing are much less likely to arise.


In an embodiment of the present invention, the aggressiveness of pre-fetching can be controlled through adjusting the length of the duration before the expiry where the content is determined to be about to expire and also the number of client requests required to trigger refreshment by the cache of the relevant object.


D. Optimizing Processing of Large Non-Cacheable Responses Using “Negative Cells” in Accordance with an Embodiment of the Present Invention


In accordance with an embodiment of the present invention, the cache recognizes and does not store objects that are above a specified size in order to improve the object hit ratio. Caches typically have limited memory space devoted to storing cached objects and therefore certain responses that exceed allocated memory space are ultimately rejected as non-cacheable and not stored by the cache. With traditional caches, the cache attempts to store the large response in its cache memory and only aborts storing the response once the cache recognizes that the response size exceeds a predefined maximum size. Traditional caches will repeatedly attempt to cache the large response each time a request for such response is received by the cache from the server. In each case, the cache will need to determine that the object is non-cacheable as exceeding the memory space. Thus, this is a manifestly inefficient approach.


In accordance with an embodiment of the present invention, the cache employs an optimization to avoid expending effort in storing such responses. Whenever the cache detects a response that becomes non-cacheable due to response size, it stores a notation regarding the corresponding request in a data structure termed a “negative cell.” The notation indicates that the request is non-cacheable. In the future, when a client requests the same object, the request is matched to the notation regarded the first request stored in the data structure. Based on the match, the cache will not try to cache the response and instead the request will completely bypass the cache.


There is no user configuration required for specifying the duration for which a negative cell should remain in the cache. In fact, the users are not even aware that this particular mechanism is being employed. In an embodiment, the cache uses the regular expiry information that it would have employed to cache the big response, to cache the “negative information” about that response.


E. Exemplary Computer System-Based Implementation


The functions of the present invention may be implemented using hardware, software, or a combination thereof and may be implemented in one or more computer systems or other processing systems. For example, FIG. 4 depicts an example computer system 400 that may be utilized to implement a the functions of the present invention.


As shown in FIG. 4, the example computer system 400 includes a processor 404 for executing software routines in accordance with embodiments of the present invention. Although a single processor is shown for the sake of clarity, the computer system 400 may also comprise a multi-processor system. The processor 404 is connected to a communication infrastructure 406 for communication with other components of the computer system 400. The communication infrastructure 406 may comprise, for example, a communications bus, cross-bar, or network.


Computer system 400 further includes a main memory 408, such as a random access memory (RAM), and a secondary memory 410. The secondary memory 410 may include, for example, a hard disk drive 412 and/or a removable storage drive 414, which may comprise a floppy disk drive, a magnetic tape drive, an optical disk drive, or the like. The removable storage drive 414 reads from and/or writes to a removable storage unit 418 in a well known manner. Removable storage unit 418 may comprise a floppy disk, magnetic tape, optical disk, or the like, which is read by and written to by removable storage drive 414. As will be appreciated by persons skilled in the relevant art(s), the removable storage unit 418 includes a computer usable storage medium having stored therein computer software and/or data.


In alternative embodiments, secondary memory 410 may include other similar means for allowing computer programs or other instructions to be loaded into computer system 400. Such means can include, for example, a removable storage unit 422 and an interface 420. Examples of a removable storage unit 422 and interface 420 include a program cartridge and cartridge interface (such as that found in video game console devices), a removable memory chip (such as an EPROM, or PROM) and associated socket, and other removable storage units 422 and interfaces 420 which allow software and data to be transferred from the removable storage unit 422 to computer system 400.


Computer system 400 further includes a display interface 402 that forwards graphics, text, and other data from the communication infrastructure 406 or from a frame buffer (not shown) for display to a user on a display unit 430.


Computer system 400 also includes at least one communication interface 424. Communication interface 424 allows software and data to be transferred between computer system 400 and external devices via a communication path 426. In embodiments of the present invention, the communication interface 424 permits data to be transferred between the computer system 400 and a data communication network, such as a public data or private data communication network. Examples of communications interface 424 can include a modem, a network interface (such as Ethernet card), a communications port, and the like. Software and data transferred via communications interface 424 are in the form of signals which can be electronic, electromagnetic, optical or other signals capable of being received by communications interface 424. These signals are provided to the communications interface via the communication path 426.


As used herein, the term “computer program product” may refer, in part, to removable storage unit 418, removable storage unit 422, a hard disk installed in hard disk drive 412, or a carrier wave carrying software over a communication path 426 (wireless link or cable) to communication interface 424. A computer useable medium can include magnetic media, optical media, or other recordable media, or media that transmits a carrier wave or other signal. These computer program products are means for providing software to computer system 400.


Computer programs (also called computer control logic) are stored in main memory 408 and/or secondary memory 410. Computer programs can also be received via communications interface 424. Such computer programs, when executed, enable the computer system 400 to perform one or more features of the present invention as discussed herein. In particular, the computer programs, when executed, enable the processor 404 to perform features of the present invention. Accordingly, such computer programs represent controllers of the computer system 400.


The present invention can be implemented as control logic in software, firmware, hardware or any combination thereof. In an embodiment where the invention is implemented using software, the software may be stored in a computer program product and loaded into computer system 400 using removable storage drive 414, hard disk drive 412, or interface 420. Alternatively, the computer program product may be downloaded to computer system 400 over communications path 426. The software, when executed by the processor 404, causes the processor 404 to perform functions of the invention as described herein.


G. Conclusion


While various embodiments of the present invention have been described above, it should be understood that they have been presented by way of example only, and not limitation. Thus, it will be understood by those skilled in the relevant art(s) that various changes in form and details may be made therein without departing from the spirit and scope of the invention as defined in the appended claims. Accordingly, the breadth and scope of the present invention should not be limited by any of the above-described exemplary embodiments, but should be defined only in accordance with the following claims and their equivalents.

Claims
  • 1. A method for performing integrated caching at a kernel level of an operating system to more efficiently process network communications comprising the steps of: (a) receiving, by a packet engine of an appliance in communication with a cache of the appliance at a kernel level, an encrypted response from a server to a request of a client, the encrypted response providing an object, the packet engine operating at the kernel level of an operating system of the appliance and providing the object as a kernel level data structure of the packet engine stored in memory of kernel space;(b) decrypting, by an Secure Socket Layer (SSL) engine of the appliance, the encrypted response from the same kernel level data structure of the packet engine stored in memory to provide a decrypted response, the SSL engine in communication with the packet engine and operating at the kernel level;(c) storing, by a cache management logic of the appliance, the object of the decrypted response in a cache, the cache management logic in communication with the packet engine and operating at the kernel level, the cache management logic operating on the object from the same kernel level data structure of the packet engine stored in memory accessed by one of the packet engine or the SSL engine;(d) determining, by a policy engine of the appliance, the client is authorized to access the object identified by the decrypted request, the policy engine in communication with the packet engine and operating at the kernel level;(e) transmitting, by the packet engine, one of the encrypted response or the decrypted response to the client.
  • 2. The method of claim 1, comprising the steps of: (f) receiving, by the packet engine, an encrypted request from the client for the object; and(g) decrypting, by the SSL engine, the encrypted request to provide a decrypted request.
  • 3. The method of claim 2, comprising the steps of: (h) determining, by the cache management logic, via the decrypted request that the encrypted request identifies the object stored in the cache;(i) encrypting, by the SSL engine, a response having the object stored in the cache; and(j) transmitting, by the packet engine, the encrypted response to the client in response to the encrypted request.
  • 4. The method of claim 1, comprising receiving a second request for the object from a second client and determining, by the policy engine, the second client is not authorized to access the object identified by the decrypted request, and the appliance not processing the second request via the cache management logic in response to the determination.
  • 5. The method of claim 1, comprising the steps determining, by the cache management logic, via the decrypted request that the encrypted request identifies the object stored in the cache, and transmitting, by the appliance, the object to the client.
  • 6. The method of claim 1, wherein step (c) comprises compressing, by a compression logic of the appliance, the object of the decrypted response to provide a compressed object, and the cache management logic storing the compressed object in the cache, the compression logic operating at a kernel level of the operating system of the appliance.
  • 7. The method of claim 6, comprising operating, by the compression logic, on the object using the same kernel level data structure of the packet engine accessed by one of the packet engine or the SSL engine.
  • 8. The method of claim 1, wherein step (d) comprises encrypting, by the SSL engine, the decrypted response, and transmitting, by the packet engine, the response encrypted by the SSL engine to the client.
  • 9. An appliance for performing integrated caching at a kernel level of an operating system to more efficiently process network communications, the appliance comprising a packet engine in communication with a cache at the kernel level, the appliance comprising: a hardware processor coupled to a memory, the hardware processor executing a packet engine, an Secure Socket Layer (SSL) engine, and a cache management logic, and a policy engine, and wherein:the packet engine is configured for receiving an encrypted response from a server to a request of a client, the encrypted response providing an object, the packet engine operating at a kernel level of the operating system of the appliance and providing the object as a kernel level data structure of the packet engine stored in a kernel space of the memory;the SSL engine is configured for decrypting the encrypted response via the same kernel level data structure of the packet engine stored in the kernel space of the memory to provide a decrypted response, the SSL engine in communication with the packet engine and operating at the kernel level;the cache management logic is configured for storing the object of the decrypted response in a cache, the cache management logic in communication with the packet engine and operating at the kernel level, the cache management logic operating on the object from the same kernel level data structure of the packet engine stored in the kernel space of the memory accessed by one of the packet engine or the SSL engine;wherein the policy engine is configured for determining the client is authorized to access the object identified by the decrypted request, the policy engine in communication with the packet engine and operating at the kernel level; andwherein the packet engine is further configured for transmitting one of the encrypted response or the decrypted response to the client.
  • 10. The appliance of claim 9, wherein the packet engine is further configured for receiving an encrypted request from the client for the object and the SSL engine is further configured for decrypting the encrypted request.
  • 11. The appliance of claim 9, wherein the cache management logic is further configured for determining via the decrypted request that the encrypted request identifies the object stored in the cache; the SSL engine is further configured for encrypting a response having the object stored in the cache; and the packet engine is further configured for transmitting the encrypted response to the client in response to the encrypted request.
  • 12. The appliance of claim 9, wherein the processor further executes a policy engine configured for determining the client is not authorized to access the object identified by the decrypted request, the policy engine in communication with the packet engine and operating at the kernel level and the appliance not processing the decrypted request via the cache management logic in response to the determination.
  • 13. The appliance of claim 9, wherein the cache management logic is further configured for determining using the decrypted request that the encrypted request identifies the object stored in the cache, and the packet engine is further configured for transmitting the object to the client.
  • 14. The appliance of claim 9, wherein the processor further executes a compression logic configured for compressing the object of the decrypted response, the compression logic operating at a kernel level of the operating system of the appliance and wherein the cache management logic stores the compressed object to the cache.
  • 15. The appliance of claim 14, wherein the compression logic is further configured for operating on the object using the same kernel level data structure of the packet engine accessed by one of the packet engine or the SSL engine.
  • 16. The appliance of claim 9, wherein the SSL engine is further configured for encrypting the decrypted response and the packet engine is further configured for transmitting the response encrypted by the SSL engine to the client.
  • 17. A method for performing integrated decryption, authentication and authorization at a kernel level of an operating system to more efficiently process network communications comprising the steps of: (a) receiving, by a packet engine of an appliance in communication with a policy engine and an Secure Socket Layer (SSL) engine of the appliance at the kernel level, an encrypted request from a client to a server, the encrypted request requesting access to an object, the packet engine operating at a kernel level of the operating system of the appliance and providing the object as a kernel level data structure of the packet engine stored in memory of kernel space;(b) decrypting, by the SSL engine of the appliance, the encrypted request from the same kernel level data structure of the packet engine stored in memory to provide a decrypted request, the SSL engine in communication with the packet engine and operating at the kernel level;(c) authenticating, by the policy engine of the appliance, the client, the policy engine in communication with the packet engine and operating at the kernel level;(d) determining, by the policy engine, a user of the client is authorized to access the object identified by the decrypted request, the policy engine operating on the object via the same kernel level data structure of the packet engine stored in memory; and(e) determining, by a cache management logic of the appliance, that the object identified by the decrypted request is stored in cache, the cache management logic in communication with the packet engine and operating at the kernel level, the cache management logic identifying the object using the same kernel level data structure of the packet engine accessed by one of the packet engine or the policy engine; and(f) transmitting, by the packet engine, to the client in response to the encrypted request a response encrypted by the SSL engine and having the object stored in the cache.
  • 18. The method of claim 17, comprising a step of: (g) determining, by a cache management logic of the appliance, that the object identified by the decrypted request is not stored in cache, the cache management logic in communication with the packet engine and operating at the kernel level, the cache management logic identifying the object using the data structure accessed by one of the packet engine or the policy engine; and(h) transmitting, by the packet engine, the decrypted request to the server.
  • 19. A method for performing integrated caching and compression at a kernel level of an operating system to more efficiently process network communications comprising the steps of: (a) receiving, by a packet engine of an appliance in communication with a cache, compression logic, and an Secure Socket Layer (SSL) engine of the appliance at the kernel level, an encrypted response from a server to a request of a client, the response providing an object, the packet engine operating at a kernel level of the operating system of the appliance and providing the object as a kernel level data structure stored in memory of kernel space, the encrypted response decrypted by the SSL engine of the appliance via the same kernel level data structure of the packet engine;(b) compressing, by the compression logic of the appliance, the object of the response, the compression logic operating at a kernel level of the operating system of the appliance to provide a compressed object, and accessing the object via the same kernel level data structure of the packet engine stored in memory;(c) storing, by the cache management logic of the appliance, the compressed object of the response in a cache, the cache management logic in communication with the packet engine and operating at the kernel level, the cache management logic operating on the object using the same kernel level data structure of the packet engine stored in memory accessed by one of the packet engine or the compression logic;(d) determining, by a policy engine of the appliance, the client is authorized to access the object identified by the decrypted request, the policy engine in communication with the packet engine and operating at the kernel level; and(e) transmitting, by the packet engine, the response to the client.
  • 20. The method of claim 19, comprising receiving, by the packet engine, a second request for the object, determining, by the cache management logic, the object is stored compressed in the cache, and transmitting, by the packet engine, the compressed object in response to the second request.
  • 21. A method for performing integrated caching at a kernel level of an operating system to more efficiently process network communications, the method comprising the steps of: (a) receiving, by a packet engine of an appliance deployed between a client and a server, an encrypted response from the server to a request of the client, the encrypted response comprising an object, the packet engine operating at a kernel level of a kernel of an operating system of the appliance;(b) providing, from a kernel level data structure stored in memory of kernel space, the object for each of the packet engine, an Secure Socket Layer (SSL) engine of the appliance and a cache management logic of the appliance, each of the packet engine, the SSL engine and the cache management logic accessing the object within the same kernel level data structure in the kernel space without using a service call, each of the packet engine, the SSL engine and the cache management logic having an equal status to access the object via the kernel level data structure stored in memory;(c) decrypting, by the SSL engine, the encrypted response from the same kernel level data structure stored in memory and without using a service call to access the kernel, the SSL engine in communication with the packet engine and operating at the kernel level;(d) storing, by a cache management logic of the appliance, the encrypted object from the encrypted response in a cache, the cache management logic in communication with the packet engine and operating on the object from the same kernel level data structure stored in memory and without using a service call to access the kernel;(e) determining, by a policy engine of the appliance, the client is authorized to access the object identified by the decrypted request, the policy engine in communication with the packet engine and operating at the kernel level; and(f) transmitting, by the packet engine, one of the encrypted response or the decrypted response to the client.
US Referenced Citations (790)
Number Name Date Kind
4479195 Herr et al. Oct 1984 A
4701844 Thompson et al. Oct 1987 A
4885680 Anthony et al. Dec 1989 A
4935870 Burk, Jr. et al. Jun 1990 A
5301270 Steinberg et al. Apr 1994 A
5307413 Denzer Apr 1994 A
5329619 Page et al. Jul 1994 A
5359712 Cohen et al. Oct 1994 A
5511208 Boyles et al. Apr 1996 A
5519699 Ohsawa May 1996 A
5521940 Lane et al. May 1996 A
5561769 Kumar et al. Oct 1996 A
5623492 Teraslinna Apr 1997 A
5625793 Mirza Apr 1997 A
5657390 Elgamal et al. Aug 1997 A
5671226 Murakami et al. Sep 1997 A
5708656 Noneman et al. Jan 1998 A
5742829 Davis et al. Apr 1998 A
5758085 Kouoheris et al. May 1998 A
5758110 Boss et al. May 1998 A
5761431 Gross et al. Jun 1998 A
5787470 DeSimone et al. Jul 1998 A
5812668 Weber Sep 1998 A
5815462 Konishi et al. Sep 1998 A
5819020 Beeler, Jr. Oct 1998 A
5822524 Chen et al. Oct 1998 A
5828840 Cowan et al. Oct 1998 A
5835726 Shwed et al. Nov 1998 A
5838920 Rosborough Nov 1998 A
5850446 Berger et al. Dec 1998 A
5852717 Bhide et al. Dec 1998 A
5864837 Maimone Jan 1999 A
5881229 Singh et al. Mar 1999 A
5889863 Weber Mar 1999 A
5893150 Hagersten et al. Apr 1999 A
5911051 Carson et al. Jun 1999 A
5918244 Percival Jun 1999 A
5925100 Drewry et al. Jul 1999 A
5931917 Nguyen et al. Aug 1999 A
5931961 Ranganathan et al. Aug 1999 A
5933605 Kawano et al. Aug 1999 A
5940074 Britt et al. Aug 1999 A
5940516 Mason et al. Aug 1999 A
5943424 Berger et al. Aug 1999 A
5956483 Grate et al. Sep 1999 A
5958016 Chang et al. Sep 1999 A
5978840 Nguyen et al. Nov 1999 A
5983208 Haller et al. Nov 1999 A
5987132 Rowney Nov 1999 A
5987482 Bates et al. Nov 1999 A
5987611 Freund Nov 1999 A
5995999 Bharadhwaj Nov 1999 A
5996076 Rowney et al. Nov 1999 A
5999179 Kekic et al. Dec 1999 A
5999525 Krishnaswamy et al. Dec 1999 A
6002767 Kramer Dec 1999 A
6021470 Frank et al. Feb 2000 A
6023724 Bhatia et al. Feb 2000 A
6026379 Haller et al. Feb 2000 A
6026413 Challenger et al. Feb 2000 A
6026440 Shrader et al. Feb 2000 A
6029175 Chow et al. Feb 2000 A
6058250 Harwood et al. May 2000 A
6061715 Hawes May 2000 A
6061769 Kapulka et al. May 2000 A
6061796 Chen et al. May 2000 A
6067569 Khaki et al. May 2000 A
6072870 Nguyen et al. Jun 2000 A
6092155 Olnowich Jul 2000 A
6101543 Alden et al. Aug 2000 A
6112085 Garner et al. Aug 2000 A
6119105 Williams Sep 2000 A
6119151 Cantrell et al. Sep 2000 A
6122403 Rhoads Sep 2000 A
6128627 Mattis et al. Oct 2000 A
6131120 Reid Oct 2000 A
6141686 Jackowski et al. Oct 2000 A
6144775 Williams et al. Nov 2000 A
6157955 Narad et al. Dec 2000 A
6161123 Renouard et al. Dec 2000 A
6163772 Kramer et al. Dec 2000 A
6170017 Dias et al. Jan 2001 B1
6173325 Kukreja Jan 2001 B1
6175869 Ahuja et al. Jan 2001 B1
6178409 Weber et al. Jan 2001 B1
6182139 Brendel Jan 2001 B1
6192408 Vahalia et al. Feb 2001 B1
6219669 Haff et al. Apr 2001 B1
6233249 Katseff et al. May 2001 B1
6233606 Dujari May 2001 B1
6233619 Narisi et al. May 2001 B1
6236652 Preston et al. May 2001 B1
6252851 Siu et al. Jun 2001 B1
6253027 Weber et al. Jun 2001 B1
6253188 Witek et al. Jun 2001 B1
6253327 Zhang et al. Jun 2001 B1
6256773 Bowman-Amuah Jul 2001 B1
6272146 Bowater et al. Aug 2001 B1
6272148 Takagi et al. Aug 2001 B1
6272556 Gish Aug 2001 B1
6282172 Robles et al. Aug 2001 B1
6289382 Bowman-Amuah Sep 2001 B1
6289450 Pensak et al. Sep 2001 B1
6301249 Mansfield et al. Oct 2001 B1
6304915 Nguyen et al. Oct 2001 B1
6308273 Goertzel et al. Oct 2001 B1
6321235 Bird et al. Nov 2001 B1
6321337 Reshef et al. Nov 2001 B1
6324525 Kramer et al. Nov 2001 B1
6324582 Sridhar et al. Nov 2001 B1
6324647 Bowman-Amuah Nov 2001 B1
6324648 Grantges, Jr. Nov 2001 B1
6327242 Amicangioli et al. Dec 2001 B1
6332163 Bowman-Amuah Dec 2001 B1
6333931 LaPier et al. Dec 2001 B1
6334664 Silverbrook Jan 2002 B1
6339832 Bowman-Amuah Jan 2002 B1
6345239 Bowman-Amuah Feb 2002 B1
6363363 Haller et al. Mar 2002 B1
6363478 Lambert et al. Mar 2002 B1
6370573 Bowman-Amuah Apr 2002 B1
6370599 Anand et al. Apr 2002 B1
6373950 Rowney Apr 2002 B1
6383478 Prokop et al. May 2002 B1
6389462 Cohen et al. May 2002 B1
6389505 Emma et al. May 2002 B1
6398359 Silverbrook et al. Jun 2002 B1
6405364 Bowman-Amuah Jun 2002 B1
6412000 Riddle et al. Jun 2002 B1
6415329 Gelman et al. Jul 2002 B1
6427132 Bowman-Amuah Jul 2002 B1
6427187 Malcolm Jul 2002 B2
6431777 Silverbrook Aug 2002 B1
6434568 Bowman-Amuah Aug 2002 B1
6434618 Cohen et al. Aug 2002 B1
6434628 Bowman-Amuah Aug 2002 B1
6438127 Le Goff et al. Aug 2002 B1
6438594 Bowman-Amuah Aug 2002 B1
6442571 Haff et al. Aug 2002 B1
6442682 Pothapragada et al. Aug 2002 B1
6442748 Bowman-Amuah Aug 2002 B1
6447113 Silverbrook et al. Sep 2002 B1
6449658 Lafe et al. Sep 2002 B1
6449695 Bereznyi et al. Sep 2002 B1
6452915 Jorgensen Sep 2002 B1
6452923 Gerszberg et al. Sep 2002 B1
6457103 Challenger et al. Sep 2002 B1
6473794 Guheen et al. Oct 2002 B1
6473802 Masters Oct 2002 B2
6477580 Bowman-Amuah Nov 2002 B1
6477665 Bowman-Amuah Nov 2002 B1
6484143 Swildens et al. Nov 2002 B1
6484206 Crump et al. Nov 2002 B2
6487641 Cusson et al. Nov 2002 B1
6496481 Wu et al. Dec 2002 B1
6496776 Blumberg et al. Dec 2002 B1
6496850 Bowman-Amuah Dec 2002 B1
6496935 Fink et al. Dec 2002 B1
6502102 Haswell et al. Dec 2002 B1
6502192 Nguyen Dec 2002 B1
6502213 Bowman-Amuah Dec 2002 B1
6507891 Challenger et al. Jan 2003 B1
6519571 Guheen et al. Feb 2003 B1
6523027 Underwood Feb 2003 B1
6529909 Bowman-Amuah Mar 2003 B1
6529948 Bowman-Amuah Mar 2003 B1
6536037 Guheen et al. Mar 2003 B1
6539396 Bowman-Amuah Mar 2003 B1
6549949 Bowman-Amuah Apr 2003 B1
6550012 Villa et al. Apr 2003 B1
6550057 Bowman-Amuah Apr 2003 B1
6553129 Rhoads Apr 2003 B1
6553377 Eschelbeck et al. Apr 2003 B1
6553409 Zhang et al. Apr 2003 B1
6571282 Bowman-Amuah May 2003 B1
6574668 Gubbi et al. Jun 2003 B1
6574688 Dale et al. Jun 2003 B1
6578068 Bowman-Amuah Jun 2003 B1
6578073 Starnes et al. Jun 2003 B1
6584569 Reshef et al. Jun 2003 B2
6590588 Lincke et al. Jul 2003 B2
6591266 Li et al. Jul 2003 B1
6594690 Cantwell Jul 2003 B2
6598075 Ogdon et al. Jul 2003 B1
6601057 Underwood et al. Jul 2003 B1
6601192 Bowman-Amuah Jul 2003 B1
6601233 Underwood Jul 2003 B1
6601234 Bowman-Amuah Jul 2003 B1
6606660 Bowman-Amuah Aug 2003 B1
6606663 Liao et al. Aug 2003 B1
6606744 Mikurak Aug 2003 B1
6609128 Underwood Aug 2003 B1
6611262 Suzuki Aug 2003 B1
6611522 Zheng et al. Aug 2003 B1
6611822 Beams et al. Aug 2003 B1
6611867 Bowman-Amuah Aug 2003 B1
6614800 Genty et al. Sep 2003 B1
6615166 Guheen et al. Sep 2003 B1
6615199 Bowman-Amuah Sep 2003 B1
6615253 Bowman-Amuah Sep 2003 B1
6618359 Chen et al. Sep 2003 B1
6629081 Cornelius et al. Sep 2003 B1
6631986 Silverbrook Oct 2003 B2
6633574 Koch et al. Oct 2003 B1
6633878 Underwood Oct 2003 B1
6636242 Bowman-Amuah Oct 2003 B2
6640238 Bowman-Amuah Oct 2003 B1
6640240 Hoffman et al. Oct 2003 B1
6640244 Bowman-Amuah Oct 2003 B1
6640248 Jorgensen Oct 2003 B1
6640249 Bowman-Amuah Oct 2003 B1
6647130 Rhoads Nov 2003 B2
6650640 Muller et al. Nov 2003 B1
6652089 Silverbrook Nov 2003 B2
6652090 Silverbrook Nov 2003 B2
6662221 Gonda et al. Dec 2003 B1
6662357 Bowman-Amuah Dec 2003 B1
6664978 Kekic et al. Dec 2003 B1
6671818 Mikurak Dec 2003 B1
6678734 Haatainen et al. Jan 2004 B1
6681029 Rhoads Jan 2004 B1
6687732 Bector et al. Feb 2004 B1
6691227 Gopal et al. Feb 2004 B1
6697824 Bowman-Amuah Feb 2004 B1
6697844 Chan et al. Feb 2004 B1
6697849 Carlson Feb 2004 B1
6700902 Meyer Mar 2004 B1
6701345 Carley et al. Mar 2004 B1
6701514 Haswell et al. Mar 2004 B1
6704873 Underwood Mar 2004 B1
6715145 Bowman-Amuah Mar 2004 B1
6718535 Underwood Apr 2004 B1
6721713 Guheen et al. Apr 2004 B1
6728787 Leigh Apr 2004 B1
6732237 Jacobs et al. May 2004 B1
6732269 Baskey et al. May 2004 B1
6732314 Borella et al. May 2004 B1
6735601 Subrahmanyam May 2004 B1
6735691 Capps et al. May 2004 B1
6742015 Bowman-Amuah May 2004 B1
6744774 Sharma Jun 2004 B2
6754699 Swildens et al. Jun 2004 B2
6760748 Hakim Jul 2004 B1
6763467 Radatti et al. Jul 2004 B1
6766373 Beadle et al. Jul 2004 B1
6772203 Feiertag et al. Aug 2004 B1
6775392 Rhoads Aug 2004 B1
6785719 Jacobson et al. Aug 2004 B1
6788315 Kekic et al. Sep 2004 B1
6789170 Jacobs et al. Sep 2004 B1
6792615 Rowe et al. Sep 2004 B1
6801927 Smith et al. Oct 2004 B1
6802020 Smith Oct 2004 B1
6807607 Lamparter Oct 2004 B1
6820125 Dias et al. Nov 2004 B1
6820133 Grove et al. Nov 2004 B1
6820974 Silverbrook Nov 2004 B2
6823374 Kausik et al. Nov 2004 B2
6826626 McManus Nov 2004 B1
6826627 Sjollema et al. Nov 2004 B2
6831898 Edsall et al. Dec 2004 B1
6842906 Bowman-Amuah Jan 2005 B1
6847892 Zhou et al. Jan 2005 B2
6850252 Hoffberg Feb 2005 B1
6870921 Elsey et al. Mar 2005 B1
6873988 Herrmann et al. Mar 2005 B2
6880086 Kidder et al. Apr 2005 B2
6883068 Tsirigotis et al. Apr 2005 B2
6888927 Cruickshank et al. May 2005 B1
6894981 Coile et al. May 2005 B1
6899420 Silverbrook May 2005 B2
6901072 Wong May 2005 B1
6904449 Quinones Jun 2005 B1
6904519 Anand et al. Jun 2005 B2
6907546 Haswell et al. Jun 2005 B1
6910074 Amin et al. Jun 2005 B1
6912522 Edgar Jun 2005 B2
6918113 Patel et al. Jul 2005 B2
6918665 Silverbrook Jul 2005 B2
6920502 Araujo et al. Jul 2005 B2
6925495 Hegde et al. Aug 2005 B2
6934288 Dempo Aug 2005 B2
6935736 Silverbrook Aug 2005 B2
6944279 Elsey et al. Sep 2005 B2
6954736 Menninger et al. Oct 2005 B2
6954877 Earl et al. Oct 2005 B2
6957186 Guheen et al. Oct 2005 B1
6957249 Salo et al. Oct 2005 B2
6959320 Shah et al. Oct 2005 B2
6974928 Bloom Dec 2005 B2
6976090 Ben-Shaul et al. Dec 2005 B2
6980962 Arganbright et al. Dec 2005 B1
6990070 Aweya et al. Jan 2006 B1
6990480 Burt Jan 2006 B1
6993016 Liva et al. Jan 2006 B1
6996628 Keane et al. Feb 2006 B2
6996631 Aiken et al. Feb 2006 B1
7000012 Moore et al. Feb 2006 B2
7007092 Peiffer Feb 2006 B2
7007103 Pinkerton et al. Feb 2006 B2
7010300 Jones et al. Mar 2006 B1
7013290 Ananian Mar 2006 B2
7016055 Dodge et al. Mar 2006 B2
7020719 Grove et al. Mar 2006 B1
7023979 Wu et al. Apr 2006 B1
7024452 O'Connell et al. Apr 2006 B1
7026954 Slemmer et al. Apr 2006 B2
7027055 Anderson et al. Apr 2006 B2
7028083 Levine et al. Apr 2006 B2
7028333 Tuomenoksa et al. Apr 2006 B2
7028334 Tuomenoksa Apr 2006 B2
7031314 Craig et al. Apr 2006 B2
7034691 Rapaport et al. Apr 2006 B1
7035427 Rhoads Apr 2006 B2
7036142 Zhang et al. Apr 2006 B1
7039606 Hoffman et al. May 2006 B2
7039708 Knobl et al. May 2006 B1
7042879 Eschbach et al. May 2006 B2
7043529 Simonoff May 2006 B1
7047279 Beams et al. May 2006 B1
7047424 Bendinelli et al. May 2006 B2
7051161 Dixit et al. May 2006 B2
7051316 Charisius et al. May 2006 B2
7054837 Hoffman et al. May 2006 B2
7054944 Tang et al. May 2006 B2
7055028 Peiffer et al. May 2006 B2
7055169 Delpuch et al. May 2006 B2
7055947 Silverbrook Jun 2006 B2
7057759 Lapstun et al. Jun 2006 B2
7069234 Cornelius et al. Jun 2006 B1
7070110 Lapstun et al. Jul 2006 B2
7072665 Blumberg et al. Jul 2006 B1
7072843 Menninger et al. Jul 2006 B2
7072965 Ryuutou et al. Jul 2006 B2
7080041 Nagel Jul 2006 B2
7082409 Cherry Jul 2006 B1
7085683 Anderson et al. Aug 2006 B2
7085834 Delany et al. Aug 2006 B2
7085854 Keane et al. Aug 2006 B2
7086728 Silverbrook Aug 2006 B2
7092370 Jiang et al. Aug 2006 B2
7096009 Mousseau et al. Aug 2006 B2
7096418 Singhal et al. Aug 2006 B1
7096495 Warrier et al. Aug 2006 B1
7097106 Silverbrook et al. Aug 2006 B2
7100195 Underwood Aug 2006 B1
7103068 Gardner et al. Sep 2006 B1
7107285 von Kaenel et al. Sep 2006 B2
7113779 Fujisaki Sep 2006 B1
7113962 Kee et al. Sep 2006 B1
7114180 DeCaprio Sep 2006 B1
7117504 Smith et al. Oct 2006 B2
7117530 Lin Oct 2006 B1
7120596 Hoffman et al. Oct 2006 B2
7120666 McCanne et al. Oct 2006 B2
7120690 Krishnan et al. Oct 2006 B1
7120852 Terry et al. Oct 2006 B2
7124101 Mikurak Oct 2006 B1
7124442 Nash-Putnam Oct 2006 B2
7126955 Nabhan et al. Oct 2006 B2
7128265 Silverbrook et al. Oct 2006 B2
7128270 Silverbrook et al. Oct 2006 B2
7130792 Tokieda et al. Oct 2006 B2
7130807 Mikurak Oct 2006 B1
7131596 Lapstun et al. Nov 2006 B2
7134131 Hendricks et al. Nov 2006 B1
7135991 Slemmer et al. Nov 2006 B2
7136645 Hanson et al. Nov 2006 B2
7137566 Silverbrook et al. Nov 2006 B2
7139999 Bowman-Amuah Nov 2006 B2
7140044 Redlich et al. Nov 2006 B2
7143153 Black et al. Nov 2006 B1
7144095 Silverbrook Dec 2006 B2
7146384 Sawafta Dec 2006 B2
7146644 Redlich et al. Dec 2006 B2
7149698 Guheen et al. Dec 2006 B2
7149897 Chincheck et al. Dec 2006 B2
7149959 Jones et al. Dec 2006 B1
7150398 Silverbrook et al. Dec 2006 B2
7152047 Nagel Dec 2006 B1
7152092 Beams et al. Dec 2006 B2
7152207 Underwood et al. Dec 2006 B1
7155518 Forslow Dec 2006 B2
7155676 Land et al. Dec 2006 B2
7159014 Kausik et al. Jan 2007 B2
7159777 Silverbrook et al. Jan 2007 B2
7162540 Jasen et al. Jan 2007 B2
7164680 Loguinov Jan 2007 B2
7165041 Guheen et al. Jan 2007 B1
7167844 Leong et al. Jan 2007 B1
7171379 Menninger et al. Jan 2007 B2
7175089 Silverbrook et al. Feb 2007 B2
7178106 Lamkin et al. Feb 2007 B2
7178719 Silverbrook et al. Feb 2007 B2
7181766 Bendinelli et al. Feb 2007 B2
7184445 Gupta et al. Feb 2007 B2
7187470 Lapstun et al. Mar 2007 B2
7188273 Allen et al. Mar 2007 B2
7188769 Silverbrook et al. Mar 2007 B2
7191252 Redlich et al. Mar 2007 B2
7194761 Champagne Mar 2007 B1
7197374 Silverbrook et al. Mar 2007 B2
7197502 Feinsmith Mar 2007 B2
7197570 Eylon et al. Mar 2007 B2
7197751 Fedotov et al. Mar 2007 B2
7199885 Dodge Apr 2007 B2
7200804 Khavari et al. Apr 2007 B1
7206811 Skurikhin et al. Apr 2007 B2
7207483 Silverbrook et al. Apr 2007 B2
7207485 Silverbrook et al. Apr 2007 B2
7212296 Dodge et al. May 2007 B2
7216173 Clayton et al. May 2007 B2
7219127 Huck et al. May 2007 B2
7221660 Simonson et al. May 2007 B1
7222152 Thompson et al. May 2007 B1
7222305 Teplov et al. May 2007 B2
7225040 Eller et al. May 2007 B2
7225244 Reynolds et al. May 2007 B2
7225382 Ramech et al. May 2007 B2
7228459 Jiang et al. Jun 2007 B2
7246233 Brabson et al. Jul 2007 B2
7249351 Songer et al. Jul 2007 B1
7260617 Bazinet et al. Aug 2007 B2
7260840 Swander et al. Aug 2007 B2
7263071 Yim Aug 2007 B2
7272853 Goodman et al. Sep 2007 B2
7284044 Teraoaka et al. Oct 2007 B2
7318100 Demmer et al. Jan 2008 B2
7321906 Green Jan 2008 B2
7328267 Bashyam et al. Feb 2008 B1
7340772 Panasyuk et al. Mar 2008 B2
7353533 Wright et al. Apr 2008 B2
7363347 Thomas Apr 2008 B2
7363378 Holmes et al. Apr 2008 B2
7370353 Yang May 2008 B2
7386631 Sibal et al. Jun 2008 B1
7389462 Wang et al. Jun 2008 B1
7392348 Dumont Jun 2008 B2
7398320 Minakuchi et al. Jul 2008 B1
7398552 Pardee et al. Jul 2008 B2
7401116 Chalfin et al. Jul 2008 B1
7404003 Noble Jul 2008 B1
7406533 Li et al. Jul 2008 B2
7409708 Goodman et al. Aug 2008 B2
7415723 Pandya Aug 2008 B2
7420992 Fang et al. Sep 2008 B1
7433314 Sharma et al. Oct 2008 B2
7441119 Brabson et al. Oct 2008 B2
7458095 Forsberg Nov 2008 B2
7464264 Goodman et al. Dec 2008 B2
7480312 Ossman Jan 2009 B2
7496659 Coverdill et al. Feb 2009 B1
7505455 Goodwin et al. Mar 2009 B1
7512702 Srivastava et al. Mar 2009 B1
7522581 Acharya et al. Apr 2009 B2
7533409 Keane et al. May 2009 B2
7536715 Markham May 2009 B2
7577144 Chang et al. Aug 2009 B2
7606902 Rao et al. Oct 2009 B2
7609721 Rao et al. Oct 2009 B2
7644188 Vlodavsky et al. Jan 2010 B2
7673048 O'Toole et al. Mar 2010 B1
7757074 Sundarrajan et al. Jul 2010 B2
7775074 Tobias et al. Aug 2010 B1
7882247 Sturniolo et al. Feb 2011 B2
7921184 Sundarrajan et al. Apr 2011 B2
7945678 Skene May 2011 B1
7979528 Eisenberg et al. Jul 2011 B2
7979694 Touitou et al. Jul 2011 B2
8005981 Tuck et al. Aug 2011 B2
8205000 Chang et al. Jun 2012 B2
8223647 Kakani et al. Jul 2012 B2
8261057 Sundarrajan et al. Sep 2012 B2
8351327 Binns Jan 2013 B1
20010030970 Wiryaman et al. Oct 2001 A1
20010037387 Gilde et al. Nov 2001 A1
20010043600 Chatterjee et al. Nov 2001 A1
20010046212 Nakajima Nov 2001 A1
20010047406 Araujo et al. Nov 2001 A1
20010049717 Freeman et al. Dec 2001 A1
20010049741 Skene et al. Dec 2001 A1
20010052016 Skene et al. Dec 2001 A1
20010052065 Alexander, III et al. Dec 2001 A1
20020004834 Guenther et al. Jan 2002 A1
20020004902 Toh et al. Jan 2002 A1
20020007402 Huston et al. Jan 2002 A1
20020007404 Vange et al. Jan 2002 A1
20020016827 McCabe et al. Feb 2002 A1
20020023210 Tuomenoksa et al. Feb 2002 A1
20020026478 Rodgers et al. Feb 2002 A1
20020026531 Keane et al. Feb 2002 A1
20020029285 Collins Mar 2002 A1
20020032725 Araujo et al. Mar 2002 A1
20020032798 Xu Mar 2002 A1
20020035451 Rothermel Mar 2002 A1
20020035596 Yang et al. Mar 2002 A1
20020037079 Duval Mar 2002 A1
20020038339 Xu Mar 2002 A1
20020042875 Shukla Apr 2002 A1
20020049608 Hartsell et al. Apr 2002 A1
20020049841 Johnson et al. Apr 2002 A1
20020055966 Border et al. May 2002 A1
20020057717 Mallory May 2002 A1
20020059274 Hartsell et al. May 2002 A1
20020059429 Carpenter et al. May 2002 A1
20020059435 Border et al. May 2002 A1
20020065864 Hartsell et al. May 2002 A1
20020065919 Taylor et al. May 2002 A1
20020069278 Forslow Jun 2002 A1
20020071438 Singh Jun 2002 A1
20020073061 Collins Jun 2002 A1
20020073298 Geiger et al. Jun 2002 A1
20020081971 Travostino Jun 2002 A1
20020083175 Afek et al. Jun 2002 A1
20020083183 Pujare et al. Jun 2002 A1
20020087729 Edgar Jul 2002 A1
20020091788 Chlan et al. Jul 2002 A1
20020095400 Johnson et al. Jul 2002 A1
20020095498 Chanda et al. Jul 2002 A1
20020098840 Hanson et al. Jul 2002 A1
20020099940 Wang Jul 2002 A1
20020101848 Lee et al. Aug 2002 A1
20020101860 Thornton et al. Aug 2002 A1
20020105972 Richter et al. Aug 2002 A1
20020107903 Richter et al. Aug 2002 A1
20020107962 Richter et al. Aug 2002 A1
20020107971 Bailey et al. Aug 2002 A1
20020107989 Johnson et al. Aug 2002 A1
20020107990 Johnson et al. Aug 2002 A1
20020108059 Canion et al. Aug 2002 A1
20020116452 Johnson et al. Aug 2002 A1
20020116582 Copeland et al. Aug 2002 A1
20020120710 Chintalapati et al. Aug 2002 A1
20020120743 Shabtay et al. Aug 2002 A1
20020126675 Yoshimura et al. Sep 2002 A1
20020133593 Johnson et al. Sep 2002 A1
20020138511 Psounis et al. Sep 2002 A1
20020138572 Delany et al. Sep 2002 A1
20020138618 Szabo Sep 2002 A1
20020138640 Raz et al. Sep 2002 A1
20020147795 Cantwell Oct 2002 A1
20020147822 Susai et al. Oct 2002 A1
20020150064 Lucidarme Oct 2002 A1
20020152373 Sun et al. Oct 2002 A1
20020152423 McCabe Oct 2002 A1
20020156927 Boucher et al. Oct 2002 A1
20020161908 Benitez et al. Oct 2002 A1
20020163882 Bornstein et al. Nov 2002 A1
20020174227 Hartsell et al. Nov 2002 A1
20020174255 Hayter et al. Nov 2002 A1
20020176532 McClelland et al. Nov 2002 A1
20020184224 Haff et al. Dec 2002 A1
20020194382 Kausik et al. Dec 2002 A1
20020196279 Bloomfield et al. Dec 2002 A1
20020199007 Clayton et al. Dec 2002 A1
20030009538 Shah et al. Jan 2003 A1
20030014623 Freed et al. Jan 2003 A1
20030014624 Maturana et al. Jan 2003 A1
20030014625 Freed et al. Jan 2003 A1
20030014628 Freed et al. Jan 2003 A1
20030014650 Freed et al. Jan 2003 A1
20030018891 Hall et al. Jan 2003 A1
20030023767 Brabson et al. Jan 2003 A1
20030026241 Ono et al. Feb 2003 A1
20030033520 Peiffer et al. Feb 2003 A1
20030037108 Peiffer et al. Feb 2003 A1
20030039354 Kimble et al. Feb 2003 A1
20030046586 Bheemarasetti et al. Mar 2003 A1
20030046587 Bheemarasetti et al. Mar 2003 A1
20030051100 Patel Mar 2003 A1
20030051102 Jacobs et al. Mar 2003 A1
20030055883 Wiles, Jr. Mar 2003 A1
20030055962 Freund et al. Mar 2003 A1
20030061505 Sperry et al. Mar 2003 A1
20030065763 Swildens et al. Apr 2003 A1
20030067874 See et al. Apr 2003 A1
20030069941 Peiffer Apr 2003 A1
20030079031 Nagano Apr 2003 A1
20030084162 Johnson et al. May 2003 A1
20030084165 Kjellberg et al. May 2003 A1
20030088698 Singh et al. May 2003 A1
20030088788 Yang May 2003 A1
20030093488 Yoshida et al. May 2003 A1
20030093566 Jardin May 2003 A1
20030105604 Ash et al. Jun 2003 A1
20030105977 Brabson et al. Jun 2003 A1
20030110296 Kirsch et al. Jun 2003 A1
20030110379 Ylonen et al. Jun 2003 A1
20030112809 Bharali et al. Jun 2003 A1
20030112823 Collins et al. Jun 2003 A1
20030115421 McHenry et al. Jun 2003 A1
20030123394 Neale et al. Jul 2003 A1
20030123481 Neale et al. Jul 2003 A1
20030131079 Neale et al. Jul 2003 A1
20030131263 Keane et al. Jul 2003 A1
20030140121 Adams Jul 2003 A1
20030149899 Boden et al. Aug 2003 A1
20030152028 Raisanen et al. Aug 2003 A1
20030154110 Walter et al. Aug 2003 A1
20030154239 Davis et al. Aug 2003 A1
20030165138 Swonk et al. Sep 2003 A1
20030167403 McCurley et al. Sep 2003 A1
20030174718 Sampath et al. Sep 2003 A1
20030177389 Albert et al. Sep 2003 A1
20030177395 Pardee et al. Sep 2003 A1
20030182423 Shafir et al. Sep 2003 A1
20030182431 Sturniolo et al. Sep 2003 A1
20030182437 Kobayashi et al. Sep 2003 A1
20030188001 Eisenberg et al. Oct 2003 A1
20030191799 Araujo et al. Oct 2003 A1
20030191812 Agarwalla et al. Oct 2003 A1
20030198189 Roberts et al. Oct 2003 A1
20030200295 Roberts et al. Oct 2003 A1
20030200353 Dogra et al. Oct 2003 A1
20030208602 Bhalla et al. Nov 2003 A1
20030212760 Chen et al. Nov 2003 A1
20030212776 Roberts et al. Nov 2003 A1
20030217105 Zircher et al. Nov 2003 A1
20030217126 Polcha et al. Nov 2003 A1
20030217149 Crichton et al. Nov 2003 A1
20030223361 Hussain et al. Dec 2003 A1
20030223418 Desai et al. Dec 2003 A1
20030223433 Lee et al. Dec 2003 A1
20030226018 Tardo et al. Dec 2003 A1
20030229718 Tock et al. Dec 2003 A1
20030229761 Basu et al. Dec 2003 A1
20030231659 DiMambro et al. Dec 2003 A1
20030233423 Dilley et al. Dec 2003 A1
20030236831 Ortiz et al. Dec 2003 A1
20030236837 Johnson et al. Dec 2003 A1
20030236861 Johnson et al. Dec 2003 A1
20030236919 Johnson et al. Dec 2003 A1
20040003101 Roth et al. Jan 2004 A1
20040003137 Callender et al. Jan 2004 A1
20040006708 Mukherjee et al. Jan 2004 A1
20040008693 Grove et al. Jan 2004 A1
20040010601 Afergan et al. Jan 2004 A1
20040010604 Tanaka et al. Jan 2004 A1
20040010621 Afergan et al. Jan 2004 A1
20040031058 Reisman Feb 2004 A1
20040034744 Karlsson et al. Feb 2004 A1
20040034794 Mayer et al. Feb 2004 A1
20040039827 Thomas et al. Feb 2004 A1
20040042487 Ossman Mar 2004 A1
20040044521 Chen et al. Mar 2004 A1
20040044731 Chen et al. Mar 2004 A1
20040049515 Haff et al. Mar 2004 A1
20040073716 Boom et al. Apr 2004 A1
20040078621 Talaugon et al. Apr 2004 A1
20040078772 Balay et al. Apr 2004 A1
20040087304 Buddhikot et al. May 2004 A1
20040098486 Gu et al. May 2004 A1
20040100976 Chang et al. May 2004 A1
20040103150 Ogdon et al. May 2004 A1
20040103225 McAlpine et al. May 2004 A1
20040103438 Yan et al. May 2004 A1
20040107360 Herrmann et al. Jun 2004 A1
20040111519 Fu et al. Jun 2004 A1
20040128252 Shirai et al. Jul 2004 A1
20040128346 Melamed et al. Jul 2004 A1
20040141525 Bhushan et al. Jul 2004 A1
20040143655 Narad et al. Jul 2004 A1
20040143734 Buer et al. Jul 2004 A1
20040146053 Nabhan et al. Jul 2004 A1
20040158705 Burdett et al. Aug 2004 A1
20040162876 Kohavi Aug 2004 A1
20040177158 Bauch et al. Sep 2004 A1
20040177359 Bauch et al. Sep 2004 A1
20040190092 Silverbrook et al. Sep 2004 A1
20040202171 Hama Oct 2004 A1
20040203296 Moreton et al. Oct 2004 A1
20040205439 Carmeli et al. Oct 2004 A1
20040210320 Pandya Oct 2004 A1
20040213248 Okuda et al. Oct 2004 A1
20040215665 Edgar et al. Oct 2004 A1
20040215746 McCanne et al. Oct 2004 A1
20040221031 Desai Nov 2004 A1
20040225898 Frost et al. Nov 2004 A1
20040225911 Smith Nov 2004 A1
20040230840 Radatti Nov 2004 A1
20040246971 Banerjee et al. Dec 2004 A1
20040249975 Tuck et al. Dec 2004 A1
20040258003 Kokot et al. Dec 2004 A1
20040260922 Goodman et al. Dec 2004 A1
20040268175 Koch et al. Dec 2004 A1
20040268357 Joy et al. Dec 2004 A1
20040268358 Darling et al. Dec 2004 A1
20050005014 Holmes et al. Jan 2005 A1
20050005107 Touboul Jan 2005 A1
20050013280 Buddhikot et al. Jan 2005 A1
20050021762 Gbadegesin Jan 2005 A1
20050022011 Swander et al. Jan 2005 A1
20050022012 Bluestone et al. Jan 2005 A1
20050022031 Goodman et al. Jan 2005 A1
20050025150 Helmy et al. Feb 2005 A1
20050027788 Koopmans et al. Feb 2005 A1
20050031058 Soong et al. Feb 2005 A1
20050033858 Swildens et al. Feb 2005 A1
20050033926 Dumont Feb 2005 A1
20050044350 White et al. Feb 2005 A1
20050050317 Kramer et al. Mar 2005 A1
20050055690 Cornillon et al. Mar 2005 A1
20050058112 Lahey et al. Mar 2005 A1
20050080850 Salesky et al. Apr 2005 A1
20050080907 Panasyuk et al. Apr 2005 A1
20050086386 Shen et al. Apr 2005 A1
20050102529 Buddhikot et al. May 2005 A1
20050108412 Sjollema et al. May 2005 A1
20050108517 Dillon et al. May 2005 A1
20050122980 Anand et al. Jun 2005 A1
20050125663 Funk Jun 2005 A1
20050132030 Hopen et al. Jun 2005 A1
20050132060 Mo et al. Jun 2005 A1
20050135250 Singh et al. Jun 2005 A1
20050144186 Hesselink et al. Jun 2005 A1
20050144481 Hopen et al. Jun 2005 A1
20050147126 Qiu et al. Jul 2005 A1
20050149481 Hesselink et al. Jul 2005 A1
20050149726 Joshi et al. Jul 2005 A1
20050165928 Shu et al. Jul 2005 A1
20050177866 Kirsch Aug 2005 A1
20050185647 Rao et al. Aug 2005 A1
20050193075 Haff et al. Sep 2005 A1
20050195780 Haverinen et al. Sep 2005 A1
20050198380 Panasyuk et al. Sep 2005 A1
20050198532 Comlekoglu et al. Sep 2005 A1
20050210150 Bahl Sep 2005 A1
20050223412 Nadalin et al. Oct 2005 A1
20050232161 Maufer et al. Oct 2005 A1
20050232298 Beverly et al. Oct 2005 A1
20050246346 Gerdes et al. Nov 2005 A1
20050246445 Panasyuk et al. Nov 2005 A1
20050256923 Adachi Nov 2005 A1
20050259634 Ross Nov 2005 A1
20050262357 Araujo et al. Nov 2005 A1
20050265315 Edgar Dec 2005 A1
20050265353 Sengupta et al. Dec 2005 A1
20050267974 Panasyuk et al. Dec 2005 A1
20050286466 Tagg et al. Dec 2005 A1
20060005240 Sundarrajan et al. Jan 2006 A1
20060015570 Khemani et al. Jan 2006 A1
20060018332 Kakani et al. Jan 2006 A1
20060031547 Tsui et al. Feb 2006 A1
20060034283 Ko et al. Feb 2006 A1
20060041635 Alexander et al. Feb 2006 A1
20060053164 Ewing et al. Mar 2006 A1
20060053253 Roth et al. Mar 2006 A1
20060059370 Asnis et al. Mar 2006 A1
20060064500 Roth et al. Mar 2006 A1
20060069719 McCanne et al. Mar 2006 A1
20060080441 Chen et al. Apr 2006 A1
20060089996 Peiffer Apr 2006 A1
20060095758 Kim et al. May 2006 A1
20060112185 van Bemmel May 2006 A1
20060123477 Raghavan et al. Jun 2006 A1
20060129689 Ho et al. Jun 2006 A1
20060133405 Fee Jun 2006 A1
20060142878 Banik et al. Jun 2006 A1
20060167975 Chan et al. Jul 2006 A1
20060185010 Sultan Aug 2006 A1
20060195840 Sundarrajan et al. Aug 2006 A1
20060225130 Chen et al. Oct 2006 A1
20060245414 Susai et al. Nov 2006 A1
20060271652 Stavrakos et al. Nov 2006 A1
20060274828 Siemens et al. Dec 2006 A1
20070008883 Kobayashi Jan 2007 A1
20070061067 Zeinstra et al. Mar 2007 A1
20070067366 Landis Mar 2007 A1
20070110046 Farrell et al. May 2007 A1
20070130324 Wang Jun 2007 A1
20070130334 Carley Jun 2007 A1
20070156852 Sundarrajan et al. Jul 2007 A1
20070156876 Sundarrajan et al. Jul 2007 A1
20070169179 Narad Jul 2007 A1
20070179955 Croft et al. Aug 2007 A1
20070192827 Maxted et al. Aug 2007 A1
20070206621 Plamondon et al. Sep 2007 A1
20070233910 Paley et al. Oct 2007 A1
20070245409 Harris et al. Oct 2007 A1
20080034416 Kumar et al. Feb 2008 A1
20080046616 Verzunov et al. Feb 2008 A1
20080049786 Ram et al. Feb 2008 A1
20080071915 Gbadegesin Mar 2008 A1
20080225720 Khemani et al. Sep 2008 A1
20080225748 Khemani et al. Sep 2008 A1
20080225753 Khemani et al. Sep 2008 A1
20080229381 Sikka et al. Sep 2008 A1
20080320151 McCanne et al. Dec 2008 A1
20090133015 Nagashima May 2009 A1
20100241846 Sundarrajan et al. Sep 2010 A1
Foreign Referenced Citations (39)
Number Date Country
1394042 Jan 2003 CN
1410905 Apr 2003 CN
0 436 365 Jul 1991 EP
0618708 Oct 1994 EP
0 838 930 Apr 1998 EP
1 045 553 Oct 2000 EP
1 134 942 Sep 2001 EP
1 427 133 Jun 2004 EP
1 432 209 Jun 2004 EP
07-302237 Nov 1995 JP
08-006910 Jan 1996 JP
10-178450 Jun 1998 JP
2002-532784 Feb 2002 JP
2002-084289 Mar 2002 JP
2004078783 Mar 2004 JP
WO-9904343 Jan 1999 WO
WO-0034885 Jun 2000 WO
WO-0147185 Jun 2001 WO
WO-0213037 Feb 2002 WO
WO-0223362 Mar 2002 WO
WO-0239301 May 2002 WO
WO-02069604 Sep 2002 WO
WO-03019876 Mar 2003 WO
WO-03026244 Mar 2003 WO
WO-03048936 Jun 2003 WO
WO-03083692 Oct 2003 WO
WO-2004088933 Oct 2004 WO
WO-2004114529 Dec 2004 WO
WO-2005013534 Feb 2005 WO
WO-2005024567 Mar 2005 WO
WO-2005048106 May 2005 WO
WO-2005088476 Sep 2005 WO
WO-2006005078 Jan 2006 WO
WO-2006012612 Feb 2006 WO
WO-2006020823 Feb 2006 WO
WO-2006074072 Jul 2006 WO
WO-2006075219 Jul 2006 WO
WO-2008112691 Sep 2008 WO
WO-2008112698 Sep 2008 WO
Non-Patent Literature Citations (84)
Entry
Krishnamurthy B.C. et al., “Key differences between HTTP/1.0 and HTTP/1.1,” Computer Networks, Elsevier Science Publishers B.V., vol. 31, No. 11-16, pp. 1737-1751, May 1999.
International Search Report for PCT/US2005/047435. mailing date Sep. 8, 2006. 5 pages.
International Search Report, PCT/US05/23914, mailed on May 7, 2007. 3 pages.
Non Final Office Action dated Nov. 21, 2007 for U.S. Appl. No. 11/323,196. 14 pages.
Non Final Office Action dated Jan. 8, 2008 for U.S. Appl. No. 11/324,146. 12 pages.
Non Final Office Action, U.S. Appl. No. 11/323,196. dated May 14, 2008.
Office Action from Sep. 25, 2008 for U.S. Appl. No. 11/323,196, 19 pages.
Office Action from Oct. 3, 2008 for U.S. Appl. No. 11/324,131, pp. 22.
Office Action from Nov. 12, 2008 for U.S. Appl. No. 11/324,146, pp. 17.
Office Action from Dec. 30, 2008 for U.S. Appl. No. 11/323,353, pp. 22.
Office Action from Feb. 13, 2009 for U.S. Appl. No. 11/323,196, pp. 13.
Office Action from Feb. 27, 2009 for U.S. Appl. No. 11/324,141, pp. 10.
Doug Allen, Layer-7 Load Balancers Pack New Punch, CMP Media LLC Network Magazine, Sep. 2003.
Office Action U.S. Appl. No. 11/324,146, filed Nov. 13, 2009.
Decision on Rejection issued Jun. 3, 2013 in Chinese Patent Application No. 201110075213.7.
European Examination Report ob 05768770.9 dated Jun. 13, 2012.
Examination Report on 3925/KOLNP/2006 dated Nov. 23, 2012.
Notice of Allowance on U.S. Appl. No. 11/324,131 dated Sep. 26, 2012.
Notice of Allowance on U.S. Appl. No. 11/324,146 dated Jun. 26, 2012.
Notice of Allowance on U.S. Appl. No. 11/324,146 dated Jul. 6, 2012.
Office Action for U.S. Appl. No. 11/324,141 dated Apr. 26, 2011.
Office Action for U.S. Appl. No. 11/324,141 dated Aug. 21, 2009.
Office Action on U.S. Appl. No. 13/031,919 dated Jun. 6, 2012.
Office Action on U.S. Appl. No. 13/031,919 dated Sep. 26, 2012.
Reynolds et al., “Aloe vera leaf gel: a review update”, Journal of Ethnopharmacology, vol. 68, pp. 3-37, 1999.
Second Chinese Office Action on 201110075213.7 dated Nov. 1, 2012.
Third Office Action issued in Chinese Application No. 200580028489.2 on Feb. 1, 2013.
US Notice of Allowance on 099011-0158 DTD Mar. 29, 2013.
US Office Action on 099011-2006579-0457-0532-US DTD Jul. 28, 2010.
US Office Action on 099011-3180 DTD May 17, 2013.
Written Opinion for PCT/US2005/047435 dated Aug. 8, 2006.
US69/34288-B2, Mar. 13, 2003, Jacobs et al.
U.S. Appl. No. 11/323,196 US Office Action Aug. 3, 2009.
U.S. Appl. No. 11/324,141 US Office Action Mar. 3, 2010.
Biagioni et al., Signature for a network Protocol Stack: A System Application of Standard ML, Jul. 1994.
Chinese Office Action on 200580028489.2 dated Mar. 5, 2012.
Chinese Office Action on 201110075213.7 dated Jan. 30, 2012.
Chinese Office Action on 201110076025.6 dated Oct. 26, 2011.
Dictionary service [online], [retrieved on Jul. 23, 2009]. Retrieved from the Internet &It;URL: http://dictionary.reference.com/browse/discard&gt.
Doug Allen, Layer-7 Load Balancers Pack New Punch, CMP Media LLC Network Magazine, Sep. 2003 [cited by examiner].
E.C. Hu et al.“Adaptive Fast Path Architecture”, IBM J. Res &Dev, vol. 45 Mar. 1, 2001, pp. 191-206, XP002663647, Retrieved from the Internet: URL:http://gec.di.uminho.pt/discip/minf/ac0203/ICCA03/IBM WebCachArch.pdf.
Chinese Office Action on 200580028489.2 dated Nov. 3, 2011.
European Examination Report on 05855924.6 dated Nov. 24, 2011.
European Supplementary Search Report on 05768770.9 dated Nov. 29, 2011.
Hasegawa et al., “A Mechanism for TCP Performance Enhancement over Assymetrical Environment,” Proceedings 8th IEEE ISCC'03, 1530-1346/03, 2003.
Ipswitch Inc., WSFTP User's Guide, Chapter 8: Security, Jun. 25, 2003, 8 pages.
J. C. Mogul, “The Case for Persistent-Connection HTTP,” 8282 Computer Communication Review 25, Oct. 1995.
Japanese Office Action on 2007-552147 dated Apr. 16, 2012.
Japanese Office Action on 2007-552147 dated Jun. 30, 2011.
Kong, et al; “Pseudo-serving: a user-responsible paradigm for internet access”; Apr. 7-11, 1997; SO Computer Networks and ISDN Systems; vol. 29; pp. 1053-1064.
Lazenka, M. “The Evolution of Software Distribution: Where Did We Come From and Where Are We Heading?”, Nov. 6-9, 2005, ACM, p. 179-183.
Noriaki Kobayashi, “Internet VPN”, N+1 Network Guide, vol. 3, No. 9, Oct. 1, 2003, pp. 44-49. (9 pages) (English Abstract).
Notice of Allowance for U.S. Appl. No. 11/323,353 dated Nov. 15, 2010.
Notice of Allowance on U.S. Appl. No. 11/324,141 dated Apr. 25, 2012.
Notice of Allowance on U.S. Appl. No. 12/020,649 dated Oct. 4, 2011.
Notice of Allowance on U.S. Appl. No. 12/020,963 dated Apr. 3, 2012.
Office Action for U.S. Appl. No. 11/324,131 dated Oct. 3, 2008.
Office Action for U.S. Appl. No. 11/324,131 dated Oct. 6, 2009.
Office Action for U.S. Appl. No. 11/324,131 dated Mar. 7, 2011.
Office Action for U.S. Appl. No. 11/324,131 dated Apr. 2, 2009.
Office Action for U.S. Appl. No. 11/324,131 dated May 27, 2010.
Office Action for U.S. Appl. No. 11/323,353 dated Jun. 3, 2009.
Office Action on U.S. Appl. No. 11/323,196 dated Nov. 19, 2009.
Office Action on U.S. Appl. No. 11/324,131 dated Jan. 4, 2012.
Office Action on U.S. Appl. No. 11/324,131 dated Aug. 11, 2011.
Office Action on U.S. Appl. No. 11/324,141 dated Nov. 10, 2010.
Office Action on U.S. Appl. No. 11/324,141 dated Dec. 15, 2011.
Office Action on U.S. Appl. No. 11/324,146 dated Jan. 24, 2012.
Office Action on U.S. Appl. No. 11/324,146 dated Feb. 15, 2011.
Office Action on U.S. Appl. No. 11/324,146 dated Jun. 22, 2011.
Office Action on U.S. Appl. No. 12/020,649 dated Mar. 28, 2011.
Office Action on U.S. Appl. No. 12/020,963 dated Jun. 10, 2011.
Office Action on U.S. Appl. No. 11/323,353 mailed on Dec. 10, 2009.
Office Action U.S. Appl. No. 11/324,146 Nov. 13, 2009.
Official Notification from Israeli Patent Office, for Appl No. 184750 mailed Apr. 7, 2011.
Pedro Hernandez, “Tacit Adds to Softricity's Spark”, Nov. 18, 2005. Available online at: www.enterpriseitplant.com/networking/news/article.php/3565401.
R. Briscoe, “The Implications of Pervasive Computing on Network Design,” BT Technology Journal, pp. 170-190, Jul. 2004.
Ranzau et al., “Softricity/Tacit, An Innovative Approach to Virtual Computing,” http://www.daboc.com/downloadnow.aspx?file=211&is=617, Oct. 20, 2005.
Teruhara Serada, “SSL VPN”, N+1 Network LAN, vol. 21, No. 3, Mar. 1, 2003, pp. 25-40. (19 pages) (English Abstract).
U.S. Appl. No. 11/323,196 Notice of Allowance Apr. 19, 2010.
U.S. Appl. No. 11/323,353 US Office Action Jun. 25, 2010.
U.S. Appl. No. 11/324,146 US Office Action Mar. 25, 2009.
Wang et al., Shield: Vulnerability-Driven Network Filters for Preventing Known Vulnerability Exploits, Feb. 2004, Microsoft Research, MSR-TR-2003-81.
Yoshikazu Kobayashi, VPN service introduction, Computer & Network LAN, vol. 21, No. 3, Mar. 1, 2003, pp. 25-40. (19 pages) (English Abstract).
Related Publications (1)
Number Date Country
20060015570 A1 Jan 2006 US
Provisional Applications (1)
Number Date Country
60583784 Jun 2004 US