Forward packet recovery with constrained network overhead

Information

  • Patent Grant
  • 10848268
  • Patent Number
    10,848,268
  • Date Filed
    Friday, April 26, 2019
    5 years ago
  • Date Issued
    Tuesday, November 24, 2020
    3 years ago
Abstract
Disclosed herein are systems and methods for forward packet recovery in a communication network with constrained network bandwidth overhead. In exemplary embodiments, a target byte protection ratio is determined. Error correcting frames are dynamically generated by a first processor such that error correcting information can be generated to approximate the target byte protection ratio. The data packets and error correcting information are then transmitted across one or more communication networks to a second processor. The second processor can use the error correcting information to regenerate or replace data packets missing or corrupted in transmission across one or more communication networks.
Description
TECHNICAL FIELD

This disclosure relates generally to network communications and more specifically to forward packet recovery.


BACKGROUND

The approaches described in this section could be pursued, but are not necessarily approaches that have previously been conceived or pursued. Therefore, unless otherwise indicated, it should not be assumed that any of the approaches described in this section qualify as prior art merely by virtue of their inclusion in this section.


Typically, data is sent between computing devices across a communications network in packets. The packets may be generated according to a variety of protocols such as Transmission Control Protocol (TCP), User Datagram Protocol (UDP), or the like. During transmission, packets may be lost, corrupted, or received out of order. In these instances, the computing device sending the packets may have to resend the packets or the receiving application may have methods for coping with missing or corrupted packets. Either way the impairments are undesirable.


Forward packet recovery methods provide for generating and transmitting additional parity packets across a communications network. The parity packet contains information that can be used to reconstruct or replace one or more corrupted or lost packets at the receiver side. Parity packets are traditionally sent in a ratio based on a number of data packets transmitted. For example, one parity packet may be sent per five data packets. However, since the number and amount of parity information is based on a number of data packets transmitted, the amount of network bandwidth needed for transmission of these packets is unpredictable, at least because data packets can be of varying sizes.


Thus, there is a need for a mechanism for transmitting error correcting information to repair or replace data based on a number of bytes transmitted, rather than based on a number of packets transmitted across one or more communication networks, to accommodate for constrained network resources.


SUMMARY

This summary is provided to introduce a selection of concepts in a simplified form that are further described in the Detailed Description below. This summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.


In various exemplary methods of the present disclosure, a first processor receives a data stream comprising a plurality of data packets for transmission across at least one communication network, selects a target byte protection ratio for the data transmission, the target byte protection ratio representing a target number of bytes of error correcting information per bytes of data in the data stream, dynamically generates one or more error correcting frames for the data stream, generates one or more error correction data packets for each error correcting frame of data in accordance with the target byte protection ratio, and transmits the plurality of data packets and the one or more error correction data packets across the at least one communication network to a second processor.


The error correcting frames can be dynamically generated by grouping data packets of the data stream in various configurations, fragmenting data packets, and/or segmenting data packets. By adjusting the error correcting frames in this way, the amount of error correcting information generated in accordance with the target byte protection ratio is reduced in comparison to a target packet protection ratio. Further, error correction information is generated in accordance with constrained network overhead resource requirements.


Other features, examples, and embodiments are described below.





BRIEF DESCRIPTION OF THE DRAWINGS

Embodiments are illustrated by way of example and not by limitation in the figures of the accompanying drawings, in which like references indicate similar elements.



FIG. 1 is a diagram of an exemplary environment in which various embodiments of the present disclosure may be practiced.



FIG. 2 is an exemplary schematic of a network appliance.



FIG. 3 depicts an exemplary error correcting frame of prior systems.



FIG. 4A depicts an exemplary embodiment of dynamic framing via grouping of data packets.



FIG. 4B depicts an exemplary flow sequence for a data protector to determine a number of error correction packets to generate for each frame.



FIG. 5A depicts an exemplary embodiment of fragmenting data packets in prior systems.



FIG. 5B depicts an exemplary embodiment of dynamic framing via fragmenting data packets.



FIG. 6 depicts an exemplary embodiment of dynamic framing via interleaving data packets.



FIG. 7 depicts an exemplary embodiment of dynamic framing via splitting data packets.



FIG. 8 depicts an exemplary embodiment of generating byte based error correction information for a plurality of data packets using a data structure.



FIG. 9A depicts another exemplary embodiment of generating byte based error correction information for a plurality of data packets using a data structure.



FIG. 9B depicts a further exemplary embodiment of generating byte based error correction information for a plurality of data packets using a data structure.





DETAILED DESCRIPTION

The following detailed description includes references to the accompanying drawings, which form a part of the detailed description. The drawings show illustrations, in accordance with exemplary embodiments. These exemplary embodiments, which are also referred to herein as “examples,” are described in enough detail to enable those skilled in the art to practice the present subject matter. The embodiments can be combined, other embodiments can be utilized, or structural, logical, and electrical changes can be made without departing from the scope of what is claimed. The following detailed description is therefore not to be taken in a limiting sense, and the scope is defined by the appended claims and their equivalents. In this document, the terms “a” and “an” are used, as is common in patent documents, to include one or more than one. In this document, the term “or” is used to refer to a nonexclusive “or,” such that “A or B” includes “A but not B,” “B but not A,” and “A and B,” unless otherwise indicated.


The embodiments disclosed herein may be implemented using a variety of technologies. For example, the methods described herein may be implemented in software executing on a computer system containing one or more computers, or in hardware utilizing either a combination of microprocessors or other specially designed application-specific integrated circuits (ASICs), programmable logic devices, or various combinations thereof. In particular, the methods described herein may be implemented by a series of computer-executable instructions residing on a storage medium, such as a disk drive, or computer-readable medium.


The embodiments described herein relate to forward packet recovery for data transmitted across one or more communication networks.


Exemplary systems and methods for forward packet recovery are provided. Forward packet recovery can be used to reconstruct missing data packets and corrupted data packets and order the received and reconstructed data packets prior to processing according to a protocol such as Transmission Control Protocol (TCP), User Datagram Protocol (UDP), or the like. In order to perform forward packet recovery, protection packets with error correcting information are sent with the data packets. Protection packets are also sometimes referred to herein as correction packets, error correction packets, error correction data packets or protection data packets. Typically, one protection packet can be used to reconstruct one missing or corrupted data packet. However, sending the protection packets in addition to the data packets uses more bandwidth in a communication network which may, in turn, slow communications between network appliances. However, not sending a sufficient number of protection packets may result in having to resend data packets. Missing data packets and out of order packets can negatively impact performance of TCP and other high level protocols.


Protection packets can be sent based on data packets in an error correcting frame of a stream of data. As used herein, a stream of data can be any sequence of data packets transmitted between two devices. A stream may be the same as a flow, as understood by persons of ordinary skill in the art, or may be different than a flow. A stream can be comprised of an aggregate of multiple flows or sessions. The stream may contain packets using TCP, UDP, and/or other protocols. In some implementations separate streams may be created for different quality of service classes or types of traffic.



FIG. 1 is a diagram of an environment 100 in which various embodiments of the present disclosure may be practiced. The environment 100 comprises a network appliance A 102 and a network appliance B 104 communicatively coupled via one or more communication network(s) 106. While two network appliances are depicted in FIG. 1 for exemplary purposes, there can be any number of network appliances communicatively coupled to one or more other network appliances.


In a wide area network, there can be multiple network appliances deployed in one or more geographic locations. Each network appliance comprises hardware and/or software elements configured to receive data and optionally perform any type of processing, including but not limited to, WAN optimization techniques to the data, before transmitting to another appliance or an endpoint device. In various embodiments, a network appliance can be configured as an additional router or gateway. If a network appliance has multiple interfaces, it can be transparent on some interfaces, act like a router on some interfaces, or be a bridge on others. Alternatively, the network appliance can be transparent on all interfaces, or appear as a router, or appear as a bridge on all interfaces. In some embodiments, network traffic can be intercepted by another device and mirrored (copied) onto a network appliance. Each network appliance may further be either physical or virtual. For example, a virtual network appliance can be in a private data center or virtual private cloud (not shown), managed by a cloud service provider, such as Amazon Web Services, or others.


The communication network(s) 106 may comprise a Wide Area Network (WAN), the Internet, MPLS (Multiprotocol Label Switching), LTE (Long Term Evolution), or any other wired or wireless network, as will be apparent to those skilled in the art. The network appliance A 102 and the network appliance B 104 may be communicatively coupled via one or more communication network(s) 106, including any combination of WAN, Internet, MPLS, or LTE. In various embodiments, network appliance A 102 and network appliance B 104 communicate via at least two communication networks.


In various embodiments, network appliance A 102 and network appliance B 104 are optionally in communication with endpoint devices 108 and 126, respectively. The network appliances can be optionally in communication with the endpoint devices either directly or through other intermediate devices such as switches and routers (not shown). Endpoint devices 108 and 126 may each comprise a computer, a server, a mobile device, or the like as will be apparent to those skilled in the art. Endpoint devices 108 and 126 may or may not be communicatively coupled directly to one or more communication network(s) 106 in addition to, or instead of, through a network appliance.


In the exemplary depiction in FIG. 1, network appliance A 102 is configured to send data packets to network appliance B 104 via one or more communication network(s) 106. Network appliance B 104 is configured to receive data packets from network appliance A 102. In other embodiments, network appliance B 104 may transmit data and network appliance A 102 may receive data. Each network appliance can comprise a computer, a server, a router, or the like. A network appliance is described further herein with respect to FIG. 2.


The network appliance A 102 comprises sub-systems that perform certain functions. In various embodiments, network appliance A 102 comprises at least an internal source 109, a pre-processor 110, data protector 112, and a transmitter 114. Certain functions of one or more sub-systems may be combined, or performed by additional sub-systems.


In an exemplary embodiment, network appliance A 102 has data to be transmitted across one or more communication network(s) 106 to network appliance B 104 or to endpoint device 126 in communication with network appliance B 104. The data for transmission may be received by network appliance A 102 from endpoint device 108, or may be generated by internal source 109 of network appliance A 102 itself.


In various embodiments, pre-processor 110 is a hardware and/or software component of network appliance A 102. The pre-processor 110 may receive data for transmission from endpoint device 108, receive data internally generated from another sub-system of network appliance A 102, or generate the data itself. Pre-processor 110 may also optionally perform other operations on the data, such as formatting, encapsulating, coalescing, or optimizing. Optimization techniques performed to the data can include such items as compression, encryption, and adding tunnel headers. While only a few examples are listed here, a person of ordinary skill in the art would understand that pre-processor 110 can perform any data optimization technique on the data.


Data protector 112 is a hardware and/or software component of network appliance A 102. In various embodiments, data protector 112 is configured to generate protection packets based on data packets received from the pre-processor 110 and a protection ratio. Protection may be used by, for example, the network appliance B 104 to reconstruct data packets that are corrupted or missing. A packet protection ratio indicates the number of protection packets per the number of data packets, and a byte protection ratio indicates the number of correction bytes per the number of data bytes in the data frame (also referred to herein as error correcting frame, or simply frame).


In some systems, one or more protection packets are sent with each frame. The protection packet may be a byte-wise exclusive-or (XOR) of all of the data packets in the frame. As is known to those skilled in the art, any single missing or corrupted data packet can be recovered using an XOR operation. Although the XOR operation is discussed, there are many ways one or more missing or corrupted data packets can be recovered, depending on the protection algorithm. The error correction packet is often referred to as a parity packet. While XOR is often used to generate parity packets, there are other methods for generating other types of protection packets (also sometimes referred to herein as protection data or protection information) which may use other operations.


In various embodiments, data protector 112 is configured to generate more than one protection packet per a defined number of data packets. For example, the packet protection ratio may be 8:40 indicating that, for a frame comprising forty data packets, eight distinct protection packets are additionally generated. This may be performed using a unique, identifiable function to generate each of the protection packets. Using a ratio of 8:40 rather than a ratio of 1:5 allows for more missing or corrupted packets that are, for example, sequential to one another, to be reconstructed and thus avoids resending multiple packets. The more than one protection packets may be generated based on algorithms using Reed-Solomon coding and finite field arithmetic, as is known to those skilled in the art, or other methods.


Further, in some embodiments, the data protector 112 is configured to change the protection ratio dynamically. This may be in response to network measurements, for example increasing the ratio when the observed loss is low and decreasing the ratio which the observed loss is high. In some circumstances the ratio will be changed when the framing algorithm “times out” because the stream has no packets to send.


Transmitter 114 is a hardware and/or software component of network appliance A 102. In various embodiments, transmitter 114 is configured to queue and transmit data packets and protection packets across one or more communication network(s) 106 to network appliance B 104. Protection packets may be transmitted substantially simultaneously as data packets, or may be sent at a delay or upon the meeting of a condition. Exemplary conditions include a frame boundary, a network condition, or a timeout parameter being met. In various embodiments, transmitter 114 can send data across multiple networks of the one or more communication network(s) 106 simultaneously, or in an alternating, interleaving or interspersed manner.


The network appliance B 104 comprises sub-systems that perform certain functions. In various embodiments, network appliance B 104 comprises at least, a receiver 116, a data corrector 118, a post-processor 122, an internal destination 124, and optionally an ordering sub-system 120. The network appliance B 104 may additionally be configured as a network appliance A 102 and may be a computer, server, router, network appliance, or the like.


Receiver 116 is a hardware and/or software component of network appliance B 104. The receiver 116 is configured to receive data packets and protection packets from the network appliance A 102 via one or more communication network(s) 106.


Data corrector 118 is a hardware and/or software component of network appliance B 104. Data corrector 118 is configured to determine if any data packets are missing, corrupted, or out of order. If any data packets are missing or corrupted, data corrector 118 is configured to determine whether the missing or corrupted packets can be reconstructed using a combination of the correctly received data packets and the received protection packets and, if the packets can be reconstructed, reconstruct the packets.


When present, the ordering sub-system 120 is configured to order the received packets, or copies of the received packets, and the reconstructed packets prior to processing by the post-processor 122. Ordering may be based on a sequence number in the header information of each packet. Ordering sub-system 120 is a hardware and/or software component of network appliance B 104. In exemplary embodiments where network appliance B 104 does not include an ordering sub-system 120, data packets may be re-ordered by endpoint device 126 or another recipient of the data. In some embodiments, the network appliance B 104 is configured to copy incoming packets, to send one copy to the data corrector 118, and send the other copy to the ordering sub-system 120.


Post-processor 122 is configured to receive data packets and reconstructed data packets from data corrector 118 and/or ordering sub-system 120, and deliver the data to endpoint device 126 or to an internal destination 124 of network appliance B 104. Post-processor 122 is a hardware and/or software component of network appliance B 104 that often performs an inverse operation to pre-processor 110 (decrypt, decompress, etc.).



FIG. 2 illustrates a block diagram of a network appliance 202, in an exemplary implementation of the disclosure. Network appliance 202 can be configured as network appliance A 102 or network appliance B 104 of FIG. 1. The network appliance 202 includes a processor 210, a memory 220, a WAN communication interface 230, a LAN communication interface 240, and a database 250. A system bus 280 links the processor 210, the memory 220, the WAN communication interface 230, the LAN communication interface 240, and the database 250. Line 260 links the WAN communication interface 230 to another device, such as another appliance, router, or gateway, and line 270 links the LAN communication interface 240 to a user computing device, or other networking device. While network appliance 202 is depicted in FIG. 2 as having these exemplary components, the appliance may have additional or fewer components.


The database 250 comprises hardware and/or software elements configured to store data in an organized format to allow the processor 210 to create, modify, and retrieve the data. The hardware and/or software elements of the database 250 may include storage devices, such as RAM, hard drives, optical drives, flash memory, and magnetic tape.


In some embodiments, some network appliances comprise identical hardware and/or software elements. Alternatively, in other embodiments, some network appliances may include hardware and/or software elements providing additional processing, communication, and storage capacity.


Each network appliance 202 can be in communication with at least one other network appliance 202, whether in the same geographic location, different geographic location, private cloud network, customer datacenter, or any other location. As understood by persons of ordinary skill in the art, any type of network topology may be used. There can be one or more secure tunnels between one or more network appliances. The secure tunnel may be utilized with encryption (e.g., IPsec), access control lists (ACLs), compression (such as header and payload compression), fragmentation/coalescing optimizations and/or error detection and correction provided by an appliance.


A network appliance 202 can further have a software program operating in the background that tracks its activity and performance. For example, information about data streams that are processed by the network appliance 202 can be collected. Any type of information about a stream can be collected, such as header information (source port, destination port, source address, destination address, protocol, etc.), packet count, byte count, timestamp, traffic type, or any other stream attribute.


Typically in forward packet recovery, an initial packet protection ratio K:N, indicating a first number of protection packets per a second number of data packets in a frame, is determined. In embodiments of the present disclosure, a byte protection ratio can be determined, indicating a number of correction bytes per a number of data bytes in a frame. The process of determining the correction bytes may be performed by the data protector 112 in the network appliance A 102.


In other systems, forward packet recovery is based on data packets, regardless of the size of each packet. An error correction packet, typically a parity packet, can be generated for each frame, with the error correcting information. FIG. 3 depicts an exemplary frame of prior systems. The frame consists of data packets 310, 320, 330 and 340, and error correction packet 350. The parity packet is generated to typically be the same size as the largest data packet in the frame. A placeholder (typically zeros) is implicitly added to the other data packets in the frame to make them all effectively the size of the largest data packet, which is packet 320 in the figure. Transmitter 114 may typically transmit the data packets across communication network(s) 106 without the placeholder zeroes. While the placeholder is described here as being a zero, in other embodiments the placeholder can be any value that is constant, such as any other number, letter, or symbol that is consistently used.


In the exemplary system of FIG. 3, one protection packet is generated for the frame of four data packets. Thus the packet protection ratio is 1:4. Since one packet out of every five total packets is a protection packet, one would expect that 20% of the network bandwidth is used to transmit error correcting information. However, since the parity packet is generated to be the size of the largest data packet, there is 1500 bytes of error correcting information for 3600 bytes of data. Thus, 1500 bytes out of 5100 total bytes used in the network is for error correcting information, resulting in a 29% network overhead for error correcting information. Thus, a packet protection ratio on a packet basis is not equivalent to a network bandwidth usage based on bytes.


Further, when data and error correcting information is transmitted across communication network(s) 106, a link of the network may have a set capacity. For example, an MPLS link may have a 10 MB capacity and an Internet link may have a 10 MB capacity. In this example, transmitting data on the MPLS link and error correcting information on the Internet link in a 1:1 ratio, could mean the transmitter could send 10 mbps of data on the MPLS link and yet have 15 mbps of error correction packets on the Internet link, thus exceeding the capacity of the Internet link. Therefore, it is advantageous to send a specified number of bytes of error correcting information that is close to a target byte protection ratio, rather than sending a target number of data packets of error correcting information.


In an example embodiment of the present disclosure, a target byte protection ratio is selected or chosen, such as 1:4. This ratio means that for every four bytes of data, one byte of error correcting recovery information is transmitted, thus meaning 20% of network capacity is being used for network overhead, i.e. one byte out of every five bytes total is non-data. While embodiments of the present disclosure provide mechanisms for approaching a target network overhead percentage, the actual overhead percentage achieved may not be exactly the target percentage.


In the example embodiment depicted in FIG. 4A, the size of a frame is dynamically selected such that a frame begins and ends when a packet size changes significantly. Packets in a stream may be of varying sizes. When a next packet in the stream is much smaller or much larger than a maximum size of packets in the frame so far, a frame boundary may be inserted. For example, a data packet may be 1500 bytes, but an acknowledgement packet may be less than 100 bytes. Multiple sequences of data packets and acknowledgement packets may be present in a single data stream.


In the exemplary embodiment of FIG. 4A, data packets 462-488 have been arranged into three frames: frame 410 consists of six data packets, frame 420 consists of five data packets, and frame 430 consists of three data packets. Since the first packet of frame 420 is significantly smaller than the packets in frame 410, a frame boundary is placed there. Similarly, the first packet of frame 430 is significantly larger than the packets in frame 420, so a frame boundary is placed there. In this way, frames can be determined dynamically and be of different sizes for the same data stream.


The determination of the placement of frame boundary can be based on packet sizes that differ from previous packet sizes in the current frame by a certain amount or by a certain percentage. That is, if a next packet is within a certain size percentage of a previous packet in the frame, then it may be considered part of the same frame. For example, packets of sizes 1400 bytes, 1500 bytes, and 1300 bytes may all be considered part of the same frame, but a packet of 1100 bytes may be considered to be in a different frame. The amount that a next packet size should differ from previous packet sizes in order to be considered part of a new frame may be determined by the user, or may be a pre-set value, or may adjusted dynamically.


Error correction packets can be generated for each frame in accordance with the target byte protection ratio. Packets 462-488 have been grouped into frames 440, 450, and 460 in exemplary FIG. 4A with their respective error correction packets. Frame 440 of FIG. 4A consists of six data packets and a target packet protection ratio is 1:4. In order to approximately meet this ratio, either one or two error correction packets may be transmitted of the size of the biggest data packet within the frame (packet 462). In the example of FIG. 4A, one error correction packet 473 is generated for this frame, of the size of packet 462.


Frame 450 consists of five smaller data packets. To approximately meet the target packet protection ratio of 1:4, either one or two error correction packets may be transmitted of the size of the biggest data packet within the frame (packet 480). In the example of FIG. 4A, one error correction packet 483 is generated for this frame, of the size of packet 480. Frame 460 consists of three data packets. To approximately meet the target packet protection ratio of 1:4, one error correction packet 489 is generated of the same size as the largest data packet in the frame (packet 486).


The determination of whether one or two error correction packets should be generated can be based on a running tally (also called running average herein) of the amount of network overhead used for error correcting information, to keep the target network overhead at around 20% for each data stream or portion thereof. For example, frame 450 might have one error correction packet generated since the packets are larger, and frame 460 might have two error correction packets since the packets are of smaller size and use less network overhead. Thus, the size of error correcting information is tracked in a running tally such that an overall network overhead used for error correcting information for all frames in the stream is kept at approximately the target network overhead usage percentage of 20%.



FIG. 4B depicts an exemplary flow sequence for a data protector 112 to determine how many error correction packets to generate for each frame in a stream. In step 490, data protector 112 receives information regarding the frame, such as number of data packets, number of data bytes, and other relevant information such as the running average of network overhead usage thus far. The data protector 112 compares in step 492 the running average of the network overhead bytes used thus far for error correcting information with the target byte protection ratio.


If the network overhead usage is above the target byte protection ratio, then for the current frame being evaluated, data protector 112 may round down the number of protection packets generated, in step 496. If the network overhead usage is below the target byte protection ratio, then data protector 112 may round up the number of protection packets generated for the frame, in step 494.


For example, in frame 440, either one or two error correction packets can be generated to approximate the target byte protection ratio of 1:4. If data protector 112 determines that the running average of the network overhead used for the stream up to that point is above the target byte protection ratio, then one protection packet 473 is generated for the frame. If data protector 112 determines that the network overhead for the stream thus far is below the target byte protection ratio, then two protection packets can be generated for the frame. The running average of network overhead is then updated accordingly in step 498. In some embodiments the rounding functions may be designed to ensure that the ratio stays within maximum and/or minimum bounds.


By dynamically selecting frame boundaries to group packets of similar size, packet recovery information generated and transmitted across the communication network(s) is closer to the target packet protection ratio, and network bandwidth can be saved.


In other systems, when a data packet is to be transmitted across a communication network that is larger than the maximum transmission unit (MTU), the packet is fragmented. For example, if a 1600 byte data packet is to be transmitted across a MTU with a capacity of 1500 bytes, then a 1500 byte packet is created and a second 100 byte packet is created, as depicted in FIG. 5A. Packet 510 is greater than the MTU, so packet 520 is generated that is of the same size as the MTU and packet 530 contains the remaining data. Note that depending on the fragmentation format, packet 530 may be slightly larger than 100 bytes due to fragmentation option overhead and/or rounding of fragmentation field values. We ignore this overhead here to simplify the explanation.


Due to the significant difference in data packet sizes of packet 520 and packet 530, having a sequence of individual large and small packets in series will cause every frame to be the size of one packet (if the previously described approach of starting a new frame when there is a significant change in the packet size). This would increase the actual error correcting ratio, effectively making it 1:1.


Thus, in embodiments of the present disclosure, a large packet that is above the capacity of the MTU over which it is to be transmitted can be fragmented into equal parts (for a packet with an even number of bytes), as depicted in FIG. 5B. That is, packet 510 is greater than the MTU, so it is fragmented into packets 540 and 550 of equal parts. If packet 510 has an odd number of bytes, then it is fragmented into packets of substantially equal size parts (e.g. one fragmented packet may have one additional byte compared to the second fragmented packet). Since these packets are of the same size, they can be grouped into a single frame, thus facilitating the dynamic framing for the stream. While packets are depicted as being fragmented into two parts here, they can be fragmented into any number of substantially equal size portions (i.e. three, four, or more parts).



FIG. 6 depicts another exemplary embodiment of dynamic framing for approximating a target packet protection ratio. An exemplary stream of data packets 610-680 is presented in the figure. Using the process described above for placing frame boundaries in accordance with packet sizes, four frames would be generated from these eight data packets—a first frame consisting of packets 610 and 620, a second frame consisting of packets 630, 640, 650, a third frame consisting of packet 660, and a fourth frame consisting of packets 670 and 680. However, since an error correction packet is generated for each frame, this would result in four error correction packets and generating and transmitting error correcting information using more than the target 20% network resources.


In order to facilitate reaching the target of 20% network overhead usage, fewer error correction packets are needed by having fewer frames. Dynamic framing is used such that each frame consists of a larger number of packets. The packets from the stream can be placed out of order and each frame grouped according to packet size. Thus the stream can be broken into two frames—a first frame consisting of packet 610, 620 and 660, and a second frame consisting of packets 630, 640, 650, 670 and 680. Error correction packet 665 can be generated for the first frame, of the size of the largest data packet in the frame. Error correction packet 685 can be generated for the second frame, of the size of the largest data packet in that frame. Simply by reorganizing the frames in this way, the amount of error correcting information for the stream is significantly reduced, thus saving network bandwidth and achieving a network overhead usage percentage that is closer to the target byte protection ratio of 20%.



FIG. 7 depicts another exemplary embodiment of generating error correcting information based on target byte protection ratio. FIG. 7 depicts an exemplary sequence of data packets 702-734 to be transmitted from a stream. In an exemplary embodiment, a target packet protection ratio may be 1:5. There are seventeen data packets in the frame, of varying sizes. Each larger data packet may be segmented into two parts, with a smaller portion being an equivalent size to the smallest data packet in the frame. That is, packet 702 can be segmented into a portion 702a and 702b, with portion 702b being the same size as packet 714. The segmentation can be virtual, and not actually generate two separate packets, as would be understood by a person of ordinary skill in the art.


In placing frame boundaries for generating protection packets, all of the packets of the smaller size and the smaller portion of the larger packets can be considered a single frame, while the remaining larger portions of the larger data packets can be considered a second frame. That is, a first frame can be 702b-712b, 714-722, 724b-726b, 728-730, and 732b-734b. A second frame can be 702a-712a, 724a-726a and 732a-734a.


The first frame consists of seventeen total packets and packet portions. Using a target byte protection ratio of 1:5, four protection packets may be generated. These error correction packets contain information for the small data packets and a small portion of the larger data packets (the “b” portions). The four protection packets generated are depicted as 740, 742, 744b, and 746b.


For the second frame consisting of the remaining portions of the larger data packets (the “a” portions), error correcting information can be similarly generated. There are ten packet portions of larger data packets remaining, thus according to the target byte protection ratio of 1:5, two error correcting portions are generated of that size. Rather than generate and transmit two new error correction packets, the error correcting information for the second frame can be added to an existing smaller protection packet, as depicted in the figure by 744a and 746a. In this way, there are a total of 4 error correction packets for the stream of seventeen data packets (instead of six error correction packets), thus keeping close to the 1:5 overall target packet and byte protection ratios for the frame.


In a further embodiment of the present disclosure, as data packets are received by the network appliance 202 for transmission, they are populated into a data structure in memory, which can be represented as a rectangle. A vertical size of the rectangular block of memory can be a byte size which is dynamic. A number of columns in the rectangular block can be based on the target packet protection ratio. Thus, for a 1:4 target byte protection ratio, a rectangular block can be generated of four columns for data, with the size of each column being variable such that all of the data fits within 4 columns only. A singular column of protection data is further generated with error correcting information for the columns. As discussed herein, protection data can use any error correcting code, such as Reed-Solomon coding and finite field arithmetic



FIG. 8 depicts an exemplary stream of data packets 802-834 for transmission. As data packets are queued for transmission, they populate the rectangular block from bottom to top, left to right. A data packet can wrap around into multiple columns, as shown by packet 826a and 826b in columns 3 and 4. Each data packet's header can be amended with information regarding its placement within the rectangular block in memory, such that data packets can be placed in order by data corrector 118 or ordering sub-system 120, even if received out of order or lost in transmission. As would be understood by a person of ordinary skill in the art, data packets can be placed in a different manner, such as from top to bottom, right to left, or in any other manner. Similarly, while a rectangular grid structure is depicted here, other data structure formats can be used.


Once the data packets are placed in the grid, the grid is read in a horizontal manner, and protection information (also referred to herein as error correcting information or packet recovery information) is generated by the data protector 112 for each horizontal line of data. In this way, error correcting information is generated in a ratio of 1:4, which is the target byte protection ratio. Protection information is generated as a column “P”, which can then be segmented into one or more data packets (P1-P3) for transmission across a communication network 106.


A receiving network appliance, such as network appliance B 104 of FIG. 1 can similarly populate a rectangular grid to determine missing or corrupt information, for repair with the protection packet(s). Once the stream has been received and the grid generated, the data corrector 118 of the receiving network appliance can read the rectangular block by each horizontal byte of data. If only one column of the four columns per each horizontal line has missing or corrupt information, the protection packet(s) can repair that information. However, if information is missing from more than one column of the rectangle for that horizontal line of bytes, then the protection packet will be unable to correct the information, since the protection ratio is 1:4 and there is only one portion of error correcting information for four portions of data. Thus, if more than one portion of data is missing or corrupt, it cannot be corrected.


In this embodiment, a vertical length, “A”, for the rectangular block can be any number of bytes. For example, A might be 10,000 bytes. When scanning the block to repair the data, the receiving network appliance runs 10,000 error correcting calculations (one for each horizontal byte line of data). When there are large amounts of data to be transmitted, A can be set to be a large value. When there are small amounts of data to be transmitted, A can be set to a small value. Similarly the number of columns can be changed dynamically based on the target byte protection ratio.



FIG. 9A depicts a further embodiment of the present disclosure for generating error correcting information in accordance with a target byte protection ratio. As data packets are received by the network appliance 202 for transmission, they are populated into a data structure in memory, which can be represented as a rectangle. A vertical size of the rectangular block of memory can be a byte size which is dynamic. A number of columns in the rectangular block can be based on the target packet protection ratio. Thus, for a 1:4 target byte protection ratio, a rectangular block can be generated of 4 columns for data, with the size of each column being variable such that all of the data fits within 4 columns only. A singular column of protection data is further generated with error correcting information for the columns. As discussed herein, protection data can use any error correcting code, such as Reed-Solomon coding and finite field arithmetic.



FIG. 9A depicts an exemplary stream of data packets 902-934 for transmission. As data packets are queued for transmission, they populate a grid with defined rows 940-952. Since the target byte protection ratio is 1:4, the grid has four columns. The columns can be of any length to accommodate the data. That is, the columns can have as many rows as necessary to accommodate the data, and each row can be of an accommodating size.


In an exemplary embodiment, the grid is populated from bottom to top, left to right, with each data packet beginning at a boundary of a row. Data packet 902 is placed in the grid starting from a boundary of row 940. The packet is of a size that it completely covers rows 940 and 942. The next packet, packet 904, is placed in the grid beginning at the next row boundary, which is row 944. Packet 904 is also of a size that it completely covers rows 944 and 946. Packets are placed in the grid in this manner. Packet 908 is placed in the grid beginning at the boundary of row 952, but wraps around to column 2 to fully accommodate all of the data, represented as 908a and 908b in exemplary FIG. 9.


If packets are of uneven length, there may be unfilled space in the rectangle between the end of a prior packet and the beginning of a subsequent packet. For example, packet 914 is placed in the grid at the beginning of the boundary of row 950. The packet is of a smaller size though and does not take up the whole space of row 950. Thus, there is some unfilled space in the grid because packet 916 is placed in the grid starting at the beginning of the next row boundary, which is row 952. Unfilled spaces in the grid are assigned a predetermined value, typically zero. All of the data packets 902-934 continue to be placed in the grid in this manner.


Each data packet's header can be amended with information regarding its placement within the rectangular block in memory, such that data packets can be placed in order by data corrector 118 or ordering sub-system 120, even if received out of order or lost in transmission.


As would be understood by a person of ordinary skill in the art, data packets can be placed in a different manner, such as from top to bottom, right to left, or in any other manner. Similarly, while a rectangular grid structure is depicted here, other data structure formats can be used.


Once the data packets are placed in the grid, the grid is read in a horizontal manner, and protection information is generated by the data protector 112 for each row of data. In this way, error correcting information is generated in a ratio of 1:4, which is the target byte protection ratio. Protection information is generated as a column “P”, which can then be segmented into one or more data packets (P1-P3) for transmission across a communication network 106.


A receiving network appliance can similarly populate a rectangular grid to determine missing or corrupt information, for repair with the protection packet(s). Once the stream has been received and the grid generated, the data corrector 118 of the receiving network appliance can read the rectangular block by each row of data. If only one column of the four columns per row has missing or corrupt information, the protection packet(s) can repair that information. However, if information is missing from more than one column of the rectangle for that row, then the protection packet will be unable to correct the information, since the protection ratio is 1:4 and there is only one portion of error correcting information for four portions of data. Thus, if more than one portion of data is missing or corrupt, it cannot be corrected.


For example, row 940 contains data from packets 902, 908, 918 and 928. If only one of these packets is missing or corrupt, then protection packet P3 can correct it. However, if more than one of these packets is missing or corrupt, then the protection packet P3 will not be able to correct the data in the row. Similarly, row 952 has data from packets 908, 916 and 926. If only one of these packets is missing or corrupt, then protection packet P1 can correct it. However, if more than one of these packets is missing or corrupt, then the protection packet will not be able to correct the data in the row.


In this embodiment, a vertical length, “A”, for the rectangular block can be any number of rows. In the exemplary embodiment of FIG. 9A, seven rows are depicted. The rows may comprise any number of bytes, such as 10,000 bytes. When scanning the block to repair the data, the receiving network appliance scans each row and thus runs seven error correcting calculations (one for each row of data). In this way, instead of running 10,000 error correcting calculations, one for each horizontal byte row of data, error correcting calculations are run for each row of data, which is a smaller value by orders of magnitude. When there are large amounts of data to be transmitted, A can be set to be a large value. When there are small amounts of data to be transmitted, A can be set to a small value.



FIG. 9B depicts a further embodiment of the present disclosure for generating error correcting information in accordance with a target byte protection ratio. As data packets are received by the network appliance 202 for transmission, they are populated into a data structure in memory, which can be represented as a rectangle. A vertical size of the rectangular block of memory can be a byte size which is dynamic. A number of columns in the rectangular block can be based on the target packet protection ratio. Thus, for a 2:4 target byte protection ratio, a rectangular block can be generated of 4 columns for data, with the size of each column being variable such that all of the data fits within 4 columns only. Two columns of protection data are further generated with error correcting information for the columns. As discussed herein, protection data can use any error correcting code, such as Reed-Solomon coding and finite field arithmetic.



FIG. 9B depicts an exemplary stream of data packets 902-934 for transmission. As data packets are queued for transmission, they populate a grid with defined rows 940-952. Since the target packet protection ratio is 2:4, the grid has 4 columns. The columns can be of any length to accommodate the data. That is, the columns can have as many rows as necessary to accommodate the data, and each row can be of an accommodating size.


In an exemplary embodiment, the grid is populated as discussed above with respect to FIG. 9A. Once the data packets are placed in the grid, the grid is read in a horizontal manner, and protection information is generated by the data protector 112 for each row of data. In this way, error correcting information is generated in a ratio of 2:4, which is the target packet protection ratio. Protection information is generated as columns 960 and 970. Each column can be segmented into one or more data protection packets, for example column 960 can be segmented into protection packets 962, 964 and 966. Column 970 can be segmented into protection packets 972, 974 and 976 for transmission across a communication network 106.


A receiving network appliance can similarly populate a rectangular grid to determine missing or corrupt information, for repair with the protection packet(s) as discussed above with respect to FIG. 9A. Once the stream has been received and the grid generated, the data corrector 118 of the receiving network appliance can read the rectangular block by each row of data. If two or fewer columns of the four columns per row have missing or corrupt information, the protection packet(s) can repair that information. However, if information is missing from more than two columns of the rectangle for that row, then the protection packets will be unable to correct the information, since there are two portions of error correcting information for four portions of data. Thus, if more than two portions of data are missing or corrupt, the affected packets cannot be reconstructed.


For example, row 940 contains data from packets 902, 908, 918 and 928. If only one of these packets is missing or corrupt, then protection packet 966 or 976 can correct it. However, if more than two of these packets is missing or corrupt, then the protection packets 966 and 976 will not be able to correct the data in the row. Similarly, row 952 has data from packets 908, 916 and 926. If only one of these packets is missing or corrupt, then protection packet 962 or 972 can correct it. However, if more than two of these packets is missing or corrupt, then the protection packets will not be able to correct the data in the row.


In this embodiment, a vertical length, “A”, for the rectangular block can be any number of rows. In the exemplary embodiment of FIG. 9B, 7 rows are depicted. The rows may comprise any number of bytes, such as 10,000 bytes. When scanning the block to repair the data, the receiving network appliance scans each row and thus runs seven error correcting calculations (one for each row of data). In this way, instead of running 10,000 error correcting calculations, one for each horizontal byte row of data, error correcting calculations are run for each row of data, which is a smaller value by orders of magnitude. When there are large amounts of data to be transmitted, A can be set to be a large value. When there are small amounts of data to be transmitted, A can be set to a small value. Similarly the number of data columns and the number of protection columns can be varied dynamically as the target byte ratio changes.


Thus, methods and systems for forward packet recovery with constrained overhead are disclosed. Although embodiments have been described with reference to specific examples, it will be evident that various modifications and changes can be made to these example embodiments without departing from the broader spirit and scope of the present application. Therefore, these and other variations upon the exemplary embodiments are intended to be covered by the present disclosure. Accordingly, the specification and drawings are to be regarded in an illustrative rather than a restrictive sense.

Claims
  • 1. A computer-implemented method for providing corrected data packets to a second processor, the method comprising: receiving at a first processor, a data stream comprising a plurality of data packets for transmission across at least one communication network, the plurality of data packets being of various sizes;selecting a target byte protection ratio for the data stream, the target byte protection ratio representing a target number of bytes of error correcting information per bytes of data in the data stream;categorizing the plurality of data packets in the data stream into a plurality of categories according to size, the categories comprising data packets of a first size range and data packets of a second size range, wherein the second size range is larger than the first size range;generating one or more error correcting frames for each category of the plurality of categories, such that each error correcting frame comprises data packets of a comparable size;generating one or more error correction data packets for each error correcting frame, in accordance with the target byte protection ratio; andtransmitting the plurality of data packets and the one or more error correction data packets across the at least one communication network to the second processor.
  • 2. The method of claim 1, wherein the target byte protection ratio for the data stream is selected to correspond to a target network bandwidth usage for the error correcting information.
  • 3. The method of claim 1, wherein the one or more error correcting frames are further generated in accordance with a timeout policy, such that all packets received within a certain period of time are determined to be within the one or more error correcting frames.
  • 4. The method of claim 1, wherein the generating one or more error correcting frames for each category further comprises generating one or more error correcting frames of information for corresponding data packets, such that the error correcting information for the data stream is in a different order than the data packets of the data stream.
  • 5. The method of claim 1, wherein the transmitting the plurality of data packets comprises transmitting the data packets from the data stream in order of receipt, and the transmitting the one or more error correction data packets comprises transmitting error correcting information out of order from an order of corresponding data packets in the data stream.
  • 6. The method of claim 1, wherein the generating one or more error correction data packets in accordance with the target byte protection ratio further comprises: keeping a running tally of a number of error correction data packets generated for all error correcting frames in the data stream; andadjusting an amount of the error correction data packets generated for each subsequent frame based on the running tally of error correction data packets for the data stream, in accordance with the target byte protection ratio for the data stream.
  • 7. The method of claim 1, wherein the receiving the plurality of data packets for transmission across the at least one communication network further comprises: comparing a size of each data packet to a maximum transmission unit capacity of the at least one communication network; andfragmenting each data packet above the capacity of the maximum transmission unit for the at least one communication network into two or more portions of substantially equal size.
  • 8. The method of claim 1, wherein the generating one or more error correcting frames for the data stream further comprises: segmenting the data packets of the second size range into a first portion equivalent in size to at least one of the data packets of the first size range and a second portion comprising a remainder of the data packets of the second size range;generating at least one error correcting frame for the data packets of the first size range and the first portion of the data packets of the second size range; andgenerating at least one error correcting frame for the second portion of the data packets of the second size range.
  • 9. The method of claim 1, wherein at least one of the plurality of data packets is received out of order at the second processor.
  • 10. The method of claim 1, wherein the error correcting information in at least one error correction data packet is generated using Reed-Solomon coding.
  • 11. The method of claim 1, further comprising regenerating the data stream at the second processor by: receiving the plurality of data packets and the one or more error correction data packets from the first processor; andutilizing information in the received plurality of data packets and the received one or more error correction data packets to regenerate the data stream by correcting at least one lost, corrupt, or out of order data packet in the data stream.
  • 12. The method of claim 1, further comprising regenerating the data stream at the second processor by: receiving the plurality of data packets and the one or more error correction data packets from the first processor; andutilizing a sequence number in a header of the received plurality of data packets and the received one or more error correction data packets to regenerate the data stream by correcting at least one lost, corrupt, or out of order data packet in the data stream.
  • 13. A computer-implemented method for regenerating a data stream received at a second processor from a first processor across at least one communication network; the method comprising: receiving a plurality of data packets and one or more error correction data packets from the first processor; wherein the one or more error correction data packets are generated by the first processor in accordance with a target byte protection ratio for the data stream, the target byte protection ratio representing a target number of bytes of error correcting information per bytes of data in the data stream; andutilizing the received plurality of data packets and the received one or more error correction data packets to regenerate the data stream by correcting at least one lost, corrupt, or out of order data packet in the data stream.
  • 14. The method of claim 13, wherein the one or more error correction data packets are generated from one or more generated error correcting frames, each of the one or more generated error correcting frame comprising data packets of a comparable size.
  • 15. The method of claim 13, wherein the one or more error correction data packets are generated by the first processor with error correcting information for the data stream that is in a different order than the data packets of the data stream.
  • 16. The method of claim 13, wherein the target byte protection ratio for the data stream is selected to correspond to a target network bandwidth usage for the error correcting information.
  • 17. A computing system for providing corrected data packets to a second network appliance, the system comprising: a first network appliance in communication with a second network appliance via at least one communication network, the first network appliance comprising at least one processor configured to: receive a data stream comprising a plurality of data packets for transmission across the at least one communication network to the second network appliance;select a target byte protection ratio for the data stream, the target byte protection ratio representing a target number of bytes of error correcting information per bytes of data in the data stream;categorize the plurality of data packets in the data stream according to size, the categories comprising a first category of data packets within a first size range and a second category of data packets within a second size range;generate one or more error correcting frames for each category, such that each error correcting frame comprises data packets of a comparable size;generate one or more error correction data packets for each generated error correcting frame, in accordance with the target byte protection ratio; andtransmit the plurality of data packets and the one or more error correction data packets across the at least one communication network to the second network appliance.
  • 18. The system of claim 17, wherein the second network appliance comprises at least one processor configured to: receive the plurality of data packets and the one or more error correction data packets from the first network appliance; andutilize the received plurality of data packets and the received one or more error correction data packets to regenerate the data stream in order by correcting at least one lost, corrupt, or out of order data packet in the data stream.
  • 19. The system of claim 17, wherein the target byte protection ratio for the data stream is selected to correspond to a target network bandwidth usage for the error correcting information.
  • 20. The system of claim 17, wherein the generating one or more error correcting frames for each category further comprises generating one or more error correcting frames of information for corresponding data packets, such that the error correcting information for the data stream is in a different order than the data packets of the data stream.
CROSS-REFERENCE TO RELATED APPLICATIONS

The present patent application is a Continuation of, and claims the priority benefit of, U.S. patent application Ser. No. 15/918,807 filed on Mar. 12, 2018 and entitled “Forward Packet Recovery with Constrained Network Overhead”, now granted as U.S. Pat. No. 10,326,551 issued on Jun. 18, 2019, which in turn is a Divisional Application of, and claims the priority benefit of, U.S. patent application Ser. No. 15/241,992 filed on Aug. 19, 2016 and entitled “Forward Packet Recovery with Constrained Overhead”, now granted as U.S. Pat. No. 9,967,056 issued on May 8, 2018. The disclosures of the above-referenced applications are hereby incorporated by reference in their entirety for all purposes.

US Referenced Citations (563)
Number Name Date Kind
4494108 Langdon, Jr. et al. Jan 1985 A
4558302 Welch Dec 1985 A
4612532 Bacon et al. Sep 1986 A
5023611 Chamzas et al. Jun 1991 A
5159452 Kinoshita et al. Oct 1992 A
5243341 Seroussi et al. Sep 1993 A
5307413 Denzer Apr 1994 A
5357250 Healey et al. Oct 1994 A
5359720 Tamura et al. Oct 1994 A
5373290 Lempel et al. Dec 1994 A
5483556 Pillan et al. Jan 1996 A
5532693 Winters et al. Jul 1996 A
5592613 Miyazawa et al. Jan 1997 A
5602831 Gaskill Feb 1997 A
5608540 Ogawa Mar 1997 A
5611049 Pitts Mar 1997 A
5627533 Clark May 1997 A
5635932 Shinagawa et al. Jun 1997 A
5652581 Furlan et al. Jul 1997 A
5659737 Matsuda Aug 1997 A
5675587 Okuyama et al. Oct 1997 A
5710562 Gormish et al. Jan 1998 A
5748122 Shinagawa et al. May 1998 A
5754774 Bittinger et al. May 1998 A
5802106 Packer Sep 1998 A
5805822 Long et al. Sep 1998 A
5883891 Williams et al. Mar 1999 A
5903230 Masenas May 1999 A
5955976 Heath Sep 1999 A
6000053 Levine et al. Dec 1999 A
6003087 Housel et al. Dec 1999 A
6054943 Lawrence Apr 2000 A
6081883 Popelka et al. Jun 2000 A
6084855 Soirinsuo et al. Jul 2000 A
6175944 Urbanke et al. Jan 2001 B1
6191710 Waletzki Feb 2001 B1
6240463 Benmohamed et al. May 2001 B1
6295541 Bodnar et al. Sep 2001 B1
6308148 Bruins Oct 2001 B1
6311260 Stone et al. Oct 2001 B1
6339616 Kovalev Jan 2002 B1
6374266 Shnelvar Apr 2002 B1
6434191 Agrawal et al. Aug 2002 B1
6434641 Haupt et al. Aug 2002 B1
6434662 Greene et al. Aug 2002 B1
6438664 McGrath et al. Aug 2002 B1
6452915 Jorgensen Sep 2002 B1
6463001 Williams Oct 2002 B1
6489902 Heath Dec 2002 B2
6493698 Beylin Dec 2002 B1
6570511 Cooper May 2003 B1
6587985 Fukushima et al. Jul 2003 B1
6614368 Cooper Sep 2003 B1
6618397 Huang Sep 2003 B1
6633953 Stark Oct 2003 B2
6643259 Borella et al. Nov 2003 B1
6650644 Colley et al. Nov 2003 B1
6653954 Rijavec Nov 2003 B2
6667700 McCanne Dec 2003 B1
6674769 Viswanath Jan 2004 B1
6718361 Basani et al. Apr 2004 B1
6728840 Shatil et al. Apr 2004 B1
6738379 Balazinski et al. May 2004 B1
6754181 Elliott et al. Jun 2004 B1
6769048 Goldberg et al. Jul 2004 B2
6791945 Levenson et al. Sep 2004 B1
6823470 Smith Nov 2004 B2
6839346 Kametani Jan 2005 B1
6842424 Key Jan 2005 B1
6856651 Singh Feb 2005 B2
6859842 Nakamichi et al. Feb 2005 B1
6862602 Guha Mar 2005 B2
6910106 Sechrest et al. Jun 2005 B2
6963980 Mattsson Nov 2005 B1
6968374 Lemieux et al. Nov 2005 B2
6978384 Milliken Dec 2005 B1
7007044 Rafert et al. Feb 2006 B1
7020750 Thiyagaranjan et al. Mar 2006 B2
7035214 Seddigh et al. Apr 2006 B1
7047281 Kausik May 2006 B1
7069268 Bums et al. Jun 2006 B1
7069342 Biederman Jun 2006 B1
7110407 Khanna Sep 2006 B1
7111005 Wessman Sep 2006 B1
7113962 Kee et al. Sep 2006 B1
7120666 McCanne et al. Oct 2006 B2
7145889 Zhang et al. Dec 2006 B1
7149953 Cameron et al. Dec 2006 B2
7177295 Sholander et al. Feb 2007 B1
7197597 Scheid et al. Mar 2007 B1
7200847 Straube et al. Apr 2007 B2
7215667 Davis May 2007 B1
7216283 Shen et al. May 2007 B2
7242681 Van Bokkelen et al. Jul 2007 B1
7243094 Tabellion et al. Jul 2007 B2
7249309 Glaise Jul 2007 B2
7266645 Garg et al. Sep 2007 B2
7278016 Detrick et al. Oct 2007 B1
7318100 Demmer et al. Jan 2008 B2
7359393 Nalawade et al. Apr 2008 B1
7366829 Luttrell et al. Apr 2008 B1
7380006 Srinivas et al. May 2008 B2
7383329 Erickson Jun 2008 B2
7383348 Seki et al. Jun 2008 B2
7388844 Brown et al. Jun 2008 B1
7389357 Duffie et al. Jun 2008 B2
7389393 Karr et al. Jun 2008 B1
7417570 Srinivasan et al. Aug 2008 B2
7417991 Crawford et al. Aug 2008 B1
7420992 Fang et al. Sep 2008 B1
7428573 McCanne et al. Sep 2008 B2
7441039 Bhardwaj Oct 2008 B2
7451237 Takekawa et al. Nov 2008 B2
7453379 Plamondon Nov 2008 B2
7454443 Ram et al. Nov 2008 B2
7457315 Smith Nov 2008 B1
7460473 Kodama et al. Dec 2008 B1
7471629 Melpignano Dec 2008 B2
7496659 Coverdill et al. Feb 2009 B1
7532134 Samuels et al. May 2009 B2
7555484 Kulkarni et al. Jun 2009 B2
7571343 Xiang et al. Aug 2009 B1
7571344 Hughes et al. Aug 2009 B2
7587401 Yeo et al. Sep 2009 B2
7596802 Border et al. Sep 2009 B2
7617436 Wenger et al. Nov 2009 B2
7619545 Samuels et al. Nov 2009 B2
7620870 Srinivasan et al. Nov 2009 B2
7624333 Langner Nov 2009 B2
7624446 Wilhelm Nov 2009 B1
7630295 Hughes et al. Dec 2009 B2
7633942 Bearden et al. Dec 2009 B2
7639700 Nabhan et al. Dec 2009 B1
7643426 Lee et al. Jan 2010 B1
7644230 Hughes et al. Jan 2010 B1
7676554 Malmskog et al. Mar 2010 B1
7698431 Hughes Apr 2010 B1
7702843 Chen et al. Apr 2010 B1
7714747 Fallon May 2010 B2
7746781 Xiang Jun 2010 B1
7764606 Ferguson et al. Jul 2010 B1
7793193 Koch Sep 2010 B2
7810155 Ravi Oct 2010 B1
7826798 Stephens et al. Nov 2010 B2
7827237 Plamondon Nov 2010 B2
7849134 McCanne et al. Dec 2010 B2
7853699 Wu et al. Dec 2010 B2
7873786 Singh et al. Jan 2011 B1
7917599 Gopalan et al. Mar 2011 B1
7924795 Wan Apr 2011 B2
7925711 Gopalan et al. Apr 2011 B1
7941606 Pullela et al. May 2011 B1
7945736 Hughes et al. May 2011 B2
7948921 Hughes et al. May 2011 B1
7953869 Demmer et al. May 2011 B2
7957307 Qiu et al. Jun 2011 B2
7970898 Clubb et al. Jun 2011 B2
7975018 Unrau et al. Jul 2011 B2
7996747 Dell Aug 2011 B2
8046667 Boyce Oct 2011 B2
8069225 McCanne Nov 2011 B2
8072985 Golan et al. Dec 2011 B2
8090027 Schneider Jan 2012 B2
8090805 Chawla et al. Jan 2012 B1
8095774 Hughes et al. Jan 2012 B1
8140757 Singh Mar 2012 B1
8171238 Hughes et al. May 2012 B1
8209334 Doerner Jun 2012 B1
8225072 Hughes et al. Jul 2012 B2
8271325 Silverman et al. Sep 2012 B2
8271847 Langner Sep 2012 B2
8307115 Hughes Nov 2012 B1
8312226 Hughes Nov 2012 B2
8352608 Keagy et al. Jan 2013 B1
8370583 Hughes Feb 2013 B2
8386797 Danilak Feb 2013 B1
8392684 Hughes Mar 2013 B2
8442052 Hughes May 2013 B1
8447740 Huang et al. May 2013 B1
8473714 Hughes et al. Jun 2013 B2
8489562 Hughes et al. Jul 2013 B1
8516158 Wu et al. Aug 2013 B1
8553757 Florencio et al. Oct 2013 B2
8565118 Shukla et al. Oct 2013 B2
8570869 Ojala Oct 2013 B2
8576816 Lamy-Bergot et al. Nov 2013 B2
8595314 Hughes Nov 2013 B1
8613071 Day et al. Dec 2013 B2
8681614 McCanne et al. Mar 2014 B1
8699490 Zheng et al. Apr 2014 B2
8700771 Ramankutty et al. Apr 2014 B1
8706947 Vincent Apr 2014 B1
8725988 Hughes et al. May 2014 B2
8732423 Hughes May 2014 B1
8738865 Hughes et al. May 2014 B1
8743683 Hughes Jun 2014 B1
8755381 Hughes et al. Jun 2014 B2
8775413 Brown et al. Jul 2014 B2
8811431 Hughes Aug 2014 B2
8843627 Baldi et al. Sep 2014 B1
8850324 Clemm et al. Sep 2014 B2
8885632 Hughes et al. Nov 2014 B2
8891554 Biehler Nov 2014 B2
8929380 Hughes et al. Jan 2015 B1
8929402 Hughes Jan 2015 B1
8930650 Hughes et al. Jan 2015 B1
9003541 Patidar Apr 2015 B1
9036662 Hughes May 2015 B1
9054876 Yagnik Jun 2015 B1
9092342 Hughes et al. Jul 2015 B2
9106530 Wang Aug 2015 B1
9130991 Hughes Sep 2015 B2
9131510 Wang Sep 2015 B2
9143455 Hughes Sep 2015 B1
9152574 Hughes et al. Oct 2015 B2
9171251 Camp et al. Oct 2015 B2
9191342 Hughes et al. Nov 2015 B2
9202304 Baenziger et al. Dec 2015 B1
9253277 Hughes et al. Feb 2016 B2
9306818 Aumann et al. Apr 2016 B2
9307442 Bachmann et al. Apr 2016 B2
9363248 Hughes Jun 2016 B1
9363309 Hughes Jun 2016 B2
9380094 Florencio et al. Jun 2016 B2
9397951 Hughes Jul 2016 B1
9438538 Hughes et al. Sep 2016 B2
9549048 Hughes Jan 2017 B1
9584403 Hughes et al. Feb 2017 B2
9584414 Sung et al. Feb 2017 B2
9613071 Hughes Apr 2017 B1
9626224 Hughes et al. Apr 2017 B2
9647949 Varki et al. May 2017 B2
9712463 Hughes et al. Jul 2017 B1
9716644 Wei et al. Jul 2017 B2
9717021 Hughes et al. Jul 2017 B2
9875344 Hughes et al. Jan 2018 B1
9906630 Hughes Feb 2018 B2
9948496 Hughes et al. Apr 2018 B1
9961010 Hughes et al. May 2018 B2
9967056 Hughes May 2018 B1
10091172 Hughes Oct 2018 B1
10164861 Hughes et al. Dec 2018 B2
10257082 Hughes Apr 2019 B2
10313930 Hughes et al. Jun 2019 B2
10326551 Hughes Jun 2019 B2
10432484 Hughes et al. Oct 2019 B2
10637721 Hughes et al. Apr 2020 B2
10719588 Hughes et al. Jul 2020 B2
10771370 Hughes et al. Sep 2020 B2
10771394 Hughes Sep 2020 B2
20010026231 Satoh Oct 2001 A1
20010054084 Kosmynin Dec 2001 A1
20020007413 Garcia-Luna-Aceves et al. Jan 2002 A1
20020009079 Jungck et al. Jan 2002 A1
20020010702 Ajtai et al. Jan 2002 A1
20020010765 Border Jan 2002 A1
20020040475 Yap et al. Apr 2002 A1
20020061027 Abiru et al. May 2002 A1
20020065998 Buckland May 2002 A1
20020071436 Border et al. Jun 2002 A1
20020078242 Viswanath Jun 2002 A1
20020101822 Ayyagari et al. Aug 2002 A1
20020107988 Jordan Aug 2002 A1
20020116424 Radermacher et al. Aug 2002 A1
20020129158 Zhang et al. Sep 2002 A1
20020129260 Benfield et al. Sep 2002 A1
20020131434 Vukovic et al. Sep 2002 A1
20020150041 Reinshmidt et al. Oct 2002 A1
20020159454 Delmas Oct 2002 A1
20020163911 Wee et al. Nov 2002 A1
20020169818 Stewart et al. Nov 2002 A1
20020181494 Rhee Dec 2002 A1
20020188871 Noehring et al. Dec 2002 A1
20020194324 Guha Dec 2002 A1
20030002664 Anand Jan 2003 A1
20030009558 Ben-Yehezkel Jan 2003 A1
20030012400 McAuliffe et al. Jan 2003 A1
20030033307 Davis et al. Feb 2003 A1
20030046572 Newman et al. Mar 2003 A1
20030048750 Kobayashi Mar 2003 A1
20030048785 Calvignac et al. Mar 2003 A1
20030067940 Edholm Apr 2003 A1
20030123481 Neale et al. Jul 2003 A1
20030123671 He et al. Jul 2003 A1
20030131079 Neale et al. Jul 2003 A1
20030133568 Stein et al. Jul 2003 A1
20030142658 Ofuji et al. Jul 2003 A1
20030149661 Mitchell et al. Aug 2003 A1
20030149869 Gleichauf Aug 2003 A1
20030204619 Bays Oct 2003 A1
20030214502 Park et al. Nov 2003 A1
20030214954 Oldak et al. Nov 2003 A1
20030233431 Reddy et al. Dec 2003 A1
20040008711 Lahti et al. Jan 2004 A1
20040047308 Kavanagh et al. Mar 2004 A1
20040083299 Dietz et al. Apr 2004 A1
20040085894 Wang et al. May 2004 A1
20040086114 Rarick May 2004 A1
20040088376 McCanne May 2004 A1
20040114569 Naden et al. Jun 2004 A1
20040117571 Chang et al. Jun 2004 A1
20040123139 Aiello et al. Jun 2004 A1
20040158644 Albuquerque et al. Aug 2004 A1
20040179542 Murakami et al. Sep 2004 A1
20040181679 Dellinger et al. Sep 2004 A1
20040199771 Morten et al. Oct 2004 A1
20040202110 Kim Oct 2004 A1
20040203820 Billhartz Oct 2004 A1
20040205332 Bouchard et al. Oct 2004 A1
20040243571 Judd Dec 2004 A1
20040250027 Heflinger Dec 2004 A1
20040255048 Lev Ran et al. Dec 2004 A1
20050010653 McCanne Jan 2005 A1
20050044270 Grove et al. Feb 2005 A1
20050053094 Cain et al. Mar 2005 A1
20050055372 Springer, Jr. et al. Mar 2005 A1
20050055399 Savchuk Mar 2005 A1
20050071453 Ellis et al. Mar 2005 A1
20050091234 Hsu et al. Apr 2005 A1
20050111460 Sahita May 2005 A1
20050131939 Douglis et al. Jun 2005 A1
20050132252 Fifer et al. Jun 2005 A1
20050141425 Foulds Jun 2005 A1
20050171937 Hughes et al. Aug 2005 A1
20050177603 Shavit Aug 2005 A1
20050182849 Chandrayana et al. Aug 2005 A1
20050190694 Ben-Nun et al. Sep 2005 A1
20050207443 Kawamura et al. Sep 2005 A1
20050210151 Abdo et al. Sep 2005 A1
20050220019 Melpignano Oct 2005 A1
20050220097 Swami et al. Oct 2005 A1
20050235119 Sechrest et al. Oct 2005 A1
20050240380 Jones Oct 2005 A1
20050243743 Kimura Nov 2005 A1
20050243835 Sharma et al. Nov 2005 A1
20050256972 Cochran et al. Nov 2005 A1
20050278459 Boucher et al. Dec 2005 A1
20050283355 Itani et al. Dec 2005 A1
20050286526 Sood et al. Dec 2005 A1
20060010243 DuRee Jan 2006 A1
20060013210 Bordogna et al. Jan 2006 A1
20060026425 Douceur et al. Feb 2006 A1
20060031936 Nelson et al. Feb 2006 A1
20060036901 Yang et al. Feb 2006 A1
20060039354 Rao et al. Feb 2006 A1
20060045096 Farmer et al. Mar 2006 A1
20060059171 Borthakur et al. Mar 2006 A1
20060059173 Hirsch et al. Mar 2006 A1
20060109805 Malamal Vadakital May 2006 A1
20060117385 Mester et al. Jun 2006 A1
20060136913 Sameske Jun 2006 A1
20060143497 Zohar et al. Jun 2006 A1
20060193247 Naseh et al. Aug 2006 A1
20060195547 Sundarrajan et al. Aug 2006 A1
20060195840 Sundarrajan et al. Aug 2006 A1
20060212426 Shakara et al. Sep 2006 A1
20060218390 Loughran et al. Sep 2006 A1
20060227717 van den Berg et al. Oct 2006 A1
20060250965 Irwin Nov 2006 A1
20060268932 Singh et al. Nov 2006 A1
20060280205 Cho Dec 2006 A1
20070002804 Xiong et al. Jan 2007 A1
20070008884 Tang Jan 2007 A1
20070011424 Sharma et al. Jan 2007 A1
20070038815 Hughes Feb 2007 A1
20070038816 Hughes et al. Feb 2007 A1
20070038858 Hughes Feb 2007 A1
20070050475 Hughes Mar 2007 A1
20070076693 Krishnaswamy Apr 2007 A1
20070076708 Kolakowski Apr 2007 A1
20070081513 Torsner Apr 2007 A1
20070097874 Hughes et al. May 2007 A1
20070110046 Farrell et al. May 2007 A1
20070115812 Hughes May 2007 A1
20070127372 Khan et al. Jun 2007 A1
20070130114 Li et al. Jun 2007 A1
20070140129 Bauer et al. Jun 2007 A1
20070150497 De La Cruz et al. Jun 2007 A1
20070160200 Ishikawa et al. Jul 2007 A1
20070174428 Lev Ran et al. Jul 2007 A1
20070179900 Daase et al. Aug 2007 A1
20070192863 Kapoor et al. Aug 2007 A1
20070195702 Yuen et al. Aug 2007 A1
20070195789 Yao Aug 2007 A1
20070198523 Hayim Aug 2007 A1
20070226320 Hager et al. Sep 2007 A1
20070237104 Alon et al. Oct 2007 A1
20070244987 Pedersen et al. Oct 2007 A1
20070245079 Bhattacharjee et al. Oct 2007 A1
20070248084 Whitehead Oct 2007 A1
20070258468 Bennett Nov 2007 A1
20070260746 Mirtorabi et al. Nov 2007 A1
20070263554 Finn Nov 2007 A1
20070276983 Zohar et al. Nov 2007 A1
20070280245 Rosberg Dec 2007 A1
20080005156 Edwards et al. Jan 2008 A1
20080013532 Gamer et al. Jan 2008 A1
20080016301 Chen Jan 2008 A1
20080028467 Kommareddy et al. Jan 2008 A1
20080031149 Hughes et al. Feb 2008 A1
20080031240 Hughes et al. Feb 2008 A1
20080037432 Cohen et al. Feb 2008 A1
20080071818 Apanowicz et al. Mar 2008 A1
20080095060 Yao Apr 2008 A1
20080133536 Bjomer et al. Jun 2008 A1
20080133561 Dubnicki et al. Jun 2008 A1
20080184081 Hama et al. Jul 2008 A1
20080205445 Kumar et al. Aug 2008 A1
20080222044 Gottlieb et al. Sep 2008 A1
20080229137 Samuels et al. Sep 2008 A1
20080243992 Jardetzky et al. Oct 2008 A1
20080267217 Colville et al. Oct 2008 A1
20080285463 Oran Nov 2008 A1
20080300887 Chen et al. Dec 2008 A1
20080313318 Vermeulen et al. Dec 2008 A1
20080320151 McCanne et al. Dec 2008 A1
20090006801 Shultz et al. Jan 2009 A1
20090024763 Stepin et al. Jan 2009 A1
20090037448 Thomas Feb 2009 A1
20090060198 Little Mar 2009 A1
20090063696 Wang et al. Mar 2009 A1
20090080460 Kronewitter et al. Mar 2009 A1
20090089048 Pouzin Apr 2009 A1
20090092137 Haigh et al. Apr 2009 A1
20090100483 McDowell Apr 2009 A1
20090158417 Khanna et al. Jun 2009 A1
20090168786 Sarkar Jul 2009 A1
20090175172 Prytz et al. Jul 2009 A1
20090182864 Khan et al. Jul 2009 A1
20090204961 DeHaan et al. Aug 2009 A1
20090234966 Samuels et al. Sep 2009 A1
20090245114 Vijayaraghavan Oct 2009 A1
20090265707 Goodman et al. Oct 2009 A1
20090274294 Itani Nov 2009 A1
20090279550 Romrell et al. Nov 2009 A1
20090281984 Black Nov 2009 A1
20100005222 Brant et al. Jan 2010 A1
20100011125 Yang et al. Jan 2010 A1
20100020693 Thakur Jan 2010 A1
20100054142 Moiso et al. Mar 2010 A1
20100070605 Hughes et al. Mar 2010 A1
20100077251 Liu et al. Mar 2010 A1
20100082545 Bhattacharjee et al. Apr 2010 A1
20100085964 Weir et al. Apr 2010 A1
20100115137 Kim et al. May 2010 A1
20100121957 Roy et al. May 2010 A1
20100124239 Hughes May 2010 A1
20100131957 Kami May 2010 A1
20100150158 Cathey et al. Jun 2010 A1
20100169467 Shukla et al. Jul 2010 A1
20100177663 Johansson et al. Jul 2010 A1
20100225658 Coleman Sep 2010 A1
20100232443 Pandey Sep 2010 A1
20100242106 Harris et al. Sep 2010 A1
20100246584 Ferguson et al. Sep 2010 A1
20100290364 Black Nov 2010 A1
20100318892 Teevan et al. Dec 2010 A1
20100333212 Carpenter et al. Dec 2010 A1
20110002346 Wu Jan 2011 A1
20110022812 Van Der Linden et al. Jan 2011 A1
20110113472 Fung et al. May 2011 A1
20110131411 Lin et al. Jun 2011 A1
20110154169 Gopal et al. Jun 2011 A1
20110154329 Arcese et al. Jun 2011 A1
20110181448 Koratagere Jul 2011 A1
20110219181 Hughes et al. Sep 2011 A1
20110225322 Demidov et al. Sep 2011 A1
20110258049 Ramer et al. Oct 2011 A1
20110261828 Smith Oct 2011 A1
20110276963 Wu et al. Nov 2011 A1
20110299537 Saraiya et al. Dec 2011 A1
20120036325 Mashtizadeh et al. Feb 2012 A1
20120069131 Abelow Mar 2012 A1
20120147894 Mulligan et al. Jun 2012 A1
20120173759 Agarwal et al. Jul 2012 A1
20120185775 Clemm et al. Jul 2012 A1
20120198346 Clemm et al. Aug 2012 A1
20120218130 Boettcher et al. Aug 2012 A1
20120221611 Watanabe et al. Aug 2012 A1
20120230345 Dvsiannikov Sep 2012 A1
20120239872 Hughes et al. Sep 2012 A1
20120290636 Kadous et al. Nov 2012 A1
20130018722 Libby Jan 2013 A1
20130018765 Fork et al. Jan 2013 A1
20130031642 Dwivedi et al. Jan 2013 A1
20130044751 Casado et al. Feb 2013 A1
20130058354 Casado et al. Mar 2013 A1
20130080619 Assuncao et al. Mar 2013 A1
20130083806 Suarez Fuentes et al. Apr 2013 A1
20130086236 Baucke et al. Apr 2013 A1
20130086594 Cottrell Apr 2013 A1
20130094501 Hughes Apr 2013 A1
20130103655 Fanghaenel et al. Apr 2013 A1
20130117494 Hughes et al. May 2013 A1
20130121209 Padmanabhan et al. May 2013 A1
20130141259 Hazarika et al. Jun 2013 A1
20130142050 Luna Jun 2013 A1
20130163594 Sharma et al. Jun 2013 A1
20130250951 Koganti Sep 2013 A1
20130263125 Shamsee et al. Oct 2013 A1
20130266007 Kumbhare et al. Oct 2013 A1
20130282970 Hughes et al. Oct 2013 A1
20130325986 Brady et al. Dec 2013 A1
20130343191 Kim et al. Dec 2013 A1
20140052864 van Der Linden et al. Feb 2014 A1
20140075554 Cooley Mar 2014 A1
20140086069 Frey et al. Mar 2014 A1
20140101426 Senthurpandi Apr 2014 A1
20140108360 Kunath et al. Apr 2014 A1
20140114742 Lamontagne et al. Apr 2014 A1
20140123213 Vank et al. May 2014 A1
20140181381 Hughes et al. Jun 2014 A1
20140269705 DeCusatis et al. Sep 2014 A1
20140279078 Nukala et al. Sep 2014 A1
20140321290 Jin et al. Oct 2014 A1
20140379937 Hughes et al. Dec 2014 A1
20150058488 Backholm Feb 2015 A1
20150074291 Hughes Mar 2015 A1
20150074361 Hughes et al. Mar 2015 A1
20150078397 Hughes et al. Mar 2015 A1
20150110113 Levy et al. Apr 2015 A1
20150120663 Le Scouamec et al. Apr 2015 A1
20150127701 Chu et al. May 2015 A1
20150143505 Border et al. May 2015 A1
20150170221 Shah Jun 2015 A1
20150281099 Banavalikar Oct 2015 A1
20150281391 Hughes et al. Oct 2015 A1
20150312054 Barabash et al. Oct 2015 A1
20150334210 Hughes Nov 2015 A1
20150365293 Madrigal et al. Dec 2015 A1
20160014051 Hughes et al. Jan 2016 A1
20160034305 Shear et al. Feb 2016 A1
20160093193 Silvers et al. Mar 2016 A1
20160112255 Li Apr 2016 A1
20160142310 Means May 2016 A1
20160218947 Hughes et al. Jul 2016 A1
20160255000 Gattani et al. Sep 2016 A1
20160255542 Hughes et al. Sep 2016 A1
20160359740 Parandehgheibi et al. Dec 2016 A1
20160380886 Blair et al. Dec 2016 A1
20170026467 Barsness et al. Jan 2017 A1
20170111692 An et al. Apr 2017 A1
20170149679 Hughes et al. May 2017 A1
20170187581 Hughes et al. Jun 2017 A1
20170359238 Hughes et al. Dec 2017 A1
20180089994 Dhondse et al. Mar 2018 A1
20180121634 Hughes et al. May 2018 A1
20180123861 Hughes et al. May 2018 A1
20180131711 Chen et al. May 2018 A1
20180205494 Hughes Jul 2018 A1
20180227216 Hughes Aug 2018 A1
20180227223 Hughes Aug 2018 A1
20190089620 Hefel et al. Mar 2019 A1
20190104207 Goel et al. Apr 2019 A1
20190149447 Hughes et al. May 2019 A1
20190230038 Hughes Jul 2019 A1
20190245771 Wu et al. Aug 2019 A1
20190260683 Hughes Aug 2019 A1
20190274070 Hughes et al. Sep 2019 A1
20190280917 Hughes et al. Sep 2019 A1
20200021506 Hughes et al. Jan 2020 A1
20200213185 Hughes et al. Jul 2020 A1
20200279029 Hughes et al. Sep 2020 A1
Foreign Referenced Citations (3)
Number Date Country
1507353 Feb 2005 EP
H05061964 Mar 1993 JP
WO0135226 May 2001 WO
Non-Patent Literature Citations (24)
Entry
“IPsec Anti-Replay Window: Expanding and Disabling,” Cisco IOS Security Configuration Guide. 2005-2006 Cisco Systems, Inc. Last updated: Sep. 12, 2006, 14 pages.
Singh et al. ; “Future of Internet Security—IPSEC”; 2005; pp. 1-8.
Muthitacharoen, Athicha et al., “A Low-bandwidth Network File System,” 2001, in Proc. of the 18th ACM Symposium on Operating Systems Principles, Banff, Canada, pp. 174-187.
“Shared LAN Cache Datasheet”, 1996, <http://www.lancache.com/slcdata.htm>, 8 pages.
Spring et al., “A protocol-independent technique for eliminating redundant network traffic”, ACM SIGCOMM computer Communication Review, vol. 30, Issue 4 (Oct. 2000) pp. 87-95, Year of Publication: 2000.
Hong, B et al. “Duplicate data elimination in a SAN file system”, In Proceedings of the 21st Symposium on Mass Storage Systems (MSS '04), Goddard, MD, Apr. 2004. IEEE, pp. 101-114.
You, L. L. and Karamanolis, C. 2004. “Evaluation of efficient archival storage techniques”, In Proceedings of the 21st IEEE Symposium on Mass Storage Systems and Technologies (MSST), pp. 1-6.
Douglis, F. et al., “Application specific Delta-encoding via Resemblance Detection”, Published in the 2003 USENIX Annual Technical Conference, pp. 1-14.
You, L. L. et al., “Deep Store an Archival Storage System Architecture” Data Engineering, 2005. ICDE 2005. Proceedings of the 21st Intl. Conf. on Data Eng.,Tokyo, Japan, Apr. 5-8, 2005, pp. 12.
Manber, Udi, “Finding Similar Files in a Large File System”, TR 93-33 Oct. 1994, Department of Computer Science, University of Arizona. <http://webglimpse.net/pubs/TR93-33.pdf>. Also appears in the 1994 winter USENIX Technical Conference, pp. 1-10.
Knutsson, Bjorn et al., “Transparent Proxy Signalling”, Journal of Communications and Networks, vol. 3, No. 2, Jun. 2001, pp. 164-174.
Definition memory (n), Webster's Third New International Dictionary, Unabridged (1993), available at <http://lionreference.chadwyck.com> (Dictionaries/Webster's Dictionary). Copy not provided in IPR2013-00402 proceedings.
Definition appliance, 2c, Webster's Third New International Dictionary, Unabridged (1993), available at <http://lionreference.chadwyck.com> (Dictionaries/Websters Dictionary). Copy not provided in IPR2013-00402 proceedings.
Newton, “Newton's Telecom Dictionary”, 17th Ed., 2001, pp. 38, 201, and 714.
Silver Peak Systems, “The Benefits of Byte-level WAN Deduplication” (2008), pp. 1-4.
Business Wire, “Silver Peak Systems Delivers Family of Appliances for Enterprise-Wide Centralization of Branch Office Infrastructure; Innovative Local Instance Networking Approach Overcomes Traditional Application Acceleration Pitfalls” (available at http://www.businesswire.com/news/home/20050919005450/en/Silver-Peak-Systems-Delivers-Family-Appliances-Enterprise-Wide#.UVzkPk7u-1 (last visited Aug. 8, 2014)), 4 pages.
Riverbed, “Riverbed Introduces Market-Leading WDS Solutions for Disaster Recovery and Business Application Acceleration” (available at http://www.riverbed.com/about/news-articles/pressreleases/riverbed-introduces-market-leading-wds-solutions-fordisaster-recovery-and-business-application-acceleration.html (last visited Aug. 8, 2014)), 4 pages.
Tseng, Josh, “When accelerating secure traffic is not secure” (available at http://www.riverbed.com/blogs/whenaccelerati.html?&isSearch=true&pageSize=38&page=2 (last visited Aug. 8, 2014)), 3 pages.
Riverbed, “The Riverbed Optimization System (RiOS) v4.0: A Technical Overview” (explaining “Data Security” through segmentation) (available at http://mediacms.riverbed.com/documents/TechOverview-Riverbed-RiOS_4_0.pdf (last visited Aug. 8, 2014)), pp. 1-18.
Riverbed, “Riverbed Awarded Patent on Core WDS Technology” (available at: http://www.riverbed.com/about/news-articles/pressreleases/riverbed-awarded-patent-on-core-wds-technology.html (last visited Aug. 8, 2014)), 2 pages.
Final Written Decision, dated Dec. 30, 2014, Inter Partes Review Case No. IPR2013-00403, pp. 1-38.
Final Written Decision, dated Dec. 30, 2014, Inter Partes Review Case No. IPR2013-00402, pp. 1-37.
“Notice of Entry of Judgement Accompanied by Opinion”, United States Court of Appeals for the Federal Circuit, Case: 15-2072, Oct. 24, 2017, 6 pages.
“Decision Granting Motion to Terminate”, Inter Partes Review Case No. IPR2014-00245, Feb. 7, 2018, 4 pages.
Related Publications (1)
Number Date Country
20190253187 A1 Aug 2019 US
Divisions (1)
Number Date Country
Parent 15241992 Aug 2016 US
Child 15918807 US
Continuations (1)
Number Date Country
Parent 15918807 Mar 2018 US
Child 16396467 US