Embodiments described herein relate generally to communication networks, and particularly to methods and apparatus for flow-based management of shared buffer resources.
A network element typically stores incoming packets for processing and forwarding. Storing the packets in a shared buffer enables to share storage resources efficiently. Methods for managing shared buffer resources are known in the art. For example, U.S. Pat. No. 10,250,530 describes a communication apparatus that includes multiple interfaces configured to be connected to a packet data network for receiving and forwarding of data packets of multiple types. A memory is coupled to the interfaces and configured as a buffer to contain packets received through the ingress interfaces while awaiting transmission to the network via the egress interfaces. Packet processing logic is configured to maintain multiple transmit queues, which are associated with respective ones of the egress interfaces, and to place both first and second queue entries, corresponding to first and second data packets of the first and second types, respectively, in a common transmit queue for transmission through a given egress interface, while allocating respective spaces in the buffer to store the first and second data packets against separate, first and second buffer allocations, which are respectively assigned to the first and second types of the data packets.
An embodiment that is described herein provides an apparatus for controlling a Shared Buffer (SB), the apparatus including an interface and a SB controller. The interface is configured to access flow-based data counts and admission states. The SB controller is configured to perform flow-based accounting of packets received by a network device coupled to a communication network, for producing flow-based data counts, each flow-based data count associated with one or more respective flows, and to generate admission states based at least on the flow-based data counts, each admission state being generated from one or more respective flow-based data counts.
In an embodiment, the SB is included in a memory accessible to the SB controller, the memory being external to the apparatus. In another embodiment, the apparatus further includes a memory, and the SB is included in the memory. In yet another embodiment, the apparatus further includes multiple ports including an ingress port, configured to connect to the communication network, and data-plane logic, configured to receive a packet from the ingress port, classify the packet into a respective flow; and, based on one or more admission states that were generated based on the flow-based data counts, decide whether to admit the packet into the SB or drop the packet.
In some embodiments, the SB controller is configured to produce an aggregated data count for packets belonging to multiple different flows, and to generate an admission state for the packets of the multiple different flows based on the aggregated data count. In other embodiments, the SB controller is configured to produce first and second flow-based data counts for packets belonging to respective first and second different flows, and to generate an admission state for the packets of the first and second flows based on both the first and the second flow-based data counts. In yet other embodiments, the SB controller is configured to generate multiple admission states based on multiple selected flows, and the data-plane logic is configured to decide whether to admit a packet belonging to one of the selected flows into the SB or drop the packet, based on the multiple admission states.
In an embodiment, the data-plane logic is configured to determine for received packets respective egress ports among the multiple ports, ingress priorities and egress priorities, and the SB controller is configured to perform occupancy accounting for (i) Rx data counts associated with respective ingress ports and ingress priorities, and (ii) Tx data counts associated with respective egress ports and egress priorities, and to generate the admission states based on the flow-based data counts and on at least one of the Rx data counts and the Tx data counts. In another embodiment, the SB controller is configured to perform the flow-based accounting and the occupancy accounting in parallel. In yet another embodiment, the SB controller is configured to identify for a received packet a corresponding flow-based data count by (i) applying a hash function to one or more fields in a header of the received packet, or (ii) processing the packet using an Access Control List (ACL).
In some embodiments, the SB controller is configured to identify for a received packet a corresponding flow-based data count based on flow-based binding used in a protocol selected from a list of protocols including: a tenant protocol, a bridging protocol, a routing protocol and a tunneling protocol. In other embodiments, the SB controller is configured to locally monitor selected flow-based data counts, to evaluate performance level of the network element based on the monitored flow-bases data counts, and based on a reporting criterion, to report information indicative of the performance level. In yet other embodiments, the SB controller is configured to calculate a drop probability based at least on a flow-based data count associated with one or more selected flows, and to generate an admission state for the one or more flows based on the flow-based data count and on the drop probability.
There is additionally provided, in accordance with an embodiment that is described herein, a method for controlling a Shared Buffer (SB), the method including, in an apparatus that includes a SB controller, accessing flow-based data counts and admission states. Flow-based accounting of packets received by a network device coupled to a communication network are performed for producing flow-based data counts, each flow-based data count associated with one or more respective flows. Admission states are generated based at least on the flow-based data counts, each admission state being generated from one or more respective flow-based data counts.
These and other embodiments will be more fully understood from the following detailed description of the embodiments thereof, taken together with the drawings in which:
Embodiments that are described herein provide methods and systems for flow-based management of shared buffer resources.
A shared buffer in a network element stores incoming packets that typically belong to multiple flows. The stored packets are processed and await transmission to their appropriate destinations.
The storage space of the shared buffer is used for storing packets received via multiple ingress ports and destined to be delivered via multiple egress ports. In some embodiments, a shared buffer controller manages the shared buffer for achieving fair allocation of the storage space among ports.
In some embodiments, the shared buffer controller manages the shared buffer resources by allocating limited amounts of storage space to entities referred to herein as “regions.” A region may be assigned to a pair comprising an ingress port and a reception priority, or to a pair comprising an egress port and a transmission priority. For each region, the shared buffer stores data up to a respective threshold that is adapted dynamically.
The shared buffer performs accounting of the amount of data currently buffered per each region and decides to admit a received packet into the shared buffer or to drop the packet, based on the accounting. In this scheme, the decision of packet admission is related to ingress/egress ports and to reception/transmission priorities but does not take into consideration the flows to which the packets traversing the network element belong.
In the disclosed embodiments, for enhancing the flexibility in managing the shared buffer storage space, a new type of a region is specified, which is referred to herein as a “flow-based” region. A flow-based region corresponds to a specific flow but is independent of any ports and the priorities assigned to ports. Using flow-based regions provides a flow-based view of the shared buffer usage, and therefore can be used for prioritizing different data flows in sharing the storage space. Moreover, complex admission schemes that combine several flow-based regions or combine a flow-based region with a port/priority region can also be used.
Consider a network element comprising multiple ports, a memory configured as a Shared Buffer (SB), SB controller and data-plane logic. The multiple ports are configured to connect to a communication network. The Shared Buffer (SB) is configured to store packets received from the communication network. The SB controller is configured to perform flow-based accounting of packets received by the network element for producing flow-based data counts, each flow-based data count associated with one or more respective flows, and to generate admission states based at least on the flow-based data counts, each admission state is generated from one or more respective flow-based. data counts. The data-plane logic is configured to receive a packet from an ingress port, to classify the packet into a respective flow, and based on one or more admission states that were generated based on the flow-based data counts, to decide whether to admit the packet into the SB or drop the packet.
The SB controller may manage the data counts and admission states in various ways. In an embodiment, the SB controller produces an aggregated data count for packets belonging to multiple different flows, and generates an admission state for the packets of the multiple different flows based on the aggregated data count. In another embodiment, the SB controller produces first and second flow-based data counts for packets belonging to respective first and second different flows, generates an admission state for the packets of the first and second flows based on both the first and the second flow-based data counts. In yet another embodiments, the SB controller generates multiple admission states based on multiple selected flows, and the data-plane logic decides whether to admit a packet belonging to one of the selected flows into the SB or drop the packet, based on the multiple admission states.
In processing the received packets, the data-plane logic determines for the received packets respective egress ports, ingress priorities and egress priorities. The SB controller performs occupancy accounting for (i) Rx data counts associated with respective ingress ports and ingress priorities, and (ii) Tx data counts associated with respective egress ports and egress priorities. The controller generates the admission states based on the flow-based data counts and on at least one of the Rx data counts and the Tx data counts. Note that the SB controller performs the flow-based accounting and the occupancy accounting in parallel.
The SB controller may link a received packet to a flow-based data count in various ways. In some embodiments, the SB controller identifies for a received packet a corresponding flow-based data count by (i) applying a hash function to one or more fields in a header of the received packet, or (ii) processing the packet using an Access Control List (ACL). In other embodiments, the SB controller identifies for a received packet a corresponding flow-based data count based on flow-based binding used. in a protocol, such as, for example, a tenant protocol, a bridging protocol, a routing protocol or a tunneling protocol.
The flow-based accounting that is used for managing the SB resources may be used for other purposes such as flow-based mirroring and flow-based congestion avoidance, as will be described further below.
In the disclosed techniques a SB controller performs flow-based accounting for selected flows. This allows sharing storage space based on individual flow prioritization. This flow-based view enables fair sharing of storage space among competing flows, regardless of the ports via which the flows arrive at the network element. Moreover, flexible admission schemes that combine flow-based data counts and occupancy data counts are also possible.
In the description that follows and in the claims, the term “network element” refers to any device in a packet network that communicates packets with other devices in the network, and/or with network nodes coupled to the network. A network element may comprise, for example, a switch, a router, or a network adapter.
Network element 20 comprises interfaces in the form of ingress ports 22 and egress ports 24 for connecting to a communication network 26. Network element 20 receives packets from the communication network via ingress ports 22 and transmits forwarded packets via egress ports 24. Although in
Communication network 26 may comprise any suitable packet network operating using any suitable communication protocols. For example, communication network 26 may comprise an Ethernet network, an IP network or an InfiniBand™ network.
Each ingress port 22 is associated with respective control logic 30 that processes incoming packets as will be described below. Although in
Upon receiving an incoming packet via an ingress port 22, the ingress port places the packet in shared buffer 34 and notifies relevant control logic 30 that the packet is ready for processing. A parser 44 parses the packet header(s) and generates for the packet a descriptor, which the parser passes to a descriptor processor 46 for further handling and generation of forwarding instructions. Based on the descriptor, descriptor processor 46 typically determines an egress port 24 through which the packet is to be transmitted. The descriptor may also indicate the quality of service (QoS) to be applied to the packet, i.e., the level of priority at reception and for transmission, and any applicable instructions for modification of the packet header. An admission decision module 48 decides on whether to drop or admit the packet. The admission decision module determines the admission decision based on admission states 62, as will be described in detail bellow.
Descriptor processor 46 places the descriptors of admitted packets in the appropriate queues in a queueing system 50 to await transmission via the designated egress ports 24. Typically, queuing system 50 contains a dedicated queue for each egress port 24 or multiple queues per egress port, one for each QoS level (e.g., transmission priority).
Descriptor processor 46 passes the descriptors of admitted packets to queueing system 50 and to a buffer (SB) controller 54, which serves as the central buffer management and accounting module for shared buffer 34. SB controller 54 performs two types of accounting, referred to herein as “occupancy accounting” and “flow-based accounting.” For the occupancy accounting, the SB controller manages “occupancy data counts” 56, whereas for the flow-based accounting, the SB controller manages “flow-based data counts” 58. SB controller 54 receives consumption information in response to control logic 30 deciding to admit a packet, and receives release , information in response to transmitting a queued packet. SB controller 54 increments or decrements the occupancy data counts and the flow-based data counts, based on the consumption and release information.
The SB controller may manage the occupancy data counts and the flow-based data counts using any suitable count units, such as numbers of bytes or packets. Based on flow-based data counts 58 and possibly on occupancy data counts 56, SB controller produces admission states 62 to be used by admission decision modules 48 for deciding on admission/drop for each received packet.
In some embodiments, SB controller 54 that manages flow-based data counts as well as occupancy data counts in association with entities that referred to herein as “regions.” An occupancy region comprises a pair of an ingress port and Rx priority or a pair of an egress port and a Tx priority. A flow-based region comprises a flow. The SB controller may determine admission states 62 based on pools 66, wherein each pool is associated with multiple regions or with their corresponding data counts. For example, a pool comprises one or more flow-based data counts, and possibly one or more Rx occupancy data counts and/or one or more Tx occupancy data counts.
In some embodiments, SB controller 54 comprises an interface 64, via which the SB controller accesses occupancy data counts 56, flow-based data counts 58, and admission states 62. In an embodiment, interface 64 serves also for accessing consumption and release information by the SB controller.
When a descriptor of a packet queued. in queueing system 50 reaches the head of its queue, queuing system 50 passes the descriptor to a packet transmitter 52 for execution. Packet transmitters 52 are respectively coupled to egress ports 24 and serve as packet transmission modules. In response to the descriptor, packet transmitter 52 reads the packet data from shared buffer 34, and (optionally) makes whatever changes are called for in the packet header for transmission to communication network 26 through egress port 24.
Upon the transmission of the packet through the corresponding egress port 24, packet transmitter 52 signals SB controller 54 that the packet has been transmitted, and in response, SB controller 54 releases the packet from SB 34, so that the packet location in SB 34 can be overwritten. This memory accounting and management process typically takes place for multiple different packets in parallel at any given time.
The configuration of network element 24 in
Some elements of network element 20, such as control logic 30 and SB controller 54 may be implemented in hardware, e.g., in one or more Application-Specific Integrated Circuits (ASICs) or Field-Programmable Gate Arrays (FPGAs). Additionally or alternatively, some elements of the network element can be implemented using software, or using a combination of hardware and software elements.
Elements that are not necessary for understanding the principles of the present application, such as various interfaces, addressing circuits, timing and sequencing circuits and debugging circuits, have been omitted from
Memory 34 may comprise any suitable storage device using any suitable storage technology, such as, for example, a Random Access Memory (RAM) . The SB may be implemented in an on-chip internal RAM or in an off-chip external RAM.
In some embodiments, the SB controller is comprised in any suitable apparatus such as a network element or a Network Interface Controller (NIC). In some embodiments, the SB is comprised in a memory accessible to the SB controller, the memory being external to the apparatus. In other embodiments, the apparatus further comprises a memory, and the SB is comprised in the memory.
In some embodiments, some of the functions of SB controller 54 may be carried out by a general-purpose processor, which is programmed in software to carry out the functions described herein. The software may be downloaded to the processor in electronic form, over a network, for example, or it may, alternatively or additionally, be provided and/or stored on non-transitory tangible media, such as magnetic, optical, or electronic memory.
In the description that follows and in the claims, elements involved in real-time packet processing and forwarding for transmission are collectively referred to as “data-plane logic.” In the example of
In some embodiments, SB controller 54 manages SB 34 for achieving a fair usage of the shared buffer. To this end, regions corresponding (PI,Rp) and (PO,Tp) are allocated respective storage spaces in the shared buffer. In the regions above, PI and PO denote respective ingress and egress ports, and Rp and Tp denote respective reception and transmission priorities. The allocated storage spaces are bounded to respective dynamic thresholds. The SB controller holds the amount of data consumed at any given time by regions (PI,Rp) and (PO,Tp) in respective occupancy data counts 56.
In some disclosed embodiments, the SB controller manages the SB resources using a flow based approach. In these embodiments, SB manages flow-based regions associated with flow-based data counts 58. Each flow-based region virtually consumes a storage space of the shared buffer bounded to a dynamic threshold. A flow-based view of SB storage consumption can be used for prioritizing SB storage among different data flows.
Admission states 62 are indicative of the amount of data consumed relative to corresponding dynamic thresholds. An admission state may have a binary value that indicates whether a data count exceeds a relevant dynamic threshold, in which case the packet should be dropped. Alternatively, an admission state may have multiple discrete values or a contiguous range, e.g., an occupancy percentage of the bounded storage space.
A packet tested by admission decision module 48 for admission may be linked to one or more regions (or corresponding data counts). For example, the packet may be linked to an occupancy data count of a region (P1, Rp), to an occupancy data count of a region (PO, Tp), and/or to a flow-based data count of a flow-based region. In general, a packet may be linked to at least one of the data count types (i) flow-based data count (ii) Rx occupancy data count, and (iii) Tx occupancy data count. Each data count type may be associated with a pool 66, depending on the SB configuration. A packet linked to a pool of multiple data counts is also associated with one or more admission states that SB controller 54 determines based or the multiple data counts.
A packet may be linked or bound to a certain data count or to a pool of multiple data counts in various ways, as described herein. In some embodiments, SB controller 54 identifies a data count (or a pool) corresponding to a received packet, e.g., a flow-based data count, by applying a hash function to one or more fields in a header of the received packet resulting in an identifier of the pool. In another embodiment, the SB controller identifies a data count (or a pool) corresponding to a received packet by processing the received packet using an Access Control List (ACI) that extracts the pool identifier.
In some embodiments, the SB controller identifies for a received packet corresponding data counts (e.g., flow-based data count) based on flow-based binding used in a protocol selected from a list of protocols comprising: a tenant protocol, a bridging protocol, a routing protocol and a tunneling protocol. In these embodiments, the flow to which the packet belongs represents the selected protocol.
Decision module 48 may decide on packet admission or drop, based on multiple admission states, in various ways. For example, when using binary admission states, decision module 48 may decide to admit a packet only when all the relevant admission states are indicative of packet admission. Alternatively, SB controller 54 may decide on packet admission when only part of the relevant admission states are indicative of packet admission, e.g., based on a majority vote criterion.
In some embodiments, the values of the admission states comprise a contiguous range, and the decision module decides on packet admission by calculating a predefined function over some or all of the relevant admission states. For example, the SB controller calculates an average data count based on two or score selected data counts, and determines the admission state by comparing the average data count to the dynamic threshold.
In general, accounting and generating admission states are tasks related to control-plane processing, whereas admission decision is a task related to the data-plane processing. The flow-based admission configurations will be described as executed by network element 20 of
In performing accounting, SB controller 54 receives consumption information indicative of admitted packets, and release information indicative of transmitted packets. SB controller 54 performs flow-based accounting to the FL1 packets to produce a flow-based data count denoted FB_DC1. In some embodiments, based on the consumption and release information, SB controller 54 performs occupancy-based accounting to produce occupancy data counts 112, depending on ingress ports, egress ports and Rx/Tx priorities determined from packets' headers. This accounting is part of the control-plane tasks.
SB controller 54 produces for the packets of FL1, based on FB_DC1, an admission state 116, denoted AS1. In the example of
In deciding on packet admission, admission decision module 48 produces respective admission decisions 124 for the packets of flow FL1. The admission decisions may be based, for example, on the flow-based admission state ASI alone, or on one or more of occupancy admission states 120 in addition to AS1.
In some embodiments, SB controller 54 comprises a visibility engine 128 that monitors flow-based data counts such as FB_DC1. Visibility engine 128 generates a visibility indication based on the behavior of FB_DC1. For example, the visibility indication may be indicative of a short-time change in the value of the flow-based data count. In some embodiments, admission decision module 48 may produce admission decisions 124 based also on the visibility indication. In some embodiments, visibility engine 128 produces a visibility indication that is used for flow-based mirroring, as will be described below.
Control logic 30 passes descriptors of packets belonging to FM for which the admission decision is positive to queueing system 50, for transmission to the communication network, using packet TX 52, via an egress port 24. Control logic 30 reports packets of FL1 that have been dropped to the SB controller, which releases the dropped packets from SB 34.
In the present example, in performing accounting, SB controller 54 performs aggregated flow-based accounting for the packets of both FL2 and FL3 to produce a common flow-based data count 136 denoted FB_DC2. The flow-based data count FB_PC2 is indicative of the amount of data currently buffered in the network element from both FL2 and FL3.
SB controller 54 produces for the packets of FL2 and FL3, based on FB_DC2, an admission state 138, denoted AS2. In the example of
In deciding on admission, admission decision modules 48 in control logic modules 30 that process packets of FL2 and FL3, produce admission decisions 142 for the packets of both FL2 and FL3. The admission decisions may be based, for example, on flow-based admission state AS2 alone, or on AS2 and on one or more of occupancy admission states 140.
In some embodiments, a visibility engine 144 (similar to visibility engine 128 above) monitors FB_DC2 and outputs a visibility indication based on FB_DC2. Admission decision module 48 may use the visibility indication in producing admission decisions 142.
Control logic modules 30 that process packets of FL2 and FL3, pass descriptors of packets belonging to these flows that have been admitted to queueing system 50, for transmission using packet Tx 52 via a common egress port 24 or via two respective egress ports. Control logic modules 30 that process packets of FL2 and FL3, report packets of FL2 and FL3 that have been dropped to the SB controller, which releases the dropped packets from SB 34.
In the present example, in performing accounting, SB controller 54 performs separate flow-based accounting to packets of FL4, FL5 and FL6, to produce respective flow-based data counts 160 denoted FB_DC3, FB_DC4 and FB_DC5.
In the present example, SB controller 54 produces, based on, data counts FB_DC3, FB_DC4 and FB_DC5 two admission states 162 denoted AS3 and AS4. Specifically, SB controller 54 produces AS3 based on data counts FB_DC3 and FB_DC4 corresponding to FL4 and FL5, and produces, AS4 based on a single data count FB_DC5 corresponding to FL6. In some embodiments, SB controller 54 also produces, based on the occupancy data counts, occupancy admission states 170 (similarly to admission states 116 of
In deciding on admission, admission decision modules 48 of control logic modules 30 that process packets of FL4, FL5 and FL6 produce admission decisions 174 for the packets of flows FL4, FL5 and FL6, based at least on one of flow-based admission states AS3 and AS4. In an embodiment, the admission decision is also based, on one or more of occupancy admission states 170.
In some embodiments, the admission decisions may be additionally based on one or more visibility indications 178 produced by monitoring one or more of flow-based data counts FB_DC3, FB_DC4 and FB_DC5 using visibility engine(s) (similar to visibility engines 128 and 144—not shown).
Control logic modules 30 that process packets of FL4, FL5 and FL6, pass descriptors of packets belonging to FL4, FL5 and FL6 that have been admitted to queueing system 50 for transmission by packet Tx 52 via a common egress port 24 or via two or three egress ports. Control logic modules 30 that process packets of FL4, FL5 and FL6, report packets of FL4, FL5 and FL6 that have been dropped to the SB controller, which releases the dropped packets from SB 34.
The method will be described as executed by network element 20 of
The method of
At a packet analysis step 204, parser 44 parses the packet header(s) to generate a descriptor for the packet. Parser 44 passes the descriptor to descriptor processor 46, which based on the descriptor determines the following parameters:
At an admission states accessing step 208, admission decision module 48 reads one or more admission states associated with (PI,Rp), (PO,Tp) and FL. As noted above, admission states associated with (PI,Rp) and with (PO, Tp) are produced by SB controller 54 based on occupancy data counts 56, and admission states associated with FL are produced by SB controller 54 based on flow-based data counts 58.
At a decision step 212, admission decision module 48 decides, based on the one or more admission states observed at step 108, whether to admit or drop the packet.
At an admission query step 216, descriptor processor 46 checks whether the packet should be admitted. When the decision at step 216 is to drop the packet, the method loops back to step 100 to receive another packet. Descriptor processor 46 also reports the dropped packet to the SB controller for releasing storage space occupied by the dropped packet. When the decision at step 216 is to admit the packet, descriptor processor 46 proceed to a queueing step 220. At step 220, the descriptor processor places the corresponding descriptor in an appropriate queue in queueing system 50 to await transmission via the designated egress ports PO at the transmission priority Tp. At a consumption reporting step 224, descriptor processor 46 reports consumption information related to the admitted packet to SB controller 54 for accounting. Following step 224, the method loops back to step 100 to receive a subsequent packet.
At a release reporting step 228, upon transmission of the queued packet via the port PO, packet Tx 52 reports the release event to SB controller 54, for accounting update and refreshing relevant admission states.
The method will be described as executed by SB controller 54 of
The method of
In response to receiving a consumption notification corresponding to a given packet, SB controller 54 increases a flow-based data count associated with a flow FL to which the given packet belongs. The SB controller also increases occupancy data counts associated with regions (PI, Rp), (PO, Tp) of the given packet. Let DC denote the amount of data corresponding to the given packet. At step 254, the SB controller calculates updated data counts as follows: Count (FL)+=DC, Count(PI,Rp)+−DC, and Count(PO,Tp)+=DC.
In response to receiving a release notification corresponding to a given packet, SB controller 54 decreases a flow-based data count associated with a flow FL to which the given packet belongs. The SB controller also decreases occupancy data counts associated with regions (PI, Rp), (PO, Tp) of the given packet. Let DC denote the amount of data corresponding to the given packet. At step 258, the SB controller calculates updated counts as follows: Count (FL)−=DC, Count(PI,Rp)−=DC, and Count(PO,Tp)−=DC.
Following each of steps 254 and 258, the method. proceeds to an admission states refreshing step 262, at which SB controller 54 updates admission states 62 associated with FL, (PI,Rp) and (PO, Tp) to reflect the effect of the consumption or release events. Following step 262, the method loops back to step to wait for a subsequent notification.
Mirroring is a technique used, for example, by network elements for reporting selected events, e.g., for the purpose of troubleshooting and performance evaluation. In mirroring, packets selected using a predefined criterion (e.g., congestion detection) may be reported to a central entity for analysis. The selected packets are duplicated and transmitted to the network, and therefore may undesirably consume a significant share of the available bandwidth.
In some embodiments, a mirroring criterion comprises a flow-based criterion. For example, packets belonging to a certain flow (FL) may be mirrored based on a flow-based count assigned to FL, e.g., using visibility engine 128 or 144. In some embodiments, packets of FL may be mirrored based on flow-based data counts of other flows. Additionally, packets belonging to FL may be mirrored based on one or more occupancy data counts that are associated with FL. In some embodiments, a flow-based mirroring criterion may be combined with another mirroring criterion such as identifying a congestion condition.
Weighted Random Early Detection (WRED) is a method that may be used for congestion avoidance. In WRED, the probability of dropping packets increases as the transmission queue builds up.
In some embodiments, admission decision module 48 comprises a flow-based WRED module (not shown) that participates in deciding on packet admission or drop. Specifically, SB controller 54 calculates a drop probability based at least on a flow-based data count associated with one or more selected flows, and generates a flow-based admission state for the one or more flows based on the flow-based data count and on the drop probability. In some embodiments, the SB controller determines the admission state also based on one or more occupancy data counts.
The embodiments described above are given by way of example, and other suitable embodiments can also be used. For example, in the embodiments described above, the flow-based accounting is carried out relative to ingress ports. In alternative embodiments, however, the flow-based accounting is carried out relative to egress ports.
Although the embodiments described herein mainly address flow-based management of a SB in a network element, the methods and systems described herein can also be used in other suitable network devices, such as in managing a SB of a Network Interface Controller (NIC).
It will be appreciated that the embodiments described. above are cited by way of example, and that the following claims are not limited to what has been particularly shown and described hereinabove. Rather, the scope includes both combinations and sub-combinations of the various features described hereinabove, as well as variations and modifications thereof which would occur to persons skilled in the art upon reading the foregoing description and which are not disclosed in the prior art. Documents incorporated by reference in the present patent application are to be considered an integral part of the application except that to the extent any terms are defined in these incorporated documents in a manner that conflicts with the definitions made explicitly or implicitly in the present specification, only the definitions in the present specification should be considered.