On-demand processing pipeline interleaved with temporal processing pipeline

Information

  • Patent Grant
  • 10437641
  • Patent Number
    10,437,641
  • Date Filed
    Friday, October 27, 2017
    7 years ago
  • Date Issued
    Tuesday, October 8, 2019
    5 years ago
Abstract
Aspects of the technology provide systems and methods for interleaving a processing task in a network monitoring appliance. A process of the technology could include steps for identifying operators in the network monitoring appliance, wherein the operators are sequentially arranged to form a processing pipeline, determining an epoch for the processing pipeline, wherein the epoch corresponds with a time duration required for a processing task to traverse the pipeline, and determining a time window in which a supplemental processing task can be interleaved into the processing pipeline, wherein the time window is based on the epoch. Machine-readable media are also provided.
Description
BACKGROUND
1. Technical Field

The present technology pertains to network configuration and troubleshooting, and more specifically increasing the throughput of a network monitoring appliance by interleaving processing requests into a temporal processing pipeline.


2. Introduction

Network configurations for large data center networks are often specified at a centralized controller. The controller can realize the intent in the network by programming switches and routers in the data center according to the specified network configurations. Network configurations are inherently very complex, and involve low level as well as high level configurations of several layers of the network such as access policies, forwarding policies, routing policies, security policies, QoS policies, etc. Given such complexity, the network configuration process is error prone.





BRIEF DESCRIPTION OF THE DRAWINGS

In order to describe the manner in which the above-recited and other advantages and features of the disclosure can be obtained, a more particular description of the principles briefly described above will be rendered by reference to specific embodiments thereof which are illustrated in the appended drawings. Understanding that these drawings depict only exemplary embodiments of the disclosure and are not therefore to be considered to be limiting of its scope, the principles herein are described and explained with additional specificity and detail through the use of the accompanying drawings in which:



FIG. 1 illustrates an example network environment in which some aspects of the technology can be implemented.



FIG. 2A illustrates an example network assurance appliance, according to some aspects of the technology.



FIG. 2B illustrates an example Directed Acrylic Graph (DAG) of operators used to form a processing pipeline for a network monitoring appliance, according to some aspects of the technology.



FIG. 3 illustrates steps of an example process interleaving a supplemental processing task into a processing pipeline, according to some aspects of the technology.



FIG. 4 illustrates an example network device in accordance with various embodiments.



FIG. 5 illustrates an example computing device in accordance with various embodiments.





DETAILED DESCRIPTION

The detailed description set forth below is intended as a description of various configurations of the disclosed technology and is not intended to represent the only configurations in which the technology can be practiced. The appended drawings are incorporated herein and constitute a part of the detailed description. The detailed description includes specific details for the purpose of providing a more thorough understanding of the subject technology. However, it will be clear and apparent that the subject technology is not limited to the specific details set forth herein and may be practiced without these details. In some instances, structures and components are shown in block diagram form in order to avoid obscuring the concepts of the subject technology.


Additional features and advantages of the disclosure will be set forth in the description which follows, and in part will be obvious from the description, or can be learned by practice of the herein disclosed principles. The features and advantages of the disclosure can be realized and obtained by means of the instruments and combinations particularly pointed out in the appended claims. These and other features of the disclosure will become more fully apparent from the following description and appended claims, or can be learned by the practice of the principles set forth herein.


Overview:


In some network implementations, one or more network systems (e.g., “network appliances”), can be configured to be connected to and monitor the network fabric. Such deployments can be used to help troubleshoot customer (e.g. tenant) networking issues, to ensure conformity with agreed-upon networking policies, and to ensure an overall high quality experience. In some monitoring appliance deployments, processing is performed sequentially by an ordered series of independent applications or “operators,” that form a processing pipeline. Each operator can be associated with a different processing task that is performed for each received request. Once processing is completed, the request is provided to a next operator in the pipeline.


One limitation of conventional monitoring appliance operation is that pipeline processing begins at predetermined points in a fixed time-period or processing “epoch.” Thus, any particular operator may sit idle, until a processing request has been completed by downstream operators in the pipeline, and a new epoch is begun.


Description


Aspects of the subject technology address the foregoing limitations by providing systems and methods increasing the pipeline throughput by interleaving processing requests into the temporal processing chain, for example, to utilize idle operator processing capabilities.


Each operator in the processing chain is configured to perform its processing capability to a received request upon receipt of a “begin-window” signal, and to end processing and return an “end-window” signal once its processing is complete. That is, idleness for any particular operator can be determined based on the receipt of an end-window signal for that operator.


Knowing the idleness state of various operators in the processing chain can be used to determine when additional requests may be interleaved into the chain. By way of example, if a first operator in the processing pipeline releases an end-window signal, then the first operator can be available to receive additional processing requests, rather than sit idle.



FIG. 1, which illustrates a diagram of an example network environment 100, such as a data center in which a processing pipeline of the technology may be implemented. Network 100 can include Fabric 120 that can represent the physical layer or infrastructure (e.g., underlay) of the network 100. Fabric 120 can include Spines 102 (e.g., spine routers or switches) and Leafs 104 (e.g., leaf routers or switches) which can be interconnected for routing traffic in the Fabric 120. The Spines 102 can interconnect the Leafs 104 in the Fabric 120, and the Leafs 104 can connect the Fabric 120 to the overlay portion of the network 100, which can include application services, servers, virtual machines, containers, endpoints, etc. Thus, network connectivity in the Fabric 120 can flow from Spines 102 to Leafs 104, and vice versa. Leafs 104 can be, for example, top-of-rack (“ToR”) switches, aggregation switches, gateways, ingress and/or egress switches, provider edge devices, and/or any other type of routing or switching device.


Leafs 104 can be responsible for routing and/or bridging tenant or customer packets and applying network policies. Network policies can be driven by the one or more controllers 116 and/or the Leafs 104. The Leafs 104 can connect Servers 106, Hypervisors 108, Virtual Machines (VMs) 110, Applications 112, Endpoints 118, External Routers 114, etc., with the Fabric 120. For example, Leafs 104 can encapsulate and decapsulate packets to and from Servers 106 in order to enable communications throughout the network 100, including the Fabric 120. Leafs 104 can also provide any other devices, services, tenants, or workloads with access to the Fabric 120.


Applications 112 can include software applications, services, operators, containers, appliances, functions, service chains, etc. For example, Applications 112 can include a firewall, a database, a CDN server, an IDS/IPS, a deep packet inspection service, a message router, a virtual switch, etc. VMs 110 can be virtual machines hosted by Hypervisors 108 running on Servers 106. VMs 110 can include workloads running on a guest operating system on a respective server. Hypervisors 108 can provide a layer of software, firmware, and/or hardware that creates and runs the VMs 110. Hypervisors 108 can allow VMs 110 to share hardware resources on Servers 106, and the hardware resources on Servers 106 to appear as multiple, separate hardware platforms. Moreover, Hypervisors 108 on Servers 106 can each host one or more VMs 110.


In some cases, VMs 110 and/or Hypervisors 108 can be migrated to other Servers 106. Servers 106 can similarly be migrated to other locations in the network environment 100. For example, a server connected to a specific leaf can be changed to connect to a different or additional leaf. Such configuration or deployment changes can involve modifications to settings and policies that are applied to the resources being migrated.


Servers 106, Hypervisors 108, and/or VMs 110 can represent a tenant or customer space. Tenant space can include workloads, services, applications, devices, and/or resources that are associated with one or more clients or subscribers. Accordingly, traffic in the network environment 100 can be routed based on specific tenant policies, spaces, agreements, configurations, etc. Moreover, addressing can vary between one or more tenants. In some configurations, tenant spaces can be divided into logical segments and/or networks and separated from logical segments and/or networks associated with other tenants. Addressing, policy, and configuration information between tenants can be managed by one or more controllers 116.


Policies, configurations, settings, etc., in the network can be implemented at the application level, the physical level, and/or both. For example, one or more controllers 116 can define a policy model at the application level which defines policies and other settings for groups of applications or services, such as endpoint groups. In some addition, the Leafs 104, as well as other physical devices such as physical servers or Spines 102, can apply specific policies to traffic. For example, Leafs 104 can apply specific policies or contracts to traffic based on tags or characteristics of the traffic, such as protocols associated with the traffic, applications or endpoint groups associated with the traffic, network address information associated with the traffic, etc.


In some examples, network 100 can be configured according to a particular software-defined network (SDN) solution. The network 100 can deploy one or more SDN solutions, such as CISCO Application Centric Infrastructure (ACI) or VMWARE NSX solutions. These example SDN solutions are briefly described below.


Application Centric Infrastructure (ACI) is an example SDN solution that can be implemented in network 100. ACI can provide an application policy-based solution through scalable distributed enforcement. ACI supports integration of physical and virtual environments under a declarative policy model for networks, servers, services, security, requirements, etc. For example, the ACI framework implements End Point Groups (EPGs), which can include a collection of endpoints or applications that share common policy requirements, such as security, QoS, services, etc. Endpoints can be virtual/logical or physical devices, such as VMs and bare-metal physical servers that are connected to the network 100. Endpoints can have one or more attributes such as VM name, guest OS name, a security tag, etc. Application policies can be applied between EPGs, instead of endpoints directly, in the form of contracts. The Leafs 104 can classify incoming traffic into different EPGs. The classification can be based on, for example, a network segment identifier such as a VLAN ID, VXLAN Network Identifier (VNID), NVGRE Virtual Subnet Identifier (VSID), MAC address, IP address, etc.


In some cases, classification in the ACI infrastructure can be implemented by Application Virtual Switches (AVS), which can run on a host, and physical hosts. For example, an AVS can classify traffic based on specified attributes, and tag packets of different attribute EPGs with different identifiers, such as network segment identifiers (e.g., VLAN ID). Finally, Leafs 104 can tie packets with their attribute EPGs based on their identifiers and enforce policies, which can be implemented and/or managed by one or more controllers 116, such as an application policy infrastructure controller (APIC). The Leaf 104 can classify to which EPG the traffic from a host belong and enforce policies accordingly.


Another example SDN solution is based on VMWare NSX. With VMWare NSX, hosts can run a distributed firewall (DFW) which can classify and process traffic. Consider a case where three types of VMs, namely, application, database and web VMs, are put into a single layer-2 network segment. Traffic protection can be provided within the network segment based on the VM type. For example, HTTP traffic can be allowed among web VMs, and disallowed between a web VM and an application or database VM. To classify traffic and implement policies, VMWARE NSX can implement security groups, which can be used to group the specific VMs (e.g., web VMs, application VMs, database VMs). DFW rules can be configured to implement policies for the specific security groups. To illustrate, from our previous example, DFW rules can be configured to block HTTP traffic between web, application, and database security groups.


Network 100 may deploy different hosts via the Leafs 104, Servers 106, Hypervisors 108, VMs 110, Applications 112, Controllers 116, and/or Endpoints 118, such as VMware ESXi hosts, Windows Hyper-V hosts, bare metal physical hosts, etc. The network 100 may interoperate with a wide variety of Hypervisors 108, Servers 106 (e.g., physical and/or virtual servers), SDN orchestration platforms, etc. The network 100 may implement a declarative model to allow its integration with application design and holistic network policy.


One or more controllers 116 can provide centralized access to fabric information, application configuration, resource configuration, application-level policy modeling for a software-defined network (SDN) infrastructure, integration with management systems or servers, etc. The one or more controllers 116 can form a control plane that interfaces with an application plane via northbound APIs and a data plane via southbound APIs. In some examples, the one or more controllers 116 can include SDN controllers or managers, such as an application policy infrastructure controller (APIC) or a vCenter NSX Manager.


Controllers 116 can define and manage application-level model(s) for policies in the network 100. In some cases, application or device policies can also be managed and/or defined by other components in the network. For example, a hypervisor or virtual appliance, such as a VM or container, can run a server or management tool to manage software and services in the network 100, including policies and settings for virtual appliances.


As referenced herein, the term “hosts” can refer to servers 106 (e.g., physical or logical), Hypervisors 108, VMs 110, containers (e.g., Applications 112), EPs 118, etc., and can run or include any type of server or application solution. Non-limiting examples of “hosts” can include DVS virtual servers, vCenter and NSX Managers, bare metal physical hosts, AVS hosts, Hyper-V hosts, VMs, Docker Containers, Virtual Routers/Switches (e.g., VPP), etc.



FIG. 2A illustrates an example network assurance appliance. In this example, Appliance 200 can include k VMs 110 operating in cluster mode. VMs are used in this example for explanation purposes. However, it should be understood that other configurations are also contemplated herein, such as use of containers, bare metal devices, Endpoints 122, or any other physical or logical systems. Moreover, while FIG. 2A illustrates a cluster mode configuration, other configurations are also contemplated herein, such as a single mode configuration (e.g., single VM, container, or server) or a service chain for example.


Appliance 200 can run on one or more Servers 106, VMs 110, Hypervisors 108, EPs 122, Leafs 104, Controllers 116, or any other system or resource. For example, Assurance Appliance 200 can be a logical service or application running on one or more VMs 110 in Network Environment 100.


Appliance 200 can include Data Framework 208, which can be based on, for example, APACHE APEX or HADOOP file system (e.g., HDFS). In some cases, assurance checks can be written as individual operators that reside in Data Framework 208. This enables a natively horizontal scale-out architecture that can scale to arbitrary number of switches in Fabric 120 (e.g., ACI fabric).


Appliance 200 can poll Fabric 120 at a configurable periodicity (e.g., an epoch). The analysis workflow can be setup as a DAG (Directed Acyclic Graph) of Operators 210, where data flows from one operator to another and eventually results are generated and persisted to Database 202 for each interval (e.g., each epoch). Operators 210 in Data Framework 208 (e.g., APEX/Hadoop) can together support assurance operations. It is understood that the topological ordering of operators in the DAG configuration proceeds sequentially, beginning with a first operator in the pipeline, and proceeding to a last operator in the pipeline.


In practice, processing tasks received by the processing pipeline proceed in a sequential order, where processing by a given operator cannot begin until processing by a previous operator is concluded. As discussed above, the processing status of a given operator can be indicated by use of signals indicating a “begin-window”, after which processing by the operator is performed, and an “end-window”, after which operator processing has ended.


Additional (supplemental) processing tasks can be interleaved into the processor pipeline, for example, at a time that it is known that the receiving operators are idle. The idleness state of a receiving operator may be inferred based on the epoch duration of the pipeline, as well as timing of end-window signals for the receiving operator. By way of example, suppose that a processing pipeline consisting of 4 operators, each with a processing time of approximately five-minutes, and that is configured to receive periodic processing tasks at 15 minute intervals. It may be inferred that the pipeline may receive supplemental processing tasks at the first operator during a time window of about 5 minutes after the first operator indicates an end-window. That is, the availability to interleave supplemental processing tasks can be determined based on the epoch time for the pipeline, a periodicity of routinely received processing tasks (e.g., from a tenant network), known processing times for each operator, and/or estimates of processing time required for the supplemental processing task.


In another implementation, operator idleness may be determined based on end-window and begin-window signals, as well as a known periodicity of routinely received processing tasks. Further to the above example, if processing tasks are received from a tenant network at approximately 20 minute intervals, then a supplemental processing task may be provided to a first operator in the pipeline up to 15 minutes after the end-window signal is received.



FIG. 2B illustrates an example of a data processing pipeline 201 formed from multiple operators sequentially organized in a Directed Acrylic Graph (DAG) topology. Pipeline 201 includes operators A, B, C, D, E and F, each configured to perform an associated processing task, and to output a processing result to a next operator in pipeline 201. Therefore, the epoch of the illustrated configuration is the time required for a processing task to begin processing at operator A, and to end operator F.


It is understood that a processing pipeline of the technology can have a greater (or fewer) number of operators than are illustrated in FIG. 2B. Additionally, the arrangement or sequence of operators can vary depending on implementation. Although operators A-F can represent different functions or applications, in some aspects, two or more operators may be configured to perform the same (or a similar) function.


Depending on the type of processing required for a supplemental processing task, the request may be interleaved into the processing pipeline at a downstream operator. By way of example, if processing required for a supplemental processing task can be performed by operator C, and it is determined that operator C is idle (e.g., operator C has issued an end-window signal), a supplemental processing task may be injected into pipeline 201 at operator C, skipping operators A and B. As discussed above with respect to FIG. 2A, the selection of an idle operator, and interleaving of a supplemental processing task into the pipeline can be based on a variety of factors, including but not limited to: determinations idleness for one or more operators in the pipeline, an epoch of the pipeline, a known cost associated with the supplemental processing task, and/or signals received by one or more operators (e.g., end-window and/or begin-window signals).


Referring back to FIG. 2A, it is understood that a network monitoring appliance 200 can perform other types of assurance operations using operators 210. Below are non-limiting examples of assurance operations that can be performed by Assurance Appliance 200 via Operators 210.


Security Policy Adherence:


Assurance Appliance 200 can check to make sure the configurations or specification that reflect a user's (e.g., administrator's) intent for the network, including for example the security policies and customer-configured contracts, are correctly implemented and/or rendered, and thus properly implemented and rendered by the fabric members (e.g., Leafs 104), and report any errors, contract violations, or irregularities found.


Static Policy Analysis:


Assurance Appliance 200 can check for issues in the specification of the user's intent or intents (e.g., identify contradictory or conflicting policies).


TCAM Utilization:


Ternary Content-Addressable Memory (TCAM) is a scarce resource in the fabric (e.g., Fabric 120). However, Assurance Appliance 200 can analyze the TCAM utilization by the network data (e.g., Longest Prefix Match (LPM) tables, routing tables, VLAN tables, BGP updates, etc.), Contracts, Logical Groups 118 (e.g., EPGs), Tenants, Spines 102, Leafs 104, and other dimensions in Network Environment 100 and/or objects in MIM 200, to provide a network operator or user visibility into the utilization of this scarce resource. This can greatly help for planning and other optimization purposes.


Endpoint Checks:


In some aspects, assurance Appliance 200 can validate that the fabric (e.g. fabric 120) has no inconsistencies in the Endpoint information registered (e.g., two leafs announcing the same endpoint, duplicate subnets, etc.), among other such checks.


Tenant Routing/Forwarding Checks:


Assurance Appliance 200 can validate that BDs, VRFs, subnets (both internal and external), VLANs, contracts, filters, applications, EPGs, etc., are correctly programmed.


Infrastructure Routing:


Assurance Appliance 200 can validate that infrastructure routing (e.g., IS-IS protocol) has no convergence issues leading to black holes, loops, flaps, and other problems.


MP-BGP Route Reflection Checks:


The network fabric (e.g., Fabric 120) can interface with other external networks and provide connectivity to them via one or more protocols, such as Border Gateway Protocol (BGP), Open Shortest Path First (OSPF), etc. The learned routes are advertised within the network fabric via, for example, MP-BGP. These checks can ensure that a route reflection service via, for example, MP-BGP (e.g., from Border Leaf) does not have health issues.



FIG. 3 illustrates steps of an example process interleaving a supplemental processing task into a processing pipeline, according to some aspects of the technology. Process 300 begins with step 302 in which multiple operators running in a monitoring appliance are identified, e.g., by a network controller. Further to the examples discussed above with respect to FIGS. 1 and 2, the operators may be processes supported by various network devices, such as VMs or network containers, and that are configured to provide a processing pipeline used to monitor a tenant network. In some aspects, operator identification can include determinations of various statistical parameters relating to one or more of the operators. By way of example, average processing times, application function, and/or operator co-dependence information may be determined for each operator in the pipeline.


In step 304, an epoch of the processing pipeline is determined. An epoch of the pipeline can be determined in different ways, depending on implementation. In some aspects, the epoch duration may be determined by multiple measurements of end-to-end processing, for example, that are performed in the course of normal operation in a tenant network deployment. In other aspects, the epoch may be determined based on processing time estimates for each of the constituent operators. Additionally, the determined epoch may be determined as a function of the processing load that is introduced. By way of example, larger processing loads may correlate with longer epochs, as compared to smaller relative processing loads.


In step 306, a time window in which a supplemental processing task can interleaved into the processing timeline is determined. Because the time window may be operator specific, such determinations can be based upon which operator in the pipeline is to receive the supplemental processing task. That is, the time window may be a function of the type of processing that is required for the supplemental processing task. As discussed above with respect to FIGS. 2A and 2B, the determined time window can represent a temporal opportunity in which supplemental processing tasks can be interleaved i.e., during otherwise idle operator periods.


In step 308, an idle operator is selected for receipt of the supplemental processing request, and the supplemental processing request is interleaved into the processing pipeline via the selected idle operator. After being interleaved into the processing pipeline, the supplemental processing task is passed sequentially to each down-stream operator in the sequential pipeline. By way of example, referring to FIG. 2B, if operator C receives the supplemental processing request, upon completion of processing by operator C, the supplemental processing task is passed to operator E and then to operator F, after which the processing is complete.



FIG. 4 illustrates an example network device 400 suitable for implementing a network appliance of the subject technology. Network device 400 includes a central processing unit (CPU) 404, interfaces 402, and a bus 410 (e.g., a PCI bus). When acting under the control of appropriate software or firmware, the CPU 404 is responsible for executing packet management, error detection, and/or routing functions. CPU 404 accomplishes all these functions under the control of software including an operating system and any appropriate applications software. CPU 404 may include one or more processors 408, such as a processor from the INTEL X86 family of microprocessors. In some cases, processor 408 can be specially designed hardware for controlling the operations of network device 400. In some cases, a memory 406 (e.g., non-volatile RAM, ROM, etc.) also forms part of CPU 404. However, there are many different ways in which memory could be coupled to the system.


The interfaces 402 are typically provided as modular interface cards (sometimes referred to as “line cards”). They can control the sending and receiving of data packets over the network and sometimes support other peripherals used with the network device 400. Among the interfaces that may be provided are Ethernet interfaces, frame relay interfaces, cable interfaces, DSL interfaces, token ring interfaces, and the like. In addition, various very high-speed interfaces may be provided such as fast token ring interfaces, wireless interfaces, Ethernet interfaces, Gigabit Ethernet interfaces, ATM interfaces, HSSI interfaces, POS interfaces, FDDI interfaces, WIFI interfaces, 3G/4G/5G cellular interfaces, CAN BUS, LoRA, and the like. Generally, these interfaces may include ports appropriate for communication with the appropriate media. In some cases, they may also include an independent processor and, in some instances, volatile RAM. The independent processors may control such communications intensive tasks as packet switching, media control, signal processing, crypto processing, and management. By providing separate processors for the communications intensive tasks, these interfaces allow the master microprocessor 404 to efficiently perform routing computations, network diagnostics, security functions, etc.


Although the system shown in FIG. 4 is one specific network device of the present invention, it is by no means the only network device architecture on which the present invention can be implemented. For example, an architecture having a single processor that handles communications as well as routing computations, etc., is often used. Further, other types of interfaces and media could also be used with the network device 400.


Regardless of the network device's configuration, it may employ one or more memories or memory modules (including memory 406) configured to store program instructions for the general-purpose network operations and mechanisms for roaming, route optimization and routing functions described herein. The program instructions may control the operation of an operating system and/or one or more applications, for example. The memory or memories may also be configured to store tables such as mobility binding, registration, and association tables, etc. Memory 406 could also hold various software containers and virtualized execution environments and data.


In some implementations, the program instructions may be configured to cause CPU 404 and/or processor 408 to perform operations for simulating failure events in a tenant network. In particular, the program instructions can cause CUP 404 and/or processor 408 to perform operations for identifying, by a network controller, a plurality of operators running in a network monitoring appliance, instantiating, by the network controller, a plurality of sensors, wherein each of the plurality of sensors is associated with a respective one of the plurality of operators, and wherein each sensor is configured to capture error log data of its corresponding operator, and aggregating, by the network controller, error log data from two or more of the plurality of sensors. In some aspects, a method of the disclosed technology may further include steps for automatically identifying, by the controller, a malfunction in at least one operator from among the plurality of operators in the network monitoring appliance. In some aspects, the error log data may be aggregated from the two or more of the plurality of sensors provides information regarding a malfunction in a tenant network coupled to the network monitoring appliance.


Network device 400 can also include an application-specific integrated circuit (ASIC), which can be configured to perform routing and/or switching operations. The ASIC can communicate with other components in the network device 400 via the bus 410, to exchange data and signals and coordinate various types of operations by the network device 400, such as routing, switching, and/or data storage operations, for example.



FIG. 5 illustrates a computing architecture 500 wherein the components of the system are in electrical communication with each other via connection 505, such as a bus. System 500 includes a processing unit (CPU or processor) 510 and a system connection 505 that couples various system components including system memory 515, such as read only memory (ROM) 520 and random access memory (RAM) 525, to processor 510. System 500 can include a cache of high-speed memory connected directly with, in close proximity to, or integrated as part of the processor 510. The system 500 can copy data from the memory 515 and/or the storage device 530 to the cache 512 for quick access by processor 510. In this way, the cache can provide a performance boost that avoids processor 510 delays while waiting for data. These and other modules can control or be configured to control the processor 510 to perform various actions. Other system memory 515 may be available for use as well. The memory 515 can include multiple different types of memory with different performance characteristics. The processor 510 can include any general purpose processor and a hardware or software service, such as service 1532, service 2534, and service 3536 stored in storage device 530, configured to control the processor 510 as well as a special-purpose processor where software instructions are incorporated into the actual processor design. The processor 510 may be a completely self-contained computing system, containing multiple cores or processors, a bus, memory controller, cache, etc. A multi-core processor may be symmetric or asymmetric.


To enable user interaction with the computing device 500, an input device 545 can represent any number of input mechanisms, such as a microphone for speech, a touch-sensitive screen for gesture or graphical input, keyboard, mouse, motion input, speech and so forth. An output device 535 can also be one or more of a number of output mechanisms known to those of skill in the art. In some instances, multimodal systems can enable a user to provide multiple types of input to communicate with the computing device 500. The communications interface 540 can generally govern and manage the user input and system output. There is no restriction on operating on any particular hardware arrangement and therefore the basic features here may easily be substituted for improved hardware or firmware arrangements as they are developed.


Storage device 530 is a non-volatile memory and can be a hard disk or other types of computer readable media which can store data that are accessible by a computer, such as magnetic cassettes, flash memory cards, solid state memory devices, digital versatile disks, cartridges, random access memories (RAMs) 525, read only memory (ROM) 520, and hybrids thereof.


The storage device 530 can include services 532, 534, 536 for controlling the processor 510. Other hardware or software modules are contemplated. Storage device 530 can be connected to the system connection 505. In one aspect, a hardware module that performs a particular function can include the software component stored in a computer-readable medium in connection with the necessary hardware components, such as the processor 510, connection 505, output device 535, and so forth, to carry out the function.


For clarity of explanation, in some instances the present technology may be presented as including individual functional blocks including functional blocks comprising devices, device components, steps or routines in a method embodied in software, or combinations of hardware and software.


In some embodiments the computer-readable storage devices, mediums, and memories can include a cable or wireless signal containing a bit stream and the like. However, when mentioned, non-transitory computer-readable storage media expressly exclude media such as energy, carrier signals, electromagnetic waves, and signals per se.


Methods according to the above-described examples can be implemented using computer-executable instructions that are stored or otherwise available from computer readable media. Such instructions can comprise, for example, instructions and data which cause or otherwise configure a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions.


Portions of computer resources used can be accessible over a network. The computer executable instructions may be, for example, binaries, intermediate format instructions such as assembly language, firmware, or source code. Examples of computer-readable media that may be used to store instructions, information used, and/or information created during methods according to described examples include magnetic or optical disks, flash memory, USB devices provided with non-volatile memory, networked storage devices, and so on.


Devices implementing methods according to these disclosures can comprise hardware, firmware and/or software, and can take any of a variety of form factors. Typical examples of such form factors include laptops, smart phones, small form factor personal computers, personal digital assistants, rackmount devices, standalone devices, and so on. Functionality described herein also can be embodied in peripherals or add-in cards. Such functionality can also be implemented on a circuit board among different chips or different processes executing in a single device, by way of further example.


The instructions, media for conveying such instructions, computing resources for executing them, and other structures for supporting such computing resources are means for providing the functions described in these disclosures.


Although a variety of examples and other information was used to explain aspects within the scope of the appended claims, no limitation of the claims should be implied based on particular features or arrangements in such examples, as one of ordinary skill would be able to use these examples to derive a wide variety of implementations. Further and although some subject matter may have been described in language specific to examples of structural features and/or method steps, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to these described features or acts. For example, such functionality can be distributed differently or performed in components other than those identified herein. Rather, the described features and steps are disclosed as examples of components of systems and methods within the scope of the appended claims.

Claims
  • 1. A computer-implemented method for interleaving a processing task in a network monitoring appliance, comprising: identifying a plurality of operators in the network monitoring appliance, wherein the plurality of operators are sequentially arranged to form a sequentially processing pipeline;determining an epoch for the processing pipeline,wherein the epoch corresponds with a time duration required for a processing task to traverse the processing pipeline; anddetermining a time window in which a supplemental processing task can be interleaved into the processing pipeline,wherein the time window is based on the epoch;interleaving the supplemental processing task into the processing pipeline during the time window.
  • 2. The computer-implemented method of claim 1, further comprising: selecting an idle operator, from among the plurality of operators in the processing pipeline; andwherein the interleaving is executed by the idle operator.
  • 3. The computer-implemented method of claim 2, wherein selecting the idle operator is further based on a type of processing associated with the supplemental processing task.
  • 4. The computer-implemented method of claim 1, wherein determining the time window in which the supplemental processing task can be interleaved into the processing pipeline is further based on an estimated time duration to complete the supplemental processing task.
  • 5. The computer-implemented method of claim 1, wherein the plurality of operators are setup as a Directed Acrylic Graph (DAG).
  • 6. The computer-implemented method of claim 1, wherein each of the plurality of operators are configured to perform a different processing task.
  • 7. The computer-implemented method of claim 1, wherein the network monitoring appliance is coupled to a tenant network, and wherein the processing pipeline is configured to analyze data received from the tenant network to facilitate identification of one or more failure events in the tenant network.
  • 8. A system comprising: one or more processors;a network interface coupled to the processors; anda computer-readable medium coupled to the processors, the computer-readable medium comprising instructions stored therein, which when executed by the processors, cause the processors to perform operations comprising:identifying a plurality of operators in the network monitoring appliance,wherein the plurality of operators are sequentially arranged to form a sequentially processing pipeline;determining an epoch for the processing pipeline, wherein the epoch corresponds with a time duration required for a processing task to traverse the processing pipeline; anddetermining a time window in which a supplemental processing task can be interleaved into the processing pipeline,wherein the time window is based on the epoch; interleaving the supplemental processing task into the processing pipeline during the time window.
  • 9. The system of claim 8, wherein the processors are further configured to perform operations comprising: selecting an idle operator, from among the plurality of operators in the processing pipeline, based on the time window; andwherein the interleaving is executed by the idle operator.
  • 10. The system of claim 9, wherein selecting the idle operator is further based on a type of processing associated with the supplemental processing task.
  • 11. The system of claim 8, wherein determining the time window in which the supplemental processing task can be interleaved into the processing pipeline is further based on an estimated time duration to complete the supplemental processing task.
  • 12. The system of claim 8, wherein the plurality of operators are setup as a Directed Acrylic Graph (DAG).
  • 13. The system of claim 8, wherein each of the plurality of operators are configured to perform a different processing task.
  • 14. The system of claim 8, wherein the network monitoring appliance is coupled to a tenant network, and wherein the processing pipeline is configured to analyze data received from the tenant network to facilitate identification of one or more failure events in the tenant network.
  • 15. A non-transitory computer-readable storage medium comprising instructions stored therein, which when executed by one or moreprocessors, cause the processors to perform operations comprising: identifying a plurality of operators in the network monitoring appliance,wherein the plurality of operators are sequentially arranged to form a sequentially processing pipeline:
  • 16. The non-transitory computer-readable storage medium claim 15, wherein the processors are further configured to perform operations comprising: selecting an idle operator, from among the plurality of operators in the processing pipeline, based on the time window; andwherein the interleaving is executed by the idle operator.
  • 17. The non-transitory computer-readable storage medium of claim 16, wherein selecting the idle operator is further based on a type of processing associated with the supplemental processing task.
  • 18. The non-transitory computer-readable storage medium of claim 15, wherein determining the time window in which the supplemental processing task can be interleaved into the processing pipeline is further based on an estimated time duration to complete the supplemental processing task.
  • 19. The non-transitory computer-readable storage medium of claim 15, wherein the plurality of operators are setup as a Directed Acrylic Graph (DAG).
  • 20. The non-transitory computer-readable storage medium of claim 15, wherein each of the plurality of operators are configured to perform a different processing task.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application claims the benefit of U.S. Application No. 62/521,685, filed Jun. 19, 2017, entitled “ON-DEMAND PROCESSING PIPELINE INTERLEAVED WITH TEMPORAL PROCESSING PIPELINE”, which is incorporated by reference in its entirety.

US Referenced Citations (180)
Number Name Date Kind
4945775 Adams Aug 1990 A
5117387 Nemirovsky May 1992 A
5204829 Lyu et al. Apr 1993 A
5835776 Tirumalai Nov 1998 A
5867711 Subramanian Feb 1999 A
6195676 Spix Feb 2001 B1
6205468 Diepstraten Mar 2001 B1
6570564 Sowizral May 2003 B1
6675189 Rehg Jan 2004 B2
6763380 Mayton et al. Jul 2004 B1
6975638 Chen Dec 2005 B1
7003562 Mayer Feb 2006 B2
7089369 Emberling Aug 2006 B2
7127686 Dreschler et al. Oct 2006 B2
7360064 Steiss et al. Apr 2008 B1
7453886 Allan Nov 2008 B1
7503046 Beaumont Mar 2009 B2
7505463 Schuba et al. Mar 2009 B2
7548967 Amyot et al. Jun 2009 B2
7552201 Areddu et al. Jun 2009 B2
7609647 Turk et al. Oct 2009 B2
7619989 Guingo et al. Nov 2009 B2
7698561 Nagendra et al. Apr 2010 B2
7743274 Langford et al. Jun 2010 B2
7765093 Li et al. Jul 2010 B2
8010952 Datla et al. Aug 2011 B2
8073935 Viswanath Dec 2011 B2
8103480 Korn et al. Jan 2012 B2
8190719 Furukawa May 2012 B2
8209738 Nicol et al. Jun 2012 B2
8261339 Aldridge et al. Sep 2012 B2
8312261 Rao et al. Nov 2012 B2
8375117 Venable, Sr. Feb 2013 B2
8441941 McDade et al. May 2013 B2
8479267 Donley et al. Jul 2013 B2
8484693 Cox et al. Jul 2013 B2
8494977 Yehuda et al. Jul 2013 B1
8554883 Sankaran Oct 2013 B2
8589934 Makljenovic et al. Nov 2013 B2
8621284 Kato Dec 2013 B2
8627328 Mousseau et al. Jan 2014 B2
8693344 Adams et al. Apr 2014 B1
8693374 Murphy et al. Apr 2014 B1
8761036 Fulton et al. Jun 2014 B2
8769543 Sandstrom Jul 2014 B2
8782182 Chaturvedi et al. Jul 2014 B2
8824482 Kajekar et al. Sep 2014 B2
8910143 Cohen et al. Dec 2014 B2
8914843 Bryan et al. Dec 2014 B2
8924798 Jerde et al. Dec 2014 B2
8954993 Kruglick Feb 2015 B2
9019840 Salam et al. Apr 2015 B2
9038151 Chua et al. May 2015 B1
9055000 Ghosh et al. Jun 2015 B1
9106555 Agarwal et al. Aug 2015 B2
9122523 Pienaar Sep 2015 B2
9137096 Yehuda et al. Sep 2015 B1
9225601 Khurshid et al. Dec 2015 B2
9246818 Deshpande et al. Jan 2016 B2
9264922 Gillot et al. Feb 2016 B2
9276877 Chua et al. Mar 2016 B1
9319300 Huynh Van et al. Apr 2016 B2
9344348 Ivanov et al. May 2016 B2
9369434 Kim et al. Jun 2016 B2
9389993 Okmyanskiy et al. Jul 2016 B1
9405553 Branson et al. Aug 2016 B2
9444842 Porras et al. Sep 2016 B2
9497207 Dhawan et al. Nov 2016 B2
9497215 Vasseur et al. Nov 2016 B2
9544224 Chu et al. Jan 2017 B2
9548965 Wang et al. Jan 2017 B2
9553845 Talmor et al. Jan 2017 B1
9571502 Basso et al. Feb 2017 B2
9571523 Porras et al. Feb 2017 B2
9594640 Chheda Mar 2017 B1
9596141 McDowall Mar 2017 B2
9641249 Kaneriya et al. May 2017 B2
9654300 Pani May 2017 B2
9654361 Vasseur et al. May 2017 B2
9654409 Yadav et al. May 2017 B2
9660886 Ye et al. May 2017 B1
9660897 Gredler May 2017 B1
9667645 Belani et al. May 2017 B1
9680875 Knjazihhin et al. Jun 2017 B2
9686180 Chu et al. Jun 2017 B2
9686296 Murchison et al. Jun 2017 B1
9690644 Anderson et al. Jun 2017 B2
9781004 Danait et al. Oct 2017 B2
9787559 Schroeder Oct 2017 B1
9998247 Choudhury et al. Jun 2018 B1
10084795 Akireddy et al. Sep 2018 B2
10084833 McDonnell et al. Sep 2018 B2
10084895 Kasat et al. Sep 2018 B2
10108516 Palmer Oct 2018 B2
10191768 Bishop Jan 2019 B2
20020143855 Traversat et al. Oct 2002 A1
20020178246 Mayer Nov 2002 A1
20030229693 Mahlik et al. Dec 2003 A1
20040073647 Gentile et al. Apr 2004 A1
20040168100 Thottan et al. Aug 2004 A1
20050108389 Kempin et al. May 2005 A1
20070011629 Shacham et al. Jan 2007 A1
20070124437 Chervets May 2007 A1
20070214244 Hitokoto et al. Sep 2007 A1
20080031147 Fieremans et al. Feb 2008 A1
20080117827 Matsumoto et al. May 2008 A1
20080133731 Bradley et al. Jun 2008 A1
20080172716 Talpade et al. Jul 2008 A1
20090240758 Pasko et al. Sep 2009 A1
20090249284 Antosz et al. Oct 2009 A1
20100191612 Raleigh Jul 2010 A1
20100198909 Kosbab et al. Aug 2010 A1
20110093612 Murakami Apr 2011 A1
20110295983 Medved et al. Dec 2011 A1
20120054163 Liu et al. Mar 2012 A1
20120198073 Srikanth et al. Aug 2012 A1
20120297061 Pedigo et al. Nov 2012 A1
20130097660 Das et al. Apr 2013 A1
20130191516 Sears Jul 2013 A1
20140019597 Nath et al. Jan 2014 A1
20140177638 Bragg et al. Jun 2014 A1
20140222996 Vasseur et al. Aug 2014 A1
20140304831 Hidlreth et al. Oct 2014 A1
20140307556 Zhang Oct 2014 A1
20140321277 Lynn, Jr. et al. Oct 2014 A1
20140379915 Yang et al. Dec 2014 A1
20150019756 Masuda Jan 2015 A1
20150113143 Stuart et al. Apr 2015 A1
20150124826 Edsall et al. May 2015 A1
20150186206 Bhattacharya et al. Jul 2015 A1
20150234695 Cuthbert et al. Aug 2015 A1
20150244617 Nakil et al. Aug 2015 A1
20150271104 Chikkamath et al. Sep 2015 A1
20150295771 Cuni et al. Oct 2015 A1
20150365314 Hiscock et al. Dec 2015 A1
20150381484 Hira et al. Dec 2015 A1
20160020993 Wu et al. Jan 2016 A1
20160021141 Liu et al. Jan 2016 A1
20160026631 Salam et al. Jan 2016 A1
20160036636 Erickson et al. Feb 2016 A1
20160048420 Gourlay et al. Feb 2016 A1
20160078220 Scharf et al. Mar 2016 A1
20160080350 Chaturvedi et al. Mar 2016 A1
20160099883 Voit et al. Apr 2016 A1
20160105317 Zimmermann et al. Apr 2016 A1
20160112246 Singh et al. Apr 2016 A1
20160112269 Singh et al. Apr 2016 A1
20160149751 Pani et al. May 2016 A1
20160164748 Kim Jun 2016 A1
20160224277 Batra et al. Aug 2016 A1
20160241436 Fourie et al. Aug 2016 A1
20160254964 Benc Sep 2016 A1
20160267384 Salam et al. Sep 2016 A1
20160323319 Gourlay et al. Nov 2016 A1
20160330076 Tiwari et al. Nov 2016 A1
20160352566 Mekkattuparamnban et al. Dec 2016 A1
20160380892 Mahadevan et al. Dec 2016 A1
20170026292 Smith et al. Jan 2017 A1
20170031800 Shani et al. Feb 2017 A1
20170031970 Burk Feb 2017 A1
20170048110 Wu et al. Feb 2017 A1
20170048126 Handige Shankar et al. Feb 2017 A1
20170054758 Maino et al. Feb 2017 A1
20170063599 Wu et al. Mar 2017 A1
20170093630 Foulkes Mar 2017 A1
20170093664 Lynam et al. Mar 2017 A1
20170093750 McBride et al. Mar 2017 A1
20170093918 Banerjee et al. Mar 2017 A1
20170111259 Wen et al. Apr 2017 A1
20170118167 Subramanya et al. Apr 2017 A1
20170126740 Bejarano Ardila et al. May 2017 A1
20170126792 Halpern et al. May 2017 A1
20170134233 Dong et al. May 2017 A1
20170163442 Shen et al. Jun 2017 A1
20170187577 Nevrekar et al. Jun 2017 A1
20170195187 Bennett et al. Jul 2017 A1
20170206129 Yankilevich et al. Jul 2017 A1
20170222873 Lee et al. Aug 2017 A1
20180069754 Dasu et al. Mar 2018 A1
20180167294 Gupta et al. Jun 2018 A1
Foreign Referenced Citations (19)
Number Date Country
105471830 Apr 2016 CN
105721193 Jun 2016 CN
105721297 Jun 2016 CN
106130766 Nov 2016 CN
106603264 Apr 2017 CN
103701926 Jun 2017 CN
WO 2015014177 Feb 2015 WO
WO 2015187337 Dec 2015 WO
WO 2016011888 Jan 2016 WO
WO 2016039730 Mar 2016 WO
WO 2016072996 May 2016 WO
WO 2016085516 Jun 2016 WO
WO 2016093861 Jun 2016 WO
WO 2016119436 Aug 2016 WO
WO 2016130108 Aug 2016 WO
WO 2016161127 Oct 2016 WO
WO 2017031922 Mar 2017 WO
WO 2017039606 Mar 2017 WO
WO 2017105452 Jun 2017 WO
Non-Patent Literature Citations (42)
Entry
Ghica et al,“Selecting Tracking Principals with Epoch Awareness”, ACM, pp. 222-231, 2010 (Year: 2010).
Sun et al, “TCP Performance Optimization for Epoch-based Execution”, ACM, pp. 1-7, 2017 (Year: 2017).
Tournavitis et al, Semi-Automatic Extraction and Exploitation of Hierarchical Pipeline Parallelism Using Profiling Information, ACM, pp. 377-388 (Year: 2010).
Chou , “Low-Cost Epoch-Based Correlation Prefetching for Commercial Applications”, IEEE, 301-313 (Year: 2007).
Quan et al, “A Task Scheduling Algorithm for Automated Parallel Processing of XQuery”, IEEE, 250-254 (Year: 2014).
Schor et al, “Expandable Process Networks to Efficiently Specify and Explore Task, Data, and Pipeline Parallelism”, IEEE, pp. 1-10 ( Year: 2013).
Cisco Systems, Inc., “Cisco Application Centric Infrastructure 9ACI Endpoint Groups (EPG) Usage and Design,” White Paper, May 2014, pp. 1-14.
Cisco Systems, Inc., “The Cisco Application Policy Infrastructure Controller Introduction: What is the Cisco Application Policy Infrastructure Controller?” Jul. 31, 2014, 19 pages.
Jain, Praveen, et al., “In-Line Distributed and Stateful Security Policies for Applications in a Network Environment,” Cisco Systems, Inc., Aug. 16, 2016, 13 pages.
Maldonado-Lopez, Ferney, et al., “Detection and prevention of firewall—rule conflicts on software-defined networking,” 2015 7th International Workshop on Reliable Networks Design and Modeling (RNDM), IEEE, Oct. 5, 2015, pp. 259-265.
Vega, Andres, et al., “Troubleshooting Cisco Application Centric Infrastructure: Analytical problem solving applied to the Policy Driven Data Center,” Feb. 15, 2016, 84 pages.
Xia, Wenfeng, et al., “A Survey on Software-Defined Networking,” IEEE Communications Surveys and Tutorials, Mar. 16, 2015, pp. 27-51.
Akella, Aditya, et al., “A Highly Available Software Defined Fabric,” HotNets-XIII, Oct. 27-28, 2014, Los Angeles, CA, USA, Copyright 2014, ACM, pp. 1-7.
Alsheikh, Mohammad Abu, et al., “Machine Learning in Wireless Sensor Networks: Algorithms, Strategies, and Application,” Mar. 19, 2015, pp. 1-23.
Author Unknown, “Aids to Pro-active Management of Distributed Resources through Dynamic Fault-Localization and Availability Prognosis,” FaultLocalization-TR01-CADlab, May 2006, pp. 1-9.
Author Unknown, “Requirements for applying formal methods to software-defined networking,” Telecommunication Standardization Sector of ITU, Series Y: Global Information Infrastructure, Internet Protocol Aspects and Next-Generation Networks, Apr. 8, 2015, pp. 1-20.
Cisco, “Verify Contracts and Rules in the ACI Fabric,” Cisco, Updated Aug. 19, 2016, Document ID: 119023, pp. 1-20.
de Silva et al., “Network-wide Security Analysis,” Semantic Scholar, Oct. 25, 2011, pp. 1-11.
Dhawan, Mohan, et al., “SPHINX: Detecting Security Attacks in Software-Defined Networks,” NDSS 2015, Feb. 8-11, 2015, San Diego, CA, USA, Copyright 2015 Internet Society, pp. 1-15.
Fayaz, Seyed K., et al., “Efficient Network Reachability Analysis using a Succinct Control Plane Representation,” 2016, ratul.org, pp. 1-16.
Feldmann, Anja, et al., “IP Network Configuration for Intradomain Traffic Engineering,” Semantic Scholar, accessed on Jul. 20, 2017, pp. 1-27.
Han, Yoonseon, et al., “An Intent-based Network Virtualization Platform for SDN,” 2016 I FIP, pp. 1-6.
Han, Wonkyu, et al., “LPM: Layered Policy Management for Software-Defined Networks,” Mar. 8, 2016, pp. 1-8.
Kazemian, Peyman, et al., “Real Time Network Policy Checking using Header Space Analysis,” USENIX Association, 10th USENIX Symposium on Networked Systems Design and Implementation (NSDI '13) pp. 99-111.
Khatkar, Pankaj Kumar, “Firewall Rule Set Analysis and Visualization, A Thesis Presented in Partial Fulfillment of the Requirements for the Degree Master of Science,” Arizona State University, Dec. 2014, pp. 1-58.
Le, Franck, et al., “Minerals: Using Data Mining to Detect Router Misconfigurations,” CyLab, Carnegie Mellon University, CMU-CyLab-06-008, May 23, 2006, pp. 1-14.
Liang, Chieh-Jan Mike, et al., “SIFT: Building an Internet of Safe Things,” Microsoft, IPSN' 15, Apr. 14-16, 2015, Seattle, WA, ACM 978, pp. 1-12.
Lindem, A., et al., “Network Device YANG Organizational Model draft-rtgyangdt-rtgwg-device-model-01,” Network Working Group, Internet—draft, Sep. 21, 2015, pp. 1-33.
Liu, Jason, et al., “A Real-Time Network Simulation Infrastracture Based on Open VPN,” Journal of Systems and Software, Aug. 4, 2008, pp. 1-45.
Lopes, Nuno P., et al., “Automatically verifying reachability and well-formedness in P4 Networks,” Microsoft, accessed on Jul. 18, 2017, pp. 1-13.
Mai, Haohui, et al., “Debugging the Data Plane with Anteater,” SIGCOMM11, Aug. 15-19, 2011, pp. 1-12.
Miller, Nancy, et al., “Collecting Network Status Information for Network-Aware Applications,” INFOCOM 2000, pp. 1-10.
Miranda, Joao Sales Henriques, “Fault Isolation in Software Defined Networks,” www.gsd.inescid.pt, pp. 1-10.
Moon, Daekyeong, et al., “Bridging the Software/Hardware Forwarding Divide,” Berkeley.edu, Dec. 18, 2010, pp. 1-15.
Panda, Aurojit, et al., “SCL: Simplifying Distributed SDN Control Planes,” people.eecs.berkeley.edu, Mar. 2017, pp. 1-17.
Shin, Seugwon, et al., “FRESCO: Modular Composable Security Services for Software-Defined Networks,” To appear in the ISOC Network and Distributed System Security Symposium, Feb. 2013, pp. 1-16.
Shukla, Apoorv, et al., “Towards meticulous data plane monitoring,” kaust.edu.sa, access on Aug. 1, 2017, pp. 1-2.
Tang, Yongning, et al., “Automatic belief network modeling via policy inference for SDN fault localization,” Journal of Internet Services and Applications, 2016, pp. 1-13.
Tomar, Kuldeep, et al., “Enhancing Network Security and Performance Using Optimized ACLs,” International Journal in Foundations of Computer Science & Technology (IJFCST), vol. 4, No. 6, Nov. 2014, pp. 25-35.
Tongaonkar, Alok, et al., “Inferring Higher Level Policies from Firewall Rules,” Proceedings of the 21st Large Installation System Administration Conference (LISA '07), Nov. 11-16, 2007, pp. 1-14.
Yu et al., “A Flexible Framework for Wireless-Based Intelligent Sensor with Reconfigurability, Dynamic adding, and Web interface,” Conference Paper, Jul. 24, 2006, IEEE 2006, pp. 1-7.
Zhou, Shijie, et al., “High-Performance Packet Classification on GPU,” 2014 IEEE, pp. 1-6.
Related Publications (1)
Number Date Country
20180365061 A1 Dec 2018 US
Provisional Applications (1)
Number Date Country
62521685 Jun 2017 US