The subject disclosure relates to the identification of network device characteristics and in particular, an automatic process for resolving network device cluster associations based on comparisons between network device and cluster characteristics.
An enterprise application is a set of workloads (e.g., computing, networking, and storage services) that are generally distributed across various nodes (or endpoints) of a network and the relationships (e.g., connectivity, dependencies, network and security policies, etc.) between the workloads. An application can include a presentation tier, an application tier, and a data tier. The presentation tier may depend on the application tier and authentication services, and the application tier may depend on the web tier and external network services (e.g., a travel reservation system, an ordering tool, a billing tool, etc.). These tiers may further depend on firewall, load balancing, wide area network (WAN) acceleration, and other network services. An enterprise can include hundreds or thousands of applications of similar and different architectures.
Certain features of the subject technology are set forth in the appended claims. However, the accompanying drawings, which are included to provide further understanding, illustrate disclosed aspects and together with the description serve to explain the principles of the subject technology. In the drawings:
The detailed description set forth below is intended as a description of various configurations of the subject technology and is not intended to represent the only configurations in which the technology can be practiced. The appended drawings are incorporated herein and constitute a part of the detailed description. The detailed description includes specific details for the purpose of providing a more thorough understanding of the technology; however, it will be clear and apparent that the subject technology is not limited to the specific details set forth herein and may be practiced without these details. In some instances, structures and components are shown in block diagram form in order to avoid obscuring certain concepts.
A thorough understanding of applications running in a data center can be critical for network management tasks such as anomaly detection (e.g., network attacks and misconfiguration), asset management (e.g., monitoring, capacity planning, consolidation, migration, and continuity planning), and compliance (e.g. conformance with governmental regulations, industry standards, and corporate policies). Despite the workload relationship complexities discussed above, traditional approaches for developing insights into an enterprise’s workloads require comprehensive knowledge on the part of human operators and the use of processes that are manual and largely customized for a particular enterprise. To the extent automated solutions exist conventional techniques for application dependency mapping (ADM) suffer from various drawbacks.
Conventional automated ADM techniques often assume stationary locations for workloads, which is an invalid assumption for many networks. For example, many enterprises implement virtualization platforms that support migration of VMs and containers, such as for optimizing distribution of workloads or for failover handling. Additionally, conventional ADM processes rely on static and structured feature vectors or signatures (i.e., bases of comparison) for discovering applications and application dependency maps. In addition, conventional ADM systems are often incapable of or fail to gather all relevant information for identifying workloads and their interdependencies.
Networks can be configured to automate discovery of applications and map the applications’ interdependencies. Using sensors deployed on (or attached to) various networking devices (e.g., routers, switches, network appliances), physical servers, hypervisors or shared kernels, and virtual partitions (e.g., VMs or containers), and/or other network elements, etc., information regarding device characteristics can be collected. Using networks of such sensors, information of traffic passing through the network can be aggregated and analyzed, providing multiple perspectives for the traffic.
In some aspects, network device information can be used to automatically form endpoint clusters. For example, endpoint characteristics can be mapped into feature vectors representing multiple various characteristic dimensions for a given node. Feature vectors representing nodes or endpoints (e.g., physical servers, VMs, or containers), can be used to assess similarities between endpoints, and to facilitate the automatic grouping of endpoints into clusters. Therefore, cluster characteristics can represent common endpoint characteristics shared by member endpoints (network devices), and can be used to identify functional device groupings, or characteristics that make up a tier or component of an application, etc. Additional details regarding endpoint cluster formation are provided by U.S. Application No. 15/145,493, entitled, “CLUSTER DISCOVERY VIA A MULTI-DOMAIN FUSION FOR APPLICATION DEPENDENCY MAPPING,” which is herein incorporated by reference in its entirety.
Although clustering techniques can be used to automate cluster creation, the dynamic nature of large enterprise networks makes it difficult to maintain static cluster characteristics. For example, initial cluster formation can be based on shared characteristics of endpoint members; however, as network and endpoint characteristics change, individual members begin to take on different properties. As a result, static clusters often result in a collection of member device that share very little in common. These difference can make it difficult to perform new cluster assignments, and can frustrate the administration of cluster based policies, e.g., by system administrators.
Aspects of the disclosed technology address the foregoing limitations of conventional clustering techniques, by providing systems and methods for automatically managing cluster membership for various endpoint devices. In some implementations, the disclosed technology provides a process of endpoint cluster management that include steps for: automatically identifying one or more device characteristics associated with a first network device, automatically identifying one or more cluster characteristics for each of a first cluster and a second cluster, and automatically comparing the device characteristics associated with the first network device with the one or more cluster characteristics for the first cluster and the second cluster. In some approaches, the process can further include steps for: automatically adding the first network device to the first cluster based on the cluster characteristics for the first cluster and the device characteristics for the first network device, automatically detecting a change to one or more device characteristics associated with a second network device, wherein the second network device is a member of the second cluster, and automatically removing the second network device from the second cluster based on changes to the one or more device characteristics associated with the second network device.
In some aspects, the processors can be further configured to perform operations for automatically detecting a change to one or more device characteristics associated with a second network device, wherein the second network device is a member of the second cluster, and automatically removing the second network device from the second cluster based on changes to the one or more device characteristics associated with the second network device.
In some aspects, the processors can be further configured to perform operations for automatically generating a first query associated with the first cluster, the first query comprising one or more labels that describe at least one of the cluster characteristics for the first cluster. In some implementations, the first query is user selectable Boolean expression that uniquely identifies the first cluster. In some aspects, the device characteristics associated with the first network device include one or more of: names, labels, metadata tags, or key value pairs. In some aspects, automatically adding the first network device to the first cluster based on the cluster characteristics is performed using a machine-learning model. In some aspects, identifying the one or more cluster characteristics for the first cluster and the second cluster is automatically performed at periodic intervals.
Endpoint cluster management techniques of the disclosed technology provide ways to sort newly instantiated endpoint devices into pre-existing clusters, and/or to manage endpoint membership in existing clusters. That is, in some aspects, the disclosed technology provides methods for identifying when endpoint or cluster characteristics have changed, and then moving endpoints in to/out of various clusters to ensure that cluster characteristics for any given cluster are sufficiently homogenous. Additionally, the disclosed technology provides automated solutions for generating user selectable queries that are composed of terms describing characteristics of a given cluster or endpoint. That is, as used herein, query can refer to a Boolean expression identifies a specific cluster. Queries can be designed to facilitate the search, identification, and selection of a given cluster, e.g., by system administrator or other user.
Referring now to the drawings,
Configuration manager 102 can be used to provision and maintain sensors 104, including installing sensor software or firmware in various network nodes, configuring sensors 104, updating the sensor software or firmware, among other sensor management tasks. For example, sensors 104 can be implemented as virtual partition images, e.g., virtual machine (VM) images or container images, and configuration manager 102 can distribute the images to host machines. The software environment may include an operating system and application software. For software running within a virtual partition, the virtual partition may appear to be, for example, one of many servers or one of many operating systems executed on a single physical server. Configuration manager 102 can instantiate a new virtual partition or migrate an existing partition to a different physical server. Configuration manager 102 can also be used to configure the new or migrated sensor.
Configuration manager 102 can also monitor the health of sensors 104. For example, configuration manager 102 can request status updates and/or receive heartbeat messages, initiate performance tests, generate health checks, and/or perform other monitoring tasks. In some embodiments, configuration manager 102 can also authenticate sensors 104. For instance, sensors 104 can be assigned a unique identifier, such as by using a one-way hash function of a sensor’s basic input/out system (BIOS) universally unique identifier (UUID) and a secret key stored by configuration image manager 102.
Sensors 104 can reside on various nodes of a network, such as a virtual partition (e.g., VM or container) 120; a hypervisor or shared kernel managing one or more virtual partitions and/or physical servers 122, an application-specific integrated circuit (ASIC) 124 of a switch, router, gateway, or other networking device, or a packet capture (pcap) 126 appliance (e.g., a standalone packet monitor, a device connected to a network devices monitoring port, a device connected in series along a main trunk of a datacenter, or similar device), or other element of a network. Sensors 104 can monitor network traffic between nodes, and send network traffic data and corresponding data (e.g., host data, process data, user data, etc.) to collectors 108 for storage. For example, sensors 104 can sniff packets being sent over its hosts’ physical or virtual network interface card (NIC), or individual processes can be configured to report network traffic and corresponding data to sensors 104.
Incorporating sensors 104 on multiple nodes and within multiple partitions of some nodes of the network can provide for robust capture of network traffic and corresponding data from each hop of data transmission. In some embodiments, each node of the network (e.g., VM, container, or other virtual partition 120, hypervisor, shared kernel, or physical server 122, ASIC 124, pcap 126, etc.) includes a respective sensor 104. However, it should be understood that various software and hardware configurations can be used to implement sensor network 104.
As sensors 104 capture communications and corresponding data, they may continuously send network traffic data to collectors 108. Network traffic data can include metadata relating to a packet, a collection of packets, a flow, a bi-directional flow, a group of flows, a session, or a network communication of another granularity. That is, the network traffic data can generally include any information describing communication on all layers of the Open Systems Interconnection (OSI) model. For example, the network traffic data can include source/destination MAC address, source/destination IP address, protocol, port number, etc. In some embodiments, the network traffic data can also include summaries of network activity or other network statistics such as number of packets, number of bytes, number of flows, bandwidth usage, response time, latency, packet loss, jitter, and other network statistics.
Sensors 104 can also determine additional data for each session, bidirectional flow, flow, packet, or other more granular or less granular network communication. The additional data can include host and/or endpoint information, virtual partition information, sensor information, process information, user information, tenant information, application information, network topology, application dependency mapping, cluster information, or other information corresponding to each flow.
In some aspects, sensors 104 can also perform preprocessing of the network traffic and corresponding data before sending the data to collectors 108. For example, sensors 104 can remove extraneous or duplicative data and/or create summaries of the data, e.g., latency, number of packets per flow, number of bytes per flow, number of flows, etc. In some embodiments, sensors 104 can be configured to only capture certain types of network information and disregard the rest. By way of example, sensors 104 can be configured to capture only a representative sample of packets (e.g., every 1,000 th packet or other suitable sample rate) and corresponding data.
Since sensors 104 can be located throughout the network, traffic and corresponding data can be collected from multiple vantage points or network perspectives to provide a more comprehensive view of network behavior. High fidelity data capture permits the data to be correlated from the various data sources, which may be used as additional data points by analytics engine 110. Further, collecting network traffic and corresponding data from multiple points of view ensures more accurate data is captured. For example, a conventional sensor network may be limited to sensors running on external-facing network devices (e.g., routers, switches, network appliances, etc.) such that east-west traffic, including VM-to-VM or container-to-container traffic on a same host may not be monitored.
Network traffic monitoring system 100 can be configured to assemble the most accurate data flow sets and corresponding sensor data. Sensors 104 can send network traffic and corresponding data to the collectors 106. Collectors 106 can be any type of storage medium that can serve as a repository for the network traffic and corresponding data captured by sensors 104. In some embodiments, data storage for collectors 106 can be located in an in-memory database, such as dashDB from IBM®, although it should be appreciated that the data storage for collectors 106 can be any software and/or hardware capable of providing rapid random access speeds typically used for analytics software. For example, the collectors 106 can utilize solid state drives, disk drives, magnetic tape drives, or a combination of the foregoing according to cost, responsiveness, and size requirements. Further, collectors 106 can utilize various database structures such as a normalized relational database or a NoSQL database, among others.
In some deployments, collectors 106 may only serve as network storage for the network traffic monitoring system 100. In such embodiments, network traffic monitoring system 100 can include a data mover module 108 for retrieving data from collectors 106 and making the data available to network clients, such as the components of analytics engine 110. In effect, data mover module 108 can serve as a gateway for presenting network-attached storage to the network clients.
Analytics engine 110 may be provided with examples of network states corresponding to an attack and network states corresponding to normal operation. The analytics engine 110 can then analyze network traffic and corresponding data to recognize when the network is under attack. In some embodiments, the network may operate within a trusted environment for a period of time so that the analytics engine 110 can establish a baseline of normal operation. Since malware is constantly evolving and changing, machine learning may be used to dynamically update models for identifying malicious traffic patterns.
In some embodiments, analytics engine 110 may be used to identify observations which differ from other examples in a dataset. For example, if a training set of example data with known outlier labels exists, supervised anomaly detection techniques may be used. Additionally, analytics engine 110 can include a data lake 130, an application dependency mapping (ADM) module 140, and elastic processing engines 150.
Data lake 130 is a large-scale storage repository that provides massive storage for various types of data, enormous processing power, and the ability to handle a nearly limitless number of concurrent tasks or jobs. In some embodiments, data lake 130 is implemented using the Hadoop® Distributed File System (HDFS™) from Apache® Software Foundation. HDFS™ is a highly scalable and distributed file system that can scale to thousands of cluster nodes, millions of files, and petabytes of data. HDFS™ is optimized for batch processing where data locations are exposed to allow computations to take place where the data resides. HDFS™ provides a single namespace for an entire cluster to allow for data coherency in a write-once, read-many access model. That is, clients can only append to existing files in the node. In HDFS™, files are separated into blocks, which are typically 64 MB in size and are replicated in multiple data nodes.
Clients access data directly from data nodes. In some embodiments, data mover 108 receives raw network traffic and corresponding data from the collectors 106 and distributes or pushes the data to the data lake 130. Data lake 130 can also receive and store out-of-band data 114, such as statuses on power levels, network availability, server performance, temperature conditions, cage door positions, and other data from internal sources, and third party data 116, such as security reports (e.g., provided by Cisco® Systems, Inc. of San Jose, California, Arbor Networks® of Burlington, Massachusetts, Symantec® Corp. of Sunnyvale, California, Sophos® Group plc of Abingdon, England, Microsoft® Corp. of Seattle, Washington, Verizon® Communications, Inc. of New York, New York, among others), geolocation data, IP watch lists, Whois data, configuration management database (CMDB) or configuration management system (CMS) as a service, and other data from external sources.
Each component of the data lake 130 can perform certain processing of the raw network traffic data and/or other data (e.g., host data, process data, user data, out-of-band data or third party data) to transform the raw data to a form useable by elastic processing engines 150. In some embodiments, the data lake 130 can include repositories for flow attributes 132, host and/or endpoint attributes 134, process attributes 136, and policy attributes 138. In some embodiments, the data lake 130 can also include repositories for VM or container attributes, application attributes, tenant attributes, network topology, application dependency maps, cluster attributes, etc.
Flow attributes 132 relate to information about flows traversing the network. A flow is generally one or more packets sharing certain attributes that are sent within a network within a specified period of time. Flow attributes 132 can include packet header fields, including but not limited to: a source address (e.g., Internet Protocol (IP) address, Media Access Control (MAC) address, Domain Name System (DNS) name, or other network address), source port, destination address, destination port, protocol type, class of service, among other fields. The source address may correspond to a first endpoint (e.g., network device, physical server, virtual partition, etc.) of the network, and the destination address may correspond to a second endpoint, a multicast group, or a broadcast domain. Flow attributes 132 can also include aggregate packet data such as flow start time, flow end time, number of packets for a flow, number of bytes for a flow, the union of TCP flags for a flow, among other flow data.
Host and/or endpoint attributes 134 describe host and/or endpoint data for each flow, and can include host and/or endpoint name, network address, operating system, CPU usage, network usage, disk space, ports, logged users, scheduled jobs, open files, and information regarding files and/or directories stored on a host and/or endpoint (e.g., presence, absence, or modifications of log files, configuration files, device special files, or protected electronic information). Host and/or endpoints attributes 134 can also include the out-of-band data 114 regarding hosts such as power level, temperature, and physical location (e.g., room, row, rack, cage door position, etc.) or third party data 116 such as whether a host and/or endpoint is on an IP watch list or otherwise associated with a security threat, Whois data, or geo-coordinates.
In some embodiments, host and/or endpoint attributes or characteristics create tags or labels to describe network device or cluster characteristics. By way of the above examples, tags can be used to identify hosts, e.g., by Domain Name System (DNS) name, IP address, Media Access Control (MAC) address or other network address, source port, protocol type, class of service, virtual partition information, process information, operating system (OS) type, tenant information, application information, network topology characteristics, cluster membership, and/or application dependencies, etc. As discussed in further detail below, tags can be represented by key-value pairs, such as, by using Boolean expressions representing concatenations of multiple tags.
Process attributes 136 relate to process data corresponding to each flow, and can include process name (e.g., bash, httpd, netstat, etc.), ID, parent process ID, path (e.g., /usr2/username/bin/, /usr/local/bin, /usr/bin, etc.), CPU utilization, memory utilization, memory address, scheduling information, nice value, flags, priority, status, start time, terminal type, CPU time taken by the process, the command that started the process, and information regarding a process owner (e.g., user name, ID, user’s real name, e-mail address, user’s groups, terminal information, login time, expiration date of login, idle time, and information regarding files and/or directories of the user).
Policy attributes 138 contain information relating to network policies. Policies establish whether a particular flow is allowed or denied by the network as well as a specific route by which a packet traverses the network. Policies can also be used to mark packets so that certain kinds of traffic receive differentiated service when used in combination with queuing techniques such as those based on priority, fairness, weighted fairness, token bucket, random early detection, round robin, among others.
Analytics engine 110 may include any number of engines 150, including for example, a flow engine 152 for identifying flows (e.g., flow engine 152) or an attacks engine 154 for identify attacks to the network. In some embodiments, the analytics engine can include a separate distributed denial of service (DDoS) attack engine 155 for specifically detecting DDoS attacks. The analytics engine 110 may further include a search engine 156. The search engine 156 may be configured, for example to perform a structured search, an NLP (Natural Language Processing) search, or a visual search. Data may be provided to the engines from one or more processing components. Analytics engine 110 can also include a policy engine 158 that manages network policy, including creating and/or importing policies, monitoring policy conformance and non-conformance, enforcing policy, simulating changes to policy or network elements affecting policy, among other policy-related tasks.
The presentation module 112 can include an application programming interface (API) or command line interface (CLI) 160, a security information and event management (SIEM) interface 162, and a web front-end 164. As the analytics engine 110 processes network traffic and corresponding data and generates analytics data, the analytics data may not be in a human-readable form or it may be too voluminous for a user to navigate. The presentation module 112 can take the analytics data generated by analytics engine 110 and further summarize, filter, and organize the analytics data as well as create intuitive presentations for the analytics data.
ADM module 140 can determine dependencies of applications of the network. That is, particular patterns of traffic may correspond to an application, and the interconnectivity or dependencies of the application can be mapped to generate a graph for the application (i.e., an application dependency mapping). In this context, an application refers to a set of networking components that provides connectivity for a given set of workloads. For example, in a conventional three-tier architecture for a web application, first endpoints of the web tier, second endpoints of the application tier, and third endpoints of the data tier make up the web application. ADM module 140 can receive input data from various repositories of the data lake 130, e.g., the flow attributes 132, the host and/or endpoint attributes 134, the process attributes 136, etc.
As discussed in further detail with respect to
Cluster grouping and management can be administered using a machine learning (ML) model, such as a classifier comprising a multilayer perceptron neural network. Machine learning solutions for matching network device characteristics with cluster characteristics help automate the process of adding (and removing) endpoint members from a cluster, and ensuring that devices sharing a common cluster also have sufficiently similar characteristic vectors. In some approaches, machine learning implementations can be used to facilitate various portions of endpoint cluster creation, update/management, and query generation. For example, machine learning (ML) techniques may be applied to the collected data to evaluate and determine optimal features across multiple domains for discovering application tendencies, and for creating clusters. Such techniques can also be used to evaluate ongoing changes to device characteristics as the network evolves, and to re-form clusters based on the updated characteristics.
Although it is understood that the described techniques can be implemented using a variety of machine learning and/or classification algorithms, the scope of the technology is not limited to a specific machine learning implementation. Implementations of the technology can include the deployment of multi-layered ML models based on one or more classification algorithms, including but not limited to: a Multinomial Naive Bayes classifier, a Bernoulli Naive Bayes classifier, a Perceptron classifier, a Stochastic Gradient Descent (SGD) Classifier, and/or a Passive Aggressive Classifier, or the like.
In some aspects, ML models can be configured to perform various types of regression, for example, using one or more regression algorithms, including but not limited to: Deep Learning, a Stochastic Gradient Descent Regressor, and/or a Passive Aggressive Regressor, etc. ML models can also be based on clustering algorithms (e.g., a Mini-batch K-means clustering algorithm), a recommendation algorithm (e.g., a Miniwise Hashing algorithm, or Euclidean LSH algorithm), and/or an anomaly detection algorithm, such as a Local outlier factor. Additionally, ML models can employ a dimensionality reduction approach, such as, one or more of: a Mini-batch Dictionary Learning algorithm, an Incremental Principal Component Analysis (PCA) algorithm, a Latent Dirichlet Allocation algorithm, and/or a Mini-batch K-means algorithm, etc.
Additionally, in some aspects, device characteristics may be represented by one or more key value pairs (KVPs), i.e., a set of two linked data items: a unique key that identifies an item of data (e.g., a device characteristic), and the value, which is either the data that is identified or a pointer to the location of that data.
Although
Network environment 200 can include a network fabric 202, a Layer 2 (L2) network 204, a Layer 3 (L3) network 206, and servers 208a, 208b, 208c, 208d, and 208e (collectively, 208). The network fabric 202 can include spine switches 210a, 210b, 210c, and 210d (collectively, “210”) and leaf switches 212a, 212b, 212c, 212d, and 212e (collectively, “212”). The spine switches 210 can connect to the leaf switches 212 in the network fabric 202. The leaf switches 212 can include access ports (or non-fabric ports) and fabric ports. The fabric ports can provide uplinks to the spine switches 210, while the access ports can provide connectivity to endpoints (e.g., the servers 208), internal networks (e.g., the L2 network 204), or external networks (e.g., the L3 network 206).
Leaf switches 212 can reside at the edge of the network fabric 202, and can thus represent the physical network edge. For instance, in some embodiments, the leaf switches 212d and 212e operate as border leaf switches in communication with edge devices 214 located in the external network 206. The border leaf switches 212d and 212e may be used to connect any type of external network device, service (e.g., firewall, deep packet inspector, traffic monitor, load balancer, etc.), or network (e.g., the L3 network 206) to the fabric 202.
Although the network fabric 202 is illustrated and described herein as an example leaf-spine architecture, one of ordinary skill in the art will readily recognize that various embodiments can be implemented based on any network topology, including any data center or cloud network fabric. Indeed, other architectures, designs, infrastructures, and variations are contemplated herein. For example, the principles disclosed herein are applicable to topologies including three-tier (including core, aggregation, and access levels), fat tree, mesh, bus, hub and spoke, etc. Thus, in some embodiments, the leaf switches 212 can be top-of-rack switches configured according to a top-of-rack architecture. In other embodiments, the leaf switches 212 can be aggregation switches in any particular topology, such as end-of-row or middle-of-row topologies. In some embodiments, the leaf switches 212 can also be implemented using aggregation switches.
Moreover, it is understood that the topology illustrated in
Network communications in the network fabric 202 can flow through the leaf switches 212. In some embodiments, the leaf switches 212 can provide endpoints (e.g., the servers 208), internal networks (e.g., the L2 network 204), or external networks (e.g., the L3 network 206) access to the network fabric 202, and can connect the leaf switches 212 to each other. In some embodiments, the leaf switches 212 can connect endpoint groups (EPGs) to the network fabric 202, internal networks (e.g., the L2 network 204), and/or any external networks (e.g., the L3 network 206). EPGs are groupings of applications, or application components, and tiers for implementing forwarding and policy logic. EPGs can allow for separation of network policy, security, and forwarding from addressing by using logical application boundaries. EPGs can be used in the network environment 200 for mapping applications in the network. For example, EPGs can comprise a grouping of endpoints in the network indicating connectivity and policy for applications.
As discussed, the servers 208 can connect to the network fabric 202 via the leaf switches 212. For example, the servers 208a and 208b can connect directly to the leaf switches 212a and 212b, which can connect the servers 208a and 208b to the network fabric 202 and/or any of the other leaf switches. The servers 208c and 208d can connect to the leaf switches 212b and 212c via the L2 network 204. The servers 208c and 208d and the L2 network 204 make up a local area network (LAN). LANs can connect nodes over dedicated private communications links located in the same general physical location, such as a building or campus.
WAN 206 can connect to the leaf switches 212d or 212e via the L3 network 206. WANs can connect geographically dispersed nodes over long-distance communications links, such as common carrier telephone lines, optical light paths, synchronous optical networks (SONET), or synchronous digital hierarchy (SDH) links. LANs and WANs can include L2 and/or L3 networks and endpoints.
The Internet is an example of a WAN that connects disparate networks throughout the world, providing global communication between nodes on various networks. The nodes typically communicate over the network by exchanging discrete frames or packets of data according to predefined protocols, such as the Transmission Control Protocol/Internet Protocol (TCP/IP). In this context, a protocol can refer to a set of rules defining how the nodes interact with each other. Computer networks may be further interconnected by an intermediate network node, such as a router, to extend the effective size of each network. The endpoints 208 can include any communication device or component, such as a computer, server, blade, hypervisor, virtual machine, container, process (e.g., running on a virtual machine), switch, router, gateway, host, device, external network, etc.
In some embodiments, the network environment 200 also includes a network controller running on the host 208a. The network controller is implemented using the Application Policy Infrastructure Controller (APIC™) from Cisco®. The APIC™ provides a centralized point of automation and management, policy programming, application deployment, and health monitoring for the fabric 202. In some embodiments, the APIC™ is operated as a replicated synchronized clustered controller. In other embodiments, other configurations or software-defined networking (SDN) platforms can be utilized for managing the fabric 202.
In some embodiments, a physical server 208 may have instantiated thereon a hypervisor 216 for creating and running one or more virtual switches (not shown) and one or more virtual machines 218, as shown for the host 208b. In other embodiments, physical servers may run a shared kernel for hosting containers. In yet other embodiments, the physical server 208 can run other software for supporting other virtual partitioning approaches. Networks in accordance with various embodiments may include any number of physical servers hosting any number of virtual machines, containers, or other virtual partitions. Hosts may also comprise blade/physical servers without virtual machines, containers, or other virtual partitions, such as the servers 208a, 208c, 208d, and 208e.
Network environment 200 can also integrate a network traffic monitoring system, such as the network traffic monitoring system 100 shown in
In some embodiments, sensors 220 can be located at each (or nearly every) network component to capture granular packet statistics and data at each hop of data transmission. In other embodiments, the sensors 220 may not be installed in all components or portions of the network (e.g., shared hosting environment in which customers have exclusive control of some virtual machines). As shown in
As illustrated, ADM engine 302 includes clustering module 303 and query generation module 305. Clustering module 303 can be configured to receive network updates and/or monitor network traffic to detect changes in network device and/or cluster characteristics, and to perform automatic clustering updates/assignments. As mentioned above, and discussed in further detail below with respect to
Outputs from clustering module 303 are then provided to query generation model 305, which is configured to construct queries uniquely identifying various endpoint groups. Queries can include strings of endpoint characteristics such as characteristics representing shared attributes of all endpoints in a particular cluster. As discussed above, tags can be used to identify hosts (e.g., by name, and/or IP address), virtual partition information, process information, operating system (OS) information, tenant information, application information, network topology characteristics, cluster membership, and/or application dependencies, and the like.
In some implementations, query generation is informed by user input. For example, user 310 can manually input/select tags to be associated with a given cluster (and therefore query), which can then be provided to query generation module 305 via inventory module 308. In some aspects, user editing of cluster tags can cause the manual inclusion (or removal) of one or more endpoints from a given cluster. That is, user inputs may be used to override endpoint classification decisions made by the machine learning model. A more detailed description of an automated endpoint clustering management process is discussed with respect to
Specifically,
In some aspects, the first network device may be a new device, such as a physical host or newly instantiated virtual machine or container process. In other aspects, the first network device may be a legacy device with changing characteristics, for example, which changes to one or more attributes associated with a label/tag.
Next, one or more cluster characteristics are identified for each of a variety of pre-existing endpoint clusters (step 404). By way of example, cluster characteristics can be identified for at least 2 different clusters, i.e., a first cluster and a second cluster representing pre-existing endpoint groups. In some approaches, cluster characteristics for the respective first cluster and the second cluster represent common characteristics shared by their respective endpoint members. Cluster characteristics for the first cluster and the second cluster can be represented by tags as well as queries that uniquely identify each cluster based on its characteristics and/or shared characteristics of the associated member devices.
Once characteristics have been identified for the pre-existing clusters, process 400 proceeds to step 406 in which device characteristics associated with the first network device are compared with the cluster characteristics for the first cluster and the second cluster. Comparisons between device characteristics and cluster characteristics can be performed using a machine learning approach. For example, device characteristics of the first network device may be represented by a vector that is provided as an input to machine learning classifier configured to evaluate a distance of the input vector from pre-existing cluster characteristic vectors in a (characteristic) vector space.
After the machine learning classifier determines the closest cluster characteristic vector to the device characteristic vector (input), the first network device is automatically added to the corresponding cluster. By way of example first network device can be automatically added to the first cluster based on similarities between cluster characteristics for the first cluster and the device characteristics for the first network device.
As discussed above, automated cluster management techniques can also be applied for endpoint/host network devices that are already assigned to a cluster. For example, changing device characteristics may result in a scenario where a given network device’s characteristics are significantly different from that of other devices in the same cluster/group. In such instances, the device may be removed from a group and/or re-assigned to another cluster.
In other examples, a newly formed cluster may provide characteristics that more closely match those of devices belonging to another group. Such changes can also trigger cluster updates. As understood by those of skill in the art, the cluster update process may be triggered by a variety of factors, including but not limited to: detected changes in the network, such as, detected changes to one or more network devices, existing clusters, network policies, network topology, and the like. Additionally, cluster updates may be performed on a predetermined schedule such as at periodic time intervals i.e., every 1 minute, every 5 minutes, or every day at a certain pre-specified time e.g. at midnight.
In some implementations, changes to cluster membership can trigger changes to associated queries, for example, that provide labels/tags representing characteristics of the associated cluster. By automatically updating queries, the system provides a convenient and automatically managed process that enables users/system administrators to identify and select a given cluster based on its associated query. As discussed above, query management may be facilitated by a graphical user interface (GUI) that enables the user to easily view, search, and select queries associated with different network/device characteristics.
Interfaces 502 are typically provided as modular interface cards (sometimes referred to as “line cards”). Generally, they control the sending and receiving of data packets over the network and sometimes support other peripherals used with the network device 500. Among the interfaces that may be provided are Ethernet interfaces, frame relay interfaces, cable interfaces, DSL interfaces, token ring interfaces, and the like. In addition, various very high-speed interfaces may be provided such as fast token ring interfaces, wireless interfaces, Ethernet interfaces, Gigabit Ethernet interfaces, ATM interfaces, HSSI interfaces, POS interfaces, FDDI interfaces, WIFI interfaces, 3G/4G/5G cellular interfaces, CAN BUS, LoRA, and the like. Generally, these interfaces may include ports appropriate for communication with the appropriate media. In some cases, they may also include an independent processor and, in some instances, volatile RAM. The independent processors may control such communications intensive tasks as packet switching, media control, signal processing, crypto processing, and management. By providing separate processors for the communications intensive tasks, these interfaces allow the master microprocessor 504 to efficiently perform routing computations, network diagnostics, security functions, etc.
Although the system shown in
Regardless of the network device’s configuration, it may employ one or more memories or memory modules (including memory 506) configured to store program instructions for the general-purpose network operations and mechanisms for roaming, route optimization and routing functions described herein. The program instructions may control the operation of an operating system and/or one or more applications, for example. The memory or memories may also be configured to store tables such as mobility binding, registration, and association tables, etc. Memory 506 could also hold various software containers and virtualized execution environments and data.
The network device 500 can also include an application-specific integrated circuit (ASIC), which can be configured to perform routing and/or switching operations. The ASIC can communicate with other components in the network device 500 via the bus 510, to exchange data and signals and coordinate various types of operations by the network device 500, such as routing, switching, and/or data storage operations, for example.
For clarity of explanation, in some instances the present technology may be presented as including individual functional blocks including functional blocks comprising devices, device components, steps or routines in a method embodied in software, or combinations of hardware and software.
In some embodiments the computer-readable storage devices, mediums, and memories can include a cable or wireless signal containing a bit stream and the like. However, when mentioned, non-transitory computer-readable storage media expressly exclude media such as energy, carrier signals, electromagnetic waves, and signals per se.
Methods according to the above-described examples can be implemented using computer-executable instructions that are stored or otherwise available from computer readable media. Such instructions can comprise, for example, instructions and data which cause or otherwise configure a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions. Portions of computer resources used can be accessible over a network. The computer executable instructions may be, for example, binaries, intermediate format instructions such as assembly language, firmware, or source code. Examples of computer-readable media that may be used to store instructions, information used, and/or information created during methods according to described examples include magnetic or optical disks, flash memory, USB devices provided with non-volatile memory, networked storage devices, and so on.
Devices implementing methods according to these disclosures can comprise hardware, firmware and/or software, and can take any of a variety of form factors. Typical examples of such form factors include laptops, smart phones, small form factor personal computers, personal digital assistants, rackmount devices, standalone devices, and so on. Functionality described herein also can be embodied in peripherals or add-in cards. Such functionality can also be implemented on a circuit board among different chips or different processes executing in a single device, by way of further example.
The instructions, media for conveying such instructions, computing resources for executing them, and other structures for supporting such computing resources are means for providing the functions described in these disclosures.
Although a variety of examples and other information was used to explain aspects within the scope of the appended claims, no limitation of the claims should be implied based on particular features or arrangements in such examples, as one of ordinary skill would be able to use these examples to derive a wide variety of implementations. Further and although some subject matter may have been described in language specific to examples of structural features and/or method steps, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to these described features or acts. For example, such functionality can be distributed differently or performed in components other than those identified herein. Rather, the described features and steps are disclosed as examples of components of systems and methods within the scope of the appended claims.
Claim language reciting “at least one of” refers to at least one of a set and indicates that one member of the set or multiple members of the set satisfy the claim. For example, claim language reciting “at least one of A and B” means A, B, or A and B.
Reference to “one embodiment” or “an embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the disclosure. The appearances of the phrase “in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. Moreover, various features are described which can be exhibited by some embodiments and not by others.
The terms used in this specification generally have their ordinary meanings in the art, within the context of the disclosure, and in the specific context where each term is used. Alternative language and synonyms can be used for any one or more of the terms discussed herein, and no special significance should be placed upon whether or not a term is elaborated or discussed herein. In some cases, synonyms for certain terms are provided. A recital of one or more synonyms does not exclude the use of other synonyms. The use of examples anywhere in this specification including examples of any terms discussed herein is illustrative only, and is not intended to further limit the scope and meaning of the disclosure or of any example term. Likewise, the disclosure is not limited to various embodiments given in this specification.
Without intent to limit the scope of the disclosure, examples of instruments, apparatus, methods and their related results according to the embodiments of the present disclosure are given below. Note that titles or subtitles can be used in the examples for convenience of a reader, which in no way should limit the scope of the disclosure. Unless otherwise defined, technical and scientific terms used herein have the meaning as commonly understood by one of ordinary skill in the art to which this disclosure pertains. In the case of conflict, the present document, including definitions will control.
Additional features and advantages of the disclosure will be set forth in the description which follows, and in part will be obvious from the description, or can be learned by practice of the herein disclosed principles. The features and advantages of the disclosure can be realized and obtained by means of the instruments and combinations particularly pointed out in the appended claims. These and other features of the disclosure will become more fully apparent from the following description and appended claims, or can be learned by the practice of the principles set forth herein.
This application claims benefit to U.S. Provisional Application No. 62/616,347, filed on Jan. 11, 2018, of which is hereby expressly incorporated by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
5742829 | Davis et al. | Apr 1998 | A |
5903545 | Sabourin et al. | May 1999 | A |
6012096 | Link et al. | Jan 2000 | A |
6144962 | Weinberg et al. | Nov 2000 | A |
6247058 | Miller et al. | Jun 2001 | B1 |
6330562 | Boden et al. | Dec 2001 | B1 |
6525658 | Streetman et al. | Feb 2003 | B2 |
6597663 | Rekhter | Jul 2003 | B1 |
6611896 | Mason, Jr. et al. | Aug 2003 | B1 |
6728779 | Griffin et al. | Apr 2004 | B1 |
6801878 | Hintz et al. | Oct 2004 | B1 |
6847993 | Novaes et al. | Jan 2005 | B1 |
6925490 | Novaes et al. | Aug 2005 | B1 |
6958998 | Shorey | Oct 2005 | B2 |
6983323 | Cantrell et al. | Jan 2006 | B2 |
6996817 | Birum et al. | Feb 2006 | B2 |
7002464 | Bruemmer et al. | Feb 2006 | B2 |
7096368 | Kouznetsov et al. | Aug 2006 | B2 |
7120934 | Ishikawa | Oct 2006 | B2 |
7162643 | Sankaran et al. | Jan 2007 | B1 |
7181769 | Keanini et al. | Feb 2007 | B1 |
7185103 | Jain | Feb 2007 | B1 |
7240106 | Cochran et al. | Jul 2007 | B2 |
7337206 | Wen et al. | Feb 2008 | B1 |
7353511 | Ziese | Apr 2008 | B1 |
7370092 | Aderton et al. | May 2008 | B2 |
7395195 | Suenbuel et al. | Jul 2008 | B2 |
7444404 | Wetherall et al. | Oct 2008 | B2 |
7466681 | Ashwood-Smith et al. | Dec 2008 | B2 |
7467205 | Dempster et al. | Dec 2008 | B1 |
7496040 | Seo | Feb 2009 | B2 |
7496575 | Buccella et al. | Feb 2009 | B2 |
7530105 | Gilbert et al. | May 2009 | B2 |
7610330 | Quinn et al. | Oct 2009 | B1 |
7633942 | Bearden et al. | Dec 2009 | B2 |
7676570 | Levy et al. | Mar 2010 | B2 |
7681131 | Quarterman et al. | Mar 2010 | B1 |
7693947 | Judge et al. | Apr 2010 | B2 |
7752307 | Takara | Jul 2010 | B2 |
7774498 | Kraemer et al. | Aug 2010 | B1 |
7783457 | Cunningham | Aug 2010 | B2 |
7844696 | Labovitz et al. | Nov 2010 | B2 |
7844744 | Abercrombie et al. | Nov 2010 | B2 |
7864707 | Dimitropoulos et al. | Jan 2011 | B2 |
7873025 | Patel et al. | Jan 2011 | B2 |
7874001 | Beck et al. | Jan 2011 | B2 |
7885197 | Metzler | Feb 2011 | B2 |
7895649 | Brook et al. | Feb 2011 | B1 |
7904420 | Ianni | Mar 2011 | B2 |
7930752 | Hertzog et al. | Apr 2011 | B2 |
7934248 | Yehuda et al. | Apr 2011 | B1 |
7957934 | Greifeneder | Jun 2011 | B2 |
7961637 | McBeath | Jun 2011 | B2 |
7970946 | Djabarov et al. | Jun 2011 | B1 |
7975035 | Popescu et al. | Jul 2011 | B2 |
8005935 | Pradhan et al. | Aug 2011 | B2 |
8040232 | Oh et al. | Oct 2011 | B2 |
8040822 | Proulx et al. | Oct 2011 | B2 |
8056134 | Ogilvie | Nov 2011 | B1 |
8135657 | Kapoor et al. | Mar 2012 | B2 |
8156430 | Newman | Apr 2012 | B2 |
8181248 | Oh et al. | May 2012 | B2 |
8185824 | Mitchell et al. | May 2012 | B1 |
8224971 | Miller et al. | Jul 2012 | B1 |
8234660 | Ari et al. | Jul 2012 | B2 |
8250657 | Nachenberg et al. | Aug 2012 | B1 |
8255972 | Azagury et al. | Aug 2012 | B2 |
8260840 | Sirota et al. | Sep 2012 | B1 |
8266697 | Coffman et al. | Sep 2012 | B2 |
8281397 | Vaidyanathan et al. | Oct 2012 | B2 |
8291495 | Burns et al. | Oct 2012 | B1 |
8296847 | Mendonca et al. | Oct 2012 | B2 |
8365286 | Poston | Jan 2013 | B2 |
8370407 | Devarajan et al. | Feb 2013 | B1 |
8381289 | Pereira et al. | Feb 2013 | B1 |
8391270 | Van Der Stok et al. | Mar 2013 | B2 |
8407164 | Malik et al. | Mar 2013 | B2 |
8413235 | Chen et al. | Apr 2013 | B1 |
8442073 | Skubacz et al. | May 2013 | B2 |
8451731 | Lee et al. | May 2013 | B1 |
8462212 | Kundu et al. | Jun 2013 | B1 |
8489765 | Vasseur et al. | Jul 2013 | B2 |
8516590 | Ranadive et al. | Aug 2013 | B1 |
8527977 | Cheng et al. | Sep 2013 | B1 |
8549635 | Muttik et al. | Oct 2013 | B2 |
8570861 | Brandwine et al. | Oct 2013 | B1 |
8572600 | Chung et al. | Oct 2013 | B2 |
8572734 | McConnell et al. | Oct 2013 | B2 |
8572735 | Ghosh et al. | Oct 2013 | B2 |
8572739 | Cruz et al. | Oct 2013 | B1 |
8588081 | Salam et al. | Nov 2013 | B2 |
8600726 | Varshney et al. | Dec 2013 | B1 |
8613084 | Dalcher | Dec 2013 | B2 |
8630316 | Haba | Jan 2014 | B2 |
8631464 | Belakhdar et al. | Jan 2014 | B2 |
8640086 | Bonev et al. | Jan 2014 | B2 |
8661544 | Yen et al. | Feb 2014 | B2 |
8677487 | Balupari et al. | Mar 2014 | B2 |
8683389 | Bar-Yam et al. | Mar 2014 | B1 |
8683554 | Martin et al. | Mar 2014 | B2 |
8706914 | Duchesneau | Apr 2014 | B2 |
8713676 | Pandrangi et al. | Apr 2014 | B2 |
8719452 | Ding et al. | May 2014 | B1 |
8719835 | Kanso et al. | May 2014 | B2 |
8752042 | Ratica | Jun 2014 | B2 |
8752179 | Zaitsev | Jun 2014 | B2 |
8755396 | Sindhu et al. | Jun 2014 | B2 |
8762951 | Kosche et al. | Jun 2014 | B1 |
8769084 | Westerfeld et al. | Jul 2014 | B2 |
8776180 | Kumar et al. | Jul 2014 | B2 |
8812725 | Kulkarni | Aug 2014 | B2 |
8813236 | Saha et al. | Aug 2014 | B1 |
8825848 | Dotan et al. | Sep 2014 | B1 |
8832013 | Adams et al. | Sep 2014 | B1 |
8832461 | Saroiu et al. | Sep 2014 | B2 |
8849926 | Marzencki et al. | Sep 2014 | B2 |
8863252 | Katzer et al. | Oct 2014 | B1 |
8881258 | Paul et al. | Nov 2014 | B2 |
8887238 | Howard et al. | Nov 2014 | B2 |
8904520 | Nachenberg et al. | Dec 2014 | B1 |
8914856 | Velummylum et al. | Dec 2014 | B1 |
8931043 | Cooper et al. | Jan 2015 | B2 |
8934378 | Novick et al. | Jan 2015 | B1 |
8954610 | Berke et al. | Feb 2015 | B2 |
8955124 | Kim et al. | Feb 2015 | B2 |
8966625 | Zuk et al. | Feb 2015 | B1 |
8973147 | Pearcy et al. | Mar 2015 | B2 |
8984331 | Quinn | Mar 2015 | B2 |
8990386 | He et al. | Mar 2015 | B2 |
8996695 | Anderson et al. | Mar 2015 | B2 |
8997227 | Mhatre et al. | Mar 2015 | B1 |
9015716 | Fletcher et al. | Apr 2015 | B2 |
9071575 | Lemaster et al. | Jun 2015 | B2 |
9088598 | Zhang et al. | Jul 2015 | B1 |
9110905 | Polley et al. | Aug 2015 | B2 |
9117075 | Yeh | Aug 2015 | B1 |
9152789 | Natarajan et al. | Oct 2015 | B2 |
9160764 | Stiansen et al. | Oct 2015 | B2 |
9178906 | Chen et al. | Nov 2015 | B1 |
9185127 | Neou et al. | Nov 2015 | B2 |
9191402 | Yan | Nov 2015 | B2 |
9197654 | Ben-Shalom et al. | Nov 2015 | B2 |
9225793 | Dutta et al. | Dec 2015 | B2 |
9237111 | Banavalikar et al. | Jan 2016 | B2 |
9246773 | Degioanni | Jan 2016 | B2 |
9258217 | Duffield et al. | Feb 2016 | B2 |
9281940 | Matsuda et al. | Mar 2016 | B2 |
9294486 | Chiang et al. | Mar 2016 | B1 |
9317574 | Brisebois et al. | Apr 2016 | B1 |
9319384 | Yan et al. | Apr 2016 | B2 |
9369479 | Lin | Jun 2016 | B2 |
9372914 | Asuncion et al. | Jun 2016 | B1 |
9396327 | Auger et al. | Jul 2016 | B2 |
9405903 | Xie et al. | Aug 2016 | B1 |
9413587 | Smith et al. | Aug 2016 | B2 |
9418222 | Rivera et al. | Aug 2016 | B1 |
9454324 | Madhavapeddi | Sep 2016 | B1 |
9461969 | Watt | Oct 2016 | B2 |
9501744 | Brisebois et al. | Nov 2016 | B1 |
9602573 | Abu-Ghazaleh et al. | Mar 2017 | B1 |
9634915 | Bley | Apr 2017 | B2 |
9645892 | Patwardhan | May 2017 | B1 |
9733973 | Prasad et al. | Aug 2017 | B2 |
9781004 | Danait et al. | Oct 2017 | B2 |
9911083 | Chamness et al. | Mar 2018 | B2 |
10031722 | Mutagi et al. | Jul 2018 | B1 |
10069689 | Derby et al. | Sep 2018 | B1 |
10574654 | Schroeder et al. | Feb 2020 | B1 |
20020053033 | Cooper et al. | May 2002 | A1 |
20020103793 | Koller et al. | Aug 2002 | A1 |
20020141343 | Bays | Oct 2002 | A1 |
20020184393 | Leddy et al. | Dec 2002 | A1 |
20030097439 | Strayer et al. | May 2003 | A1 |
20030145232 | Poletto et al. | Jul 2003 | A1 |
20030154399 | Zuk et al. | Aug 2003 | A1 |
20040030776 | Cantrell et al. | Feb 2004 | A1 |
20040167921 | Carson et al. | Aug 2004 | A1 |
20040243533 | Dempster et al. | Dec 2004 | A1 |
20040268149 | Aaron | Dec 2004 | A1 |
20050039104 | Shah et al. | Feb 2005 | A1 |
20050063377 | Bryant et al. | Mar 2005 | A1 |
20050166066 | Ahuja et al. | Jul 2005 | A1 |
20050185621 | Sivakumar et al. | Aug 2005 | A1 |
20050207376 | Ashwood-Smith et al. | Sep 2005 | A1 |
20050257244 | Joly et al. | Nov 2005 | A1 |
20050289244 | Sahu et al. | Dec 2005 | A1 |
20060048218 | Lingafelt et al. | Mar 2006 | A1 |
20060080733 | Khosmood et al. | Apr 2006 | A1 |
20060095968 | Portolani et al. | May 2006 | A1 |
20060156408 | Himberger et al. | Jul 2006 | A1 |
20060195448 | Newport | Aug 2006 | A1 |
20060235972 | Asnis | Oct 2006 | A1 |
20060272018 | Fouant | Nov 2006 | A1 |
20060274659 | Ouderkirk | Dec 2006 | A1 |
20060294219 | Ogawa et al. | Dec 2006 | A1 |
20070044147 | Choi et al. | Feb 2007 | A1 |
20070097976 | Wood et al. | May 2007 | A1 |
20070169179 | Narad | Jul 2007 | A1 |
20070195729 | Li et al. | Aug 2007 | A1 |
20070195797 | Patel et al. | Aug 2007 | A1 |
20070211637 | Mitchell | Sep 2007 | A1 |
20070300061 | Kim et al. | Dec 2007 | A1 |
20080022385 | Crowell et al. | Jan 2008 | A1 |
20080056124 | Nanda et al. | Mar 2008 | A1 |
20080082662 | Dandliker et al. | Apr 2008 | A1 |
20080101234 | Nakil et al. | May 2008 | A1 |
20080126534 | Mueller et al. | May 2008 | A1 |
20080250122 | Zsigmond et al. | Oct 2008 | A1 |
20080270199 | Chess et al. | Oct 2008 | A1 |
20080301765 | Nicol et al. | Dec 2008 | A1 |
20090064332 | Porras et al. | Mar 2009 | A1 |
20090133126 | Jang et al. | May 2009 | A1 |
20090240789 | Dandabany | Sep 2009 | A1 |
20090241170 | Kumar et al. | Sep 2009 | A1 |
20090307753 | Dupont et al. | Dec 2009 | A1 |
20090313373 | Hanna et al. | Dec 2009 | A1 |
20090313698 | Wahl | Dec 2009 | A1 |
20090323543 | Shimakura | Dec 2009 | A1 |
20090328219 | Narayanaswamy | Dec 2009 | A1 |
20100005288 | Rao et al. | Jan 2010 | A1 |
20100077445 | Schneider et al. | Mar 2010 | A1 |
20100095293 | O‘Neill et al. | Apr 2010 | A1 |
20100095367 | Narayanaswamy | Apr 2010 | A1 |
20100138810 | Komatsu et al. | Jun 2010 | A1 |
20100148940 | Gelvin et al. | Jun 2010 | A1 |
20100153316 | Duffield et al. | Jun 2010 | A1 |
20100153696 | Beachem et al. | Jun 2010 | A1 |
20100220584 | DeHaan et al. | Sep 2010 | A1 |
20100235514 | Beachem | Sep 2010 | A1 |
20100235915 | Memon et al. | Sep 2010 | A1 |
20100303240 | Beachem et al. | Dec 2010 | A1 |
20100319060 | Aiken et al. | Dec 2010 | A1 |
20100332539 | Mohan et al. | Dec 2010 | A1 |
20110010585 | Bugenhagen et al. | Jan 2011 | A1 |
20110055381 | Narasimhan et al. | Mar 2011 | A1 |
20110055388 | Yumerefendi et al. | Mar 2011 | A1 |
20110066719 | Miryanov et al. | Mar 2011 | A1 |
20110069685 | Tofighbakhsh | Mar 2011 | A1 |
20110083125 | Komatsu et al. | Apr 2011 | A1 |
20110126136 | Abella et al. | May 2011 | A1 |
20110126275 | Anderson et al. | May 2011 | A1 |
20110145885 | Rivers et al. | Jun 2011 | A1 |
20110170860 | Smith et al. | Jul 2011 | A1 |
20110173490 | Narayanaswamy et al. | Jul 2011 | A1 |
20110185423 | Sallam | Jul 2011 | A1 |
20110196957 | Ayachitula et al. | Aug 2011 | A1 |
20110202655 | Sharma et al. | Aug 2011 | A1 |
20110225207 | Subramanian et al. | Sep 2011 | A1 |
20110228696 | Agarwal et al. | Sep 2011 | A1 |
20110277034 | Hanson | Nov 2011 | A1 |
20110302652 | Westerfeld | Dec 2011 | A1 |
20110314148 | Petersen et al. | Dec 2011 | A1 |
20120005542 | Petersen et al. | Jan 2012 | A1 |
20120079592 | Pandrangi | Mar 2012 | A1 |
20120102361 | Sass et al. | Apr 2012 | A1 |
20120102543 | Kohli et al. | Apr 2012 | A1 |
20120117226 | Tanaka et al. | May 2012 | A1 |
20120136996 | Seo et al. | May 2012 | A1 |
20120137278 | Draper et al. | May 2012 | A1 |
20120137361 | Yi et al. | May 2012 | A1 |
20120140626 | Anand et al. | Jun 2012 | A1 |
20120197856 | Banka et al. | Aug 2012 | A1 |
20120198541 | Reeves | Aug 2012 | A1 |
20120216271 | Cooper et al. | Aug 2012 | A1 |
20120233473 | Vasseur et al. | Sep 2012 | A1 |
20120240232 | Azuma | Sep 2012 | A1 |
20120246303 | Petersen et al. | Sep 2012 | A1 |
20120266220 | Brudnicki et al. | Oct 2012 | A1 |
20120278021 | Lin et al. | Nov 2012 | A1 |
20130003538 | Greenberg et al. | Jan 2013 | A1 |
20130006935 | Grisby | Jan 2013 | A1 |
20130038358 | Cook et al. | Feb 2013 | A1 |
20130086272 | Chen et al. | Apr 2013 | A1 |
20130103827 | Dunlap et al. | Apr 2013 | A1 |
20130145099 | Liu et al. | Jun 2013 | A1 |
20130159999 | Chiueh et al. | Jun 2013 | A1 |
20130174256 | Powers | Jul 2013 | A1 |
20130179487 | Lubetzky et al. | Jul 2013 | A1 |
20130179879 | Zhang et al. | Jul 2013 | A1 |
20130198839 | Wei et al. | Aug 2013 | A1 |
20130205215 | Dunn et al. | Aug 2013 | A1 |
20130246925 | Ahuja et al. | Sep 2013 | A1 |
20130247201 | Alperovitch et al. | Sep 2013 | A1 |
20130254879 | Chesla et al. | Sep 2013 | A1 |
20130268994 | Cooper et al. | Oct 2013 | A1 |
20130275579 | Hernandez et al. | Oct 2013 | A1 |
20130283374 | Zisapel et al. | Oct 2013 | A1 |
20130290521 | Labovitz et al. | Oct 2013 | A1 |
20130297771 | Osterloh et al. | Nov 2013 | A1 |
20130304900 | Trabelsi et al. | Nov 2013 | A1 |
20130305369 | Karta et al. | Nov 2013 | A1 |
20130318357 | Abraham et al. | Nov 2013 | A1 |
20130326623 | Kruglick | Dec 2013 | A1 |
20130333029 | Chesla et al. | Dec 2013 | A1 |
20130347103 | Veteikis et al. | Dec 2013 | A1 |
20140006610 | Formby et al. | Jan 2014 | A1 |
20140006871 | Lakshmanan et al. | Jan 2014 | A1 |
20140012814 | Bercovici et al. | Jan 2014 | A1 |
20140033193 | Palaniappan | Jan 2014 | A1 |
20140047185 | Peterson et al. | Feb 2014 | A1 |
20140047372 | Gnezdov et al. | Feb 2014 | A1 |
20140059200 | Nguyen et al. | Feb 2014 | A1 |
20140089494 | Dasari et al. | Mar 2014 | A1 |
20140096058 | Molesky et al. | Apr 2014 | A1 |
20140115219 | Ajanovic et al. | Apr 2014 | A1 |
20140143825 | Behrendt et al. | May 2014 | A1 |
20140149490 | Luxenberg et al. | May 2014 | A1 |
20140156814 | Barabash et al. | Jun 2014 | A1 |
20140164607 | Bai et al. | Jun 2014 | A1 |
20140165207 | Engel et al. | Jun 2014 | A1 |
20140173623 | Chang et al. | Jun 2014 | A1 |
20140192639 | Smirnov | Jul 2014 | A1 |
20140201717 | Mascaro et al. | Jul 2014 | A1 |
20140215573 | Cepuran | Jul 2014 | A1 |
20140215621 | Xaypanya et al. | Jul 2014 | A1 |
20140281030 | Cui et al. | Sep 2014 | A1 |
20140289854 | Mahvi | Sep 2014 | A1 |
20140298461 | Hohndel et al. | Oct 2014 | A1 |
20140317737 | Shin et al. | Oct 2014 | A1 |
20140331276 | Frascadore et al. | Nov 2014 | A1 |
20140331280 | Porras et al. | Nov 2014 | A1 |
20140331304 | Wong | Nov 2014 | A1 |
20140351203 | Kunnatur et al. | Nov 2014 | A1 |
20140351415 | Harrigan et al. | Nov 2014 | A1 |
20140359695 | Chari et al. | Dec 2014 | A1 |
20150009840 | Pruthi et al. | Jan 2015 | A1 |
20150026809 | Altman et al. | Jan 2015 | A1 |
20150033305 | Shear et al. | Jan 2015 | A1 |
20150036533 | Sodhi et al. | Feb 2015 | A1 |
20150039751 | Harrigan et al. | Feb 2015 | A1 |
20150046882 | Menyhart et al. | Feb 2015 | A1 |
20150058976 | Carney et al. | Feb 2015 | A1 |
20150067143 | Babakhan et al. | Mar 2015 | A1 |
20150082151 | Liang et al. | Mar 2015 | A1 |
20150082430 | Sridhara et al. | Mar 2015 | A1 |
20150085665 | Kompella et al. | Mar 2015 | A1 |
20150095332 | Beisiegel et al. | Apr 2015 | A1 |
20150112933 | Satapathy | Apr 2015 | A1 |
20150113133 | Srinivas et al. | Apr 2015 | A1 |
20150124608 | Agarwal et al. | May 2015 | A1 |
20150128133 | Pohlmann | May 2015 | A1 |
20150138993 | Forster et al. | May 2015 | A1 |
20150142962 | Srinivas et al. | May 2015 | A1 |
20150195291 | Zuk et al. | Jul 2015 | A1 |
20150249622 | Phillips et al. | Sep 2015 | A1 |
20150256555 | Choi et al. | Sep 2015 | A1 |
20150261842 | Huang et al. | Sep 2015 | A1 |
20150261886 | Wu et al. | Sep 2015 | A1 |
20150271255 | Mackay et al. | Sep 2015 | A1 |
20150295945 | Canzanese, Jr. et al. | Oct 2015 | A1 |
20150347554 | Vasantham et al. | Dec 2015 | A1 |
20150356297 | Guri et al. | Dec 2015 | A1 |
20150358352 | Chasin et al. | Dec 2015 | A1 |
20160006753 | McDaid et al. | Jan 2016 | A1 |
20160021131 | Heilig | Jan 2016 | A1 |
20160026552 | Holden et al. | Jan 2016 | A1 |
20160036837 | Jain et al. | Feb 2016 | A1 |
20160050132 | Zhang | Feb 2016 | A1 |
20160072815 | Rieke et al. | Mar 2016 | A1 |
20160094529 | Mityagin | Mar 2016 | A1 |
20160103692 | Guntaka et al. | Apr 2016 | A1 |
20160105350 | Greifeneder et al. | Apr 2016 | A1 |
20160112270 | Danait et al. | Apr 2016 | A1 |
20160119234 | Valencia Lopez et al. | Apr 2016 | A1 |
20160127395 | Underwood et al. | May 2016 | A1 |
20160134533 | Janardhanan | May 2016 | A1 |
20160147495 | Alpert | May 2016 | A1 |
20160147585 | Konig et al. | May 2016 | A1 |
20160162308 | Chen et al. | Jun 2016 | A1 |
20160162312 | Doherty et al. | Jun 2016 | A1 |
20160173446 | Nantel | Jun 2016 | A1 |
20160191476 | Schütz et al. | Jun 2016 | A1 |
20160205002 | Rieke et al. | Jul 2016 | A1 |
20160216994 | Sefidcon et al. | Jul 2016 | A1 |
20160277488 | Fallon et al. | Sep 2016 | A1 |
20160294691 | Joshi | Oct 2016 | A1 |
20160308908 | Kirby et al. | Oct 2016 | A1 |
20160321455 | Deng et al. | Nov 2016 | A1 |
20160352526 | Adler et al. | Dec 2016 | A1 |
20160357424 | Pang et al. | Dec 2016 | A1 |
20160357546 | Chang et al. | Dec 2016 | A1 |
20160357587 | Yadav et al. | Dec 2016 | A1 |
20160357957 | Deen et al. | Dec 2016 | A1 |
20160359592 | Kulshreshtha et al. | Dec 2016 | A1 |
20160359628 | Singh et al. | Dec 2016 | A1 |
20160359658 | Yadav et al. | Dec 2016 | A1 |
20160359673 | Gupta et al. | Dec 2016 | A1 |
20160359677 | Kulshreshtha et al. | Dec 2016 | A1 |
20160359678 | Madani et al. | Dec 2016 | A1 |
20160359679 | Parandehgheibi et al. | Dec 2016 | A1 |
20160359680 | Parandehgheibi et al. | Dec 2016 | A1 |
20160359686 | Parandehgheibi et al. | Dec 2016 | A1 |
20160359696 | Yadav et al. | Dec 2016 | A1 |
20160359697 | Scheib et al. | Dec 2016 | A1 |
20160359698 | Deen et al. | Dec 2016 | A1 |
20160359699 | Gandham et al. | Dec 2016 | A1 |
20160359700 | Pang et al. | Dec 2016 | A1 |
20160359701 | Pang et al. | Dec 2016 | A1 |
20160359703 | Gandham et al. | Dec 2016 | A1 |
20160359704 | Gandham et al. | Dec 2016 | A1 |
20160359705 | Parandehgheibi et al. | Dec 2016 | A1 |
20160359708 | Gandham et al. | Dec 2016 | A1 |
20160359709 | Deen et al. | Dec 2016 | A1 |
20160359711 | Deen et al. | Dec 2016 | A1 |
20160359712 | Alizadeh Attar et al. | Dec 2016 | A1 |
20160359740 | Parandehgheibi et al. | Dec 2016 | A1 |
20160359759 | Singh et al. | Dec 2016 | A1 |
20160359872 | Yadav et al. | Dec 2016 | A1 |
20160359877 | Kulshreshtha et al. | Dec 2016 | A1 |
20160359878 | Prasad et al. | Dec 2016 | A1 |
20160359879 | Deen et al. | Dec 2016 | A1 |
20160359880 | Pang et al. | Dec 2016 | A1 |
20160359881 | Yadav et al. | Dec 2016 | A1 |
20160359888 | Gupta et al. | Dec 2016 | A1 |
20160359889 | Yadav et al. | Dec 2016 | A1 |
20160359890 | Deen et al. | Dec 2016 | A1 |
20160359891 | Pang et al. | Dec 2016 | A1 |
20160359897 | Yadav et al. | Dec 2016 | A1 |
20160359912 | Gupta et al. | Dec 2016 | A1 |
20160359913 | Gupta et al. | Dec 2016 | A1 |
20160359914 | Deen et al. | Dec 2016 | A1 |
20160359915 | Gupta et al. | Dec 2016 | A1 |
20160359917 | Rao et al. | Dec 2016 | A1 |
20160373481 | Sultan et al. | Dec 2016 | A1 |
20170034018 | Parandehgheibi et al. | Feb 2017 | A1 |
20170079044 | Zhu et al. | Mar 2017 | A1 |
20170094706 | Kim et al. | Mar 2017 | A1 |
20170270315 | Kuo et al. | Sep 2017 | A1 |
20180006829 | Kravitz et al. | Jan 2018 | A1 |
20180006911 | Dickey | Jan 2018 | A1 |
20180018602 | DiMaggio et al. | Jan 2018 | A1 |
20180069688 | Arregui De La Cruz et al. | Mar 2018 | A1 |
Number | Date | Country |
---|---|---|
101093452 | Dec 2007 | CN |
101770551 | Jul 2010 | CN |
102521537 | Jun 2012 | CN |
103023970 | Apr 2013 | CN |
103716137 | Apr 2014 | CN |
104065518 | Sep 2014 | CN |
0811942 | Dec 1997 | EP |
1383261 | Jan 2004 | EP |
1450511 | Aug 2004 | EP |
2045974 | Apr 2009 | EP |
2887595 | Jun 2015 | EP |
2009-016906 | Jan 2009 | JP |
1394338 | May 2014 | KR |
2007014314 | Feb 2007 | WO |
2007070711 | Jun 2007 | WO |
2008069439 | Jun 2008 | WO |
2013030830 | Mar 2013 | WO |
2015042171 | Mar 2015 | WO |
2015099778 | Jul 2015 | WO |
2016004075 | Jan 2016 | WO |
2016019523 | Feb 2016 | WO |
Entry |
---|
“Application Discovery Manager User‘s Guide,” VMWare, Inc. pp. 1-74. |
“Cisco IT Tetration Deployment, Part 1 of 2,” Oct. 25, 2017, pp. 1-30. |
Australian Government Department of Defence, Intelligence and Security, “Top 4 Strategies to Mitigate Targeted Cyber Intrusions,” Cyber Security Operations Centre Jul. 2013, http://www.asd.gov.au/infosec/top-mitigations/top-4-strategies-explained.htm. |
Author Unkown, “Blacklists & Dynamic Reputation: Understanding Why the Evolving Threat Eludes Blacklists,” www.dambala.com, 9 pages, Dambala, Atlanta, GA, USA. |
Aydein, Galip, et al., “Architecture and Implementation of a Scalable Sensor Data Storage and Analysis Using Cloud Computing and Big Data Technologies,” Journal of Sensors, Vol. 2015, Article ID 834217, Feb. 2015, 11 pages. |
Backes, Michael, et al., “Data Lineage in Malicious Environments,” IEEE 2015, pages 1-13. |
Bauch, Petr, “Reader‘s Report of Master‘s Thesis, Analysis and Testing of Distributed NoSQL Datastore Riak,” May. 28, 2015, Brno. 2 pages. |
Bayati, Mohsen, et al., “Message-Passing Algorithms for Sparse Network Alignment,” Mar. 2013, 31 pages. |
Berezinski, Przemyslaw, et al., “An Entropy-Based Network Anomaly Detection Method,” Entropy, 2015, Vol. 17, www.mdpi.com/joumal/entropy, pages 2367-2408. |
Berthier, Robin, et al. “Nfsight: Netflow-based Network Awareness Tool,” 2010, 16 pages. |
Bhuyan, Dhiraj, “Fighting Bots and Botnets,” 2006, pages 23-28. |
Bosch, Greg, “Virtualization,” 2010, 33 pages. |
Breen, Christopher, “MAC 911, How to dismiss Mac App Store Notifications,” Macworld.com, Mar. 24, 2014, 3 pages. |
Chandran, Midhun, et al., “Monitoring in a Virtualized Environment,” GSTF International Journal On Computing, Vol. 1, No. 1, Aug. 2010. |
Chari, Suresh, et al., “Ensuring continuous compliance through reconciling policy with usage,” Proceedings of the 18th ACM symposium on Access control models and technologies (SACMAT ’13). ACM, New York, NY, USA, 49-60. |
Chen Xu, et al., “Automating network application dependency discovery: experiences, limitations, and new solutions,” 8th USENIX conference on Operating systems design and implementation (OSDI’08), USENIX Association, Berkeley, CA, USA, 117-130. |
Chou, C.W., et al., “Optical Clocks and Relativity,” Science Vol. 329, Sep. 24, 2010, pages 1630-1633. |
Cisco Systems, “Cisco Network Analysis Modules (NAM) Tutorial,” Cisco Systems, Inc., Version 3.5. |
Cisco Systems, Inc., “A Cisco Guide to Defending Against Distributed Denial of Service Attacks,” May. 3, 2016, 34 pages. |
Cisco Systems, Inc., “Addressing Compliance from One Infrastructure: Cisco Unified Compliance Solution Framework,” 2014. |
Cisco Systems, Inc., “Cisco Application Dependency Mapping Service,” 2009. |
Cisco Systems, Inc., “Cisco Application Visibility and Control,” Oct. 2011, 2 pages. |
Cisco Systems, Inc., “Cisco Tetration Platform Data Sheet”, Updated Mar. 5, 2018, 21 pages. |
Cisco Systems, Inc., “White Paper - New Cisco Technologies Help Customers Achieve Regulatory Compliance,” 1992-2008. |
Cisco Technology, Inc., “Cisco Lock-and-Key:Dynamic Access Lists,” http://www/cisco.com/c/en/us/support/docs/security-vpn/lock-key/7604-13.html; Updated Jul. 12, 2006, 16 pages. |
Di Lorenzo, Guisy, et al., “EXSED: An Intelligent Tool for Exploration of Social Events Dynamics from Augmented Trajectories,” Mobile Data Management (MDM), pages 323-330, Jun. 3-6, 2013. |
Duan, Yiheng, et al., “Detective: Automatically Identify and Analyze Malware Processes in Forensic Scenarios via DLLs, IEEE ICC 2015 - Next Generation Networking Symposium, pages 5691-5696. |
Feinstein, Laura, et al., “Statistical Approaches to DDoS Attack Detection and Response,” Proceedings of the DARPA Information Survivability Conference and Exposition (DISCEX ’03), Apr. 2003, 12 pages. |
George, Ashley, et al., “NetPal: A Dynamic Network Administration Knowledge Base,” 2008, pages 1-14. |
Goldsteen, Abigail, et al., “A Tool for Monitoring and Maintaining System Trustworthiness at Run Time,” REFSQ (2015), pages 142-147. |
Hamadi, S., et al., “Fast Path Acceleration for Open vSwitch in Overlay Networks,” Global Information Infrastructure and Networking Symposium (GIIS), Montreal, QC, pages 1-5, Sep. 15-19, 2014. |
Heckman, Sarah, et al., “On Establishing a Benchmark for Evaluating Static Analysis Alert Prioritization and Classification Techniques,” IEEE, 2008; 10 pages. |
Hewlett-Packard, “Effective use of reputation intelligence in a security operations center,” Jul. 2013, 6 pages. |
Hideshima, Yusuke, et al., “Starmine: A Visualization System for Cyber Attacks,” https://www.researchgate.net/publication/221536306, Feb. 2006, 9 pages. |
Huang, Hing-jie, et al., “Clock Skew Based Node Identification in Wireless Sensor Networks,” IEEE, 2008, 5 pages. |
InternetPerils, Inc., “Control Your Internet Business Risk,” 2003-2015, https://www.internetperils.com . |
Ives, Herbert, E., et al., “An Experimental Study of the Rate of a Moving Atomic Clock,” Journal of the Optical Society of America, Vol. 28, No. 7, Jul. 1938, pages 215-226. |
Janoff, Christian, et al., “Cisco Compliance Solution for HIPAA Security Rule Design and Implementation Guide,” Cisco Systems, Inc., Updated Nov. 14, 2015, part 1 of 2, 350 pages. |
Janoff, Christian, et al., “Cisco Compliance Solution for HIPAA Security Rule Design and Implementation Guide,” Cisco Systems, Inc., Updated Nov. 14, 2015, part 2 of 2, 588 pages. |
Kerrison, Adam, et al., “Four Steps to Faster, Better Application Dependency Mapping - Laying the Foundation for Effective Business Service Models,” BMCSoftware, 2011. |
Kim, Myung-Sup, et al. “A Flow-based Method for Abnormal Network Traffic Detection,“ IEEE, 2004, pages 599-612. |
Lab Sku, “VMware Hands-on Labs - HOL-SDC-1301” Version: 20140321-160709, 2013; http://docs.hol.vmware.com/HOL-2013/holsdc-1301htmlen/ (part 1 of 2). |
Lab Sku, “VMware Hands-on Labs - HOL-SDC-1301” Version: 20140321-160709, 2013; http://docs.hol.vmware.com/HOL-2013/holsdc-1301htmlen/ (part 2 of 2). |
Lachance, Michael, “Dirty Little Secrets of Application Dependency Mapping,” Dec. 26, 2007. |
Landman, Yoav, et al., “Dependency Analyzer,” Feb. 14, 2008, http://jfrog.com/confluence/display/DA/Home. |
Lee, Sihyung, “Reducing Complexity of Large-Scale Network Configuration Management,” Ph.D. Dissertation, Carniege Mellon University, 2010. |
Li, Ang, et al., “Fast Anomaly Detection for Large Data Centers,” Global Telecommunications Conference (GLOBECOM 2010, Dec. 2010, 6 pages. |
Liu, Ting, et al., “Impala: A Middleware System For Managing Autonomic, Parallel Sensor Systems,” In Proceedings of the Ninth ACM SIGPLAN Symposium On Principles And Practice Of Parallel Programming(PPoPP ’03), ACM, New York, NY, USA, Jun. 11-13, 2003, pages 107-118. |
Lu, Zhonghai, et al., “Cluster-based Simulated Annealing for Mapping Cores onto 2D Mesh Networks on Chip,” Design and Diagnostics of Electronic Circuits and Systems, pp.1, 6, 16-18, Apr. 2008. |
Matteson, Ryan, “Depmap: Dependency Mapping of Applications Using Operating System Events: a Thesis,” Master‘s Thesis, California Polytechnic State University, Dec. 2010. |
Natarajan, Arun, et al., “NSDMiner: Automated Discovery of Network Service Dependencies,” Institute of Electrical and Electronics Engineers INFOCOM, Feb. 2012, 9 pages. |
Navaz, A.S. Syed, et al., “Entropy based Anomaly Detection System to Prevent DDoS Attacks in Cloud,” International Journal of computer Applications (0975-8887), Vol. 62, No. 15, Jan. 2013, pages 42-47. |
Neverfail, “Neverfail IT Continuity Architect,” 2015, https://web.archive.org/web/20150908090456/http://www.neverfailgroup.com/products/it-continuity-architect . |
Nilsson, Dennis K., et al., “Key Management And Secure Software Updates In Wireless Process Control Environments,” In Proceedings of the First ACM Conference On Wireless Network Security (WiSec ’08), ACM, New York, NY, USA, Mar. 31-Apr. 2, 2008, pages 100-108. |
Nunnally, Troy, et al., “P3D: A Parallel 3D Coordinate Visualization for Advanced Network Scans,” IEEE 2013, Jun. 9-13, 2013, 6 pages. |
O‘Donnell, Glenn, et al., “The CMDB Imperative: How to Realize the Dream and Avoid the Nightmares,” Prentice Hall, Feb. 19, 2009. |
Ohta, Kohei, et al., “Detection, Defense, and Tracking of Internet-Wide Illegal Access in a Distributed Manner,” 2000, pages 1-16. |
Pathway Systems International Inc., “How Blueprints does Integration,” Apr. 15, 2014, 9 pages, http://pathwaysystems.com/company-blog/ . |
Pathway Systems International Inc., “What is Blueprints?” 2010-2016, http://pathwaysystems.com/blueprints-about/. |
Popa, Lucian, et al., “Macroscope: End-Point Approach to Networked Application Dependency Discovery,” CoNEXT’09, Dec. 1-4, 2009, Rome, Italy, 12 pages. |
Prasad, K. Munivara, et al., “An Efficient Detection of Flooding Attacks to Internet Threat Monitors (ITM) using Entropy Variations under Low Traffic,” Computing Communication & Networking Technologies (ICCCNT ‘12), Jul. 26-28, 2012, 11 pages. |
Sachan, Mrinmaya, et al., “Solving Electrical Networks to incorporate Supervision in Random Walks,” May. 13-17, 2013, pages 109-110. |
Sammarco, Matteo, et al., “Trace Selection for Improved WLAN Monitoring,” Aug. 16, 2013, pages 9-14. |
Shneiderman, Ben, et al., “Network Visualization by Semantic Substrates,” Visualization and Computer Graphics, Vol.12, No.5, pp.733,740, Sep.-Oct. 2006. |
Thomas, R., “Bogon Dotted Decimal List,” Version 7.0, Team Cymru NOC, Apr. 27, 2012, 5 pages. |
Wang, Ru, et al., “Learning directed acyclic graphs via bootstarp aggregating,” 2014, 47 pages, http://arxiv.org/abs/1406.2098 . |
Wang, Yongjun, et al., “A Network Gene-Based Framework for Detecting Advanced Persistent Threats,” Nov. 2014, 7 pages. |
Witze, Alexandra, “Special relativity aces time trial, ‘Time dilation’ predicted by Einstein confirmed by lithium ion experiment,” Nature, Sep. 19, 2014, 3 pages. |
Woodberg, Brad, “Snippet from Juniper SRX Series” Jun. 17, 2013, 1 page, O’Reilly Media, Inc. |
Zatrochova, “Analysis and Testing of Distributed NoSQL Datastore Riak,” Spring, 2015, 76 pages. |
Zhang, Yue, et al., “Cantina: A Content-Based Approach to Detecting Phishing Web Sites,” May. 8-12, 2007, pages 639-648. |
Arista Networks, Inc., “Application Visibility and Network Telemtry using Splunk,” Arista White Paper, Nov. 2013, 11 pages. |
Blair, Dana, et al., U.S. Appl. No. 62/106,006, tiled Jan. 21, 2015, entitled “Monitoring Network Policy Compliance.” |
Kraemer, Brian, “Get to know your data center with CMDB,” TechTarget, Apr. 5, 2006, http://searchdatacenter.techtarget.com/news/118820/Get-to-know-your-data-center-with-CMDB . |
Li, Bingbong, et al., “A Supervised Machine Learning Approach to Classify Host Roles On Line Using sFlow,” in Proceedings of the first edition workshop on High performance and programmable networking, 2013, ACM, New York, NY, USA, 53-60. |
Number | Date | Country | |
---|---|---|---|
62616347 | Jan 2018 | US |