Embodiments generally relate to monitoring a network, such as for security. One or more embodiments regard network security monitoring in a telecommunications (e.g., Third Generation Partnership Project (3GPP), Long Term Evolution (LTE), or the like), or other network.
Some current widely used mechanisms for traffic monitoring/telemetry include using established protocols like sampled flow (sFlow), from the sFlow.org consortium, Netflow, from Cisco Systems®, Inc. of San Jose Calif., United States, or the like for collecting telemetry and delivering the collected telemetry to an analysis system. The analysis system can perform network security analytics on the telemetry. Traffic data is typically collected by physical switches using port mirroring, basic filtering/statistics aggregation using dedicated hardware engines within the physical switches/routers, or through special-purpose traffic duplication hardware.
At present, the inventors are not aware of a known solution for providing line-rate, secure network security monitoring capability on a general purpose communication platform, at high packet processing rates, and without visible effects to the workloads in question.
In the drawings, which are not necessarily drawn to scale, like numerals may describe similar components in different views. Like numerals having different letter suffixes may represent different instances of similar components. The drawings illustrate generally, by way of example, but not by way of limitation, various embodiments discussed in the present document.
Discussed herein, are systems and methods directed to accelerating NSM. One or more embodiments include hardware solutions to improve such acceleration, such as a hardware queue manager (HQM). Embodiments can help deliver platform capabilities (communication platform core, component blocks, CSME/ME/IE), enhancements for securely deploying NSM, and advanced non-intrusive debug capabilities within scalable operator networks.
Communication platforms can be optimized for single stream processing from network interface controllers (NICs)/switches. However, operational security mechanisms like Network Security Monitoring (NSM), secure delivery of telemetry out of platform, maintaining SLA sensitivity while performing network monitoring, and the like, are all tasks that can help Network Function Virtualization (NFV), LTE, 3GPP, Virtual Evolved Packet Core (vEPC), and/or Virtual Customer Premises Equipment (vCPE) network visibility. Thus, solving these problems can help reduce workloads on the platform.
Embodiments discussed herein can help address this workload and visibility gap by providing a general mechanism that can address one or more of these gaps. One or more embodiments can securely bind platform trusted elements like Converged Security and Manageability Engine (CSME), Management Engine (ME), IE, and Software Guard Extensions (SGX), and electrical or electronic components (e.g., HQM, cache controller, crypto processor, or the like)/Field Programmable Gate Array (FPGA) based security and compression processing.
As switching functions extend into the servers between Virtual Machines (VMs), such as using Open vSwitch (OvS®), from Apache® Software Foundation of Forest Hill, Md., United States, or router software, performing the same actions in software is highly inefficient and results in platform packet processing degradation. This degradation may be noticeable to tenants (e.g., infrastructure as a service (IaaS) tenants) or service consumers, for example. Performing port mirroring, using switched port analyzer (SPAN), and/or Terminal Access Point (TAP), and then using NSM traffic protection (e.g., encryption, source authentication, etc.) using current communication platform software-only mechanisms consumes many compute cycles across many cores and increases core-to-core communication. This is not sustainable given the increasing rate of packet processing on modern servers, exceeding 100 Gigabits per second (Gbps) and projected to reach 400 Gbps by 2020.
Without hardware acceleration, the current performance level of plaintext traffic ingress to traffic egress is disrupted, and performance degrades, which can affect the entire platform performance, limit scalability, and restrict tenant load capabilities (e.g., bandwidth) or lead to poor Service Level Agreement (SLA) delivery.
By offloading these capabilities in a carefully designed software framework and through modifications to the existing HQM, secure network monitoring and debug capabilities can be constructed with a minimum of communication platform software cycles spent, leveraging HQM offload and providing improved customer value, via improved system performance. In addition, communication platform security technologies, like SGX and CSME, can be used for securely provisioning the HQM traffic policies and memory areas for forking (e.g., “teeing”) the duplicate traffic. SGX and VM-based memory encryption technologies can provide some measure of confidentiality, but do not protect a policy path and the policy processing from other VMs/Virtual Network Functions (VNFs), System Administrator, NFV Infrastructure Managers, or the like. NFV generally decouples software implementations of network functions from the hardware the network functions used to perform the operations of the software.
NSM can be a useful security-provider in existing 3GPP/LTE networks. NSM may include an ability to securely monitor network telemetry for security threats, such as malware encroachments, anomalies, and zero-day attacks, among others. As existing LTE/3GPP systems move to NFV, and Software Defined Systems (SDS), NSM can be provided while maintaining a same line rate and same functionality as in existing systems.
As a number of cores on a platform increases, it may be expected that all the telecommunications infrastructure like the vEPC or vCPE are running on the same platform. Thus, it can be useful to monitor the vEPC and vCPE networks on the platform. Currently this entire processing, on at least some platforms is not visible on the physical networks outside of the platform. Hence, adding components like NSM on platforms may help increase security. Stated another way, in environments where VNFs (VMs performing specific network functions) are chained into Service Function Chains (SFC), the ability to inspect and debug traffic between each stage can help improve security for full-scale deployments. One or more of these problems may be addressed by one or more embodiments.
To save compute (e.g., central processing unit (CPU)) cycles, it can help to accelerate operations in a general-purpose manner, such as to not take away cycles from higher-level compute jobs or packet processing workload. Using the techniques described in the following sections, embodiments can help achieve these goals.
Some requirements and problems that should be addressed for NFV/SDN systems, over and above existing physical network functions/systems, can be summarized as follows. These security and performance requirements are taken from the European Telecommunications Standards Institute (ETSI) NFV specifications (e.g., ETSI NFV SEC013, SEC012, SEC001) that are being developed with the operators for NFV/SDN deployments on communications platforms:
1. NFV Workloads can migrate across platforms. This means that NSM capabilities can be replicated/migrated along with the workloads. Hence, NSM can consistently exist on every platform to which a workload could migrate.
2. NFV/SDN platforms support heterogeneous workloads (e.g., including vEPC, vCPE, video, web, tenant applications, etc.). Plus, NFV control and data planes are on entirely different platforms. Together, these conditions can degrade NSM performance, impacting SLA and traffic, and making it useful for NSM to be performed on all platforms in a deterministic manner, supporting SLAs with low jitter.
3. Security requirements for NSM (from ETSI NFV SEC013 normative specification) demands utilization of platform trust capabilities, such as to be able to securely deliver and implement the NSM policy.
Embodiments can function by collecting telemetry, such as from dynamic, virtual networks that overlay onto physical networks. One or more embodiments integrate policy engines from open virtual switches/routers into a communication platform. One or more embodiments may, additionally or alternatively, use packet forking (sometimes known as “teeing” in the art) capabilities in an HQM device to provide these functions. Such embodiments can help save CPU cycles and enable additional NSM capabilities, improving overall system processing speed. Traffic forking into multiple streams could include a division of the traffic after decryption, de-encapsulation, and processing, before re-encryption or compression, for example.
At present, the inventors are unaware of a known solution for providing line-rate, secure NSM capability on general purpose communication platforms at high packet processing rates, such as without visible effects to the workloads in question. The current communication platforms are optimized for single stream processing from NICs/switches. Operational security mechanisms like NSM, secure delivery of telemetry out of platform, maintaining SLA sensitivity, while performing NSM, are tasks will help secure NFV/LTE vEPC, vCPE, network visibility.
The NFV infrastructure 102 virtualizes network node functions that may be chained together to create, for example, communication services. The NFV infrastructure 102 as illustrated includes virtual network functions (VNFs) 116A, 116B, 116C, and 116D, a switch/router 118, and an operator infrastructure 120.
VNFs 116A-116D are virtualized tasks formerly carried out by dedicated hardware. The VNFs 116A-116D move these operations previously performed by the dedicated hardware to software instantiations. The VNFs 116A-116D operate by virtue of the hardware and/or software programming of the host 122.
The switch/router 118 provides communication capability between platforms 126A-126B and/or VNFs 116A-116D. A switch connects computers or other network devices. A router connects a first network to a second network, such as by connecting a network to the internet. The router generally manages a path of data (e.g., a request, message, or a response, among others) provided to another network.
The operator infrastructure 120 hosts other operator's VNFs and includes the host 122 (e.g., operating system (OS), cloud OS, and/or hypervisor), a firmware interface 124 (e.g., a unified extensible firmware interface (UEFI) or basic input/output system (BIOS)), a plurality of platforms 126A and 126B (for example), and interconnect circuitry 128 (e.g., input/output (I/O) ports, network interface controller (NIC), switch, host fabric interface (HFI), or the like).
The host 122 can include hardware or software for providing resources used by the VNFs 116A-116D for performing their operations. The host can include an OS, cloud OS, hypervisor, or the like on which the VNFs can operate.
The firmware interface 124 can include a UEFI, BIOS, or the like. The firmware interface 124 defines a software interface between the host 122 and the firmware (e.g., the VNF 116A-116D).
The platform 126A-126B provides communication functionality for devices connected to a communication network. Examples of functionality provided by the platform 126A-126B include voice lines and internet access, and operations for supporting the voice and internet access, among others. The platform 126A includes memory 130 (so does platform 126B, but the storage is not shown in order not to obscure the view of the system 100. The platforms 126A-126B may be communicatively coupled by interconnect circuitry 128.
The orchestrator 104 performs resource and/or network service orchestration. The orchestrator 104 binds functions provided by the VNFs 116A-116D, such as to create a service in an otherwise dispersed NFV environment. The orchestrator 104 can help ensure that adequate compute, storage, and/or other network resources are available to provide a network service. The orchestrator 104 can authorize, coordinate, release, and govern VNFs 116A-116D sharing resources of the infrastructure 102.
The OSS/BSS 106 are computer systems used by telecom service providers to manage their networks. The OSS/BSS 106 can support network inventory, configuration, fault management, and/or service provisioning.
The virtual network function manager (VNFM) 108 works with the orchestrator 104 and VIM 110 to provide VNF capability. The VNFM 108 may instantiate VNFs 116A-116D, scale VNFs 116A-116D, update and/or upgrade the VNFs 116A-116D, and/or terminate VNFs 116A-116D. The VNFM 108 can manage a single VNF or multiple VNFs. The VNFM 108 maintains the virtualized resources that support the VNF 116A-116D.
The VIM 110 as illustrated includes an SDN controller 114. The VIM 110 controls and manages compute, storage, and other network resources of the infrastructure 102. The VIM 110 can handle infrastructure in one or a plurality of infrastructures, such as can include the infrastructure 102. The VIM 110 can maintain a list of which virtual resources are allocated which physical resources, manage security group policies (for access control), manage a repository of NFV hardware resources and software resources, such as to help improve and optimize the use of system resources.
The SDN controller 114 is an application that manages flow control, such as within networks. The SDN controller 114 allows a server or other network resource to select a switched packet destination, indicating to a switch where to send a packet. The SDN controller 114 can take a control plane of network hardware and runs it as software. The SDN controller 114 can distribute monitoring policies.
The network analysis tools 112 can provide functionality for analyzing telemetry data, such as NSM functionality. The network analysis tools 112 can include vEPC, vCPE NFV analytics, data storage, network anomaly detection, and/or malware detection, or the like.
Network security monitoring (NSM) traffic is indicated by arrows 132, 134, 136, and 138. The flow of this NSM traffic may be “accelerated” by embodiments discussed herein. The SDN controller 114 of the VIM 110 may provide a policy to the VNF 116D (e.g., a security monitoring (SecMon) VNF), as indicated by a combination of arrows 138 and 132. Traffic provided to the VNF 116D may be securely terminated. The VNF 116D can monitor traffic (e.g., all traffic) on the switch/router 118 (as indicated by arrow 132). The VNF 116D can apply a provided monitoring policy to the traffic. The VNF 116D provides, based on the monitoring policy, monitored traffic to the interconnect circuitry 128 (as indicated by the arrow 134 or the arrows 140 and 138). The interconnect circuitry 128 provides the monitored traffic to network analysis tools 112, as indicated by arrow 136. The network analysis tools 112 may include security and networking analytics systems, meta data collectors, network profiling, per-tenant and/or per-flow monitoring systems, and/or onto tenant monitoring systems.
The NSM scheme of
Other arrows (not labelled with a reference number in
Consider a complex environment, such as that of
NSM, using the system 200, may include providing a monitoring policy to the VNF 116D and the security circuitry 210, as indicated by arrows 138 and 132, and 202 or 204, respectively. In one or more embodiments, the security circuitry 210 may monitor all traffic on the switch/router 118, thus offloading the CPU work from the host 122.
The platform security circuitry 210 (e.g., SGX, ME, IE, etc.) may have secure out of band on-die or on-package or intra-fabric channel(s) that can be used to deliver a policy from the security circuitry to the various VNFs operating on the processor or chipset. The security circuitry 210 may receive policies from the SDN controller 114 and translate into on-processor blocks (e.g., caches, three dimensional xpoint (3DXP) or other memory, NICs, switche(s), storage, etc.) specific configuration commands and send different commands to different component blocks.
Each component (including those mentioned above) may include one or more queues 208. Queues are logical and configurable sets that include resources offered by each component. Thus, a cache can serve multiple queues and be allocated to one or more VNFs/VMs. The same arrangement may be put in place for memory, NIC, ports, etc. The SDN controller 114 and/or security circuitry 210 may configure these queues 208 at the start of the platform and optionally while the platform 126A-126B executes one or more VNFs 116A-116D.
Using the queues 208, the platform component resources can thus be “segmented” into sets of platform resource buckets which can be assigned to each VNF. One criteria of allocation being the projected or contracted SLA delivery. Platform segmented resources can be allocated per tenant, per flow, and/or per paid subscription model.
The queues 208 can process data (including network traffic). The queues 214 can be replicated per policy into one or more other queues. In other words, per the policy delivered into the component blocks from the security circuitry 210, the component blocks can replicate the entire queues onto other queues that support, or are owned and/or dedicated for NSM.
The VNF 116D can either read the queues 214 for configurable (e.g., software configurable) further processing and analysis (e.g., using the VNF 116D as a front-end software processor, such as for analytics). In other embodiments, (e.g., per policy) the queue(s) 214 may be flushed directly on to the external network for external analysis and monitoring by the analysis tools 112. In yet other embodiments, the analytics and analysis may run as part of the VNF 116D which can perform front end software processing and then the analysis tools 112.
In one or more embodiments, the interconnect circuitry 128 can be designed to write to two duplicate/mirrored descriptor tables instead of the single table that is used in the system 100. One of the tables can be used for general I/O tasks and the other can be used by the VNF 116D to review based on a relevant policy. By modifying the design of both the hardware and/or software, full line-rate security monitoring can be achieved with very little overhead in either the network device or software stack processing. A similar approach can be made available for storage or other applications.
The queue policy registers 216 include data indicating which queues (of the memory queues 208) provide data to which cores 218. The queue policy registers 216 can be read to determine a core to which duplicated traffic and other traffic is to be provided.
The cores 218 are processing cores, such as can include processing cores and/or debug cores. The cores 218 can implement the VNFs 116A-116D. The memory queues 208. The caches 220 can store instructions and/or data to be used by the cores 218 in performing their operations.
The system 200 provides a way to build an interface where a data center-level administrator (or other entity) can debug, monitor, or otherwise view data in a complicated service chain. This is a convenient way to be able to reach in from the orchestrator 104, the SDN controller 114, and, for example, monitor traffic between VM7 and VM8, and determine whether their traffic pattern is normal, or whether there is a misconfiguration in the system 100.
The security circuitry 210 can duplicate the traffic flows between a pair of VMs, for example. Such a configuration could provide the capability to do this sort of debug without necessarily having to go build a custom interface for each of the pieces of software that need to operate on the system.
The system 200 can provide a work around for software versioning issues, which can be problematic. Consider an OSS of some version that supports one interface, that is changed in the next version of the interface, creating havoc in the monitoring software systems. An encryption function, for example, SGX, can help retain encryption of the VMs and yet still provide visibility. So, at that point the security circuitry 210 can be trusted to perform the duplication.
In one or more embodiments, the interconnect circuitry 128 can create duplicated descriptor tables. Such duplication of descriptor table can be one way to help perform the debug, monitoring, or the like, perhaps without the presence of an HQM. This is one way in which a NIC could be used to duplicate traffic flows, for example.
The NIC 302 connects a compute device, such as the platform 126A-126B, to a network. The NIC, in one or more embodiments can include an HFI. The NIC 302 receives data to be processed, as indicated by data 301. The data 301 can be provided by the NIC 302 to the cache 318 and the producer 304. The producer 304 converts the data 301 to a format compatible with the workers 308A-308I. The producer 304 can add descriptors to the data to indicate what operations are to be performed on the data. The converted data can be provided to the HQM 306, such as on line 307.
The HQM 306, 310, and 312 includes a queue 320, 322, and 324. The HQMs 306, 310, and 312 manage communication requests and can aggregate data between cores without relying on software routines. The HQMs 306, 310, and 312 may manage the replication and distribution of the NSM data to workers 308A-308I. The HQM 306, 310, and 312 can replicate and tee data, such as for debug and/or detection (or other NSM) purposes. As will be seen later,
The HQM 306, 310, and 312 form an offload acceleration engine, so that a CPU core can send a packet by creating a 16-bit descriptor and sending data using an instruction. The HQM 306, 310, and 312 then can perform up to two operations (1) forking the flow from one source to two output queues, and (2) gathering performance metrics and logging. One or both operations may be performed without disturbing the core operations. The forked output queue is then read by either another core (to be sent out, effectively a version of port mirroring with the HQM 306, 310, and 312 involved to hardware accelerate the introspection) or to the IE to be sent out over a management interface (e.g., an SDN controller or other interface). The HQM 306, 310, and 312 can offload queue management, and can help fork (e.g., “tee”) traffic for debug, security monitoring (for NFV standard ETSI NFV SEC013), network performance monitoring (ETSI NFV specification PER001), traffic engineering, or the like. Forking is receiving data on a first connection and providing, by a component coupled to the first connection, at least a portion of the received data to at least two other components. Similarly, the statistics can help detect dropped packets, for example, without requiring the full traffic stream to be duplicated. Many statistics may be maintained in the HQM 306, 310, and 312 already; new performance counters for the number of queue entries may be split across queues. In one or more embodiments, the statistics and/or other counters may be accessible to the OS/VMM only, and not guests, such as for security reasons. Applicable, existing protocols could include simple network management protocol (SNMP), reliable event logging (RLOG), Netflow, internet protocol flow information export (IPFIX), sFlow, Internet engineering task force (IETF) standards, or the like.
The workers 308A-308D are cores 218. The workers 308A-308D are first-level workers. The workers 308A-308D receive data from the queue 320, as indicated by line 309. The operations performed by the workers 308A-308D can include, for example, per-packet work, such as can include decryption, encapsulation, de-encapsulation, firewall, encryption, decryption, compression, and/or decompression, among others. The result of operations performed by the workers 308A-308D are provided to a second-level HQM 310, as indicated by line 311. The queue temporarily stores the results and the HQM 310 provides the results to second-level workers 308E-308I. The second-level workers 308E-308H can perform packet inspection (e.g., deep packet inspection), network address translation, intrusion detection, ad insertion, routing, or the like. The results from the second-level workers 308E-H can be provided to a third level HQM 312, as indicated by line 317. A thread can be associated with an individual worker, such that the worker is responsible for the execution of the thread.
In one or more embodiments, results from the workers 308A-308D can be provided to the worker 308I, such as for debug purposes, such as indicated by line 315. In cases in which the first stage of cores has a functional bug, it is difficult to debug these cores, such as at high speeds with traditional software-based virtual switch (vswitch) solutions (e.g., the system 100). To make debug of network service function chains (SFCs) where multiple sets of complex multi-core VMs are chained together, hardware-supported debug capabilities can be important, such as to verify that the operations being performed are what is expected. The HQM 310 can be modified with hardware extensions to enable debug—more details are provided in
The HQM 312 provides data to a consumer 314, as indicated by line 319. The consumer 314 converts the data from the workers 308E-308H to a form compatible with a receiver. The consumer 314 provides the converted data to the NIC 316, as indicated by line 333.
The workers 308A-308D can provide data to the cache 318, as indicated by line 321. The workers 308A-308D can receive data from the cache 318, as indicated by line 323. The workers 308E-308H can provide data to the cache 318, as indicated by line 327. The workers 308E-308H can receive data from the cache 318, as indicated by line 325. The NIC 316 can receive data from the cache 318, such as to be transmitted to a device on a network which the system 300 supports, on line 331.
The queue entry duplication circuitry 402 can help ensure that data in an entry to which a pointer points does not change by the time the data is provided to the worker 308I. The queue entry duplication circuitry 402 and data copy circuitry 404 can retrieve an entry to which a pointer points and provide the duplicated traffic 406 to the worker 308I. The copy circuitry 404 can be part of the HQM 306, 310, or 312 or external to the HQM 306, 310, or 312.
The cores 218A-218E receive instructions and perform operations based on the instructions. A set of instructions, when executed by one of the processor cores 218A-218E, can allow a software program to perform a specific function via the physical actions of the cores 218A-218E. The workers 308A-308D can include respective cores 218A-218C. The input queues 506 include queues 506A, 506B, 506C, and 506D. The input queues 506 can be a part of the queue 320.
The HQM 504 as illustrated includes a sequencer circuitry 508, control circuitry, the duplication circuitry 402, and the copy circuitry 404. Items of the HQM 504 can be a part of the HQM 310. The sequencer circuitry 508 generates addresses to step through a program. The addresses generated by the sequencer circuitry 508 can be generated based on a counter, a field from an instruction, or other data from the input queues 506. The control circuitry 510 manages operations of a program. The control circuitry 510 can respond to control commands and data, such as from a user and/or the sequencer circuitry 508.
The output queues 512 can include 512A, 512B, 512C, and 512D. The output queues 512 can be a part of the queue 322. The output queues 512 manage a flow of data to the processing cores 218D and the debug cores 218E (among other cores). The debug cores 218E can include the worker 308I. The processing cores 218D can include the workers 308E-308H.
The configuration interface 514 as illustrated includes configuration registers and queue routing control circuitry 518. The configuration registers 516 include data indicating which data from the input queues 506 is to be provided for NSM operations, such as to the debug cores 218E. The queue routing control circuitry 518 can be configured to control the sequencer and cause the sequencer circuitry 508 to provide data to be debugged to the output queue 512D. The data to be debugged can be provided to one of the output queues 512A-512C and the output queue 512D, such as to provide the data to the processing cores 218D and the debug cores 218E. Data that is not to be debugged, can be provided to the processing cores 218D, without being provided to the debug cores 218E. The configuration registers 516 and queue routing control circuitry 518 allow software to reconfigure or, hardware to reconfigure, for example, out-of-band, and allow a particular queue, which is set up to route traffic from a first core to a second core (e.g., first VM to a second VM), is now set up to route traffic from the first core to the second core as well as from the first core to the third core to, for instance, enable the debug, or other monitoring.
The HQM 504 can function as a memory-mapped IO device which routes queue entries from input queues 506 through sequencing by sequencer circuitry 508, arbitration, optional reordering, and/or optional atomic flow management (e.g., by the control circuitry 510 and/or duplication circuitry 402), then to output queues 512. The queue entries in the output queues 512 can be delivered to platform threads, such as can be executed by the processing cores 218D and debug cores 218E. Such a system 500 can offload queue management and eliminate shared memory based queue structures for inter-thread data exchange, reducing overheads and improving performance. Specific changes to support flow duplication, including config registers 516, such as can be used to enable the duplication (e.g., via the teeing mode) of the sequencer circuitry 508, changes to the queue sequencer circuitry 508 and new logic to copy the queue entries to duplicate the packets (e.g., the duplication circuitry 402, which can copy the pointer of one or more of the input queues 506). Optionally, a copy circuitry 404 (internal or external to the HQM 504) can be used to copy the packets themselves rather than just the pointer as can be done by the duplication circuitry 402, so the debug cores 218E can consume the packets later without risk of the contents changing, which could lead to incorrect debug results.
There are some interface options shown in
Current ARM and million instructions per second (MIPS) systems can help perform physical network functions, and use dedicated fixed function component engines for traffic separation. Embodiments discussed herein can be implemented using general purpose ARM systems or others as platform architects start addressing dynamic, migration-driven, and heterogeneous NFV/SDN workloads and systems, such as software-driven cloud-portable NFV systems.
One or more embodiments discussed herein may be used in a 5G/Mobile Edge Cloud (MEC) environment, such as for a wireless base station customer. An SoC including an HQM as discussed herein can be useful in an SGX or other security enabled environment.
Elements of this invention can enable an interface architecture or other communication platform to be operationally ready for NFV and SDN deployments. As a number of cores increase on interface architecture (IA) platforms, multiple, heterogeneous VNFs/VMs may be deployed on same platform. NSM provides an operational visibility component to an operator network deployment. Embodiments herein may help assure that IA provides line rate and secure network traffic metrics and traffic elements for off-platform or on-platform monitoring.
To help explain the advantages of some embodiments, a few use cases are provided herein. Consider multiple compute devices (e.g., Box A, Box B, and Box C) connected via a standard Ethernet switch. Box A communicates with Box B, and Box B communicates with Box C, and then, to the rest of a network. However, when the functionality of the Boxes A-C is consolidated onto a server as VMs or VNFs, there are only a few different options for how to connect the functionality. One is via raw shared memory, which has security and robust operational problems. Another is a virtual switch, which is slow. A third is using hardware acceleration, like the HQM discussed herein. But, when using HQM, there is generally bug visibility into the functionality of the Boxes A-C. Assume Box B is in the middle of the service function chain (i.e., the functionality of Boxes A-C connected together), and Box B is misconfigured and as a result, dropping packets.
The visibility in a typical configuration would be at the input to Box A and at the output from Box C. Such a configuration makes it difficult to determine which of the Boxes A-C is misconfigured, or malfunctioning, which makes it difficult to figure out which Box to go debug in more detail. With one or more embodiments, the data (e.g., packet stream) can be forked into a debug stream. The packet stream between Box A and Box B could be inspected, for example, such as to make sure that those packets are properly being routed, or that Box A is operating properly. Additionally, or alternatively, a test point can be inserted between Boxes B and C. Using one or more embodiments, the operations of Box B could be debugged. Worker 8, in the embodiments of
The method 700 can further include, wherein the first worker thread executes on a first processing core, the second worker thread executes on a second processing core, and the third worker thread executes on a third processing core, the first, second, and third processing cores comprising separate processing cores. The method 700 can further include, wherein the first worker thread executes on a first virtual machine, the second worker thread executes on a second virtual machine, and the third worker thread executes on a third virtual machine, the first, second, and third processing virtual machines comprising separate virtual machines.
The method 700 can further include, wherein the operation 704 includes copying a pointer to the first data. The method 700, can further include routing the first data and the first copied data to respective output queues coupled between second and third processing cores and a hardware queue manager. The method 700, can further include receiving indication information to indicate data to be copied from a plurality of input queues and providing the data to be copied to data duplication circuitry.
“Circuitry” as used herein means electrical and/or electronic components, such as can include one or more transistors, resistors, capacitors, inductors, diodes, logic gates, multiplexers, oscillators, buffers, modulators, regulators, amplifiers, demodulators, radios (e.g., transmit or receive radios or transceivers), sensors (e.g., a transducer that converts one form of energy (e.g., light, heat, electrical, mechanical, or other energy) to another form of energy), or the like.
In one embodiment, processor 810 has one or more processing cores 812 and 812N, where 812N represents the Nth processing core inside processor 810 where N is a positive integer. In one embodiment, system 800 includes multiple processors including 810 and 805, where processor 805 has logic similar to or identical to the logic of processor 810. In some embodiments, processing core 812 includes, but is not limited to, pre-fetch logic to fetch instructions, decode logic to decode the instructions, execution logic to execute instructions and the like. In some embodiments, processor 810 has a cache memory 816 to cache instructions and/or data for system 800. Cache memory 816 may be organized into a hierarchal structure including one or more levels of cache memory.
In some embodiments, processor 810 includes a memory controller 814, which is operable to perform functions that enable the processor 810 to access and communicate with memory 830 that includes a volatile memory 832 and/or a non-volatile memory 834. In some embodiments, processor 810 is coupled with memory 830 and chipset 820. Processor 810 may also be coupled to a wireless antenna 878 to communicate with any device configured to transmit and/or receive wireless signals. In one embodiment, the wireless antenna interface 878 operates in accordance with, but is not limited to, the IEEE 802.11 standard and its related family, Home Plug AV (HPAV), Ultra Wide Band (UWB), Bluetooth, WiMax, or any form of wireless communication protocol.
In some embodiments, volatile memory 832 includes, but is not limited to, Synchronous Dynamic Random Access Memory (SDRAM), Dynamic Random Access Memory (DRAM), RAMBUS Dynamic Random Access Memory (RDRAM), and/or any other type of random access memory device. Non-volatile memory 834 includes, but is not limited to, flash memory, phase change memory (PCM), read-only memory (ROM), electrically erasable programmable read-only memory (EEPROM), or any other type of non-volatile memory device.
Memory 830 stores information and instructions to be executed by processor 810. In one embodiment, memory 830 may also store temporary variables or other intermediate information while processor 810 is executing instructions. The memory 830 is an example of a machine-readable medium. While a machine-readable medium may include a single medium, the term “machine-readable medium” may include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers).
The term “machine-readable medium” may include any medium that is capable of storing, encoding, or carrying instructions for execution by a machine (e.g., circuitry, or the like) and that cause the machine to perform any one or more of the techniques of the present disclosure; or that is capable of storing, encoding or carrying data structures used by or associated with such instructions. In other words, the circuitry discussed herein can include instructions and can therefore be termed a machine-readable medium in the context of various embodiments. Other non-limiting machine-readable medium examples may include solid-state memories, and optical and magnetic media. Specific examples of machine-readable media may include: non-volatile memory, such as semiconductor memory devices (e.g., Electrically Programmable Read-Only Memory (EPROM), Electrically Erasable Programmable Read-Only Memory (EEPROM)) and flash memory devices; magnetic disks, such as internal hard disks and removable disks; magneto-optical disks; and CD-ROM and DVD-ROM disks.
In the illustrated embodiment, chipset 820 connects with processor 810 via Point-to-Point (PtP or P-P) interfaces 817 and 822. Chipset 820 enables processor 810 to connect to other elements in system 800. In some embodiments of the invention, interfaces 817 and 822 operate in accordance with a PtP communication protocol such as the Intel® QuickPath Interconnect (QPI) or the like. In other embodiments, a different interconnect may be used.
In some embodiments, chipset 820 is operable to communicate with processor 810, 805N, display device 840, and other devices. Chipset 820 may also be coupled to a wireless antenna 878 to communicate with any device configured to transmit and/or receive wireless signals.
Chipset 820 connects to display device 840 via interface 826. Display device 840 may be, for example, a liquid crystal display (LCD), a plasma display, cathode ray tube (CRT) display, or any other form of visual display device. In some embodiments of the invention, processor 810 and chipset 820 are merged into a single SOC. In addition, chipset 820 connects to one or more buses 850 and 855 that interconnect various elements 874, 860, 862, 864, and 866. Buses 850 and 855 may be interconnected together via a bus bridge 872. In one embodiment, chipset 820 couples with a non-volatile memory 860, a mass storage device(s) 862, a keyboard/mouse 864, and a network interface 866 via interface 824 and/or 804, etc.
In one embodiment, mass storage device 862 includes, but is not limited to, a solid state drive, a hard disk drive, a universal serial bus flash memory drive, or any other form of computer data storage medium. In one embodiment, network interface 866 is implemented by any type of well-known network interface standard including, but not limited to, an Ethernet interface, a universal serial bus (USB) interface, a Peripheral Component Interconnect (PCI) Express interface, a wireless interface and/or any other suitable type of interface. In one embodiment, the wireless interface operates in accordance with, but is not limited to, the IEEE 802.11 standard and its related family, Home Plug AV (HPAV), Ultra Wide Band (UWB), Bluetooth, WiMax, or any form of wireless communication protocol.
While the components shown in
The present subject matter may be described by way of several examples.
Example 1 can include a hardware queue manager comprising an input interface coupled to a plurality of input queues, the input interface to receive first data from at least a first worker thread through an input queue of the plurality of input queues, queue duplication circuitry to generate a copy of at least a portion of the first data to create first copied data, and an output interface coupled to a plurality of output queues, the output interface to (a) provide the first copied data to a second worker thread coupled to a first output queue of the plurality of output queues, and (b) provide at least a portion of the first data to a third worker thread coupled to a second output queue of the plurality of output queues.
In Example 2, Example 1 may further include, wherein the first worker thread executes on a first processing core, the second worker thread executes on a second processing core, and the third worker thread executes on a third processing core, the first, second, and third processing cores comprising separate processing cores.
In Example 3, Example 2 may further include, wherein the first worker thread executes on a first virtual machine, the second worker thread executes on a second virtual machine, and the third worker thread executes on a third virtual machine, the first, second, and third processing virtual machines comprising separate virtual machines.
In Example 4, at least one of Examples 1-3 may further include, wherein the queue duplication circuitry is to copy a pointer to the first data.
In Example 5, at least one of Examples 1-4 may further include data copy circuitry to copy the first data.
In Example 6, at least one of Examples 1-5 may further include queue routing control circuitry to route the first data and the first copied data to respective output queues of the plurality of output queues coupled between the second and third processing cores and the hardware queue manager.
In Example 7, Example 6 may further include sequencer circuitry to receive indication information to indicate data to be copied from the plurality of input queues, and provide the data to be copied to the third processing core, and a copy of the data to be copied, to the duplication circuitry.
Example 8 includes a non-transitory machine-readable medium including instructions that, when executed on a machine, cause the machine to perform operations including receiving, from an input of a plurality of input queues coupled to a hardware queue manager, first data from at least a first worker thread, generating a copy of at least a portion of the first data to create first copied data, and (a) providing the first copied data to a second worker thread coupled to a first output queue of a plurality of output queues, and (b) providing at least a portion of the first data to a third worker thread coupled to a second output queue of the plurality of output queues.
In Example 9, Example 8 may further include, wherein the second worker thread is a debug thread and the third worker thread includes packet inspection, network address translation, intrusion detection, ad insertion, or routing.
In Example 10, at least one of Examples 8-9 may further include, wherein the first worker thread executes on a first processing core, the second worker thread executes on a second processing core, and the third worker thread executes on a third processing core, the first, second, and third processing cores comprising separate processing cores.
In Example 11, at least one of Example 8-9 may further include, wherein the first worker thread executes on a first virtual machine, the second worker thread executes on a second virtual machine, and the third worker thread executes on a third virtual machine, the first, second, and third processing virtual machines comprising separate virtual machines.
In Example 12, at least one of Examples 8-11 may further include, wherein generating the copy of at least the portion of the first includes copying a pointer to the first data.
In Example 13, at least one of Examples 8-12 may further include, wherein the operations further comprise routing the first data and the first copied data to respective output queues of the plurality of output queues coupled between second and third processing cores and a hardware queue manager.
In Example 14, Example 13 may further include receiving indication information to indicate data to be copied from the plurality of input queues and providing the data to be copied to data duplication circuitry.
Example 15 includes a method for network security monitoring, the method comprising receiving, at an input interface of a hardware queue manager and through an input queue of a plurality of input queues coupled to the hardware queue manager, first data from a first worker thread, generating, at the hardware queue manager, a copy of at least a portion of the first data to create first copied data, and (a) providing the first copied data to a second worker thread to perform network security monitoring, the second worker thread coupled to a first output queue of a plurality of output queues coupled to the hardware queue manager, and (b) providing at least a portion of the first data to a third worker thread through a second output queue of the plurality of output queues coupled to the hardware queue manager.
In Example 16, Example 15 may further include, wherein the second worker thread is a debug thread and the third worker thread includes packet inspection, network address translation, intrusion detection, ad insertion, or routing.
In Example 17, at least one of Examples 15-16 may further include, wherein the first worker thread executes on a first processing core, the second worker thread executes on a second processing core, and the third worker thread executes on a third processing core, the first, second, and third processing cores comprising separate processing cores.
In Example 18, at least one of Examples 15-16 may further include, wherein the first worker thread executes on a first virtual machine, the second worker thread executes on a second virtual machine, and the third worker thread executes on a third virtual machine, the first, second, and third processing virtual machines comprising separate virtual machines.
In Example 19, at least one of Examples 15-18 may further include, wherein generating the copy of at least the portion of the first includes copying a pointer to the first data.
In Example 20, at least one of Examples 15-19 may further include, wherein the operations further comprise routing the first data and the first copied data to respective output queues of the plurality of output queues coupled between second and third processing cores and a hardware queue manager.
In Example 21, Example 20 may further include receiving indication information to indicate data to be copied from the plurality of input queues and providing the data to be copied to data duplication circuitry.
Example 22 includes a system comprising a network function virtualization infrastructure comprising a plurality of processing cores to perform operations of virtual network functions and security circuitry to monitor and copy traffic of the processing cores, a controller coupled to the security circuitry to provide a security policy defining traffic to be monitored and indicating whether the security circuitry is to copy data or a pointer to the data of the traffic, and a network analysis tool to determine when the copied traffic includes a bug, a dropped packet, or a security threat.
In Example 23, Example 22 may further include, wherein the controller is further to determine which core of the plurality of processing cores performs a specific virtual network function of the plurality of virtual network functions, and wherein the security policy, when executed, is to indicate the determined core, and the security circuitry is to copy the traffic from the determined core.
In Example 24, Example 23 may further include a switch or router and wherein the controller is further to provide the security policy to the security circuitry through the switch or router.
In Example 25, Example 24 may further include a plurality of memory queues in a memory, to receive the copied traffic from the security circuitry and provide the copied traffic to the network analysis circuitry.
Each of these non-limiting examples can stand on its own, or can be combined in various permutations or combinations with one or more of the other examples. “Non-transitory” merely means that the medium comprises a form of tangible media.
The above detailed description includes references to the accompanying drawings, which form a part of the detailed description. The drawings show, by way of illustration, specific embodiments in which methods, apparatuses, and systems discussed herein can be practiced. These embodiments are also referred to herein as “examples.” Such examples can include elements in addition to those shown or described. However, this document is intended to include examples in which only those elements shown or described are provided. Moreover, examples using any combination or permutation of those elements shown or described (or one or more aspects thereof) herein, either with respect to a particular example (or one or more aspects thereof), or with respect to other examples (or one or more aspects thereof) shown or described herein, are also included as example embodiments.
In this document, the terms “a” or “an” are used, as is common in patent documents, to include one or more than one, independent of any other instances or usages of “at least one” or “one or more.” In this document, the term “or” is used to refer to a nonexclusive or, such that “A or B” includes “A but not B,” “B but not A,” and “A and B,” unless otherwise indicated. In this document, the terms “including” and “in which” are used as the plain-English equivalents of the respective terms “comprising” and “wherein.” Also, in the following claims, the terms “including” and “comprising” are open-ended, that is, a system, device, article, composition, formulation, or process that includes elements in addition to those listed after such a term in a claim are still deemed to fall within the scope of that claim. Moreover, in the following claims, the terms “first,” “second,” and “third,” etc. are used merely as labels, and are not intended to impose numerical requirements on their objects.
The above description is intended to be illustrative, and not restrictive. For example, the above-described examples (or one or more aspects thereof) may be used in combination with each other. Other embodiments can be used, such as by one of ordinary skill in the art upon reviewing the above description. The Abstract is provided to allow the reader to quickly ascertain the nature of the technical disclosure. It is submitted with the understanding that it will not be used to interpret or limit the scope or meaning of the claims. Also, in this Description of Embodiments, various features may be grouped together to streamline the disclosure. This should not be interpreted as intending that an unclaimed disclosed feature is essential to any claim. Rather, inventive subject matter may lie in less than all features of a particular disclosed embodiment. Thus, the following claims are hereby incorporated into the Description of Embodiments as examples or embodiments, with each claim standing on its own as a separate embodiment, and it is contemplated that such embodiments can be combined with each other in various combinations or permutations. The scope of the invention should be determined regarding the appended claims, along with the full scope of equivalents to which such claims are entitled.
This application is a divisional of U.S. patent application Ser. No. 15/607,121, filed on May 26, 2017. This application is herein incorporated by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
Parent | 15607121 | May 2017 | US |
Child | 17213132 | US |