The present disclosure pertains to systems and methods for integration of a parallel redundancy protocol (“PRP”) in a software defined network (“SDN”).
Non-limiting and non-exhaustive embodiments of the disclosure are described, including various embodiments of the disclosure, with reference to the figures, in which:
PRP is a network protocol standard for Ethernet that provides redundancy and protects against single points of failure. A PRP-enabled device has two communication ports, each of which is attached to a separate local area network (“LAN”). To avoid a single point of failure, the two LANs may use distinct physical links, and thus may be assumed to be fail-independent. As long as one path is operational, the destination application always receives at least one frame. The redundancy provided by PRP may be implemented by network devices, and thus may be invisible to applications. Similar protocols, such as High-availability Seamless Redundancy (HSR), may provide benefits in terms of redundancy and reliability. PRP offers certain advantages over HSR, such as the ability to connect non-PRP devices to the same network. In contrast, when a non-HSR device is connected to a HSR network, the device typically connects through a redundancy box (RedBox).
High reliability and redundant communication networks offer advantages that may be utilized in an infrastructure system (e.g., electric power systems, telecommunication systems, etc.), manufacturing systems, alarm systems, and a variety of other applications. Such networks, which may be referred to as operational technology (“OT”) networks, may manage, monitor, and control a wide range of devices. OT networks may comprise a large number of machine-to-machine communications, and as such, large volumes of data may be generated and transmitted. Management of such networks may present a variety of challenges.
OT networks may utilize a variety of technologies, including SDN networking technologies. In an SDN, a controller may regulate communications on the network. SDN networking technologies offer a variety of advantages, such as deny-by-default security, latency guarantees, deterministic transport capabilities, redundancy, and fail-over planning, etc. An SDN allows a programmatic change control platform, which allows an entire communication network to be managed as a single asset, simplifies the understanding of the network, and enables continuous monitoring of a network. In an SDN, the systems that decide where the traffic is sent (i.e., the control plane) are separated from the systems that perform the forwarding of the traffic in the network (i.e., the data plane).
The control plane may be used to optimize the usage of network resources by creating specific data flows through the communication network. A data flow, as the term is used herein, refers to a set of parameters used to match and take action based on network packet contents. Data flows may permit dedicated paths based on a variety of criteria that offer significant control and precision to operators of the network. In contrast, in large traditional networks, trying to match a network-discovered path with an application-desired data path may be a challenging task involving changing configurations in many devices. To compound this problem, the management interfaces and feature sets used on many devices are not standardized. Further, network administrators often need to reconfigure the network to avoid loops, gain route convergence speed, and prioritize certain classes of applications.
Significant complexity in managing a traditional network in the context of an OT network arises from the fact that each network device (e.g., a switch or router) includes both control logic and data forwarding logic. For example, in a traditional network router, routing protocols such as Routing Information Protocol (RIP) or Open Shortest Path First (OSPF) constitute the control logic that determines how a packet should be forwarded. The paths determined by the routing protocol are encoded in routing tables, which are then used to forward packets. Similarly, in a Layer 2 device such as a network bridge (or network switch), configuration parameters and/or a Spanning Tree Algorithm (STA) constitute the control logic that determines the path of the packets. Thus, the control plane in a traditional network is distributed in the switching fabric (network devices), and as a consequence, changing the forwarding behavior of a network involves changing configurations of many (potentially all) network devices.
In contrast, in an SDN, a controller embodies the control plane and determines how packets (or frames) should flow (or be forwarded) in the network. The controller communicates this information to the network devices, which constitute the data plane. The controller may set forwarding tables in network devices that establish how data is to be routed. This enables centralized configuration and management of a network. In addition to simplifying the management of a network, an SDN architecture may also enable monitoring and troubleshooting features that may be beneficial for use in OT networks. Such features may include, but are not limited to: mirroring a data-selected flow rather than mirroring a whole port; alarming on bandwidth when it gets close to saturation; providing metrics (e.g., counters and meters for quality of service, packet counts, errors, drops, or overruns, etc.) for a specified flow; and permitting the monitoring of specified applications rather than monitoring based on virtual local area networks (VLAN) or media access control (MAC) addresses.
Use of PRP or other redundant communication protocols in an SDN may involve configuration of parallel redundant communication flows between devices. If such flows are not correctly implemented, PRP communications may not be correctly routed in the data plane, and the benefits afforded by use of PRP may be lost. Implementing parallel redundant paths adds to the burden of configuring and maintaining an SDN.
The inventors of the present application have recognized that certain advantages may be realized by automating the identification and configuration of PRP devices in an SDN. The techniques disclosed herein may be used in a variety of specific applications, including:
The embodiments of the present disclosure will be best understood by reference to the drawings, wherein like parts are designated by like numerals throughout. It will be readily understood that the components of the disclosed embodiments, as generally described and illustrated in the figures herein, could be arranged and designed in a wide variety of different configurations. Thus, the following detailed description of the embodiments of the systems and methods of the disclosure is not intended to limit the scope of the disclosure, as claimed, but is merely representative of possible embodiments of the disclosure. In addition, the steps of a method do not necessarily need to be executed in any specific order, or even sequentially, nor need the steps be executed only once unless otherwise specified.
In some cases, well-known features, structures, or operations are not shown or described in detail. Furthermore, the described features, structures, or operations may be combined in any suitable manner in one or more embodiments. It will also be readily understood that the components of the embodiments as generally described and illustrated in the figures herein could be arranged and designed in a wide variety of different configurations.
Several aspects of the embodiments described may be implemented as software modules or components. As used herein, a software module or component may include any type of computer instruction or computer-executable code located within a memory device and/or transmitted as electronic signals over a system bus or wired or wireless network. A software module or component may, for instance, comprise one or more physical or logical blocks of computer instructions, which may be organized as a routine, program, object, component, data structure, etc. that performs one or more tasks or implements particular abstract data types.
In certain embodiments, a particular software module or component may comprise disparate instructions stored in different locations of a memory device, which together implement the described functionality of the module. Indeed, a module or component may comprise a single instruction or many instructions, and may be distributed over several different code segments, among different programs, and across several memory devices. Some embodiments may be practiced in a distributed computing environment where tasks are performed by a remote processing device linked through a communications network. In a distributed computing environment, software modules or components may be located in local and/or remote memory storage devices. In addition, data being tied or rendered together in a database record may be resident in the same memory device, or across several memory devices, and may be linked together in fields of a record in a database across a network.
Embodiments may be provided as a computer program product including a non-transitory computer and/or machine-readable medium having stored thereon instructions that may be used to program a computer (or another electronic device) to perform processes described herein. For example, a non-transitory computer-readable medium may store instructions that, when executed by a processor of a computer system, cause the processor to perform certain methods disclosed herein. The non-transitory computer-readable medium may include, but is not limited to, hard drives, floppy diskettes, optical disks, CD-ROMs, DVD-ROMs, ROMs, RAMs, EPROMs, EEPROMs, magnetic or optical cards, solid-state memory devices, or other types of machine-readable media suitable for storing electronic and/or processor-executable instructions.
The data plane 104 includes the plurality of network devices 106a-106d in communication with one another via a plurality of physical links 120a-120d. In various embodiments, the network devices 106a-106d may be embodied as switches, multiplexers, and other types of network devices. The physical links 120a-120d may be embodied as Ethernet, fiber optic, and other forms of data communication channels. As illustrated, the physical links 120a-120d between the network devices 106a-106d may provide redundant connections such that a failure of one of the physical links 120a-120d is incapable of completely blocking communication with an affected network device. In some embodiments, the physical links 120a-120d may provide an N−1 redundancy or better.
Data consuming/producing devices 116a-116c may represent a variety of devices within that produce or consume data. For example, data consuming/producing devices 116a-116c may, for example, be embodied as a pair of transmission line relays configured to monitor an electrical transmission line. The transmission line relays may monitor various aspects of the electric power flowing through the transmission line (e.g., voltage measurements, current measurements, phase measurements, synchrophasors, etc.) and may communicate the measurements to implement a protection strategy for the transmission line. Traffic between the transmission line relays may be routed through the data plane 104 using a plurality of data flows 118 implemented by controller 112. Data consuming/producing devices 116a-116c may be embodied by a wide range of devices consistent with embodiments of the present disclosure.
Applications 110a-110c may represent a variety of applications operating in an applications plane. In the SDN architecture illustrated in
Data consuming/producing devices 116a-116c may transmit information to one or more network devices embodied as SDN switches 106a-106d. The data from data consuming/producing devices 116a-116c may be routed by data plane 104 according to the plurality of data flows 118 specified by controller 112.
Controller 112 may support automated configuration 108 based on approved services, parameters, and data metrics. Upon detection of a communication associated with an approved service and parameters, controller 112 may generate corresponding data flow(s) 118. In some embodiments, the automated configuration may occur during a commissioning phase, while in other embodiments, automated configuration 108 may be ongoing while system 100 is in operation. Information about automating configuration during a commissioning phase is described in U.S. Pat. No. 9,923,779, which is incorporated herein by reference.
A controller 202 is in communication with SDN switch fabric 204. Controller 202 may program devices in SDN switch fabric 204 using communication flows. The communication flows may be used by SDN switch fabric 204 to route traffic between devices in the SDN switch fabric 204 and devices connected to the SDN switch fabric 204. Controller 202 may utilize a variety of protocols (e.g., OpenFlow) and technologies to configure SDN switch fabric 204.
Host 206 may be configured to implement PRP. Duplicate packets may be transmitted to host 206 using communication links 208 and 210. If host 206 receives duplicate packets, one of the duplicate packets may be retained and one may be discarded. The retained packet may be passed along to an application operating on host 206. Host 206 may embody a variety of types of equipment. In some embodiments, host 206 may be embodied as an intelligent electronic device (IED). As used herein, an IED may refer to any microprocessor-based device that monitors, controls, automates, and/or protects monitored equipment within an electric power system. Such devices may include, for example, remote terminal units, differential relays, distance relays, directional relays, feeder relays, overcurrent relays, transformer relays, voltage regulator controls, voltage relays, breaker failure relays, generator relays, motor relays, automation controllers, bay controllers, meters, recloser controls, communications processors, computing platforms, programmable logic controllers (PLCs), programmable automation controllers, input and output modules, and the like. The term IED may be used to describe an individual IED or a system comprising multiple IEDs.
In operation, controller 202 may determine whether host 206 is configured to implement PRP by monitoring traffic from host 206. For example, when Port A egresses a packet into Port 1, controller 202 may intercept and analyze the packet. Analysis of the packet may allow controller 202 to determine that Port A is emitting PRP traffic. In addition, traffic from Port B may be analyzed by controller 202. In various embodiments, Port A and Port B may be connected to separate LANs, which may be comprised within SDN switch fabric 204. Systems consistent with the present disclosure may identify that the host is in a PRP configuration and identify which port is connected to each LAN so that packets are routed to the correct interface (e.g., packets for LAN A are routed to the port associated with LAN A, and packets for LAN B are routed to the port associated with LAN B).
In some embodiments, controller 202 may monitor traffic from a newly discovered device to determine whether the device is configured to communicate using PRP. For example, upon connection of host 206, traffic from host 206 may be routed to controller 202, and controller 202 may determine whether the traffic includes pairs of packets that conform to PRP. Such embodiments may allow for the passive detection of PRP devices. In other embodiments described in greater detail below, a more active process may be employed to identify devices configured to communicate using PRP.
Based on detection of PRP traffic emitted by host 206, controller 202 may implement various settings and communication flows. In contrast, typical systems require that users know PRP device configurations, correctly configure the SDN to pass the redundant traffic through separate LANs and to the correct ports on the remote device, and correctly connect this device's ports to the network. Manual configuration is error prone, and mistakes are time consuming to troubleshoot. In contrast, by automatically determining that host 206 is using PRP, the configuration burden may be eased by implementing PRP communication between host 206 and another device (not shown) using PRP without user intervention.
At 252, traffic from a host may be monitored. If a system implementing method 250 determines at 254 that the traffic emitted by the host comprises non-PRP packets, the method may end because the host is not using PRP. A determination of whether the traffic comprises PRP packets at 256 may be made based on both the number of packets received as well as the packet structure. PRP packets include a redundancy control trailer (RCT) that is added to each frame and that includes a sequence number to allow a receiving device to discard duplicate frames. Upon receipt of a pair of PRP packets from a host, method 250 may determine that the host is using PRP at 258.
At 260, a system implementing method 250 may determine whether a destination for the PRP traffic is also configured using PRP. In some embodiments, traffic from the destination may be monitored to determine if such traffic is PRP traffic. Alternative, techniques discussed in greater detail in connection with
Based on the determination that a host and a destination are both using PRP, a system implementing method 250 may determine whether to establish PRP communication flows at 262. In some embodiments, communication flows may be established without user intervention. In other words, a system implementing method 250 may automatically establish the communication flows associated with PRP communication from the host. In other embodiments, an operator may be prompted to accept a communication flow before creation of the communication flow.
If PRP communication flows are to be established, they may be identified and implemented at 264. Automation of the communication PRP flows may reduce the configuration burden associated with implementing PRP. A system implementing method 250 may automatically implement communication flows using paths through an SDN switch fabric that rely on physically distinct devices to avoid single points of failure and to achieve other benefits.
In some embodiments, part of the automated configuration may include configuring supervisory frames transmitted between PRP devices. PRP uses supervisory frames with Ethertype 0x88FB to manage how traffic is sent between PRP devices. The proper routing of supervisory frames is necessary to ensure proper communication between PRP devices. Supervisory frames may facilitate use of PRP devices and non-PRP in the same system.
At 266, method 250 may determine if there are other destinations to configure. Using this approach, communications for each PRP-enabled destination may be configured, and communications for each non-PRP-enabled destination may also be configured. Stated in other terms, each communication circuit may be identified and configured, whether or not the destination in the communication circuit is PRP-enabled.
Using either packet that is received on port 5 or port 6 from 306, controller 302 may identify that the traffic is PRP and determine that host 308 is currently unknown. Controller 302 may attempt to discover host 308 and determine if host 308 is in a PRP configuration. To make this determination, controller 302 may generate a custom ARP packet that includes a redundancy control trailer (RCT) that elicits a response from host 308. If the response consists of a single non-PRP packet (received on either port 3 or port 4), controller 302 can determine that host 308 is not configured to use PRP. If the response consists of a pair of PRP packets, with one packet being received by port 3 and the other by port 4, controller 302 can determine that host 308 is operating using PRP.
Upon receipt of PRP traffic destined for a previously unknown host, an SDN controller implementing method 350 may generate an ARP request including an RCT trailer and directed to the previously unknown host at 354. The ARP request may cause the previously known host to respond and provide its link layer address, such as a MAC address.
By causing the previously unknown host to respond to the ARP request, a device implementing method 350 may determine whether the previously unknown host is using PRP. If a single non-PRP packet is received at 356, a system implementing method 350 may confirm that the host is not using PRP at 360.
At 368, appropriate non-PRP communication flows may be identified and implemented. Where one device uses PRP and another device does not use PRP, only a single communication path between the devices may be implemented because both devices must use PRP to obtain the advantages provided by PRP.
If a pair of PRP packets are received in response to the ARP request at 358, a system implementing method 350 may confirm that the previously unknown host is using PRP at 362. Based on the determination that the host is using PRP, a system implementing method 350 may determine whether to establish PRP communication flows at 364. If PRP communication flows are to be established, appropriate PRP communication flows may be identified and implemented at 366.
In various embodiments, a controller may operate on a host in system 400 that also performs other tasks. An in-band SDN controller 402 is unlikely to determine that it is operating a host device 406 configured for PRP communication. To detect configuration, a packet may be generated and egressed from controller 402 such that it is forwarded back to the controller 402. Flows that are forwarded to the SDN controller are communicated through the transport layer security (“TLS”) connection the SDN controller has with the switch receiving the traffic. As such, the packets are not visible in their original form once they reach C1 or C2, but rather the packets are a byte stream in an encrypted TCP connection. The packets are received on port 1 and/or port 2 then are forwarded via the SDN protocol.
In one specific embodiment, controller 402 may send a PRP packet from SDN switch fabric 404 to itself to that will cause its network stack to respond. If PRP is configured, indicators will be returned (e.g., the packet will be formatted according to the PRP protocol) with the response that will allow controller 402 to determine whether communication according to the PRP protocol is enabled. PRP communication is handled Layer 2 in the Open Systems Interconnection (“OSI”) model, and as such indicators of PRP communication are removed by the Network Interface Card (“NIC”) before the packet is passed up the network stack. If a packet is returned over the TLS controller channel, however, the indicators of PRP communication are preserved can be analyzed by the controller.
Determining whether an SDN is operating on a host configured to communicate using PRP may be useful in a variety of situations. For example, controller 402 may run a syslog server, and host 406 may have a message to forward to controller 402 to include in the syslog. In this case, the syslog messages will be duplicated by PRP to provide redundancy. If the PRP path is not properly configured, the benefit of this redundancy cannot be utilized. Upon discovering the PRP configuration of the controller 402, network planning can be performed correctly to utilize this redundancy.
Using SDN allows a controller (not shown) to configure the switches, ensuring that network traffic follows a specified path, while still providing alternative paths for purposes of additional fault tolerance. For example, in typical operation, packets from PRP port 1A may be routed to SDN switch 1, then to SDN switch 2, and then to PRP port 2A. In the event of a failure of the link between SDN switch 1 and SDN switch 2, traffic may be routed from PRP port 1A to SDN switch 1, to SDN switch 4, to SDN switch 2, and finally to PRP port 2A.
Various embodiments consistent with the present disclosure may facilitate planning and visualization of PRP systems to route traffic through distinct nodes. For example, PRP port 1A is forwarded through switches 1 and 2 to PRP port 2A. PRP port 1B's outgoing traffic is forwarded through switch 3 and then switch 4 before being delivered to PRP port 2B. Using this scheme, any one switch can fail and PRP will still function as intended. This is also functionally equivalent to the typical practice of having two physically separate networks for PRP traffic.
The interconnections illustrated using dashed lines offer many paths through SDN switch fabric 504 that could result in a single point of failure. For example, traffic sent from PRP port 1A may be routed to switch 1, then to switch 2, then to PRP port 2A; and traffic sent from PRP port 1B may be routed to switch 3, then to switch 2, then to switch 4, and then to PRP port 2B. This routing scheme includes a single point of failure (i.e., switch 2). As such, if switch 2 fails, both PRP paths fail. In various embodiments consistent with the present disclosure, a visualization system may identify single points of failure and/or suggest paths that utilize physical distinct nodes.
While system 600 provides physically separate paths for PRP communications, systems and methods consistent with the present disclosure may optimize the configuration to improve efficiency using SDN. As described in connection with
All of the redundant communication paths in system 650 pass through distinct nodes, so that there is no single point of failure for redundant paths. For example, if SDN switch 1 fails, traffic between the station buses of host 602 and host 604 may still pass between PRP port 1B and PRP port 3B using SDN switch 2 and SCN switch 4. Similarly, if any other SDN switch fails, traffic may still pass through a redundant path without interruption. This level of redundancy is achieved using fewer SDN switches, and therefore can be implemented at a lower cost in comparison to system 600 illustrated in
SDN controller 702 includes a communications interface 718 to communicate with SDN switch fabric 720. Communications interface 718 may facilitate communications with multiple devices (not shown). SDN controller 702 may further include a time input 704, which may be used to receive a time signal (e.g., a common time reference) allowing SDN controller 702 to apply a time-stamp to received data. In certain embodiments, a common time reference may be received via communications interface 718, and accordingly, a separate time input may not be required. One such embodiment may employ the IEEE 1588 protocol. A data bus 726 may facilitate communication among various components of SDN controller 702.
Processor 706 may be configured to process communications received via communications interface 718 and time input 704 and to coordinate the operation of the other components of SDN controller 702. Processor 706 may operate using any number of processing rates and architectures. Processor 706 may be configured to perform any of the various algorithms and calculations described herein. Processor 706 may be embodied as a general-purpose integrated circuit, an application-specific integrated circuit, a field-programmable gate array, and/or any other suitable programmable logic device.
Instructions to be executed by processor 706 may be stored in random access memory (RAM) 714. Such instructions may include information for processing routing and processing data packets received via communications interface 718 based on a plurality of traffic flows. In some embodiments, traffic that does not match a plurality of traffic flows may be routed to SDN controller 702 for additional analysis and/or action.
A user-interface subsystem 716 may be configured to receive from a user various types of information relating to configuring SDN switch fabric 720. In some embodiments, the user-interface subsystem 716 may be configured to confirm or reject the creation of communication flows related to PRP communication. The communication flows to be confirmed may be identified by SDN controller 702 during the operation of system 700.
A traffic routing subsystem 712 may be configured to generate a variety of communication flows implemented by SDN switch fabric 720. The traffic routing subsystem 712 may specify the configuration of a variety of intermediate devices (e.g., routers, switches, multiplexers, etc.), separating communicating hosts 722 and 724. The traffic routing subsystem 712 may be configured to generate physically distinct paths for traffic flows among PRP traffic in system 700. For example, host 722 may provide a redundant stream of data to host 724. A first redundant stream of data may be transmitted from PRP port 1A to port 5 and from port 3 to PRP port 2A. A second redundant stream of data may be transmitted from PRP port 1B to port 6 and from port 4 to PRP port 2B. In some embodiments the communication flows may be implemented such that parallel communication flows between host 722 and host 724 pass through distinct physical links and/or distinct switches in SDN switch fabric 720.
A PRP identification subsystem 708 may identify traffic in system 700. In some embodiments, PRP identification subsystem 708 may monitor traffic transmitted by host 722 to host 724 to determine that the traffic comprises at least one pair of redundant data packets that conforms to PRP. Packets that conform to PRP may include information that allows a receiving device to identify and discard duplicate frames. Such information may include a sequence number and/or other type of information. In various embodiments, PRP identification subsystem 708 may allow for passive detection of PRP traffic. Alternatively, SDN controller 702 may interrogate host 722 and/or host 724 to solicit a response in PRP format. In one specific embodiment, the interrogation may comprise generating an ARP request directed to one of host 722 and 724 and analyzing a response.
A PRP optimization subsystem 710 may generate configurations of traffic flows that utilize distinct communication links and/or switches to avoid single points of failure. Such optimization may improve redundancy within a system without requiring additional hardware, as discussed in connection with
A PRP visualization subsystem 726 may provide a visual representation of networks to facilitate design and planning of a network configurations. In some embodiments, a visualization may identify single points of failure and/or suggest paths that utilize physical distinct nodes for PRP communications. Such a visualization may allow a user to ensure that communication flows avoid single points of failure.
While specific embodiments and applications of the disclosure have been illustrated and described, it is to be understood that the disclosure is not limited to the precise configurations and components disclosed herein. Accordingly, many changes may be made to the details of the above-described embodiments without departing from the underlying principles of this disclosure. The scope of the present invention should, therefore, be determined only by the following claims.
Number | Name | Date | Kind |
---|---|---|---|
6747957 | Pithawala | Jun 2004 | B1 |
7218632 | Bechtolsheim | May 2007 | B1 |
7266849 | Gregory | Sep 2007 | B1 |
7376831 | Kollmyer | May 2008 | B2 |
7872983 | Lai | Jan 2011 | B2 |
8347079 | Cho | Jan 2013 | B2 |
8520670 | Giniger | Aug 2013 | B1 |
8553544 | Lai | Oct 2013 | B2 |
8800044 | Raad | Aug 2014 | B2 |
8989380 | Belser | Mar 2015 | B1 |
9038151 | Chua | May 2015 | B1 |
9237129 | Ling | Jan 2016 | B2 |
9286171 | Cardona | Mar 2016 | B2 |
9432255 | Hasan | Aug 2016 | B1 |
9432380 | Margalit | Aug 2016 | B2 |
9680588 | Connolly | Jun 2017 | B2 |
9686125 | Smith | Jun 2017 | B2 |
9760504 | Chinnakkonda Vidyapoornachary | Sep 2017 | B2 |
9769060 | Dearien | Sep 2017 | B2 |
10863558 | Powers | Dec 2020 | B2 |
11165685 | Mullis | Nov 2021 | B2 |
11228521 | Mullis | Jan 2022 | B2 |
20020023217 | Wheeler | Feb 2002 | A1 |
20020172157 | Rhodes | Nov 2002 | A1 |
20030112821 | Cleveland | Jun 2003 | A1 |
20030125924 | Lines | Jul 2003 | A1 |
20030133443 | Klinker | Jul 2003 | A1 |
20030188159 | Josset | Oct 2003 | A1 |
20050025141 | Chao | Feb 2005 | A1 |
20050078672 | Caliskan | Apr 2005 | A1 |
20050192008 | Desai | Sep 2005 | A1 |
20070143838 | Milligan | Jun 2007 | A1 |
20070217344 | Krywaniuk | Sep 2007 | A1 |
20080005558 | Hadley | Jan 2008 | A1 |
20080080384 | Atkins | Apr 2008 | A1 |
20080184030 | Kelly | Jul 2008 | A1 |
20080204248 | Cam Winget | Aug 2008 | A1 |
20090252309 | Beyer | Oct 2009 | A1 |
20090257743 | Chung | Oct 2009 | A1 |
20090265470 | Shen | Oct 2009 | A1 |
20090285093 | Bolt | Nov 2009 | A1 |
20090313189 | Sun | Dec 2009 | A1 |
20100241608 | Huang | Sep 2010 | A1 |
20110085567 | Beecroft | Apr 2011 | A1 |
20110087952 | Marin | Apr 2011 | A1 |
20130035066 | Nylander | Feb 2013 | A1 |
20130067552 | Hawkes | Mar 2013 | A1 |
20130077477 | Daraiseh | Mar 2013 | A1 |
20130108259 | Srinivas | May 2013 | A1 |
20130159865 | Smith | Jun 2013 | A1 |
20130212285 | Hoffmann | Aug 2013 | A1 |
20130227646 | Haggerty | Aug 2013 | A1 |
20130250770 | Zou | Sep 2013 | A1 |
20130263247 | Jungck | Oct 2013 | A1 |
20130290735 | Rombouts | Oct 2013 | A1 |
20130294228 | Ahuja | Nov 2013 | A1 |
20140025945 | McGrew | Jan 2014 | A1 |
20140029451 | Nguyen | Jan 2014 | A1 |
20140064100 | Edwards | Mar 2014 | A1 |
20140112130 | Yang | Apr 2014 | A1 |
20140115706 | Silva | Apr 2014 | A1 |
20140129700 | Mehta | May 2014 | A1 |
20140153572 | Hampel | Jun 2014 | A1 |
20140160939 | Arad | Jun 2014 | A1 |
20140226467 | Park | Aug 2014 | A1 |
20140241345 | DeCusatis | Aug 2014 | A1 |
20140245387 | Colpo | Aug 2014 | A1 |
20140280834 | Medved | Sep 2014 | A1 |
20140325038 | Kis | Oct 2014 | A1 |
20140325649 | Zhang | Oct 2014 | A1 |
20140371941 | Keller | Dec 2014 | A1 |
20140376406 | Kim | Dec 2014 | A1 |
20150081762 | Mason | Mar 2015 | A1 |
20150112933 | Satapathy | Apr 2015 | A1 |
20150195190 | Shah Heydari | Jul 2015 | A1 |
20150312658 | Winzer | Oct 2015 | A1 |
20150363522 | Maurya | Dec 2015 | A1 |
20160043996 | Syed Mohamed | Feb 2016 | A1 |
20160119299 | Amulothu | Apr 2016 | A1 |
20160142427 | de los Reyes | May 2016 | A1 |
20160165454 | Li | Jun 2016 | A1 |
20160277195 | Maeda | Sep 2016 | A1 |
20160330076 | Tiwari | Nov 2016 | A1 |
20160337247 | Yao | Nov 2016 | A1 |
20160344592 | Cook | Nov 2016 | A1 |
20160359629 | Nadathur | Dec 2016 | A1 |
20170026187 | Ramatchandirane | Jan 2017 | A1 |
20170026225 | Smith | Jan 2017 | A1 |
20170026226 | Grussling | Jan 2017 | A1 |
20170026243 | Berner | Jan 2017 | A1 |
20170026252 | Dearien | Jan 2017 | A1 |
20170026276 | Dearien | Jan 2017 | A1 |
20170026291 | Smith | Jan 2017 | A1 |
20170026292 | Smith | Jan 2017 | A1 |
20170026349 | Smith | Jan 2017 | A1 |
20170201488 | Krywaniuk | Jul 2017 | A1 |
20170289117 | Powers | Oct 2017 | A1 |
20170302656 | Ramatchandirane | Oct 2017 | A1 |
20170317780 | Wood | Nov 2017 | A1 |
20170346630 | Nadathur | Nov 2017 | A1 |
20180027405 | Holtmanns | Jan 2018 | A1 |
20190116053 | Allan | Apr 2019 | A1 |
20190273717 | Dearien | Sep 2019 | A1 |
20200021513 | Hegde | Jan 2020 | A1 |
20200136999 | Levy-Abegnoli | Apr 2020 | A1 |
20210135972 | Mullis | May 2021 | A1 |
20210194791 | Mullis | Jun 2021 | A1 |
20220271854 | Maruyama | Aug 2022 | A1 |
Number | Date | Country |
---|---|---|
2765751 | Aug 2014 | EP |
20150051107 | May 2015 | KR |
2011133912 | Oct 2011 | WO |
2015038040 | Mar 2015 | WO |
Entry |
---|
Braun, Wolfgang, Menth, Michael, Software-Defined Networking Using OpenFlow: Protocols, Applications and Architectural Design Choices, Future Internet, May 12, 2014. |
Cahn, Adam, Hoyos, Juan, Hulse, Matthew, Keller, Eric, Software-Defined Energy Communication Networks: From Substation Automation to Future Smart Grids, Smart Grid Communications, IEEE Oct. 2013. |
Dally, William J., Virtual-Channel Flow Control, IEEE Transactions on Parallel and Distributed Systems, vol. 3, No. 2, Mar. 1992. |
Jain, Sushant, et al., B4: Experience with a Globally-Deployed Software Defined WAN, ACM SIGCOMM Computer Communication Review, vol. 43 Issue 4, pp. 3-14. Oct. 2013. |
Monaco, Matthew, Michel, Oliver, Keller, Eric, Applying Operating System Principles to SDN Controller Design, Hotnets '13, Nov. 2013. |
Drutskoy, Dmitry, Keller, Eric, Rexford, Jennifer, Scalable Network Virtualization in Software-Defined Networks, IEEE Internet Computing, vol. 17, Issue: 2, Nov. 27, 2012. |
Kuzniar, Maciej, et al., Automatic Failure Recovery for Software-Defined Networks, HotSDN '13, Aug. 16, 2013. |
Mizrahi, Tal, Moses, Yoram. ReversePTP: A Software Defined Networking Approach to Clock Synchronization, HotSDN '14, Aug. 22, 2014. |
Ramos, Ramon Marques, et al. SlickFlow: Resilient Source Routing in Data Centere Networks Unlocked by OpenFlow, 2013 IEEE 38th Conference on Local Computer Networks, Oct. 2013. |
Torhonen, Ville, Designing a Software-Defined Datacenter, Master of Science Thesis, Tampere University of Technology, May 2014. |
Yang, Qiaoyin and Smith, Rhett, Improve Protection Communications Network Reliability Through Software-Defined Process Bus, Jan. 2018. |
Dearien, Jason: “Setting Up a Fully Redundant RSTP-to-SDN Tie Point” Application Guide, vol. II AG2017-28, Sep. 22, 2017. |
Number | Date | Country | |
---|---|---|---|
20230066212 A1 | Mar 2023 | US |