This present disclosure relates generally to networks, and more particularly to speeding up of the transfer of data over medium to long distances by fundamentally modifying the nature of exchanges between sender and receiver devices. Methodology shift is from round-trip packet transfers to one-way slingshot-like file sending, as well as adopting principles of high performance computing to improve throughput.
This disclosure specifically describes a one-way information slingshot that saves complete, any-sized data files at near wire speed over a network tapestry to remote storage. The information slingshot can be integrated with existing network fabrics or it can be a standalone mechanism independently operating, or a blend of both.
The present disclosure also relates generally to software and networks, and more particularly, to the granularity of a tick of time governing the operations of higher level functionality.
Information Slingshot over a Network Tapestry
The conventional technologies and the limitations or drawbacks of those technologies serve to demonstrate why a different approach can be beneficial. This present disclosure's features, methods and utility described herein outline how to build and use this technology. It also offers perspective of how this present disclosure imbues certain improvements to overcome the limitations of the conventional technologies.
This present disclosure can be beneficial to meet various networking needs. The utilization of this present disclosure by the financial industry is hereby offered as an example to illustrate features and how they overcome the drawbacks of the conventional technologies. However, the scope and scale of the application and use of this present disclosure transcends this narrow focus and are applicable for many other use cases.
The internet is a network of networks. A network fabric may be defined as either a network under the control of one administrative body, or as a type of network protocol such is Internet Protocol (IP), or for a network within a regional area, etc. Fabrics can be weaved together through the peering of network fabrics joined together at the edges. Methods such as network address translation (NAT) and others can be used.
A Global Virtual Network (GVN) is a network type which exists over-the-top (OTT) of the internet and it weaves layers of fabrics into a network tapestry with Egress Ingress Points (EIP) at each fabric edge.
One type of network fabric can be a Wide Area Network (WAN) which consists of a dedicated line such as Multiprotocol Label Switching (MPLS) or Digital Data Network (DDN) or other between two Local Area Network (LANS) points. These dedicated links can also be peered to other networks either on a one to one basis defined as single honed or as one to many networks relationship via a multi-honed peering, with many possible paths or routes through the middle between two or more end points.
A network fabric may also define the scale and scope of a network fabric type from end-to-end. Ethernet defines a type of network but this can also be further classified by protocols such Internet Protocol (IP) over Ethernet, such as Internet Protocol version 4 (IPv4), or Internet Protocol version 6 (IPv6), and other network types. Built on top of IP are protocols such as Transmission Control Protocol (TCP) and User Datagram Protocol (UDP). TCP over IP (TCP/IP) is more verbose and has greater built-in error checking and handling for reliability of data sent versus UPD over IP (UDP/IP) which is geared more for streaming data such as audio or video where one lost packet will not have a dramatically adverse effect.
In addition to different protocols and IP versions built on top of Ethernet, Ethernet itself has Ethernet, Gigabit Ethernet (available at speeds of 1 or 10 or 40 or 100 Gigabit) and other versions of it are expected to be introduced. Ethernet is at the time of writing the popular protocol of choice to connect various devices together which are able to communicate with each other.
Infiniband (IB) is an alternate protocol to Ethernet with IB utilizing different physical NIC ports, plugs, and cabling as well as IB operating in a different manner. To connect many nodes into a high performance computing/computer (HPC) environment, IB is the preferred choice. IB allows for native Remote Direct Memory Access (RDMA) between nodes by-passing the network and operating system (O/S) stacks of the host devices for access where the RDMA storage (or other) devices are connected. There are also alternatives such as Fiber Channel and other network technologies which may offer the ability to sustain fast native RDMA transfers and to save files over distance to a parallel file system (PFS) or into RAM or save to another remote storage device.
While fiber channel is an alternative to IB, for the sake of disclosure, this disclosure focuses on IB as being the network type of choice for connecting HPC nodes, especially over great distances. IB is used as an example for demonstration sake but any alternative network type which can provide similar base connectivity can be used as a substitute to support an Information Slingshot.
When comparing network types, reliability is of paramount importance. Another main driver affecting the choice of type of network and network protocol also has to do with time for communications to travel across the distance of the physical path. Latency is a measure of time for data to travel in one direction or can also be expressed as a round trip time (RTT) over a specified distance between two points. Bandwidth is a measure of the data transfer rate of a network segment, communications route, network fabric, or other network path.
There are rules of physics which are hereby being referenced as a foundation. As time and distance are significant, this disclosure uses the following baseline:
In computing, the current main or standard measures of time for networking are milliseconds (ms) and for processing are microseconds (μs). The granularity of a tick of time can for example be measured either as a fraction (Every 1/20th or 1/10th or 1/100th) or as decimals (0.05, 0.1, 0.01) of a millisecond. The following table is a reference of some possible values and their corresponding equivalent conversion.
The following table compares the speed of light in a vacuum versus the speed of light inside of the glass core of optical fiber. This illustrates the physical limitation of fiber efficiency to establish a baseline for the theoretical best speed that one can hope to achieve for light to travel through fiber.
While the Refractive Index of fiber optic cables may vary slightly, an average is herein assumed as follows: Average of approx. 203 m to 204 m/μs vs. speed of light of 299.792 m/μs for an efficiency of 68.05%.
There exist specialized industries whose applications have an absolute requirement for the best quality of service (QoS) with lowest possible latency and zero packet loss over distance. These specialized industries need to exchange information for applications such as financial info, transaction execution to confirmation, scientific computing, and more. In many cases, a dedicated fiber optic line will be used for transport to offer more reliability and higher throughput than what could be expected for transport over the open internet.
One alternative to terrestrial fiber optic line communications is relay by satellite(s) transmission. Slingshot can operate over satellite links and it may be possible for communication speeds to approach the speed of light using such base transport medium. Satellite links are prone to bandwidth limitations, weather risks, cosmic interference, and other challenges. For this disclosure, the focus for Information Slingshot is for its demonstration to be deployed over terrestrial fiber optic line networks.
Ethernet and its combination of networking technologies is the most widely used network type deployed from the local area networks within offices, data centers and other clusters of devices to the global backbones linking networks operated by Internet Service Providers (ISPs) across the global internet. Internet Protocol—IPv4 while widely deployed is a victim of its own success as it faces a situation where the number of available IP Address is almost completely exhausted. Technologies such as NAT for devices in a LAN have helped mitigate this but have still not solved the problem and IPv4 addresses are consequently in very short supply. IPv6 has at its core what appears to be an ample supply of available IP addresses, however, IPv6 has not been deployed universally due to a number of factors, one of them being lock in of installed legacy equipment which only handles IPv4 and not both IPv4 and IPv6. Ethernet became the dominant network type and its wide use is prevalent both in the LAN and across the broader internet because it was a relatively easy standard to implement and to globally deploy. Network effects also come into play as more and more devices utilize a protocol or network type, it makes the decision easier for others to adopt the same, with a further incentive of high volumes leading to commoditization having the effect of driving down prices.
In the data center where concentrated computing, storage, processing and other operations are spread over various rack-mounted servers, a faster transport than Ethernet was required to back-channel connect these servers together for them to share data and communicate. Fiber channel and IB are two such technologies offering ultra-low latency and high capacity bandwidth. IB's lossless and parallel transfers offer strong advantages particularly alongside the deployment of Remote Direct Memory Access (RDMA) and also the opportunity to deploy and utilize a globally dispersed Parallel File System (PFS). The limitation of IB was that it was only deployable at a relatively short distance measured in meters. This was then extended to only a few kilometers. Until recently, IB “long-distance” links were limited to within a Metro Area Network (MAN) or between two nearby metro areas connecting data centers to each other via superfast IB over dedicated lines between them. Recent technologies allow IB to extend up to 20,000 kilometers between two IB gateway devices over a dark fiber line.
There are known drawbacks in the conventional techniques of Ethernet's IP protocol especially over long distance. IP's problems are inherently caused by how packets are sent. IP communications using its most popular protocol TCP/IP entail roundtrips and verbosity resulting in higher than necessary RTT.
IP transmits packetized data. Each packet includes headers, data payload, and possibly other information. The size of a packet allowed by an IP network fabric is defined as maximum transmission unit (MTU) and may be limited to hundreds of bytes to 4 KB or more but if the payload is larger than that set packet limit allows, to transport that data payload from point A to point B, the payload will need to be split into multiple-parts with two or more packets and subsequent roundtrips required per transmission and confirmation. Packet size also has an impact on whether or not a packet experiences fragmentation. It is common for a fragmented packet to get mangled if the fragmentation is forced by a middle-device which has a lower MTU than the other devices between network segments in a transmission path through a chain of devices. This drawback of IP can lead to dropped packets which have to be resent. Even if all packets arrive, fragmentation adds to congestion in the middle and can also flood receiving devices with more packets than expected. For example—if a sending device uses a common setting of 1500 bytes MTU, and breaks a payload into a stream of packets which are each 1500 bytes, the sending device will stream them out at the speed that the stack and NIC can handle up to the available bandwidth of the “last-mile” connection. And if a device in the middle on the internet has an MTU setting of 1492, then each packet flowing through this device will be fragmented or split into multiple packets. The result is that the receiving device is receiving double the number of packets that the sending device streamed. This can saturate and flood the receiving device leading to congestion and potential loss, or at least a slowdown. Furthermore, network devices are rated to send so many packets per second, regardless of how “full” a packet is. Fragmented packets are typically not full to capacity further lowering potential bandwidth potential with the knock-on effect of having a drag on network performance.
The reliability and verification of TCP/IP packets offer peace of mind but at great cost to speed and a forced rigidity because some disadvantages to IP over distance conventional techniques are that there are limits due to physical limitations and the nature of network protocol(s).
Specifically, packetized transport such as multipart packet transfer splitting up a larger payload risks being fragmented by a device in the middle which may not be able to or configured to handle a large payload.
Roundtrip transport of Ethernet/IP is known to have QoS degradation over distance with delays, known problems such as congestion which leads to delays, especially over distance. TCP/IP may catch a lost packet and will retransmit. While this is good from a reliability standpoint, it adds latency and congestion to networks.
It is important to also note how a data packet is handled by Ethernet because it has a store and forward model where a packet is received, examined and then forwarded only after the entire packet including full payload has been completely received and examined. Latency for this operation within a computer/router/network device to handle a packet of Ethernet traffic is approximately 100 microseconds (μs) per device. This drag of 100 μs delay per middle device for a packet to be received and passed on may seem miniscule but over a long distance pathway, this delay can be multiplied by passing through many devices. NAT also adds another small but still measurable friction point of resistance. Delay due to client-server methodology—prepare packet, send, transport, receive, combine packets, use—add bloat and unnecessary steps to a communication flow within the stack of an O/S of that device.
There is room for improvement in current IP protocols including need to lower latency, eliminate packet loss, and to solve other problems such as inefficiencies in packetization and bi-directional transfers, to address Bandwidth Delay Product (BDP) and related algorithms governing file transfers, and more.
While current trading methodologies have been engineered to be as efficient as possible with market information broadcast via UDP/IP multi-cast and trade orders & confirmations sent in the smallest possible packets via TCP/IP at the lowest latency, problems still exist. UPD/IP does not have the same error correction mechanism of TCP/IP and as such, neither the receiver nor the sender know if a packet is lost. And it is during times of heavy traffic where loss of packets is more likely, and this is precisely the time when it is imperative for traders to have the most complete information as possible.
There are also changes to computing which incumbent network protocols are going to have trouble addressing the need to move large amounts of data.
A new paradigm is required due to emerging technologies such as blockchain (and other distributed register technologies), industrial internet of things (HOT), remote medicine, and others where there exists a requirement for fast, reliable transmission of ever growing data payloads. Slingshot is a mechanism for information to be sent in one direction as a single file of virtually any size.
An Information Slingshot uses RDMA over IB or similar base network type to save a data file of unlimited size on to a remote storage device such as a PFS, and for a File Queue Manager (FQM) or similar mechanism on a device in that target remote region to pull the file from the PFS. Error correction is built into the storage device and subsequently governs the process writing of the file, just as a storage device connected locally interacts with an O/S within a host device. Slingshot is in a sense distributing what was an internal process to a globally distributed one. One-way transport is another feature of Slingshot realizing the efficiency of UDP/IP while achieving the reliability of TCP/IP addressing the need for confirmation of packets sent. Slingshot uses parallel stream file saves which have the effect of last byte arrival at or near the same time as the arrival of the first byte transmitted.
Slingshot can utilize any base network which can support RDMA and which has attributes similar to IB. Principals of high performance computing (HPC) such as RDMA, IB and other attributes provide foundation on top of which to build for a novel approach. Noting storage on a PFS device can be physical or virtual, or an HDD, SAS, SSD, RAM disk, or any other device which can provide fast storage medium and allows RDMA or equivalent type of direct remote access.
For this disclosure's demonstration, IB is utilized as an example network to run Slingshot on top of. IB bandwidth under load can achieve 90 to 95 percent of theoretical BW maximum. IB features cut through switching where IB devices receive the headers of a packet, use logic for forwarding decision and pipe the payload onwards, even before the complete payload is received. IB therefore has extremely low latency compared with Ethernet. It is also much less verbose than TCP/IP or UDP/IP. It runs on top of dark fiber connections. Compared with Ethernet over dark fiber it still is relatively faster and if native RDMA is utilized, latency can be measured as one-way for effective transmission rather than two-way as RTT is for Ethernet TCP/IP.
In the scope of financial data transmission, this disclosure is a framework to harness tapestry and utilize advanced connectivity approaches to address the need to speed up global transaction execution and other demanding communications needs.
An important distinction is that Slingshot does not make a traditional file transfer but rather executes a file saved to remote storage. As a transport for a clump of various data payload “files”, this slingshot method moves a larger block of data faster from one region to another region in a very efficient manner.
Significant gains across long distances. See enclosed Example Embodiments Figures and their explanations for more features.
Information slingshot can be integrated into a tapestry in a network type such as a GVN via egress ingress points (EIP) at the edges between fabrics such as EIP 2970 in
HPC cluster with all nodes able to act as a control node putting files onto PFS storage locations for other nodes to retrieve. While it is possible for a node to pull files from a PFS in a remote region, the added latency for pulling and marking and moving files does not make sense. It is best for a sender to save the file into a PFS as close to the target as possible. A feature of this disclosure is for nodes to be aware of each other, of PFS storage instances, as well as the proximity of the PFS storage instance to the target device the sender wishes to have read the saved file.
Any file presented to the information slingshot mechanism, from any source, will be saved remotely in target region as a complete file. The File Queue Manager will retrieve the file from the queue on the PFS or from RAM or any other substitute storage medium. It will process the file, checking that it is complete at process time. Complete files will be available for the next computing chain.
An information slingshot overcomes the drawbacks of conventional techniques as it provides an efficient, fast viable framework for the sharing of large volumes of information between devices over large distances.
Transaction latency is significantly reduced. Long distance transmission is at near wire speed, without the need to split file into multiple parts nor to packetize at send or to have to reassemble packets. This disclosure also eliminates packet bloat of headers and multipart delimiter boundaries. Cut through transfer is vastly more efficient than store and forward of IP packet handling. Advanced error detection and correction on file write further provides higher efficiency.
The nature of information exchange transforms from sequential packet transfer to parallel multi-thread file transfer over distance and the parallel file save to storage in the remote region is more efficient.
There is also an added benefit of higher availability of files accessible via RDMA containing structure and extended payload facilitating much larger data transfers. RDMA has superfast read and write to/from storage devices without having to go through the host as this bypasses the stack both at origin and at destination.
Bottlenecks and points of resistance due to the nature of IP protocols are also avoided by this disclosure. No more need for packetization by saving complete files to a remote PFS for another device to pull and use.
Decoupling of roundtrip RTT nature of packetized communication to the one-way sending file put reduces transmission time. Another subtle advantage is that body data can be variable in file size.
For conventional technologies, if data to transport is larger than specified by the payload size of a packet, it will be split up into multi-parts and thus will require multiple roundtrips for transmission. This adds significant time to completion of transmission and more processing to create and then validate and re-assemble a multiple part message. Information slingshot offers a known and consistent transmission time between points without need of the sending device to split up the file nor for the receiving device to have to reassemble it.
It overcomes the delays due to cumulative total of various transmission times and processing times for a chain of packets. Where packetized Ethernet IP packet transmission time is calculated as RTT time*total number of packets, Information Slingshot gives a more precise timing mechanism.
Information Slingshot can send a file whole without the need to break data into multi-part payloads sent by a stream of packets as is done by Ethernet IP. This also has the positive effect of keeping data movement at a very low level, without having to task the CPU, RAM, or other elements of the stack of a device.
Information Slingshot further avoids the need to receive a stream of packets. Ethernet consumes valuable time and resources when it has to receive a stream of packets and then validates each one, makes sure that they are in the correct order, and then reassembles the data file.
The core advantages of Information Slingshot (RDMA OTT IB) are parallel streams when transferring, low latency, lossless transfer, zero-copy networking, reduced load on hardware via cut through switching, RDMA bypassing the stack of the host devices at both ends, native utilization of parallel file system (PFS), one-way transfer without need for confirmation packets, no RTT, just one way transfer, no “packet size limit”.
The nature of a transfer via Slingshot can change how the composition of the transmission is approached. Data file header and body data can be more verbose, and can also include a file header and distinctive end-of-file (EOF) marker. Body Data including parameters, info array, algorithmic logic framework, and more can permit more feature rich data transmission. Data File Footer w/checksums, security and other integrity related items can also be included within the transmitted file.
This present disclosure may offer communications advantages. At long distance—near real time, with near wire-line speed may be achieved. At short distance can save μs (microseconds) from prior methods and the result is fast availability of data in as remote region. Information Slingshot changes how transactions are executed by speeding up transmission, simplifying addressing, allowing larger file payloads, and offering other advantages.
The enclosed figures and example embodiments describe more about the conventional techniques, the drawbacks of the conventional techniques and the features of this disclosure as well as how these features overcome the drawbacks of the conventional techniques.
The network slingshot and derivative technologies may provide benefits, such as, but not limited to: a) The network slingshot is a one-way fast delivery of data in one shot; b) There may be no need to packetize traffic, unlimited file size, arrival of first byte transmitted at near wire speed, with last bit arrival Δtime as fast as possible via dynamic parallel transmission streams scaled based on available BW headroom; c) High performance networking (HPN) based on Sling offers numerous advantages; and d) Other sling related tech can be discussed during meeting in a more formal disclosure.
Financial and other Network Technologies Based on Information Slingshot may provide benefits, such as, but not limited to:
The Granularity of a Tick
This disclosure describes the methodology to utilize very fine grained granularity for ticks of time to govern either fixed or variable data processing cycles. The time allotted for a cycle can be calculated dynamically based on a number of factors including the time required to process an item during a cycle. This information can be utilized to algorithmically calculate the limit of the maximum number of items which can be processed taking into account post-processed during a cycle of time. Information is available both to and from the application layer.
Software in the form of scripts or programs or routines or other types of blocks of code (SW) may be executed with a one-time trigger. In some instances, this is sufficient. However, there are types of SW which need to be run at regular and repeating time intervals.
Built into the UNIX and LINUX system is a scheduler called Command Run On (CRON) which will execute SW to run at a specific point in time. It can also set interval times. An alternative to a CRON job would be to have a Daemon running as a background process which has an internal clock with set time or other triggers at which point other SW jobs may be launched. In theory, these would operate like a CRON but could be set up in a more customizable manner.
A script may also be crafted in such a way that once executed, it keeps looping based on a condition that never becomes true to force an exit, or is based on a very large number of times which it is allowed to loop with a decreasing number every time it loops.
A combination of CRON launch and internal loop may be deployed in an instance when granularity of loop time is less than the minimum one minute required by a CRON tab set up. For example, if a certain event should occur every five seconds, a one minute CRON can be triggered and within that SW script, a loop which iterates twelve times (60 seconds/12 loops=5 second intervals).
A script may also run and then at the end of its execution, may call a new instance of itself to run again. This may have a sleep function at end to pause between its next runtime or can be set with a sleep of zero seconds implying a new call to itself to begin immediately at the end.
A related manager may be acting as a watchdog by doing something like monitoring a directory and when its file list is populated by a certain number of files, this may trigger the launch of another script for processing of the files. Therefore, the conditional launch of a SW script is based on a non-time based factor to be reached. A variation of this can be based on when a queue is full by quantity of items in the queue, by total memory size in RAM or in storage of the queue, or based on remaining free space, or other factors.
With regards to loops or execution of a SW script, there can be both internal and external conditions set such as a maximum execution time or a mechanism to try and make sure that a script or function or other routine is only run once. This checking to make sure that no duplicate instances are running is done to avoid two overlapping operations running at the same time which may interfere with one another.
A key HW operation is the internal system clock upon which many functions rely. Internal clocks are subject to drift over time, other running slower or faster than they should and therefore, updating them via regular running SW script.
Set execution time and intervals either hard coded into a script or as dynamic variables so that codependent scripts/routines/managers/modules or other software can utilize the same ticks parameters. These dynamic variables for ticks may either be set in a database, in an included flat file, may be calculated or otherwise generated.
One time triggers may have their uses but at times need to keep running, re-running code, scripts or routines and therefore the triggers need to be automated. In order to preserve system resources and also to have more control over these triggers, a framework tied to time to allow scheduling is required.
CRONs can be run at a specific time on a specific day or every X time unit or by various other calendar related timeframes. The smallest unit of time specified by a CRON is one minute. This is acceptable for certain system related tasks like log rotation, scheduled backups, or other routine procedures which can and should be automated. However, some tasks require granularity which is much finer than one minute as well as requiring more monitoring or control at the userland or user space software application layer.
In the instances of counting or conditional looping scripts or scripts which call themselves at termination of current cycle, there exist a few problems but one of the prime issues is that the next call is made at an indeterminate time. There are ways to keep track of run time and adding a sleep function to try to fire the next loop or call iteration at a regular interval. However, this is only applicable to a cycle which runs in less time that the interval time. In the instance that the cycle takes longer than the intended interval time, the next tick will be late and therefore may throw off the next cycles timing. Managing this process can become complicated very quickly. This therefore creates issues with regularity and dependability.
A daemon running a background process is very similar to a CRON but requires its own scheduler and manager to govern run times.
Conditional execution based on factors such as a list length, queue item quantity, free disk space or other factors may be efficient in some instances, are inefficient and inconsistent in other instances. Some items can get stuck in a queue for an inordinate amount of time. Also this does not take into account other factors so at times server resources are under-utilized and then at other times pounded by overutilization potentially causing other negative knock-on issues.
Loops within a script triggered per minute by a CRON are sequential and so if 12 loops in 60 seconds are prescribed then five seconds are allocated per loop. As noted above, if a loop only takes 3 seconds, a 2 second sleep can be issued at the end of the loop to keep it to 5 seconds regularity. The real problems arise when one or more loops is greater than 5 seconds which can push the script time to longer than 60 seconds. At the next minute, the CRON scheduler will call the next instance of the script which will begin its 12 loop cycles. Negative issues of concurrently running scripts can cause significant problems due to overlap handling of files, or in the case of transaction processing, if two scripts are executing a transaction, there can be unintended duplicates, or orphaned steps belonging to separate processes or other issues.
There are other negatives to the conventional techniques and these mainly deal with granularity in milliseconds (ms) or even finer granularity of time such as microseconds (μs).
In computing, the current main or standard measures of time for networking are milliseconds (ms) and for some processing are microseconds (μs). The granularity of a tick of time can for example be measured either as a fraction (Every 1/20th or 1/10th or 1/100th) or as decimals (0.05, 0.1, 0.01) of a millisecond.
In scheduling the running of software scripts, routines, etc. there exists a need for greater control over when it is triggered to run. For very fine granularity of time such as in the case when interval times are measured in ms or μs, an intelligent manager is required to regulate the granularity of ticks of time.
This manager requires an accurate and up-to-date system clock. Built into the manager are time updaters to keep the system clock as synchronized as possible to actual time. Utilization of best of breed agreement algorithms such as intersection algorithm or Marzullo's algorithm can be utilized.
This tick manager also relies on system resources monitors, on understanding system processing tasks, and other physical capabilities (and constraints), as well as other factors to base its estimation on how much time it should take for a cycle of a script or routine or other software call to complete its run.
By also looking at the items to process prior to processing, a more reasoned time estimate can be made. In the algorithmic calculation of how long a tick should take there are three main factors utilized in calculating the delta t or the amount of time from start of tick to the end.
The Granularity of a Tick
Δt=P+Q+R
t=Delta time from the start of tick to end of the tick
P=Time for batch processing of items for this tick in the cycle.
Q=Time for post batch processing computing.
R=Time for delay to ensure no overlap between batch items and/or to set offset between ticks.
This granularity of a tick is based on a number of factors. The start time of a tick is either based on completion of last cycle called by a tick or according to a fixed time interval scheduled.
There may be two types of cycles:
A fixed time cycle is based on an estimate for time per item to process (P) and then handle post processing (Q), a limited quantity of items is set. This limit ensures that each cycle does not run over time.
A variable time cycle allows for all items to be processed in that batch. The R delay time ensures that next cycle pulls a fresh batch and that all items in the last processed batch have been processed (P) and that post processing (Q) has been completed. For example, for files in a queue—a list can be processed at the P stage and then the files can be marked, moved, deleted or otherwise touched so that they do not get pulled in the next batch. The time delay (R) ensures that there is enough delay between last item processed and post-processed in last batch and the first item to be processed in next batch.
A component part of this tick manager is to maintain historical records of runtime data for each cycle. These logged records can be used to analyze periods of under usage and peak usage. It can also detect delays due to unprocessed items cut off because of limited time to deal with quantities per cycle. If this occurs too often, then it is an indicator that either more processing power is required, that the software is not efficient, or may indicate other issues such as database slowdowns or other needs for software refactoring, or other remedies to fix issues.
The tick can have granularity as fine as required by the application, can be set to run at a fixed interval, at variable intervals based on cycle length or other factors. The granularity of a tick allows for consistent network beacon flashes, slingshot related functionality, and other time reliant functionality.
For more features, see example embodiments as presented in the enclosed figures and their descriptions. Granularity of a Tick relies on timers, time circuits, time-keeping software, logic, time updaters and other time keeping and synchronization technologies, to be working properly in order that they provide accurate system time for internal clocks and other mechanisms. Granularity of a Tick is built on top of these time mechanisms.
Fixed time cycle tick granularity is set by dynamic variables and based on ever changing conditions. The number of items in a queue can also be set by dynamic variable or other factors. Constant testing, monitoring of resources, logging of current performance, comparison against past performances, determination of system load requirements, and other intelligent approaches offers significant power to determine how long each tick should take for processing (P), post processing (Q), and delay (R). And knowing how long each item should take to be processed during a cycle called by a tick permits for the setting of a limit of items to be processed during subsequent tick cycles.
The adjustable on-the-fly tick manager and metrics can indicate system health, identify need to upgrade, and also can be worked into a distributed processing configuration to pull resources on demand as required and as available. Algorithmic analysis on the fly governs how many items can be processed balancing available resources with how much those resources can physically process. Predictability and reliability are the result.
The intervals between ticks can also be adjusted accordingly based on both dynamic user requirements and the contextual reality of the situation.
This highly controlled tick manager at very fine time granularities such as ms or μs is advantageous for a number of reasons such as;
The granularity of a tick forms the basis of reliable time intervals over top of which other software applications can be built. Such as for a network information beacon set to pulse at X ms regularity, or a slinghop, or other type of time sensitive application.
The monitoring data can also be made available at the software layer so that it can adapt its operations to current conditions and have time related information to utilize at a very high level.
Option of fixed time cycles or time adjustable cycles can benefit different scenarios. By having a manager running and monitoring cycles, during peak use where cycles run longer than intended, during less demanding times, those subsequent cycles can be run shorter to bring the cycles back into a regular interval to catch up with the schedule where it was previously delayed.
This manager and its data are easy to integrate as a flow manager into existing architecture to realize benefits. There also exist other applications and benefits which can be realized by having more control over the granularity of a tick.
Systems and methods for file transfer and processing in a network environment are disclosed. In one embodiment, the system may comprise one or more processors. The one or more processors may be coupled to a first device. The one or more processors may be configured to retrieve a file from a file queue. The file may be stored in a local store of the first device. The file may be transferred from a second remote device via Remote Direct Memory Access. The one or more processors may further be configured to determine if the file is complete. The one or more processors may further be configured to remove the file from the file queue, if the file is determined to be complete.
In accordance with other aspects of this embodiment, the second remote device may be a backbone exchange server.
In accordance with other aspects of this embodiment, the system may further comprise a dedicated link between the first device and the second remote device.
In accordance with other aspects of this embodiment, the file may comprise a header, a body, and a footer.
In accordance with other aspects of this embodiment, the footer may comprise a marker that indicates an end of the file.
In accordance with other aspects of this embodiment, the file may comprise a key value pair.
In accordance with other aspects of this embodiment, the footer may comprise checksum to validate integrity of the body.
In accordance with other aspects of this embodiment, the Remote Direct Memory Access may be implemented based on at least one of Infiniband or fibre channel, or an alternative and equivalent network type.
In accordance with other aspects of this embodiment, the file may comprise executable code.
In accordance with other aspects of this embodiment, the one or more processors may further be configured to: if the file is determined to be complete, at least one of set a flag on the file, move the file to a folder or delete the file from the local store.
In another embodiment, systems and methods for governing granularity of time allocated to process a task are disclosed. In one embodiment, the system may comprise one or more processors. The one or more processors may be configured to determine a time duration needed to process a task. The time duration may be calculated based on a first duration to process the task, a second duration for post-processing related to the task, and a third duration as a delay.
In accordance with other aspects of this embodiment, the task may comprise a batch processing.
In accordance with other aspects of this embodiment, the time duration may be at least in one of microsecond or millisecond.
In accordance with other aspects of this embodiment, the time duration may be adjusted based on a historical record of runtime data for the time duration.
In accordance with other aspects of this embodiment, the second duration may comprise time needed for updating a status of the task.
In order to facilitate a fuller understanding of the present disclosure, reference is now made to the accompanying drawings, in which like elements are referenced with like numerals or references. These drawings should not be construed as limiting the present disclosure, but are intended to be illustrative only.
This example embodiment demonstrates a linking of a network in Region A 0100 to a network in Region B 0120, connected by a dedicated path over a fiber backbone connection FBB 0130. The FBB 0130 is connected at both ends by end point devices (EP) 0102 and EP 0122 that act as anchors at either end 0132 and 0138 of the bridge over the FBB 0130.
The shortest distances are within each region Local Region A 0150 and Local Region B 0152. The longest distance is over the Long Haul WAN segment 0154 and it is over this segment where the greatest improvements can be realized (e.g., method of transport—most significant distance over the longest segment—FBB 0130)
Per the timing of end to end transfer of data from a device in network in Region A 0100 to devices in a network in Region B 0120 consists of three blocks of network segments. Local Region A 0100 from devices in 0100 to end point 0102 via communications path 0106.
The long haul between regions is from EP 0102 to FBB 0130 via communications path 0132 and at the other end via 0138 to EP 0122. Finally, the local network segment 0126 from EP 0122 to devices in Region B 0120.
In a scenario such as financial trade instruction or information transmission, Long Haul WAN 0154 segment could represent a significant and major portion of time required for end-to-end 0100 to 0120. Therefore, it is essential to improve the connectivity of the Long Haul WAN 0154 segment.
At each end in Region A where EP 02100 is located and Region B where EP 02200 is located, it links EP to a LAN 02102/02202 and also illustrates an egress ingress point (EIP) 02106/02206 on each LAN to the open internet 02108/02208 at EIP 02106 and EIP 02206 via connecting paths such as 02116 and 02216, following standard gateway to point of presence (POP) to Internet methodology used in widespread practice today.
For example, each LAN 02102/02202 and Internet zone 02108/02208 illustrated has various devices connected to those networks. In the case of each LAN for example LAN 02102, devices are connected to it such as a server SRV_A 02122, a client C_A 02126, and a parallel file system node (PFS) PFS 02128 via paths 02152, 02156 and 02158 respectively. The internet for example Internet 02108 has two servers SRV_A 02132 and SRV_A 02136 connected to it via communications paths 02162 and 02166.
In an actual use case, there may be the same quantity of or more or less devices connected to each network zone LAN 02102, Internet 02108, LAN 02202, Internet 02208 than what is described herein. The quantity and types of these connected devices are only for illustrative purposes.
Other devices can be utilized such as routers, firewalls, managed switches, servers and other types of devices can govern how the traffic can flow within and between various networks.
As described in the Legend box 03000 at the bottom right, each financial market noted herein is from a networking perspective described as a Global Node 03008. Around the Global Node two rings denote the type of connectivity quality zone in for example purposes a radius from the center where the market or exchange is located. This is for simplification only as many factors determine the size and shape of these zones. However, the two zones can be differentiated from each other as the closest one being a High Performance Zone 03006 and the other being an Optimal Service Area 03002.
Markets are connected to each other via High Performance Network links to share pricing information and also to facilitate the processing of trade orders and to provide transaction confirmation of trades, and other information from remote locations to a market in another location.
The further away a market is from a querying client or server or other type of device, the longer it takes for information to flow, including pricing, volume or other real-time market information, transactional requests, fulfillment confirmations whether order filled or not, or other types of data exchange.
Zones indicated herein are:
There are many other markets around the world which are significant but those noted were indicated for illustrative purposes. There are also paths indicated between each market such as 03200 between NYC 03100 and LDN 03110. In reality, there are a multitude of paths representing undersea cables between the two points. This was meant to simplify the example illustrated herein.
The lower the latency between the points, the faster market information can be known and more rapid transactional related information can be executed and processed.
The communications path described herein is from Start 04110 to End 04210.
LAN_A 04100 is connected to WAN 04000 to LAN_B 04200 as a process flow illustrating packetized information flow. The information is prepared on a device at step 04112 and sent to a server or via a server 04120. This figure demonstrates the sending of a stream of packets in an IP-send/IP-receive process.
Paths 04322, 04332, 04352, 04522, 04532, 04552, 04622, 04632, 04652 are the send of one of a multiple number of multi-packet transmit packets. Paths 04326, 04336, 04356, 04526, 04536, 04556, 04626, 04636, 04656 are the packet confirmation denoting that it was received. A lost packet will not arrive at destination device and therefore this confirmation will not be sent back to the origin device. It is also possible for a packet to arrive at destination but for the confirmation to be lost in transit.
Take total packet size allowed minus header(s) size, minus footer(s) size, and that equals the largest size allowed for the payload of a single packet.
To calculate the total payload size:
payload size=allowed payload size−header size−footer size
The total number of packets needed to send a data file sized X is:
Therefore, a larger data size needs to be packetized into a stream of packets.
This figure illustrates an example where the packet may be relayed between devices rather than just having packets simply just passed on where at more than one stages, the data is broken into various packets and then reassembled into a data file once all packets are received at process as illustrated in IP Send Packet 04116 to IP receive packet 04118.
For example, SRV_A 04120 can query a market or other data source for information and will buffer that locally before replicating that information via IP Send Packet 04006 to IP Receive Packet 04008 to SRV_B 04220.
And further down the line, a client may Process Info 04212 by querying SRV_B 04220 via IP Send Packet 04216 to IP Receive Packet 04218.
The key point is that packetizing of data into packet allowed sized payloads and then re-combination at destination or source can occur at multiple times along a communications path.
This disclosure can be utilized within the topology as described in
SRV_BBX 05508 in region or zone ZD205500 can be connected to SRV_BBX 05608 in other region or zone ZD305600 via a dark fiber connection 05818 over dark fiber fabric 05808.
SRV_BBX 05508 uses this disclosure to directly write a file to parallel file storage PFS 05620 via remote direct memory access (RDMA) over 05818 bypassing the stack of SRV_BBX 05608 via path 05650.
SRV_IBX 05608 uses this disclosure to directly write a file to parallel file storage PFS 05520 via remote direct memory access (RDMA) over 05818 bypassing the stack of SRV_IBX 05508 via path 05550.
Path 05812 can be IPv4 or some kind of standardized internet protocol over which traffic flows from SRV_AP 05510 to and or from SRV_AP 05610 via path 05812 over-the-top of the GVN 05802 via a tunnel or other type of communication path.
This embodiment illustrates that various types of network fabrics can be combined into a greater network tapestry. These fabrics can be seamlessly woven together as described in U.S. Provisional Patent Application No. 62/174,394. The information slingshot can be either a standalone method or can be integrated as a network segment within a greater network path comprised of various network segments.
This example embodiment illustrates the topology of a global virtual network (GVN), its various devices, communications paths, and other embodiments. It shows how various geographic regions or zones or territory are linked together over various types of paths.
In this case, additional header and possibly more footer information is required to denote which piece of the multipart file is contained in the payload section of the packet. When a file is “saved” it is to a hierarchical file system (HFS) 06120, 06320 and the send is from reading saved local file from this HFS and packetizing a multi-part file transfer across as many packets are required via Send File 06200. As the file is being received, process Rec. File 06220 will read each packet, make sure they are in order and then will take each part of the multi-part transfer and stitch the file back together part-by-part.
The re-assembled file is saved HFS 06320 in the host device 06310.
It is a simplification of this network but designed to illustrate how a PFS can distribute storage devices across four various regions.
Each communication path in the tapestry can run various fabric protocols. A file storage on a PFS such as PFS 07810, PFS 07812, inside of a LAN 07120 or LAN 07812, or PFS 07800 or PFS 07802 in the cloud can be accessed via remote direct memory access (RDMA) via a protocol such as InfiniBand (IB) or other protocol which will allow native RDMA access to distributed nodes in a parallel file system (PFS).
The key point is that a cornerstone of high performance computing (HPC) can be achieved through a topology afforded by a GVN.
The steps in the flow from Start 08000 which could be a device generating a file, passing a file, pulling a file or otherwise handling a file such as File 08100. Path 08120 describes the saving a file directly to a PFS device via RDMA with a direct write to the storage node device from one region to another region.
The Retrieve Files from Queue 08300 on PFS 08200 step via path 08220 can be locally or also remotely with respect to the region where the PFS device is located.
End 08600 implies access to the file such as read, load, or otherwise use the file.
Files saved on a local HFS such as HFS 09040 by SRV_APP_109020 and on HFS 09240 by SRV_APP_209220. SRV_APP can represent an application server or other type of device and is labeled for illustrative purposes. WAN 09100 can be LAN/Internet/Dedicated line or other type of network connection.
MMDBR# refers to multiple master database replication. In such a deployment, database table and fields structure definition are equivalent and identical. Each host has a starting number on rows with auto-increment offset value equal to the number of hosts. For example, in the case of only two databases attached to master hosts, DB_1 would have row numbers 1,3,5, and on. And DB_2 would have ID row numbering of 0, 2, 4, 6, 8, 10, etc. Another anti-collision methodology may also be in place.
Database replication can be an automated or manual or other process where contents of each DB_109050 and DB_209250 are replicated to the other. To protect data from being overwritten, the row offsets accomplish this protection by ensuring no identical row IDs between the two sides.
However, this approach is not efficient for single or small of rows and can also lead to issues with synchronization.
Therefore, before using this method, a locking of the database during transfer is required at MMDBR110510 and MMDBR210512 to message between the Db's 10050 and 10250 to make sure that they are in synch. This adds unwanted drag to the system.
This example may not be an ideal utilization of Slingshot.
This figure illustrates a SRV_BBX_111080 directly writing a file via RDMA_111600 to PFS 11802, and SRV_IBX_212280 cross-posting a different file via RDMA_211610 to PFS 11800.
The data payloads are efficiently transferred. A record of the transfer is kept at each side respectively noting which files were received. Multiple master DB replication with DB_111050 using odd rows and DB_211250 using even rows. In this example there is a reliance on the database replication mechanism to keep track of multi-master replication events. Synchronization issues can therefore ensue. There is also potential for significant lag between time a remote write can be executed by a device and long delay to get confirmation which itself can be significantly detrimental. There is therefore a speed differential as well as reliability gap to overcome between the coupling of superfast RDMA writes and traditional database replication of information.
This example embodiment does not illustrate the ideal implementation of Slingshot. The save to remote PFS devices is efficient but the making the SRV_APP in the target region aware of the file is not fast enough nor is it reliable.
Because of the reliability of RDMA, once the file is saved into a PFS storage node there is confidence it is there. Accounting servers (SRV_ACC) such as SRV_ACC_112060 and SRV_ACC_212260 communicate with each other to catalog received transfers.
This again is not an ideal approach to slingshot and points to the need to completely decouple the Slingshot approach from current methods deployed over IP and to embrace the power of Slingshot.
Using current packetized IP technologies, even a UDP/IP multicast approach is not the most efficient.
This figure contains example embodiments describing the gathering of information by a device D013100 in one region Region A 13000 and then the onward flow of conveying that info to a device in Region B 13200.
The step Gather Market Info 13120 to Prepare File Contents 13130 to Slingshot File to Remote PFS 13180 are on D013100. Slingshot via 13882 to parallel file system (PFS) node PFS 13880 on storage device S213800. There are numerous advantages of using RDMA over IB 13882 or equivalent network type versus transferring files by TCP/IP or UPD/IP over Ethernet, mainly due to the fact that the file does not need to be packetized, but also due to how packets are handled, and the efficiency of RDMA over IB.
Saved file on PFS 13880 is retrieved on device D213300 by process Retrieve File by File from PFS 13380 via path 13382. A single or batch of multiple files can be retrieved simultaneously.
The follow on steps 13332 to 13325 to 13320 describe using of the file in the remote region from file origin. There is no further need for the devices to D013100 and D213300 to communicate with each other regarding the scheduling of or the transfer or the use of file itself.
The key point is that both devices are operating independently of each other. Only when D214300 wants to put in an instruction for a trade to be executed on D114100 is there a communications event but it too is not directly from D214300 to D114100 but rather for D214300 to save one or more data files on parallel file system (PFS) 14000. Once the file is written, D214300 can move on to other things.
D114100 has an automated process which runs according to an interval of time. See
This approach is efficient for trade orders such as a trade execution request. Device D1 can be as close to the market as possible. Device D2 can be the client terminal or a server which is in very close proximity to clients which access it. The key point is that trade payload can be sent at as close to wire-speed as possible.
This is advantageous and because it is able to send a large file size, this mechanism can be used to send/receive blockchains of ever growing height (data size), algo-bots (smart algorithms to be remotely executed), and other larger than packet size payloads.
The difference in this figure are example embodiments Confirmation from Market 15120 and Notification of Trade Exec. 15320.
D115100 and D215300 could all occur within the same device or via distributed operations between different devices in a zone represented by D115100 and D215300.
This figure starts the process from Start 16000 via path 16010 to Information 16100 step to evaluating and preparing the Information 16200 via 16110 either as a copying of a file or to structure the information into a file specifically designed for saving remotely via RDMA to a PFS 16300 via 16210. This saved file will then be fetched by process Retrieve File from Queue on PFS 16500 via path 16310. The file contents will be parsed by step Parse File Contents 16600 via path 16510. End step 16900 reached via path 16610 implies that the file can be used. The key point is that this process can be simplified into a few key steps.
This file could be stored in RAM, memory, saved to disk, or otherwise stored in another form of memory or storage on a device or otherwise.
Header 19100 can contain information about host origin, host destination, timestamp, and other info.
Security information can be stored in fields in both the header 19100 and the footer section 19300. This security information may hold references to keys to use for decryption of payload 19200, as well as other information. Payload (Body Data) may be encrypted in whole or in part or sent unencrypted.
Payload checksum(s) in the footer is used to validate the integrity of the body data.
Other features can take advantage of the flexibility offered.
A significant advantage is that there can be an algorithm in payload which gets sent as part of the payload.
EOF notation in the Footer 19300 will indicate that the file has arrived, is complete and ready to be validated/verified for accuracy and then ultimately use.
In the scope of the financial industry, a remotely executed algorithm in payload 19200 for example can contain an exit condition which is instructed to take a prescribed action if market conditions change with algorithmic instructions to evaluate the direction of market change and then modify instructions accordingly, to:
Connectivity of the FQM to the file 20888 on the PFS 20880 is via path FQM 20386 and coordination between the FQM and Retrieve file from queue on PFS 20380 is via path FQM 20388.
This figure also drills down each section, Header 20852, Payload 20856, and Footer 20858 of a file 20850 to demonstrate example fields. In the Header 20862, Payload 20866, Footer 20868.
These example fields are demonstrated as key=value pairs of information and are for illustrative purposes only. The data may be stored in another data structure, and fields can be different than described in this figure.
This figure is an algorithmic representation of what has been described in
The decision gate Check to see if each file is complete 22300 implies that it will look at each of the files pulled by File Queue Manager 22100 from PFS 22800. If a file is complete 22802, 22804, 22806, 22806, it will be processed via path Yes BFP 22510 and ultimately used. If it is not complete 22410, then the file will be Ignored 22440 and its process will end via BFP 22448 at END 22990.
End 22990 implies the end of the operations of the File Queue Manager 22100.
END 23940 implies that a file has been ignored because it is incomplete and 23400 the implication is that it will be read again during subsequent file batch pulls from the PFS 23800.
Interval A 24500 and Batch Pull A 24600 occur at the same time. USE 24602 happens at the end of this interval. Interval B 24510 and Batch Pull B 24610 occur at the same time. USE 24612 happens at the end of this interval. Interval C 24520 and Batch Pull C 24620 occur at the same time. USE 24622 happens at the end of this interval. There is a problem which exists by having one interval begin right after one has ended because there may not be enough time for a file to be marked as read, even though it has been read by a previous batch process.
For example, File 25802 is pulled by both Batch Pull A 24600 and Batch Pull B 24610. This also occurs again when File 25808 which was read by Batch Pull B 24610 is also pulled by Batch Pull C 24620.
This is a very dangerous flaw as for example, a trade request being executed twice could cause significant financial or other damage due to unintended consequences.
Marking as read may involve setting a flag on the file record on the storage device, or moving the file into a different folder, or deleting it from the storage device after it was moved, or other way to “mark” it so that subsequent batches know that the file has been read and “used” by a previous Batch Pull.
There is a further Delay B 25515 between Interval B 25510 and Interval C 25520 (and 25615 between Batch Pull B 25610 and Batch Pull C 25620).
The key point is that this delay allows for the current batch to evaluate and process all of the files it has pulled and where it has utilized complete files, to mark those files as read.
This delay added to the mechanism is fully dynamic and can be lengthened if more processing time is required for a batch or shortened to a batch processing is completed. The interval times can also be dynamically adjusted based on a number of factors. Granularity of a Tick describes the Δt=P+Q+R. t=Delta time from the start of tick to end of the tick, P=Time for batch processing of items for this tick in the cycle. Q=Time for post batch processing computing. R=Time for delay to ensure no overlap between batch items and/or to set offset between ticks.
Path CP 26150 from junction point 26120 to a use process such as Write file to remote PFS 26860. The write is made by Write file to remote PFS 26300 via path CP 26810.
The other path CP 26130 is for database record saving and logging. The Db record is saved locally in Db 26250 via process Save in DB 26200. Log records backups, or other related files are stored in a local HFS 26240 via Save in Log 26220 process. The remote write ends at END 26998. The administrative flow of database and log saves ends at END 26990.
Parallel and concurrent to processing, logging via Save in Db 27320 to Db_127350 and Save in Log 27600 writing notation records to logs on HFS 27650.
Only if the calculated checksum for the payload of a file at Is the File complete 27200 is equal to the checksum embedded in footer, can the file be marked as read and okay 27188.
Use this file 27220 implies that something will be done with the information or instructions or other data sent within the file structure. If market pricing information displayed or processed to algorithm for processing. If a trade instruction is ordered, then 27220 will feed this to the market for execution. The response can then be remotely written via same methodology. This may be utilized in the financial world or for other applications.
One key point is that each batch of files at a file processing cycle list pulled as a list at the start 27100. And each file is subsequently pulled 27110 and checked for completeness 27200. If a file is complete “yes” 27222, it is used 27220. The stream forks after this step. Mark file CP 27188 marks the file as read on the PFS 27180. Path CP 27302 leads to junction point 27300. The path is forked here to CP27322 leading to saving in a database 27350 and a log on HFS 27650. The other path CP 27312 is to use the file.
For the sake of simplification this demonstration, the file pulls 27110 are described as a looped pull. The step CP 27102 between Pull file list from PFS 27100 and Pull a file from PFS 27110 can also spawn parallel and concurrent file pulls each running at same time, following this process.
Other variations and possibilities exist and those skilled in this area will be able to make slight variations to this algorithm to achieve the save results.
The key point is that file batches can be pulled, each file individually checked if complete, if not complete 27242, the file is not used 27240 but this is still saved 27320 in Db 27350 and logged 27600 on HFS 27650. If a file is complete 27222, it is used 27220, marked 27188, processed 27512, and a record of its pull, check, use, and other information is saved in database 27350 and files are also logged on HFS 27650.
SRV_BBX servers can also connect to each other such as SRV_BBX 28300 and SRV_BBX 28732 via path 28738. This figure also includes a Sling Node Module 28388 on an SRV_BBX 28300. This sling node can either reside on the SRV_BBX 28300 or a sling node can be an independent device/server working in unison with the SRV_BBX 28300.
OTT2 represents a second degree over-the-top and these modules are built on top of a GVN 29008 such as OTT2 GVN end-to-end fabric 29010 or a OTT2 MPFWM multi-perimeter firewall mechanism 29012.
UTI1 is the Slingshot layer 29020 “under the internet” or in parallel with the internet 29000. UTI2 is where sling modules 29030 exist such as Sling Hop, Sling Route, Information Beacon and other related tech.
At each boundary between one network fabric and another are egress ingress points (EIPs). EIPs such as EIP 29052 imply a bridge between a LAN and a GVN, typically via an end point device (EPD) at the edge of the LAN 29002 building a secure tunnel with an access point server (SRV_AP) in the GVN 29008. An EIP 29058 between the GVN 29008 and the Internet 29000 implies an EIP on an SRV_AP. EIPs between other layers can represent other topologies and involve various other devices.
These various network fabrics when linked by EIPs together constitute a network tapestry.
New elements are also GVN Level 2—Logic 30050, GVN Level 4—Logic 30150, GVN Sub-Level 1—Logic 30650, and GVN Sub-Level 2—Logic 30750. These logic layers are where certain settings, operations, routines, and other methods are utilized to power EIPs between layers.
CRON jobs may be set either by manually adding items to the CRON list 31108 or this list may be managed by a software manager layer. Some problems exist relating to knowing if this process is running properly or not. And is there only one process running as intended or are there many spawned and unintentional parallel instances running. Environment items 31318 need to be to date and functioning within expected parameters. Foundation items 31680 need to be running optimally and within expected parameters.
System Time/Internal Clock 31600 has to be up-to-date 31360 and accurate to a very fine time granularity. This update regularity needs to be more frequent than the drift that the system is susceptible to. Hardware and associated drivers 31620 have some dependency on available resources.
When the file is called by a trigger at Start 32000. In this example, preparation code is run between example lines 004 and 009. In the case of a counter loop, the conditional loop will run while the Loop Condition 32200 evaluates of the current loop cycle counter $ii is less than the number of times that this loop should be run. At line 27 of the example script, the Counter or (external) condition 32150 will add one to $ii via $ii++ or some equivalent. This loop will continue to run as long as the condition remains true 32220.
If the counter $ii is equal to or greater than the number of times the loop should cycle, the loop will end and the processing flow will continue. Post loop code to execute is located between lines 033 and 037.
The script/routine/software ends at line 038 and its execution terminates at END 32990.
While the script is externally called once and loops as many times as prescribed either by a counter or by a condition, there can also be problems such as:
It begins at Start 33000. Code 33100 is listed from lines 001 through 020. To prepare for execution code is located between lines 002 and 004. The code to run during that one instance is between lines 005 and 014. Post execution code to run is between 015 and 018. At line 020 of the figure, the script calls itself and for example it opens a shell CLI> to re-launch script 33200. If there is a condition to end the running of the code (logic not illustrated), then at line 016, code info can be saved 33800. The script ends at END 33990.
If no condition or logic to end the looping of the script running exists, then it will loop run indefinitely.
34400 Can't launch
34402 Too many instances
34404 Uncaught exception/crash
34406 Execution time>tick allowed time
34408 Corrupted status info
34410 Can't re-launch
34412 Corrupted script file
34414 Unknown problems
The launcher which called the script may think that it is running but in reality it is not.
The above problems are included to highlight some of the potential issues. Not all or only some of these could present themselves at any time. More problems not outlined can exist.
The Granularity of a Tick
Δt=P+Q+R
Δt=Delta time from the start of tick to end of the tick 35100
P=Time for batch processing of items for this tick in the cycle.
Q=Time for post batch processing computing.
R=Time for delay to ensure no overlap between batch items and/or to set offset between ticks.
The period 35000 is the delta time or Δt for the duration of a tick.
Δt=x 36100 and Δt=y 36200 do not equal each other as described by x≠y 36010.
In this example, the delays R 36130 and R 36230 are equal to each other. The processing times P 36110 and P 36210 are not equal reflecting the different amount of tasks each is required to run. The post processing times Q 36120 and Q 36220 are also respectively different based on the amount of tasks each has to complete.
The duration of time for each cycle therefore changes based on how many items are to be processed.
The ticks are kept into equal Δt duration 37010 in at least two ways:
This fixed time is possible if there is a limit on the quantity of the number of items processed based on how much time is required for each P and Q.
It does not take into account reliability or loss or related congestion or other issues beyond latency which may have an effect on the IP path.
There are inefficiencies of using Slinghop of Δt at each bridgehead 38610 and 38620 of the Slinghop transparent hop segment 38600 versus the gain achieved 38370 via the utilization of Slinghop.
A minimum distance is needed to compensate for the requisite processing time of the Slinghop mechanism at 38615 and 38625.
Measure Latency of IP 38100 uses the following equation to measure
Total Internet Path Time Δt ms=TX 38510+TW 38500+TY 34590
Measure Latency of SL 38200 uses the following equation to measure:
Total Internet-Slinghop Hybrid Path Time Δt=TX 38630+TV 38610+TZ 38600+TV 38620+TY 38640
Latency comparison of IP vs SL uses Equation
Δt{TV+TZ+TV}<Δt{TW ms}
If True that Slinghop enhanced path (Δt{TV+TZ+TV}) has a lower latency SL is Lower 38370 than internet path then “SL is most efficient transport, use” 38380 is the most optimal path.
If False and internet (Δt{TW ms}) is faster than slinghop “IP is Lower” 38320 then “IP is most efficient transport, use” 38360 is the most optimal path.
The importance of exact timing is not just that files are received and processed without duplicates but also that there is a limited delay—that the file is processed with minimal delay.
The ability to patch pull and process in parallel allows for the Slingshot mechanism to massively scale.
This is a critically important point because the next batch file pull Read Queue+Process 40610 from PFS Incoming Files 40800 via path 40612 should only include unmarked files or files not filled by previous batches. Then at Post P 40620 the files pulled and used are marked via path 40622 so that they will not be inadvertently pulled by a subsequent batch file pull.
There are three cycles noted herein; 41100, 41200 and 41300 and each cycle contains different respective quantities of Quantity 58 in 41810, of Quantity 22 in 41830, and of Quantity 8 in 48860.
At junction 42020, two parallel processes execute to Gather tick parameters 42030 sourced from Tick settings 42050 and Check system monitors 42100 sourced from System monitor data 42150 accumulated from System monitor data 42160 source.
If all is okay for the tick to continue running at All okay? 42140, then the flow will continue along the path Yes 42142. If not okay, flow will go to error junction 42200 via path No 42144 to Error Handling 42400.
The two types of tick are either variable or fixed tick? 42170. A variable time tick 42176 begins at Script Start 42300. A fixed-time tick 42172 begins at Script Start 42500 using information from Use tick parameters 42175. There is also an additional logic gate Remaining tick time? 42550 to ensure that there are no time overruns. Common tick running completion processes, used by both fixed and variable processes, include prep script end 42600 and reporting 42680.
Per period calculated metrics with data from current period 43100, short term 43200 and long term 43300 historical data can be plotted on a standard deviation curve based on the various cyclical trends. Notation of lows, highs, averages and other analysis can indicate whether current performance is in line with expectations or is better, worse, or otherwise different than previous experience.
This data gathering and contextual analysis assists artificial intelligence (AI) algorithms in decision making. Current period 43100 can be compared with short term 43200 and long term 43300 performance data.
On the central, control server (SRV_CNTRL) 44200 systemic information is stored in the database (Db) 44295 of the Central Info Repository 44295. This information can be used by tick related managers, monitors and other system components. See
To update the time of its internal clock, a Device 44100 can query a SRV_TIME's 44600 NTP server listener 44650 via an API call or another type of request-response framework via paths REQ 44660 for the request and RESP 44680 via the response. The databases on the Device 44100 Db 44155 and on the SRV_TIME 44600 database Db 44655 can store information about time requests, time drift of the device, another other tick, time related, and other information.
Device(s) 44100 save information about time in their local databases Db 44155 and also report their time related information from their System and time manager 44120 to the SRV_CNTRL 44200 via a REQ-RESP posting to SRV_CNTRL 44200 via REQ 44600 with confirmation back via RESP 44620.
The NTP Time Module 44250 on SRV_CNTRL 44200 analyses systemic time related information.
The pulses are 45300 through 45318 and the duration of time between each pulse is noted as ts0045100 through ts1845118. These pulses should occur at defined intervals so that each flash can be relied on. These can carry simple to complex data. The key point is that the beacon is reliable.
The combination timing of multiple Beacon flashes is called the Pulse rate. And the frequency is the Δt between two flashes. Regarding tolerance, this has to do with the latency and the demand for information currency. How current information is when received is a function of the latency. The freshness of the information is a function of the latency plus the delay between pulses.
For example, if latency is 50 ms from 45000 to 45800 and pulse frequency and tick granularity are every 1 ms, then the info is always fresh. However, if the frequency is 3 ms between pulses and tick granularity is 1 ms, then there is a chance that the information could be stale by 2 ms.
The freshness of the information assumes sending of pulse at best possible wire speed and as most efficient sending and receiving as possible.
The assumption is that there is a one-way transport of information as a file sent at as close to wire speed as possible from SRV BB 46000 to be written via remote direct memory access (RDMA) to each PFS. The beacon ensures that information is conveyed globally and combined with multiple pulses ensures that information is always as fresh as possible by combining known latency between two points with consistent granularity of ticks for the sending of the Beacon information pulse/flash and the receiving and then the granularity of a tick for the processing and then subsequent use of received files.
There are more applications for this beyond that which is illustrated.
The first origin of information transmitted by a pulsing information beacon is at NYC 47100. Pulses are sent to LDN 47110 via path 74200. The second origin of information transmitted by a pulsing information beacon is at TOK 47120. Pulses are sent to LDN 47110 via path 47208.
In this example, pulses from both origins have been synchronized to overlap. This can be by adjusting start times or by aligning ticks and sending rates.
For this example, NYC represents New York, N.Y., USA, LDN represents London, England, UK, and TOK represents Tokyo, Japan.
The segments between the locations are indicated as well as the best wire speed in one direction for information.
The distance between New York and London is 3,456 miles and the best one-way wire speed 47000 is equal to 27.3 ms. The distance between Tokyo and London is 5,936 miles and the best one-way wire speed is 47020 is equal to 46.8 ms. Not illustrated on the upper part of the diagram but compared below is the distance between New York and Tokyo which is 6,737 miles and the best one-way wire speed between them is 47080 is 53.1 ms.
For example, in financial markets where pricing data, other information, and related news from different global regions is relevant and can influence trading in a local market, having the freshest and up to date information is mission critical. A super computer node (SCN) in London is able to receive information about New York and Tokyo and can react accordingly. If for example pricing in Tokyo starts fluctuating, London can react faster than New York by a Location Distance Advantage of 801 miles 47082 which is equal to 6.3 ms.
The granularity of a tick and its impact on timing and sequencing therefore is of mission critical importance to traders seeking an advantage.
There are many other examples of how multiple pulses from diverse locations can benefit from reliable granularity of ticks.
The O/S Monitor+Manager 48918 evaluates the processes and other system related items running which can have an influence on timing. The Time Updater 48970 interacts with the time server (SRV_TIME) 4831800 via NTP listener 4831850 to request and receive make the most current time.
Tick Software 48950 on All Devices—with Tick 48900 operates to ensure that the granularity of a tick is as efficient as possible. It interacts with monitors and managers. Built in software and plugins and other methods can be utilized to extend this out. The Resources Monitor 48980 and Process Mon.+Manager 48985 can provide information to the Tick Manager 48988.
The modules, managers, HW components, and other constituent parts demonstrated herein are for example only and may differ in real world use. Other devices may also utilize granularity of a tick for various kinds of functionality.
A key unifying thread between all of them is the importance of time. And therefore system time must be accurate and the granularity of a tick reliable so that events and especially sequential events can be processed and executed in a predictable fashion.
This application is a continuation of U.S. patent application Ser. No. 15/781,664, filed Jun. 5, 2018, which is a U.S. National Stage application under 35 U.S.C. § 371 of International Patent Application No. PCT/IB2016/001867, which claims the benefit of and priority to U.S. Provisional Application No. 62/266,060 filed on Dec. 11, 2015 and U.S. Provisional Application No. 62/296,257 filed on Feb. 17, 2016. The entire contents of each of these applications are incorporated herein by reference. International Patent Application Nos., PCT/US15/64242, PCTUS16/12178, PCT/US16/15276, PCT/US 16/15278, PCT/US16/26467, PCT/US16/26481, PCT/US16/26489, PCT/US16/37180, and PCT/IB16/00110 are hereby incorporated by reference in their entirety.
Number | Name | Date | Kind |
---|---|---|---|
4890281 | Balboni et al. | Dec 1989 | A |
5828847 | Gehr et al. | Oct 1998 | A |
5893089 | Kikinis | Apr 1999 | A |
5940838 | Schmuck et al. | Aug 1999 | A |
6209039 | Albright et al. | Mar 2001 | B1 |
6289201 | Weber et al. | Sep 2001 | B1 |
6374302 | Glasso et al. | Apr 2002 | B1 |
6463465 | Nieuwejaar | Oct 2002 | B1 |
6477166 | Sanzi et al. | Nov 2002 | B1 |
6593863 | Pitio | Jul 2003 | B2 |
6611587 | Brown et al. | Aug 2003 | B2 |
6671361 | Goldstein | Dec 2003 | B2 |
6678241 | Gai et al. | Jan 2004 | B1 |
6693876 | Zey | Jan 2004 | B1 |
6690223 | Wan | Feb 2004 | B1 |
6735207 | Prasad et al. | May 2004 | B1 |
6785295 | Graf et al. | Aug 2004 | B1 |
6879995 | Chinta et al. | Apr 2005 | B1 |
6973048 | Pitio | Dec 2005 | B2 |
6996117 | Lee et al. | Feb 2006 | B2 |
7006505 | Bleszynski et al. | Feb 2006 | B1 |
7039701 | Wesley | May 2006 | B2 |
7069318 | Burbeck et al. | Jun 2006 | B2 |
7145882 | Limaye et al. | Dec 2006 | B2 |
7145922 | Pitio | Dec 2006 | B2 |
7161899 | Limaye et al. | Jan 2007 | B2 |
7161965 | Pitio | Jan 2007 | B2 |
7173902 | Daniell et al. | Feb 2007 | B2 |
7177929 | Burbeck et al. | Feb 2007 | B2 |
7221687 | Shugard | May 2007 | B2 |
7224706 | Loeffler-Lejeune | May 2007 | B2 |
7254833 | Cornelius et al. | Aug 2007 | B1 |
7269130 | Pitio | Sep 2007 | B2 |
7310348 | Trinh et al. | Dec 2007 | B2 |
7349403 | Lee et al. | Mar 2008 | B2 |
7349411 | Pitio | Mar 2008 | B2 |
7349435 | Giacomini | Mar 2008 | B2 |
7389312 | Ohran | Jun 2008 | B2 |
7433964 | Raguram et al. | Oct 2008 | B2 |
7551623 | Feroz et al. | Jun 2009 | B1 |
7577691 | Novik et al. | Aug 2009 | B2 |
7587487 | Gunturu | Sep 2009 | B1 |
7633909 | Jones et al. | Dec 2009 | B1 |
7689722 | Timms et al. | Mar 2010 | B1 |
7742405 | Trinh et al. | Jun 2010 | B2 |
7742411 | Trinh et al. | Jun 2010 | B2 |
7801030 | Aggarwal et al. | Sep 2010 | B1 |
7822877 | Chong et al. | Oct 2010 | B2 |
7870418 | Sekaran et al. | Jan 2011 | B2 |
7886305 | Ahmed et al. | Feb 2011 | B2 |
7930339 | Tobita et al. | Apr 2011 | B2 |
7957311 | Trinh et al. | Jun 2011 | B2 |
8010751 | Yang et al. | Aug 2011 | B2 |
8064909 | Spinelli et al. | Nov 2011 | B2 |
8069258 | Howell | Nov 2011 | B1 |
8069435 | Lai | Nov 2011 | B1 |
8073777 | Barry et al. | Dec 2011 | B2 |
8107363 | Saluja | Jan 2012 | B1 |
8239915 | Satish et al. | Aug 2012 | B1 |
8259571 | Raphel et al. | Sep 2012 | B1 |
8266672 | Moore | Sep 2012 | B2 |
8401028 | Mihaly et al. | Mar 2013 | B2 |
8422397 | Ansari et al. | Apr 2013 | B2 |
8437641 | Lee et al. | May 2013 | B2 |
8458786 | Kailash et al. | Jun 2013 | B1 |
8544065 | Archer et al. | Sep 2013 | B2 |
8611335 | Wu et al. | Dec 2013 | B1 |
8611355 | Sella et al. | Dec 2013 | B1 |
8625411 | Srivivasan et al. | Jan 2014 | B2 |
8687791 | Cordell et al. | Apr 2014 | B1 |
8699683 | Jackson et al. | Apr 2014 | B1 |
8769057 | Breau et al. | Jul 2014 | B1 |
8798060 | Vautrin et al. | Aug 2014 | B1 |
8854965 | Richards | Sep 2014 | B1 |
8861344 | Trinh et al. | Oct 2014 | B2 |
8874680 | Das | Oct 2014 | B1 |
8966075 | Chickering et al. | Feb 2015 | B1 |
8976798 | Border et al. | Mar 2015 | B2 |
9015310 | Ochi | Apr 2015 | B2 |
9038151 | Chua et al. | May 2015 | B1 |
9110820 | Bent et al. | Aug 2015 | B1 |
9164702 | Nesbit | Oct 2015 | B1 |
9164795 | Vincent | Oct 2015 | B1 |
9167501 | Kempf et al. | Oct 2015 | B2 |
9172603 | Padmanabhan et al. | Oct 2015 | B2 |
9213594 | Strasser et al. | Dec 2015 | B2 |
9241004 | April | Jan 2016 | B1 |
9253028 | DeCusatis et al. | Feb 2016 | B2 |
9277452 | Aithal et al. | Mar 2016 | B1 |
9294304 | Sindhu | Mar 2016 | B2 |
9294497 | Ben-Or et al. | Mar 2016 | B1 |
9298719 | Noronha et al. | Mar 2016 | B2 |
9350644 | Desai et al. | May 2016 | B2 |
9350710 | Herle et al. | May 2016 | B2 |
9351193 | Raleigh et al. | May 2016 | B2 |
9369433 | Paul et al. | Jun 2016 | B1 |
9432258 | Van der Merwe et al. | Aug 2016 | B2 |
9432336 | Ostrowski | Aug 2016 | B2 |
9450817 | Bahadur et al. | Sep 2016 | B1 |
9455924 | Cicic et al. | Sep 2016 | B2 |
9461996 | Hayton et al. | Oct 2016 | B2 |
9525663 | Yuan et al. | Dec 2016 | B2 |
9525696 | Kapoor et al. | Dec 2016 | B2 |
9544137 | Brandwine | Jan 2017 | B1 |
9554061 | Proctor et al. | Jan 2017 | B1 |
9565117 | Dahod et al. | Feb 2017 | B2 |
9569587 | Ansari et al. | Feb 2017 | B2 |
9590820 | Shukla | Mar 2017 | B1 |
9590902 | Lin et al. | Mar 2017 | B2 |
9609003 | Chmielewski et al. | Mar 2017 | B1 |
9609482 | Want et al. | Mar 2017 | B1 |
9641612 | Yu | May 2017 | B2 |
9699001 | Addanki et al. | Jul 2017 | B2 |
9699135 | Dinha | Jul 2017 | B2 |
9729539 | Agrawal et al. | Aug 2017 | B1 |
9858559 | Raleigh et al. | Jan 2018 | B2 |
9888042 | Annamalaisami et al. | Feb 2018 | B2 |
9898317 | Nakil et al. | Feb 2018 | B2 |
9948649 | Zhao et al. | Apr 2018 | B1 |
10044678 | Van der Merwe et al. | Aug 2018 | B2 |
10061664 | Verkaik et al. | Aug 2018 | B2 |
10070369 | Lynn, Jr. et al. | Sep 2018 | B2 |
10078754 | Brandwine et al. | Sep 2018 | B1 |
10079839 | Bryan et al. | Sep 2018 | B1 |
10091304 | Hoffmann | Oct 2018 | B2 |
10237253 | Chen | Mar 2019 | B2 |
10275267 | De Kadt et al. | Apr 2019 | B1 |
10331472 | Wang | Jun 2019 | B2 |
10574482 | Ore et al. | Feb 2020 | B2 |
10673712 | Gosar et al. | Jun 2020 | B1 |
10756929 | Knutsen et al. | Aug 2020 | B2 |
10904201 | Ermagan et al. | Jan 2021 | B1 |
10922286 | Rubenstein | Feb 2021 | B2 |
11032187 | Hassan | Jun 2021 | B2 |
11092447 | Aiello et al. | Aug 2021 | B2 |
11108595 | Rubenstein | Dec 2021 | B2 |
20020007350 | Yen | Jan 2002 | A1 |
20020029267 | Sankuratripati et al. | Mar 2002 | A1 |
20020046253 | Uchida et al. | Apr 2002 | A1 |
20020049901 | Carvey | Apr 2002 | A1 |
20020087447 | McDonald et al. | Jul 2002 | A1 |
20020186654 | Tornar | Dec 2002 | A1 |
20030023351 | Fukui | Jan 2003 | A1 |
20030046529 | Loison et al. | Mar 2003 | A1 |
20030110214 | Sato | Jun 2003 | A1 |
20030072433 | Brown et al. | Aug 2003 | A1 |
20030147403 | Border et al. | Aug 2003 | A1 |
20030195973 | Savarda | Oct 2003 | A1 |
20030233551 | Kouznetsov et al. | Dec 2003 | A1 |
20040205339 | Medin | Oct 2004 | A1 |
20040268151 | Matsuda | Dec 2004 | A1 |
20050180319 | Hutnik et al. | Aug 2005 | A1 |
20050203892 | Wesley et al. | Sep 2005 | A1 |
20050208926 | Hamada | Sep 2005 | A1 |
20050235352 | Staats et al. | Oct 2005 | A1 |
20060020793 | Rogers et al. | Jan 2006 | A1 |
20060031407 | Dispensa et al. | Feb 2006 | A1 |
20060031483 | Lund et al. | Feb 2006 | A1 |
20060047944 | Kilian-Kehr | Mar 2006 | A1 |
20060075057 | Gildea et al. | Apr 2006 | A1 |
20060179150 | Farley et al. | Aug 2006 | A1 |
20060195896 | Fulp et al. | Aug 2006 | A1 |
20060225072 | Lari et al. | Oct 2006 | A1 |
20060288397 | Uchida | Dec 2006 | A1 |
20070083482 | Rath et al. | Apr 2007 | A1 |
20070112812 | Harvey et al. | May 2007 | A1 |
20070165672 | Keels et al. | Jul 2007 | A1 |
20070168486 | McCoy et al. | Jul 2007 | A1 |
20070168517 | Weller et al. | Jul 2007 | A1 |
20070226043 | Pietsch et al. | Sep 2007 | A1 |
20080010676 | Dosa Racz et al. | Jan 2008 | A1 |
20080043742 | Pong et al. | Feb 2008 | A1 |
20080091598 | Fauleau | Apr 2008 | A1 |
20080117927 | Donhauser et al. | May 2008 | A1 |
20080130891 | Sun et al. | Jun 2008 | A1 |
20080168377 | Stallings et al. | Jul 2008 | A1 |
20080191598 | Yang et al. | Aug 2008 | A1 |
20080240121 | Xiong et al. | Oct 2008 | A1 |
20080247386 | Wildfeuer | Oct 2008 | A1 |
20080256166 | Branson et al. | Oct 2008 | A1 |
20080260151 | Fluhrer et al. | Oct 2008 | A1 |
20080301794 | Lee | Dec 2008 | A1 |
20090003223 | McCallum et al. | Jan 2009 | A1 |
20090092043 | Lapuh et al. | Apr 2009 | A1 |
20090100165 | Wesley, Sr. et al. | Apr 2009 | A1 |
20090106569 | Roh et al. | Apr 2009 | A1 |
20090122990 | Gundavelli et al. | May 2009 | A1 |
20090129386 | Rune | May 2009 | A1 |
20090132621 | Jensen et al. | May 2009 | A1 |
20090141734 | Brown et al. | Jun 2009 | A1 |
20090144416 | Chatley et al. | Jun 2009 | A1 |
20090144443 | Vasseur et al. | Jun 2009 | A1 |
20090193428 | Dalberg et al. | Jul 2009 | A1 |
20090213754 | Melamed | Aug 2009 | A1 |
20090217109 | Sekaran et al. | Aug 2009 | A1 |
20090259798 | Wang et al. | Oct 2009 | A1 |
20100017603 | Jones | Jan 2010 | A1 |
20100131616 | Walter et al. | May 2010 | A1 |
20100250700 | O'Brien et al. | Sep 2010 | A1 |
20100316052 | Petersen | Dec 2010 | A1 |
20100325309 | Cicic et al. | Dec 2010 | A1 |
20110007652 | Bai | Jan 2011 | A1 |
20110170613 | Tanaka | Jul 2011 | A1 |
20110185006 | Raghav et al. | Jul 2011 | A1 |
20110231917 | Chaturvedi et al. | Sep 2011 | A1 |
20110247063 | Aabye et al. | Oct 2011 | A1 |
20110268435 | Mizutani et al. | Nov 2011 | A1 |
20110314473 | Yang et al. | Dec 2011 | A1 |
20120005264 | McWhirter et al. | Jan 2012 | A1 |
20120005307 | Das et al. | Jan 2012 | A1 |
20120082057 | Welin et al. | Apr 2012 | A1 |
20120105637 | Yousef et al. | May 2012 | A1 |
20120158882 | Oehme et al. | Jun 2012 | A1 |
20120179904 | Dunn et al. | Jul 2012 | A1 |
20120185559 | Wesley, Sr. et al. | Jul 2012 | A1 |
20120188867 | Fiorone et al. | Jul 2012 | A1 |
20120196646 | Crinon et al. | Aug 2012 | A1 |
20120210417 | Shieh | Aug 2012 | A1 |
20120210434 | Curtis et al. | Aug 2012 | A1 |
20120270580 | Anisimov et al. | Oct 2012 | A1 |
20120320916 | Sebastian | Dec 2012 | A1 |
20130032990 | Hattori | Feb 2013 | A1 |
20130070751 | Atwal et al. | Mar 2013 | A1 |
20130110787 | Garimella et al. | May 2013 | A1 |
20130173900 | Liu | Jul 2013 | A1 |
20130247167 | Paul et al. | Sep 2013 | A1 |
20130259465 | Blair | Oct 2013 | A1 |
20130283118 | Rayner | Oct 2013 | A1 |
20130286835 | Plamondon et al. | Oct 2013 | A1 |
20130287037 | Bush et al. | Oct 2013 | A1 |
20130308471 | Krzanowski et al. | Nov 2013 | A1 |
20130318233 | Biswas et al. | Nov 2013 | A1 |
20130322255 | Dillon | Dec 2013 | A1 |
20130343180 | Kini et al. | Dec 2013 | A1 |
20140020942 | Cho et al. | Jan 2014 | A1 |
20140026179 | Deverajan et al. | Jan 2014 | A1 |
20140071835 | Sun et al. | Mar 2014 | A1 |
20140086253 | Yong | Mar 2014 | A1 |
20140101036 | Phillips et al. | Apr 2014 | A1 |
20140108665 | Arora et al. | Apr 2014 | A1 |
20140149549 | Fu | May 2014 | A1 |
20140149552 | Carney et al. | May 2014 | A1 |
20140169214 | Nakajima | Jun 2014 | A1 |
20140181248 | Deutsch et al. | Jun 2014 | A1 |
20140199962 | Mohammed et al. | Jul 2014 | A1 |
20140210693 | Bhamidipati et al. | Jul 2014 | A1 |
20140215059 | Astiz Lezaun et al. | Jul 2014 | A1 |
20140226456 | Khan et al. | Aug 2014 | A1 |
20140229945 | Barkai et al. | Aug 2014 | A1 |
20140237464 | Waterman et al. | Aug 2014 | A1 |
20140250066 | Calkowski et al. | Sep 2014 | A1 |
20140269712 | Kidambi | Sep 2014 | A1 |
20140269728 | Jalan et al. | Sep 2014 | A1 |
20140278543 | Kasdon | Sep 2014 | A1 |
20140280911 | Wood et al. | Sep 2014 | A1 |
20140289826 | Croome | Sep 2014 | A1 |
20140304728 | Wendling | Oct 2014 | A1 |
20140310243 | McGee et al. | Oct 2014 | A1 |
20140324931 | Grube et al. | Oct 2014 | A1 |
20140331309 | Spiers et al. | Nov 2014 | A1 |
20140337459 | Kuang et al. | Nov 2014 | A1 |
20140341023 | Kim et al. | Nov 2014 | A1 |
20140351939 | Moore et al. | Nov 2014 | A1 |
20140359704 | Chen | Dec 2014 | A1 |
20140362712 | Agrawal et al. | Dec 2014 | A1 |
20140366119 | Floyd et al. | Dec 2014 | A1 |
20140369230 | Nallur | Dec 2014 | A1 |
20150006596 | Fukui et al. | Jan 2015 | A1 |
20150056960 | Egner et al. | Feb 2015 | A1 |
20150063117 | DiBurro et al. | Mar 2015 | A1 |
20150063360 | Thakkar et al. | Mar 2015 | A1 |
20150086018 | Harjula et al. | Mar 2015 | A1 |
20150089582 | Dilley et al. | Mar 2015 | A1 |
20150095384 | Antony et al. | Apr 2015 | A1 |
20150121532 | Barel | Apr 2015 | A1 |
20150128246 | Feghali et al. | May 2015 | A1 |
20150207812 | Back et al. | Jul 2015 | A1 |
20150222633 | Smith et al. | Aug 2015 | A1 |
20150222637 | Hung et al. | Aug 2015 | A1 |
20150248434 | Avati et al. | Sep 2015 | A1 |
20150271104 | Chikkamath et al. | Sep 2015 | A1 |
20150281176 | Banfield | Oct 2015 | A1 |
20150326588 | Vissamsetty et al. | Nov 2015 | A1 |
20150341223 | Shen et al. | Nov 2015 | A1 |
20150363230 | Kasahara et al. | Dec 2015 | A1 |
20160006695 | Prodoehl et al. | Jan 2016 | A1 |
20160028586 | Blair | Jan 2016 | A1 |
20160028770 | Raleigh et al. | Jan 2016 | A1 |
20160048938 | Jones et al. | Feb 2016 | A1 |
20160055323 | Stuntebeck et al. | Feb 2016 | A1 |
20160077745 | Patel et al. | Mar 2016 | A1 |
20160105530 | Shribman et al. | Apr 2016 | A1 |
20160117277 | Raindel et al. | Apr 2016 | A1 |
20160119279 | Maslak et al. | Apr 2016 | A1 |
20160127492 | Malwankar et al. | May 2016 | A1 |
20160134528 | Lin et al. | May 2016 | A1 |
20160134543 | Zhang et al. | May 2016 | A1 |
20160165463 | Zhang | Jun 2016 | A1 |
20160224460 | Bryant et al. | Aug 2016 | A1 |
20160226755 | Hammam et al. | Aug 2016 | A1 |
20160255556 | Michel et al. | Sep 2016 | A1 |
20160261575 | Maldaner | Sep 2016 | A1 |
20160285977 | Ng et al. | Sep 2016 | A1 |
20160308762 | Teng et al. | Oct 2016 | A1 |
20160330736 | Polehn et al. | Nov 2016 | A1 |
20160337223 | Mackay | Nov 2016 | A1 |
20160337484 | Tola | Nov 2016 | A1 |
20160352628 | Reddy et al. | Dec 2016 | A1 |
20160364158 | Narayanan et al. | Dec 2016 | A1 |
20160366233 | Le et al. | Dec 2016 | A1 |
20170063920 | Thomas et al. | Mar 2017 | A1 |
20170078922 | Raleigh et al. | Mar 2017 | A1 |
20170105142 | Hecht et al. | Apr 2017 | A1 |
20170201556 | Fox et al. | Jul 2017 | A1 |
20170230821 | Chong et al. | Aug 2017 | A1 |
20170344703 | Ansari et al. | Nov 2017 | A1 |
20180013583 | Rubenstein et al. | Jan 2018 | A1 |
20180024873 | Milliron et al. | Jan 2018 | A1 |
20180091417 | Ore et al. | Mar 2018 | A1 |
20180198756 | Dawes | Jul 2018 | A1 |
20200382341 | Ore et al. | Dec 2020 | A1 |
20210044453 | Knutsen et al. | Feb 2021 | A1 |
20210342725 | Marsden et al. | Nov 2021 | A1 |
20210345188 | Shaheen | Nov 2021 | A1 |
Number | Date | Country |
---|---|---|
2014381693 | Aug 2016 | AU |
1315088 | Sep 2001 | CN |
1392708 | Jan 2003 | CN |
1536824 | Oct 2004 | CN |
1754161 | Mar 2006 | CN |
1829177 | Sep 2006 | CN |
101079896 | Nov 2007 | CN |
101282448 | Oct 2008 | CN |
101478533 | Jul 2009 | CN |
101599888 | Dec 2009 | CN |
101765172 | Jun 2010 | CN |
101855865 | Oct 2010 | CN |
101969414 | Feb 2011 | CN |
102006646 | Apr 2011 | CN |
102209355 | Oct 2011 | CN |
102255794 | Nov 2011 | CN |
102340538 | Feb 2012 | CN |
102457539 | May 2012 | CN |
102687480 | Sep 2012 | CN |
102739434 | Oct 2012 | CN |
103118089 | May 2013 | CN |
103384992 | Nov 2013 | CN |
103828297 | May 2014 | CN |
104320472 | Jan 2015 | CN |
1498809 | Jan 2005 | EP |
1530761 | May 2005 | EP |
1635253 | Mar 2006 | EP |
2154834 | Feb 2010 | EP |
2357763 | Aug 2011 | EP |
6430499 | Nov 2018 | JP |
WO-0233551 | Apr 2002 | WO |
WO-2003025709 | Mar 2003 | WO |
WO-03041360 | May 2003 | WO |
WO-2003090018 | Oct 2003 | WO |
WO-2003088047 | Oct 2003 | WO |
WO-2003090017 | Oct 2003 | WO |
WO-2005065035 | Jul 2005 | WO |
WO-2006055838 | May 2006 | WO |
WO-2008058088 | May 2008 | WO |
WO-2008067323 | Jun 2008 | WO |
WO-2010072030 | Jul 2010 | WO |
WO-2012100087 | Jul 2012 | WO |
WO-2013068530 | May 2013 | WO |
WO-2013120069 | Aug 2013 | WO |
WO-2013135753 | Sep 2013 | WO |
WO-2015021343 | Feb 2015 | WO |
WO-2016073361 | May 2016 | WO |
WO-2016094291 | Jun 2016 | WO |
WO-2016110785 | Jul 2016 | WO |
WO-2016123293 | Aug 2016 | WO |
WO-2016162748 | Oct 2016 | WO |
WO-2016162749 | Oct 2016 | WO |
WO-2016164612 | Oct 2016 | WO |
WO-2016198961 | Dec 2016 | WO |
WO-2018049649 | Mar 2018 | WO |
Entry |
---|
“Open Radio Equipment Interface (ORI); ORI Interface Specification; Part 2: Control and Management (Release 4),” Group Specification, European Telecommunications Standards Institute (ETSI), 650, Route des Lucioles; F-06921 Sophia-Antipolis; France, vol. ORI, No. V4.1.1, Oct. 1, 2014 (185 pages). |
“Operations and Quality of Service Telegraph Services, Global Virtual Network Service,” ITU-T Standard, International Telecommunication Union, Geneva, Switzerland, No. F.16, Feb. 21, 1995, pp. 1-23 (23 pages). |
Baumgartner, A., et al., “Mobile Core Network Virtualization: A Model for combined virtual Core Network Function Placement and Topology Optimization,” Proceedings of the 2015 1st IEEE Conference on Network Softwarization (NetSoft), London, UK, 2015, pp. 1-9, doi: 10.1109/NETSOFT.2015.7116162 (9 pages). |
Chen, Y., et al., “Resilient Virtual Network Service Provision in Network Virtualization Environments,” 2010 IEEE 16th International Conference on Parallel and Distributed Systems, Shanghai, China, 2010, pp. 51-58, doi: 10.1109/ICPADS.2010.26, 2010 (8 pages). |
Chowdhury, N.M.M.K. et al., “Virtual Network Embedding with Coordinated Node and Link Mapping”, IEEE Communications Society Subject Matter Experts for Publication in the IEEE Infocom 2009, pp. 783-791. (Year: 2009) (9 pages). |
Definition of “backbone” in Microsoft Computer Dictionary, Mar. 2002, Fifth Edition, Microsoft Press (2 pages). |
Definition of “server” in Microsoft Computer Dictionary, Mar. 2002, Fifth Edition, Microsoft Press (3 pages). |
Examination Report, dated Aug. 2, 2018, for European Patent Application No. 16734942.2 (8 pages). |
Examination Report, dated Jul. 20, 2017, for Chinese Application No. 201680004969.3 (1 page). |
Examination Report, dated Mar. 3, 2020, for Chinese Application No. 201680020937.2 (9 pages). |
Examination Report, dated Mar. 5, 2020, for Chinese Patent Application No. 201580066318.2 (10 pages). |
Examination Report, dated Oct. 19, 2018, for European Patent Application No. 16727220.2 (11 pages). |
Extended European Search Report dated Sep. 7, 2018 received in related European Patent Application No. 16744078.3 (7 pages). |
Extended European Search Report, dated Aug. 2, 2018, for European Patent Application No. 15866542.2 (8 pages). |
Extended European Search Report, dated Sep. 7, 2018, for European Patent Application No. 16777297.9 (4 pages). |
Extended Search Report, dated Nov. 29, 2018, for European Patent Application No. 16806960.7 (10 pages). |
Figueiredo, R. J., et al., “Social VPNs: Integrating Overlay and Social Networks for Seamless P2P Networking,” 2008 IEEE 17th Workshop on Enabling Technologies: Infrastructure for Collaborative Enterprises, Rome, Italy, 2008, pp. 93-98, doi: 10.1109/WETICE.2008.43, 2008 (6 pages). |
First Office Action, dated Jun. 3, 2020, for Chinese Patent Application No. 201680066545.X (11 pages). |
Gong, L. et al., “Revenue-Driven Virtual Network Embedding Based on Global Resource Information”, Globecom 2013, Next Generation Networking Symposium, pp. 2294-2299. (Year: 2013) (6 pages). |
Haeri, S. et al., “Global Resource Capacity Algorithm with Path Splitting for Virtual Network Embedding”, 2016 IEEE International Symposium on Circuits and Systems (ISCAS), doi:10.1109/ISCAS.2016.7527328, pp. 666-669. (Year: 2016) (4 pages). |
International Search Report and Written Opinion, issued by U.S. Patent and Trademark Office as International Searching Authority, dated Apr. 8, 2016, for International Application No. PCT/US2016/015278 (9 pages). |
International Search Report and Written Opinion, issued by European Patent Office as International Searching Authority, dated Aug. 10, 2016, for International Application No. PCT/IB2016/000531 (20 pages). |
International Search Report and Written Opinion, issued by Chinese Patent Office as International Searching Authority, dated Aug. 23, 2017, for International Application No. PCT/IB2017/000580 (6 pages). |
International Search Report and Written Opinion, issued by Chinese Patent Office as International Searching Authority, dated Dec. 28, 2016, for International Application No. PCT/IB2016/001161 (7 pages). |
International Search Report and Written Opinion, issued by U.S. Patent and Trademark Office as International Searching Authority, dated Feb. 12, 2016, for International Application No. PCT/US2015/064242 (9 pages). |
International Search Report and Written Opinion, issued by Chinese Patent Office as International Searching Authority, dated Jul. 28, 2017, for International Application No. PCT/IB2017/000557 (6 pages). |
International Search Report and Written Opinion, dated Jul. 7, 2016, for International Application No. PCT/US2016/026489 (7 pages). |
International Search Report and Written Opinion, issued by Chinese Patent Office as International Searching Authority, dated Jun. 7, 2016, for International Application No. PCT/IB2016/000110 (8 pages). |
International Search Report and Written Opinion, issued by Chinese Patent Office as International Searching Authority, dated May 11, 2017, for International Application No. PCT/IB2016/001867 (13 pages). |
International Search Report and Written Opinion, issued by Chinese Patent Office as International Searching Authority, dated Sep. 1, 2017, for International Application No. PCT/IB2017/000613 (7 pages). |
International Search Report and Written Opinion, issued by European Patent Office as International Searching Authority, dated Sep. 23, 2016, for International Application No. PCT/IB2016/000528 (11 pages). |
Office Action, dated Mar. 12, 2020, for Chinese Patent Application No. 201680032657.3 (5 pages). |
Office Action, dated Mar. 13, 2020, received in related Chinese Patent Application No. 201680021239.4 (9 pages). |
Office Action, dated May 7, 2020, for Chinese Patent Application No. 201680020878.9 (7 pages). |
Robert Russell, “Introduction to RDMA Programming,” Apr. 17, 2014, XP055232895, last retrieved from the Internet Oct. 5, 2021: URL:web.archive.org/web/20140417205540/http://www.cs.unh.edu/˜rdr/rdma-intro-module.ppt (76 pages). |
Supplementary European Search Report, dated Dec. 11, 2019, for European Patent Application No. 17788882.3 (8 pages). |
Supplementary Partial European Search Report, dated May 20, 2019, for European Patent Application No. 16872483.9 (8 pages). |
Szeto, W. et al., “A Multi-Commodity Flow Based Approach to Virtual Network Resource Allocation,” GLOBECOM 2003, IEEE Global Telecommunications Conference (IEEE Cat. No. 03CH37489), San Francisco, CA, USA, 2003, pp. 3004-3008, vol. 6, doi: 10.1109/GLOCOM.2003.1258787, 2003 (5 pages). |
Number | Date | Country | |
---|---|---|---|
20220300466 A1 | Sep 2022 | US |
Number | Date | Country | |
---|---|---|---|
62296257 | Feb 2016 | US | |
62266060 | Dec 2015 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15781664 | US | |
Child | 17837625 | US |