The project leading to this application has received funding from the European Union's Horizon 2020 research and innovation program under grant agreement No. 957403.
The present invention relates to computer networks, and in particular, but not exclusively to, clock synchronization in computer networks.
The Precision Time Protocol (PTP) is a protocol used to synchronize clocks throughout a computer network. A master clock device sends a clock synchronization marker message to a slave clock device across the network and measures when the marker message egresses the master clock device on to a network link. The master clock device then sends a clock synchronization follow-up message, which includes a timestamp indicating the time when the clock synchronization marker message egressed the master clock device. The reason for using the clock synchronization follow-up message to carry the time when the clock synchronization marker message leaves the master clock device is that is very difficult to measure when a message is leaving a device and to add a timestamp of that time in the same message. The slave clock then synchronizes its local clock based on the received messages and optionally other message interactions. The above process is known as 2-step PTP.
In one example, a user-space application running on a host device generates the clock synchronization marker message and provides the message to a kernel space driver which provides the message to a network interface controller (NIC). The NIC sends the message to the slave clock device and measures the time when the message egresses the NIC to the network, sometime known as “hitting the wire”. The NIC returns the measured time to the kernel space driver. The user-space application polls the kernel space driver for the measured time. The user-space application generates the clock synchronization follow-up message and inserts the measured time in the clock synchronization follow-up message when received from the kernel space driver. The user-space application then provides the clock synchronization follow-up message to the kernel space, which provides the clock synchronization follow-up message to NIC, which sends the clock synchronization follow-up message to the slave clock device.
There is provided in accordance with an embodiment of the present disclosure, a system including a network interface controller including a device interface to connect to a processing device and receive a time synchronization marker message from an application running on the processing device, a network interface to send packets over a network, and packet processing circuitry to process the time synchronization marker message for sending via the network interface over the network to a slave clock device, generate a time synchronization follow-up message including a timestamp indicative of when the synchronization marker message egressed the network interface, and process the time synchronization follow-up message for sending via the network interface over the network to the slave clock device.
Further in accordance with an embodiment of the present disclosure the packet processing circuitry is configured to generate the time synchronization follow-up message autonomously of the application running on the processing device.
Still further in accordance with an embodiment of the present disclosure, the system includes the processing device, wherein the application is configured to generate the time synchronization marker message but not the time synchronization follow-up message.
Additionally in accordance with an embodiment of the present disclosure the packet processing circuitry is configured to duplicate the time synchronization marker message in a transmission pipeline of the packet processing circuitry as the time synchronization follow-up message for sending to the slave clock device.
Moreover in accordance with an embodiment of the present disclosure the packet processing circuitry is configured to perform a packet loopback of the time synchronization marker message in a transmission pipeline of the packet processing circuitry yielding a loop-backed packet directed towards a receive pipeline of the packet processing circuitry with the timestamp added to metadata of the loop-backed packet, and perform a hairpin operation to move the loop-backed packet from the receive pipeline to the transmission pipeline for sending to the slave clock device as the time synchronization follow-up message.
Further in accordance with an embodiment of the present disclosure the packet processing circuitry is configured to change an operation code field of the loop-backed packet from marker message to follow-up message.
Still further in accordance with an embodiment of the present disclosure the packet processing circuitry is configured to insert the timestamp into a header of the loop-backed packet responsively to the timestamp in the metadata of the loop-backed packet.
Additionally in accordance with an embodiment of the present disclosure the packet processing circuitry is configured to insert the timestamp into a preciseOriginTimestamp field of the loop-backed packet.
Moreover, in accordance with an embodiment of the present disclosure, the system includes a kernel space driver disposed between the application and the network interface controller.
Further in accordance with an embodiment of the present disclosure, the system includes the slave clock device including clock synchronization circuitry to receive the time synchronization marker message, and receive the time synchronization follow-up message while being unaware that the time synchronization follow-up message was generated by the network interface controller and not the application running on the processing device.
There is also provided in accordance with another embodiment of the present disclosure, a time synchronization method, including connecting to a processing device, receiving a time synchronization marker message from an application running on the processing device, sending packets over a network, processing by a network interface controller the time synchronization marker message for sending via a network interface over the network to a slave clock device, generating by the network interface controller a time synchronization follow-up message including a timestamp indicative of when the synchronization marker message egressed the network interface, and processing by the network interface controller the time synchronization follow-up message for sending via the network interface over the network to the slave clock device.
Still further in accordance with an embodiment of the present disclosure the generating includes generating the time synchronization follow-up message autonomously of the application running on the processing device.
Additionally in accordance with an embodiment of the present disclosure, the method includes the application generating the time synchronization marker message but not the time synchronization follow-up message.
Moreover, in accordance with an embodiment of the present disclosure, the method includes duplicating by the network interface controller the time synchronization marker message in a transmission pipeline as the time synchronization follow-up message for sending to the slave clock device.
Further in accordance with an embodiment of the present disclosure, the method includes performing by the network interface controller a packet loopback of the time synchronization marker message in a transmission pipeline yielding a loop-backed packet directed towards a receive pipeline with the timestamp added to metadata of the loop-backed packet, and performing a hairpin operation to move the loop-backed packet from the receive pipeline to the transmission pipeline for sending to the slave clock device as the time synchronization follow-up message.
Still further in accordance with an embodiment of the present disclosure, the method includes changing an operation code field of the loop-backed packet from marker message to follow-up message.
Additionally in accordance with an embodiment of the present disclosure, the method includes inserting the timestamp into a header of the loop-backed packet responsively to the timestamp in the metadata of the loop-backed packet.
Moreover, in accordance with an embodiment of the present disclosure the inserting includes inserting the timestamp into a preciseOriginTimestamp field of the loop-backed packet.
Further in accordance with an embodiment of the present disclosure, the method includes receiving by the slave clock device the time synchronization marker message, and receiving by the slave clock device the time synchronization follow-up message while being unaware that the time synchronization follow-up message was generated by the network interface controller and not the application running on the processing device.
The present invention will be understood from the following detailed description, taken in conjunction with the drawings in which:
As previously mentioned, a user-space application running on a host device generates a clock synchronization marker message and provides the message to a kernel space driver which provides the message to a network interface controller (NIC). The NIC sends the message to the slave clock device and measures the time when the message egresses the NIC over the network, sometime known as “hitting the wire”. The NIC returns the measured time to the kernel space driver. The user-space application polls the kernel space driver for the measured time. The user-space application generates the clock synchronization follow-up message and inserts the measured time in the clock synchronization follow-up message when received from the kernel space driver. The user-space application then provides the clock synchronization follow-up message to the kernel space, which provides the clock synchronization follow-up message to NIC, which sends the clock synchronization follow-up message to the slave clock device. Generating the clock synchronization messages is a very processing intensive and is compounded by the number of slave clock devices in the network.
Embodiments of the present invention provide a system to generate clock synchronization follow-up messages in a NIC thereby offloading this task from a user-space application running on a host to the NIC. The clock synchronization marker messages are still generated by the user-space application for sending by the NIC to the slave clock devices.
In some embodiments, the clock synchronization marker message is duplicated in the transmission pipeline of the NIC as a clock synchronization follow-up message. The time that the clock synchronization marker message egresses the NIC is added to a header of the clock synchronization follow-up message and the operational code of the clock synchronization follow-up message is generally updated to “follow-up message”.
In other embodiments, a packet loopback of the clock synchronization marker message is performed towards the end of the transmission pipeline whereby the original clock synchronization marker message continues to egress the NIC whereas a copy of the clock synchronization marker message (i.e., a loop-backed packet) is directed towards a receive pipeline of the NIC with a timestamp (indicative of the time that the original clock synchronization marker message egresses the NIC) added to metadata of the loop-backed packet. The original clock synchronization marker message is included in the payload of the loop-backed packet. A hairpin operation is performed on the loop-backed packet moving the loop-backed packet into the transmission pipeline. The timestamp is taken from the metadata of the loop-backed packet and inserted into the header of the loop-backed packet and the operation code field of the loop-backed packet is changed from marker message to “follow-up message”. The above insertions and changes may be performed using steering actions in the transmission pipeline that identify a loop-backed packet with a status of “marker message”. Optionally, other fields of the loop-backed packet are modified, added or removed (e.g., optional TLV (type-length-value) fields).
In some embodiments, the user-space application may be aware that follow-up messages are being sent by the NIC (e.g., 2-step PTP where the generation of the follow-up messages is offloaded to the NIC), whereas in other embodiments, the user-space application may be unaware that the NIC is generating follow-up messages.
Reference is now made to
The system 10 may include a host device 16 (or any suitable processing device such as a central processing unit (CPU) or graphics processing unit (GPU)), a network interface controller 12 and a slave clock device 14. The host device 16 and the network interface controller 12 may also be known as a master clock device.
The host device 16 may include a processor 18 which is configured to run an application 20 and a kernel space driver 22, which is disposed logically between the application 20 and the network interface controller 12. The application 20 is configured to generate a time synchronization marker message 24, which is provided to the kernel space driver 22 and then to the network interface controller 12 for sending to the slave clock device 14 over a network 32.
The network interface controller 12 includes a device interface 26, packet processing circuitry 28, and a network interface 30. The packet processing circuitry 28 may include a physical layer (PHY) module and a MAC layer module, by way of example.
Reference is now made to
The packet processing circuitry 28 is configured to generate a time synchronization follow-up message 34 including a timestamp indicative of when the synchronization marker message 24 egressed the network interface 30 (block 206).
The packet processing circuitry is configured to generate the time synchronization follow-up message 34 autonomously of the application 20 running on the host device 16. In some embodiments, the application 20 running on the host device 16 is configured to generate the time synchronization marker message 24, but not the time synchronization follow-up message 34.
The packet processing circuitry 28 is configured to process the time synchronization follow-up message 34 for sending via the network interface 30 over the network 32 to the slave clock device 14 (block 208).
In practice, some or all of the functions of the packet processing circuitry 28 may be combined in a single physical component or, alternatively, implemented using multiple physical components. These physical components may comprise hard-wired or programmable devices, or a combination of the two. In some embodiments, at least some of the functions of the packet processing circuitry 28 may be carried out by a programmable processor under the control of suitable software. This software may be downloaded to a device in electronic form, over a network, for example. Alternatively, or additionally, the software may be stored in tangible, non-transitory computer-readable storage media, such as optical, magnetic, or electronic memory.
The slave clock device 14 includes: a physical hardware clock 36 to maintain a clock time; and clock synchronization circuitry 38. The clock synchronization circuitry 38 is configured to receive the time synchronization marker message 24 and the time synchronization follow-up message 34 while being unaware that the time synchronization follow-up message 34 was generated by the network interface controller 12 and not by the application 20 running on the host device 16. The clock synchronization circuitry 38 is configured to adjust the clock time of the physical hardware clock 36 responsively to receiving the messages 24, 34 and optionally other message interactions, e.g., to measure transit time between the master clock device and the slave clock device 14.
Reference is now made to
In some embodiments, the packet processing circuitry 28 is configured to duplicate (arrow 48) the time synchronization marker message 24 in a transmission pipeline 40 of the packet processing circuitry 28 as the time synchronization follow-up message 34 for sending to the slave clock device 14 (block 210). The packet processing circuitry 28 is configured to change an operation code field 42 of the time synchronization follow-up message 34 from marker message to follow-up message (block 212). The packet processing circuitry 28 is configured to insert a timestamp 44 (indicative of the time that the time synchronization marker message 24 egressed the network interface 30) into a header 46 of the time synchronization follow-up message 34 (block 214). In some embodiments, the packet processing circuitry 28 is configured to insert the timestamp 44 into a preciseOriginTimestamp field of the time synchronization follow-up message 34.
Reference is now made to
The packet processing circuitry 28 is configured to perform a packet loopback (block 54) of the time synchronization marker message 24 in the transmission pipeline 40 yielding a loop-backed packet 50 directed towards a receive pipeline 52 of the packet processing circuitry 28 with the timestamp 44 added to metadata of the loop-backed packet 50 (block 216). The time synchronization marker message 24 is typically included in a payload of the loop-backed packet 50.
The packet processing circuitry 28 is configured to perform a hairpin operation (block 56) to move the loop-backed packet 50 from the receive pipeline 52 to the transmission pipeline 40 for sending to the slave clock device 14 as the time synchronization follow-up message 34 (block 218).
In some embodiments, the packet processing circuitry 28 is configured to change an operation code field 42 of the loop-backed packet 50 (i.e., time synchronization follow-up message 34) from marker message to follow-up message (typically in the transmission pipeline 40) (block 212). In some embodiments, the packet processing circuitry 28 is configured to insert the timestamp 44 into the header 46 of the loop-backed packet 50 (i.e., the time synchronization follow-up message 34), typically in the transmission pipeline 40, responsively to the timestamp 44 in the metadata of the loop-backed packet 50 (block 214). In some embodiments, the packet processing circuitry 28 is configured to insert the timestamp into a preciseOriginTimestamp field of the loop-backed packet 50.
Various features of the invention which are, for clarity, described in the contexts of separate embodiments may also be provided in combination in a single embodiment. Conversely, various features of the invention which are, for brevity, described in the context of a single embodiment may also be provided separately or in any suitable sub-combination.
The embodiments described above are cited by way of example, and the present invention is not limited by what has been particularly shown and described hereinabove. Rather the scope of the invention includes both combinations and sub-combinations of the various features described hereinabove, as well as variations and modifications thereof which would occur to persons skilled in the art upon reading the foregoing description and which are not disclosed in the prior art.
Number | Name | Date | Kind |
---|---|---|---|
6212197 | Christensen et al. | Apr 2001 | B1 |
6564267 | Lindsay | May 2003 | B1 |
6816492 | Turner et al. | Nov 2004 | B1 |
6954432 | Bychowsky et al. | Oct 2005 | B1 |
7224669 | Kagan et al. | May 2007 | B2 |
7245627 | Goldenberg et al. | Jul 2007 | B2 |
7676597 | Kagan et al. | Mar 2010 | B2 |
8370675 | Kagan | Feb 2013 | B2 |
8879552 | Zheng | Nov 2014 | B2 |
9390056 | Noureddine et al. | Jul 2016 | B1 |
9397960 | Arad et al. | Jul 2016 | B2 |
9450846 | Huang | Sep 2016 | B1 |
9628209 | Kingsley | Apr 2017 | B2 |
9667446 | Worrell | May 2017 | B2 |
10019333 | Regev | Jul 2018 | B2 |
10027601 | Narkis et al. | Jul 2018 | B2 |
10075253 | Olofsson | Sep 2018 | B2 |
10084559 | Devineni | Sep 2018 | B1 |
10187400 | Castro et al. | Jan 2019 | B1 |
10284383 | Bloch et al. | May 2019 | B2 |
10320508 | Shimizu et al. | Jun 2019 | B2 |
10727966 | Izenberg et al. | Jul 2020 | B1 |
10778361 | Almog et al. | Sep 2020 | B1 |
11271874 | Levi et al. | Mar 2022 | B2 |
11283454 | Ravid et al. | Mar 2022 | B2 |
11336383 | Mula et al. | May 2022 | B2 |
11388263 | Levi et al. | Jul 2022 | B2 |
11476928 | Levi et al. | Oct 2022 | B2 |
11543852 | Levi et al. | Jan 2023 | B2 |
11831403 | Lee | Nov 2023 | B2 |
20030002483 | Zwack | Jan 2003 | A1 |
20040174820 | Ricciulli | Sep 2004 | A1 |
20060007862 | Sayeedi et al. | Jan 2006 | A1 |
20060064579 | Aaron et al. | Mar 2006 | A1 |
20060095732 | Tran et al. | May 2006 | A1 |
20060153179 | Ho et al. | Jul 2006 | A1 |
20080298380 | Rittmeyer et al. | Dec 2008 | A1 |
20120201248 | Aoki | Aug 2012 | A1 |
20130027611 | Closset | Jan 2013 | A1 |
20130173950 | Banerjea et al. | Jul 2013 | A1 |
20140092918 | Jost | Apr 2014 | A1 |
20140379714 | Hankins | Dec 2014 | A1 |
20150046741 | Yen et al. | Feb 2015 | A1 |
20150055508 | Ashida et al. | Feb 2015 | A1 |
20150318015 | Bose et al. | Nov 2015 | A1 |
20170125026 | Lawrence | May 2017 | A1 |
20170171362 | Bolotov et al. | Jun 2017 | A1 |
20170188092 | Lawrence | Jun 2017 | A1 |
20170331926 | Raveh et al. | Nov 2017 | A1 |
20180124812 | Thubert et al. | May 2018 | A1 |
20180191629 | Biederman et al. | Jul 2018 | A1 |
20180309821 | Byers et al. | Oct 2018 | A1 |
20180351676 | Butterworth | Dec 2018 | A1 |
20190056972 | Zhou et al. | Feb 2019 | A1 |
20190087352 | Lee et al. | Mar 2019 | A1 |
20190087652 | Townsend et al. | Mar 2019 | A1 |
20190124524 | Gormley | Apr 2019 | A1 |
20190149365 | Chatterjee et al. | May 2019 | A1 |
20190319730 | Webb et al. | Oct 2019 | A1 |
20190379714 | Levi et al. | Dec 2019 | A1 |
20200026656 | Liao et al. | Jan 2020 | A1 |
20200252320 | Zemach et al. | Aug 2020 | A1 |
20200259759 | Shalev et al. | Aug 2020 | A1 |
20210081352 | Yang et al. | Mar 2021 | A1 |
20210152267 | Chan | May 2021 | A1 |
20210328900 | Sattinger et al. | Oct 2021 | A1 |
20220116132 | Kasichainula | Apr 2022 | A1 |
20220360423 | Levi et al. | Nov 2022 | A1 |
20220416925 | Levi et al. | Dec 2022 | A1 |
20230262281 | Kasichainula | Aug 2023 | A1 |
Number | Date | Country |
---|---|---|
2012216611 | Mar 2013 | AU |
103354989 | Oct 2013 | CN |
0607412 | Jan 1999 | EP |
2005529523 | Sep 2005 | JP |
2007042145 | Feb 2007 | JP |
9418627 | Aug 1994 | WO |
9857497 | Dec 1998 | WO |
2001037489 | May 2001 | WO |
2010117359 | Oct 2010 | WO |
Entry |
---|
IEEE Standard 1588-2008, “IEEE Standard for a Precision Clock Synchronization Protocol for Networked Measurement and Control Systems”, pp. 1-289, year 2008. |
Weibel et al., “Implementation and Performance of Time Stamping Techniques”, 2004 Conference on IEEE 1588, pp. 1-29, Sep. 28, 2004. |
InfiniBandTM Architecture Specification vol. 1, Release 1.2.1, pp. 1-1727, Nov. 2007. |
Lu et al., “A Fast CRC Update Implementation”, Computer Engineering Laboratory, Electrical Engineering Department, pp. 113-120, Oct. 8, 2003. |
Mellette et al., “Toward Optical Switching in the Data Center”, IEEE 19th International Conference on High Performance Switching and Routing (HPSR), pp. 1-6, Bucharest, Romania, Jun. 18-20, 2018. |
Bakopoulos et al., “NEPHELE: an end-to-end scalable and dynamically reconfigurable optical architecture for application-aware SDN cloud datacenters”, IEEE Communications Magazine, vol. 56, issue 2, pp. 1-26, Feb. 2018. |
O-RAN Alliance, “O-RAN Fronthaul Working Group; Control, User and Synchronization Plane Specification”, ORAN-WG4.CUS.0-v01.00, Technical Specification, pp. 1-189, year 2019. |
Vattikonda et al., “Practical TDMA for Datacenter Ethernet”, EuroSys conference, Bern, Switzerland, pp. 225-238, Apr. 10-13, 2012. |
Ericsson Ab et al., “Common Public Radio Interface: eCPRI Interface Specification”, V2.0, pp. 1-109, May 10, 2019. |
Xilinx Inc., “Radio over Ethernet Framer v2.1”, PB056 (v2.1), pp. 1-9, Oct. 30, 2019. |
Weibel, H., “High Precision Clock Synchronization according to IEEE 1588 Implementation and Performance Issues”, Zurich University of Applied Sciences, pp. 1-9, Jan. 17, 2005. |
Sanchez-Palencia, J., “[RFC,v3,net-next,00/18] Time based packet transmission”, pp. 1-14, Mar. 7, 2018. |
IEEE Std 802.1Qaz™, “IEEE Standard for Local and metropolitan area networks—Media Access Control (MAC) Bridges and Virtual Bridged Local Area Networks—Amendment 18: Enhanced Transmission Selection for Bandwidth Sharing Between Traffic Classes”, IEEE Computer Society, pp. 1-110, Sep. 30, 2011. |
Crupnicoff et al., “Deploying Quality of Service and Congestion Control in InfiniBand-based Data Center Networks” White Paper, Mellanox Technologies Inc, Rev 1.0, pp. 1-19, Oct. 15, 2005. |
Mathey et al., “Scalable Deadlock-Free Deterministic Minimal-Path Routing Engine for InfiniBand-Based Dragonfly Networks”, IEEE Transactions on Parallel and Distributed Systems, vol. 29, No. 1, pp. 183-197, Jan. 2018. |
Wikipedia, Precision Time Protocol, pp. 1-9, Apr. 20, 2020. |
SMPTE Standard, “Professional Media Over Managed IP Networks:Traffic Shaping and Delivery Timing for Video”, The Society of Motion Picture and Television Engineers, pp. 1-17, Nov. 22, 2017. |
Wikipedia, “Time-Sensitive Networking”, pp. 1-12, Mar. 5, 2020. |
IEEE Standard 1588-2002, “IEEE Standard for a Precision Clock Synchronization Protocol for Networked Measurement and Control Systems”, pp. 1-154, year 2002. |
Yau et al., “Development of Situation-Aware Application Software for Ubiquitous Computing Environments,” Proceedings of the 26th Annual International Computer Software and Applications Conference (COMSAC'02), IEEE Computer Society, pp. 1-6, year 2002. |
Davie et al., “An Expedited Forwarding PHB (Per-Hop-Behavior)”, RFC 3246, Network Working Group, pp. 1-32, Mar. 2002. |
Mills, “Network Time Protocol (NTP),” RFC 0958, pp. 2-15, Sep. 1985. |
Mills, “Network Time Protocol (Version 1): Specification and Implementation,” RFC 1059, pp. 2-59, Jul. 1988. |
Mills, “Internet Time Synchronization: The Network Time Protocol,” IEEE Transactions on Communication, vol. 39, No. 10, pp. 1482-1493, Oct. 1991. |
Mills, “Network Time Protocol (Version 3): Specification, Implementation and Analysis,” RFC 1305, pp. 1-96, Mar. 1992. |
Baboescu et al., “Scalable Packet Classification,” Proceedings of SIGCOMM'01, ACM, pp. 199-210, year 2001. |
Bader et al., “Survey and Comparison of Open Source Time Series Databases,” BTW 2017, Workshopband, Lecture Notes in Informatics (LNI), Gesellschaft fuer Informatik, Bonn, pp. 249-268, year 2017. |
Li et al., “Sundial: Fault-Tolerant Clock Synchronization for Datacenters,” Proceedings of the 14th USENIX Symposium on Operating Systems Design and Implementation, Open Access, USENIX Association, pp. 1171-1186, Nov. 2020. |
Levi et al., U.S. Appl. No. 17/667,600, filed Feb. 9, 2022. |
“Time-Appliance-Project / Open-Time-Server,” Spec Revision No. 1.0, GitHub, Inc., pp. 1-9, last updated Oct. 3, 2022, as downloaded from http://www.opentimeserver.com/. |
Manevich et al., U.S. Appl. No. 17/579,630, filed Jan. 20, 2022. |
Levi et al., U.S. Appl. No. 17/871,937, filed Jul. 24, 2022. |
Levi et al., U.S. Appl. No. 17/868,841, filed Jul. 20, 2022. |
Levi et al., U.S. Appl. No. 17/983,427, filed Nov. 9, 2022. |
Manevich et al., U.S. Appl. No. 17/885,604, filed Aug. 11, 2022. |
Manevich et al., U.S. Appl. No. 18/110,788, filed Feb. 16, 2023. |
Wertheimer et al., U.S. Appl. No. 17/886,606, filed Aug. 12, 2022. |
U.S. Appl. No. 17/667,600 Office Action dated Oct. 23, 2024. |
U.S. Appl. No. 17/667,600 Office Action dated Dec. 16, 2024. |
Number | Date | Country | |
---|---|---|---|
20240146431 A1 | May 2024 | US |