The movement of data within an electronic system generally involves moving data from a source location to a destination location. Direct memory access (DMA) is a technique whereby a DMA controller is programmed to a move a specified amount of data starting at a source address to a destination starting at a destination address. The movement of the data traverse the communication infrastructure of the electronic system. Some systems, such as systems-on-chip (Sots), are relatively highly segmented meaning that there are multiple bus interconnects and bridges through which data is moved. Traversing a bridge coupled between two bus segments can involve significant latency as the data coming into the bridge is temporarily buffered before it is then written out to the destination bus while also adhering to the timing requirements of the various buses and bridges comprising the communication infrastructure. Depending on the use of the data being moved, excessive latency can be problematic. For example, some devices have high speed serial ports that have internal buffers which may be too small to compensate for the round-trip latency. That is, data may be received into a buffer and the buffer may trigger a DMA request upon the buffer being filled to a threshold point. The DMA engine, however, may be coupled to the buffer over numerous bridges and interconnect segments, and thus a delay occurs as the DMA request is in transit from the buffer to the DMA engine. During the delay, the buffer may undesirably overflow.
In one example, an integrated circuit (IC) includes first and second memory devices and a bridge. The IC also includes a first interconnect segment coupled between the first memory device and the bridge. The IC further includes a second interconnect segment coupled between the first and second memory devices, and a third interconnect segment coupled between the bridge and the second memory device. The IC includes a first DMA circuit coupled to the first interconnect segment, and a second DMA circuit coupled to the second interconnect segment. A fourth interconnect segment is coupled between the first and second DMA circuits.
For a detailed description of various examples, reference will now be made to the accompanying drawings in which:
The source and target devices 106, 114 are coupled together by a series of interconnect segments and bridges. In the example of
The DMA circuit 104 can be programmed by commands from the CPU 102 to move data from the source device 106 to the target device 114 to thereby alleviate the CPU 102 itself having to read data from the source device 106 and write such data to the target device 114. The CPU 102, for example, may program a source address, a destination address, and a count (e.g., byte count, word count, etc.) into the DMA circuit 104. The source address may correspond to a starting address within the source device 106 where the data begins that is to be written to the target device 114, and the destination address corresponds to the address within the target device to which the data is to be written. The count indicates the amount of data to be written. Arrows 150 and 152 indicate the flow of data during a DMA write operation. Initially, a read engine 160 within the DMA circuit 104 reads data from the source device 106 as indicated by arrow 150. The data is read into a buffer 161. A write engine 162 (also within the DMA circuit 104) writes the data from the buffer 161 to the target device 114 as indicated by arrow 152. The read engine 160 and the write engine 162 are both part of the same DMA circuit 104. As such, the DMA architecture of
The system 100 of
The latency of the read transaction from the source device 106 into the DMA circuit 104 is fairly low as the data only traverses one interconnect segment 108 in this example. However, the latency of the write transaction from the DMA circuit 104 to the target device 114 may be fairly high as the data traverses three interconnect segments 108, 110, and 112 and two bridges 109 and 111.
Arrows 250, 251, and 252 illustrate the data flow of a DMA write operation for the example of
Further, the streaming interconnect 215 implements a “non-blocking” communication protocol. A non-blocking protocol means that, upon the master DMA circuit 210 attempting to initiate a data transaction (251) through the streaming interconnect 215 to the master DMA circuit 210, the transaction is guaranteed to complete without taking more than a threshold amount of time and without being blocked or otherwise interrupted by other transactions that may flow through the streaming interconnect. The latency experienced in a non-blocking fabric is primarily due to any variation of rate (the combination of clock speed and data path width) at various points in the fabric and arbitration pushback which occurs when more than one source tries to use a specific path in the fabric. These causes of latency are fully bounded in a non-blocking fabric. In a blocking fabric, the response latency of the target itself is not bounded. If the target of a data transfer does not have sufficient buffer capacity in which to place the data which is being transferred, then the target must push back on the fabric for as long as necessary until buffering frees up. In a non-blocking fabric, sufficient buffer capacity is guaranteed.
In one example, the system implements a dynamic mode in which the CPU 102 programs the master DMA circuit 210, and the master DMA circuit 210 transmits a transfer control parameter set across the non-blocking streaming interconnect 215 to the remote DMA circuit 220 to program the remote DMA circuit 220. A proxy is provided by the master DMA circuit 210 which maps accesses to memory mapped registers for the streaming interconnect 215 and converts the accesses to configuration read/write commands. Such configuration read/write commands are transmitted across the streaming interconnect 215 to the remote DMA circuit 220.
The examples of
The DMA read operation in the example of
Because the communication pathway between the master and remote DMA circuits 210, 220 comprises the streaming interconnect 215, and not bridge 109, interconnect segment 110, and bridge 111, fewer interconnect hops are required in performing a DMA read operation with the split-DMA architecture of
As shown in
The term “couple” is used throughout the specification. The term may cover connections, communications, or signal paths that enable a functional relationship consistent with the description of the present disclosure. For example, if device A generates a signal to control device B to perform an action, in a first example device A is coupled to device B, or in a second example device A is coupled to device B through intervening component C if intervening component C does not substantially alter the functional relationship between device A and device B such that device B is controlled by device A via the control signal generated by device A.
This application is a continuation of U.S. application Ser. No. 17/099,896, filed Nov. 17, 2020, which is a continuation of U.S. application Ser. No. 16/600,881, filed Oct. 14, 2019, now U.S. Pat. No. 10,838,896, which claims priority to U.S. Provisional Application No. 62/745,892, filed Oct. 15, 2018, each of which is incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
5664142 | Boldt et al. | Sep 1997 | A |
5748945 | Ng | May 1998 | A |
6081851 | Futral et al. | Jun 2000 | A |
6493803 | Pham et al. | Dec 2002 | B1 |
6675200 | Cheriton et al. | Jan 2004 | B1 |
6996655 | Lee | Feb 2006 | B1 |
7603488 | Gravenstein et al. | Oct 2009 | B1 |
10838896 | Karguth et al. | Nov 2020 | B2 |
10853308 | Subramanian | Dec 2020 | B1 |
20080109604 | Reilly | May 2008 | A1 |
20090037900 | von Khurja et al. | Feb 2009 | A1 |
20100082850 | Yokomizo | Apr 2010 | A1 |
20120191964 | Lee et al. | Jul 2012 | A1 |
20150127994 | Sankar et al. | May 2015 | A1 |
20160350246 | Steedman | Dec 2016 | A1 |
Number | Date | Country | |
---|---|---|---|
20230042413 A1 | Feb 2023 | US |
Number | Date | Country | |
---|---|---|---|
62745892 | Oct 2018 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 17099896 | Nov 2020 | US |
Child | 17971707 | US | |
Parent | 16600881 | Oct 2019 | US |
Child | 17099896 | US |