This disclosure relates to the field of data processing systems. More particularly, this disclosure relates to data processing systems including a processing pipeline and the control of the flow of instructions through that processing pipeline.
It is known to provide data processing systems including processing pipelines comprising multiple stages, such as stages corresponding to instruction fetch, instruction decode, instruction issue, instruction execution, and write back. Such pipelined processing contributes to increased instruction throughput through parallelism. The instruction fetch stage of the processing pipeline serves to fetch instructions to be executed from a memory system storing those instructions. The time taken to fetch an instruction can vary considerably, such as in dependence upon factors such as whether the instruction is cached, whether the instruction must be fetched from a slower main memory, whether fetching the instruction triggers a virtual to physical page mapping fault, etc. Thus, the time taken to fetch an instruction may vary by many orders of magnitude.
This variation introduces differences in the times it may take to execute a given sequence of instructions as in one case all of the instructions may be cached and available with no delay, whereas in another instance of the same instructions the instruction fetch stage may be required to perform multiple high latency fetch operations involving main memory and potentially page table walking. Fetches can be from different sources, or different types of main memories with varying delays, such as Flash, DDR memory, etc.
At least some embodiments of the disclosure provide apparatus for processing data comprising:
a processing pipeline having fetch circuitry to fetch instructions to be executed from a memory;
a one or more buffers to store instructions fetched from said memory by said fetch circuitry;
At least some embodiments of the disclosure provide apparatus for processing data comprising:
At least some embodiments of the disclosure provides a method of processing data comprising:
Further aspects, features and advantages of the present technique will be apparent from the following description of examples, which is to be read in conjunction with the accompanying drawings.
In operation, instructions are fetched by the fetch circuitry 12 from the instruction cache 8. If a miss occurs in the instruction cache 8, then the instruction is fetched from the main memory 6. Such a main memory instruction fetch will take more time than a fetch from the instruction cache 8. The fetch circuitry 12 stores fetched instructions into a first buffer 30. Decoder circuitry 14 reads instructions from the first buffer 30 and decodes them to form decoded instructions which are then stored within a second buffer 32. Issue circuitry 16 reads the decoded instructions from the second buffer 32 and issues these to the execute circuitry 18, 20, 22, 24, 26 when slots are available to execute those instructions within the relevant functional unit. Finally, write back circuitry 28 serves to write back the results of executing the instructions to update the state of the processor core with those results.
Buffer control circuitry 34 coupled to the decoder circuitry 14 receives a signal from the decoder circuitry 14 when this decodes a program instruction serving as a programmable trigger to stall a stallable portion of the processing pipeline, accumulate within one or more buffers a burst of instructions, and then, when a number (which may be predetermined) of program instructions have been accumulated within the one or more buffers, to restart the stallable portion of the processing pipeline.
In this example embodiment, the stallable portion of the processing pipeline comprises the issue circuitry 16. When this is stalled in response to the programmable trigger, instructions (and decoded instructions) are accumulated within buffers 30 and 32. It will be appreciated that in other embodiments different portions of the processing pipeline may serve as the stallable portion. For example, the decoder circuitry 14 could be stalled such that the instructions accumulate within the first buffer 30. In other embodiments it could be that the execute circuitry 18, or individual functional processing blocks such as the integer pipeline 20, the floating point pipeline 22, the single instruction multiple data pipeline 24 or the load store unit 26, could be stalled with the effect that instructions are buffered upstream of such stalled portions at least in respect of instructions intended to be executed by those stalled portions (e.g. it would be possible to stall and accumulate instructions intended to be executed by the integer pipeline 20 while instructions to be executed by other pipelines 22, 24, 26 within the execution circuitry were allowed to continue).
The stalling of the processing pipeline in order to accumulate bursts of fetched instructions can have a variety of uses including uses which seek to test the operation of the data processing system. For example, built in self test library code 36 stored within the main memory 6 may be executed to test the correct functioning of portions of the data processing system. In order to ensure that the test is properly performed and is not influenced by the timing of the fetching of program instructions, the built in self test library code 36 may include instructions which serve as a programmable trigger for the stalling, accumulation and restarting operations as described above such that particular sequences/bursts of instructions within that built in self test code may be executed with a deterministic timing which is independent of any instruction fetch delay variation. Another example use of the present technique is executing a particular sequence/burst of instructions with a deterministic timing (independent of any instruction fetch delay) to sample multiple data points from different peripherals/devices in close timing proximity.
As an example, the load store unit 26 may include stall forwarding circuitry 38 which serves to buffer given write data of a pending data write to a given memory address before that given write data is written out to the memory (such as the data cache 10 or the main memory 6) and, while such a write is pending and the given write data is stored within the store forwarding circuitry, serves to service a subsequent data read for that given address using the copy of the data stored within the store forwarding circuitry 38 rather than and incurring the delay of waiting for that write data to be written out to the memory system and then reading that data back from the memory system. However, in order to properly test such store forwarding circuitry, it is important that the write operation is rapidly followed by the read operation such that the store out to memory will not have taken place and the store forwarding circuitry will be active in servicing the read. This may be achieved by including the write instruction and the subsequent read instruction within a burst of instructions which is gathered together within the processing pipeline downstream of the fetch circuitry 12 and then released as a burst of instructions to be processed such that the write instruction will be followed by the read instruction with a deterministic timing between the write instruction and the read instruction which is independent of variable fetch delays.
The buffer control circuitry 34 includes escape circuitry 40 which serves to detect an escape event and when such an escape event is detected stops any accumulating of instructions into the one or more buffers 30, 32 and restarts the stallable portion (e.g. the issue circuitry 16 in this example embodiment). Such escape circuitry 40 may, for example, be useful to ensure that pathological conditions such as, for example, deadlocks do not arise whereby fetched instructions never arrive due to other processes preventing those instructions being fetched and so the processing pipeline is permanently stalled. Another example use of the escape circuitry 40 is where there is no predetermined length for the burst of instructions accumulated while the pipeline is stalled, rather instructions are accumulated until, for example: a hardware trigger event arises (such as a buffer full signal—buffer sizes may different between implementations); or a certain programmed monitor event arises within the design (or buffer control circuitry); or an interrupt or abort is triggered; or combinations of the preceding. The escape circuitry can thus provide either or both of a backdoor to stop a stall or a main design feature to stop a stall in an intended manner.
The escape events which are detected by the escape circuitry 40 may take a variety of different forms. For example, the escape event may comprise that a time taken for the number of instructions to be fetched from the memory exceeds a threshold time, e.g. a time out timer. Alternatively, the escape event may serve to detect one or more monitor events having occurred (e.g. a predetermined number of attempted memory fetches, a predetermined number of memory aborts, etc) which are indicative of a problem such that the processing pipeline should be restarted/unstalled.
The one or more buffers, 30, 32 into which the number of instructions is accumulated subsequent to the programmable trigger may be buffers which are already provided within the processing pipeline for the normal operation of that processing pipeline. In this case, the additional overhead associated with providing the facility to stall, accumulate and restart a stallable portion of the processing pipeline for instruction fetch determinism may be relatively low. In other example embodiments it is possible that dedicated buffers may be provided to serve to store the accumulated instructions before the stalled portion of the pipeline is restarted.
When the stalled portion is restarted, the burst of instructions which have been accumulated within the one or more buffers may be executed atomically e.g. without intervening interrupts or other delays. Such atomic execution is assisted by the accumulation of the instructions within the one or more buffers as this removes the possibility of memory aborts arising partway through executing the atomic sequence due to memory faults within instruction fetches.
The programmable trigger for controlling the buffer control circuitry 34 to control the processing pipeline to perform the stall, accumulate and restart steps may take a variety of different forms as will be described further below. Some of the forms of control and the provision of the programmable trigger may utilise a configuration register 42 into which parameters relating to the programmable trigger may be stored. For example, a programmable flag may be set within the configuration register 42 to prime the buffer control circuitry 34 to perform the actions of the stalling, accumulating and restarting in synchronism with a synchronising instruction which has yet to be received. The configuration register 42 may also in some embodiments store a parameter specifying the length of the burst of instructions to be accumulated.
It will be appreciated that the examples discussed above are only some forms of the circuitry and programmable triggers which may be used in embodying the present techniques. Other forms of circuitry and programmable trigger are also possible.
In the present application, the words “configured to . . . ” are used to mean that an element of an apparatus has a configuration able to carry out the defined operation. In this context, a “configuration” means an arrangement or manner of interconnection of hardware or software. For example, the apparatus may have dedicated hardware which provides the defined operation, or a processor or other processing device may be programmed to perform the function. “Configured to” does not imply that the apparatus element needs to be changed in any way in order to provide the defined operation.
Although illustrative embodiments of the invention have been described in detail herein with reference to the accompanying drawings, it is to be understood that the invention is not limited to those precise embodiments, and that various changes and modifications can be effected therein by one skilled in the art without departing from the scope and spirit of the invention as defined by the appended claims.
Number | Date | Country | Kind |
---|---|---|---|
1709064 | Jun 2017 | GB | national |
This application is a Continuation Application of Application Ser. No. 15/974,769, filed May 9, 2018 and claims priority to GB Patent Application No. 1709064.8 filed Jun. 7, 2017, the entire content of each of which are hereby incorporated by reference.
Number | Name | Date | Kind |
---|---|---|---|
5579498 | Ooi | Nov 1996 | A |
5765037 | Morrison | Jun 1998 | A |
6006032 | Blandy | Dec 1999 | A |
6212603 | McInerney et al. | Apr 2001 | B1 |
6230263 | Ryan | May 2001 | B1 |
6353805 | Zahir et al. | Mar 2002 | B1 |
6370638 | Yamasaki | Apr 2002 | B1 |
6629167 | Undy | Sep 2003 | B1 |
6671795 | Marr | Dec 2003 | B1 |
6691240 | Stotzer | Feb 2004 | B1 |
6859871 | Batten | Feb 2005 | B1 |
6988187 | Magoshi | Jan 2006 | B2 |
7139898 | Nemirovsky | Nov 2006 | B1 |
7469334 | Chaudhry | Dec 2008 | B1 |
7490221 | Evans | Feb 2009 | B2 |
9251117 | Hanai | Feb 2016 | B2 |
9417916 | Stark | Aug 2016 | B1 |
20030084273 | Moore | May 2003 | A1 |
20030149860 | Becker | Aug 2003 | A1 |
20040039896 | Pechanek | Feb 2004 | A1 |
20040059904 | Sonnekalb | Mar 2004 | A1 |
20040068640 | Jacobson et al. | Apr 2004 | A1 |
20040098577 | Sane | May 2004 | A1 |
20050154860 | Arimilli | Jul 2005 | A1 |
20060010309 | Chaudhry | Jan 2006 | A1 |
20070074012 | Graham | Mar 2007 | A1 |
20070083742 | Abernathy | Apr 2007 | A1 |
20070226465 | Chaudhry | Sep 2007 | A1 |
20080126770 | Morrow | May 2008 | A1 |
20090198980 | Alexander et al. | Aug 2009 | A1 |
20090210679 | Tsai et al. | Aug 2009 | A1 |
20110264862 | Karlsson | Oct 2011 | A1 |
20120023314 | Crum | Jan 2012 | A1 |
20120233441 | Barreh | Sep 2012 | A1 |
20140032828 | Khailany | Jan 2014 | A1 |
20140075156 | Blasco-Allue | Mar 2014 | A1 |
20140181482 | Smaus | Jun 2014 | A1 |
20150301830 | Wiencke | Oct 2015 | A1 |
20150378732 | Gschwind | Dec 2015 | A1 |
20160378496 | Gray | Dec 2016 | A1 |
20170083337 | Burger | Mar 2017 | A1 |
20170109251 | Das | Apr 2017 | A1 |
20170139716 | Caulfield et al. | May 2017 | A1 |
20180095753 | Bai | Apr 2018 | A1 |
20180136955 | Takeo | May 2018 | A1 |
20180357065 | Bhartia et al. | Dec 2018 | A1 |
Number | Date | Country |
---|---|---|
0 987 624 | Mar 2000 | EP |
H6-131178 | May 1994 | JP |
H11-15661 | Jan 1999 | JP |
2002-149404 | May 2002 | JP |
2009-110209 | May 2009 | JP |
2011-508309 | Mar 2011 | JP |
2006083542 | Aug 2006 | WO |
2006112190 | Oct 2006 | WO |
Entry |
---|
Combined Search and Examination Report for GB 1709064.8, dated Nov. 8, 2017, 6 pages. |
Extended European Search Report dated Oct. 30, 2018 in EP Application No. 18170259.8, 6 pages. |
Office Action issued in JP 2018-101271, dated May 31, 2022 with English translation (9 pages). |
Ryan, Bob, “M1 of Cyrix to challenge to Pentium”, Nikkei Byte, Mar. 1, 1994, Nikkei Business Publications, Inc., No. 123, pp. 295-301, with English Translation (11 pages). |
Office Action issued in KR 10-2018-0062243, dated Nov. 20, 2022, with English translation (12 pages). |
Number | Date | Country | |
---|---|---|---|
20210089323 A1 | Mar 2021 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15974769 | May 2018 | US |
Child | 16950936 | US |