1. Field of the Invention
This invention relates generally to computer processing techniques and, more particularly, to methods involving branch prediction in computer processing.
2. Description of Background
Branch prediction is used to improve the performance of processors. When a processor detects a conditional branch, an uncertainty is temporarily introduced into the processor pipeline. If the branch is taken, the next instruction is fetched from an address usually specified in the branch instruction. If the branch is not taken, execution proceeds to the instruction following the branch.
Large amounts of chip area are dedicated to the processor branch prediction mechanism. In practice, the branch address and target address of each branch encountered by the processor are saved in a table, called a Branch History Table (BHT). During the instruction fetch phase of a processor pipeline, the BHT is searched for a matching branch address, and if found, the target address is fetched, and the instruction located at the target address becomes the next instruction decoded. If no matching branch address is found in the instruction fetch segment, instruction fetching and decoding continue down the sequential path. Branch prediction errors occur when the table is incorrect. If the BHT predicted the branch to be taken and the predicted branch is not taken, the BHT entry is deleted or updated to record the correct branch action pattern. If a branch predicted not taken is taken, typically a new entry is made in the BHT, or the new branch pattern is updated. If the predicted target address is wrong, the correct address is saved in the BHT.
The prediction accuracy of the branch prediction mechanism (BHT) is proportional to the size of the BHT. Branch prediction accuracy may reach 80 to 90 percent, but there is a severe penalty when the BHT is incorrect. Prediction errors may cause the processor pipeline to be flushed, causing pipeline drain. A single prediction error may introduce a pipeline bubble (idle cycles) of 10 to 20 cycles or longer, depending on the pipeline length.
Increasing the size of the BHT reduces the penalty caused by prediction errors. The BHT may contain 4K to 16K entries, and with each entry approximately 8 bytes, (4 bytes for the branch address and 4 bytes for the target address) the overall size of the BHT may be 32K to 128K bytes. Although a larger BHT would reduce the percentage of wrong predictions and resultant penalties, the table hardware must be packaged in the speed critical instruction fetch and decode regions of the processor. The location of the table hardware limits the size of the BHT.
It is desirable to reduce the table hardware associated with the BHT without sacrificing prediction accuracy.
The shortcomings of the prior art are overcome and additional advantages are achieved through an exemplary method for branch prediction, the method comprising, receiving a load instruction including a first data location in a first memory area, retrieving data including a branch address and a target address from the first data location data location, and saving the data in a branch prediction memory.
An alternate exemplary method for branch prediction, the method comprising, receiving an unload instruction including a first data location in a first memory area, retrieving data including a branch address and a target address from the branch prediction memory, and saving the data in the first data location data location.
Additional features and advantages are realized through the techniques of the present invention. Other embodiments and aspects of the invention are described in detail herein and are considered a part of the claimed invention. For a better understanding of the invention with advantages and features, refer to the description and to the drawings.
The subject matter that is regarded as the invention is particularly pointed out and distinctly claimed in the claims at the conclusion of the specification. The foregoing and other aspects, features, and advantages of the invention are apparent from the following detailed description taken in conjunction with the accompanying drawings in which:
a and 3b illustrate an exemplary embodiment of a high level block diagram of a program executed by a processor.
a-c illustrate exemplary embodiments of Load BHT instructions.
a-c illustrate exemplary embodiments of Unload BHT instructions.
a-b illustrate alternate exemplary embodiments of Unload BHT instructions.
The detailed description explains the preferred embodiments of the invention, together with advantages and features, by way of example with reference to the drawings.
Methods involving prefetching into a branch prediction mechanism are provided.
In this regard, it is desirable to reduce the size of a branch history table (BHT) size while improving branch prediction accuracy and processor performance. To accomplish this, a mechanism is described that allows future branch information to be prefetched into the BHT.
Prefetching future branch information into a prefetched branch history table (PBHT) allows a designer to reduce the size of the BHT without sacrificing branch prediction accuracy or performance. The branch information used to process the upcoming (future) instructions and branches of a program is held in the speed critical area (PBHT), while the full program and branch information are maintained in program save areas or main memory. The full size of the branch prediction mechanism is no longer limited in size by the hardware or cycle time considerations. The size of the PBHT need only by limited to the amount of branch information needed to capture the imminent working set of a program and the ability of the processor to stage (prefetch) the upcoming information into the PBHT on a timely basis.
Prefetching will improve processor performance if the items prefetched are used during branch prediction, and are fetched far enough in advance of their use.
It is desirable to provide a mechanism to determine that the entries prefetched into the BHT are actually used by the processor. This further increases the prefetching accuracy for the items prefetched. This is achieved by providing each entry prefetched with a confirmation (used) bit as part of the BHT entry. The bit is initially unset (equals 0) when a prefetch occurs and reset (equals 1) if the prefetch is used during branch prediction. If the processor does not use the prefetched entry, the confirmation bit associated with that entry would prevent that entry from being prefetched in the future.
The above mechanisms are achieved by providing two new instructions: unload BHT (UBHT) and load BHT (LBHT). The unload BHT instruction takes the most recent (youngest) BHT entries and saves them (writes them) in a branch information save area, while the load BHT instruction reverses the direction that information is transferred. The load BHT instruction takes information from the branch information save area and writes it into the BHT. Since program execution patterns are repetitive, reloading the BHT with information collected during a previous execution phase of a program will produce a higher prefetch accuracy (prefetch to used ratio) than allowing the compiler to statically generate prefetches at compile time.
The contents of the PBHT are not based on the spatial and temporal referencing patterns, but on a relative certainty of future use as specified in the repetitive execution flow of a program. The PBHT control mechanism dynamically captures (saves) the branch information of a program and then prefetches (reloads) the relevant information back into the PBHT ahead of its use. Thus the processor requires fast access only to the branch information which is in close (logical) proximity to the current instructions being processed. This represents only a small subset of the information that would be contained in a full BHT. The information contained in a full BHT can be prefetched from main memory (collected in program save areas or compiler analysis) and prefetched into the PBHT, ahead of its use, to maintain a high prediction accuracy.
In operation, instructions requiring operands are sent to the address generation unit 106 via the decode module 104. The operands are fetched from the data cache 108 and the instructions are processed in the execution unit 110. Stores are sent from the execution unit 110 to the data cache 108.
The instruction fetch, and branch prediction mechanism processes occur in parallel to the decode, cache access, and execution flow described above. An instruction fetch address 112 is sent to the instruction fetch generator 114. The instruction fetch generator 114 generates a next instruction fetch address and saves the next instruction fetch address in the instruction fetch address 112. The instruction fetch generator 114 also sends the instruction fetch address to the instruction cache 118. The instruction cache 118 sends the contents of the instruction fetch address the decode module 104. Branch prediction occurs in parallel. The address of the instruction fetch is sent to the branch predict unit 116.
The branch prediction unit 116 includes the BHT 120 that saves the branch address and target address of the most recently executed branches encountered by the processor 100. The address from the instruction fetch generator 114 is compared with each branch address saved in the BHT. If a match occurs (a branch is contained in the most recent instruction fetch) the next instruction fetch address will be the predicted target address of the matching branch. This address then becomes the next instruction fetch and is sent back to the instruction fetch address 112 and instruction fetch generator 114. If no match occurs, the next instruction fetch is the next-sequential address from the current instruction fetch address. Finally, BHT updates are sent from the execution unit 110 to the BHT 120.
A prior art example of the BHT 120 format is illustrated in
a and 3b illustrate an exemplary embodiment of a high level block diagram of a program executed by a processor.
b further illustrates an exemplary embodiment of a block diagram showing the prefetching of information into the branch predictor (loading a BHT entry) and the saving of information that was recently in the branch predictor (unloading a BHT entry). The BHT entry loaded includes a branch address and a target address of an upcoming branch.
New BHT updates are staged (prefetched) into the BHT 320 when a BHT load operation is added to the entry to each program logic instruction 304. The BHT load operation avoids initial branch prediction errors caused by new (first time) taken branches. Initially the information contained in the BISA 308 is generated by a compiler, however future executions of the program will prefetch BHT entries that have been written to the BISA 308 by a previously executed unload BHT instruction. The prefetched BHT entries represent dynamically collected BHT entries that were saved the last time the program was executed. Dynamically collecting branch information and subsequent reloading of the branch information improves branch prediction accuracy while allowing the BHT 320 to be reduced in size.
The value in the new entry count (NEC) 404 register (of
The simplest form of a LBHT instruction prefetches a single BHT entry. The instruction consists of an opcode and branch/target address pair, where the branch/target address pair specifies the branch and target address for the upcoming branch. These addresses may be expressed as a relative distance from the LBHT instruction, base/displacement encoding of each address, register value plus offset, or represent the actual BHT entry settings (branch and target address values).
a-c illustrate exemplary embodiments of Load BHT instructions.
b illustrates an alternative exemplary embodiment of an encoding for the Load BHT instruction. The opcode field 702, the length field 704, and the address identifier 706 are similar to the fields described above regarding
Other examples of mask values 708 may specify a thread or program space identifier where only entries with matching tag information 606 (from the BISA 308) will be loaded into the BHT 320. Finally, other mask values 708 may select BHT entries according to an opcode type, or branch type (conditional, unconditional). All entries loaded into the NEB 402 have the confirmation bit set to zero. The confirmation bit will determine if the entry is unloaded into the BISA 308 when an unload BHT instruction is encountered.
c illustrates another alternative exemplary embodiment of an encoding for the Load BHT instruction. The opcode field 702 and the address identifier 706 are similar to the fields described above regarding
The formats for the Unload BHT instructions are illustrated in
Referring to
b illustrates an alternate exemplary embodiment of an unload BHT instruction similar to the instruction of
c illustrates another exemplary embodiment of an encoding of the unload BHT instruction. The mask 808, and BISA address 806 are similar to the fields described above regarding
a and 9b illustrate exemplary embodiments of the Unload-with-Confirmation BHT (UCBHT) instruction. In operation, only NEB 402 entries that have a confirmation bit set to 1 will be unloaded into the BISA 308.
The mask field 908 is similar to the mask value 908 illustrated in
b illustrates an alternate exemplary embodiment of the Unload BHT instruction. The mask value 908, and BISA address 906 shown are similar to the fields described in
The technical effects and benefits of the above described embodiments provide a method and system for accurate branch prediction with a reduced amount of hardware.
While the preferred embodiment to the invention has been described, it will be understood that those skilled in the art, both now and in the future, may make various improvements and enhancements which fall within the scope of the claims which follow. These claims should be construed to maintain the proper protection for the invention first described.
Number | Name | Date | Kind |
---|---|---|---|
4370711 | Smith | Jan 1983 | A |
4679141 | Pomerene et al. | Jul 1987 | A |
5606675 | Sakamura et al. | Feb 1997 | A |
5642493 | Burgess | Jun 1997 | A |
5682531 | Nakamura | Oct 1997 | A |
5742804 | Yeh et al. | Apr 1998 | A |
5774685 | Dubey | Jun 1998 | A |
5796971 | Emberson | Aug 1998 | A |
5859999 | Morris et al. | Jan 1999 | A |
5987495 | Ault et al. | Nov 1999 | A |
6029228 | Cai et al. | Feb 2000 | A |
6055621 | Puzak | Apr 2000 | A |
6092188 | Corwin et al. | Jul 2000 | A |
6108775 | Shiell et al. | Aug 2000 | A |
6189091 | Col et al. | Feb 2001 | B1 |
6272623 | Talcott | Aug 2001 | B1 |
6308322 | Serocki et al. | Oct 2001 | B1 |
6324643 | Krishnan et al. | Nov 2001 | B1 |
6408325 | Shaylor | Jun 2002 | B1 |
6427192 | Roberts | Jul 2002 | B1 |
6560693 | Puzak et al. | May 2003 | B1 |
6611910 | Sharangpani et al. | Aug 2003 | B2 |
6671762 | Soni et al. | Dec 2003 | B1 |
6725365 | Cofler et al. | Apr 2004 | B1 |
6760835 | Yu | Jul 2004 | B1 |
6877089 | Sinharoy | Apr 2005 | B2 |
6957327 | Gelman et al. | Oct 2005 | B1 |
7197630 | Alsup et al. | Mar 2007 | B1 |
7272664 | Arimilli et al. | Sep 2007 | B2 |
7441110 | Puzak et al. | Oct 2008 | B1 |
7870371 | Mutlu et al. | Jan 2011 | B2 |
20020194462 | Henry et al. | Dec 2002 | A1 |
20050278513 | Aristodemou et al. | Dec 2005 | A1 |
20060218385 | Smith et al. | Sep 2006 | A1 |
20070083739 | Glew | Apr 2007 | A1 |
20070088915 | Archambault et al. | Apr 2007 | A1 |
20070250666 | Sartorius et al. | Oct 2007 | A1 |
20080052499 | Koc | Feb 2008 | A1 |
20090089564 | Brickell et al. | Apr 2009 | A1 |
20100287358 | Emma et al. | Nov 2010 | A1 |
Number | Date | Country | |
---|---|---|---|
20090313462 A1 | Dec 2009 | US |