A cache in a central processing unit is a data storage structure that is used by the central processing unit of a computer to reduce the average time that it takes to access memory. It is a memory which stores copies of data that is located in the most frequently used main memory locations. Moreover, cache memory is memory that is smaller and that may be accessed more quickly than main memory. There are several different types of caches. These include physically indexed physically tagged (PIPT), virtually indexed virtually tagged (VIVT) and virtually indexed physically tagged (VIPT).
Caches that can accommodate multiple accesses in a single cycle provide performance advantages. In particular, such caches feature reduced access latencies. Conventional approaches to accommodating multiple accesses in a single cycle include the use of multi-ported caches and the provision of caches that include a plurality of tag and data banks.
A multi-ported cache is a cache which can serve more than one request at a time. In accessing some conventional caches a single memory address is used, whereas in a multi-ported cache, N memory addresses can be requested at a time, where N is the number of ports that is possessed by the multi-ported cache. An advantage of a multi ported cache is that greater throughput (e.g., a greater number of load and store requests) may be accommodated. However, the number of cache ports that are needed to accommodate increasingly high levels of throughput may not be practical.
Caches that include a plurality of tag and data banks can serve more than one request at a time as each bank can serve at least one request. However, when more than one request attempts to access the same bank, the request that will be allowed to access the bank must be determined. In one conventional approach, serial arbitrations are used to determine which request will be allowed to access the data bank and to access the tag bank for the tag that corresponds to the request. However, the time that it takes to execute such arbitrations can delay access to the bank and thus delay the triggering of the critical Load Hit signal. Moreover, because a plurality of both loads and stores may need to be accommodated, the additional banks, ports and arbitration that may be needed can exact unsatisfactory latency penalties.
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter.
Conventional approaches to accommodating throughput that involves multiple accesses can result in unsatisfactory latency penalties as it relates to the timing of load hit signals. A method for supporting a plurality of load and store accesses of a cache is disclosed that addresses these shortcomings. However, the claimed embodiments are not limited to implementations that address any or all of the aforementioned shortcomings. As a part of the aforementioned method, responsive to a request of a plurality of requests to access a block of a plurality of blocks of a load cache, the block of the load cache and a logically and physically paired block of a store coalescing cache are accessed in parallel. The data that is accessed from the block of the load cache is overwritten by the data that is accessed from the block of the store coalescing cache by merging on a per byte basis. Thereafter, access is provided to the merged data. Because the data is merged locally and block access arbitrations involving a plurality of access requests to the load cache and the store coalescing cache are executed in parallel with the identification of associated tags, a plurality of loads and stores are provided quick access to the data maintained by the load cache and the store coalescing cache.
The invention, together with further advantages thereof, may best be understood by reference to the following description taken in conjunction with the accompanying drawings in which:
It should be noted that like reference numbers refer to like elements in the figures.
Although the present invention has been described in connection with one embodiment, the invention is not intended to be limited to the specific forms set forth herein. On the contrary, it is intended to cover such alternatives, modifications, and equivalents as can be reasonably included within the scope of the invention as defined by the appended claims.
In the following detailed description, numerous specific details such as specific method orders, structures, elements, and connections have been set forth. It is to be understood however that these and other specific details need not be utilized to practice embodiments of the present invention. In other circumstances, well-known structures, elements, or connections have been omitted, or have not been described in particular detail in order to avoid unnecessarily obscuring this description.
References within the specification to “one embodiment” or “an embodiment” are intended to indicate that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the present invention. The appearance of the phrase “in one embodiment” in various places within the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. Moreover, various features are described which may be exhibited by some embodiments and not by others. Similarly, various requirements are described which may be requirements for some embodiments but not other embodiments.
Some portions of the detailed descriptions, which follow, are presented in terms of procedures, steps, logic blocks, processing, and other symbolic representations of operations on data bits within a computer memory. These descriptions and representations are the means used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. A procedure, computer executed step, logic block, process, etc., is here, and generally, conceived to be a self-consistent sequence of steps or instructions leading to a desired result. The steps are those requiring physical manipulations of physical quantities. Usually, though not necessarily, these quantities take the form of electrical or magnetic signals of a computer readable storage medium and are capable of being stored, transferred, combined, compared, and otherwise manipulated in a computer system. It has proven convenient at times, principally for reasons of common usage, to refer to these signals as bits, values, elements, symbols, characters, terms, numbers, or the like.
It should be borne in mind, however, that all of these and similar terms are to be associated with the appropriate physical quantities and are merely convenient labels applied to these quantities. Unless specifically stated otherwise as apparent from the following discussions, it is appreciated that throughout the present invention, discussions utilizing terms such as “accessing” or “comparing” or “merging” or “providing access” or the like, refer to the action and processes of a computer system, or similar electronic computing device that manipulates and transforms data represented as physical (electronic) quantities within the computer system's registers and memories and other computer readable media into other data similarly represented as physical quantities within the computer system memories or registers or other such information storage, transmission or display devices.
Referring to
Referring to
In one embodiment, store coalescing cache 103a and load cache 103b are organized such that related entries are maintained in blocks of the respective caches that are positioned similarly within the caches as is shown in
In one embodiment, the data maintained in store coalescing cache 103a is considered to be more up to date (e.g., is considered to be the latest version) than the data that is maintained in the logically and physically paired load cache 103b. Accordingly, the data that is maintained in store coalescing cache 103a is given priority over data that is maintained in load cache 103b in merge operations involving such data. This is shown in
Referring again to
In one embodiment, system 101, is able to facilitate the execution of a plurality of loads and stores in a single cycle based on the organization of load cache 103b and store coalescing cache 103a. In one embodiment, the parallelism facilitated by system 101 operates to provide quicker load and store accesses. In one embodiment, system 101 can be located in a cache controller (not shown). In other embodiments, system 101 can be separate from a cache controller, but operate cooperatively therewith.
Main memory 111 includes physical addresses that store the information that is copied into cache memory. In one embodiment, when the information that is contained in the physical addresses of main memory that have been cached is changed, the corresponding cached information is updated to reflect the changes made to the information stored in main memory. Also shown in
Operation
Referring to
At B, the block of the store coalescing cache that logically and physically corresponds to the specific block of the level one load cache is simultaneously identified. In one embodiment, because both the load cache and the store coalescing cache are similarly organized, the bits of the virtual address associated with the request that addresses the aforementioned specific block of the load cache, simultaneously addresses the similarly situated, logically and physically paired, block of the store coalescing cache.
At C, a read enable that corresponds to the block of the store coalescing cache and a read enable that corresponds to the block of the load cache are asserted in parallel (because arbitration for both blocks are the same) and data from each of the blocks is accessed.
At D, the data accessed from the accessed block of the store coalescing cache and the data accessed from the accessed block of the load cache is locally compared.
At E, the data accessed from the accessed block of the store coalescing cache and the data accessed from the accessed block of the load cache is merged. And, the merged data is made available to the load requestor.
In exemplary embodiments stores can access the store coalescing cache without a corresponding access of the load cache. However, as discussed herein, load requests trigger a parallel access of both the load cache and the store coalescing cache. Exemplary embodiments accommodate a throughput of many loads and stores, do not penalize the timing of the “load hit” signal and maintain consistency between loads and stores.
Load request accessor 201 accesses requests to access blocks of a load cache (e.g., load cache 103b in
Block accessor 203, responsive to a request of a plurality of requests to access a block of a plurality of blocks in a load cache (e.g., 103b in
Data comparer/merger 205 compares and merges data accessed from the accessed block of the load cache to data that is accessed from the accessed block of the store coalescing cache on a per byte basis. In one embodiment, the data that is maintained in the store coalescing cache is given priority over the data that is maintained in the load cache. As such, data that is inconsistent with that which is supplied by the store coalescing cache is replaced by the data supplied from the store coalescing cache. In one embodiment, data comparer/merger 203 can be implemented using a multiplexer that compares and merges the data supplied by the store coalescing cache and the data supplied by the load cache. In one embodiment, the multiplexer can be a 2:1 multiplexer. In one embodiment, the comparison is locally performed (carried out inside the level one data cache).
Data access provider 207 provides access to the merged data to the load requestor. In one embodiment, the herein described processes ensure that the data provided can be a part of a throughput of many loads and stores, is provided in a manner that does not penalize the timing of the “load hit” signal and maintains consistency between loads and stores.
It should be appreciated that the aforementioned components of system 101 can be implemented in hardware or software or in a combination of both. In one embodiment, components and operations of system 101 can be encompassed by components and operations of one or more computer components or programs (e.g., a cache controller). In another embodiment, components and operations of system 101 can be separate from the aforementioned one or more computer components or programs but can operate cooperatively with components and operations thereof.
Referring to
At 303, data accessed from the block of the load cache is compared and merged with data that is accessed from the block of the store coalescing cache on a per byte basis. In one embodiment, as a part of the data comparison and merger process, the data that is maintained in the block of the store coalescing cache is given priority over the data that is maintained in the block of the load cache. Consequently, where the data in the block of the load cache and the data in the block of the store coalescing cache are different, the data from the block of the store coalescing cache is included in the merged data result.
At 305, access to the result of the data merger is provided to access requestors. In one embodiment, operations described herein ensure that access to the latest version of data that is associated with an address is provided to the requestor.
With regard to exemplary embodiments thereof, systems and methods for supporting a plurality of load and store accesses of a cache are disclosed. Responsive to a request of a plurality of requests to access a block of a plurality of blocks of a load cache, the block of the load cache and a logically paired block of the store coalescing cache are accessed in parallel. The data that is accessed from the block of the load cache is compared to the data that is accessed from the block of the store coalescing cache and merged on a per byte basis. Access is provided to the merged data.
Although many of the components and processes are described above in the singular for convenience, it will be appreciated by one of skill in the art that multiple components and repeated processes can also be used to practice the techniques of the present invention. Further, while the invention has been particularly shown and described with reference to specific embodiments thereof, it will be understood by those skilled in the art that changes in the form and details of the disclosed embodiments may be made without departing from the spirit or scope of the invention. For example, embodiments of the present invention may be employed with a variety of components and should not be restricted to the ones mentioned above. It is therefore intended that the invention be interpreted to include all variations and equivalents that fall within the true spirit and scope of the present invention.
This application is a continuation of, claims the benefit of, and priority to the copending non-provisional patent application Ser. No. 13/561,570, entitled “SYSTEMS AND METHODS FOR SUPPORTING A PLURALITY OF LOAD AND STORE ACCESSES OF A CACHE,” with filing date Jul. 30, 2012, and hereby incorporated by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
4075704 | O'Leary | Feb 1978 | A |
4245344 | Richter | Jan 1981 | A |
4356550 | Katzman et al. | Oct 1982 | A |
4414624 | Summer, Jr. et al. | Nov 1983 | A |
4524415 | Mills, Jr. et al. | Jun 1985 | A |
4527237 | Frieder et al. | Jul 1985 | A |
4577273 | Hopper et al. | Mar 1986 | A |
4597061 | Cline et al. | Jun 1986 | A |
4600986 | Scheuneman et al. | Jul 1986 | A |
4633434 | Scheuneman | Dec 1986 | A |
4682281 | Woffinden et al. | Jul 1987 | A |
4816991 | Watanabe et al. | Mar 1989 | A |
4920477 | Colwell et al. | Apr 1990 | A |
5294897 | Notani et al. | Mar 1994 | A |
5317705 | Gannon et al. | May 1994 | A |
5317754 | Blandy et al. | May 1994 | A |
5548742 | Wang et al. | Aug 1996 | A |
5559986 | Alpert et al. | Sep 1996 | A |
5574878 | Onodera et al. | Nov 1996 | A |
5581725 | Nakayama | Dec 1996 | A |
5634068 | Nishtala et al. | May 1997 | A |
5752260 | Liu | May 1998 | A |
5754818 | Mohamed | May 1998 | A |
5787494 | DeLano et al. | Jul 1998 | A |
5793941 | Pencis et al. | Aug 1998 | A |
5802602 | Rahman et al. | Sep 1998 | A |
5806085 | Berliner | Sep 1998 | A |
5813031 | Chou et al. | Sep 1998 | A |
5835951 | McMahan | Nov 1998 | A |
5852738 | Bealkowski et al. | Dec 1998 | A |
5860146 | Vishin et al. | Jan 1999 | A |
5864657 | Stiffler | Jan 1999 | A |
5872985 | Kimura | Feb 1999 | A |
5881277 | Bondi et al. | Mar 1999 | A |
5903750 | Yeh et al. | May 1999 | A |
5905509 | Jones et al. | May 1999 | A |
5918251 | Yamada et al. | Jun 1999 | A |
5956753 | Glew et al. | Sep 1999 | A |
5974506 | Sicola et al. | Oct 1999 | A |
6016533 | Tran | Jan 2000 | A |
6073230 | Pickett et al. | Jun 2000 | A |
6075938 | Bugnion et al. | Jun 2000 | A |
6088780 | Yamada et al. | Jul 2000 | A |
6092172 | Nishimoto et al. | Jul 2000 | A |
6101577 | Tran | Aug 2000 | A |
6115809 | Mattson, Jr. et al. | Sep 2000 | A |
6134634 | Marshall, Jr. et al. | Oct 2000 | A |
6138226 | Yoshioka et al. | Oct 2000 | A |
6157998 | Rupley, II et al. | Dec 2000 | A |
6167490 | Levy et al. | Dec 2000 | A |
6205545 | Shah et al. | Mar 2001 | B1 |
6212613 | Belair | Apr 2001 | B1 |
6226732 | Pei et al. | May 2001 | B1 |
6247097 | Sinharoy | Jun 2001 | B1 |
6253316 | Tran et al. | Jun 2001 | B1 |
6256727 | McDonald | Jul 2001 | B1 |
6256728 | Witt et al. | Jul 2001 | B1 |
6260131 | Kikuta et al. | Jul 2001 | B1 |
6260138 | Harris | Jul 2001 | B1 |
6272662 | Jadav | Aug 2001 | B1 |
6275917 | Okada | Aug 2001 | B1 |
6321298 | Hubis | Nov 2001 | B1 |
6332189 | Baweja et al. | Dec 2001 | B1 |
6341324 | Caulk, Jr. et al. | Jan 2002 | B1 |
6437789 | Tidwell | Aug 2002 | B1 |
6449671 | Patkar et al. | Sep 2002 | B1 |
6457120 | Sinharoy | Sep 2002 | B1 |
6557083 | Sperber et al. | Apr 2003 | B1 |
6594755 | Nuechterlein et al. | Jul 2003 | B1 |
6604187 | McGrath et al. | Aug 2003 | B1 |
6609189 | Kuszmaul et al. | Aug 2003 | B1 |
6658549 | Wilson et al. | Dec 2003 | B2 |
6681395 | Nishi | Jan 2004 | B1 |
6907600 | Neiger et al. | Jun 2005 | B2 |
6912644 | O'Connor et al. | Jun 2005 | B1 |
7007108 | Emerson et al. | Feb 2006 | B2 |
7111145 | Chen et al. | Sep 2006 | B1 |
7143273 | Miller et al. | Nov 2006 | B2 |
7149872 | Rozas et al. | Dec 2006 | B2 |
7213106 | Koster et al. | May 2007 | B1 |
7278030 | Chen et al. | Oct 2007 | B1 |
7380096 | Rozas et al. | May 2008 | B1 |
7406581 | Southwell | Jul 2008 | B2 |
7546420 | Shar et al. | Jun 2009 | B1 |
7680988 | Nickolls et al. | Mar 2010 | B1 |
7783868 | Ukai | Aug 2010 | B2 |
7856530 | Mu | Dec 2010 | B1 |
7913058 | Rozas et al. | Mar 2011 | B2 |
8145844 | Bruce | Mar 2012 | B2 |
8239656 | Rozas et al. | Aug 2012 | B2 |
8301847 | Dantzig | Oct 2012 | B2 |
8522253 | Rozas et al. | Aug 2013 | B1 |
8868838 | Glasco et al. | Oct 2014 | B1 |
8930674 | Avudaiyappan et al. | Jan 2015 | B2 |
9047178 | Talagala et al. | Jun 2015 | B2 |
20010049782 | Hsu et al. | Dec 2001 | A1 |
20020069326 | Richardson et al. | Jun 2002 | A1 |
20020082824 | Neiger et al. | Jun 2002 | A1 |
20020099913 | Steely, Jr. | Jul 2002 | A1 |
20030065887 | Maiyuran et al. | Apr 2003 | A1 |
20030088752 | Harman | May 2003 | A1 |
20040034762 | Kacevas | Feb 2004 | A1 |
20040044850 | George et al. | Mar 2004 | A1 |
20040064668 | Kjos et al. | Apr 2004 | A1 |
20040093483 | Nguyen et al. | May 2004 | A1 |
20040103251 | Alsup | May 2004 | A1 |
20040117593 | Uhlig et al. | Jun 2004 | A1 |
20040117594 | VanderSpek | Jun 2004 | A1 |
20040143727 | McDonald | Jul 2004 | A1 |
20040193857 | Miller et al. | Sep 2004 | A1 |
20040205296 | Bearden | Oct 2004 | A1 |
20040215886 | Cargnoni et al. | Oct 2004 | A1 |
20040225872 | Bonanno et al. | Nov 2004 | A1 |
20050005085 | Miyanaga | Jan 2005 | A1 |
20050027961 | Zhang | Feb 2005 | A1 |
20050060457 | Olukotun | Mar 2005 | A1 |
20050108480 | Correale, Jr. et al. | May 2005 | A1 |
20050154867 | DeWitt, Jr. et al. | Jul 2005 | A1 |
20060004964 | Conti et al. | Jan 2006 | A1 |
20060026381 | Doi et al. | Feb 2006 | A1 |
20060190707 | McIlvaine et al. | Aug 2006 | A1 |
20060236074 | Williamson et al. | Oct 2006 | A1 |
20060277365 | Pong | Dec 2006 | A1 |
20080077813 | Keller et al. | Mar 2008 | A1 |
20080091880 | Vishin | Apr 2008 | A1 |
20080126771 | Chen et al. | May 2008 | A1 |
20080195844 | Shen et al. | Aug 2008 | A1 |
20080215865 | Hino et al. | Sep 2008 | A1 |
20080235500 | Davis et al. | Sep 2008 | A1 |
20080270758 | Ozer et al. | Oct 2008 | A1 |
20080270774 | Singh et al. | Oct 2008 | A1 |
20080282037 | Kusachi et al. | Nov 2008 | A1 |
20090138659 | Lauterbach | May 2009 | A1 |
20090157980 | Bruce | Jun 2009 | A1 |
20090158017 | Mutlu et al. | Jun 2009 | A1 |
20090164733 | Kim et al. | Jun 2009 | A1 |
20090172344 | Grochowski et al. | Jul 2009 | A1 |
20090287912 | Sendag | Nov 2009 | A1 |
20100138607 | Hughes et al. | Jun 2010 | A1 |
20100169578 | Nychka | Jul 2010 | A1 |
20100169611 | Chou et al. | Jul 2010 | A1 |
20100211746 | Tsukishiro | Aug 2010 | A1 |
20110010521 | Wang et al. | Jan 2011 | A1 |
20110082980 | Gschwind et al. | Apr 2011 | A1 |
20110082983 | Koktan | Apr 2011 | A1 |
20110153955 | Herrenschmidt et al. | Jun 2011 | A1 |
20120005462 | Hall et al. | Jan 2012 | A1 |
20120042126 | Krick et al. | Feb 2012 | A1 |
20130019047 | Podvalny et al. | Jan 2013 | A1 |
20130046934 | Nychka | Feb 2013 | A1 |
20130086417 | Sivaramakrishnan et al. | Apr 2013 | A1 |
20130097369 | Talagala et al. | Apr 2013 | A1 |
20130238874 | Avudaiyappan et al. | Sep 2013 | A1 |
20130304991 | Bottcher et al. | Nov 2013 | A1 |
20130311759 | Abdallah | Nov 2013 | A1 |
20130346699 | Walker | Dec 2013 | A1 |
20140032844 | Avudaiyappan et al. | Jan 2014 | A1 |
20140032845 | Avudaiyappan et al. | Jan 2014 | A1 |
20140032856 | Avudaiyappan et al. | Jan 2014 | A1 |
20140075168 | Abdallah | Mar 2014 | A1 |
20140108730 | Avudaiyappan et al. | Apr 2014 | A1 |
20140156947 | Avudaiyappan et al. | Jun 2014 | A1 |
20140281242 | Abdallah et al. | Sep 2014 | A1 |
20160041908 | Avudaiyappan et al. | Feb 2016 | A1 |
20160041930 | Avudaiyappan et al. | Feb 2016 | A1 |
Number | Date | Country |
---|---|---|
1305150 | Jul 2001 | CN |
0596636 | May 1994 | EP |
0706133 | Apr 1996 | EP |
2343270 | May 2000 | GB |
200707284 | Mar 1995 | TW |
539996 | Jul 2003 | TW |
200401187 | Jan 2004 | TW |
591530 | Jun 2004 | TW |
I233545 | Jun 2005 | TW |
I281121 | May 2007 | TW |
WO0125921 | Apr 2001 | WO |
Entry |
---|
Barham et al., “Xen and the Art of Virtualization,” ACM Symposium on Operating Systems Principles, Oct. 19, 2003, pp. 164-177. |
Cooperman, Gene; Cache Basics, 2003, http://www.ccs.neu.edu/course/com3200/parent/Notesicache-basics.html, pp. 1-3. |
Final Office Action from U.S. Appl. No. 13/561,441, dated Oct. 14, 2016, 14 pages. |
Final Office Action from U.S. Appl. No. 13/561,441, dated Sep. 26, 2014, 16 pages. |
Final Office Action from U.S. Appl. No. 13/561,491, dated Jan. 14, 2015, 14 pages. |
Final Office Action from U.S. Appl. No. 13/561,491, dated Nov. 12, 2014, 14 pages. |
Final Office Action from U.S. Appl. No. 13/561,491, dated Oct. 13, 2015, 16 pages. |
Final Office Action from U.S. Appl. No. 13/561,528, dated Nov. 10, 2014, 19 pages. |
Final Office Action from U.S. Appl. No. 14/173,602, dated Jan. 8, 2016, 32 pages. |
Final Office Action from U.S. Appl. No. 14/173,602, dated Jan. 9, 2017, 18 pages. |
Final Office Action from U.S. Appl. No. 14/922,042, dated Oct. 14, 2016, 15 pages. |
Final Office Action from U.S. Appl. No. 14/922,053, dated Oct. 14, 2016, 16 pages. |
Garmany J., “The Power of Indexing,” archived on Mar. 9, 2009, 7 pages. |
International Preliminary Report on Patentability for Application No. PCT/US2013/051128, dated Feb. 12,2015, 8 pages. |
International Search Report and Written Opinion for Application No. PCT/US2013/051128, dated Oct. 30, 2013, 9 pages. |
Jacobson et al., “Path-based Next Trace Prediction,” IEEE, 1997, pp. 14-23. |
Nanda Ax., et al., “The Misprediction Recovery Cache,” International Journal of Parallel Programming, Plenum Publishing Corporation, 1998, vol. 26 (4), pp. 383-415. |
Non-Final Office Action from U.S. Appl. No. 13/561,441, dated Dec. 22, 2014, 13 pages. |
Non-Final Office Action from U.S. Appl. No. 13/561,441, dated Jan. 39, 2017, 17 pages. |
Non-Final Office Action from U.S. Appl. No. 13/561,441, dated Jun. 4, 2014, 12 pages. |
Non-Final Office Action from U.S. Appl. No. 13/561,441, dated May 17, 2016, 17 pages. |
Non-Final Office Action from U.S. Appl. No. 13/561,491, dated Feb. 8, 2016, 16 pages. |
Non-Final Office Action from U.S. Appl. No. 13/561,491, dated Jun. 16, 2014, 13 pages. |
Non-Final Office Action from U.S. Appl. No. 13/561,491, dated May 5, 2015, 15 pages. |
Non-Final Office Action from U.S. Appl. No. 13/561,528, dated Dec. 19, 2014, 13 pages. |
Non-Final Office Action from U.S. Appl. No. 13/561,528, dated Jun. 17, 2014 , 11 pages. |
Non-Final Office Action from U.S. Appl. No. 14/173,602, dated Jul. 29, 2016, 19 pages. |
Non-Final Office Action from U.S. Appl. No. 14/173,602, dated Sep. 8, 2015, 32 pages. |
Non-Final Office Action from U.S. Appl. No. 14/922,042, dated Apr. 7, 2016, 16 pages. |
Non-Final Office Action from U.S. Appl. No. 14/922,042, dated Dec. 14, 2016, 20 pages. |
Non-Final Office Action from U.S. Appl. No. 14/922,053, dated Apr. 7, 2016, 15 pages. |
Non-Final Office Action from U.S. Appl. No. 14/922,053, dated Dec. 15, 2016, 16 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,441, dated Feb. 2, 2016, 10 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,441, dated Jun. 23, 2015, 14 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,441, dated Mar. 18, 2015, 8 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,441, dated Oct. 21, 2015, 10 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,491, dated Dec. 19, 2016, 13 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,491, dated Jun. 8, 2016, 12 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,491, dated Sep. 14, 2016, 18 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,528, dated Aug. 3, 2016, 6 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,528, dated Feb. 8, 2016, 15 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,528, dated Jul. 13, 2015, 14 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,528, dated Mar. 31, 2015, 5 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,528, dated May 2, 2016, 10 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,528, dated Oct. 21, 2015, 9 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,570, dated Aug. 27, 2015, 9 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,570, dated Dec. 7, 2015, 7 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,570, dated Jan. 29, 2015, 9 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,570, dated Jun. 20, 2014, 8 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,570, dated May 20, 2015, 3 pages. |
Notice of Allowance from U.S. Appl. No. 13/561,570, dated Oct. 22, 2014, 16 pages. |
Rotenberg E., et al.,“Trace Cache: a Low Latency Approach to High Bandwidth Instruction Fetching,” Apr. 11, 1996, 48 pages. |
Techopedia, Cache Memory Definition, www.techopedia.com/definition/6307/cache-memory. |
Wallace S., et al.,“Multiple Branch and Block Prediction,” Third International symposium on High-Performance Computer Architecture, IEEE, Feb. 1997, pp. 94-103. |
Ye J., et al.,“A New Recovery Mechanism in Superscalar Microprocessors by Recovering Critical Misprediction,”IEICE Transactions on Fundamentals of Electronics, Communications and Computer Sciences, 2011, vol. E94-A (12), pp. 2639-2648. |
Yeh T., et al.,“Increasing the Instruction Fetch Rate Via Multiple Branch Prediction and a Branch Address Cache,” 7th International Conference on Supercomputing, ACM, 1993, pp. 67-76. |
Number | Date | Country | |
---|---|---|---|
20160041913 A1 | Feb 2016 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13561570 | Jul 2012 | US |
Child | 14922035 | US |