The present patent application is related to copending applications:
1. Technical Field
The present invention relates to multiprocessor systems in general, and in particular to memory controllers for multiprocessor systems. Still more particularly, the present invention relates to a method and apparatus for supporting low-overhead memory locks within a multiprocessor system.
2. Description of Related Art
A multiprocessor system typically requires a mechanism for synchronizing operations of various processors within the multiprocessor system in order to allow interactions among those processors that work on a task. Thus, the instruction set of processors within a multiprocessor system are commonly equipped with explicit instructions for handling task synchronization. For example, the instruction set of PowerPC® processors, which are manufactured by International Business Machines Corporation of Armonk, N.Y., provides instructions such as lwarx or ldwarx and stwcx or stdwx (hereafter referred to as larx and stcx, respectively) for building synchronization primitives.
The larx instruction loads an aligned word of memory into a register within a processor. In addition, the larx instruction places a “reservation” on the block of memory that contains the word of memory accessed. The reservation contains the address of the memory block and a flag. The flag is made active, and the address of the memory block is loaded when a larx instruction successfully reads the word of memory referenced. If the reservation is valid (i.e., the flag is active), the processor and the memory hierarchy are obligated to monitor the entire processing system cooperatively for any operation that attempts to write to the memory block at which the reservation exists.
The reservation flag is used to control the behavior of a stcx instruction that is the counterpart to the larx instruction. The stcx instruction first determines if the reservation flag is valid. If so, the stcx instruction performs a Store to the word of memory specified, sets a condition code register to indicate that the Store has succeeded, and resets the reservation flag. If, on the other hand, the reservation flag in the reservation is not valid, the stcx instruction does not perform a Store to the word of memory and sets a condition code register indicating that the Store has failed. The stcx instruction is often referred to as a “Conditional Store” due to the fact that the Store is conditional on the status of the reservation flag.
The general concept underlying the larx/stcx instruction sequence is to allow a processor to read a memory location, to modify the memory location in some way, and to store the new value to the memory location while ensuring that no other processor within a multiprocessor system has altered the memory location from the point in time when the larx instruction was executed until the stcx instruction completes. Such a sequence is usually referred to as an “atomic read-modify-write” sequence because a processor was able to read a memory location, modify a value within the memory location, and then write a new value without any interruption by another processor writing to the same memory location. The larx/stcx sequence of operations do not occur as one uninterruptable sequence, but rather, the fact that the processor is able to execute a larx instruction and then later successfully complete the stcx instruction ensures a programmer that the read/modify/write sequence did, in fact, occur as if it were atomic. This atomic property of a larx/stcx sequence can be used to implement a number of synchronization primitives well-known to those skilled in the art.
The larx/stcx sequence of operations work well with cache memories that are in close proximity with processors. However, the larx/stcx sequence of operations are not efficient for accessing a system memory, especially when many processors, which are located relatively far away from the system memory, are attempting to access the same memory block. Consequently, it would be desirable to provide a method and apparatus for supporting low-overhead memory locks for a system memory within a multiprocessor system.
In accordance with a preferred embodiment of the present invention, a lock control section is assigned to a data block within a system memory of a multiprocessor system. In response to a request for accessing the data block by a processing unit within the multiprocessor system, a determination is made by a memory controller whether or not the lock control section of the data block has been set. If the lock control section of the data block has been set, the request for accessing the data block is denied. Otherwise, if the lock control section of the data block has not been set, the lock control section of the data block is set, and the request for accessing the data block is allowed by the memory controller.
All features and advantages of the present invention will become apparent in the following detailed written description.
The invention itself, as well as a preferred mode of use, further objects, and advantages thereof, will best be understood by reference to the following detailed description of an illustrative embodiment when read in conjunction with the accompanying drawings, wherein:
With reference now to the drawings, and in particular to
Processing units 11a-11n, which may have homogeneous or heterogeneous processor architectures, use a common set of instructions to operate. As a general example of processing units 11a-11n, processing unit 11a includes a processor core 12 having multiple execution units (not shown) for executing program instructions. Processing unit 11a has one or more level-one caches, such as an instruction cache 13 and a data cache 14, which are implemented with high-speed memory devices. Instruction cache 13 and data cache 14 are utilized to store instructions and data, respectively, that may be repeatedly accessed by processing unit 11a in order to avoid long delay time associated with loading the same information from system memory 19. Processing unit 11a may also include level-two caches, such as an L2 cache 15 for supporting caches 13 and 14.
With reference now to
For example, as shown in
Referring now to
Lock control section 31 of data block 30 allows a memory controller, such as memory controller 18 from
With reference now to
The request is preferably made by a requesting processing unit to a memory controller via a Memory-Lock Load instruction, which is distinguished from a conventional Load instruction. As will be explained below, the Memory-Lock Load instruction allows the memory controller to set a lock control section of the requested data block (such as lock control section 31 of data block 30 from
The determination is preferably made by the memory controller via a checking of a lock control section of the requested data block. As shown in
If the requested data block is being accessed by another processing unit within the multiprocessor system, the requesting processing unit is not allowed to access the requested data block, and the requesting processing unit is invited to retry, as shown in block 43, and the process returns to block 42. Otherwise, if the requested data block is not being accessed by another processing unit within the multiprocessor system, the lock control section of the requested data block is set to a logical “1” to prevent other processing unit from accessing the requested data block, as depicted in block 44, and the requesting processing unit is allowed to access the requested data block.
After the access of the requested data block has been completed, as shown in block 45, the lock control section of the requested data block is reset to a logical “0” to allow other processing unit to access the requested data block, as depicted in block 46.
The requesting processing unit preferably signifies the completion of access to the memory controller via a Memory-Unlock Store instruction, which is distinguished from a conventional Store instruction. The Memory-Unlock Store instruction allows the memory controller to reset the lock control section of the requested data block (i.e., unlocking the requested data block) such that other processing units can access the requested data block again. After the requesting processing unit has initially gained control of the requested data block via a Memory-Lock Load instruction, the requesting processing unit can perform many Load or Memory-Lock Load instructions. However, the requesting processing unit can only perform one Memory-Unlock Store instruction for the memory controller to release the lock on the request data block.
Although it has been explained that the lock control section is to be implemented in the first bit of the first byte within a data block, it is understood by those skilled in the art that the lock control section can be implemented by more than one bit, and/or can be located anywhere within the first or other byte of a data block.
In block 43 of
When there are more than one processing units requesting for the same data block that is currently being accessed by another processing unit, instead of inviting all requesting processing units to retry, it may be more beneficial to inform a requesting processing unit located relatively far away from memory controller 18 to perform more useful operations other than retry. This is because the retry time is relatively long for requesting processing units that are located farther away from memory controller 18 than those that are closer. The relative distance of a requesting processing unit to memory controller 18 can be found in distance field 22 of processing unit tracking table 20 from
Alternatively, instead of inviting a requesting processing unit to retry, the memory controller can also place the access request from the requesting processing unit in a queue when the requested data is being accessed by another processing unit. Referring back to
After processing unit 11b has completed its access to data block 1234ABCD, memory controller 18 will allow processing unit 11c to gain access to data block 1234ABCD, and the processing unit number of processing unit 11a will be moved from slot 2 to slot 1. Similarly, after processing unit 11c has completed its access to data block 1234ABCD, memory controller 18 will allow processing unit 11a to gain access to data block 1234ABCD, and the address of data block 1234ABCD along with the processing unit number of processing unit 11a will be removed from queue table 25. Although each entry of queue table 25 is shown to have a queue depth of two, it is understood by those skilled in the art that a queue depth of less or more than two is also permissible.
All of the above-mentioned options for handling data block access denial are implemented from the perspective of a memory controller, such as memory controller 18 from
As has been described, the present invention provides a method and apparatus for supporting low-overhead memory locks within a system memory of a multiprocessor system.
While an illustrative embodiment of the present invention has been described in the context of a fully functional data processing system, those skilled in the art will appreciate that the software aspects of an illustrative embodiment of the present invention are capable of being distributed as a program product in a variety of forms, and that an illustrative embodiment of the present invention applies equally regardless of the particular type of media used to actually carry out the distribution. Examples of the types of media include recordable type media such as thumb drives, floppy disks, hard drives, CD ROMs, DVDs.
While the invention has been particularly shown and described with reference to a preferred embodiment, it will be understood by those skilled in the art that various changes in form and detail may be made therein without departing from the spirit and scope of the invention.
This invention was made with United States Government support under Agreement number HR0011-07-9-0002 awarded by DARPA. The Government has certain rights in the invention.
Number | Name | Date | Kind |
---|---|---|---|
5179669 | Peters | Jan 1993 | A |
5206946 | Brunk | Apr 1993 | A |
5263155 | Wang | Nov 1993 | A |
5353425 | Malamy et al. | Oct 1994 | A |
5487162 | Tanaka et al. | Jan 1996 | A |
5592641 | Fandrich et al. | Jan 1997 | A |
6393533 | Mende et al. | May 2002 | B1 |
6460124 | Kagi et al. | Oct 2002 | B1 |
6490662 | Pong et al. | Dec 2002 | B1 |
6546443 | Kakivaya et al. | Apr 2003 | B1 |
6622189 | Bryant et al. | Sep 2003 | B2 |
6633960 | Kessler et al. | Oct 2003 | B1 |
6801986 | Steely, Jr. et al. | Oct 2004 | B2 |
7058948 | Hoyle | Jun 2006 | B2 |
20030088755 | Gudmunson et al. | May 2003 | A1 |
20030126381 | Vo | Jul 2003 | A1 |
20040039962 | Ganesh et al. | Feb 2004 | A1 |
20040073905 | Emer et al. | Apr 2004 | A1 |
20040073909 | Arimilli et al. | Apr 2004 | A1 |
20040143712 | Armstrong et al. | Jul 2004 | A1 |
20040181636 | Martin et al. | Sep 2004 | A1 |
20040186988 | Polyudov | Sep 2004 | A1 |
20050033948 | Wei et al. | Feb 2005 | A1 |
20060218382 | Chen et al. | Sep 2006 | A1 |
20070094430 | Speier et al. | Apr 2007 | A1 |
20070204121 | O'Connor | Aug 2007 | A1 |
20070214333 | Nijhawan et al. | Sep 2007 | A1 |
20070260827 | Zimmer et al. | Nov 2007 | A1 |
20080235690 | Ang et al. | Sep 2008 | A1 |
20090198849 | Arimilli et al. | Aug 2009 | A1 |
20090198916 | Arimilli et al. | Aug 2009 | A1 |
20090198920 | Arimilli et al. | Aug 2009 | A1 |
20090198933 | Arimilli et al. | Aug 2009 | A1 |
Entry |
---|
Krofcheck, M.—Non-final Office Action, U.S. Appl. No. 12/024,181, dated Dec. 23, 2010. |
Patel, K.—Non-final Office Action, U.S. Appl. No. 12/024,234, dated Jul. 19, 2010. |
Patel, K.—Final Office Action, U.S. Appl. No. 12/024,234, dated Nov. 24, 2010. |
Krofcheck, M.—Non-final Office Action, U.S. Appl. No. 12/024,223, dated Dec. 28, 2010. |
IBM TDB—‘Shared Memory Addressing Structure’; Technical Disclosure Bulletin, Aug. 1978; vol. 21 iss. 3, pp. 998-1000. |
Patel, K.—Nonfinal Office Action dated Nov. 26, 2010, U.S. Appl. No. 12/024,245. |
Number | Date | Country | |
---|---|---|---|
20090198849 A1 | Aug 2009 | US |