This application is related to U.S. patent application Ser. No. 13/693,869, filed Dec. 4, 2012, entitled “SYSTEMS AND METHODS FOR AUTOMATICALLY GENERATING MASTER-SLAVE LATCH STRUCTURES WITH FULLY REGISTERED FLOW CONTROL”, assigned to the assignee of the present disclosure, which is hereby incorporated herein by reference in its entirety.
In electronics, logic synthesis is a computerized process by which a description of a circuit's operation, design and organization is automatically converted into a design implementation that comprises logic gates. Common examples of this process involve the synthesis of designs expressed using hardware description languages (HDLs), including very high speed integrated circuit HDL (VHDL) and Verilog, typically at the register transfer level (RTL), that describe a circuit's operation, design and organization. Typically the synthesis tool takes RTL code, a standard cell library and user defined constraints and produces a gate level representation of the design. In producing the representation, it attempts to best meet the user defined constraints. Logic synthesis is one aspect of electronic design automation (EDA).
Logic synthesis systems can be configured to generate specific types of circuits. For example, a conventional synthesis system that is configured to generate random access memory (RAM) arrays, generates an 8-deep latch array (includes 8 latches) as is shown in
Circuit design characteristics that are used to determine the merits of circuit designs that are generated by logic synthesis systems include testability, circuit area, timing and power. These characteristics can depend on the components that are used to implement the designs. Synthesis systems such as those that generate latch arrays such as is shown in
Latch arrays generated by some conventional logic synthesis systems include a flip-flop and readout multiplexor that can occupy excessive space. A latch array that is generated by a logic synthesis system that uses latch pairs instead flip flops is disclosed that addresses these shortcomings. However, the claimed embodiments are not limited to implementations that address any or all of the aforementioned shortcomings.
In one embodiment, the latch array that is generated by the logic synthesis system is a synchronous sequential latch array that includes N/2 rows of master-slave latch pairs. In one embodiment, the logic synthesis system that generates the master-slave latch pairs can be invoked from another logic synthesis system that generates first-in-first-out (FIFO) structures based on configuration parameter settings. In one embodiment, the latch array provides fully-registered flow control. In one embodiment, the latch array that is generated utilizes latch pairs instead of the flip flops that are used in designs generated by some conventional logic synthesis systems. The elimination of a flip flop from the design, as compared to designs using such, saves area and latency. The latch array can be scan-tested and thus does not require a test harness which also saves area and improves timing. In addition, the read-out multiplexor can be smaller because it can be N/2 to 1 instead of N to 1. The use of a smaller multiplexor results in less consumption of power.
In one embodiment, a synchronous sequential access latch array generated by an automated system for generating master-slave latch structures is disclosed. A master-slave latch array includes N/2 rows of master-slave latch pairs, an N/2-to-1 multiplexer and control logic. N is equal to the number of addresses or latches that are included in the latch array.
In one embodiment, a method of operating a sequential access memory that includes a master-slave latch pair array is disclosed. The method includes, in a second half of a clock cycle, receiving data into an empty and open master latch in a first row of the master-slave latch pair array, if master-slave latch pairs of other rows are empty or if master-slave latch pairs of other rows are full, or, receiving data into an empty and open master latch of a subsequent row of the master-slave latch pair array if a master latch and a slave latch of a row or rows preceding the subsequent row are full. Additionally, the method includes, in a first half of a cycle subsequent to a clock cycle in which data is received by a master latch in a row of the master-slave latch pair array, receiving data into a slave latch of the row of the master-slave latch pair array and providing access to the data received into the slave latch of the row of the master-slave latch pair array.
In one embodiment, a method of operating a sequential access memory that includes a master-slave latch pair array is disclosed. The method includes receiving data into respective master latches in a second half of first respective clock cycles, wherein data that is received in the second half of the first respective clock cycles is allowed to flow uninhibited from respective master latches to respective corresponding slave latches. Additionally, the method includes receiving data into respective master latches in a second half of second respective clock cycles and providing access to data stored in the respective corresponding slave latches.
The embodiments of the present invention, together with further advantages thereof, may best be understood by reference to the following description taken in conjunction with the accompanying drawings in which:
It should be noted that like reference numbers refer to like elements in the figures.
Although the present invention has been described in connection with one embodiment, the invention is not intended to be limited to the specific forms set forth herein. On the contrary, it is intended to cover such alternatives, modifications, and equivalents as can be reasonably included within the scope of the invention as defined by the appended claims.
In the following detailed description, numerous specific details such as specific method orders, structures, elements, and connections have been set forth. It is to be understood however that these and other specific details need not be utilized to practice embodiments of the present invention. In other circumstances, well-known structures, elements, or connections have been omitted, or have not been described in particular detail in order to avoid unnecessarily obscuring this description.
References within the specification to “one embodiment” or “an embodiment” are intended to indicate that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the present invention. The appearance of the phrase “in one embodiment” in various places within the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. Moreover, various features are described which may be exhibited by some embodiments and not by others. Similarly, various requirements are described which may be requirements for some embodiments but not other embodiments.
Some portions of the detailed descriptions, which follow, are presented in terms of procedures, steps, logic blocks, processing, and other symbolic representations of operations on data bits within a computer memory. These descriptions and representations are the means used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. A procedure, computer executed step, logic block, process, etc., is here, and generally, conceived to be a self-consistent sequence of steps or instructions leading to a desired result. The steps are those requiring physical manipulations of physical quantities. Usually, though not necessarily, these quantities take the form of electrical or magnetic signals of a computer readable storage medium and are capable of being stored, transferred, combined, compared, and otherwise manipulated in a computer system. It has proven convenient at times, principally for reasons of common usage, to refer to these signals as bits, values, elements, symbols, characters, terms, numbers, or the like.
It should be borne in mind, however, that all of these and similar terms are to be associated with the appropriate physical quantities and are merely convenient labels applied to these quantities. Unless specifically stated otherwise as apparent from the following discussions, it is appreciated that throughout the present invention, discussions utilizing terms such as “receiving” or “generating” or “providing access” or the like, refer to the action and processes of a computer system, or similar electronic computing device that manipulates and transforms data represented as physical (electronic) quantities within the computer system's registers and memories and other computer readable media into other data similarly represented as physical quantities within the computer system memories or registers or other such information storage, transmission or display devices.
Referring to
In one embodiment, sequential latch array design 205 is a gate level design representation of a sequential latch array. In one embodiment, sequential latch array design 205 can be physically implemented to form an actual sequential access latch array 211.
In one embodiment, the physically implemented sequential access latch array 211 can be used in GPU 209. For example, in one embodiment, sequential access latch array 211 can be used as a flow control device that controls flow between data sending and data receiving components that are a part GPU 209. In particular, sequential access latch array 211 can be used as an interface between components that work at different speeds or irregularly. This use can prevent a slower operating component from determining the operating speed of other operating components that are involved in a transfer of data. In addition, sequential access latch array 211 can be used to couple peripheral systems to GPU 209. For example, it can be used so that the processing speed of GPU 209 need not be reduced when it exchanges data with a peripheral.
Sequential Access Latch Array
Referring to
In the
In one embodiment, the master and slave latches of sequential access latch array 211 receive data according to rules. In one embodiment, master latch 251 of first row master-slave pair, 251 and 253, can receive data, in a second half of a clock cycle, if master latch 251 is empty and master-slave latch pairs of other rows are empty or if master latch 251 is empty and master-slave latch pairs of other rows are full.
In one embodiment, master latches 255, 259 and 263 of second, third and fourth row master-slave latch pairs can receive data in a second half of a clock cycle, if they are empty and master-slave latch pairs of the preceding row or rows are full.
Slave latches 253, 257, 261 and 265 of the first, second, third and fourth row master-slave latch pairs, when empty, can receive data in the first half of a clock cycle that is subsequent to a clock cycle in which data is received by the corresponding master latch in the row. Rows of the master-slave latch pairs are filled in order from the first row to the fourth.
In another embodiment, respective master latches 251, 255, 259 and 263 of first, second, third and fourth row master-slave latch pairs can receive data in the second half of first respective clock cycles. The received data is allowed to flow uninhibited into the slave latches of the respective master-slave latch pairs. Then, in the second half of second respective clock cycles the empty master latches 251, 255, 259 and 263 can again receive data. In this embodiment, respective master latches of respective master-slave latch pairs are filled, in a round robin manner, in the second half of first and second respective clock cycles. In one embodiment, because the slave latches are transparent, when data is stored in the slave latches of the respective master-slave latch pairs, the data becomes available for readout.
In one embodiment, slave latches of sequential access latch array 211 can be continuously kept open, wherein the data received by the slave latches flow from master latches of sequential access latch array 211 into slave latches of sequential access latch array 211 uninhibited. In this embodiment, power savings can be realized (since the power that is used to open and close the slave latches is saved).
Referring again to
Operation
Referring to
Referring to
Referring to
Referring to
Referring to
Referring to
Referring to
Referring to
Referring to
At 403, in a first half of a cycle, subsequent to a clock cycle in which data is received by a master latch in a row of the sequential access latch array, data is received into a slave latch of that row (if the slave latch is open). And, at 405, read access is provided to the data that is stored in the slave latch.
Referring to
At 503, data is received into respective master latches in a second half of second respective clock cycles.
At 505, access is provided to data in respective corresponding slave latches.
Table 1 shows an exemplary automatically generated, designer annotated script, that describes a sequential access latch array design generated in accordance with embodiments of the present invention (e.g., sequential latch array design 205 in
In one embodiment, the automatic generation of the structure described in Table 1 can be instantiated via a logic synthesis system or component (e.g., 204 in
With regard to exemplary embodiments thereof, a synchronous sequential latch array generated by an automated system for generating master-slave latch structures is disclosed. A master-slave latch structure includes N/2 rows of master-slave latch pairs, an N/2-to-1 multiplexer and control logic. N is equal to the number of latches that are included in the latch array.
In one embodiment, master and slave latches of the latch array can be separately enabled. In one embodiment, the latch array provides fully-registered flow control. In one embodiment, the latch array is a FIFO buffer.
Although many of the components and processes are described above in the singular for convenience, it will be appreciated by one of skill in the art that multiple components and repeated processes can also be used to practice the techniques of the present invention. Further, while the invention has been particularly shown and described with reference to specific embodiments thereof, it will be understood by those skilled in the art that changes in the form and details of the disclosed embodiments may be made without departing from the spirit or scope of the invention. For example, embodiments of the present invention may be employed with a variety of components and should not be restricted to the ones mentioned above. It is therefore intended that the invention be interpreted to include all variations and equivalents that fall within the true spirit and scope of the present invention.
Number | Name | Date | Kind |
---|---|---|---|
3241122 | Bardell | Mar 1966 | A |
3421026 | Herbert | Jan 1969 | A |
5305258 | Koshizuka | Apr 1994 | A |
5448257 | Margeson, III et al. | Sep 1995 | A |
5600598 | Skjaveland et al. | Feb 1997 | A |
5694355 | Skjaveland et al. | Dec 1997 | A |
5696990 | Rosenthal et al. | Dec 1997 | A |
5835941 | Pawlowski | Nov 1998 | A |
5867443 | Linderman | Feb 1999 | A |
5870329 | Foss | Feb 1999 | A |
5949706 | Chang et al. | Sep 1999 | A |
6041008 | Marr | Mar 2000 | A |
6055590 | Pettey et al. | Apr 2000 | A |
6084856 | Simmons et al. | Jul 2000 | A |
6163500 | Wilford et al. | Dec 2000 | A |
6263331 | Liu et al. | Jul 2001 | B1 |
6263410 | Kao et al. | Jul 2001 | B1 |
6265894 | Reblewski | Jul 2001 | B1 |
6265922 | Kirsch | Jul 2001 | B1 |
6304505 | Forbes et al. | Oct 2001 | B1 |
6366529 | Williams et al. | Apr 2002 | B1 |
6400735 | Percey | Jun 2002 | B1 |
6438024 | Gold et al. | Aug 2002 | B1 |
6501677 | Rau et al. | Dec 2002 | B1 |
6646938 | Kodama | Nov 2003 | B2 |
6803610 | Koolhaas et al. | Oct 2004 | B2 |
6885589 | Cioaca | Apr 2005 | B2 |
6987775 | Haywood | Jan 2006 | B1 |
7051169 | Ganton | May 2006 | B2 |
7106098 | Zack et al. | Sep 2006 | B1 |
7111133 | Ishikawa et al. | Sep 2006 | B2 |
7283404 | Khan et al. | Oct 2007 | B2 |
7333516 | Sikkink et al. | Feb 2008 | B1 |
7346861 | Lee | Mar 2008 | B1 |
7414903 | Noda | Aug 2008 | B2 |
7498850 | Hendrickson | Mar 2009 | B2 |
7499347 | Chen et al. | Mar 2009 | B2 |
7603246 | Newcomb et al. | Oct 2009 | B2 |
7613030 | Iwata et al. | Nov 2009 | B2 |
7630389 | Alfieri et al. | Dec 2009 | B1 |
7724606 | Osawa et al. | May 2010 | B2 |
7739538 | Fee et al. | Jun 2010 | B2 |
7783911 | Chen et al. | Aug 2010 | B2 |
8164934 | Watanabe et al. | Apr 2012 | B2 |
8332957 | Hayasaka | Dec 2012 | B2 |
8363492 | Ishino et al. | Jan 2013 | B2 |
8369177 | Hold et al. | Feb 2013 | B2 |
8453096 | Magee et al. | May 2013 | B2 |
8488360 | Okuda | Jul 2013 | B2 |
8565034 | Lu et al. | Oct 2013 | B1 |
8570818 | Jung et al. | Oct 2013 | B2 |
8760208 | Dike et al. | Jun 2014 | B2 |
8830766 | Sahu | Sep 2014 | B2 |
20010018734 | Lie | Aug 2001 | A1 |
20020089364 | Goldgeisser et al. | Jul 2002 | A1 |
20030025217 | Song | Feb 2003 | A1 |
20030120886 | Moller et al. | Jun 2003 | A1 |
20050128844 | Yamagami | Jun 2005 | A1 |
20060132209 | Meltzer et al. | Jun 2006 | A1 |
20060136656 | Conley et al. | Jun 2006 | A1 |
20070180006 | Gyoten et al. | Aug 2007 | A1 |
20070253263 | Noda | Nov 2007 | A1 |
20070300095 | Fee | Dec 2007 | A1 |
20090119631 | Cortadella et al. | May 2009 | A1 |
20090168499 | Kushida et al. | Jul 2009 | A1 |
20100174877 | Yagihashi | Jul 2010 | A1 |
20100306426 | Boonstra et al. | Dec 2010 | A1 |
20110040906 | Chung et al. | Feb 2011 | A1 |
20110168875 | Okuda | Jul 2011 | A1 |
20120147680 | Koike | Jun 2012 | A1 |
20120163068 | Houston | Jun 2012 | A1 |
20130129083 | Fujino | May 2013 | A1 |
20140003160 | Trivedi et al. | Jan 2014 | A1 |
20140056050 | Yang et al. | Feb 2014 | A1 |
20140129745 | Alfieri | May 2014 | A1 |
20140156891 | Alfieri | Jun 2014 | A1 |
20140184268 | Alfieri et al. | Jul 2014 | A1 |
20140244921 | Alfieri et al. | Aug 2014 | A1 |
20140354330 | Gotterba et al. | Dec 2014 | A1 |
20140355334 | Gotterba et al. | Dec 2014 | A1 |
20140355362 | Wang et al. | Dec 2014 | A1 |
20150016183 | Sinangil et al. | Jan 2015 | A1 |
Entry |
---|
Rouse, Margaret; “Cache Memory”, Sep. 2005, www.whatis.com. |
Sakai et al., “Prototype Implementation of a Highly Parallel Dataflow Machine em-4”, Parallel Processing Symposium, 1991 Proceedings, Fifth International IEEE, 1991. |
“Electronic/Flip Flops” Nov. 29, 2011, Wikibooks, <http://en.wikibooks.org/wiki/Electronics/Flip—Flops>, retrieved Sep. 30, 2014. |
“What is the Difference Between Static RAM and Dynamic RAM?”, Aug. 24, 2000, HowStuffWorks.com, <http:///computer.howstuffworks.com/question452.htm>Retrieved Sept 29 2014. |
Number | Date | Country | |
---|---|---|---|
20140153341 A1 | Jun 2014 | US |