Apparatuses and methods for providing data to a configurable storage area

Information

  • Patent Grant
  • 10860482
  • Patent Number
    10,860,482
  • Date Filed
    Monday, February 11, 2019
    5 years ago
  • Date Issued
    Tuesday, December 8, 2020
    4 years ago
Abstract
Apparatuses and methods for providing data to a configurable storage area are described herein. An example apparatus may include an extended address register including a plurality of configuration bits indicative of an offset and a size, an array having a storage area, a size and offset of the storage area based, at least in part, on the plurality of configuration bits, and a buffer configured to store data, the data including data intended to be stored in the storage area. A memory control unit may be coupled to the buffer and configured to cause the buffer to store the data intended to be stored in the storage area in the storage area of the array responsive, at least in part, to a flush command.
Description
TECHNICAL FIELD

Embodiments of the present invention relate generally to memories, and more specifically, in one or more described embodiments, to providing data from a buffer.


BACKGROUND

Memories may be included in a variety of apparatuses, such as computers or other devices, including but not limited to portable memory devices, solid state drives, personal digital assistants, music players, cameras, phones, wireless devices, displays, chip sets, set top boxes, gaming systems, vehicles, and appliances. There are many different types of memory including random-access memory (RAM), read only memory (ROM), dynamic random access memory (DRAM), synchronous dynamic random access memory (SDRAM), flash memory, and resistance variable memory, among others.


Memories, such as resistance variable memory devices, may be used as non-volatile memory for a wide range of electronic devices. Resistance variable memory devices may include, for example, phase change memory (PCM) or resistive memory, among others.


In some instances, it may be desirable to use PCM to implement random access memory. However, implementation in this manner is not seamless as read operations cannot be performed at a memory location undergoing a write operation and write operations may be relatively slower than write operations of conventional RAM. For example, PCMs may require additional time to manage and complete write operations, such as preparing data to be written to memory, suspending a write operation for a read access request, and monitoring progress of a write operation.


As a result of the limitations of using PCM as RAM replacement, operation of PCMs in memory systems may be restricted and/or require additional mechanisms for operation. For example, write and read operations may not be performed at any time on any memory location, and as a result, status registers most be regularly queried to determine whether a particular write operation has completed before another write operation is performed. Additionally, in some applications write operations are necessarily suspended for read operations and resumed upon completion of the read operation.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a block diagram of an apparatus according to an embodiment of the present invention.



FIG. 2 is a block diagram of a memory according to an embodiment of the present invention.



FIG. 3 is a block diagram of a cache line according to an embodiment of the present invention.



FIG. 4 is a table illustrating example storage area offsets according to an embodiment of the present invention.



FIG. 5 is a table illustrating example storage area sizes according to an embodiment of the present invention.



FIG. 6 is a flowchart of a method for performing a write operation according to an embodiment of the present invention.



FIG. 7 is a flowchart of a method for performing a read operation according to an embodiment of the present invention.





DETAILED DESCRIPTION

Apparatuses and methods for providing data to a configurable storage area are disclosed herein. Certain details are set forth below to provide a sufficient understanding of embodiments of the invention. However, it will be clear to one having skill in the art that embodiments of the invention may be practiced without these particular details. Moreover, the particular embodiments of the present invention described herein are provided by way of example and should not be used to limit the scope of the invention to these particular embodiments. In other instances, well-known circuits, control signals, timing protocols, and software operations have not been shown in detail in order to avoid unnecessarily obscuring the invention.


Embodiments described herein are directed generally to memory and storing data therein. For example, reference is made to storing data and is intended to include both write operations for storing data in volatile memories and programming operations for storing data in non-volatile memories. Moreover, while reference is made herein to write data, write data is intended to include any data provided to a memory for storing in one or more of volatile and non-volatile memory.



FIG. 1 is a block diagram of an apparatus 100 according to as embodiment of the present invention. The apparatus may comprise circuitry, one or more semiconductor dice, a packaged semiconductor, a device including such circuitry, die, or package, and/or a system including such a device. The apparatus 100 may include a host 110 (e.g., memory controller) and a memory 120. The host 110 and the memory 120 may be coupled by a command and address (CA) bus 130 and a data bus 135. The memory 120 may be configured to receive commands and/or addresses from the host 110 over the CA bus 130, and the memory may be configured to receive data and/or provide data over the data bus 135. Although description is made herein with respect to the host 110 and memory 120 providing signals over the CA bus 130 and the data bus 135, in some examples, a single bus may be used to provide signals. For example, in at least one embodiment, the host 110 may be configured to provide commands, addresses, and data to the memory 120 and receive data from the memory 120, over a single, shared bus.


The memory 120 may be configured to perform operations (e.g., read operations, write operations, or flush operations) in response to received commands and/or addresses that are provided by the host 110. For example, the memory 120 may provide data to the host 110 over the data bus 135 responsive to a read command, and may store data received over the data bus 135 responsive to a write command. The memory 120 may internally provide data stored in a buffer to a storage area responsive to a flush command. Additionally, the memory 120 may further provide information to the host 110 over the data bits 135 in response to particular commands. The information may indicate, for instance, whether the memory 120 is available to perform a memory operation and/or an amount of time before the memory 120 may become available to perform a memory operation.


By way of example, the information provided by the memory 120 may include information related to a memory operation performed in response to a command, remaining time of a variable latency period tLAT following receipt of a command, an acknowledgement ACK indicating an end of a variable latency period tLAT, as well as other information that may be related to the memory 120. By way of example, the memory 120 may indicate it is busy performing an operation or may indicate it is available to perform an operation. In some examples, the memory 120 may indicate an amount of time needed before the memory 120 will become available to perform a memory command.


During the tLAT period, the memory 120 may manage memory operations. For example, during the tLAT period the memory 120 may prepare itself for an operation (e.g., by preparing memory circuits for performing an operation), complete a memory operation in progress, suspend an operation in progress to begin performing a new operation, or manage other operations. As such, the duration of the variable latency period tLAT may vary. In some embodiments, the memory 120 may provide information related to the time for the tLAT period. The memory 120 may, for instance, provide a remaining duration for the tLAT period.


In some embodiments, the memory 120 may be configured such that the tLAT period is guaranteed to end within one or more particular amounts of time. For example, in at least one embodiment, when a command comprises a read command, the memory 120 may be configured such that the tLAT period ends whhin a time tMAXLATR. When a command comprises a write command, the memory 130 may be confirmed the tLAT period will end within tMAXLATW. Values for both tMAXLATR and tMAXLATW may be stored as parameters in a register (not shown in FIG. 1), and in some instances, may be predetermined and/or equal or unequal. The values for tMAXLATR and tMAXLATW may be adjusted in real-time, for instance, based on priority of commands. By way of example, tMAXLATW may be increased when read commands are prioritized over write commands.



FIG. 2 is a block diagram of a memory 200 according to an embodiment of the present invention. The memory 200 may be used to implement, at least in part, the memory 120 of FIG. 1. The memory 200 may include a memory control unit 210, a buffer 220, and an array 230. The memory 200 may include elements that have been previously described with respect to the apparatus 100 of FIG. 1. Those elements have been identified in FIG. 2 using the same reference numbers used in FIG. 1 and operation of the common elements is as previously described. Consequently, a detailed description of the operation of these elements will not be repeated in the interest of brevity.


The memory control unit 210 may be coupled to the buffer 220 and the array 230 and may be configured to control operation of the buffer 220 and the array 230. The memory control unit 210 may be configured to receive commands and addresses from the CA bus 130 and the buffer 220 may be configured to receive data from the data bus 135 and provide data to the data bus 135. The memory control unit 210 may be configured to cause data (e.g., write data or read data) to be provided between the buffer 220 and the array 230 over the data bus 240 in response to received commands and addresses. For example, in response to a write command, the memory control unit 210 may cause write data to be provided from the buffer 220 to the array 230. Similarly, in response to a read command, the memory control unit 210 may cause read data to be provided from the array 230 to the buffer 220. Responsive to a flush command, the memory control unit 210 may cause write data to be provided from the buffer 220 to a storage area of the array 230, for example, a storage area 232 configured to provide persistent storage.


The buffer 220 may include a phiialiiy of cache lines 222, each of which may comprise volatile memory (e.g., registers) and may be configured to store data received by the buffer 220 on one or more of the data busses 135, 240. Each of the plurality of cache lines 222 may further be configured to srore an address and/or information associated with the data.


By way of example, with reference to FIG. 3, a cache line 300 according to an embodiment of the invention may be used to implement one or more of the cache lines 222 of FIG. 2. The cache line 300 may include a plurality of fields in which respective addresses, data and/or information associated with the data may be stored. The cache line 300 may, for instance, include an address field 302, a data field 304, a data flag field 306, a usage field 308, and an age field 310. The address field 302 and the data field 304 may include an address (e.g., row address) and data associated with the address respectively. For example, the data may comprise read data received from tire array 230 (e.g., in response to a read command) and the address may comprise the address at which the data is stored in the array 230, or the data may comprise write data received for storing in the array 230 (e.g., in response to a write command) and the address may comprise the address at which the data is to be stored. The address included in the address field 302 may be a memory address or information related to a memory address. The data flag field 306 may include information (e.g., flag bits) indicating whether the data stored is the cache line 300 has been modified since the cache line 300 was associated with a particular address. In some examples, the data flag field 306 may include 1 flag bit for each byte of data stored, and in this manner, the data flag field 306 may indicate which portions of data stored in a data field 304 have been modified. The usage field 308 may include information indicating whether the cache line 300 is in use. For example, a cache line 300 may be in use when the data comprises write data not yet stored in the array 230. In some examples, the usage field 308 may further indicate whether the data comprises write data to be stored in a storage area 232 of the array 230, as described in more detail below. The age field 310 may include information indicating how long the data has been stored in the data field 304 of the cache line 300. Information stored in the age field 310 may indicate, for example, a duration of time, a number of operations (e.g., transparent write operations), a number of clock cycles, or a combination thereof.


Referring once again to FIG. 2, the buffet 220 further include a buffer control unit 224 that may be configured to control operation of the buffet 220. The buffet control unit 224 may be configured to select a cache line 222 for storing data received by the buffer 220, for instance, during a write operation, and/or may be configured to update one or more fields of a cache line 222. The buffer control unit 224 may further be configured to provide data from the buffer 220 during respective operations. During a read operation, for example, the buffer control unit 224 may provide data from a cache line 222 to the data bus 135. As will be explained in more detail below, during a flush command, the buffer control unit 224 may provide data from one or more cache lines 222 to the array 230. In some examples, the buffer control unit 224 may be configured to provide data from the buffer 220 response, at least in part, to one or more control signals provided by the memory control unit 210.


The array 230 may comprise any array known in the art, now or in the future, and may include volatile memory cells, non-volatile memory cells (e.g., PCM memory cells), or a combination thereof.


In some instances, a memory, such as the memory 200 of FIG. 2, may be configured to provide data from one or more cache lines 222 to the array 230 responsive, at least in part, to a command. Accordingly, a host such as the host 110 of FIG. 1, may be configured to provide a flush command to the memory control unit 210 of the memory 200, and in response, the memory control unit 210 may cause the buffer control unit 224 to perform a flush operation. A flush operation may comprise providing write data from one or more of the cache lines 222 to the memory 230.


Additionaly, in at least one embodiment, the array 230 may comprise a storage area 232 configured to provide persistent storage of data. The storage area 232 may include non-volatile memory. The memory 200 may include a register, such as the extended address register 260 of FIG. 2, that may be configured to store a plurality of configuration bits that define the partition (e.g., size and/or offset relative to an address) of the storage area 232 within the array 230.


With reference to FIG. 4, table 400 illustrates example storage area offsets according to an embodiment of the invention. As shown, a set of the configuration bits of the extended address register 260 may be assigned a particular binary value, thereby determining the offset of the storage area 232 within the array 230. In one embodiment, the configuration bits may be configured to indicate an offset relative to the size of the array 230. For example, a binary value of ‘0001’ may indicate an offset of the storage area 232 of one-sixteenth of the size of the array 230. Additionally or alternatively, respective binary values may indicate an offset using a specific row address (e.g., the row address at which the storage area 232 begins or the row address at which the storage area 232 ends).


With reference to FIG. 5, table 500 illustrates example storage area sizes according to an embodiment of the invention. As shown, a set of the configuration hits of the extended address register 260 may be assigned a particular binary value, thereby determining the size of the storage area 232. In one embodiment, the configuration bits may be configured to indicate a size based on the size of the array 230. For example, a binary value of ‘0000’ may indicate that the storage area 232 comprises the entire array 230. In another example, a binary value of ‘0010’ may indicate a size of one-eighth of the size of the array 230. Additionally or alternatively, respective binary values may indicate a size for the storage area 232 using, for instance, in bytes, blocks, and/or row address ranges.


In at least one embodiment, the memory control unit 210 may be configured to update the configuration bits to increase or decrease the size of the storage area 232 and/or adjust the offset of the storage area 232 relative to a particular location in the array 230. The memory control unit 210 may be configured to update the configuration bits, for instance, in response to a command (e.g., resize command) provided by the host 110.


Referring hack to FIG. 2, in some instances, the memory control unit 210 may be configured to identify write data intended for storing in the storage area 232. For example, the memory control unit 210 may identify write data intended for storing in the storage area 232 based, at least in part, on an address associated with the write data (e.g., whether the address comprises an address located in the storage area 232). In another example, the memory control unit 210 may identify write data intended for storing in the storage area 232, for instance, based, at least in part, on a usage flag 308 associated with the write data.


If the memory control unit 210 identifies write data intended for storing in the storage area 232, the memory control unit 210 my set a flush bit of a register, such as the flag status register 250 of FIG. 2, to a first state to indicate that one or more cache lines 222 include data intended to be stored in the storage area 232 of the array 230.


In response to a flush command, provided, for instance by the host 110, the memory control unit 210 may be configured to cause the buffer control unit 224 to provide write data from all cache lines 222 in use or may be configured to provide write data from fewer than all cache lines 222 in use. For example, write data may be provided from only from cache lines 222 in use storing write data intended to be stored in the storage area 232. In another example, write data may be first provided from cache lines 222 in use storing write data intended to be stored in the storage area 232 and subsequently provided from all other from cache lines 222 in use. Once all write data intended to be stored in the storage area 232 has been provided from the buffer 220 to the storage area 232, for instance, in response to a flush command, the memory control unit 210 may set the flush bit of the flag status register 250 to a second state, thereby indicating that none of the cache lines 222 include write data intended to be stored in the storage area 232 not yet stored in the array 230.


As described, the memory 230 may include a storage area 232 that may comprise non-volatile memory. Thus, by providing a flush command to the memory 230, the host 110 may assure that particular data (e.g., write data associated, with an address located in the storage area 232) is maintained in persistent memory. Other data, such as data not intended to be maintained as persistent, may be stored in other portions of the array 230. In some examples, the other portions of the array 230 may comprise non-volatile memory, volatile memory, or a combination thereof. As described, the memory 200 may be configured to provide information to the host 110 in response to particular commands. In some examples, a host, such as the host 110 of FIG. 1, may be configured to provide a read flag status register (RFSR) command, and in response, the memory control unit 210 may provide the state of the flush bit of the flag status register 250 to the host 110. In at least one embodiment, the memory control unit 210 may cause the buffer control unit 224 to proxide the state of the flush bit to the host 110 using the data bus 135. Based, at least in part, on the state of the flush bit, the host 110 may provide a flush command to the memory control unit 210. For example, after providing a flush command to the memory 200, the host 110 may provide one or more RFSR commands to determine (e.g., periodically determine) the state of the flush bit. In this manner, the host 110 may monitor progress of a flush operation to ensure that data is properly provided to the array 230, as described.


Moreover, the memory 200 may be configured to guarantee that a variable latency period tLAT will end within a particular time tLATMAXR and tLATMAXW for read and write commands, respectively. Accordingly, if the memory 200 is performing a flush operation during the period tLAT, in some instances, the memory 200 may need to interrupt the flush command and prepare itself to perform the respective read command or write command. In at least one embodiment, once the read or write command is performed, the memory 200 may resume the flush command. In other embodiments, the memory 200 may require that the host 110 subsequently provide a new flush command.


As described, one or more of the cache lines 222 may store write data received by the buffer 220 from the data bus 135. Because the write data is first stored in a cache line 222, a usage field of the cache line 222 may include information indicating that the cache line is in use, e.g., storing write data, that has not yet been stored in the array 230. In some instances, it may be desirable to delay providing write data from a cache line 222 to the array 230 such that other operations may be performed before the write data is provided. However, as a result, during operation of the buffer 220, the number of cache lines 222 not in use may decrease. Accordingly, the buffer control unit 224 may be configured to perform a transfer operation when a threshold has been met. The threshold may comprise, for instance, a particular number of cache lines 222 being in use. By way of example, in performing a transfer operation, the buffer control unit 224 may provide data from a first number of cache lines 222 (e.g., M cache lines 222) to the array 230 in response to a second number of cache lines (e.g., N cache lines 222) being in use. In some examples, M and N may comprise a same number of cache lines 222 or may comprise a different number of cache lines 222. Once the transfer operation is complete, those cache lines 222 having data provided to the array 230 during the transfer operation may be considered to not be in use, and the buffer control unit 224 may update the information in a usage field (e.g., usage field 308) of these cache tines 222 accordingly. Subsequently, those cache lines 222 may be reused, such as to store read data and/or store write data in accordance with embodiments described herein.


In some examples, the sequence in which write data stored in cache lines 222 is provided during a transfer operation may be based, at least in part, on information in one or more fields of the cache lines 222. In at least one embodiment, the buffer control unit 224 may be configured to prioritize write data of respective cache lines 222 based, at least in part on the age field 310. The buffer control unit 224 may, for instance, be configured to first provide data from a cache line 222 having stored write data that has been stored in the buffer 220 for the longest duration as indicated by information in respective age fields 310.


In some examples, the memory control unit 210 and/or the buffer control unit 224 may be configured to prioritize internal memory operations, such as transfer operations. Accordingly, based on pending Internal memory operations, the time at which a transfer operation is performed may be adjusted (e.g., hastened or delayed) based, at least in part, on the respective priority of the transfer operation.



FIG. 6 is a flowchart, of a method 600 for performing a write operation according to an embodiment of the present invention. The method may be implemented using, for instance, the memory 200 of FIG. 2. At step 605, the memory control, unit 210 may receive a write command and an address on the CA bus 130, for instance, from the host 110. The memory control unit 210 may provide the address to the buffer 220, and in particular, to the buffer control unit 224. At step 610, the buffer control unit 224 may determine whether a cache line 222 of the buffer 220 is associated with the address. For example, the buffer control unit 224 may determine whether an address stored in an address field 302 of a cache line 222 matches the address received from the memory control unit 210. If a cache line 222 is associated with the address, the buffer control unit 224 may select the associated cache line 222 at step 615.


If no cache line 222 is associated with the address as step 620 the buffer control unit 224 may select a new cache line 222. The buffer control unit 224 may select, for instance, any cache line 222 not in use, and further may store the address reveled from the memory control unit 210 in the address field 302 of the selected cache line 222. At step 625, the memory control unit 210 may cause data stored at the address of the array 230 to be provided to the buffer 220, and the buffer 220 may store the data in the selected cache line 222.


At step 630, the buffer 220 may receive write data from the data bus 135. At step 635, the write data may be stored in the selected cache line 222. The buffer control unit 224 may also update fields of the selected cache line 222. For example, the buffer control unit 224 may update the data flag field 306 to indicate which, if any, of the data stored in the selected cache line 222 has been modified as a result of storing the write data. The buffer control unit 224 further may update the usage field 308 to indicate that the cache line 222 is in use and includes write data not yet stored in the array 230.


As described, the buffer control unit 224 may be configured to provide data from one or more cache lines 222 responsive, at least in part, to a threshold being met. Accordingly, at step 640, the buffer control unit 224 may determine the number of cache lines 222 in use. If the number does not exceed a threshold, the write operation ends. If the number satisfies the threshold, the buffer control unit 224 may perform a transfer operation. As described, this may cause write data not yet stored in the array 230 to be provided from one or more cache lines 222 to the array 230. For each cache line 222 having respective write data provided to the array 230 in this manner, the buffer control unit 224 further may update the usage field 308 to indicate that each of these cache lines 222 is no longer in use.



FIG. 7 is a flowchart of a method 700 for performing a read operation according to an embodiment of the present invention. The method may be implemented, for instance, using the memory 200 of FIG. 2. At step 705, the memory control unit 210 may receive a read command and an address on the CA bus 130. The memory control unit 210 may provide the address to the buffer 220, and in particular, to the buffer control unit 224. At step 710, the buffer control unit 224 may determine whether a cache line 222 of the buffer 220 is associated with the address. For example, the buffer control unit 224 may determine whether an address stored in an address field 302 of a cache line 222 matches the address received from the memory control unit 210. If a cache line 222 is associated with the address, the buffer control unit 224 may select the associated cache line 222 at step 715.


If no cache line 222 is associated with the address, at step 720 the buffer control unit 224 may select a new cache line 222. The buffer control unit 224 may select, for instance, any cache line 222 not in use and may store the address received from the memory control unit 210 in the address field 302 of the selected cache line 222. At step 725, the memory control unit 210 may cause data stored at the address of the array 230 to be provided to the buffer 220, and the buffer 220 may store the data in the selected cache line 222. At step 730, the data stored in the selected cache line 122 may be provided to the data bus 135.


From the foregoing it will be appreciated that, although specific embodiment of the invention have been described herein for purposes of illustration, various modifications may be made without deviating from the spirit and scope of the invention. Accordingly, the invention is not limited except as by the appended claims.

Claims
  • 1. An apparatus, comprising: an array having a storage area;a register configured to store at least one bit;a buffer configured to store data; anda memory control circuit configured to control the storage area of the array to store the data stored in the buffer based on the at least one bit stored in the register in response to a flush command, the memory control circuit further configured to,responsive to a latency period for a read command or a write command, interrupt the flush command.
  • 2. The apparatus of claim 1, wherein the at least one bit defines a partition of the array as the storage area.
  • 3. The apparatus of claim 1, wherein the at least one bit indicates an offset for the storage area relative to a size of the array.
  • 4. The apparatus of claim 1, wherein the at least one bit defines a partition of the array as the storage area by a size and/or offset of the storage area relative to an address of the storage area.
  • 5. The apparatus of claim 1, wherein the memory control circuit is further configured to receive a command among a plurality of commands, wherein the received command is a flush command, the plurality of commands include at least one other command, andwherein the memory control circuit is further configured to interrupt the flush command to prepare for the at least one other command.
  • 6. The apparatus of claim 1, wherein the array comprises non-volatile memory.
  • 7. The apparatus of claim 1, wherein the memory control circuit is further configured to access one of a random access memory and a read only memory.
  • 8. An apparatus, comprising: an array;a register configured to store at least one bit; anda control circuit configured to: control a partition of the array to receive data and store the received data,control the partition of the array to store the received data responsive to a flush command, andresponsive to a latency period for a read command or a write command, interrupt the flush command.
  • 9. The apparatus of claim 8, further comprising a buffer configured to store and provide the data to the partition of the array, wherein the partition of the array is defined based on the at least one bit stored in the register.
  • 10. The apparatus of claim 9, wherein the at least one bit defines the partition of the array as a storage area by a size and/or offset of the storage area relative to an address of the storage area.
  • 11. The apparatus of claim 8, wherein the array comprises non-volatile memory.
  • 12. The apparatus of claim 8, wherein the control circuit is further configured to access one of a random access memory and a read only memory.
  • 13. A wireless device, comprising: a storage area;a register configured to store at least one bit;a buffer configured to store data; anda control circuit configured to control the storage area to receive and store the data based on the at least one bit and in response to a flush command, and further configured to interrupt the flush command responsive to a latency period for a read command or a write command.
  • 14. The wireless device of claim 13, wherein the buffer includes a plurality of cache lines configured to store the data to be provided to, and stored in, the storage area.
  • 15. The wireless device of claim 13, further comprising an array having the storage area, wherein the at least one bit indicates an offset for the storage area relative to a size of the array.
  • 16. The wireless device of claim 13, wherein the control circuit is further configured to access one of a random access memory and a read only memory.
CROSS-REFERENCED TO RELATED APPLICATION(S)

This application is a continuation of U.S. patent application Ser. No. 15/857,435 filed Dec. 28, 2017, issued as U.S. Pat. No. 10,223,263 on March 5. 2019, which is a continuation of U.S. patent application Ser. No. 15/637,961, filed Jun. 29, 2017, issued as U.S. Pat. No. 9,928,171 on Mar. 27, 2018, which is a continuation of U.S. patent application Ser. No. 13/967,206, filed Aug. 14, 2013, issued as U.S. Pat. No. 9,727,493 on Aug. 8, 2017. These applications and patents are incorporated by reference herein, in their entirety, and for all purposes.

US Referenced Citations (223)
Number Name Date Kind
4833642 Ooi May 1989 A
5287469 Tsuboi Feb 1994 A
5369616 Wells et al. Nov 1994 A
5546344 Fawcett Aug 1996 A
5668972 Liu et al. Sep 1997 A
5778432 Rubin et al. Jul 1998 A
5822244 Hansen et al. Oct 1998 A
5892961 Trimberger Apr 1999 A
5893149 Hagersten et al. Apr 1999 A
5953522 Fox et al. Sep 1999 A
6026034 Suzuki et al. Feb 2000 A
6201739 Brown et al. Mar 2001 B1
6219384 Kliza et al. Apr 2001 B1
6260128 Ohshima et al. Jul 2001 B1
6275077 Tobin et al. Aug 2001 B1
6412045 DeKoning et al. Jun 2002 B1
6438055 Taguchi et al. Aug 2002 B1
6459313 Godbee et al. Oct 2002 B1
6460114 Jeddeloh Oct 2002 B1
6510095 Matsuzaki et al. Jan 2003 B1
6546462 Palanca et al. Apr 2003 B1
6552956 Shibata Apr 2003 B2
6629224 Suzuki et al. Sep 2003 B1
6687185 Keeth et al. Feb 2004 B1
6710726 Kim et al. Mar 2004 B2
6744285 Mangum et al. Jun 2004 B2
6861901 Prexl et al. Mar 2005 B2
6879540 Maruyama et al. Apr 2005 B2
6973008 Krause Dec 2005 B2
6980479 Park Dec 2005 B2
6988218 Drexler Jan 2006 B2
6996016 Oh Feb 2006 B2
7042799 Cho May 2006 B2
7058799 Johnson Jun 2006 B2
7111185 Gomm et al. Sep 2006 B2
7139673 Vashi et al. Nov 2006 B1
7170819 Szczypinski Jan 2007 B2
7173863 Conley et al. Feb 2007 B2
7187599 Schnell et al. Mar 2007 B2
7209396 Schnell Apr 2007 B2
7219251 Adkisson May 2007 B2
7268605 Fang et al. Sep 2007 B2
7269754 Ramaswamy et al. Sep 2007 B2
7280430 Lee Oct 2007 B2
7340632 Park Mar 2008 B2
7363540 Patel et al. Apr 2008 B2
7375560 Lee et al. May 2008 B2
7411852 Nishioka et al. Aug 2008 B2
7443216 Gomm et al. Oct 2008 B2
7451338 Lemos Nov 2008 B2
7463534 Ku et al. Dec 2008 B2
7509517 Matsumoto et al. Mar 2009 B2
7579683 Falik et al. Aug 2009 B1
7590013 Yu et al. Sep 2009 B2
7593273 Chu et al. Sep 2009 B2
7609584 Kim et al. Oct 2009 B2
7616040 Motomura Nov 2009 B2
7624209 Ansari et al. Nov 2009 B1
7631248 Zakharchenko et al. Dec 2009 B2
7643334 Lee et al. Jan 2010 B1
7656745 Kwak Feb 2010 B2
7660187 Johnson et al. Feb 2010 B2
7663946 Kim Feb 2010 B2
7671648 Kwak Mar 2010 B2
7715260 Kuo et al. May 2010 B1
7716510 Kwak May 2010 B2
7751261 Cho Jul 2010 B2
7773435 Cho Aug 2010 B2
7802039 Morita Sep 2010 B2
7826305 Fujisawa Nov 2010 B2
7826583 Jeong et al. Nov 2010 B2
7865658 Lasser et al. Jan 2011 B2
7872924 Ma Jan 2011 B2
7885365 Hagleitner et al. Feb 2011 B2
7913103 Gold et al. Mar 2011 B2
7945800 Gomm et al. May 2011 B2
7948817 Coteus et al. May 2011 B2
7969813 Bringivijayaraghavan et al. Jun 2011 B2
7983094 Roge et al. Jul 2011 B1
8004884 Franceschini et al. Aug 2011 B2
8030981 Kim Oct 2011 B2
8195891 Trika Jun 2012 B2
8358546 Kim et al. Jan 2013 B2
8441888 Bringivijayaraghavan et al. May 2013 B2
8489820 Ellard Jul 2013 B1
8509011 Bringivijayaraghavan Aug 2013 B2
8539129 Wang Sep 2013 B2
8582356 Abdulla Nov 2013 B2
8595441 Nakibly et al. Nov 2013 B1
8902684 Kwean Dec 2014 B2
9042156 Nakura et al. May 2015 B2
9268651 Salyers Feb 2016 B1
9563565 Bernardi et al. Feb 2017 B2
9710192 Mirichigni et al. Jul 2017 B2
9727493 Mirichigni et al. Aug 2017 B2
9734097 Mirichigni et al. Aug 2017 B2
9740485 Mirichigni et al. Aug 2017 B2
9754648 Mirichigni et al. Sep 2017 B2
9928171 Mirichigni et al. Mar 2018 B2
10067764 Mirichigni et al. Sep 2018 B2
10067890 Mirichigni et al. Sep 2018 B2
10163472 Mirichigni et al. Dec 2018 B2
10223263 Mirichigni et al. Mar 2019 B2
20010015924 Arimoto et al. Aug 2001 A1
20020006074 Roohparvar Jan 2002 A1
20020168316 Shibata Nov 2002 A1
20030117864 Hampel et al. Jun 2003 A1
20030212865 Hicken et al. Nov 2003 A1
20040098549 Dorst May 2004 A1
20050024107 Takai et al. Feb 2005 A1
20050068812 Oh Mar 2005 A1
20060028883 Shimoyama Feb 2006 A1
20060069812 Osborne Mar 2006 A1
20060155948 Ruckerbauer Jul 2006 A1
20060268619 Chen Nov 2006 A1
20060282717 Suda Dec 2006 A1
20070005829 Fujimoto Jan 2007 A1
20070005922 Swaminathan et al. Jan 2007 A1
20070033427 Correale et al. Feb 2007 A1
20070088903 Choi Apr 2007 A1
20070145151 Nakamura et al. Jun 2007 A1
20070150688 Osborne et al. Jun 2007 A1
20070156991 Gould et al. Jul 2007 A1
20070192651 Schoch Aug 2007 A1
20070217253 Kim et al. Sep 2007 A1
20080080267 Lee Apr 2008 A1
20080080271 Kim Apr 2008 A1
20080082707 Gupta et al. Apr 2008 A1
20080082769 Bouchou Apr 2008 A1
20080117682 Byeon May 2008 A1
20080155301 Ahvenainen et al. Jun 2008 A1
20080209109 Lasser Aug 2008 A1
20080270811 Chow et al. Oct 2008 A1
20080304351 Lim et al. Dec 2008 A1
20080307165 Kazuhiro Dec 2008 A1
20080320214 Ma et al. Dec 2008 A1
20090024778 Morita Jan 2009 A1
20090044190 Tringali et al. Feb 2009 A1
20090157974 Lasser Jun 2009 A1
20090232250 Yamada et al. Sep 2009 A1
20090235025 Kondo Sep 2009 A1
20090265514 Biles et al. Oct 2009 A1
20090315600 Becker et al. Dec 2009 A1
20100001762 Kim Jan 2010 A1
20100013675 Bennett Jan 2010 A1
20100023675 Chen et al. Jan 2010 A1
20100124090 Arai May 2010 A1
20100124102 Lee et al. May 2010 A1
20100161880 You Jun 2010 A1
20100161932 Stern et al. Jun 2010 A1
20100165769 Kuroki Jul 2010 A1
20100165780 Bains et al. Jul 2010 A1
20100172179 Gorobets et al. Jul 2010 A1
20100174955 Carnevale et al. Jul 2010 A1
20100195429 Sonoda Aug 2010 A1
20100254198 Bringivijayaraghavan et al. Oct 2010 A1
20100269115 Arimilli et al. Oct 2010 A1
20100293343 Perego Nov 2010 A1
20110035536 Shim et al. Feb 2011 A1
20110047319 Jeon et al. Feb 2011 A1
20110055453 Bennett Mar 2011 A1
20110055671 Kim et al. Mar 2011 A1
20110072190 Borracini et al. Mar 2011 A1
20110153900 Zitlaw Jun 2011 A1
20110170354 DeCaro et al. Jul 2011 A1
20110197014 Yeh Aug 2011 A1
20110228625 Bringivijayaraghavan et al. Sep 2011 A1
20110238866 Zitlaw Sep 2011 A1
20110238941 Xu et al. Sep 2011 A1
20110289255 Wang et al. Nov 2011 A1
20110314324 Ozdemir Dec 2011 A1
20120017053 Yang et al. Jan 2012 A1
20120054419 Chen et al. Mar 2012 A1
20120066434 Zitlaw Mar 2012 A1
20120117303 Carannante et al. May 2012 A1
20120124317 Mirichigni et al. May 2012 A1
20120131263 Yeh May 2012 A1
20120131304 Franceschini et al. May 2012 A1
20120173809 Ko et al. Jul 2012 A1
20120179860 Falanga et al. Jul 2012 A1
20120216085 Weingarten et al. Aug 2012 A1
20120254873 Bringivijayaraghavan Oct 2012 A1
20120265922 Post et al. Oct 2012 A1
20120269015 Bringivijayaraghavan Oct 2012 A1
20120324141 Seong et al. Dec 2012 A1
20120331207 Lassa et al. Dec 2012 A1
20130036253 Baltar Feb 2013 A1
20130058175 Lin et al. Mar 2013 A1
20130159766 D'Abreu et al. Jun 2013 A1
20130191601 Peterson Jul 2013 A1
20130191704 Eleftheriou et al. Jul 2013 A1
20130194013 Kwak Aug 2013 A1
20130250701 Bringivijayaraghavan et al. Sep 2013 A1
20130262746 Srinivasan Oct 2013 A1
20130282967 Ramanujan Oct 2013 A1
20130321052 Huber et al. Dec 2013 A1
20130329503 Bringivijayaraghavan Dec 2013 A1
20130342254 Mazumder et al. Dec 2013 A1
20140010025 Bringivijayaraghavan Jan 2014 A1
20140013028 Rametta Jan 2014 A1
20140035640 Kwak et al. Feb 2014 A1
20140055184 Vankayala Feb 2014 A1
20140082259 Yeh Mar 2014 A1
20140089562 Deguchi et al. Mar 2014 A1
20140122814 Mirichigni et al. May 2014 A1
20140122822 Mirichigni et al. May 2014 A1
20140136771 Cornwell et al. May 2014 A1
20140215129 Kuzmin et al. Jul 2014 A1
20140281182 Mirichigni et al. Sep 2014 A1
20150052288 Mincnigni et al. Feb 2015 A1
20150052299 Mirichigni et al. Feb 2015 A1
20150095551 Confalonieri et al. Apr 2015 A1
20150100744 Mirichigni Apr 2015 A1
20150347038 Monteleone et al. Dec 2015 A1
20170083263 Mirichigni et al. Mar 2017 A1
20170147234 Frans et al. May 2017 A1
20170300413 Mirichigni et al. Oct 2017 A1
20170308382 Mirichigni et al. Oct 2017 A1
20170309318 Mirichigni et al. Oct 2017 A1
20170329534 Mirichigni et al. Nov 2017 A1
20180121356 Mirichigni et al. May 2018 A1
20180349302 Mirichigni et al. Dec 2018 A1
20190080733 Mirichigni et al. Mar 2019 A1
Foreign Referenced Citations (14)
Number Date Country
1497604 May 2004 CN
1751360 Mar 2006 CN
101256539 Sep 2008 CN
101752009 Jun 2010 CN
101820450 Sep 2010 CN
1860460 Jan 2011 CN
2005108395 Apr 2005 JP
2009025866 Feb 2009 JP
2011503753 Jan 2011 JP
2013524388 Jun 2013 JP
201303735 Jan 2013 TW
2011130547 Oct 2011 WO
2014149831 Sep 2014 WO
2015183443 Dec 2015 WO
Non-Patent Literature Citations (13)
Entry
U.S. Appl. No. 15/637,961, entitled “Apparatuses and Methods for Providing Data to a Configurable Storage Area”, filed Jun. 29, 2017, pp. all.
U.S. Appl. No. 15/643,361, entitled “Apparatuses and Methods for Memory Operations Having Variable Latencies”, filed Jul. 6, 2017; pp. all.
U.S. Appl. No. 15/667,358, entitled “Apparatuses and Methods for Variable Latency Memory Operations”, filed Aug. 2, 2017; pp. all.
U.S. Appl. No. 15/857,435, entitled “Apparatuses and Methods for Providing Data to a Configurable Storage Area”, filed Dec. 28, 2017; pp. all.
U.S. Appl. No. 16/105,846 titled “Apparatuses and Methods for Memory Operations Having Variable Latencies” filed Aug. 20, 2018; pp. all.
U.S. Appl. No. 16,058,793 titled “Apparatuses and Methods for Variable Latency Memory Operations” filed Aug. 8, 2018; pp. all.
Definition from Wiktionary, the free dictionary—Archived revision by Tuvalkin (talk | contribs) as of 21:49, Jan. 5, 2012. https://en.wiktionary.org/w/index/php?title=available&oldid=15765189 Downloaded Jan. 13, 2016.
U.S. Appl. No. 16/898,865 titled “Apparatuses and Methods for Memory Operations Having Variable Latencies” filed Nov. 13, 2018, pp. all.
“Interfacing DDR SDRAM with Stratix II Devices”, Version 3.2, Altera Corp, Sep. 2008, pp. all.
“JEDEC—Low Power Double Data Rate 2 (LPDDR2)”, Feb. 2010, pp. all.
Handy, J. “The Cache Memory Book Second Edition”, Academic Press Prof., Inc., San Diego, CA, Jan. 1993, p. 67.
Lee “How to Implement DDR SGRAM in Graphic System” Samsung Electric, 4Q1998 (Oct.-Dec. 1998).
Qureshi, et al., “Enhancing Lifetime and Security of PCM-Based Main Memory with Start-Gap Wearing Leveling”, Microarchitecture, 2009 42nd Annual IEEE/ACM International Symposium, Dec. 12-16, 2009, pp. 14-23.
Related Publications (1)
Number Date Country
20190171567 A1 Jun 2019 US
Continuations (3)
Number Date Country
Parent 15857435 Dec 2017 US
Child 16272945 US
Parent 15637961 Jun 2017 US
Child 15857435 US
Parent 13967206 Aug 2013 US
Child 15637961 US