Coherent memory access

Information

  • Patent Grant
  • 11620228
  • Patent Number
    11,620,228
  • Date Filed
    Friday, June 17, 2022
    2 years ago
  • Date Issued
    Tuesday, April 4, 2023
    a year ago
Abstract
Apparatuses and methods related to providing coherent memory access. An apparatus for providing coherent memory access can include a memory array, a first processing resource, a first cache line and a second cache line coupled to the memory array, a first cache controller, and a second cache controller. The first cache controller coupled to the first processing resource and to the first cache line can be configured to provide coherent access to data stored in the second cache line and corresponding to a memory address. A second cache controller coupled through an interface to a second processing resource external to the apparatus and coupled to the second cache line can be configured to provide coherent access to the data stored in the first cache line and corresponding to the memory address. Coherent access can be provided using a first cache line address register of the first cache controller which stores the memory address and a second cache line address register of the second cache controller which also stores the memory address.
Description
TECHNICAL FIELD

The present disclosure relates generally to memory, and more particularly, to apparatuses and methods associated with coherent memory access.


BACKGROUND

Memory devices are typically provided as internal, semiconductor, integrated circuits in computers or other electronic devices. There are many different types of memory including volatile and non-volatile memory. Volatile memory can require power to maintain its data and includes random-access memory (RAM), dynamic random access memory (DRAM), and synchronous dynamic random access memory (SDRAM), among others. Non-volatile memory can provide persistent data by retaining stored data when not powered and can include NAND flash memory, NOR flash memory, read only memory (ROM), Electrically Erasable Programmable ROM (EEPROM), Erasable Programmable ROM (EPROM), and resistance variable memory such as phase change random access memory (PCRAM), resistive random access memory (RRAM), and magnetoresistive random access memory (MRAM), among others.


Memory is also utilized as volatile and non-volatile data storage for a wide range of electronic applications. Non-volatile memory may be used in, for example, personal computers, portable memory sticks, digital cameras, cellular telephones, portable music players such as MP3 players, movie players, and other electronic devices. Memory cells can be arranged into arrays, with the arrays being used in memory devices.


Various computing systems include a number of processing resources that are coupled to memory (e.g., a memory system), which is accessed in association with executing a set of instructions (e.g., a program, applications, etc.). The number of processing resources can access data stored in the memory.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a block diagram of an apparatus in the form of a computing system including a memory system and capable of providing cache coherency for memory cache in accordance with a number of embodiments of the present disclosure.



FIG. 2 is a block diagram of a memory bank capable of providing cache coherency for memory cache in accordance with a number of embodiments of the present disclosure.



FIG. 3 is a block diagram of a device of a memory system capable providing cache coherency for memory cache in accordance with a number of embodiments of the present disclosure.





DETAILED DESCRIPTION

The present disclosure includes apparatuses and methods related to synchronizing access to memory cache. An example apparatus can include a first processing resource, a first cache line, and a second cache line coupled to a memory array. The example apparatus can also include a first cache controller coupled to the first processing resource and the first cache line and configured to synchronize access to a first data stored in the first cache line with a second data stored in the second cache line. The example apparatus can further include a second cache controller coupled through an interface to a second processing resource external to the memory system and coupled to the second cache line, wherein the second cache controller is configured to synchronize access to the second data with the first data.


Computing systems can include cache as part of a host and/or a processing resource such as a central processing unit (CPU). Cache stores data to fulfil future requests for the data in a smaller duration of time than if the data is retrieved from memory. In a number of examples, cache can be part of a memory system instead of a host and/or a processing resource such as a CPU.


The cache incorporated into the memory system can be used in accessing data stored in a memory array of the memory system, where the memory array is an array of memory cells. For example, each of a plurality of processing resources that are part of the memory system can be associated with a different cache line and/or cache unit. As used herein, cache can refer to multiple cache units. A cache unit refers to a plurality of cache lines. A cache line (e.g., a cache block) describes data having a particular size that is stored in cache. The cache line size often corresponds with a data transfer size of a memory. A cache line can also refer to a physical portion of the cache (e.g., a portion of a cache unit) configured to store a block of data having the particular cache line size such that a cache line can refer to a physical portion of a cache unit storing data and/or to the stored data itself. The cache incorporated in the memory system can also be used to access data stored in the memory array from different banks of the memory array. The cache incorporated in the memory system can further be used to access data from processing resources external to the one or more banks of the memory system.


The different cache units and/or cache lines may store data having a same address. As such, different cache units and/or cache lines can store outdated data. Outdated data describes data which has been altered in a first cache line but has not been altered in a second cache line. The data stored in the second cache line can be outdated. The outdated data can also be referred to as invalid data while updated data is referred to as valid data. The cache lines storing the data may also be referred to as valid and invalid cache lines. A cache line may be valid if it stores valid data and invalid if it stores invalid data.


The data stored by different cache lines can be accessed by a plurality of processing resources. For example, a first cache line can be accessed by a processing resource internal to a bank of a memory device while a second cache line can be accessed by a processing resource external to the bank of the memory device. To ensure that the processing resources have access to valid data, one or more controllers of the cache lines can be configured to provide coherency (e.g., cache coherency) among the different cache lines of the cache corresponding to an array of a memory device.


As used herein, cache coherency refers to uniformity of data that can be stored in different cache lines (e.g., cache lines designated for access from different processing resources). For example, cache coherency ensures that changes to data corresponding to memory addresses (e.g., of an array) shared by multiple processing resources are propagated throughout a system (e.g., memory system). In this manner, cache coherency ensures that access requests (e.g., reads/writes) to particular addresses within a shared memory address space return valid (e.g., up to date) data despite the fact that invalid copies of the data corresponding to the memory addresses may exist within the memory system (e.g., within cache comprising multiple cache lines).


Coherent access can describe providing access to a coherent cache. For example, coherent access to cache lines can be described as providing valid data to different access commands provided from different processing resources, where the access commands access a same memory address. For instance, providing coherent access to a memory address corresponding to an array can result in providing valid data regardless of whether an access command was processed by a first controller coupled to a first cache line or a second controller coupled to a second cache line.


In the following detailed description of the present disclosure, reference is made to the accompanying drawings that form a part hereof, and in which is shown by way of illustration how a number of embodiments of the disclosure may be practiced. These embodiments are described in sufficient detail to enable those of ordinary skill in the art to practice the embodiments of this disclosure, and it is to be understood that other embodiments may be utilized and that process, electrical, and/or structural changes may be made without departing from the scope of the present disclosure. As used herein, the designator “N” indicates that a number of the particular feature so designated can be included with a number of embodiments of the present disclosure.


As used herein, “a number of” something can refer to one or more of such things. For example, a number of memory devices can refer to one or more of memory devices. A “plurality” of something intends two or more. Additionally, designators such as “N,” as used herein, particularly with respect to reference numerals in the drawings, indicates that a number of the particular feature so designated can be included with a number of embodiments of the present disclosure.


The figures herein follow a numbering convention in which the first digit or digits correspond to the drawing figure number and the remaining digits identify an element or component in the drawing. Similar elements or components between different figures may be identified by the use of similar digits. As will be appreciated, elements shown in the various embodiments herein can be added, exchanged, and/or eliminated so as to provide a number of additional embodiments of the present disclosure. In addition, the proportion and the relative scale of the elements provided in the figures are intended to illustrate various embodiments of the present disclosure and are not to be used in a limiting sense.



FIG. 1 is a block diagram of a computing system 100 including a memory system 116 and capable of providing cache coherency for memory cache 114 in accordance with a number of embodiments of the present disclosure. The computing system 100 includes a host 103 and a memory system 116. The host 103 includes a processing resource 102, a cache controller 104, and cache 106-1 and 106-2. The memory system 116 includes a processing resource 108, a memory array 110, cache controllers 112, and cache 114. As used herein, an “apparatus” can refer to, but is not limited to, a variety of structures or combinations of structures, such as a circuit or circuitry, a die or one or more die, a device or devices, or a system or systems. For example, computing system 100, memory system 116, host 103, controllers 104 and/or 112, and/or the memory array 110 may separately be referred to as an “apparatus.”


The system 100 can comprise multiple memory systems (e.g., the memory system 116) each having a number of devices corresponding thereto. Each of the devices of the memory system 116 can be, for example, a “chip” comprising multiple banks. For example, the memory system 116 can comprise a plurality of modules (e.g., dual in-line memory modules (DIMMs)).


In this example, the computing system 100 includes the host 103 coupled to the controllers 112 (e.g., cache controllers 112) via a bus 105, the controllers 112 being part of the memory system 116. The computing system 100 can be a laptop computer, personal computers, digital camera, digital recording and playback device, mobile telephone, PDA, memory card reader, interface hub, sensor, Internet-of-Things (IoT) enabled device, an on-chip application microprocessor, and/or a microcontroller, among other systems, and the processing resource 102 can be a number of processing resources (e.g., one or more processors) capable of accessing the memory 110 (e.g., via the controllers 112). The host 103 may be responsible for execution of an operating system (OS) and/or various applications that can be loaded thereto (e.g., from memory system 116 via the controllers 104 and 112).


The controller 104 may receive memory access requests (e.g., in the form of read and write commands, which may be referred to as load and store commands, respectively) from the processing resource 102. The controller 104 can determine if the memory access requests can be fulfilled using the cache 106-1 and 106-2, referred to as cache 106. If the access requests can be fulfilled using the cache 106, then the controller 104 can fulfill the requests using the cache 106. The controller 104 controls the cache 106 as shown by the dotted line surrounding the controller 104 and the cache 106.


The controller 104 can control different levels of the cache 106. For example, the cache 106-1 can be a first level of cache (L1) and the cache 106-2 can be a second level of cache (L2). Although two levels of cache are shown in FIG. 1, other examples, can include more or fewer levels of cache. The controller 104 can be a single controller that manages multiple different cache units. As used herein, a cache unit can include a grouping of cache lines. For example, the cache 106-1 can be a first cache unit and cache 106-2 can be a second cache unit both of which are controlled/managed by controller 104.


The cache 106-1 and the cache 106-2 can have a hierarchy scheme. For example, the cache 106-1 can retrieve data from the cache 106-2. The cache 106-2 can retrieve data from the memory system 116. As such, the cache 106-1 has a higher level than the cache 106-2. The controller 104, which is a single controller, can control the movement of data between the processing resource and the cache 106-1 and the movement of data between the cache 106-1 and the cache 106-2.


The data stored in the cache 106 can be transferred via the bus 105. The bus 105 (e.g., buses) can transfer data utilizing protocols. Such protocol may be custom or proprietary, and may employ a standardized protocol, such as Peripheral Component Interconnect Express (PCIe), Gen-Z, CCIX, or the like.


The bus transfers data between the host 103 and the memory system 116. The bus 105 can include a data bus and/or an address bus, among other buses such as a timing bus. The characteristics of the bus 105 are the characteristics of the data bus and/or the address bus. The bus 105 can transfer data in words referred to as packets and/or blocks which are a segmentation of a stream of data. If the data transferred by the bus 105 includes X bits of data, then the X bits can be grouped into a plurality of packets, each packet comprising Y bits where Y is less than X. The bus 105 can transfer one of the packets at a time.


There can be a timing component to the transfer of data via the bus 105. The controller 104 can synchronize the transfer of data packets with the receipt of the data packets by the cache controllers 112 and/or the processing resource 108. The synchronization can be carried out using timing information. The timing information can be provided using a timing bus (e.g., timing channel). The timing bus can transmit clock pulses to the receiver (e.g., the cache controllers 112 and/or the processing resource 108, among other possible receivers). The receiver can utilize the timing information to determine a timing associated with receiving the data. The data can be provided and received at a constant rate that corresponds to the synchronization and the timing information.


In some examples, cache coherency can be carried out without the use of timing channels. For example, the cache controller 104 and the receiver (e.g., the cache controllers 112 and/or the processing resource 108) can establish synchronization by providing a predefined baud rate. The term baud describes a common measure of the rate of data over a data bus. In some examples, the baud rate provided as bits per unit of time such as a second (bps). In some examples, the bus 105 further comprises combined address, command, and data buses or separate buses for respective address, command, and data signals.


In FIG. 1, the memory system 116 shows a bank comprising an array 110, the cache 114, the cache controllers 112, and the processing resource 108. Although a single bank is shown, the memory system 116 can comprise multiple banks, each bank comprising an array, cache, cache controllers, and a processing resource as shown in FIG. 3. Although a single array 110 and a single processing resource 108 are shown, the memory system 116 can include multiple arrays 110 and/or processing resources 108. The memory system 116 can comprise the array 110 (e.g., banks) of memory cells and can comprise corresponding support circuitry (e.g., address circuitry, I/O circuitry, control circuitry, read/write circuitry, etc.) associated with accessing the array(s) (e.g., to read data from the arrays and write data to the arrays). As an example, the memory banks (e.g., memory devices) can include a number of DRAM devices, SRAM devices, PCRAM devices, RRAM devices, FeRAM, phase-change memory, 3DXP, and/or Flash memory devices. As used herein, a memory device and/or a memory system can be referred to as a memory apparatus. In a number of embodiments, the memory system 116 can serve as main memory for the computing system.


The controllers 112 can comprise control circuitry, in the form of hardware, firmware, or software, or any combination of the three. As an example, the controllers 112 can comprise a state machine, a sequencer, and/or some other type of control circuitry. In a number of embodiments, the controllers 112 may be co-located with the memory system 116 (e.g., in a system-on-chip (SOC) configuration) and/or can be part of the memory system 116. The controllers 112 can be configured to control the access to data stored in the cache 114. Each of the controllers 112 can be coupled to each other and can be coupled to a different cache unit from the cache 114. The controllers 112 can also provide coherency (e.g., cache coherency) among the different cache lines of the cache 114 corresponding to array 110. The controllers 112 can provide cache coherency utilizing a plurality of cache line address registers shared between the cache 114. The cache line address registers can be internal to the cache 114.


The processing resource 108 can also include a state machine, a sequencer, a microprocessor, and/or some other type of control circuitry, which may be implemented in the form of an application specific integrated circuit (ASIC) coupled to a printed circuit board. The processing resource 108 can be a controller such as a bank controller. A bank controller can, in some examples, perform operations on data stored in the array 110 and/or the cache 114. The bank controller can perform read operations and write operations to provide read access and write access to the cache 114 and/or the array 110. The bank controller may also perform logical operations and/or vector operations utilizing the data stored in the cache 114 and/or the array 110. The function of the cache 114 and the cache controllers 112 is further described in FIGS. 2 and 3.



FIG. 2 is a block diagram of a memory bank 220 capable of providing cache coherency for memory cache in accordance with a number of embodiments of the present disclosure. The memory bank 220 can be a bank in a memory system such as memory system 116 in FIG. 1.


The bank 220 includes a processing resource 208, an array 210, cache controllers 212-1 and 212-2 referred to as cache controllers 212, and cache 214-1, 214-2, and 214-3 referred to as cache 214. The processing resource 208, the array 210, and the cache controllers 212 are analogous to the processing resource 108, the array 110, and the cache controllers 112 of FIG. 1, respectively. The cache 214 is analogous to cache 114.


In some examples, the bank 220 can be coupled to a plurality of interfaces. The interfaces can be, for example, an interconnect 222-1 and a stream interconnect 222-2. The interconnect 222-1 and the stream interconnect 222-2 can couple the bank 220 to a plurality of different banks and/or processing resources external to the bank 220. In some examples, the interconnect 222-1 and the stream interconnect 222-2 can be an advanced extensible interface (AXI) interface and an AXI stream interface, respectively.


The interconnect 222-1 and the stream interconnect 222-2 can be different from the bus 105 in FIG. 1. For example, the interconnect 222-1 and/or the stream interconnect 222-2 can transfer streams of data without dividing the data into packets and/or words. Further, the interconnect 222-1 and/or the stream interconnect 222-2 can transfer data without relying on a timing component. For example, the interconnect interface can function independent of timing limitations.


The stream interconnect 222-2 can provide data to the cache controller 212-2. The cache controller 212-2 can store the data in the cache 214-2. The interconnect 222-1 can provide and receive data from the cache controllers 212.


The controllers 212 can provide coherent access to data stored in cache 214 and/or the array 210. For example, the controllers 212 can provide coherent access to access commands provided by processing resources (e.g., processing resource 208) internal to the bank 220 and to access commands provided by processing resources external to the bank 220. The access requests can be provided via a bus 205 analogous to bus 105 in FIG. 1, an interconnect 222-1, and/or a stream interconnect 222-2.


The cache 214-1 can be referred to a first cache unit which can correspond to cache utilized to store data accessed by the processing resource 208. The cache 214-1 can be referred to as a data cache. Each of the cache units can comprise cache lines. For example, the first cache unit (e.g., cache 214-1) comprises four cache lines in FIG. 1. The processing resource 208 can issue read operations and/or write operations to access the data stored in the cache lines of the cache 214-1. The cache 214-1 is controlled by the cache controller 212-1.


The cache 214-2 is a second cache unit which can be referred to as a bank-to-bank cache. Processing resource external to the bank 220 can utilize the cache 214-2 to store data. Processing resources external to the bank 220 can include, for example, processing resources of different banks or processing resources that are external to any bank but that are internal to the memory system of which the bank 220 is a part of. The cache 214-2 can comprise eight cache lines. However, the number of cache lines provided for each of the cache 214 is illustrative and should not be construed as limiting given that each of the cache 214 can include a different number of cache lines than those described herein. The cache 214-2 is controlled by the cache controller 212-2


The cache 214-3 is a third cache unit which can be referred to as a double data rate (DDR) cache. Processing resources external to the memory system 116 of FIG. 1 can utilize the cache 214-2 to store data. The processing resources external to the memory system 116 can include, for example, processing resources of a host such as a CPU, graphical processing unit (GPU), and/or different types of processing resources external the memory system. The cache 214-3 can be coupled to the processing resource external to the memory system via the bus 205. The bus 205 is analogous to the bus 105 in FIG. 1. In some examples, the cache 214-3 can be controlled by one of the controllers 212.


In some examples, a bank 220 can include more or less cache units than those shown in FIG. 2. For instance, a bank 220 can include three cache units or more than three cache units. Access requests can provide access to coherent cache units. For example, valid data stored in one of a plurality of different cache units can be copied to the other cache units of the plurality of cache units to service write commands and/or read commands (e.g., access commands). A bank can include multiple processing resources each of which communicates with a different cache controller coupled to different cache units. The valid data stored in one of the cache units can be copied to cache units storing invalid data to provide cache coherency. The valid data can be copied from the cache unit storing the valid data to the cache units storing invalid data if both the cache units are associated with a same memory address. The cache units can be associated with a same memory address if a cache line address register of the cache unit storing the valid data and a cache line address register of the cache units storing the invalid data store data having a same memory address.


The cache 214-1, 214-2, and 214-3 can be coupled to each other. For example, the cache 214-1 can be coupled to the cache 214-2. The cache 214-2 can be coupled to the cache 214-3. Data can be moved from any one of the cache 214 to any other of the cache 214. For example, data stored in a cache line of the cache 214-1 can be moved to a cache line of the cache 214-2. The cache 214 can also be coupled to the array 210.


The controllers 212 can store memory addresses corresponding to data stored in their respective cache lines in a plurality of cache line address registers (e.g., cache line address registers 213-1 and 213-2). The memory addresses can be stored in the cache line address registers using translation lookaside buffer (TLB) entries, where the TLB entries include the memory addresses.


Each of the controllers 212 can include a different plurality of cache line address registers 213-1 and 213-2, referred to as cache line address registers 213. For example, the controller 212-1 can comprise the cache line address registers 213-1 to store memory addresses corresponding to data stored in the cache lines of the cache 214-1. The controller 212-2 can comprise the cache line address registers 213-2 to store memory addresses corresponding to data stored in the cache lines of the cache 214-2.


Each of the cache lines of the cache 214 can be associated with a different one of the cache line address registers 213 of the controllers 212. For example, a first cache line address register of the controller 212-1 can store a memory address(es) corresponding to a first cache line of the cache 214-1, a second cache line address register of the controller 212-1 can store memory addresses corresponding to a second cache line of the cache 214-1, etc. In some instances, there can be a one-to-one relationship between the controllers 212 and the cache units corresponding to the cache 214 and between the cache line address registers 213 of each of each of the controllers 212 and the cache lines of each of the cache units.


The relationship between the controllers 212 and the cache 214 is such that the cache 214 does not have a corresponding hierarchy associated therewith. Given that in some examples each controller 212 is associated with a single cache unit of the cache 214, there is no other cache from which to develop a hierarchy. For example, the cache 214-1 does not have a higher or a lower priority and/or level as compared to the cache 214-2 and/or the cache 214-3 because the cache 214-1 has a one-to-one relationship with the controller 212-1. Likewise, the cache 214-2 does not have a higher or a lower priority and/or level as compared to the cache 214-1 and/or the cache 214-3 because the cache 214-2 has a one-to-one relationship with the controller 212-2.


Each of the controllers 212 can be coupled to the other controllers. For example, the controller 212-1 is coupled to the controller 212-2 and the controller 212-2 is coupled to the controller 212-1 such that the controllers 212 can communicate with each other. The controllers 212 can communicate with each other to share memory addresses stored in their respective cache line address registers 213. The controller 212-1 can have access to the cache line address registers 213-2 of the controller 212-2. The controller 212-2 can have access to the cache line address registers 213-1 of the controller 212-1. As such, the controllers 212 can share the cache line address registers 213.


By sharing the memory addresses stored in the cache line address registers 213, the controllers 212 can determine whether the other controllers store data corresponding to a memory address. For example, the controller 212-1 can determine whether a memory address is stored in the cache line address registers 213-2 of the controller 212-2 and the controller 212-2 can determine whether a memory address is stored in the cache line address registers 213-1 of the controller 212-1.


A cache controller from the controllers 212 can receive a read access command requesting data. The read access command can be received from a processing resource. For example, the controller 212-1 can receive a read access command from the processing resource 208. The controller 212-2 can receive the read access command from an application processing resource and/or a bank controller which are external to the bank 220.


A cache controller (e.g., cache controllers 212-1 or 212-2) can determine whether the data is stored in a cache line corresponding to the cache controller. For example, the cache controller 212-1 can determine whether the data that is requested is stored in a first cache line corresponding to the cache 214-1 by determining whether an address of the data is stored in the cache line address register 213-1 of the cache controller 212-1. The cache controller 212-2 can determine whether the data that is requested is stored in a second cache line corresponding to the cache 214-2 by determining whether an address of the data is stored in the cache line address registers 213-2 of the cache controller 212-2.


Responsive to determining that the data is not stored in the cache line corresponding to the controller, the cache controller can determine whether the data is stored in a different cache line corresponding to a different cache controller. For example, the cache controller 212-1 can determine whether the data that is requested is stored in a second cache line of the cache controller 212-2 by determining whether an address of the data is stored in the cache line address registers 213-2 of the cache controller 212-2. The cache controller 212-2 can determine whether the data that is requested is stored in the first cache line of the cache controller 212-1 by determining whether an address of the data is stored in the cache line address registers 213-1 of the cache controller 212-1. The cache controllers 212 can determine whether an address is stored in the cache line address registers 213 because the cache controllers 212 share access to their respective cache line address registers 213.


Responsive to determining that the data is stored in the different cache line, the cache controller can store the data in the cache line and can respond to the read access command by providing the data stored in the cache line. For example, the cache controller 212-1 can store the data in the first cache line and can provide the data stored in the first cache line. The cache controller 212-2 can store the data in the second cache line and can provide the data stored in the second cache line. The data can be provided to the processing resource 208 that requested the data or a processing resource external to the bank 220.


In some instances, the read access command can include a memory address of the data. The memory address can be used to identify the data that is sought. For example, the memory address can identify the data that is to be read.


Determining whether the data is stored in the cache line can further include determining whether a value stored in a cache line address register of the cache controller comprising the cache line is equal to the memory address. For example, the cache controller 212-1 can determine whether the value stored in the cache line address registers 213-1 is equivalent to the memory address. In some instances, the value can be equivalent to the memory address if the value represents the memory address and/or if the memory address can be generated from the value. In some examples, a response to the read access command can include the data stored in the cache line corresponding to the controller that received the read access command.


Determining whether that the data is stored in the different cache line can further comprise determining whether the memory address is stored in a cache line address register corresponding to the different cache controller. Responsive to determining that the memory address is stored in the second cache line address register, the cache controller can lock the different cache line corresponding to the different cache controller. For example, the cache controller 212-1 can lock a cache line corresponding to the cache 214-2 responsive to determining that the cache line is associated with the memory address via a corresponding cache line address register. The cache controller 212-2 can lock a cache line corresponding to the cache 214-1 responsive to determining that the cache line is associate with the memory address. As used herein, a cache line can be locked by preventing the cache line from being written to and/or read. For example, a cache controller 214-1 can lock write access commands from writing to the second cache line coupled to the cache controller 214-2. The cache controller 214-1 can lock read access commands, not provided by the cache controller 214-1, to the second cache line.


In some examples, the cache controller can determine whether the cache line that is to be written to is populated. A cache line can be populated if the cache line actively stores data. The cache line can actively store data if the cache line address register associated with the cache line stores a memory address of the memory array 210 or a memory address of a different memory array of a different bank. Responsive to determining that the cache line is populated, the cache controller can flush the cache line or write the cache line to a memory array 210 coupled to the cache line. The data stored in the cache line can be written to the memory array 210 if the data stored in the cache line is populated and/or if the data stored in the cache line has not been written to the array 210. Flushing data from a cache line can include deleting the data from the cache line. Data can be flushed from a cache line if the data stored in the memory array 210 is the data stored in the cache line. There is no need to write the data stored in the cache line if the data is already stored in the memory array 210.


The controller can transfer the data stored in the different cache line to the cache line by requesting a transfer of the data stored in the second cache line from the abstract state machine (ASM) 207, wherein the transfer is a direct transfer from the second cache line to the first cache line. A direct transfer of data from one cache line to a different cache line can include the transferring of data without storing the data in a different location as part of transferring the data. The ASM 207 can be coupled to the array 210 and to the cache controllers 212. For example, the ASM 207 can be coupled to the cache controller 212-1 and the cache controller 212-2, where the cache controllers 212 are also coupled to each other.


Responsive to transferring the data to the cache line, the cache controller can unlock the different cache line. The different cache line can be unlocked by allowing data to be written to the cache line and/or read from the cache line.


In some example, access to data can also include write access. A first cache controller can receive a write access command having a memory address and a first data to be written. The first cache controller can determine whether the memory address is associated with a first cache line corresponding to the first cache controller. Responsive to determining that the memory address is not associated with the first cache line, the first cache controller can also determine whether the memory address is associated with a second cache line corresponding to a second cache controller. Responsive to determining that the memory address is associated with the second cache line, the first cache controller can copy a second data stored in the second cache line to the first cache line and can write the first data to the first cache line. The first cache controller can be one of the cache controllers 212. The second cache controller can be the other cache controller from the cache controllers 212.


In some examples, responsive to determining that the memory address is associated with the first cache line, the first cache controller can store the first data in the first cache line. Responsive to determining that the memory address is associated with the first cache line, the first cache controller can determine whether the second cache line corresponding to the second cache controller is also associated with the memory address. The second cache controller can also be associated with the memory address if, for example, both the first cache line and the second cache line store data having a same memory address.


In some instances, responsive to determining that the memory address is associated with the second cache controller, the first cache controller can provide an invalidation command to the second cache controller to invalidate the second cache line. An invalidation command can include invalidate the data stored in a cache line. Invalidating data can indicate that the data stored in the cache line is not synchronized with the array 210 and that the data needs to be updated before the data can be used. The data is invalidated due to the writing of the first data to the first cache line having the memory address and due to the second write line also being associated with the memory address.



FIG. 3 is a block diagram of a device 330 of the memory system capable of providing cache coherency for memory cache in accordance with a number of embodiments of the present disclosure. The device 330 can include banks 320-1 and 320-2 referred to as banks 320. The device 330 can also include a processing resource 308-3.


The banks 320 can each include a bank controller, cache controllers, and cache. For example, the bank 320-1 includes the bank controller 308-1, cache controllers 312-1 and 312-2, and cache 314-1 and 314-2. The bank 320-2 includes the bank controller 308-2, cache controllers 312-3 and 312-4, and cache 314-3 and 314-4.


The banks 320 and/or the processing resource 308-3 can be coupled through the interconnect 322-1 and the stream interconnect 322-2. For example, the processing resource 308-3 can provide access commands to the cache controllers 312-2 and/or 312-4 through the stream interconnect 322-2. The cache controllers 312-2 and/or 312-4 can provide data to the processing resource 308-3 through the stream interconnect 322-2.


The processing resource 308-3 can be an application processing resource. The application processing resource provides application processing in contrast with display processing, wireless communications processing, and/or power management processing, among other types of processes. The banks 320 can be coupled to each other through the interconnect 322-1.


Although specific embodiments have been illustrated and described herein, those of ordinary skill in the art will appreciate that an arrangement calculated to achieve the same results can be substituted for the specific embodiments shown. This disclosure is intended to cover adaptations or variations of various embodiments of the present disclosure. It is to be understood that the above description has been made in an illustrative fashion, and not a restrictive one. Combinations of the above embodiments, and other embodiments not specifically described herein will be apparent to those of skill in the art upon reviewing the above description. The scope of the various embodiments of the present disclosure includes other applications in which the above structures and methods are used. Therefore, the scope of various embodiments of the present disclosure should be determined with reference to the appended claims, along with the full range of equivalents to which such claims are entitled.


In the foregoing Detailed Description, various features are grouped together in a single embodiment for the purpose of streamlining the disclosure. This method of disclosure is not to be interpreted as reflecting an intention that the disclosed embodiments of the present disclosure have to use more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive subject matter lies in less than all features of a single disclosed embodiment. Thus, the following claims are hereby incorporated into the Detailed Description, with each claim standing on its own as a separate embodiment.

Claims
  • 1. An apparatus, comprising: a memory array;a first processing resource;a first cache line and a second cache line coupled to the memory array;a first cache controller coupled to the first processing resource and to the first cache line and configured to provide coherent access to data stored in the second cache line and corresponding to a memory address; anda second cache controller coupled to the second cache line and coupled through an interconnect to a second processing resource, the second cache controller configured to provide coherent access to data stored in the first cache line and corresponding to the memory address,wherein coherent access is provided by, responsive to determining that the memory address is stored in a cache line address register of the second cache controller, locking the second cache line prior to copying the data from the second cache line to the first cache line.
  • 2. The apparatus of claim 1, wherein the apparatus is a bank of a memory system.
  • 3. The apparatus of claim 2, wherein the first processing resource is a bank controller.
  • 4. The apparatus of claim 1, wherein the first cache controller and the second cache controller are further configured to provide coherent access to the data responsive to receiving a read access command and a write access command.
  • 5. The apparatus of claim 1, wherein the second cache controller is coupled to the second processing resource through a stream interconnect.
  • 6. The apparatus of claim 5, wherein the second processing resource is an application processing resource.
  • 7. The apparatus of claim 1, wherein the second cache controller is coupled to a plurality of processing resources, external to a bank including the first cache controller and the second cache controller, including the second processing resource through the interconnect.
  • 8. The apparatus of claim 7, wherein the second cache controller is further configured to provide coherent access to the data stored in the second cache line via the interconnect without dividing the data into packets or words.
  • 9. The apparatus of claim 1, wherein the first cache controller and the second cache controller are further configured to provide coherent access to data via a direct coupling between the first cache controller and the second cache controller.
  • 10. A method for providing coherent data, comprising: receiving, at a first cache controller of a memory device, an access command requesting data from a memory address via a stream interconnect;determining whether the memory address is associated with a first cache line corresponding to the first cache controller by referencing a first cache line address register;responsive to determining that the memory address is not associated with the first cache line, determining whether the memory address is associated with a second cache line corresponding to a second cache controller by referencing a second cache line address register;responsive to determining that the memory address is stored in the second cache line address register, locking the second cache line;copying the requested data from the second cache line to the first cache line to provide data coherency between the first cache line and the second cache line; andresponding to the access command by providing the requested data stored in the first cache line via the stream interconnect coupling a processing resource of the memory device to the first cache controller.
  • 11. The method of claim 10, wherein the access command includes the memory address of the requested data.
  • 12. The method of claim 11, wherein determining whether the memory address is associated with the first cache line further comprises determining whether a value stored in the first cache line address register of the first cache controller is equal to the memory address.
  • 13. The method of claim 12, wherein responsive to determining that the value stored in the first cache line address register is equal to the memory address, responding to the access command by providing the requested data stored in the first cache line via the stream interconnect without relying on a timing component.
  • 14. The method of claim 10, wherein responsive to determining that the first cache line is populated, flushing the first cache line or writing the first cache line to a memory array coupled to the first cache line and the second cache line.
  • 15. The method of claim 10, wherein responsive to locking the second cache line, requesting a copying of the requested data stored in the second cache line from an abstract state machine (ASM), wherein the copying is a direct transfer of the requested data from the second cache line to the first cache line.
  • 16. The method of claim 10, wherein responsive to transferring the requested data to the first cache line, unlocking the second cache line.
  • 17. A memory system, comprising: a first bank comprising a first bank controller;a second bank coupled to the first bank and comprising: a second bank controller;a first cache controller, coupled to the first bank controller via an interconnect, and configured to provide data coherency to a first cache line with a second data stored in a second cache line responsive to receiving a first access command from the first bank controller, wherein the first cache line is associated with the first cache controller and a bus coupling a processing resource external to the memory system with first cache line;a second cache controller coupled to the second bank controller and the first cache controller and configured to provide data coherency to the second cache line with a first data stored in the first cache line responsive to receiving a second access command from the second bank controller, wherein the second cache line is associated with the second cache controller.
  • 18. The memory system of claim 17, wherein the interconnect is an interface between the first bank controller and the first cache controller and the second cache controller.
  • 19. The memory system of claim 18, wherein the interconnect interface transfers a stream of data.
  • 20. The memory system of claim 19, wherein the interconnect interface functions independent of timing limitations.
PRIORITY INFORMATION

This application is a Continuation of U.S. application Ser. No. 17/013,313, filed Sep. 4, 2020, which is a Continuation of U.S. application Ser. No. 16/156,654, filed Oct. 10, 2018, which issued as U.S. Pat. No. 10,769,071 on Sep. 8, 2020, the contents of which are included herein by reference.

US Referenced Citations (363)
Number Name Date Kind
4380046 Fung Apr 1983 A
4435792 Bechtolsheim Mar 1984 A
4435793 Ochii Mar 1984 A
4727474 Batcher Feb 1988 A
4843264 Galbraith Jun 1989 A
4958378 Bell Sep 1990 A
4977542 Matsuda et al. Dec 1990 A
5023838 Herbert Jun 1991 A
5034636 Reis et al. Jul 1991 A
5201039 Sakamura Apr 1993 A
5210850 Kelly et al. May 1993 A
5253308 Johnson Oct 1993 A
5276643 Hoffmann et al. Jan 1994 A
5325519 Long et al. Jun 1994 A
5367488 An Nov 1994 A
5379257 Matsumura et al. Jan 1995 A
5386379 Ali-Yahia et al. Jan 1995 A
5398213 Yeon et al. Mar 1995 A
5440482 Davis Aug 1995 A
5446690 Tanaka et al. Aug 1995 A
5473576 Matsui Dec 1995 A
5481500 Reohr et al. Jan 1996 A
5485373 Davis et al. Jan 1996 A
5506811 McLaury Apr 1996 A
5615404 Knoll et al. Mar 1997 A
5638128 Hoogenboom Jun 1997 A
5638317 Tran Jun 1997 A
5654936 Cho Aug 1997 A
5678021 Pawate et al. Oct 1997 A
5724291 Matano Mar 1998 A
5724366 Furutani Mar 1998 A
5751987 Mahant-Shetti et al. May 1998 A
5787458 Miwa Jul 1998 A
5854636 Watanabe et al. Dec 1998 A
5867429 Chen et al. Feb 1999 A
5870504 Nemoto et al. Feb 1999 A
5915084 Wendell Jun 1999 A
5935263 Keeth et al. Aug 1999 A
5986942 Sugibayashi Nov 1999 A
5991209 Chow Nov 1999 A
5991785 Alidina et al. Nov 1999 A
6005799 Rao Dec 1999 A
6009020 Nagata Dec 1999 A
6092186 Betker et al. Jul 2000 A
6122211 Morgan et al. Sep 2000 A
6125071 Kohno et al. Sep 2000 A
6134164 Lattimore et al. Oct 2000 A
6147514 Shiratake Nov 2000 A
6151244 Fujino et al. Nov 2000 A
6157578 Brady Dec 2000 A
6163862 Adams et al. Dec 2000 A
6166942 Vo et al. Dec 2000 A
6172918 Hidaka Jan 2001 B1
6175514 Henderson Jan 2001 B1
6181698 Hariguchi Jan 2001 B1
6208544 Beadle et al. Mar 2001 B1
6226215 Yoon May 2001 B1
6301153 Takeuchi et al. Oct 2001 B1
6301164 Manning et al. Oct 2001 B1
6304477 Naji Oct 2001 B1
6389507 Sherman May 2002 B1
6418498 Martwick Jul 2002 B1
6466499 Blodgett Oct 2002 B1
6490659 McKean et al. Dec 2002 B1
6510098 Taylor Jan 2003 B1
6563754 Lien et al. May 2003 B1
6578058 Nygaard Jun 2003 B1
6731542 Le et al. May 2004 B1
6754746 Leung et al. Jun 2004 B1
6768679 Le et al. Jul 2004 B1
6807614 Chung Oct 2004 B2
6816422 Hamade et al. Nov 2004 B2
6819612 Achter Nov 2004 B1
6894549 Eliason May 2005 B2
6943579 Hazanchuk et al. Sep 2005 B1
6948056 Roth et al. Sep 2005 B1
6950771 Fan et al. Sep 2005 B1
6950898 Merritt et al. Sep 2005 B2
6956770 Khalid et al. Oct 2005 B2
6961272 Schreck Nov 2005 B2
6965648 Smith et al. Nov 2005 B1
6985394 Kim Jan 2006 B2
6987693 Cernea et al. Jan 2006 B2
7020017 Chen et al. Mar 2006 B2
7028170 Saulsbury Apr 2006 B2
7045834 Tran et al. May 2006 B2
7054178 Shiah et al. May 2006 B1
7061817 Raad et al. Jun 2006 B2
7079407 Dimitrelis Jul 2006 B1
7173857 Kato et al. Feb 2007 B2
7187585 Li et al. Mar 2007 B2
7196928 Chen Mar 2007 B2
7260565 Lee et al. Aug 2007 B2
7260672 Garney Aug 2007 B2
7372715 Han May 2008 B2
7400532 Aritome Jul 2008 B2
7406494 Magee Jul 2008 B2
7447720 Beaumont Nov 2008 B2
7454451 Beaumont Nov 2008 B2
7457181 Lee et al. Nov 2008 B2
7535769 Cernea May 2009 B2
7546438 Chung Jun 2009 B2
7562198 Noda et al. Jul 2009 B2
7574466 Beaumont Aug 2009 B2
7602647 Li et al. Oct 2009 B2
7663928 Tsai et al. Feb 2010 B2
7685365 Rajwar et al. Mar 2010 B2
7692466 Ahmadi Apr 2010 B2
7752417 Manczak et al. Jul 2010 B2
7791962 Noda et al. Sep 2010 B2
7796453 Riho et al. Sep 2010 B2
7805587 Van Dyke et al. Sep 2010 B1
7808854 Takase Oct 2010 B2
7827372 Bink et al. Nov 2010 B2
7869273 Lee et al. Jan 2011 B2
7898864 Dong Mar 2011 B2
7924628 Danon et al. Apr 2011 B2
7937535 Ozer et al. May 2011 B2
7957206 Bauser Jun 2011 B2
7979667 Allen et al. Jul 2011 B2
7996749 Ding et al. Aug 2011 B2
8042082 Solomon Oct 2011 B2
8045391 Mokhlesi Oct 2011 B2
8059438 Chang et al. Nov 2011 B2
8095825 Hirotsu et al. Jan 2012 B2
8117462 Snapp et al. Feb 2012 B2
8164942 Gebara et al. Apr 2012 B2
8208328 Hong Jun 2012 B2
8213248 Moon et al. Jul 2012 B2
8223568 Seo Jul 2012 B2
8238173 Akerib et al. Aug 2012 B2
8274841 Shimano et al. Sep 2012 B2
8279683 Klein Oct 2012 B2
8310884 Iwai et al. Nov 2012 B2
8332367 Bhattacherjee et al. Dec 2012 B2
8339824 Cooke Dec 2012 B2
8339883 Yu et al. Dec 2012 B2
8347154 Bahali et al. Jan 2013 B2
8351292 Matano Jan 2013 B2
8356144 Hessel et al. Jan 2013 B2
8417921 Gonion et al. Apr 2013 B2
8462532 Argyres Jun 2013 B1
8484276 Carlson et al. Jul 2013 B2
8495438 Roine Jul 2013 B2
8503250 Demone Aug 2013 B2
8526239 Kim Sep 2013 B2
8533245 Cheung Sep 2013 B1
8555037 Gonion Oct 2013 B2
8599613 Abiko et al. Dec 2013 B2
8605015 Guttag et al. Dec 2013 B2
8625376 Jung et al. Jan 2014 B2
8644101 Jun et al. Feb 2014 B2
8650232 Stortz et al. Feb 2014 B2
8873272 Lee Oct 2014 B2
8964496 Manning Feb 2015 B2
8971124 Manning Mar 2015 B1
9015390 Klein Apr 2015 B2
9047193 Lin et al. Jun 2015 B2
9165023 Moskovich et al. Oct 2015 B2
9659605 Zawodny et al. May 2017 B1
9659610 Hush May 2017 B1
9697876 Tiwari et al. Jul 2017 B1
9761300 Willcock Sep 2017 B1
9997212 Finkbeiner et al. Jun 2018 B1
10068664 Penney et al. Sep 2018 B1
20010007112 Porterfield Jul 2001 A1
20010008492 Higashiho Jul 2001 A1
20010010057 Yamada Jul 2001 A1
20010028584 Nakayama et al. Oct 2001 A1
20010043089 Forbes et al. Nov 2001 A1
20020059355 Peleg et al. May 2002 A1
20030167426 Slobodnik Sep 2003 A1
20030222879 Lin et al. Dec 2003 A1
20040073592 Kim et al. Apr 2004 A1
20040073773 Demjanenko Apr 2004 A1
20040085840 Vali et al. May 2004 A1
20040095826 Perner May 2004 A1
20040154002 Ball et al. Aug 2004 A1
20040205289 Srinivasan Oct 2004 A1
20040240251 Nozawa et al. Dec 2004 A1
20050015557 Wang et al. Jan 2005 A1
20050078514 Scheuerlein et al. Apr 2005 A1
20050097417 Agrawal et al. May 2005 A1
20060026355 Okawa et al. Feb 2006 A1
20060047937 Selvaggi et al. Mar 2006 A1
20060069849 Rudelic Mar 2006 A1
20060146623 Mizuno et al. Jul 2006 A1
20060149804 Luick et al. Jul 2006 A1
20060181917 Kang et al. Aug 2006 A1
20060215432 Wickeraad et al. Sep 2006 A1
20060225072 Lari et al. Oct 2006 A1
20060291282 Liu et al. Dec 2006 A1
20070103986 Chen May 2007 A1
20070171747 Hunter et al. Jul 2007 A1
20070180006 Gyoten et al. Aug 2007 A1
20070180184 Sakashita et al. Aug 2007 A1
20070195602 Fong et al. Aug 2007 A1
20070285131 Sohn Dec 2007 A1
20070285979 Turner Dec 2007 A1
20070291532 Tsuji Dec 2007 A1
20080025073 Arsovski Jan 2008 A1
20080037333 Kim et al. Feb 2008 A1
20080052711 Forin et al. Feb 2008 A1
20080133845 Blumrich et al. Jun 2008 A1
20080137388 Krishnan et al. Jun 2008 A1
20080165601 Matick et al. Jul 2008 A1
20080178053 Gorman et al. Jul 2008 A1
20080215937 Dreibelbis et al. Sep 2008 A1
20090067218 Graber Mar 2009 A1
20090083493 Kinter Mar 2009 A1
20090154238 Lee Jun 2009 A1
20090154273 Borot et al. Jun 2009 A1
20090254697 Akerib Oct 2009 A1
20100067296 Li Mar 2010 A1
20100091582 Vali et al. Apr 2010 A1
20100172190 Lavi et al. Jul 2010 A1
20100210076 Gruber et al. Aug 2010 A1
20100226183 Kim Sep 2010 A1
20100308858 Noda et al. Dec 2010 A1
20100332895 Billing et al. Dec 2010 A1
20110051523 Manabe et al. Mar 2011 A1
20110063919 Chandrasekhar et al. Mar 2011 A1
20110093662 Walker et al. Apr 2011 A1
20110103151 Kim et al. May 2011 A1
20110119467 Cadambi et al. May 2011 A1
20110122695 Li et al. May 2011 A1
20110140741 Zerbe et al. Jun 2011 A1
20110219260 Nobunaga et al. Sep 2011 A1
20110267883 Lee et al. Nov 2011 A1
20110317496 Bunce et al. Dec 2011 A1
20120005397 Lim et al. Jan 2012 A1
20120017039 Margetts Jan 2012 A1
20120023281 Kawasaki et al. Jan 2012 A1
20120120705 Mitsubori et al. May 2012 A1
20120134216 Singh May 2012 A1
20120134225 Chow May 2012 A1
20120134226 Chow May 2012 A1
20120140540 Agam et al. Jun 2012 A1
20120182798 Hosono et al. Jul 2012 A1
20120195146 Jun et al. Aug 2012 A1
20120198310 Tran et al. Aug 2012 A1
20120246380 Akerib et al. Sep 2012 A1
20120265964 Murata et al. Oct 2012 A1
20120281486 Rao et al. Nov 2012 A1
20120303627 Keeton et al. Nov 2012 A1
20130003467 Klein Jan 2013 A1
20130061006 Hein Mar 2013 A1
20130107623 Kavalipurapu et al. May 2013 A1
20130117541 Choquette et al. May 2013 A1
20130124783 Yoon et al. May 2013 A1
20130132702 Patel et al. May 2013 A1
20130138646 Sirer et al. May 2013 A1
20130163362 Kim Jun 2013 A1
20130173888 Hansen et al. Jul 2013 A1
20130205114 Badam et al. Aug 2013 A1
20130219112 Okin et al. Aug 2013 A1
20130227361 Bowers et al. Aug 2013 A1
20130283122 Anholt et al. Oct 2013 A1
20130286705 Grover et al. Oct 2013 A1
20130326154 Haswell Dec 2013 A1
20130332707 Gueron et al. Dec 2013 A1
20140185395 Seo Jul 2014 A1
20140215185 Danielsen Jul 2014 A1
20140250279 Manning Sep 2014 A1
20140344934 Jorgensen Nov 2014 A1
20150029798 Manning Jan 2015 A1
20150042380 Manning Feb 2015 A1
20150063052 Manning Mar 2015 A1
20150078108 Cowles et al. Mar 2015 A1
20150120987 Wheeler Apr 2015 A1
20150134713 Wheeler May 2015 A1
20150270015 Murphy et al. Sep 2015 A1
20150279466 Manning Oct 2015 A1
20150324290 Leidel Nov 2015 A1
20150325272 Murphy Nov 2015 A1
20150356009 Wheeler et al. Dec 2015 A1
20150356022 Leidel et al. Dec 2015 A1
20150357007 Manning et al. Dec 2015 A1
20150357008 Manning et al. Dec 2015 A1
20150357019 Wheeler et al. Dec 2015 A1
20150357020 Manning Dec 2015 A1
20150357021 Hush Dec 2015 A1
20150357022 Hush Dec 2015 A1
20150357023 Hush Dec 2015 A1
20150357024 Hush et al. Dec 2015 A1
20150357047 Tiwari Dec 2015 A1
20160062672 Wheeler Mar 2016 A1
20160062673 Tiwari Mar 2016 A1
20160062692 Finkbeiner et al. Mar 2016 A1
20160062733 Tiwari Mar 2016 A1
20160063284 Tiwari Mar 2016 A1
20160064045 La Fratta Mar 2016 A1
20160064047 Tiwari Mar 2016 A1
20160098208 Willcock Apr 2016 A1
20160098209 Leidel et al. Apr 2016 A1
20160110135 Wheeler et al. Apr 2016 A1
20160125919 Hush May 2016 A1
20160154596 Willcock et al. Jun 2016 A1
20160155482 La Fratta Jun 2016 A1
20160162407 Pierson et al. Jun 2016 A1
20160188250 Wheeler Jun 2016 A1
20160196142 Wheeler et al. Jul 2016 A1
20160196856 Tiwari et al. Jul 2016 A1
20160225422 Tiwari et al. Aug 2016 A1
20160266873 Tiwari et al. Sep 2016 A1
20160266899 Tiwari Sep 2016 A1
20160267951 Tiwari Sep 2016 A1
20160292080 Leidel et al. Oct 2016 A1
20160306584 Zawodny et al. Oct 2016 A1
20160306614 Leidel et al. Oct 2016 A1
20160349322 Shibahara Dec 2016 A1
20160350230 Murphy Dec 2016 A1
20160365129 Willcock Dec 2016 A1
20160371033 La Fratta et al. Dec 2016 A1
20170052906 Lea Feb 2017 A1
20170178701 Willcock et al. Jun 2017 A1
20170192844 Lea et al. Jul 2017 A1
20170228192 Willcock et al. Aug 2017 A1
20170235515 Lea et al. Aug 2017 A1
20170236564 Zawodny et al. Aug 2017 A1
20170242902 Crawford et al. Aug 2017 A1
20170243623 Kirsch et al. Aug 2017 A1
20170262369 Murphy Sep 2017 A1
20170263306 Murphy Sep 2017 A1
20170269865 Willcock et al. Sep 2017 A1
20170269903 Tiwari Sep 2017 A1
20170277433 Willcock Sep 2017 A1
20170277440 Willcock Sep 2017 A1
20170277581 Lea et al. Sep 2017 A1
20170277637 Willcock et al. Sep 2017 A1
20170278559 Hush Sep 2017 A1
20170278584 Rosti Sep 2017 A1
20170285988 Dobelstein Oct 2017 A1
20170293434 Tiwari Oct 2017 A1
20170301379 Hush Oct 2017 A1
20170309314 Zawodny et al. Oct 2017 A1
20170329577 Tiwari Nov 2017 A1
20170336989 Zawodny et al. Nov 2017 A1
20170337126 Zawodny et al. Nov 2017 A1
20170337953 Zawodny et al. Nov 2017 A1
20170352391 Hush Dec 2017 A1
20170371539 Mai et al. Dec 2017 A1
20180012636 Alzheimer et al. Jan 2018 A1
20180024769 Howe et al. Jan 2018 A1
20180024926 Penney et al. Jan 2018 A1
20180025759 Penney et al. Jan 2018 A1
20180025768 Hush Jan 2018 A1
20180032458 Bell Feb 2018 A1
20180033478 Lea et al. Feb 2018 A1
20180039484 La Fratta et al. Feb 2018 A1
20180046405 Hush et al. Feb 2018 A1
20180046461 Tiwari Feb 2018 A1
20180060069 Rosti et al. Mar 2018 A1
20180074754 Crawford Mar 2018 A1
20180075899 Hush Mar 2018 A1
20180088850 Willcock Mar 2018 A1
20180102147 Willcock et al. Apr 2018 A1
20180108397 Venkata et al. Apr 2018 A1
20180130515 Zawodny et al. May 2018 A1
20180136871 Leidel May 2018 A1
20180239531 Lea Aug 2018 A1
20180239712 Lea Aug 2018 A1
20180240510 Hush et al. Aug 2018 A1
Foreign Referenced Citations (15)
Number Date Country
102141905 Aug 2011 CN
0214718 Mar 1987 EP
2026209 Feb 2009 EP
H0831168 Feb 1996 JP
2009259193 Mar 2015 JP
10-0211482 Aug 1999 KR
10-2010-0134235 Dec 2010 KR
10-1155127 Jun 2012 KR
10-2013-0049421 May 2013 KR
0075786 Dec 2000 WO
2001065359 Sep 2001 WO
2010079451 Jul 2010 WO
2013062596 May 2013 WO
2013081588 Jun 2013 WO
2013095592 Jun 2013 WO
Non-Patent Literature Citations (19)
Entry
“4.9.3 MINLOC and MAXLOC”, Jun. 12, 1995, (5pgs.), Message Passing Interface Forum 1.1, retrieved from http://www.mpi-forum.org/docs/mpi-1.1/mpi-11-html/node79.html.
Stojmenovic, “Multiplicative Circulant Networks Topological Properties and Communication Algorithms”, (25 pgs.), Discrete Applied Mathematics 77 (1997) 281-305.
Boyd et al., “On the General Applicability of Instruction-Set Randomization”, Jul.-Sep. 2010, (14 pgs.), vol. 7, Issue 3, IEEE Transactions on Dependable and Secure Computing.
Elliot, et al., “Computational RAM: Implementing Processors in Memory”, Jan.-Mar. 1999, (10 pgs.), vol. 16, Issue 1, IEEE Design and Test of Computers Magazine.
Extended European Search Report from related European Patent Application No. 19871002.2, dated Jun. 13, 2022, 11 pages.
Dybdahl, et al., “Destructive-Read in Embedded DRAM, Impact on Power Consumption,” Apr. 2006, (10 pgs.), vol. 2, Issue 2, Journal of Embedded Computing-Issues in embedded single-chip multicore architectures.
Kogge, et al., “Processing In Memory: Chips to Petaflops,” May 23, 1997, (8 pgs.), retrieved from: http://www.cs.ucf.edu/courses/cda5106/summer02/papers/kogge97PIM.pdf.
Draper, et al., “The Architecture of the DIVA Processing-In-Memory Chip,” Jun. 22-26, 2002, (12 pgs.), ICS '02, retrieved from: http://www.isi.edu/˜draper/papers/ics02.pdf.
Adibi, et al., “Processing-In-Memory Technology for Knowledge Discovery Algorithms,” Jun. 25, 2006, (10 pgs.), Proceeding of the Second International Workshop on Data Management on New Hardware, retrieved from: http://www.cs.cmu.edu/˜damon2006/pdf/adibi06inmemory.pdf.
U.S. Appl. No. 13/449,082, entitled, “Methods and Apparatus for Pattern Matching,” filed Apr. 17, 2012, (37 pgs.).
U.S. Appl. No. 13/743,686, entitled, “Weighted Search and Compare in a Memory Device,” filed Jan. 17, 2013, (25 pgs.).
U.S. Appl. No. 13/774,636, entitled, “Memory as a Programmable Logic Device,” filed Feb. 22, 2013, (30 pgs.).
U.S. Appl. No. 13/774,553, entitled, “Neural Network in a Memory Device,” filed Feb. 22, 2013, (63 pgs.).
U.S. Appl. No. 13/796,189, entitled, “Performing Complex Arithmetic Functions in a Memory Device,” filed Mar. 12, 2013, (23 pgs.).
International Search Report and Written Opinion for PCT Application No. PCT/US2013/043702, dated Sep. 26, 2013, (11 pgs.).
Pagiamtzis, et al., “Content-Addressable Memory (CAM) Circuits and Architectures: A Tutorial and Survey”, Mar. 2006, (16 pgs.), vol. 41, No. 3, IEEE Journal of Solid-State Circuits.
Pagiamtzis, Kostas, “Content-Addressable Memory Introduction”, Jun. 25, 2007, (6 pgs.), retrieved from: http://www.pagiamlzis.com/cam/camintro.
Debnath, Biplob, Bloomflash: Bloom Filter on Flash-Based Storage, 2011 31st Annual Conference on Distributed Computing Systems, Jun. 20-24, 2011, 10 pgs.
Derby, et al., “A High-Performance Embedded DSP Core with Novel SIMD Features”, Apr. 6-10, 2003, (4 pgs), vol. 2, pp. 301-304, 2003 IEEE International Conference on Accoustics, Speech, and Signal Processing.
Related Publications (1)
Number Date Country
20220318148 A1 Oct 2022 US
Continuations (2)
Number Date Country
Parent 17013313 Sep 2020 US
Child 17843387 US
Parent 16156654 Oct 2018 US
Child 17013313 US