The present disclosure relates generally to semiconductor memory and methods, and more particularly, to apparatuses and methods for performing operations by a memory device in a self-refresh state.
Memory devices are typically provided as internal, semiconductor, integrated circuits in various computing systems. There are many different types of memory including volatile and non-volatile memory. Volatile memory can require power to maintain its data (e.g., host data, error data, etc.) and includes random access memory (RAM), dynamic random access memory (DRAM), static random access memory (SRAM), synchronous dynamic random access memory (SDRAM), and thyristor random access memory (TRAM), among others. Non-volatile memory can provide persistent data by retaining stored data when not powered and can include NAND flash memory, NOR flash memory, and resistance variable memory such as phase change random access memory (PCRAM), resistive random access memory (RRAM), and magnetoresistive random access memory (MRAM), such as spin torque transfer random access memory (STT RAM), among others.
Computing systems often include a number of processing resources (e.g., one or more processors), which may retrieve and execute instructions and store the results of the executed instructions to a suitable location. A processor can comprise a number of functional units such as arithmetic logic unit (ALU) circuitry, floating point unit (FPU) circuitry, and a combinatorial logic block, for example, which can be used to execute instructions by performing an operation on data (e.g., one or more operands). As used herein, an operation can be, for example, a Boolean operation, such as AND, OR, NOT, NAND, NOR, and XOR, and/or other operations (e.g., invert, shift, arithmetic, statistics, among many other possible operations). For example, functional unit circuitry may be used to perform the arithmetic operations, such as addition, subtraction, multiplication, and division on operands, via a number of operations.
A number of components in a computing system may be involved in providing instructions to the functional unit circuitry for execution. The instructions may be executed, for instance, by a processing resource such as a controller and/or host processor. Data (e.g., the operands on which the instructions will be executed) may be stored in a memory array that is accessible by the functional unit circuitry. The instructions and/or data may be retrieved from the memory array and sequenced and/or buffered before the functional unit circuitry begins to execute instructions on the data. Furthermore, as different types of operations may be executed in one or multiple clock cycles through the functional unit circuitry, intermediate results of the instructions and/or data may also be sequenced and/or buffered. A sequence to complete an operation in one or more clock cycles may be referred to as an operation cycle. Time consumed to complete an operation cycle costs in terms of processing and computing performance and power consumption, of a computing apparatus and/or system.
In many instances, the processing resources (e.g., processor and associated functional unit circuitry) may be external to the memory array, and data is accessed via a bus between the processing resources and the memory array to execute a set of instructions. Processing performance may be improved in a processing-in-memory device, in which a processor may be implemented internally and near to a memory (e.g., directly on a same chip as the memory array). A processing-in-memory device may save time by reducing and eliminating external communications and may also conserve power.
The present disclosure includes apparatuses and methods for performing operations by a memory device in a self-refresh state. An example includes an array of memory cells and a controller coupled to the array of memory cells. The controller is configured to direct performance of compute operations, e.g., read, write, copy, and/or erase operations, on data stored in the array when the array of memory cells is in a self-refresh state.
In the following detailed description of the present disclosure, reference is made to the accompanying drawings that form a part hereof, and in which is shown by way of illustration how one or more embodiments of the disclosure may be practiced. These embodiments are described in sufficient detail to enable those of ordinary skill in the art to practice the embodiments of this disclosure, and it is to be understood that other embodiments may be utilized and that process, electrical, and structural changes may be made without departing from the scope of the present disclosure.
As used herein, designators such as “X”, “Y”, “N”, “M”, etc., particularly with respect to reference numerals in the drawings, indicate that a number of the particular feature so designated can be included. It is also to be understood that the terminology used herein is for the purpose of describing particular embodiments only, and is not intended to be limiting. As used herein, the singular forms “a”, “an”, and “the” can include both singular and plural referents, unless the context clearly dictates otherwise. In addition, “a number of”, “at least one”, and “one or more” (e.g., a number of memory arrays) can refer to one or more memory arrays, whereas a “plurality of” is intended to refer to more than one of such things. Furthermore, the words “can” and “may” are used throughout this application in a permissive sense (i.e., having the potential to, being able to), not in a mandatory sense (i.e., must). The term “include,” and derivations thereof, means “including, but not limited to”. The terms “coupled” and “coupling” mean to be directly or indirectly connected physically or for access to and movement (transmission) of commands and/or data, as appropriate to the context. The terms “data” and “data values” are used interchangeably herein and can have the same meaning, as appropriate to the context.
The figures herein follow a numbering convention in which the first digit or digits correspond to the figure number and the remaining digits identify an element or component in the figure. Similar elements or components between different figures may be identified by the use of similar digits. For example, 150 may reference element “50” in
System 100 includes a host 110 coupled (e.g., connected) to memory device 120, which includes a memory array 130. Host 110 can be a host system such as a personal laptop computer, a desktop computer, a digital camera, a smart phone, or a memory card reader, among various other types of hosts. Host 110 can include a system motherboard and/or backplane and can include a number of processing resources (e.g., one or more processors, microprocessors, etc.). A more detailed diagram of one example of host 110 interaction with the memory device 120 is described in association with
The system 100 can include separate integrated circuits or both the host 110 and the memory device 120 can be on the same integrated circuit. The system 100 can be, for instance, a server system and/or a high performance computing (HPC) system and/or a portion thereof. Although the example shown in
For clarity, the system 100 has been simplified to focus on features with particular relevance to the present disclosure. The memory array 130 can be a DRAM array, SRAM array, STT RAM array, PCRAM array, TRAM array, RRAM array, NAND flash array, and/or NOR flash array, for instance. The array 130 can comprise memory cells arranged in rows coupled by access lines, which may be referred to herein as word lines and/or select lines, and columns coupled by sense lines, which may be referred to herein as data lines and/or digit lines. Although a single array 130 is shown in
The memory device 120 includes address circuitry 142 to latch address signals provided over a bus 156 through I/O circuitry 144. Bus 156 can serve as a data bus (e.g., an I/O bus) and as an address bus; however, embodiments are not so limited. Status and/or exception information can be provided from the controller 140 on the memory device 120 to host 110 through an interface, e.g., as shown at 141 and described in connection with
Memory refresh involves periodically reading information, e.g. data, from an area of computer memory and rewriting the read data to the same area without modification, e.g., using sensing circuitry 150 as described in connection with
While a refresh cycle is occurring, the memory being refreshed has formerly not been available for compute operations. However, during compute operations commanded by a host, e.g., via a central processing unit (CPU), this “overhead” time may not be large enough to significantly slow down a compute operation. For instance, less than 0.4% of the time for a memory chip, e.g., a memory device or array, may be occupied by refresh cycles. In DRAM memory arrays, for example, the memory cells in each memory device may be divided into banks, e.g., as shown at 121-1, . . . , 121-7 and described in connection with
Refresh circuitry may include a refresh counter, e.g., a counter register shown at 136 and described in connection with
As described herein, selection of a different mode, e.g., from a mode register as shown at 138-1 and described in connection with
To enable faster performance of compute operations on data stored in the array when the array of memory cells is in the self-refresh state, a modulated self-refresh mode may be selected to cause the default frequency to be shortened from 15.6 μs to, for example, 7.8 μs. Performance of a compute operation may correspond to a time point at which data from a row in the memory device is read by sensing circuitry 150, e.g., a sense amplifier 406 of the sensing circuitry as described in connection with
A logical operation is intended to mean a processing-in-memory (PIM) operation performed using one bit vector processing, as described further herein. Such one bit vector processing may be performed with the sensing circuitry 150 including a sense amplifier and a compute component, as shown at 431 and described in connection with
A counter register 136-1, or part of the counter register, may be associated with, e.g., coupled to, the circuitry of a controller, e.g., as shown at 140 and described in connection with
During compute operations commanded by the host 110, signals may be transmitted between the host 110 and a memory device 120 and/or the controller 140 of the memory device 120. In some instances, a microprocessor associated with the host 110 may control refresh of the memory cells in the memory array 130 when they are interesting, e.g., other signals are being transmitted between them, with a timer triggering a periodic interrupt to run a subroutine that performs the refresh. Allowing the microprocessor to enter, for example, an energy-saving “sleep mode” when no operations are being performed involving input and/or output (I/O) of data and/or commands between the host 110 and the memory device 120, however, may stop the refresh process and result in loss of the data in memory.
Hence, memory devices 120, as described herein, may have a counter register 136-1 associated with, e.g., coupled to, the controller 140 and/or a counter register 136-2 associated with, e.g., coupled to, the memory array 120 itself. These internal counter registers may be used to generate refresh cycles when the memory device 120 is in a self-refresh state. The self-refresh state of the memory cells of the memory device 120 may correspond to the sleep mode of the host 110. For example, a counter register 136 may include an on-chip oscillator that internally generates refresh cycles such that a corresponding external counter, e.g., a timer associated with the host microprocessor, may be disconnected, e.g., shut down.
Such a sleep mode of the host, e.g., a CPU of the host, may be a low power state, e.g., mode, for a computing system in which associated memory devices, e.g., DDR SDRAM memory devices, among others, enter a self-refresh state. As described herein, this low power state may be used to perform operations by, e.g., in, memory devices in the self-refresh state. The controller 140 may be configured to direct, e.g., via a counter register 136, the performance of the compute operations and/or logical operations described herein at a rate corresponding to a frequency of performance of a memory refresh cycle for the data stored in the memory cells. The compute operations and/or logical operations may be performed on the data using sensing circuitry 150 coupled to the array of memory cells 130 during performance of a self-refresh operation by the sensing circuitry on the data.
In various embodiments, compute operations and/or logical operations, as described herein, may be performed while the memory device is in the self-refresh state even though the clock rate of the compute and/or logical operations may be reduced by, for example, a factor of 1000 times, e.g., from around 15 ns to around 15 μs. This reduced rate for performing such operations may be acceptable because a functionality that is operated during the self-refresh state may be a functionality that can operate with high latency, as described further herein, and/or that does not involve I/O of data and/or commands between the host 110 and the memory device 120.
The counter registers 136 and/or mode registers 138 described herein may include one or more separate registers, e.g., separate and/or in addition to other array control registers such as DDR registers to a DRAM array. For example, counter registers 136 and/or mode registers 138 may be coupled to an interface (e.g., 141 in
Controller 140 may decode signals provided by address and control (A/C) bus 154 from the host 110. According to various embodiments, the controller 140 can be a reduced instruction set computer (RISC) type controller operating on 32 and/or 64 bit length instructions. These signals can include chip enable signals, read enable signals, write enable signals, and address latch signals, among other signals, that are used to control operations performed on the memory array 130, including data read, data write, and data erase operations. In various embodiments, the controller 140 is responsible for executing instructions from the host 110. The controller 140 can include firmware in the form of executable microcode instructions and/or hardware in the form of an application specific integrated circuit (ASIC) and transistor circuitry. As described herein, the A/C bus 154 and the output bus 157 coupled to the host 110 to send signals to the controller 140 and/or receive signals from the controller 140, along with the I/O circuitry 144 used for bi-directional data communication with host 110 over the data bus 156, may be idle during the performance of the compute operations and/or logical operations in the self-refresh state.
In various embodiments, the controller 140 is responsible for executing instructions from the host 110 and sequencing access to the array 130, among other functions. For example, executing instructions from host 110 can include performing operations, e.g., by executing microcode instructions, using processing resources corresponding to the counter registers 136, mode registers 138, sensing circuitry 150, and/or logic 170, as described further herein. The controller 140 can include a state machine, e.g., firmware and/or hardware in the form of an ASIC, a sequencer, and/or some other type of controlling circuitry. In various embodiments the controller 140 can control shifting data, e.g., right or left, in an array 130.
In the example shown in
As described further below, in a number of embodiments, the sensing circuitry 150 can comprise a number of sense amplifiers and a number of compute components, which may serve as an accumulator, and can be used to perform various compute operations, e.g., to perform logical operations on data associated with complementary sense lines. In a number of embodiments, storage locations, e.g., latches, corresponding to the compute components can serve as stages of a shift register. For example, clock signals can be applied to the compute components to shift data from one compute component to an adjacent compute component.
In a number of embodiments, the sensing circuitry 150 can be used to perform logical operations using data stored in array 130 as inputs and store the results of the logical operations back to the array 130 without transferring data via a sense line address access. e.g., without firing a column decode signal. As such, various compute functions can be performed using, and within, sensing circuitry 150 rather than (or in association with) being performed by processing resources external to the sensing circuitry, e.g., by a processor associated with host 110 and/or other processing circuitry, such as ALU circuitry, located on device 120, e.g., on controller 140 or elsewhere.
In various previous approaches, data associated with an operand, for instance, would be read from memory via sensing circuitry and provided to external ALU circuitry via I/O lines, e.g., via local I/O lines and/or global I/O lines. The external ALU circuitry could include a number of registers and would perform compute functions using the operands, and the result would be transferred back to the array via the I/O lines. In contrast, in a number of embodiments of the present disclosure, sensing circuitry 150 may be configured to perform logical operations on data stored in memory array 130 and store the result back to the memory array 130 without enabling an I/O line, e.g., a local I/O line, coupled to the sensing circuitry 150. Additional logic circuitry 170 can be coupled to the sensing circuitry 150 and can be used to store, e.g., cache and/or buffer, results of operations described herein.
The sensing circuitry 150 can be formed on pitch with the memory cells of the array. In some instances, circuitry of processing resource(s), e.g., a compute engine, may not conform to pitch rules associated with a memory array. For example, the memory cells of a memory array may have a 4F2 or 6F2 cell size, where “F” is a feature size corresponding to the cells. As such, the devices, e.g., logic gates, associated with ALU circuitry of previous PIM systems may not be capable of being formed on pitch with the memory cells, which can affect chip size and/or memory density, for example. A number of embodiments of the present disclosure can include the control circuitry and/or the sensing circuitry, e.g., including sense amplifiers and/or compute components, as described herein, being formed on pitch with the memory cells of the array and being configured to, e.g., being capable of performing, compute functions, e.g., memory and/or PIM operations, on pitch with the memory cells. The sensing circuitry can, in some embodiments, be capable of performing data sensing and compute functions and at least temporary storage, e.g., caching, of data local to the array of memory cells.
PIM capable device operations can use bit vector based operations. As used herein, the term “bit vector” is intended to mean a number of bits on a bit vector memory device, e.g., a PIM device, stored in a row of an array of memory cells and/or in sensing circuitry. Thus, as used herein a “bit vector operation” is intended to mean an operation that is performed on a bit vector that is a portion of virtual address space and/or physical address space, e.g., used by a PIM device. In some embodiments, the bit vector may be a physically contiguous number of bits on the bit vector memory device stored physically contiguous in a row and/or in the sensing circuitry such that the bit vector operation is performed on a bit vector that is a contiguous portion of the virtual address space and/or physical address space. For example, a row of virtual address space in the PIM device may have a bit length of 16K bits, e.g., corresponding to 16K complementary pairs of memory cells in a DRAM configuration. Sensing circuitry 150, as described herein, for such a 16K bit row may include a corresponding 16K processing elements, e.g., compute components as described herein, formed on pitch with the sense lines selectably coupled to corresponding memory cells in the 16 bit row. A compute component in the PIM device may operate as a one bit vector processing element on a single bit of the bit vector of the row of memory cells sensed by the sensing circuitry 150, e.g., sensed by and/or stored in a sense amplifier 406 paired with the compute component 431, as described further in connection with
As such, in a number of embodiments, circuitry external to array 130 and sensing circuitry 150 is not needed to perform compute functions as the sensing circuitry 150 can perform the appropriate memory and/or logical operations in order to perform such compute functions without the use of an external processing resource. Therefore, the sensing circuitry 150 may be used to complement and/or to replace, at least to some extent, such an external processing resource (or at least the bandwidth consumption of such an external processing resource).
However, in a number of embodiments, the sensing circuitry 150 may be used to perform operations, e.g., to execute instructions, in addition to operations performed by an external processing resource, e.g., host 110. For instance, host 110 and/or sensing circuitry 150 may be limited to performing only certain logical operations and/or a certain number of logical operations.
Enabling an I/O line can include enabling, e.g., turning on, a transistor having a gate coupled to a decode signal, e.g., a column decode signal, and a source/drain coupled to the I/O line. However, embodiments are not limited to performing operations using sensing circuitry, e.g., 150, without enabling column decode lines of the array. Whether or not local I/O lines are used in association with performing operations via sensing circuitry 150, the local I/O line(s) may be enabled in order to transfer a result to a suitable location other than back to the array 130, e.g., to an external register.
Each column 122 is configured to be coupled to sensing circuitry 150, as described in connection with
Each of the of the subarrays 125-0, 125-1, . . . , 125-N−1 can include a plurality of rows 119 shown vertically as Y, e.g., each subarray may include 512 rows in an example DRAM bank. Example embodiments are not limited to the example horizontal and vertical orientation of columns and rows described herein or the example numbers thereof.
As shown in
As shown in
Bank 121 can, for example, represent an example bank of a memory device 120 such one of the plurality of banks, e.g., banks 121-0, . . . , 121-7, shown in
As shown in the example embodiment of
In some embodiments, the controller 140 may be configured to provide instructions (control signals based on commands) and data to a plurality of locations of a particular bank 121 in the memory array 130 and to the sensing component stripes 124-0, 124-1, . . . , 124-N−1 via a write path 149 and/or the data path 155 with control and data registers 151. For example, the control and data registers 151 can provide instructions to be executed using by the sense amplifiers and the compute components of the sensing circuitry 150 in the sensing component stripes 124-0, 124-1, . . . , 124-N−1.
The channel controller 143 can also be coupled to each of the plurality of memory devices, 120-1, . . . , 120-N via a data bus 156, as described in
As shown in
In the example of
For example, each of the plurality of banks, e.g., 121-0, . . . , 121-7, in the plurality of memory devices 120-1, . . . , 120-N can include address circuitry 142 to latch address signals provided over a data bus 156 (e.g., an I/O bus) through I/O circuitry 144. Status and/or exception information can be provided from the controller 140 on the memory device 120 to the channel controller 143, using the bus 157, which in turn can be provided from the plurality of memory devices 120-1, . . . , 120-N to the host 110 and vice versa.
For each of the plurality of banks, e.g., 121-0, . . . , 121-7, address signals can be received through address circuitry 142 and decoded by a row decoder 146 and a column decoder 152 to access the memory array 130. Data can be read from memory array 130 by sensing voltage and/or current changes on the data lines using sensing circuitry 150. The sensing circuitry 150 can read and latch a page, e.g., row, of data from the memory array 130. The I/O circuitry 144 can be used for bi-directional data communication with host 110 over the data bus 156. The write circuitry 148 is used to write data to the memory array 130 and the bus 157 can be used to report status, exception and other data information to the channel controller 143.
The channel controller 143 can include one or more local buffers 161 to store microcode instructions and can include logic 160 to allocate a plurality of locations, e.g., subarrays or portions of subarrays, in the arrays of each respective bank to store microcode instructions, e.g., bank commands and arguments, PIM commands, etc., for the various banks associated with the operation of each of the plurality of memory devices 120-1, . . . , 120-N. The channel controller 143 can send microcode instructions, e.g., bank commands and arguments, PIM commands, status and exception information, etc., to the plurality of memory devices 120-1, . . . , 120-N to store those microcode instructions within a given bank of a memory device. For example, the channel controller 143 and/or bank arbiter 145 may send, e.g., as received from host 110, mode selection instructions to mode registers 138-2-1, . . . , 138-2-N associated with, e.g., via interfaces 141-1, . . . , 141-N, the respective plurality of banks 121-1, . . . , 121-7 in each of the respective plurality of memory devices 120-1, . . . , 120-N.
As described above in connection with
In the example of
In the example shown in
The control logic 131 may receive microcode instructions from cache 171 and/or host 110 and may decode microcode instructions into function calls, e.g., microcode function calls (uCODE), implemented by the sequencers 132. The microcode function calls can be the operations that the sequencer 132 receives and executes to cause the memory device 120 to perform particular compute and/or logical operations using the sensing circuitry such as sensing circuitry 150 in
In the example shown in
As described in connection with
As such, the control logic 131, sequencer 132, and timing circuitry 133 may operate to generate sequences of operation cycles for a DRAM array. In the memory device 120 example, each sequence may be designed to perform operations, such as a Boolean logical operations AND, OR, XOR, etc., which together achieve a specific function. For example, the sequences of operations may repetitively perform a logical operation for a one (1) bit add in order to calculate a multiple bit sum. Each sequence of operations may be fed into a first in/first out (FIFO) buffer coupled to the timing circuitry 133 to provide timing coordination with the sensing circuitry 150 and/or additional logic circuitry 170 associated with the array of memory cells 130, e.g., DRAM arrays, shown in
In the example memory device 120 shown in
The D mode 235 in the mode register 238 can be used for performance of the compute and/or logical operations at a rate corresponding to the default frequency for a memory refresh cycle for the data stored in the memory cells, as described above. The D mode 235 can be used to refresh data based on the row in the memory array in which the data is stored, as determined by the reference. In some embodiments, a refresh cycle interval of around 15 μs may be a default frequency, e.g., the default self-refresh mode, for a memory refresh cycle in the self-refresh state for data stored in the memory cells. In the self-refresh state, the data may be read from each row, e.g., to perform computation and/or logical operations, at a frequency of the refresh cycle interval, which may be around 15 μs in the default self-refresh mode.
The N mode 239 in the mode register 238 may be selected to prevent computations, e.g., computation and/or logical operations, from being performed in the self-refresh state. N mode 239 may, for example, be selected to protect data in particular locations in the memory, to specify by exclusion which locations in the memory are usable for computations in the self-refresh state, and/or to ensure that a mobile device including the memory devices described herein remains in a state of relatively reduced power consumption, among other possible reasons for selecting N mode 239.
The M mode 237 in the mode register 238 can be selected to enable adjustment of the D mode 235 and/or the N mode 239 by changing a refresh frequency setting, e.g., via microcode instructions, in the counter register, e.g., as shown at 136-1 in
The set 334 of mode instructions may be saved in a number of mode registers. For example, the set 334 of mode instructions may be saved in mode register 138-1 in the controller 140 described in connection with
The set 334 of mode instructions for banks 321-0, . . . , 121-N of the memory device 120 can include D mode 335 being selected for bank 321-0 via bits in microcode instructions. For example, a microcode instruction may include a bit that causes a D mode 335 column to store a data unit, e.g., 1 in binary, corresponding to a row to designate bank 321-0 and bits that cause M mode 337 and N mode 339 columns to store a different data unit, e.g., 0 in binary, corresponding to the row to designate bank 321-0. As such, the microcode instructions can enable selection of the D mode 335, e.g., the default mode for a memory refresh cycle in the self-refresh state, for data stored in the memory cells of bank 321-0.
The same microcode instruction or a different microcode instruction may include a bit that causes an M mode 337 column to store a data unit corresponding to a row to designate bank 321-1 and bits that cause D mode 335 and N mode 339 columns to store a different data unit corresponding to the row to designate bank 321-1. As such, the microcode instructions can enable selection of the M mode 337, e.g., the modulated mode for adjustment of the memory refresh cycle relative to the D mode 335 and/or the N mode 339, for data stored in the memory cells of bank 321-1.
The same microcode instruction or a different microcode instruction may include a bit that causes an N mode 339 column to store a data unit corresponding to a row to designate bank 321-2 and bits that cause D mode 335 and M mode 337 columns to store a different data unit corresponding to the row to designate bank 321-2. As such, the microcode instructions can enable selection of the N mode 339, e.g., to prevent computations from being performed in the self-refresh state, for data stored in the memory cells of bank 321-2.
In some embodiments, the microcode instructions just described may be sent by the host 110. Updated microcode instructions, e.g., to change modes selected for particular banks, also may be sent by the host 110. In various embodiments, the microcode instructions may be decoded by the controller 140 and setting of values in the mode register may be directed by the controller 140 and/or the microcode instructions may be sent directly to the mode register to set the modes for the banks. In some embodiments, the row, column, and/or subarray indexes of a counter register 136 may be utilized to further specify to which row, column, and/or subarray in a particular bank the selected mode is to be applied.
A memory cell can include a storage element, e.g., capacitor, and an access device, e.g., transistor. For instance, a first memory cell can include transistor 402-1 and capacitor 403-1, and a second memory cell can include transistor 402-2 and capacitor 403-2, etc. In this embodiment, the memory array 430 is a DRAM array of 1T1C (one transistor one capacitor) memory cells, although other embodiments of configurations can be use, e.g., 2T2C with two transistors and two capacitors per memory cell. In a number of embodiments, the memory cells may be destructive read memory cells, e.g., reading the data stored in the cell destroys the data such that the data originally stored in the cell is refreshed after being read. The data stored in the memory cells of the memory array 430 also can be refreshed in a self-refresh state as instructed by circuitry, as described herein, located in, or associated with, the memory array 430 and/or a controller 140 coupled thereto, e.g., as opposed to being instructed to refresh by a functionality in the host 110.
The cells of the memory array 430 can be arranged in rows coupled by access (word) lines 404-X (Row X), 404-Y (Row Y), etc., and columns coupled by pairs of complementary sense lines, e.g., digit lines DIGIT(D) and DIGIT(D)_ shown in
Although rows and columns are illustrated as orthogonally oriented in a plane, embodiments are not so limited. For example, the rows and columns may be oriented relative to each other in any feasible three-dimensional configuration. For example, the rows and columns may be oriented at any angle relative to each other, may be oriented in a substantially horizontal plane or a substantially vertical plane, and/or may be oriented in a folded topology, among other possible three-dimensional configurations.
Memory cells can be coupled to different digit lines and word lines. For example, a first source/drain region of a transistor 402-1 can be coupled to digit line 405-1 (D), a second source/drain region of transistor 402-1 can be coupled to capacitor 403-1, and a gate of a transistor 402-1 can be coupled to word line 404-Y. A first source/drain region of a transistor 402-2 can be coupled to digit line 405-2 (D)_, a second source/drain region of transistor 402-2 can be coupled to capacitor 403-2, and a gate of a transistor 402-2 can be coupled to word line 404-X. A cell plate, as shown in
The memory array 430 is configured to couple to sensing circuitry 450 in accordance with a number of embodiments of the present disclosure. In this embodiment, the sensing circuitry 450 comprises a sense amplifier 406 and a compute component 431 corresponding to respective columns of memory cells, e.g., coupled to respective pairs of complementary digit lines. The sense amplifier 406 can be coupled to the pair of complementary digit lines 405-1 and 405-2. The compute component 431 can be coupled to the sense amplifier 406 via pass gates 407-1 and 407-2. The gates of the pass gates 407-1 and 407-2 can be coupled to operation selection logic 413.
The operation selection logic 413 can be configured to include pass gate logic for controlling pass gates that couple the pair of complementary digit lines un-transposed between the sense amplifier 406 and the compute component 431 and swap gate logic for controlling swap gates that couple the pair of complementary digit lines transposed between the sense amplifier 406 and the compute component 431. The operation selection logic 413 can also be coupled to the pair of complementary digit lines 405-1 and 405-2. The operation selection logic 413 can be configured to control continuity of pass gates 407-1 and 407-2 based on a selected operation.
The sense amplifier 406 can be operated to determine a data value, e.g., logic state, stored in a selected memory cell. The sense amplifier 406 can comprise a cross coupled latch, which can be referred to herein as a primary latch. In the example illustrated in
In operation, when a memory cell is being sensed, e.g., read, the voltage on one of the digit lines 405-1 (D) or 405-2 (D)_ will be slightly greater than the voltage on the other one of digit lines 405-1 (D) or 405-2 (D)_. An ACT 465 signal and an RNiF 428 signal can be driven low to enable, e.g., fire, the sense amplifier 406. The digit lines 405-1 (D) or 405-2 (D)_ having the lower voltage will turn on one of the PMOS transistor 429-1 or 429-2 to a greater extent than the other of PMOS transistor 429-1 or 429-2, thereby driving high the digit line 405-1 (D) or 405-2 (D)_ having the higher voltage to a greater extent than the other digit line 405-1 (D) or 405-2 (D)_ is driven high.
Similarly, the digit line 405-1 (D) or 405-2 (D)_ having the higher voltage will turn on one of the NMOS transistor 427-1 or 427-2 to a greater extent than the other of the NMOS transistor 427-1 or 427-2, thereby driving low the digit line 405-1 (D) or 405-2 (D)_ having the lower voltage to a greater extent than the other digit line 405-1 (D) or 405-2 (D)_ is driven low. As a result, after a short delay, the digit line 405-1 (D) or 405-2 (D)_ having the slightly greater voltage is driven to the voltage of the supply voltage VCC through a source transistor, and the other digit line 405-1 (D) or 405-2 (D)_ is driven to the voltage of the reference voltage, e.g., ground, through a sink transistor. Therefore, the cross coupled NMOS transistors 427-1 and 427-2 and PMOS transistors 429-1 and 429-2 serve as a sense amplifier pair, which amplify the differential voltage on the digit lines 405-1 (D) and 405-2 (D)_ and operate to latch a data value sensed from the selected memory cell.
Embodiments are not limited to the sense amplifier 406 configuration illustrated in
The sense amplifier 406 can, in conjunction with the compute component 431, be operated to perform various operations using data from an array as input. In a number of embodiments, the result of an operation can be stored back to the array without transferring the data via a digit line address access and/or moved between banks without using an external data bus, e.g., without firing a column decode signal such that data is transferred to circuitry external from the array and sensing circuitry via local I/O lines. As such, a number of embodiments of the present disclosure can enable performing operations and compute functions associated therewith using less power than various previous approaches. Additionally, since a number of embodiments eliminate the need to transfer data across local and global I/O lines and/or external data buses in order to perform compute functions, e.g., between memory and discrete processor, a number of embodiments can enable an increased, e.g., faster, processing capability as compared to previous approaches.
The sense amplifier 406 can further include equilibration circuitry 414, which can be configured to equilibrate the digit lines 405-1 (D) and 405-2 (D)_. In this example, the equilibration circuitry 414 comprises a transistor 424 coupled between digit lines 405-1 (D) and 405-2 (D)_. The equilibration circuitry 414 also comprises transistors 425-1 and 425-2 each having a first source/drain region coupled to an equilibration voltage, e.g., VDD 438, where VDD is a supply voltage associated with the array. A second source/drain region of transistor 425-1 can be coupled digit line 405-1 (D), and a second source/drain region of transistor 425-2 can be coupled digit line 405-2 (D)_. Gates of transistors 424, 425-1, and 425-2 can be coupled together, and to an equilibration (EQ) control signal line 426. As such, activating EQ 426 enables the transistors 424, 425-1, and 425-2, which effectively shorts digit lines 405-1 (D) and 405-2 (D)_ together and to the equilibration voltage, e.g., VDD/2 438.
Although
As described further below, in a number of embodiments, the sensing circuitry 450, e.g., sense amplifier 406 and compute component 431, can be operated to perform a selected operation and initially store the result in one of the sense amplifier 406 or the compute component 431. For example, the result may be initially stored in one of the sense amplifier 406 or the compute component 431 without transferring data from the sensing circuitry via a local or global I/O line and/or moved between banks without using an external data bus, e.g., without performing a sense line address access via activation of a column decode signal, for instance.
Performance of operations, e.g., Boolean logical operations involving data values, is fundamental and commonly used. Boolean logical operations are used in many higher level operations. Consequently, speed and/or power efficiencies that can be realized with improved operations can translate into speed and/or power efficiencies of higher order functionalities.
As shown in
The memory device can include a sensing component stripe, e.g., as shown at 124-0, 124-1, . . . , 124-N−1 and described in connection with
The gates of the pass gates 507-1 and 507-2 can be controlled by a logical operation selection logic signal, Pass. For example, an output of the logical operation selection logic can be coupled to the gates of the pass gates 507-1 and 507-2. The compute component 531 can comprise a loadable shift register configured to shift data values left and right.
According to the embodiment illustrated in
The sensing circuitry shown in
According to various embodiments, the logical operation selection logic 513 can include four logic selection transistors: logic selection transistor 562 coupled between the gates of the swap transistors 542 and a TF signal control line, logic selection transistor 552 coupled between the gates of the pass gates 507-1 and 507-2 and a TT signal control line, logic selection transistor 554 coupled between the gates of the pass gates 507-1 and 507-2 and a FT signal control line, and logic selection transistor 564 coupled between the gates of the swap transistors 542 and a FF signal control line. Gates of logic selection transistors 562 and 552 are coupled to the true sense line through isolation transistor 550-1, e.g., having a gate coupled to an ISO signal control line. Gates of logic selection transistors 564 and 554 are coupled to the complementary sense line through isolation transistor 550-2, e.g., also having a gate coupled to an ISO signal control line.
Data values present on the pair of complementary sense lines 505-1 and 505-2 can be loaded into the compute component 531 via the pass gates 507-1 and 507-2. The compute component 531 can comprise a loadable shift register. When the pass gates 507-1 and 507-2 are OPEN, data values on the pair of complementary sense lines 505-1 and 505-2 are passed to the compute component 531 and thereby loaded into the loadable shift register. The data values on the pair of complementary sense lines 505-1 and 505-2 can be the data value stored in the sense amplifier 506 when the sense amplifier is fired. The logical operation selection logic signal, Pass, is high to OPEN the pass gates 507-1 and 507-2.
The ISO, TF, TT, FT, and FF control signals can operate to select a logical function to implement based on the data value (“B”) in the sense amplifier 506 and the data value (“A”) in the compute component 531. In particular, the ISO, TF, TT, FT, and FF control signals are configured to select the logical function to implement independent from the data value present on the pair of complementary sense lines 505-1 and 505-2, although the result of the implemented logical operation can be dependent on the data value present on the pair of complementary sense lines 505-1 and 505-2. For example, the ISO, TF, TT, FT, and FF control signals may select the logical operation to implement directly because the data value present on the pair of complementary sense lines 505-1 and 505-2 is not passed through logic to operate the gates of the pass gates 507-1 and 507-2.
Additionally,
The logical operation selection logic signal Pass can be activated, e.g., high, to OPEN the pass gates 507-1 and 507-2, e.g., conducting, when the ISO control signal line is activated and either the TT control signal is activated, e.g., high, and data value on the true sense line is “1” or the FT control signal is activated, e.g., high, and the data value on the complement sense line is “1.”
The data value on the true sense line being a “1” OPENs logic selection transistors 552 and 562. The data value on the complimentary sense line being a “1” OPENs logic selection transistors 554 and 564. If the ISO control signal or either the respective TT/FT control signal or the data value on the corresponding sense line, e.g., sense line to which the gate of the particular logic selection transistor is coupled, is not high, then the pass gates 507-1 and 507-2 will not be OPENed by a particular logic selection transistor.
The logical operation selection logic signal Pass* can be activated, e.g., high, to OPEN the swap transistors 542, e.g., conducting, when the ISO control signal line is activated and either the TF control signal is activated, e.g., high, and data value on the true sense line is “1,” or the FF control signal is activated, e.g., high, and the data value on the complement sense line is “1.” If either the respective control signal or the data value on the corresponding sense line, e.g., sense line to which the gate of the particular logic selection transistor is coupled, is not high, then the swap transistors 542 will not be OPENed by a particular logic selection transistor.
The Pass* control signal is not necessarily complementary to the Pass control signal. It is possible for the Pass and Pass* control signals to both be activated or both be deactivated at the same time. However, activation of both the Pass and Pass* control signals at the same time shorts the pair of complementary sense lines together, which may be a disruptive configuration to be avoided.
The sensing circuitry illustrated in
Logic Table 6-1 illustrated in
Via selective control of the continuity of the pass gates 507-1 and 507-2 and the swap transistors 542, each of the three columns of the upper portion of Logic Table 6-1 can be combined with each of the three columns of the lower portion of Logic Table 6-1 to provide 3×3=9 different result combinations, corresponding to nine different logical operations, as indicated by the various connecting paths shown at 675. The nine different selectable logical operations that can be implemented by the sensing circuitry, e.g., 150 in
The columns of Logic Table 6-2 illustrated in
As described herein, sensing circuitry 150, e.g., as described in connection with
A result of a selected logical operation is based on a first data value that may be stored in a sense amplifier and a second data value that may be stored in a compute component, e.g., an accumulator, a shift circuit. The result of the selected logical operation may be initially stored in the sense amplifier for some selected logical operations, and may be initially stored in the compute component for some selected logical operations. Some selected logical operations may be implemented so as to have the result stored in either the sense amplifier or the compute component. In a number of embodiments, whether a result of a logical operation is initially stored in the sense amplifier or the compute component can depend on when logical selection control signals corresponding to a selected logical operation to be performed are provided to logical selection logic of the sensing circuitry, e.g., whether the logic selection control signals are fired before or after the sense amplifier is fired. According to some embodiments, logical operation selection logic may be configured to control pass gates, e.g., control continuity of the pass gates, based on a data value stored in the compute component and the selected logical operation. Controlling continuity of a gate, e.g., transistor, may be used herein to refer to controlling whether or not the gate is conducting, e.g., whether a channel of the transistor is in a conducting or non-conducting state.
The timing diagram 760 shown in
At time to for both the refresh cycle 766 and the XOR cycle 767, EQ is disabled and the ROW X data value may be latched in the compute component, e.g., 431. At time t1, ROW Y is enabled, e.g., goes high, to access, e.g., select, the ROW Y memory cell. At time t2, the sense amplifier, e.g., 406, is enabled, e.g., goes high, is fired, which drives the complementary sense lines, e.g., 405-1 and 405-2 and 505-1 and 505-2 in
For the XOR cycle 767 only, at time t4, the PHASE 2R and PHASE 2L, e.g., as shown at 583 and 592, respectively, and described in connection with
In this example, if PASS goes high at time t4, the pass transistors 507-1 and 507-2 are enabled such that the DIGIT and DIGIT_ signals, which correspond to the ROW Y data value, are provided to the respective compute component nodes ST2 and SF2. As such, the value stored in the compute component 531 (e.g., the ROW X data value) may be flipped, depending on the value of DIGIT and DIGIT_ (e.g., the ROW Y data value). In this example, if PASS stays low at time t4, the pass transistors 507-1 and 507-2 are not enabled such that the DIGIT and DIGIT_ signals, which correspond to the ROW Y data value, remain isolated from the nodes ST2 and SF2 of the compute component 531. As such, the data value in the compute component (e.g., the ROW X data value) would remain the same. In this example, if PASS* goes high at time t4, the swap transistors 542 are enabled such that the DIGIT and DIGIT signals, which correspond to the ROW Y data value, are provided to the respective compute component nodes ST2 and SF2 in a transposed manner, e.g., the “true” data value on DIGIT(n) would be provided to node SF2 and the “complement” data value on DIGIT(n) would be provided to node ST2. As such, the value stored in the compute component 53, e.g., the ROW X data value, may be flipped, depending on the value of DIGIT and DIGIT_, e.g., the ROW Y data value. In this example, if PASS* stays low at time t4, the swap transistors 542 are not enabled such that the DIGIT and DIGIT_ signals, which correspond to the ROW Y data value, remain isolated from the nodes ST2 and SF2 of the compute component 531. As such, the data value in the compute component, e.g., the ROW X data value, would remain the same.
At time t5, TF and FT are disabled, which results in PASS and PASS* going (or remaining) low, such that the pass transistors 507-1 and 507-2 and swap transistors 542 are disabled. At time t5, ROW Y is disabled, and PHASE 2R, PHASE 2L, and ISO are enabled. Enabling PHASE 2R and PHASE 2L at time t5 enables feedback on the latch of the compute component 531 such that the result of the XOR operation (e.g., “A” XOR “B”) is latched therein. Enabling ISO 558 at time t5 again couples nodes ST2 and SF2 to the gates of the enable transistors 552, 554, 562, and 564.
At time t7 for both the refresh cycle 766 and the XOR cycle 767, equilibration is enabled, e.g., EQ goes high such that DIGIT and DIGIT are driven to an equilibrate voltage and the sense amplifier 506 is disabled, e.g., goes low. The sense (read) operation included in both the refresh cycle 766 and the XOR cycle 767 coupling to the row to access data values from the memory cells therein destroys the data such that the data originally stored in the memory cell may be refreshed after being read. In the case of a compute operation, e.g., a read operation, performed during the self-refresh state, the data values may be utilized for performance of the high latency operations described herein and transferred back, e.g., refreshed, to the memory array, e.g., to a memory cell coupled to ROW X, ROW Y, and/or a different row via the complementary sense lines. In the case of a logical operation, e.g., an XOR operation, performed during the self-refresh state, the result of the XOR operation, which is initially stored in the compute component 531 in this example, can be transferred to the memory array e.g., to a memory cell coupled to ROW X, ROW Y, and/or a different row via the complementary sense lines.
Initiation at t0 of the refresh cycle 766 and/or the XOR cycle 767 just described coincides with initiation of performance of the compute operations and/or logical operations in the self-refresh state. As described herein, a refresh cycle interval which may be around 15 μs, e.g., in a default self-refresh mode, which determines the rate at which the data may be read from each row to perform the compute and/or logical operations described herein. When no longer in the self-refresh state, the results of performance of the high latency compute and/or logical operations may be sent to and/or accessed by an external location, e.g., an external processing component of the host 110, via I/O lines.
Embodiments described herein provide a method of operating an apparatus that may be in the form of a computing system 100 including a memory device 120 for performing operations, as described herein, by the memory device in a self-refresh state. As described herein, the method can include selecting from a plurality of modes, e.g., as shown at 235, 237, and 239 and described in connection with
The method can include adjusting a frequency of performance of a memory refresh cycle for the data stored in the memory cells and performing the compute operations at a rate corresponding to the adjusted frequency of performance of the memory refresh cycle, as described in connection with
The method can include performing the compute and/or logical operations described herein in the self-refresh state, in which high latency is not a burden for performing such operations. A battery-powered mobile device, for example, may be in a low power state quite often, e.g., while in a user's pocket or purse, while the user is asleep, etc. During those periods, the data stored in memory cells may be retained there because the memory device is in the self-refresh state. The compute operations described herein may be performed with high latency in the self-refresh state because, for example, the user is not actively interacting with the mobile device so lack of low latency and/or presence of high latency in performance of operations is not noticeable. The high latency may not be a burden because the data is processed in memory during the low power and/or self-refresh state and the processed data is available for access by the user at some later time.
Examples of applications that may be operated to take advantage of the low power and/or self-refresh state may include operations intended to run as background operations that may not involve user interaction, e.g., with the host. Such high latency background operations may include: facial detection in images; feature extraction from images; security scan of in-memory threats, such as viruses, worms, Trojan horses, etc.; neural network processing; and parsing of large data sets; among other types of operations. Other operations that may be performed in the low power and/or self-refresh state may include operations that may not use a full computing potential, e.g., accuracy and/or speed, of a computing system, even though a user may be actively interacting with the system. Such operations may include: electronic games; video playback; and camera input; among other types of operations. In some examples, for error-tolerant applications, e.g., graphics applications, data may be self-refreshed and operations performed at a rate lower than the default frequency for a memory refresh cycle in the self-refresh state, e.g., in order to reduce power consumption, with minor quality loss, e.g., as approximate computing. At least some of these operations may include performing logical operations, e.g., PIM operations such as Boolean operations, as described herein. Whereas these operations may be performed in the self-refresh state, performance as such may be intended to provide a result that is accessible when the memory device is not in the self-refresh state and is interacting with the host.
While example embodiments including various combinations and configurations of sensing circuitry, sense amplifiers, compute component, dynamic latches, isolation devices, and/or shift circuitry have been illustrated and described herein, embodiments of the present disclosure are not limited to those combinations explicitly recited herein. Other combinations and configurations of the memory device, controller, counter register, mode register, memory array, sensing circuitry, logic circuitry, and/or cache disclosed herein are expressly included within the scope of this disclosure.
Although specific embodiments have been illustrated and described herein, those of ordinary skill in the art will appreciate that an arrangement calculated to achieve the same results can be substituted for the specific embodiments shown. This disclosure is intended to cover adaptations or variations of one or more embodiments of the present disclosure. It is to be understood that the above description has been made in an illustrative fashion, and not a restrictive one. Combination of the above embodiments, and other embodiments not specifically described herein will be apparent to those of skill in the art upon reviewing the above description. The scope of the one or more embodiments of the present disclosure includes other applications in which the above structures and methods are used. Therefore, the scope of one or more embodiments of the present disclosure should be determined with reference to the appended claims, along with the full range of equivalents to which such claims are entitled.
In the foregoing Detailed Description, some features are grouped together in a single embodiment for the purpose of streamlining the disclosure. This method of disclosure is not to be interpreted as reflecting an intention that the disclosed embodiments of the present disclosure have to use more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive subject matter lies in less than all features of a single disclosed embodiment. Thus, the following claims are hereby incorporated into the Detailed Description, with each claim standing on its own as a separate embodiment.
This application is a Divisional of U.S. application Ser. No. 15/222,514, filed Jul. 28, 2016, which issues as U.S. Pat. No. 10,468,087 on Nov. 5, 2019, the contents of which are included herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
4380046 | Fung | Apr 1983 | A |
4435792 | Bechtolsheim | Mar 1984 | A |
4435793 | Ochii | Mar 1984 | A |
4727474 | Batcher | Feb 1988 | A |
4843264 | Galbraith | Jun 1989 | A |
4958378 | Bell | Sep 1990 | A |
4977542 | Matsuda et al. | Dec 1990 | A |
5023838 | Herbert | Jun 1991 | A |
5034636 | Reis et al. | Jul 1991 | A |
5201039 | Sakamura | Apr 1993 | A |
5210850 | Kelly et al. | May 1993 | A |
5253308 | Johnson | Oct 1993 | A |
5276643 | Hoffmann et al. | Jan 1994 | A |
5325519 | Long et al. | Jun 1994 | A |
5367488 | An | Nov 1994 | A |
5379257 | Matsumura et al. | Jan 1995 | A |
5386379 | Ail-Yahia et al. | Jan 1995 | A |
5398213 | Yeon et al. | Mar 1995 | A |
5440482 | Davis | Aug 1995 | A |
5446690 | Tanaka et al. | Aug 1995 | A |
5473576 | Matsui | Dec 1995 | A |
5481500 | Reohr et al. | Jan 1996 | A |
5485373 | Davis et al. | Jan 1996 | A |
5506811 | McLaury | Apr 1996 | A |
5615404 | Knoll et al. | Mar 1997 | A |
5638128 | Hoogenboom | Jun 1997 | A |
5638317 | Tran | Jun 1997 | A |
5654936 | Cho | Aug 1997 | A |
5678021 | Pawate et al. | Oct 1997 | A |
5724291 | Matano | Mar 1998 | A |
5724366 | Furutani | Mar 1998 | A |
5751987 | Mahant-Shetti et al. | May 1998 | A |
5787458 | Miwa | Jul 1998 | A |
5854636 | Watanabe et al. | Dec 1998 | A |
5867429 | Chen et al. | Feb 1999 | A |
5870504 | Nemoto et al. | Feb 1999 | A |
5915084 | Wendell | Jun 1999 | A |
5935263 | Keeth et al. | Aug 1999 | A |
5986942 | Sugibayashi | Nov 1999 | A |
5991209 | Chow | Nov 1999 | A |
5991785 | Alidina et al. | Nov 1999 | A |
6005799 | Rao | Dec 1999 | A |
6009020 | Nagata | Dec 1999 | A |
6092186 | Betker et al. | Jul 2000 | A |
6122211 | Morgan et al. | Sep 2000 | A |
6125071 | Kohno et al. | Sep 2000 | A |
6134164 | Lattimore et al. | Oct 2000 | A |
6147514 | Shiratake | Nov 2000 | A |
6151244 | Fujino et al. | Nov 2000 | A |
6157578 | Brady | Dec 2000 | A |
6163862 | Adams et al. | Dec 2000 | A |
6166942 | Vo et al. | Dec 2000 | A |
6172918 | Hidaka | Jan 2001 | B1 |
6175514 | Henderson | Jan 2001 | B1 |
6181698 | Hariguchi | Jan 2001 | B1 |
6208544 | Beadle et al. | Mar 2001 | B1 |
6226215 | Yoon | May 2001 | B1 |
6301153 | Takeuchi et al. | Oct 2001 | B1 |
6301164 | Manning et al. | Oct 2001 | B1 |
6304477 | Naji | Oct 2001 | B1 |
6389507 | Sherman | May 2002 | B1 |
6418498 | Martwick | Jul 2002 | B1 |
6449204 | Arimoto | Sep 2002 | B1 |
6466499 | Blodgett | Oct 2002 | B1 |
6510098 | Taylor | Jan 2003 | B1 |
6563754 | Lien et al. | May 2003 | B1 |
6578058 | Nygaard | Jun 2003 | B1 |
6731542 | Le et al. | May 2004 | B1 |
6754746 | Leung et al. | Jun 2004 | B1 |
6768679 | Le et al. | Jul 2004 | B1 |
6807614 | Chung | Oct 2004 | B2 |
6816422 | Hamade et al. | Nov 2004 | B2 |
6819612 | Achter | Nov 2004 | B1 |
6894549 | Eliason | May 2005 | B2 |
6943579 | Hazanchuk et al. | Sep 2005 | B1 |
6948056 | Roth et al. | Sep 2005 | B1 |
6950771 | Fan et al. | Sep 2005 | B1 |
6950898 | Merritt et al. | Sep 2005 | B2 |
6956770 | Khalid et al. | Oct 2005 | B2 |
6961272 | Schreck | Nov 2005 | B2 |
6965648 | Smith et al. | Nov 2005 | B1 |
6985394 | Kim | Jan 2006 | B2 |
6987693 | Cernea et al. | Jan 2006 | B2 |
7020017 | Chen et al. | Mar 2006 | B2 |
7028170 | Saulsbury | Apr 2006 | B2 |
7045834 | Tran et al. | May 2006 | B2 |
7054178 | Shiah et al. | May 2006 | B1 |
7061817 | Raad et al. | Jun 2006 | B2 |
7079407 | Dimitrelis | Jul 2006 | B1 |
7173857 | Kato et al. | Feb 2007 | B2 |
7187585 | Li et al. | Mar 2007 | B2 |
7196928 | Chen | Mar 2007 | B2 |
7260565 | Lee et al. | Aug 2007 | B2 |
7260672 | Gamey | Aug 2007 | B2 |
7372715 | Han | May 2008 | B2 |
7400532 | Aritome | Jul 2008 | B2 |
7406494 | Magee | Jul 2008 | B2 |
7447720 | Beaumont | Nov 2008 | B2 |
7454451 | Beaumont | Nov 2008 | B2 |
7457181 | Lee et al. | Nov 2008 | B2 |
7535769 | Cernea | May 2009 | B2 |
7546438 | Chung | Jun 2009 | B2 |
7562198 | Noda et al. | Jul 2009 | B2 |
7574466 | Beaumont | Aug 2009 | B2 |
7602647 | Li et al. | Oct 2009 | B2 |
7663928 | Tsai et al. | Feb 2010 | B2 |
7685365 | Rajwar et al. | Mar 2010 | B2 |
7692466 | Ahmadi | Apr 2010 | B2 |
7752417 | Manczak et al. | Jul 2010 | B2 |
7791962 | Noda et al. | Sep 2010 | B2 |
7796453 | Riho et al. | Sep 2010 | B2 |
7805587 | Van Dyke et al. | Sep 2010 | B1 |
7808854 | Takase | Oct 2010 | B2 |
7827372 | Bink et al. | Nov 2010 | B2 |
7869273 | Lee et al. | Jan 2011 | B2 |
7898864 | Dong | Mar 2011 | B2 |
7924628 | Danon et al. | Apr 2011 | B2 |
7937535 | Ozer et al. | May 2011 | B2 |
7957206 | Bauser | Jun 2011 | B2 |
7979667 | Allen et al. | Jul 2011 | B2 |
7996749 | Ding et al. | Aug 2011 | B2 |
8042082 | Solomon | Oct 2011 | B2 |
8045391 | Mokhlesi | Oct 2011 | B2 |
8059438 | Chang et al. | Nov 2011 | B2 |
8095825 | Hirotsu et al. | Jan 2012 | B2 |
8117462 | Snapp et al. | Feb 2012 | B2 |
8164942 | Gebara et al. | Apr 2012 | B2 |
8208328 | Hong | Jun 2012 | B2 |
8213248 | Moon et al. | Jul 2012 | B2 |
8223568 | Seo | Jul 2012 | B2 |
8238173 | Akerib et al. | Aug 2012 | B2 |
8274841 | Shimano et al. | Sep 2012 | B2 |
8279683 | Klein | Oct 2012 | B2 |
8310884 | Iwai et al. | Nov 2012 | B2 |
8332367 | Bhattacherjee et al. | Dec 2012 | B2 |
8339824 | Cooke | Dec 2012 | B2 |
8339883 | Yu et al. | Dec 2012 | B2 |
8347154 | Bahali et al. | Jan 2013 | B2 |
8351292 | Matano | Jan 2013 | B2 |
8356144 | Hessel et al. | Jan 2013 | B2 |
8417921 | Gonion et al. | Apr 2013 | B2 |
8462532 | Argyres | Jun 2013 | B1 |
8484276 | Carlson et al. | Jul 2013 | B2 |
8495438 | Roine | Jul 2013 | B2 |
8503250 | Demone | Aug 2013 | B2 |
8526239 | Kim | Sep 2013 | B2 |
8533245 | Cheung | Sep 2013 | B1 |
8555037 | Gonion | Oct 2013 | B2 |
8599613 | Abiko et al. | Dec 2013 | B2 |
8605015 | Guttag et al. | Dec 2013 | B2 |
8625376 | Jung et al. | Jan 2014 | B2 |
8644101 | Jun et al. | Feb 2014 | B2 |
8650232 | Stortz et al. | Feb 2014 | B2 |
8873272 | Lee | Oct 2014 | B2 |
8964496 | Manning | Feb 2015 | B2 |
8971124 | Manning | Mar 2015 | B1 |
9015390 | Klein | Apr 2015 | B2 |
9047193 | Lin et al. | Jun 2015 | B2 |
9165023 | Moskovich et al. | Oct 2015 | B2 |
20010007112 | Porterfield | Jul 2001 | A1 |
20010008492 | Higashiho | Jul 2001 | A1 |
20010010057 | Yamada | Jul 2001 | A1 |
20010028584 | Nakayama et al. | Oct 2001 | A1 |
20010043089 | Forbes et al. | Nov 2001 | A1 |
20020021609 | Kitamoto et al. | Feb 2002 | A1 |
20020059355 | Peleg et al. | May 2002 | A1 |
20030167426 | Slobodnik | Sep 2003 | A1 |
20030222879 | Lin et al. | Dec 2003 | A1 |
20040042333 | Shore et al. | Mar 2004 | A1 |
20040073592 | Kim et al. | Apr 2004 | A1 |
20040073773 | Demjanenko | Apr 2004 | A1 |
20040085840 | Vali et al. | May 2004 | A1 |
20040095826 | Perner | May 2004 | A1 |
20040154002 | Ball et al. | Aug 2004 | A1 |
20040165465 | Kashiwazaki | Aug 2004 | A1 |
20040184335 | Afghahi et al. | Sep 2004 | A1 |
20040205289 | Srinivasan | Oct 2004 | A1 |
20040221098 | Ito et al. | Nov 2004 | A1 |
20040240251 | Nozawa et al. | Dec 2004 | A1 |
20050015557 | Wang et al. | Jan 2005 | A1 |
20050078514 | Scheuerlein et al. | Apr 2005 | A1 |
20050097417 | Agrawal et al. | May 2005 | A1 |
20060047937 | Selvaggi et al. | Mar 2006 | A1 |
20060069849 | Rudelic | Mar 2006 | A1 |
20060146623 | Mizuno et al. | Jul 2006 | A1 |
20060149804 | Luick et al. | Jul 2006 | A1 |
20060181917 | Kang et al. | Aug 2006 | A1 |
20060215432 | Wickeraad et al. | Sep 2006 | A1 |
20060225072 | Lari et al. | Oct 2006 | A1 |
20060291282 | Liu et al. | Dec 2006 | A1 |
20070011596 | Suh | Jan 2007 | A1 |
20070103986 | Chen | May 2007 | A1 |
20070171747 | Hunter et al. | Jul 2007 | A1 |
20070180006 | Gyoten et al. | Aug 2007 | A1 |
20070180184 | Sakashita et al. | Aug 2007 | A1 |
20070195602 | Fong et al. | Aug 2007 | A1 |
20070253268 | Kim et al. | Nov 2007 | A1 |
20070285131 | Sohn | Dec 2007 | A1 |
20070285979 | Turner | Dec 2007 | A1 |
20070291532 | Tsuji | Dec 2007 | A1 |
20080025073 | Arsovski | Jan 2008 | A1 |
20080037333 | Kim et al. | Feb 2008 | A1 |
20080052711 | Forin et al. | Feb 2008 | A1 |
20080137388 | Krishnan et al. | Jun 2008 | A1 |
20080165601 | Matick et al. | Jul 2008 | A1 |
20080178053 | Gorman et al. | Jul 2008 | A1 |
20080215937 | Dreibelbis et al. | Sep 2008 | A1 |
20090067218 | Graber | Mar 2009 | A1 |
20090154238 | Lee | Jun 2009 | A1 |
20090154273 | Borot et al. | Jun 2009 | A1 |
20090248972 | Ware et al. | Oct 2009 | A1 |
20090254697 | Akerib | Oct 2009 | A1 |
20100067296 | Li | Mar 2010 | A1 |
20100091582 | Vali et al. | Apr 2010 | A1 |
20100124138 | Lee et al. | May 2010 | A1 |
20100172190 | Lavi et al. | Jul 2010 | A1 |
20100210076 | Gruber et al. | Aug 2010 | A1 |
20100226183 | Kim | Sep 2010 | A1 |
20100308858 | Noda et al. | Dec 2010 | A1 |
20100332895 | Billing et al. | Dec 2010 | A1 |
20110051523 | Manabe et al. | Mar 2011 | A1 |
20110063919 | Chandrasekhar et al. | Mar 2011 | A1 |
20110093662 | Walker et al. | Apr 2011 | A1 |
20110103151 | Kim et al. | May 2011 | A1 |
20110119467 | Cadambi et al. | May 2011 | A1 |
20110122695 | Li et al. | May 2011 | A1 |
20110140741 | Zerbe et al. | Jun 2011 | A1 |
20110219260 | Nobunaga et al. | Sep 2011 | A1 |
20110267883 | Lee et al. | Nov 2011 | A1 |
20110317496 | Bunce et al. | Dec 2011 | A1 |
20120005397 | Lim et al. | Jan 2012 | A1 |
20120017039 | Margetts | Jan 2012 | A1 |
20120023281 | Kawasaki et al. | Jan 2012 | A1 |
20120075947 | Kang et al. | Mar 2012 | A1 |
20120120705 | Mitsubori et al. | May 2012 | A1 |
20120134216 | Singh | May 2012 | A1 |
20120134225 | Chow | May 2012 | A1 |
20120134226 | Chow | May 2012 | A1 |
20120140540 | Agam et al. | Jun 2012 | A1 |
20120182798 | Hosono et al. | Jul 2012 | A1 |
20120195146 | Jun et al. | Aug 2012 | A1 |
20120198310 | Tran et al. | Aug 2012 | A1 |
20120246380 | Akerib et al. | Sep 2012 | A1 |
20120265964 | Murata et al. | Oct 2012 | A1 |
20120281486 | Rao et al. | Nov 2012 | A1 |
20120303627 | Keeton et al. | Nov 2012 | A1 |
20130003467 | Klein | Jan 2013 | A1 |
20130061006 | Hein | Mar 2013 | A1 |
20130107623 | Kavalipurapu et al. | May 2013 | A1 |
20130117541 | Choquette et al. | May 2013 | A1 |
20130124783 | Yoon et al. | May 2013 | A1 |
20130132702 | Patel et al. | May 2013 | A1 |
20130138646 | Sirer et al. | May 2013 | A1 |
20130163362 | Kim | Jun 2013 | A1 |
20130173888 | Hansen et al. | Jul 2013 | A1 |
20130194857 | Miyamoto | Aug 2013 | A1 |
20130205114 | Badam et al. | Aug 2013 | A1 |
20130219112 | Okin et al. | Aug 2013 | A1 |
20130227361 | Bowers et al. | Aug 2013 | A1 |
20130283122 | Anholt et al. | Oct 2013 | A1 |
20130286705 | Grover et al. | Oct 2013 | A1 |
20130326154 | Haswell | Dec 2013 | A1 |
20130332707 | Gueron et al. | Dec 2013 | A1 |
20140185395 | Seo | Jul 2014 | A1 |
20140215185 | Danielsen | Jul 2014 | A1 |
20140250279 | Manning | Sep 2014 | A1 |
20140344934 | Jorgensen | Nov 2014 | A1 |
20150029798 | Manning | Jan 2015 | A1 |
20150042380 | Manning | Feb 2015 | A1 |
20150063052 | Manning | Mar 2015 | A1 |
20150078108 | Cowles et al. | Mar 2015 | A1 |
20150120987 | Wheeler | Apr 2015 | A1 |
20150134713 | Wheeler | May 2015 | A1 |
20150270015 | Murphy et al. | Sep 2015 | A1 |
20150279466 | Manning | Oct 2015 | A1 |
20150324290 | Leidel | Nov 2015 | A1 |
20150325272 | Murphy | Nov 2015 | A1 |
20150356009 | Wheeler et al. | Dec 2015 | A1 |
20150356022 | Leidel et al. | Dec 2015 | A1 |
20150357007 | Manning et al. | Dec 2015 | A1 |
20150357008 | Manning et al. | Dec 2015 | A1 |
20150357019 | Wheeler et al. | Dec 2015 | A1 |
20150357020 | Manning | Dec 2015 | A1 |
20150357021 | Hush | Dec 2015 | A1 |
20150357022 | Hush | Dec 2015 | A1 |
20150357023 | Hush | Dec 2015 | A1 |
20150357024 | Hush et al. | Dec 2015 | A1 |
20150357047 | Tiwari | Dec 2015 | A1 |
20160062672 | Wheeler | Mar 2016 | A1 |
20160062673 | Tiwari | Mar 2016 | A1 |
20160062692 | Finkbeiner et al. | Mar 2016 | A1 |
20160062733 | Tiwari | Mar 2016 | A1 |
20160063284 | Tiwari | Mar 2016 | A1 |
20160064045 | La Fratta | Mar 2016 | A1 |
20160064047 | Tiwari | Mar 2016 | A1 |
20160098208 | Willcock | Apr 2016 | A1 |
20160098209 | Leidel et al. | Apr 2016 | A1 |
20160110135 | Wheeler et al. | Apr 2016 | A1 |
20160125919 | Hush | May 2016 | A1 |
20160154596 | Willcock et al. | Jun 2016 | A1 |
20160155482 | La Fratta | Jun 2016 | A1 |
20160163377 | Oh et al. | Jun 2016 | A1 |
20160188250 | Wheeler | Jun 2016 | A1 |
20160196142 | Wheeler et al. | Jul 2016 | A1 |
20160196856 | Tiwari et al. | Jul 2016 | A1 |
20170148504 | Saifuddin et al. | May 2017 | A1 |
Number | Date | Country |
---|---|---|
102141905 | Aug 2011 | CN |
0214718 | Mar 1987 | EP |
2026209 | Feb 2009 | EP |
H0831168 | Feb 1996 | JP |
2009259193 | Mar 2015 | JP |
10-0211482 | Aug 1998 | KR |
10-2010-0134235 | Dec 2010 | KR |
10-2013-0049421 | May 2013 | KR |
2001065359 | Sep 2001 | WO |
2010079451 | Jul 2010 | WO |
2013062596 | May 2013 | WO |
2013081588 | Jun 2013 | WO |
2013095592 | Jun 2013 | WO |
2015187609 | Dec 2015 | WO |
Entry |
---|
Boyd et al., “On the General Applicability of Instruction-Set Randomization”, Jul.-Sep. 2010, (14 pgs.), vol. 7, Issue 3, IEEE Transactions on Dependable and Secure Computing. |
Stojmenovic, “Multiplicative Circulant Networks Topological Properties and Communication Algorithms”, (25 pgs.), Discrete Applied Mathematics 77 (1997) 281-305. |
“4.9.3 MINLOC and MAXLOC”, Jun. 12, 1995, (5pgs), Message Passing Interface Forum 1.1, retrieved from http://www.mpi-forum.org/docs/mpi-1.1/mpi-11-html/node79.html. |
Derby, et al., “A High-Performance Embedded DSP Core with Novel SIMD Features”, Apr. 6-10, 2003, (4 pgs), vol. 2, pp. 301-304, 2003 IEEE International Conference on Accoustics, Speech, and Signal Processing. |
Debnath, Biplob, Bloomflash: Bloom Filter on Flash-Based Storage, 2011 31st Annual Conference on Distributed Computing Systems, Jun. 20-24, 2011, 10 pgs. |
Pagiamtzis, Kostas, “Content-Addressable Memory Introduction”, Jun. 25, 2007, (6 pgs.), retrieved from: http://www.pagiamtzis.com/cam/camintro. |
Pagiamtzis, et al., “Content-Addressable Memory (CAM) Circuits and Architectures: A Tutorial and Survey”, Mar. 2006, (16 pgs.), vol. 41, No. 3, IEEE Journal of Solid-State Circuits. |
International Search Report and Written Opinion for PCT Application No. PCT/US2013/043702, dated Sep. 26, 2013, (11 pgs.). |
Elliot, et al., “Computational RAM: Implementing Processors in Memory”, Jan.-Mar. 1999, (10 pgs.), vol. 16, Issue 1, IEEE Design and Test of Computers Magazine. |
Dybdahl, et al., “Destructive-Read in Embedded DRAM, Impact on Power Consumption,” Apr. 2006, (10 pgs.), vol. 2, Issue 2, Journal of Embedded Computing-Issues in embedded single-chip multicore architectures. |
Kogge, et al., “Processing In Memory: Chips to Petaflops,” May 23, 1997, (8 pgs.), retrieved from: http://www.cs.ucf.edu/courses/cda5106/summer02/papers/kogge97PIM.pdf. |
Draper, et al., “The Architecture of the DIVA Processing-In-Memory Chip,” Jun. 22-26, 2002, (12 pgs.), ICS '02, retrieved from: http://www.isi.edu/˜draper/papers/ics02.pdf. |
Adibi, et al., “Processing-In-Memory Technology for Knowledge Discovery Algorithms,” Jun. 25, 2006, (10 pgs.), Proceeding of the Second International Workshop on Data Management on New Hardware, retrieved from: http://www.cs.cmu.edu/˜damon2006/pdf/adibi06inmemory.pdf. |
U.S. Appl. No. 13/449,082, entitled, “Methods and Apparatus for Pattern Matching,” filed Apr. 17, 2012, (37 pgs.). |
U.S. Appl. No. 13/743,686, entitled, “Weighted Search and Compare in a Memory Device,” filed Jan. 17, 2013, (25 pgs.). |
U.S. Appl. No. 13/774,636, entitled, “Memory as a Programmable Logic Device,” filed Feb. 22, 2013, (30 pgs.). |
U.S. Appl. No. 13/774,553, entitled, “Neural Network in a Memory Device,” filed Feb. 22, 2013, (63 pgs.). |
U.S. Appl. No. 13/796,189, entitled, “Performing Complex Arithmetic Functions in a Memory Device,” filed Mar. 12, 2013, (23 pgs.). |
Extended European Search Report and Written Opinion for related EP Application No. 17835246.4, dated Feb. 17, 2020, 8 pages. |
International Search Report and Written Opinion for related PCT Application No. PCT/US2017/044089, dated Nov. 8, 2017, 11 pages. |
Number | Date | Country | |
---|---|---|---|
20200058347 A1 | Feb 2020 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15222514 | Jul 2016 | US |
Child | 16665648 | US |