The present disclosure relates generally to semiconductor memory apparatuses and methods, and more particularly, to apparatuses and methods for computing reduction and prefix sum operations in memory.
Memory devices are typically provided as internal, semiconductor, integrated circuits in computers or other electronic systems. There are many different types of memory including volatile and non-volatile memory. Volatile memory can require power to maintain its data (e.g., host data, error data, etc.) and includes random access memory (RAM), dynamic random access memory (DRAM), static random access memory (SRAM), synchronous dynamic random access memory (SDRAM), and thyristor random access memory (TRAM), among others. Non-volatile memory can provide persistent data by retaining stored data when not powered and can include NAND flash memory, NOR flash memory, and resistance variable memory such as phase change random access memory (PCRAM), resistive random access memory (RRAM), and magnetoresistive random access memory (MRAM), such as spin torque transfer random access memory (STT RAM), among others.
Electronic systems often include a number of processing resources (e.g., one or more processors), which may retrieve and execute instructions and store the results of the executed instructions to a suitable location. A processor can comprise a number of functional units such as arithmetic logic unit (ALU) circuitry, floating point unit (FPU) circuitry, and/or a combinatorial logic block, for example, which can be used to execute instructions by performing logical operations such as AND, OR, NOT, NAND, NOR, and XOR, and invert (e.g., inversion) logical operations on data (e.g., one or more operands). For example, functional unit circuitry (FUC) may be used to perform arithmetic operations such as addition, subtraction, multiplication, and/or division on operands via a number of logical operations.
A number of components in an electronic system may be involved in providing instructions to the FUC for execution. The instructions may be generated, for instance, by a processing resource such as a controller and/or host processor. Data (e.g., the operands on which the instructions will be executed) may be stored in a memory array that is accessible by the FUC. The instructions and/or data may be retrieved from the memory array and sequenced and/or buffered before the FUC begins to execute instructions on the data. Furthermore, as different types of operations may be executed in one or multiple clock cycles through the FUC, intermediate results of the operations and/or data may also be sequenced and/or buffered.
In many instances, the processing resources (e.g., processor and/or associated FUC) may be external to the memory array, and data can be accessed via a bus between the processing resources and the memory array to execute instructions. Processing performance may be improved in a processor-in-memory (PIM) device, in which a processor may be implemented internal and/or near to a memory (e.g., directly on a same chip as the memory array), which may conserve time and power in processing. However, such PIM devices may still have various drawbacks. For example, such PIM devices may have a limited topology, which can make it difficult to shift (e.g., move) data in the memory. For instance, such PIM devices may only be able to shift bits of data one place at a time. As such, performing logical operations that involve a large amount of data shifting, such as, for instance, reduction and prefix sum operations, using such PIM devices can be a slow process.
The present disclosure includes apparatuses and methods for computing reduction and prefix sum operations in memory. A number of embodiments include processing circuitry configured to compute a reduction operation on data stored in a group of memory cells by splitting the data into a plurality of elements, copying each of the plurality of elements into elements that are wider than before being copied, and performing a logical operation associated with the reduction operation on each of the copied elements.
A number of embodiments of the present disclosure can provide faster computations of reduction and prefix sum operations in memory than previous approaches. For example, a number of embodiments of the present disclosure can use fewer data shifts to compute reduction and prefix sum operations than previous approaches, which can reduce the amount of time needed to compute the reduction and prefix sum operations as compared with previous approaches. For instance, in some previous approaches, reduction and prefix sum operations may be computed by shifting data in one direction (e.g., to the right) by increasing powers of two (e.g., shift the data to the right by one element and perform an addition operation, then shift the data to the right by two elements and perform an addition operation, then shift the data to the right by four elements and perform an addition operation, etc.), which can involve a larger number of data shifts, and therefore take longer, than reduction and prefix sum operations computed in accordance with embodiments of the present disclosure.
Further, in previous approaches, data may be transferred from and/or to the array and sensing circuitry (e.g., via a bus comprising input/output (I/O) lines) to and/or from a processing resource such as a processor, microprocessor, and/or compute engine, which may comprise ALU circuitry and/or other functional unit circuitry configured to perform the appropriate logical operations. However, transferring data to and/or from memory from and/or to such processing resource(s) can involve significant power consumption and time requirements. Even if the processing resource is located on the same chip as the array, significant power can be consumed in moving data out of the array to the compute circuitry (e.g., ALU), which can involve performing a sense line address access (e.g., firing of a column decode signal) in order to transfer data from sense lines onto I/O lines, moving the data to the array periphery, and providing the data to a register in association with performing a comparison operation, for instance.
Furthermore, the circuitry of the processing resource(s) (e.g., compute engine) may not conform to pitch rules associated with a memory array. For example, the cells of a memory array may have a 4F2 or 6F2 cell size, where “F” is a feature size corresponding to the cells. As such, the devices (e.g., logic gates) associated with ALU circuitry of previous PIM systems may not be capable of being formed on pitch with the memory cells, which can affect chip size and/or memory density, for example. A number of embodiments of the present disclosure include sensing circuitry formed on pitch with memory cells of the array and capable of performing operations such as those described herein below to compute reduction and prefix sum operations.
As used herein, the designators “X” and “Y”, particularly with respect to reference numerals in the drawings, indicate that a number of the particular feature so designated can be included with a number of embodiments of the present disclosure. Additionally, as used herein, “a number of” something can refer to one or more of such things. For example, a number of memory cells can refer to one or more memory cells.
The figures herein follow a numbering convention in which the first digit or digits correspond to the drawing figure number and the remaining digits identify an element or component in the drawing. Similar elements or components between different figures may be identified by the use of similar digits. For example, 130 may reference element “30” in
In the example illustrated in
Although system 100 includes a host (e.g., host 110) in the example illustrated in
For clarity, the system 100 has been simplified to focus on features with particular relevance to the present disclosure. The memory array 130 can be a DRAM array, SRAM array, STT RAM array, PCRAM array, TRAM array, RRAM array, NAND flash array, and/or NOR flash array, for instance. The array 130 can comprise memory cells arranged in rows coupled by access lines (which may be referred to herein as word lines or select lines) and columns coupled by sense lines (which may be referred to herein as digit lines or data lines). Although a single array 130 is shown in
The memory device 120 includes address circuitry 142 to latch address signals provided over an I/O bus 156 (e.g., a data bus) through I/O circuitry 144. Address signals are received and decoded by a row decoder 146 and a column decoder 152 to access the memory array 130. Data can be sensed (e.g., read) from memory array 130 by sensing voltage and/or current changes on the data lines using processing circuitry, such as sensing circuitry 150, controller 140, address circuitry 142, row decoder 146, and/or column decoder 152. The sensing circuitry 150 can read and latch a page (e.g., row) of data from the memory array 130. The I/O circuitry 144 can be used for bi-directional data communication with host 110 over the I/O bus 156. The write circuitry 148 is used to program (e.g., write) data to the memory array 130.
Control circuitry (e.g., controller) 140 decodes signals provided by control bus 154 from the host 110. These signals can include chip enable signals, write enable signals, and address latch signals that are used to control operations performed on the memory array 130, including data read, data write, and data erase operations. In various embodiments, the control circuitry 140 is responsible for executing instructions from the host 110. The control circuitry 140 can be a state machine, a sequencer, or some other type of controller.
The controller 140 can include a shift controller 170 that can control signals provided to, for instance, shift circuitry (e.g., shift circuitry 223 described in connection with
Examples of the sensing circuitry 150 are described further below in association with
In a number of embodiments, the sensing circuitry (e.g., 150) can be used to compute reduction and/or prefix sum operations on data stored in array 130 without performing (e.g., transferring data via) a sense line address access (e.g., without firing a column decode signal). As such, a reduction and/or prefix sum operation can be computed using, and within, sensing circuitry 150 rather than (or in association with) being computed by processing resources external to the sensing circuitry 150 (e.g., by a processor associated with host 110 and/or other processing circuitry, such as ALU circuitry, located on device 120 (e.g., on control circuitry 140 or elsewhere)). However, embodiments of the present disclosure are not so limited. For example, in a number of embodiments, reduction and/or prefix sum operations in accordance with the present disclosure can be computed by processing resources (e.g., processing circuitry) external to sensing circuitry 150. For instance, reduction and/or prefix sum operations in accordance with the present disclosure can be computed by an array of processors, such as bit-serial processors or nibble-serial processors, among others, connected in a linear or other type of connection topology. A prefix sum operation can refer to an operation that computes all the partial sums of an array of numbers (e.g., that computes sums over all prefixes of the array), and can include, for example, a segmented prefix sum operation. A reduction operation can refer to an operation that reduces an input array to a single element, which is the sum of all the elements in the input array. In some instances, a reduction operation can be a part of a prefix sum operation. Examples of reduction and prefix sum operations will be further described herein.
In various previous approaches, data associated with a reduction or prefix sum operation (e.g., the data on which the reduction or prefix sum operation is to be computed), for instance, would be read from memory via sensing circuitry and provided to external ALU circuitry via I/O lines (e.g., via local and/or global I/O lines) and/or busses. The external ALU circuitry could include a number of registers, and would perform the computation and transfer the result back to the array via the local I/O lines and/or busses. In contrast, in a number of embodiments of the present disclosure, sensing circuitry (e.g., 150) is configured to compute a reduction and/or prefix sum operation on data stored in memory cells in memory array 130, including storing the result of the computation back to the array 130, without transferring the data via (e.g., enabling) a local I/O line and/or bus coupled to the sensing circuitry, which can be formed on pitch with the memory cells of the array. Enabling an I/O line can include enabling (e.g., turning on) a transistor having a gate coupled to a decode signal (e.g., a column decode signal) and a source/drain coupled to the I/O line. Embodiments are not so limited. For instance, in a number of embodiments, the sensing circuitry (e.g., 150) can be used to compute reduction and/or prefix sum operations without enabling column decode lines of the array; however, the local I/O line(s) may be enabled in order to transfer a result to a suitable location other than back to the array (e.g., to an external register and/or external hardware).
As such, in a number of embodiments, circuitry (e.g., registers and/or an ALU) external to array 130 and sensing circuitry 150 may not be needed to compute reduction and/or prefix sum operations, as the sensing circuitry 150 can perform the appropriate computations and logical operations without the use of an external processing resource. Therefore, the sensing circuitry 150 may be used to complement and/or to replace, at least to some extent, such an external processing resource (or at least the bandwidth of such an external processing resource). However, in a number of embodiments, the sensing circuitry 150 may be used to perform logical operations (e.g., to execute instructions) in addition to logical operations performed by an external processing resource (e.g., host 110). For instance, host 110 and/or sensing circuitry 150 may be limited to performing only certain logical operations and/or a certain number of logical operations.
In the example illustrated in
In a number of embodiments, the memory cells of array 230 may be destructive read memory cells (e.g., reading the data stored in the cell destroys the data such that the data originally stored in the cell is refreshed after being read). The cells of the memory array 230 are arranged in rows coupled by word lines (e.g., 204-X (Row X), 204-Y (Row Y)) and columns coupled by pairs of complementary data lines DIGIT(n−1)/DIGIT(n−1)_, DIGIT(n)/DIGIT(n)_, DIGIT(n+1)/DIGIT(n+1)_. The individual data lines corresponding to each pair of complementary data lines can also be referred to as data lines 205-1 (D) and 205-2 (DJ respectively. Although only three pairs of complementary data lines are shown in
Memory cells can be coupled to different data lines and/or word lines. For example, as shown in
The memory array 230 is coupled to sensing circuitry 250 in accordance with a number of embodiments of the present disclosure. For example, the sensing circuitry 250 can be coupled to the array of memory cells by one or more data lines (e.g., complementary data lines D and DJ through shift circuitry 223. As such, the sensing circuitry can be configured and/or operated to shift data stored in a memory cell coupled to a first data line to a memory cell coupled to a second data line via (e.g., through) shift circuitry 223.
The sensing circuitry 250, including the sense amplifier 206 and the compute component 231, can be coupled to the array 230 via the shift circuitry 223. The shift circuitry 223 can include a pair of isolation transistors 221-1 and 221-2 having gates coupled to a first control signal 222 (e.g., NORM) that, when activated, enables (e.g., turns on) the isolation transistors 221-1 and 221-2 to connect the sensing circuitry 250 (e.g., sense amplifier 206 and the compute component 231) to a column of memory cells with which the sensing circuitry 250 is associated. According to various embodiments, conduction of isolation transistors 221-1 and 221-2 can be referred to as a “normal” configuration.
As illustrated in
Although shift circuitry 223 is shown in
According to some embodiments, shift circuitry 223 can be configured to connect the sensing circuitry 250 to a non-adjacent pair of complementary data lines. According to various embodiments, shift circuitry 223 can be configured to connect the sensing circuitry 250 to a pair of complementary data lines from among a plurality of pairs of complementary data lines (e.g., selected from among adjacent pairs of complementary data lines to the left and right of the pair of complementary data lines to which isolation transistors 221-1 and 221-2 are coupled).
In the example illustrated in
The sense amplifier 306 can be operated to determine a data value (e.g., logic state) stored in a selected memory cell. The sense amplifier 306 can comprise a cross coupled latch, which can be referred to herein as a primary latch. Embodiments are not limited to the example sense amplifier 306. As an example, the sense amplifier 306 can be current-mode sense amplifier and/or single-ended sense amplifier (e.g., sense amplifier coupled to one data line). Also, embodiments of the present disclosure are not limited to a folded data line architecture.
According to various embodiments, the compute component 331 can comprise a latch, which can be referred to herein as a secondary latch, and which can serve as, and be referred to as, an accumulator. The secondary latch can be a static latch, such as a cross coupled latch, and/or a dynamic latch.
In the example illustrated in
In the example illustrated in
The transistors 307-1 and 307-2 can be referred to as pass transistors, which can be enabled via respective signals 311-1 (Passd) and 311-2 (Passdb) in order to pass the voltages or currents on the respective data lines D and D_to the inputs of the cross coupled latch comprising transistors 308-1, 308-2, 309-1, and 309-2 (e.g., the input of the secondary latch). In this example, the second source/drain region of transistor 307-1 is directly coupled to a first source/drain region of transistors 308-1 and 309-1 as well as to the gates of transistors 308-2 and 309-2. Similarly, the second source/drain region of transistor 307-2 is directly coupled to a first source/drain region of transistors 308-2 and 309-2 as well as to the gates of transistors 308-1 and 309-1.
A second source/drain region of transistor 308-1 and 308-2 is commonly directly coupled to a negative control signal 312-1 (Accumb). A second source/drain region of transistors 309-1 and 309-2 both share a common connection direct to a positive control signal 312-2 (Accum). The Accum signal 312-2 can be a supply voltage (e.g., VDD) and the Accumb signal can be a reference voltage (e.g., ground). Activating signals 312-1 and 312-2 enables the cross coupled latch comprising transistors 308-1, 308-2, 309-1, and 309-2 corresponding to the secondary latch.
The enabled sense amplifier pair can operate to amplify a differential voltage between common node 317-1 and common node 317-2 such that node 317-1 is driven to one of the Accum signal voltage and the Accumb signal voltage (e.g., to one of VDD and ground), and node 317-2 is driven to the other of the Accum signal voltage and the Accumb signal voltage.
The compute component 331 also includes inverting transistors 314-1 and 314-2 having a first source/drain region directly coupled to the respective data lines D and D_. A second source/drain region of the transistors 314-1 and 314-2 is directly coupled to a first source/drain region of pull-down transistors 316-1 and 316-2, respectively. The gates of transistors 314-1 and 314-2 are coupled to a signal 313 (InvD). The gate of transistor 316-1 is coupled to the common node 317-1 to which the gate of transistor 308-2, the gate of transistor 309-2, and the first source/drain region of transistor 308-1 are also directly coupled. In a complementary fashion, the gate of transistor 316-2 is directly coupled to the common node 317-2 to which the gate of transistor 308-1, the gate of transistor 309-1, and the first source/drain region of transistor 308-2 are also directly coupled. As such, activating signal InvD serves to invert the data value stored in the secondary latch and drives the inverted value onto data lines 305-1 D and D.
Sensing circuitry 350 (e.g., sense amplifier 306 and compute component 331) can be used to compute reduction and prefix sum operations on data stored in the memory cells (e.g., in a group of the memory cells) of arrays 130 and/or 230. For example, sensing circuitry 350 can be used to compute a reduction operation on the data stored in a group of the memory cells by sensing the data stored in the group of memory cells and splitting the data (e.g., vector) into a plurality of elements (e.g., blocks) by, for example, taking every nth bit of data for some n greater than or equal to two. For example, the data can be split into even and odd indexed elements. The group of memory cells may be coupled to a single (e.g., only one) sense line, or to a plurality of sense lines. For example, the data may include data stored in a single column of memory cells, or data stored in a plurality of columns of memory cells. Further, the group of memory cells may be coupled to a plurality of access lines. For example, the data may include data stored in a plurality of rows of memory cells. Sensing circuitry 350 may split the data using shift circuitry 223 previously described in connection with
Sensing circuitry 350 (e.g., via shift circuitry 223) can then copy each of the plurality of elements (e.g., the even indexed elements and the odd indexed elements) into elements that are twice as wide and half as tall than before being copied. For example, the even indexed elements can be copied to memory cells coupled to twice as many sense lines and half as many access lines, and the odd indexed elements can be copied to memory cells coupled to twice as many sense lines and half as many access lines. Accordingly, the data of the elements may be copied into double the number of columns and half the number of rows. The elements can be copied (e.g., programmed) to the same group of memory cells, an additional (e.g., different) group of memory cells, or a partially overlapping group of memory cells. Moreover, the elements can be copied to the same part of the memory array, a different part of the memory array, or a partially overlapping part of the memory array.
Sensing circuitry 350 can then perform a logical operation associated with the reduction operation on each of the copied elements (e.g., on the copied even indexed elements and the copied odd indexed elements) to combine the copied elements (e.g., to recombine the even and odd indexed elements into a single vector). The logical operation associated with the reduction operation can be, for example, the underlying operation of the reduction operation. For instance, if the reduction operation is an addition operation, the logical operation can be an addition operation, and if the reduction operation is a multiplication operation, the logical operation can be a multiplication operation. The reduction operation (e.g., the logical operation) can also be a minimum or maximum location operation, an OR operation, an AND operation, or an XOR operation, among others. The operation may apply to corresponding elements in a plurality of vectors (e.g., the operation may combine the first element of each vector, the second element of each vector, etc.)
In some embodiments, sensing circuitry 350 can output the result of the reduction operation to external hardware. An example of a computation of a reduction operation will be described in connection with
As an additional example, sensing circuitry 350 can compute a prefix sum operation on the data stored in a group of the memory cells by sensing the data stored in the group of memory cells and splitting the data into a plurality of elements in a manner analogous to the reduction operation. Sensing circuitry 350 can then copy each of the plurality of elements into elements twice as wide and half as tall, in a manner analogous to the reduction operation. Sensing circuitry 350 can then perform a logical operation (e.g., addition) associated with the prefix sum operation on the copied elements to combine the copied elements. The operation may apply to corresponding elements in a plurality of vectors, in a manner analogous to the reduction operation.
The logical operation can be performed using, for example, algorithms based on the structure of an adder such as a Kogge-Stone adder or a Brent-Kung adder, or a recursive call to one of the prefix sum algorithms described herein. Further, the prefix sum operation can be computed using an identity element, or without using an identity element.
As an example in which the prefix sum operation is computed using an identity element, sensing circuitry 350 can shift (e.g., via shift circuitry 223) the output of the logical operation by one element, and fill in the identity element in the element of the output of the logical operation that is empty as a result of the shift. The direction the output of the logical operation is shifted may be based on the direction of the prefix sum operation. For example, if the prefix sum operation is a forward prefix sum operation, the output of the logical operation may be shifted right by one element, and the identity element may be filled in the left-most (e.g., first) element of the output. If the prefix sum operation is a backward prefix sum operation, the output of the logical operation may be shifted left by one element, and the identity element may be filled in the right-most (e.g., last) element of the output.
Sensing circuitry 350 can then perform another logical operation (e.g., addition) on the shifted output and one of the previously copied elements (e.g., the even indexed elements) to combine the shifted output and the one of the copied elements, and pack the output of this logical operation (e.g., the combined shifted output and copied even indexed elements) into elements half as wide and twice as tall as before the logical operation. An example of such a computation of a prefix sum operation will be described in connection with
As an example in which the prefix sum operation is computed without using an identity element, sensing circuitry 350 can shift (e.g., via shift circuitry 223) the output of the logical operation by one element, and fill in an arbitrary value in the element of the output of the logical operation that is empty as a result of the shift. The direction the output of the logical operation is shifted may be based on the direction of the prefix sum operation, in a manner analogous to that previously described in the example in which the prefix sum operation is computed using an identity element.
Sensing circuitry 350 can then perform another logical operation (e.g., addition) on the shifted output and one of the previously copied elements (e.g., the even indexed elements) to combine the shifted output and the one of the copied elements, and store the output of this logical operation (e.g., the combined shifted output and copied even indexed elements) in a temporary vector T. The left-most (e.g., first) element of temporary vector T can then be replaced with the first element of the previously copied even indexed elements. This can be done by, for example, an if-then-else operation, hardware multiplexor, masked vector operations, a vector merge operation, or any other process capable of performing such a replacement operation. The output of this replacement operation can then be packed into elements half as wide as twice as tall as before the logical operation.
Although embodiments of the present disclosure have described copying blocks of data into elements twice as wide and half as tall, embodiments of the present disclosure are not limited to a particular fraction into which the blocks of data may be copied. For example, embodiments of the present disclosure may include copying blocks of data into other fractional elements. For example, vertical operations can be performed on the blocks of data until the data has been spread out sufficiently for horizontal operations to be performed on the data. Upon the data being spread out sufficiently for horizontal operations to be performed, the data can be gathered into a single row, and a horizontal reduction or prefix sum operation can be computed on the single row of data. As such, instead of performing block packing operations iteratively, the block packing operations can be saved up and performed at once after the data has been spread out sufficiently. As an example, for 16-bit data, the data may be converted to a horizontal format only once, with only every 16th element used for the rest of the computation operation. The result of the horizontal reduction or prefix sum operation can then be converted back directly into vertical 16-bit elements later in the computing operation.
Reduction operation 460 can be computed on data 462 illustrated in
As shown in
As shown in
As shown in
As shown in
As shown in
As shown in
As shown in
As shown in
As shown in
As shown in
As shown in
The below example of code (e.g., pseudocode representing executable instructions) can be used for performing additions of block-format data as described in connection with
1. Create two new arrays gen and prop, each of the same size as a and b.
2. Set gen[i] to the bitwise AND of a[i] and b[i] for each i in 0..vsize-1.
3. Set prop[i] to the bitwise XOR of a[i] and b[i] for each i in 0..vsize-1.
4. For each power of 2 from 1 to the greatest power of 2 less than or equal to hsize*vsize:
In the above example, shifting a block means to shift its numerical value (e.g., sequence of bits) right, which means the bits need to wrap between rows. To shift a vsize by hsize block named “a” right by n bits:
1. Create an array “out” of the same size as a, initializing it to zero.
2. Create a mask with (n mod hsize) 0's and hsize−(n mod hsize) 1's, repeating that pattern for the width of the row.
3. If n mod hsize is not equal to 0, repeat the following for each i from 0..vsize-1:
The below example of code (e.g., pseudocode) can be used for performing splitting and/or packing operations as described in connection with
1. Create a mask containing blocks of hsize ones and zeros alternating (for example, for length 16 and hsize 4, the mask is 1111000011110000) from LSB to MSB.
2. Bitwise AND both rows with the mask and its logical complement (NOT); this produces four rows as results: row0_and_mask, row1_and_mask, row0_and_not_mask, row1_and_not_mask.
3. Shift row1_and_mask right by hsize bits, then bitwise OR in row0_and_mask; this produces the first row of the result.
4. Shift row0_and_not_mask right by hsize bits, then bitwise OR in row1_and_not_mask; this produces the second row of the result.
Widening data from block size m×n to block size m/2×2n can use the above transpose example on adjacent pairs of rows from the input data (e.g., 0 and 1, 2 and 3, etc.) with hsize=n. Narrowing the data back down from m/2×2n to m×n can use hsize=n as well on corresponding rows of the input data being combined into the narrower format.
Although not shown, memory cells can coupled to the pairs of complementary sense lines 605-1 and 605-2 (e.g., columns). The memory cells can be, for example, the memory cells of memory array 230 previously described in connection with
As shown in
In the example illustrated in
The gates of the pass gates 607-1 and 607-2 can be controlled by a logical operation selection logic signal, Pass. For example, an output of the logical operation selection logic 636 can be coupled to the gates of the pass gates 607-1 and 607-2, as shown in
The sensing circuitry 650 shown in
According to various embodiments, the logical operation selection logic 636 can include four logic selection transistors: logic selection transistor 663 coupled between the gates of the swap transistors 643 and a TF signal control line, logic selection transistor 653 coupled between the gates of the pass gates 607-1 and 607-2 and a TT signal control line, logic selection transistor 655 coupled between the gates of the pass gates 607-1 and 607-2 and a FT signal control line, and logic selection transistor 667 coupled between the gates of the swap transistors 643 and a FF signal control line. Gates of logic selection transistors 663 and 653 are coupled to the true sense line through isolation transistor 651-1 (having a gate coupled to an ISO signal control line). Gates of logic selection transistors 667 and 655 are coupled to the complementary sense line through isolation transistor 651-2 (also having a gate coupled to an ISO signal control line).
Data values present on the pair of complementary sense lines 605-1 and 605-2 can be loaded into the compute component 631 via the pass gates 607-1 and 607-2. When the pass gates 607-1 and 607-2 are OPEN, data values on the pair of complementary sense lines 605-1 and 605-2 are passed to the compute component 631 and thereby loaded into the loadable shift register. The data values on the pair of complementary sense lines 605-1 and 605-2 can be the data value stored in the sense amplifier 606 when the sense amplifier is enabled (e.g., fired). The logical operation selection logic signal, Pass, is activated to OPEN (e.g., turn on) the pass gates 607-1 and 607-2.
The ISO, TF, TT, FT, and FF control signals can operate to select a logical operation to implement based on the data value (“B”) in the sense amplifier 606 and the data value (“A”) in the compute component 631 (e.g., as used herein, the data value stored in a latch of a sense amplifier is referred to as a “B” data value, and the data value stored in a latch of a compute component is referred to as an “A” data value). In particular, the ISO, TF, TT, FT, and FF control signals are configured to select the logical operation (e.g., function) to implement independent from the data value present on the pair of complementary sense lines 605-1 and 605-2 (although the result of the implemented logical operation can be dependent on the data value present on the pair of complementary sense lines 605-1 and 605-2. For example, the ISO, TF, TT, FT, and FF control signals select the logical operation to implement directly since the data value present on the pair of complementary sense lines 605-1 and 605-2 is not passed through logic to operate the gates of the pass gates 607-1 and 607-2.
Additionally,
As an example, the logical operation selection logic signal Pass can be activated (e.g., high) to OPEN (e.g., turn on) the pass gates 607-1 and 607-2 when the ISO control signal line is activated and either the TT control signal is activated (e.g., high) with the data value on the true sense line being “1” or the FT control signal is activated (e.g., high) with the data value on the complement sense line being “1.”
The data value on the true sense line being a “1” OPENs logic selection transistors 653 and 663. The data value on the complementary sense line being a “1” OPENs logic selection transistors 655 and 667. If the ISO control signal or either the respective TT/FT control signal or the data value on the corresponding sense line (e.g., sense line to which the gate of the particular logic selection transistor is coupled) is not high, then the pass gates 307-1 and 307-2 will not be OPENed by a particular logic selection transistor.
The logical operation selection logic signal Pass* can be activated (e.g., high) to OPEN (e.g., turn on) the swap transistors 643 when the ISO control signal line is activated and either the TF control signal is activated (e.g., high) with data value on the true sense line being “1,” or the FF control signal is activated (e.g., high) with the data value on the complement sense line being “1.” If either the respective control signal or the data value on the corresponding sense line (e.g., sense line to which the gate of the particular logic selection transistor is coupled) is not high, then the swap transistors 643 will not be OPENed by a particular logic selection transistor.
The sensing circuitry 650 illustrated in
Although not shown in
As noted above, the compute components 631 can comprise a loadable shift register. In this example, each compute component 631 is coupled to a corresponding pair of complementary data lines 605-1/605-2, with a node ST2 being coupled to the particular data line (e.g., DIGIT(n)) communicating a “true” data value and with node SF2 being coupled to the corresponding complementary data line (e.g., DIGIT(n)_) communicating the complementary data value (e.g., “false” data value).
In this example, the loadable shift register comprises a first right-shift transistor 681 of a particular compute component 631 having a gate coupled to a first right-shift control line 683 (e.g., PHASE 1R), and a second right-shift transistor 691 of the particular compute component 631 having a gate coupled to a second right-shift control line 685 (e.g., PHASE 2R). Node ST2 of the particular control component is coupled to an input of a first inverter 687, whose output (e.g., node SF1) is coupled to a first source/drain region of transistor 691. The second source/drain region of transistor 691 is coupled to the input (e.g., node SF2) of a second inverter 689. The output (e.g., node ST1) of inverter 689 is coupled to a first source/drain region of transistor 681, and a second source/drain region of transistor 681 the particular compute component 631 is coupled to an input (e.g., node ST2) of a first inverter 687 of an adjacent compute component 631. The loadable shift register shown in
In operation, a data value on a pair of complementary data lines (e.g., 605-1/605-2) can be loaded into a corresponding compute component 631 (e.g., by operating logical operation selection logic as described above). As an example, a data value can be loaded into a compute component 631 via overwriting of the data value currently stored in the compute component 631 with the data value stored in the corresponding sense amplifier 606. Alternatively, a data value may be loaded into a compute component by deactivating the control lines 683, 685, 697, and 699.
Once a data value is loaded into a compute component 631, the “true” data value is separated from the complement data value by the first inverter 687. Shifting data to the right (e.g., to an adjacent compute component 631) can include alternating operation of the first right-shift transistor 681 and the second right-shift transistor 691, for example, via the PHASE 1R and PHASE 2R control signals being periodic signals that go high out of phase from one another (e.g., non-overlapping alternating square waves 180 out of phase). The transistor 693 can be turned on to latch the shifted data value.
An example of shifting data left via the shift register shown in
Embodiments of the present disclosure are not limited to the shifting capability described in association with the compute components 631. For example, a number of embodiments and include shift circuitry in addition to and/or instead of the shift circuitry described in association with a loadable shift register.
The sensing circuitry in
In a number of examples, the sense amplifier 606 and the compute component 631 can be in at least one of two states associated with the first mode and the second mode. As used herein, a state of a sense amplifier 606 and/or the compute component 631 describes a transfer of data between the sense amplifier 606 and/or the compute component 631. The state of the sense amplifier 606 and the compute component 631 can also be described as the state of a sensing component. The state of a sensing component can be based on whether the sense amplifier 606 is in an equilibration state or is storing a data value (e.g., logic “0” or logic “1”). For example, a sense amplifier can be configured to be in an initial state, wherein the initial state is one of an equilibration state and a data storage state. An equilibration state includes the sense amplifier 606 being in an equilibration state. A data storage state includes the sense amplifiers 606 storing a data value. As used herein, a data value can be referred to as a bit and/or a digit value. Data can be transferred from a compute component 631 to a sense amplifier 606 in response to enabling a pass gate (e.g., activating the PASS and/or PASS* control signals via the TF 663, TT 653, FT 655, and/or FF 667 control signals that are referred to herein as a logical operation selection logic) and the sense amplifier 606 being in a equilibration state. Data can be transferred from a sense amplifier 606 to a compute component 631 in response to enabling the pass gate (e.g., activating the PASS and/or PASS* control signals via the TF 663, TT 653, FT 655, and/or FF 667 control signals that are referred to herein as a logical operation selection logic) and the sense amplifier 606 being in a data storage state. The direction of the transfer of data between the sense amplifier 606 and the compute component 631 is determined by whether the sense amplifier 606 is in an equilibration state or stores a data value before the PASS and/or PASS* control signals are activated and by a particular operation selected via the logical operation selection logic (e.g., TF 663, TT 653, FT 655, and FF 667 control signals).
For example, if the sense amplifier 606 is equilibrated and the PASS and/or PASS* control signals are activated to provide a conduction path (e.g., electrical continuity) between the sense amplifier 606 and the compute component 631, then a data value stored in the compute component 631 can be transferred from the compute component 631 to the sense amplifier 606.
If the sense amplifier 606 is configured to store a first bit (e.g., first data value) and the PASS and/or PASS* control signals are activated to provide a conduction path between the sense amplifier 606 and the compute component 631, then a second bit (e.g., second data value) that is stored in the compute component 631 before the activation of the PASS and/or PASS* control signals can be replaced by the first bit and the sense amplifier 606 retains the first bit. Furthermore, a number of operations can be performed using the first bit and the second bit using the logical operation selection logic and the result of the operation can be stored in the compute component 631.
Using an equilibration signal to direct the transfer of data between the sense amplifier 606 and the compute component 631 can provide the ability to selectively perform an operation in sense amplifiers that are not equilibrated without performing the operation in sense amplifiers that are equilibrated. For example, a PASS and/or a PASS* control signal can be activated in a plurality of sensing components to move data between a first group of a plurality of sense amplifiers that are equilibrated and a first group of a plurality of compute components. The PASS and/or PASS* control signals can also be activated to move data between a second group of the plurality of sense amplifiers and a second group of the plurality of component components that are not equilibrated to selectively perform an operation in a second group of sense components while not performing the operation on a first group of sense components.
Logic Table 601-1 illustrated in
The logic tables illustrated in
Via selective control of the state of the pass gates 607-1 and 607-2 and the swap transistors 643, each of the three columns of the upper portion of Logic Table 601-1 can be combined with each of the three columns of the lower portion of Logic Table 601-1 to provide 3×3=9 different result combinations, corresponding to nine different logical operations, as indicated by the various connecting paths shown at 627. The nine different selectable logical operations that can be implemented by the sensing circuitry 650 are summarized in Logic Table 601-2 illustrated in
The columns of Logic Table 601-2 illustrated in
For example, the results for the values of FF, FT, TF, and TT of “0000” are summarized as “A” since the result (initially stored in the compute component after the sense amplifier fires) is the same as the starting value in the compute component. Other columns of results are similarly annotated in row 634, where “A*B” intends A AND B, “A+B” intends A OR B, and “AXB” intends A XOR B. By convention, a bar over a data value or a logical operation indicates an inverted value of the quantity shown under the bar. For example, AXB bar intends not A XOR B, which is also A XNOR B.
Although specific embodiments have been illustrated and described herein, those of ordinary skill in the art will appreciate that an arrangement calculated to achieve the same results can be substituted for the specific embodiments shown. This disclosure is intended to cover adaptations or variations of a number of embodiments of the present disclosure. It is to be understood that the above description has been made in an illustrative fashion, and not a restrictive one. Combination of the above embodiments, and other embodiments not specifically described herein will be apparent to those of ordinary skill in the art upon reviewing the above description. The scope of a number of embodiments of the present disclosure includes other applications in which the above structures and methods are used. Therefore, the scope of a number of embodiments of the present disclosure should be determined with reference to the appended claims, along with the full range of equivalents to which such claims are entitled.
In the foregoing Detailed Description, some features are grouped together in a single embodiment for the purpose of streamlining the disclosure. This method of disclosure is not to be interpreted as reflecting an intention that the disclosed embodiments of the present disclosure have to use more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive subject matter lies in less than all features of a single disclosed embodiment. Thus, the following claims are hereby incorporated into the Detailed Description, with each claim standing on its own as a separate embodiment.
This application is a Continuation of U.S. application Ser. No. 15/810,880 filed Nov. 13, 2017, which issues as U.S. Pat. No. 10,261,691 on Apr. 16, 2019, which is a Continuation of U.S. application Ser. No. 14/874,068 filed Oct. 2, 2015, which issued as U.S. Pat. No. 9,836,218 on Dec. 5, 2017, which claims the benefit of U.S. Provisional Application No. 62/059,256, filed Oct. 3, 2014, the contents of which are incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
4380046 | Fung | Apr 1983 | A |
4435792 | Bechtolsheim | Mar 1984 | A |
4435793 | Ochii | Mar 1984 | A |
4727474 | Batcher | Feb 1988 | A |
4843264 | Galbraith | Jun 1989 | A |
4939687 | Hartley et al. | Jul 1990 | A |
4958378 | Bell | Sep 1990 | A |
4977542 | Matsuda et al. | Dec 1990 | A |
5023838 | Herbert | Jun 1991 | A |
5034636 | Reis et al. | Jul 1991 | A |
5058072 | Kashimura | Oct 1991 | A |
5121197 | Yamada et al. | Jun 1992 | A |
5201039 | Sakamura | Apr 1993 | A |
5210850 | Kelly et al. | May 1993 | A |
5253308 | Johnson | Oct 1993 | A |
5276643 | Hoffman et al. | Jan 1994 | A |
5325519 | Long et al. | Jun 1994 | A |
5367488 | An | Nov 1994 | A |
5379257 | Matsumura et al. | Jan 1995 | A |
5386379 | Ali-Yahia et al. | Jan 1995 | A |
5398213 | Yeon et al. | Mar 1995 | A |
5440482 | Davis | Aug 1995 | A |
5446690 | Tanaka et al. | Aug 1995 | A |
5473576 | Matsui | Dec 1995 | A |
5475631 | Parkinson et al. | Dec 1995 | A |
5481500 | Reohr et al. | Jan 1996 | A |
5485373 | Davis et al. | Jan 1996 | A |
5506811 | McLaury | Apr 1996 | A |
5615404 | Knoll et al. | Mar 1997 | A |
5638128 | Hoogenboom | Jun 1997 | A |
5638317 | Tran | Jun 1997 | A |
5654936 | Cho | Aug 1997 | A |
5678021 | Pawate et al. | Oct 1997 | A |
5724291 | Matano | Mar 1998 | A |
5724366 | Furutani | Mar 1998 | A |
5751987 | Mahant-Shetti et al. | May 1998 | A |
5787458 | Miwa | Jul 1998 | A |
5854636 | Watanabe et al. | Dec 1998 | A |
5867429 | Chen et al. | Feb 1999 | A |
5870504 | Nemoto et al. | Feb 1999 | A |
5915084 | Wendell | Jun 1999 | A |
5935263 | Keeth et al. | Aug 1999 | A |
5978827 | Ichikawa | Nov 1999 | A |
5986942 | Sugibayashi | Nov 1999 | A |
5991209 | Chow | Nov 1999 | A |
5991785 | Alidina et al. | Nov 1999 | A |
6005799 | Rao | Dec 1999 | A |
6009020 | Nagata | Dec 1999 | A |
6092186 | Betker et al. | Jul 2000 | A |
6122211 | Morgan et al. | Sep 2000 | A |
6125071 | Kohno et al. | Sep 2000 | A |
6134164 | Lattimore et al. | Oct 2000 | A |
6147514 | Shiratake | Nov 2000 | A |
6151244 | Fujino et al. | Nov 2000 | A |
6157578 | Brady | Dec 2000 | A |
6163862 | Adams et al. | Dec 2000 | A |
6166942 | Vo et al. | Dec 2000 | A |
6172918 | Klidaka | Jan 2001 | B1 |
6175514 | Henderson | Jan 2001 | B1 |
6181698 | Hariguchi | Jan 2001 | B1 |
6208544 | Beadle et al. | Mar 2001 | B1 |
6226215 | Yoon | May 2001 | B1 |
6301153 | Takeuchi et al. | Oct 2001 | B1 |
6301164 | Manning et al. | Oct 2001 | B1 |
6304477 | Naji | Oct 2001 | B1 |
6389507 | Sherman | May 2002 | B1 |
6418498 | Martwick | Jul 2002 | B1 |
6466499 | Blodgett | Oct 2002 | B1 |
6510098 | Taylor | Jan 2003 | B1 |
6563754 | Lien et al. | May 2003 | B1 |
6578058 | Nygaard | Jun 2003 | B1 |
6731542 | Le et al. | May 2004 | B1 |
6754746 | Leung et al. | Jun 2004 | B1 |
6768679 | Le et al. | Jul 2004 | B1 |
6807614 | Chung | Oct 2004 | B2 |
6816422 | Hamade et al. | Nov 2004 | B2 |
6819612 | Achter | Nov 2004 | B1 |
6894549 | Eliason | May 2005 | B2 |
6943579 | Hazanchuk et al. | Sep 2005 | B1 |
6948056 | Roth | Sep 2005 | B1 |
6950771 | Fan et al. | Sep 2005 | B1 |
6950898 | Merritt et al. | Sep 2005 | B2 |
6956770 | Khalid et al. | Oct 2005 | B2 |
6961272 | Schreck | Nov 2005 | B2 |
6965648 | Smith et al. | Nov 2005 | B1 |
6985394 | Kim | Jan 2006 | B2 |
6987693 | Cernea et al. | Jan 2006 | B2 |
7020017 | Chen et al. | Mar 2006 | B2 |
7028170 | Saulsbury | Apr 2006 | B2 |
7045834 | Tran et al. | May 2006 | B2 |
7054178 | Shiah et al. | May 2006 | B1 |
7061817 | Raad et al. | Jun 2006 | B2 |
7079407 | Dimitrelis | Jul 2006 | B1 |
7173857 | Kato et al. | Feb 2007 | B2 |
7187585 | Li et al. | Mar 2007 | B2 |
7196928 | Chen | Mar 2007 | B2 |
7260565 | Lee et al. | Aug 2007 | B2 |
7260672 | Garney | Aug 2007 | B2 |
7372715 | Han | May 2008 | B2 |
7400532 | Aritome | Jul 2008 | B2 |
7406494 | Magee | Jul 2008 | B2 |
7447720 | Beaumont | Nov 2008 | B2 |
7454451 | Beaumont | Nov 2008 | B2 |
7457181 | Lee et al. | Nov 2008 | B2 |
7535769 | Cernea | May 2009 | B2 |
7546438 | Chung | Jun 2009 | B2 |
7562198 | Noda et al. | Jul 2009 | B2 |
7574466 | Beaumont | Aug 2009 | B2 |
7602647 | Li et al. | Oct 2009 | B2 |
7663928 | Tsai et al. | Feb 2010 | B2 |
7685365 | Rajwar et al. | Mar 2010 | B2 |
7692466 | Ahmadi | Apr 2010 | B2 |
7752417 | Manczak et al. | Jul 2010 | B2 |
7791962 | Noda et al. | Sep 2010 | B2 |
7796453 | Riho et al. | Sep 2010 | B2 |
7805587 | Van Dyke et al. | Sep 2010 | B1 |
7808854 | Takase | Oct 2010 | B2 |
7827372 | Bink et al. | Nov 2010 | B2 |
7869273 | Lee et al. | Jan 2011 | B2 |
7898864 | Dong | Mar 2011 | B2 |
7924628 | Danon et al. | Apr 2011 | B2 |
7937535 | Ozer et al. | May 2011 | B2 |
7957206 | Bauser | Jun 2011 | B2 |
7979667 | Allen et al. | Jul 2011 | B2 |
7996749 | Ding et al. | Aug 2011 | B2 |
8042082 | Solomon | Oct 2011 | B2 |
8045391 | Mohklesi | Oct 2011 | B2 |
8059438 | Chang et al. | Nov 2011 | B2 |
8095825 | Hirotsu et al. | Jan 2012 | B2 |
8117462 | Snapp et al. | Feb 2012 | B2 |
8164942 | Gebara et al. | Apr 2012 | B2 |
8208328 | Hong | Jun 2012 | B2 |
8213248 | Moon et al. | Jul 2012 | B2 |
8223568 | Seo | Jul 2012 | B2 |
8238173 | Akerib et al. | Aug 2012 | B2 |
8274841 | Shimano et al. | Sep 2012 | B2 |
8279683 | Klein | Oct 2012 | B2 |
8310884 | Iwai et al. | Nov 2012 | B2 |
8332367 | Bhattacherjee et al. | Dec 2012 | B2 |
8339824 | Cooke | Dec 2012 | B2 |
8339883 | Yu et al. | Dec 2012 | B2 |
8347154 | Bahali et al. | Jan 2013 | B2 |
8351292 | Matano | Jan 2013 | B2 |
8356144 | Hessel et al. | Jan 2013 | B2 |
8417921 | Gonion et al. | Apr 2013 | B2 |
8462532 | Argyres | Jun 2013 | B1 |
8484276 | Carlson et al. | Jul 2013 | B2 |
8495438 | Roine | Jul 2013 | B2 |
8503250 | Demone | Aug 2013 | B2 |
8526239 | Kim | Sep 2013 | B2 |
8533245 | Cheung | Sep 2013 | B1 |
8555037 | Gonion | Oct 2013 | B2 |
8599613 | Abiko et al. | Dec 2013 | B2 |
8605015 | Guttag et al. | Dec 2013 | B2 |
8625376 | Jung et al. | Jan 2014 | B2 |
8644101 | Jun et al. | Feb 2014 | B2 |
8650232 | Stortz et al. | Feb 2014 | B2 |
8873272 | Lee | Oct 2014 | B2 |
8964496 | Manning | Feb 2015 | B2 |
8971124 | Manning | Mar 2015 | B1 |
9015390 | Klein | Apr 2015 | B2 |
9047193 | Lin et al. | Jun 2015 | B2 |
9047205 | Chen | Jun 2015 | B1 |
9165023 | Moskovich et al. | Oct 2015 | B2 |
20010007112 | Porterfield | Jul 2001 | A1 |
20010008492 | Higashiho | Jul 2001 | A1 |
20010010057 | Yamada | Jul 2001 | A1 |
20010028584 | Nakayama et al. | Oct 2001 | A1 |
20010043089 | Forbes et al. | Nov 2001 | A1 |
20020059355 | Peleg et al. | May 2002 | A1 |
20020174158 | Sutherland et al. | Nov 2002 | A1 |
20030023646 | Lin et al. | Jan 2003 | A1 |
20030167426 | Slobodnik | Sep 2003 | A1 |
20030222879 | Lin et al. | Dec 2003 | A1 |
20040073592 | Kim et al. | Apr 2004 | A1 |
20040073773 | Demjanenko | Apr 2004 | A1 |
20040085840 | Vali et al. | May 2004 | A1 |
20040095826 | Perner | May 2004 | A1 |
20040154002 | Ball et al. | Aug 2004 | A1 |
20040205289 | Srinivasan | Oct 2004 | A1 |
20040240251 | Nozawa et al. | Dec 2004 | A1 |
20050015557 | Wang et al. | Jan 2005 | A1 |
20050078514 | Scheuerlein et al. | Apr 2005 | A1 |
20050097417 | Agrawal et al. | May 2005 | A1 |
20050285862 | Noda et al. | Dec 2005 | A1 |
20060047937 | Selvaggi et al. | Mar 2006 | A1 |
20060069849 | Rudelic | Mar 2006 | A1 |
20060146623 | Mizuno et al. | Jul 2006 | A1 |
20060149804 | Luick et al. | Jul 2006 | A1 |
20060181917 | Kang et al. | Aug 2006 | A1 |
20060215432 | Wickeraad et al. | Sep 2006 | A1 |
20060225072 | Lari et al. | Oct 2006 | A1 |
20060291282 | Liu et al. | Dec 2006 | A1 |
20070103986 | Chen | May 2007 | A1 |
20070171747 | Hunter et al. | Jul 2007 | A1 |
20070180006 | Gyoten et al. | Aug 2007 | A1 |
20070180184 | Sakashita et al. | Aug 2007 | A1 |
20070195602 | Fong et al. | Aug 2007 | A1 |
20070285131 | Sohn | Dec 2007 | A1 |
20070285979 | Turner | Dec 2007 | A1 |
20070291532 | Tsuji | Dec 2007 | A1 |
20080025073 | Arsovski | Jan 2008 | A1 |
20080037333 | Kim et al. | Feb 2008 | A1 |
20080052711 | Forin et al. | Feb 2008 | A1 |
20080137388 | Krishnan et al. | Jun 2008 | A1 |
20080165601 | Matick et al. | Jul 2008 | A1 |
20080178053 | Gorman et al. | Jul 2008 | A1 |
20080215937 | Dreibelbis et al. | Sep 2008 | A1 |
20090067218 | Graber | Mar 2009 | A1 |
20090154238 | Lee | Jun 2009 | A1 |
20090154273 | Borot et al. | Jun 2009 | A1 |
20090254697 | Akerib | Oct 2009 | A1 |
20090295817 | Yu | Dec 2009 | A1 |
20100067296 | Li | Mar 2010 | A1 |
20100091582 | Vali et al. | Apr 2010 | A1 |
20100164972 | Akerib | Jul 2010 | A1 |
20100172190 | Lavi et al. | Jul 2010 | A1 |
20100210076 | Gruber et al. | Aug 2010 | A1 |
20100226183 | Kim | Sep 2010 | A1 |
20100308858 | Noda et al. | Dec 2010 | A1 |
20100332895 | Billing et al. | Dec 2010 | A1 |
20110051523 | Manabe et al. | Mar 2011 | A1 |
20110063919 | Chandrasekhar et al. | Mar 2011 | A1 |
20110093662 | Walker et al. | Apr 2011 | A1 |
20110103151 | Kim et al. | May 2011 | A1 |
20110119467 | Cadambi et al. | May 2011 | A1 |
20110122695 | Li et al. | May 2011 | A1 |
20110140741 | Zerbe et al. | Jun 2011 | A1 |
20110141830 | Chi | Jun 2011 | A1 |
20110219260 | Nobunaga et al. | Sep 2011 | A1 |
20110267883 | Lee et al. | Nov 2011 | A1 |
20110317496 | Bunce et al. | Dec 2011 | A1 |
20120005397 | Lim et al. | Jan 2012 | A1 |
20120017039 | Margetts | Jan 2012 | A1 |
20120023281 | Kawasaki et al. | Jan 2012 | A1 |
20120120705 | Mitsubori et al. | May 2012 | A1 |
20120134216 | Singh | May 2012 | A1 |
20120134225 | Chow | May 2012 | A1 |
20120134226 | Chow | May 2012 | A1 |
20120140540 | Agam et al. | Jun 2012 | A1 |
20120182798 | Hosono et al. | Jul 2012 | A1 |
20120195146 | Jun et al. | Aug 2012 | A1 |
20120198310 | Tran et al. | Aug 2012 | A1 |
20120246380 | Akerib et al. | Sep 2012 | A1 |
20120265964 | Murata et al. | Oct 2012 | A1 |
20120281486 | Rao et al. | Nov 2012 | A1 |
20120303627 | Keeton et al. | Nov 2012 | A1 |
20130003467 | Klein | Jan 2013 | A1 |
20130061006 | Hein | Mar 2013 | A1 |
20130107623 | Kavalipurapu et al. | May 2013 | A1 |
20130117541 | Choquette et al. | May 2013 | A1 |
20130124783 | Yoon et al. | May 2013 | A1 |
20130132702 | Patel et al. | May 2013 | A1 |
20130138646 | Sirer et al. | May 2013 | A1 |
20130163362 | Kim | Jun 2013 | A1 |
20130173888 | Hansen et al. | Jul 2013 | A1 |
20130205114 | Badam et al. | Aug 2013 | A1 |
20130207170 | Kurokawa | Aug 2013 | A1 |
20130219112 | Okin et al. | Aug 2013 | A1 |
20130227361 | Bowers et al. | Aug 2013 | A1 |
20130283122 | Anholt et al. | Oct 2013 | A1 |
20130286705 | Grover et al. | Oct 2013 | A1 |
20130326154 | Haswell | Dec 2013 | A1 |
20130332707 | Gueron et al. | Dec 2013 | A1 |
20140016420 | Jeong | Jan 2014 | A1 |
20140185395 | Seo | Jul 2014 | A1 |
20140215185 | Danielsen | Jul 2014 | A1 |
20140250279 | Manning | Sep 2014 | A1 |
20140344934 | Jorgensen | Nov 2014 | A1 |
20150134713 | Wheeler | May 2015 | A1 |
20150324290 | Leidel | Nov 2015 | A1 |
20150325272 | Murphy | Nov 2015 | A1 |
Number | Date | Country |
---|---|---|
102141905 | Aug 2011 | CN |
0214718 | Mar 1987 | EP |
2026209 | Feb 2009 | EP |
H0831168 | Feb 1996 | JP |
2009259193 | Mar 2015 | JP |
10-0211482 | Aug 1998 | KR |
10-2010-0134235 | Dec 2010 | KR |
10-2013-0049421 | May 2013 | KR |
2001065359 | Sep 2001 | WO |
2010079451 | Jul 2010 | WO |
2013062596 | May 2013 | WO |
2013081588 | Jun 2013 | WO |
2013095592 | Jun 2013 | WO |
Entry |
---|
Boyd et al., “On the General Applicability of Instruction-Set Randomization”, Jul.-Sep. 2010, (14 pgs.), vol. 7, Issue 3, IEEE Transactions on Dependable and Secure Computing. |
Stojmenovic, “Multiplicative Circulant Networks Topological Properties and Communication Algorithms”, (25 pgs.), Discrete Applied Mathematics 77 (1997) 281-305. |
“4.9.3 MINLOC and MAXLOC”, Jun. 12, 1995, (5pgs.), Message Passing Interface Forum 1.1, retrieved from http://www.mpi-forum.org/docs/mpi-1.1/mpi-11-html/node79.html. |
Derby, et al., “A High-Performance Embedded DSP Core with Novel SIMD Features”, Apr. 6-10, 2003, (4 pgs), vol. 2, pp. 301-304, 2003 IEEE International Conference on Accoustics, Speech, and Signal Processing. |
Debnath, BIPLOB, Bloomflash: Bloom Filter on Flash-Based Storage, 2011 31st Annual Conference on Distributed Computing Systems, Jun. 20-24, 2011, 10 pgs. |
Pagiamtzis, Kostas, “Content-Addressable Memory Introduction”, Jun. 25, 2007, (6 pgs.), retrieved from: http://www.pagiamtzis.com/cam/camintro. |
Pagiamtzis, et al., “Content-Addressable Memory (CAM) Circuits and Architectures: A Tutorial and Survey”, Mar. 2006, (16 pgs.), vol. 41, No. 3, IEEE Journal of Solid-State Circuits. |
International Search Report and Written Opinion for PCT Application No. PCT/US2013/043702, dated Sep. 26, 2013, (11 pgs.). |
Elliot, et al., “Computational RAM: Implementing Processors in Memory”, Jan.-Mar. 1999, (10 pgs.), vol. 16, Issue 1, IEEE Design and Test of Computers Magazine. |
Dybdahl, et al., “Destructive-Read in Embedded DRAM, Impact on Power Consumption,” Apr. 2006, (10 pgs.), vol. 2, Issue 2, Journal of Embedded Computing-Issues in embedded single-chip multicore architectures. |
Kogge, et al., “Processing in Memory: Chips to Petaflops,” May 23, 1997, (8 pgs.), retrieved from: http://www.cs.ucf.edu/courses/cda5106/summer02/papers/kogge97PIM.pdf. |
Draper, et al., “The Architecture of the DIVA Processing-In-Memory Chip,” Jun. 22-26, 2002, (12 pgs.), ICS '02, retrieved from: http://www.isi.edu/˜draper/papers/ics02.pdf. |
Adibi, et al., “Processing-In-Memory Technology for Knowledge Discovery Algorithms,” Jun. 25, 2006, (10 pgs.), Proceeding of the Second International Workshop on Data Management on New Hardware, retrieved from: http://www.cs.cmu.edu/˜damon2006/pdf/adibi06inmemory.pdf. |
U.S. Appl. No. 13/449,082, entitled, “Methods and Apparatus for Pattern Matching,” filed Apr. 17, 2012, (37 pgs.). |
U.S. Appl. No. 13/743,686, entitled, “Weighted Search and Compare in a Memory Device,” filed Jan. 17, 2013, (25 pgs.). |
U.S. Appl. No. 13/774,636, entitled, “Memory as a Programmable Logic Device,” filed Feb. 22, 2013, (30 pgs.). |
U.S. Appl. No. 13/774,553, entitled, “Neural Network in a Memory Device,” filed Feb. 22, 2013, (63 pgs.). |
U.S. Appl. No. 13/796,189, entitled, “Performing Complex Arithmetic Functions in a Memory Device,” filed Mar. 12, 2013, (23 pgs.). |
Number | Date | Country | |
---|---|---|---|
20190243556 A1 | Aug 2019 | US |
Number | Date | Country | |
---|---|---|---|
62059256 | Oct 2014 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15810880 | Nov 2017 | US |
Child | 16384212 | US | |
Parent | 14874068 | Oct 2015 | US |
Child | 15810880 | US |