Target architecture determination

Information

  • Patent Grant
  • 10795653
  • Patent Number
    10,795,653
  • Date Filed
    Monday, November 26, 2018
    5 years ago
  • Date Issued
    Tuesday, October 6, 2020
    4 years ago
Abstract
Examples of the present disclosure provide apparatuses and methods for target architecture determination. An example method comprises receiving an indication of a type of target architecture in a portion of source code and creating compiled code for the type of target architecture based on the indication.
Description
TECHNICAL FIELD

The present disclosure relates generally to semiconductor memory apparatuses and methods, and more particularly, to apparatuses and methods related to target architecture determination.


BACKGROUND

Memory devices are typically provided as internal, semiconductor, integrated circuits in computers or other electronic systems. There are many different types of memory including volatile and non-volatile memory. Volatile memory can require power to maintain its data (e.g., host data, error data, etc.) and includes random access memory (RAM), dynamic random access memory (DRAM), static random access memory (SRAM), synchronous dynamic random access memory (SDRAM), and thyristor random access memory (TRAM), among others. Non-volatile memory can provide persistent data by retaining stored data when not powered and can include NAND flash memory, NOR flash memory, and resistance variable memory such as phase change random access memory (PCRAM), resistive random access memory (RRAM), and magnetoresistive random access memory (MRAM), such as spin torque transfer random access memory (STT RAM), among others.


Electronic systems often include a number of processing resources (e.g., one or more processors), which may retrieve and execute instructions and store the results of the executed instructions to a suitable location. A processor can comprise a number of functional units (e.g., herein referred to as functional unit circuitry such as arithmetic logic unit (ALU) circuitry, floating point unit (FPU) circuitry, and/or a combinatorial logic block, for example, which can execute instructions to perform logical operations such as AND, OR, NOT, NAND, NOR, and XOR logical operations on data (e.g., one or more operands).


A number of components in an electronic system may be involved in providing instructions to the functional unit circuitry for execution. The instructions may be generated, for instance, by a processing resource such as a controller and/or host processor. Data (e.g., the operands on which the instructions will be executed to perform the logical operations) may be stored in a memory array that is accessible by the functional unit circuitry. The instructions and/or data may be retrieved from the memory array and sequenced and/or buffered before the functional unit circuitry begins to execute instructions on the data. Furthermore, as different types of operations may be executed in one or multiple clock cycles through the functional unit circuitry, intermediate results of the operations and/or data may also be sequenced and/or buffered.


In many instances, the processing resources (e.g., processor and/or associated functional unit circuitry) may be external to the memory array, and data is accessed via a bus between the processing resources and the memory array to execute a set of instructions. Processing performance may be improved in a processor-in-memory (PIM) device, in which a processor may be implemented internal and/or near to a memory (e.g., directly on a same chip as the memory array). A PIM device may save time by reducing and/or eliminating external communications and may also conserve power. Data can be moved from the memory array to registers external to the memory array via a bus.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a block diagram of an apparatus in the form of a computing system including a memory device in accordance with a number of embodiments of the present disclosure.



FIG. 2 illustrates a block diagram of a heterogeneous target architecture system in accordance with a number of embodiments of the present disclosure.



FIG. 3 illustrates examples of source code for heterogeneous target architecture determination in accordance with a number of embodiments of the present disclosure.



FIG. 4 illustrates a table showing address space indicators and target architectures in accordance with a number of embodiments of the present disclosure.





DETAILED DESCRIPTION

Examples of the present disclosure provide apparatuses and methods for target architecture determination. An example method comprises receiving an indication of a type of target architecture in a portion of source code and creating compiled code for the type of target architecture based on the indication.


Computing architectures can include a variety of hardware and software configurations that can perform a variety of computing functions. The code that the various computing architectures can execute to perform their functions can vary among the various computing architectures. Therefore, code can be created for execution by a particular computing architecture (e.g., target architecture). The code for the target architecture can include instructions that are unique and/or particular to a target architecture.


In a number of embodiments, code for a target architecture can include sets of instructions that are specific to a particular target architecture, therefore code can be generated that use instructions specific to a target architecture. The code that is generated for a particular target architecture can be executed properly by the particular target architecture, but another target architecture may not be able to execute the code and/or may generate errors when executing the code. A target architecture that executes code with instructions that are specific to a particular target architecture can be a heterogeneous target architecture. A target architecture that executes code with instructions that are not specific to a particular target architecture can be a homogeneous target architecture.


In a number of embodiments, source code can include an attribute that indicates a type of target architecture for a portion of instructions in the source code by using a named address space that is associated with a type of target architecture. For example, source code can include a named address space that a compiler will recognize and create compiled code (e.g., assembly code and/or object code) for the type of target architecture indicated by the named address space. Also, source code can include a number of portions of source code that define attributes that include named address spaces. Compiled code can be created by a compiler for each of the portions of source code according to the type of target architecture indicated by the named address spaces.


In the following detailed description of the present disclosure, reference is made to the accompanying drawings that form a part hereof, and in which is shown by way of illustration how one or more embodiments of the disclosure may be practiced. These embodiments are described in sufficient detail to enable those of ordinary skill in the art to practice the embodiments of this disclosure, and it is to be understood that other embodiments may be utilized and that process, electrical, and/or structural changes may be made without departing from the scope of the present disclosure. As used herein, “a number of” a particular thing can refer to one or more of such things (e.g., a number of memory arrays can refer to one or more memory arrays).


The figures herein follow a numbering convention in which the first digit or digits correspond to the drawing figure number and the remaining digits identify an element or component in the drawing. Similar elements or components between different figures may be identified by the use of similar digits. For example, 270 may reference element “70” in FIG. 2, and a similar element may be referenced as 370 in FIG. 3. As will be appreciated, elements shown in the various embodiments herein can be added, exchanged, and/or eliminated so as to provide a number of additional embodiments of the present disclosure. In addition, as will be appreciated, the proportion and the relative scale of the elements provided in the figures are intended to illustrate certain embodiments of the present invention, and should not be taken in a limiting sense



FIG. 1 is a block diagram of an apparatus in the form of a computing system 100 including a target architecture (e.g., memory device 120) in accordance with a number of embodiments of the present disclosure. The target architecture in FIG. 1 is a memory device 120. In a number of embodiments, target architectures can be memory devices (e.g., a PIM capable device), graphics processing units (GPUs), digital signal processors (DSPs), field programmable gate arrays (FPGAS), among other types of target architectures. As used herein, a memory device 120, a memory array 130, controller 140, compiler 114, and/or sensing circuitry 150 might also be separately considered an “apparatus.”


System 100 includes a host 110 coupled to memory device 120, which includes a memory array 130. System 100 can be a PIM capable system. Host 110 can be a host system such as a personal laptop computer, a desktop computer, a digital camera, a mobile telephone, or a memory card reader, among various other types of hosts. Host 110 can include a system motherboard and/or backplane and can include a number of processing resources (e.g., one or more processors, microprocessors, or some other type of controlling circuitry). The system 100 can include separate integrated circuits or both the host 110 and the device associated with a type of target architecture, memory device 120 in FIG. 1, can be on the same integrated circuit. The device associated with a type of target architectures that is part of system 100 and coupled to host 110 can be memory devices, graphics processing units (GPUs), digital signal processors (DSPs), field programmable gate arrays (FPGAS), The system 100 can be, for instance, a server system and/or a S performance computing (HPC) system and/or a portion thereof. Although the example shown in FIG. 1 illustrates a system having a Von Neumann architecture, embodiments of the present disclosure can be implemented in non-Von Neumann architectures (e.g., a Turing machine), which may not include one or more components (e.g., CPU, ALU, etc.) often associated with a Von Neumann architecture.


The host 110 can include a processor 112 and a compiler 114. Processor 112 can comprise a number of functional units such as arithmetic logic unit (ALU) circuitry, floating point unit (FPU) circuitry, and/or a combinatorial logic block, for example, which can comprise a number of registers and can be used to execute instructions by performing logical operations such as AND, OR, NOT, NAND, NOR, and XOR logical operations on operands. For example, the number of functional units, which may be referred to as functional unit circuitry, may be used to perform arithmetic operations such as addition, subtraction, multiplication, and/or division on operands. In many instances, the processing resources may be external to memory device 120 comprising memory array 130 in which the instructions to be executed (e.g., the executable file) are stored. The instructions to be executed can be accessed (e.g., retrieved from the memory array 130) via a bus 154 between the processor 112 and the memory device 120, for example.


Compiler 114 can include a non-transitory computer readable medium that stores instructions executable by processor 112 to perform various embodiments of the present disclosure. For example, compiler 114 can include instructions to identify target architectures and to create compile code for particular types of target architectures. Also, compiler 114 can include a table that correlates named address spaces with types of target architectures.


Compiler 114 can receive source code and use the processor 112 to generate compiled code based on the source code. The compiler 114 can include instructions stored on a non-transitory computer readable medium that can be executed by the processor to generate the compiled code. Compiler 114 can recognize commands from the source code that are associated with a particular architecture and create compiled code for the particular architecture.


For clarity, the system 100 has been simplified to focus on features with particular relevance to the present disclosure. The memory array 130 can be a DRAM array, SRAM array, STT RAM array, PCRAM array, TRAM array, processor in memory (PIM) array, RRAM array, NAND flash array, and/or NOR flash array, for instance. The array 130 can comprise memory cells arranged in rows coupled by access lines (which may be referred to herein as word lines or select lines) and columns coupled by sense lines (which may be referred to herein as digit lines or data lines). Although a single array 130 is shown in FIG. 1, embodiments are not so limited. For instance, memory device 120 may include a number of arrays 130 (e.g., a number of banks of DRAM cells). An example DRAM array is described in association with FIGS. 2 and 3.


The memory device 120 includes address circuitry 142 to latch address signals provided over an I/O bus 156 (e.g., a data bus) through I/O circuitry 144. Address signals are received and decoded by a row decoder 146 and a column decoder 152 to access the memory array 130. Data can be read from memory array 130 by sensing voltage and/or current changes on the sense lines using sensing circuitry 150. The sensing circuitry 150 can read and latch a page (e.g., row) of data from the memory array 130. The I/O circuitry 144 can be used for bi-directional data communication with host 110 over the I/O bus 156. The write circuitry 148 is used to write data to the memory array 130.


Controller 140 decodes signals provided by control bus 154 from the host 110. These signals can include chip enable signals, write enable signals, and address latch signals that are used to control operations (e.g., PIM operations) performed on the memory array 130, including data read, data write, and data erase operations. In various embodiments, the control circuitry 140 is responsible for executing instructions from the host 110. The controller 140 can be a state machine, a sequencer, or some other type of controller.


In a number of embodiments, the sensing circuitry 150 can comprise a number of sense amplifiers and a number of compute components, which may comprise a latch serving as a Comp_Compulator and can be used to perform logical operations (e.g., on data associated with complementary sense lines).



FIG. 2 illustrates a block diagram of a heterogeneous target architecture system in accordance with a number of embodiments of the present disclosure. In FIG. 2, source code 260 illustrates an example portion of source code that can be used to determine a type of target architecture for the instruction in the source code. Source code 260 defines an attribute 271 that includes a named address space 270 to indicate the type of target architecture for the instructions of this portion of the source code 260. The named address space 270 can include an integer to indicate on which type of target architecture the instructions will be used. The address space 270 in FIG. 2 is address space “102”. The various target architectures can include graphics processing units (GPUs), processor in memory (PIM) devices, digital signal processors (DSPs), field programmable gate arrays (FPGAS), among other architectures. Also, a named address space can be assigned to various versions of an architecture. For example, a first named address space can be assigned to a PIM device having a first storage capacity and a second named address space can be assigned to PIM device having a second storage capacity. Also, a first named address space can be assigned to horizontal operations on a PIM device and a second named address space can be assigned to vertical operations on a PIM device. Each of the devices having a different architecture can be coupled via a host that includes a host processor, such as processor 112 in FIG. 1. The attribute can also include variable 272. Variable 272, indicated as “MyVariable” in FIG. 2, can be associated with the instructions for the portion of the source code that has named address space 270.


In FIG. 2, a compiler 262 can receive and interpret the source code 260. The compiler 262 can include parser 266 to separate the source code in a number of portions based on the named address space of the source code 260. The parser 266 can use a number of rules, which are described below in association with FIG. 3, to group portions of source code together that can be compiled into compiled code for a particular architecture indicated by the named address space. For example, parser 266 can separate the source code 260 from other portions of source code (not shown) based on the named address space 270 of the source code, which indicates source code 260 is for the type of target architecture indicated by named address space “102”.


Once the parser 266 has separated the source code based on the named address spaces of the source code, the target identifier 268 can determine which target architecture is associated with the named address space 270 of the source code 260. The target identifier 268 can use a table, such as table 495 from FIG. 4 and described below in association with FIG. 4, to correlate the type of target architecture associated with the named address space 270. Code generator 270 can include instructions for creating compiled code that is specific for the various target architectures associated with a named address space 270. In FIG. 2, the address space “102” is associated with “Architecture_X”, therefore code generator 270 can create compiled code that includes instructions for the target architecture of “Architecture_X”.


Once the target identifier 268 has identified the target architecture for the source code, the code generator 270 can create compiled code for the instructions from the source code. The code generator 270 can create compiled code 264 and compiled code 266. The compiled code 264 can include instructions for a base processor. The source code (not shown) from which the compiled code 264 was created can include an address space indicator of “0” or not include an address space indicator. The compiled code 266 can include instructions for Architecture_X, as indicated by named address space 270 in source code 260. Code generator 270 can include instructions for creating code that is specific for the various target architectures associated with an address space indicator 270.



FIG. 3 illustrates examples of source code for target architecture determination in accordance with a number of embodiments of the present disclosure. Examples 380, 382, and 384 in FIG. 3 each include example of source code that can be compiled according to a number of embodiments of the present disclosure. Example 380 illustrates source code that can be compiled together because the attributes include named address spaces that are associated with the same type of target architecture. Example 382 illustrates source code that can be compiled into two separate portions of compiled code because the attributes include named address spaces that are associated with different types of target architectures. Example 384 illustrates source code that includes complex flow control statements that can be compiled together because the attribute includes a single named address space associated with a particular type of target architecture.


In example 380, attributes 371-1 and 371-2 both include named address space 370-1, defined by integer “102”. Therefore, source code portion 381 having variables 372-1 and 372-2 in the argument, both associated with named address space 370-1, can be compiled together into compiled code that uses instructions associated with a type of target architecture correlated to named address space “102”. The source code in example 380 includes a first attribute 371-1 with address space indicator 370-1, which is “102” in FIG. 3, having an associated variable 372-1, which is “XX” in FIG. 3; and a second attribute 371-2 with address space indicator 370-1, which is “102” in FIG. 3, having an associated variable 372-2, which is “YY” in FIG. 3. In example 380, the first attribute 371-1 and the second attribute 371-2 both have the same address space indicator 370-1, therefore the source code portion 381 can be compiled together by a compiler into compiled code for the target architecture associated with address space indicator 370-1. The compiled code based on the source code from example 380 can use variables 372-1 and 372-2 defined in the first and second attributes 371-1 and 371-2.


In example 382, attribute 371-3 includes named address space 370-1 and 370-2, defined by integer “102”, and attribute 371-4 includes named address space 370-3, defined by integer “103”. Therefore, source code portion 383 having variables 372-1, 372-2, 372-3, and 372-4 in the argument, associated with named address spaces 370-1 and 370-3, can be compiled into separate portions of compiled code, each using instructions associated with a particular type of target architecture. The source code in example 382 includes a first attribute with address space indicator 370-1, which is “102” in FIG. 3, having associated variables 372-1, which is “X” in FIG. 3, and 372-2, which is “Y” in FIG. 3; and a second attribute with address space indicator 370-3, which is “103” in FIG. 3, having associated variables 372-3, which is “S” in FIG. 3, and 372-4, which is “T” in FIG. 3. In example 382, the first attribute and the second attribute have different address space indicators, 370-1 and 370-3, therefore the source code portion 383 can be compiled by a compiler into two different portions of compiled code. A first portion can be compiled for target architecture associated with address space 370-1 and a second portion can be complied for the target architecture associated with address space 370-3.


In example 384, attribute 371-5 includes named address space 370-1, defined by integer “102”. Source code portion 385 includes complex flow control statements that include variables 372-1 and 372-2 in the argument, both associated with named address space 370-1, can be compiled together into compiled code that uses instructions associated with a type of target architecture correlated to named address space “102”. The source code in example 384 includes a first attribute with address space indicator 370-1, which is “102” in FIG. 3, having associated variables 372-1, which is “X” in FIG. 3, and 372-2, which is “Y” in FIG. 3. In example 384, source code portion 385 includes instructions having complex flow control. In a number of embodiments, when source code that includes complex flow control instructions is associated with a single target architecture, as indicated by an attribute having a single named address space, the source code can be compiled together into compiled code. In example 384, the complex flow control instructions, such as if-then-else statements and switch statements, among others, can be compiled together into compiled code for the target architecture associated with named address space 370-1 in response to the source code associating only named address space 370-1 with the complex flow control instructions in source code portion 385 of example 384.



FIG. 4 illustrates a table showing address space indicators and target architectures in accordance with a number of embodiments of the present disclosure. Table 495 in FIG. 4 can include a named address space 470 and a target architecture 492. In Table 495, named address space 470-1, which is “100” in FIG. 4, can be associated with PIM type 1492-1 target architecture, named address space 470-2, which is “101” in FIG. 4, can be associated with PIM type 2492-2 target architecture, named address space 470-3, which is “102” in FIG. 4, can be associated with PIM type 3492-3 target architecture, and named address space 470-4, which is “0” in FIG. 4, can be associated with host processor 492-4 target architecture. In a number of embodiments, the target architectures that are identified and associated with a named address space can be a number of heterogeneous target architectures that include one or more disparate instruction set architectures (e.g., instructions that are only properly executed by a particular target architecture). Table 495 can be stored in a compiler, such as in the target identifier 268 on compiler 262 in FIG. 2. Table 495 can be used by a compiler to determine which target architecture will be the basis for creating compiled code from source code.


Although specific embodiments have been illustrated and described herein, those of ordinary skill in the art will appreciate that an arrangement calculated to achieve the same results can be substituted for the specific embodiments shown. This disclosure is intended to cover adaptations or variations of one or more embodiments of the present disclosure. It is to be understood that the above description has been made in an illustrative fashion, and not a restrictive one. Combination of the above embodiments, and other embodiments not specifically described herein will be apparent to those of skill in the art upon reviewing the above description. The scope of the one or more embodiments of the present disclosure includes other applications in which the above structures and methods are used. Therefore, the scope of one or more embodiments of the present disclosure should be determined with reference to the appended claims, along with the full range of equivalents to which such claims are entitled.


In the foregoing Detailed Description, some features are grouped together in a single embodiment for the purpose of streamlining the disclosure. This method of disclosure is not to be interpreted as reflecting an intention that the disclosed embodiments of the present disclosure have to use more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive subject matter lies in less than all features of a single disclosed embodiment. Thus, the following claims are hereby incorporated into the Detailed Description, with each claim standing on its own as a separate embodiment.

Claims
  • 1. A method, comprising: separating a source code from other portions of source code based on a named address space defined in the source code;determining a type of target architecture based on the named address space defined in source code;creating a first portion of compiled code that includes instructions for the type of target architecture, wherein the first portion of compiled code includes a number of homogeneous commands associated with homogeneous target architectures and a first number of heterogeneous commands associated with heterogeneous target architectures;creating a second portion of compiled code that includes instructions for a base processor, wherein the second portion of compiled code includes the number of homogeneous commands and a second number of heterogeneous commands associated with the homogenous target architectures;the target architecture only executing the first portion of compiled code; andthe base processor only executing the second portion of compiled code.
  • 2. The method of claim 1, wherein an attribute includes the named address space.
  • 3. The method of claim 2, wherein the attribute is defined by the source code.
  • 4. The method of claim 3, wherein the attribute includes a variable.
  • 5. The method of claim 4, wherein the variable is associated with instructions for the first portion of the compiled code.
  • 6. The method of claim 1, wherein the named address space includes an integer to indicate the type of the target architecture.
  • 7. The method of claim 1, further comprising creating a third portion of compiled code that includes instructions for a different target architecture.
  • 8. The method of claim 1, wherein a third portion of compiled code is created in response to source code having different attributes.
  • 9. The method of claim 1, wherein a third portion of compiled code is created in response to the source code having different address spaces.
  • 10. The method of claim 1, wherein processing in memory (PIM) is the target architecture.
  • 11. An apparatus, comprising: a processor;a compiler that includes: a parser, wherein the parser is configured to separate a source code from other portions of source code based on a named address space defined in the source code;a target identifier, wherein the target identifier is configured to determine a type of target architecture based on the named address space defined in the source code, and wherein the target architecture is dependent on at least one of the target architecture being configured to perform horizontal operations or the target architecture being configured to perform vertical operations; anda code generator, wherein the code generator is configured to create a first portion of compiled code that includes instructions for the type of target architecture and a second portion of compiled code that includes instructions for a base processor;the target architecture configured to execute only the first portion of compiled code; andthe base processor configured to execute only the second portion of compiled code.
  • 12. The apparatus of claim 11, wherein the compiler can receive and interpret the source code.
  • 13. The apparatus of claim 11, wherein the parser separates the source code in a number of portions based on a number of named address spaces, including the named address space of the source code.
  • 14. The apparatus of claim 11, wherein the parser groups portions of the source code together based on the type of architecture.
  • 15. The apparatus of claim 11, wherein the target identifier includes a table correlating the type of the target architecture with the named address space.
  • 16. An apparatus comprising: a processor;a compiler that includes: a parser, wherein the parser is configured to separate a source code from other portions of source code based on a named address space defined in the source code and group portions of the source code together based on type of target architecture;a target identifier, wherein the target identifier is configured to determine the type of target architecture is a processing in memory (PIM) device based on the named address space defined in the source code; anda code generator, wherein the code generator is configured to create a first portion of compiled code that includes instructions for the PIM device from the source code and a second portion of complied code that includes instructions for a base processor from the source code;the PIM device configured to execute only the first portion of compiled code; andthe base processor configured to execute only the second portion of compiled code.
  • 17. The apparatus of claim 16, wherein the instructions are received from a memory array.
PRIORITY INFORMATION

This application is a Continuation of U.S. application Ser. No. 15/097,764, filed Apr. 13, 2016, which issues as U.S. Pat. No. 10,140,104 on Nov. 27, 2018, which claims the benefit of U.S. Provisional Application No. 62/147,341, filed Apr. 14, 2015, the contents of which are included herein by reference.

US Referenced Citations (275)
Number Name Date Kind
4380046 Fung Apr 1983 A
4435792 Bechtolsheim Mar 1984 A
4435793 Ochii Mar 1984 A
4727474 Batcher Feb 1988 A
4843264 Galbraith Jun 1989 A
4958378 Bell Sep 1990 A
4977542 Matsuda et al. Dec 1990 A
5023838 Herbert Jun 1991 A
5034636 Reis et al. Jul 1991 A
5201039 Sakamura Apr 1993 A
5210850 Kelly et al. May 1993 A
5253308 Johnson Oct 1993 A
5276643 Hoffmann et al. Jan 1994 A
5325519 Long et al. Jun 1994 A
5367488 An Nov 1994 A
5379257 Matsumura et al. Jan 1995 A
5386379 Ali-Yahia et al. Jan 1995 A
5398213 Yeon et al. Mar 1995 A
5440482 Davis Aug 1995 A
5446690 Tanaka et al. Aug 1995 A
5473576 Matsui Dec 1995 A
5481500 Reohr et al. Jan 1996 A
5485373 Davis et al. Jan 1996 A
5506811 McLaury Apr 1996 A
5615404 Knoll et al. Mar 1997 A
5638128 Hoogenboom Jun 1997 A
5638317 Tran Jun 1997 A
5654936 Cho Aug 1997 A
5678021 Pawate et al. Oct 1997 A
5724291 Matano Mar 1998 A
5724366 Furutani Mar 1998 A
5751987 Mahant-Shetti et al. May 1998 A
5774728 Breslau et al. Jun 1998 A
5787458 Miwa Jul 1998 A
5854636 Watanabe et al. Dec 1998 A
5867429 Chen et al. Feb 1999 A
5870504 Nemoto et al. Feb 1999 A
5915084 Wendell Jun 1999 A
5935263 Keeth et al. Aug 1999 A
5986942 Sugibayashi Nov 1999 A
5991209 Chow Nov 1999 A
5991785 Alidina et al. Nov 1999 A
6005799 Rao Dec 1999 A
6009020 Nagata Dec 1999 A
6092186 Betker et al. Jul 2000 A
6122211 Morgan et al. Sep 2000 A
6125071 Kohno et al. Sep 2000 A
6134164 Lattimore et al. Oct 2000 A
6147514 Shiratake Nov 2000 A
6151244 Fujino et al. Nov 2000 A
6157578 Brady Dec 2000 A
6163862 Adams et al. Dec 2000 A
6166942 Vo et al. Dec 2000 A
6172918 Hidaka Jan 2001 B1
6175514 Henderson Jan 2001 B1
6181698 Hariguchi Jan 2001 B1
6208544 Beadle et al. Mar 2001 B1
6226215 Yoon May 2001 B1
6301153 Takeuchi et al. Oct 2001 B1
6301164 Manning et al. Oct 2001 B1
6304477 Naji Oct 2001 B1
6389507 Sherman May 2002 B1
6418498 Martwick Jul 2002 B1
6466499 Blodgett Oct 2002 B1
6510098 Taylor Jan 2003 B1
6563754 Lien et al. May 2003 B1
6578058 Nygaard Jun 2003 B1
6731542 Le et al. May 2004 B1
6754746 Leung et al. Jun 2004 B1
6768679 Le et al. Jul 2004 B1
6807614 Chung Oct 2004 B2
6816422 Hamade et al. Nov 2004 B2
6819612 Achter Nov 2004 B1
6894549 Eliason May 2005 B2
6943579 Hazanchuk et al. Sep 2005 B1
6948056 Roth et al. Sep 2005 B1
6950771 Fan et al. Sep 2005 B1
6950898 Merritt et al. Sep 2005 B2
6956770 Khalid et al. Oct 2005 B2
6961272 Schreck Nov 2005 B2
6965648 Smith et al. Nov 2005 B1
6985394 Kim Jan 2006 B2
6987693 Cernea et al. Jan 2006 B2
7020017 Chen et al. Mar 2006 B2
7028170 Saulsbury Apr 2006 B2
7045834 Tran et al. May 2006 B2
7054178 Shiah et al. May 2006 B1
7061817 Raad et al. Jun 2006 B2
7079407 Dimitrelis Jul 2006 B1
7173857 Kato et al. Feb 2007 B2
7187585 Li et al. Mar 2007 B2
7196928 Chen Mar 2007 B2
7260565 Lee et al. Aug 2007 B2
7260672 Gamey Aug 2007 B2
7372715 Han May 2008 B2
7400532 Aritome Jul 2008 B2
7406494 Magee Jul 2008 B2
7447720 Beaumont Nov 2008 B2
7454451 Beaumont Nov 2008 B2
7457181 Lee et al. Nov 2008 B2
7535769 Cernea May 2009 B2
7546438 Chung Jun 2009 B2
7562198 Noda et al. Jul 2009 B2
7574466 Beaumont Aug 2009 B2
7602647 Li et al. Oct 2009 B2
7663928 Tsai et al. Feb 2010 B2
7685365 Rajwar et al. Mar 2010 B2
7692466 Ahmadi Apr 2010 B2
7752417 Manczak et al. Jul 2010 B2
7791962 Noda et al. Sep 2010 B2
7796453 Riho et al. Sep 2010 B2
7805587 Van Dyke et al. Sep 2010 B1
7808854 Takase Oct 2010 B2
7827372 Bink et al. Nov 2010 B2
7869273 Lee et al. Jan 2011 B2
7898864 Dong Mar 2011 B2
7924628 Danon et al. Apr 2011 B2
7937535 Ozer et al. May 2011 B2
7957206 Bauser Jun 2011 B2
7979667 Allen et al. Jul 2011 B2
7996749 Ding et al. Aug 2011 B2
8042082 Solomon Oct 2011 B2
8045391 Mohklesi Oct 2011 B2
8059438 Chang et al. Nov 2011 B2
8095825 Hirotsu et al. Jan 2012 B2
8117462 Snapp et al. Feb 2012 B2
8164942 Gebara et al. Apr 2012 B2
8208328 Hong Jun 2012 B2
8213248 Moon et al. Jul 2012 B2
8223568 Seo Jul 2012 B2
8238173 Akerib et al. Aug 2012 B2
8274841 Shimano et al. Sep 2012 B2
8279683 Klein Oct 2012 B2
8310884 Iwai et al. Nov 2012 B2
8332367 Bhattacherjee et al. Dec 2012 B2
8339824 Cooke Dec 2012 B2
8339883 Yu et al. Dec 2012 B2
8347154 Bahali et al. Jan 2013 B2
8351292 Matano Jan 2013 B2
8356144 Hessel et al. Jan 2013 B2
8417921 Gonion et al. Apr 2013 B2
8418155 McAllister et al. Apr 2013 B2
8462532 Argyres Jun 2013 B1
8484276 Carlson et al. Jul 2013 B2
8495438 Roine Jul 2013 B2
8503250 Demone Aug 2013 B2
8526239 Kim Sep 2013 B2
8533245 Cheung Sep 2013 B1
8555037 Gonion Oct 2013 B2
8599613 Abiko et al. Dec 2013 B2
8605015 Guttag et al. Dec 2013 B2
8625376 Jung et al. Jan 2014 B2
8644101 Jun et al. Feb 2014 B2
8650232 Stortz et al. Feb 2014 B2
8873272 Lee Oct 2014 B2
8964496 Manning Feb 2015 B2
8971124 Manning Mar 2015 B1
9015390 Klein Apr 2015 B2
9047193 Lin et al. Jun 2015 B2
9165023 Moskovich et al. Oct 2015 B2
20010007112 Porterfield Jul 2001 A1
20010008492 Higashiho Jul 2001 A1
20010010057 Yamada Jul 2001 A1
20010028584 Nakayama et al. Oct 2001 A1
20010043089 Forbes et al. Nov 2001 A1
20020059355 Peleg et al. May 2002 A1
20030167426 Slobodnik Sep 2003 A1
20030222879 Lin et al. Dec 2003 A1
20040073552 Bailey et al. Apr 2004 A1
20040073592 Kim et al. Apr 2004 A1
20040073773 Demjanenko Apr 2004 A1
20040083467 Hanley et al. Apr 2004 A1
20040085840 Vali et al. May 2004 A1
20040095826 Perner May 2004 A1
20040154002 Ball et al. Aug 2004 A1
20040205289 Srinivasan Oct 2004 A1
20040240251 Nozawa et al. Dec 2004 A1
20050015557 Wang et al. Jan 2005 A1
20050078514 Scheuerlein et al. Apr 2005 A1
20050097399 Bliss et al. May 2005 A1
20050097417 Agrawal et al. May 2005 A1
20060047937 Selvaggi et al. Mar 2006 A1
20060069849 Rudelic Mar 2006 A1
20060146623 Mizuno et al. Jul 2006 A1
20060149804 Luick et al. Jul 2006 A1
20060181917 Kang et al. Aug 2006 A1
20060215432 Wickeraad et al. Sep 2006 A1
20060225072 Lari et al. Oct 2006 A1
20060291282 Liu et al. Dec 2006 A1
20070103986 Chen May 2007 A1
20070171747 Hunter et al. Jul 2007 A1
20070180006 Gyoten et al. Aug 2007 A1
20070180184 Sakashita et al. Aug 2007 A1
20070195602 Fong et al. Aug 2007 A1
20070234316 Bayerlein Oct 2007 A1
20070285131 Sohn Dec 2007 A1
20070285979 Turner Dec 2007 A1
20070291532 Tsuji Dec 2007 A1
20080025073 Arsovski Jan 2008 A1
20080037333 Kim et al. Feb 2008 A1
20080052711 Forin et al. Feb 2008 A1
20080137388 Krishnan et al. Jun 2008 A1
20080165601 Matick et al. Jul 2008 A1
20080178053 Gorman et al. Jul 2008 A1
20080215937 Dreibelbis et al. Sep 2008 A1
20090067218 Graber Mar 2009 A1
20090154238 Lee Jun 2009 A1
20090154273 Borot et al. Jun 2009 A1
20090204785 Yates, Jr. Aug 2009 A1
20090254697 Akerib Oct 2009 A1
20100067296 Li Mar 2010 A1
20100091582 Vali et al. Apr 2010 A1
20100172190 Lavi et al. Jul 2010 A1
20100210076 Gruber et al. Aug 2010 A1
20100226183 Kim Sep 2010 A1
20100293342 Morfey et al. Nov 2010 A1
20100308858 Noda et al. Dec 2010 A1
20100332895 Billing et al. Dec 2010 A1
20110051523 Manabe et al. Mar 2011 A1
20110063919 Chandrasekhar et al. Mar 2011 A1
20110093662 Walker et al. Apr 2011 A1
20110103151 Kim et al. May 2011 A1
20110119467 Cadambi et al. May 2011 A1
20110122695 Li et al. May 2011 A1
20110140741 Zerbe et al. Jun 2011 A1
20110219260 Nobunaga et al. Sep 2011 A1
20110267883 Lee et al. Nov 2011 A1
20110317496 Bunce et al. Dec 2011 A1
20120005397 Lim et al. Jan 2012 A1
20120017039 Margetts Jan 2012 A1
20120023281 Kawasaki et al. Jan 2012 A1
20120120705 Mitsubori et al. May 2012 A1
20120134216 Singh May 2012 A1
20120134225 Chow May 2012 A1
20120134226 Chow May 2012 A1
20120140540 Agam et al. Jun 2012 A1
20120182798 Hosono et al. Jul 2012 A1
20120195146 Jun et al. Aug 2012 A1
20120198310 Tran et al. Aug 2012 A1
20120246380 Akerib et al. Sep 2012 A1
20120265964 Murata et al. Oct 2012 A1
20120281486 Rao et al. Nov 2012 A1
20120303627 Keeton et al. Nov 2012 A1
20130003467 Klein Jan 2013 A1
20130007712 Agarwal Jan 2013 A1
20130061006 Hein Mar 2013 A1
20130107623 Kavalipurapu et al. May 2013 A1
20130117541 Choquette et al. May 2013 A1
20130124783 Yoon et al. May 2013 A1
20130132702 Patel et al. May 2013 A1
20130138646 Sirer et al. May 2013 A1
20130163362 Kim Jun 2013 A1
20130173888 Hansen et al. Jul 2013 A1
20130205114 Badam et al. Aug 2013 A1
20130219112 Okin et al. Aug 2013 A1
20130227361 Bowers et al. Aug 2013 A1
20130283122 Anholt et al. Oct 2013 A1
20130286705 Grover et al. Oct 2013 A1
20130326154 Haswell Dec 2013 A1
20130332707 Gueron et al. Dec 2013 A1
20130339928 Trofin Dec 2013 A1
20140089905 Hux et al. Mar 2014 A1
20140185395 Seo Jul 2014 A1
20140189679 Adams Jul 2014 A1
20140215185 Danielsen Jul 2014 A1
20140250279 Manning Sep 2014 A1
20140344934 Jorgensen Nov 2014 A1
20150029798 Manning Jan 2015 A1
20150042380 Manning Feb 2015 A1
20150063052 Manning Mar 2015 A1
20150078108 Cowles et al. Mar 2015 A1
20150134713 Wheeler May 2015 A1
20150324290 Leidel Nov 2015 A1
20150325272 Murphy Nov 2015 A1
20150347108 Munshi Dec 2015 A1
Foreign Referenced Citations (13)
Number Date Country
102141905 Aug 2011 CN
0214718 Mar 1987 EP
2026209 Feb 2009 EP
H0831168 Feb 1996 JP
2009259193 Mar 2015 JP
10-0211482 Aug 1998 KR
10-2010-0134235 Dec 2010 KR
10-2013-0049421 May 2013 KR
2001065359 Sep 2001 WO
2010079451 Jul 2010 WO
2013062596 May 2013 WO
2013081588 Jun 2013 WO
2013095592 Jun 2013 WO
Non-Patent Literature Citations (29)
Entry
Brockman et al., “Programming, Compilation, and Runtime Support for Processing-In-Memory Based Parallel Architectures” (Year: 2003).
Beemster et al., “The Embedded C Extension to C” (Year: 2005).
Wang et al., “OpenMP Memkind: An Extension for Heterogeneous Physical Memories” (Year: 2017).
Schwartz et al., “OpenMP Memkind: An Extension for Heterogeneous Physical Memories” (Year: 1984).
Michael Meissner, “Adding named address space support to the GCC compiler” (Year: 2009).
Ryoo et al., “Optimization Principles and Application Performance Evaluation of a Multithreaded GPU Using CUDA” (Year: 2008).
Hong et al., “MapCG: Writing Parallel Program Portable between CPU and GPU” (Year: 2010).
Boyd et al., “On the General Applicability of Instruction-Set Randomization”, Jul.-Sep. 2010, (14 pgs.), vol. 7, Issue 3, IEEE Transactions on Dependable and Secure Computing.
Stojmenovic, “Multiplicative Circulant Networks Topological Properties and Communication Algorithms”, (25 pgs.), Discrete Applied Mathematics 77 (1997) 281-305.
“4.9.3 MINLOC and MAXLOC”, Jun. 12, 1995, (5pgs.), Message Passing Interface Forum 1.1, retrieved from http://www.mpi-forum.org/docs/mpi-1.1/mpi-11-html/node79.html.
Derby, et al, “A High-Performance Embedded DSP Core with Novel SIMD Features”, Apr. 6-10, 2003, (4 pgs), vol. 2, pp. 301-304, 2003 IEEE International Conference on Accoustics, Speech, and Signal Processing.
Debnath, Biplob, Bloomflash: Bloom Filter on Flash-Based Storage, 2011 31st Annual Conference on Distributed Computing Systems, Jun. 20-24, 2011, 10 pgs.
Pagiamtzis, Kostas, “Content-Addressable Memory Introduction”, Jun. 25, 2007, (6 pgs.), retrieved from: http://www.pagiamtzis.com/cam/camintro.
Pagiamtzis, et al., “Content-Addressable Memory (CAM) Circuits and Architectures: A Tutorial and Survey”, Mar. 2006, (16 pgs.), vol. 41, No. 3, IEEE Journal of Solid-State Circuits.
International Search Report and Written Opinion for PCT Application No. PCT/US2013/043702, dated Sep. 26, 2013, (11 pgs.).
Elliot, et al., “Computational RAM: Implementing Processors in Memory”, Jan.-Mar. 1999, (10 pgs.), vol. 16, Issue 1, IEEE Design and Test of Computers Magazine.
Dybdahl, et al., “Destructive-Read in Embedded DRAM, Impact on Power Consumption,” Apr. 2006, (10 pgs.), vol. 2, Issue 2, Journal of Embedded Computing-Issues in embedded single-chip multicore architectures.
Kogge, et al., “Processing in Memory: Chips to Petaflops,” May 23, 1997, (8 pgs.), retrieved from: http://www.cs.ucf.edu/courses/cda5106/summer02/papers/kogge97PIM.pdf.
Draper, et al., “The Architecture of the DIVA Processing-In-Memory Chip,” Jun. 22-26, 2002, (12 pgs.), ICS '02, retrieved from: http://www.isi.edu/˜draper/papers/ics02.pdf.
Adibi, et al., “Processing-In-Memory Technology for Knowledge Discovery Algorithms,” Jun. 25, 2006, (10 pgs.), Proceeding of the Second International Workshop on Data Management on New Hardware, retrieved from: http://www.cs.cmu.edu/˜damon2006/pdf/adibi06inmemory.pdf.
U.S. Appl. No. 13/449,082, entitled, “Methods and Apparatus for Pattern Matching,” filed Apr. 17, 2012, (37 pgs.).
U.S. Appl. No. 13/743,686, entitled, “Weighted Search and Compare in a Memory Device,” filed Jan. 17, 2013, (25 pgs.).
U.S. Appl. No. 13/774,636, entitled, “Memory as a Programmable Logic Device,” filed Feb. 22, 2013, (30 pgs.).
U.S. Appl. No. 13/774,553, entitled, “Neural Network in a Memory Device,” filed Feb. 22, 2013, (63 pgs.).
U.S. Appl. No. 13/796,189, entitled, “Performing Complex Arithmetic Functions in a Memory Device,” filed Mar. 12, 2013, (23 pgs.).
Muttreja et al., “Hybrid Simulation for Embedded Software Energy Estimation”, Proceedings of the 2005 42nd Design Automation Conference, Jun. 13-17, 2005, Anaheim, CA, pp. 23-26.
Strumpen, Volker, “Compiler Technology for Portable Checkpoints”, Laboratory for Computer Science, Massachusetts Institute of Technology, May 1998, pp. 1-11.
Bokan et al., “Adjustment of GCC Compiler Frontend for Embedded Processors”, 2014 22nd Telecommunications Forum Telfor (TELFOR), Nov. 25-27, 2014, Belgrade, Serbia, 5 pgs.
Wells et al., “Architecture of an Open Object-Oriented Database Management System”, Computer (IEEE Computer Society Publication), vol. 25, Issue 10, Oct. 1992, 10 pgs.
Related Publications (1)
Number Date Country
20190095182 A1 Mar 2019 US
Provisional Applications (1)
Number Date Country
62147341 Apr 2015 US
Continuations (1)
Number Date Country
Parent 15097764 Apr 2016 US
Child 16199477 US