Embodiments described herein are related to circuitry to perform matrix and vector operations in processor-based systems.
A variety of workloads being performed in modern computing systems rely on massive amounts of vector multiplications. For example, certain long short term memory (LSTM) learning algorithms are used in a variety of contexts such as language detection, card readers, natural language processing, and handwriting processing, among other things. LSTM processing includes numerous vector multiplications. The multiplications may be small integers or floating point numbers, for example, but very large numbers of them. Additionally, many of these workloads make significant use of outer product operations. The outer product operation is the matrix result of two input vectors (X and Y), where each element (i, j) of the matrix is the product of element i of the vector X and element j of the vector Y: Mij=XiYj. The performance of such operations on a general purpose central processing unit (CPU), even a CPU with vector instructions, is very low; while the power consumption is very high. Low performance, high power workloads are problematic for any computing system, but are especially problematic for battery-powered systems.
In an embodiment, a computation engine is configured to perform vector multiplications, producing either vector results or outer product (matrix) results. The instructions provided to the computation engine specify a matrix mode or a vector mode for the instructions. The computation engine performs the specified operation. The computation engine may perform numerous computations in parallel, in an embodiment. In an embodiment, the instructions may also specify an offset with the input memories, providing additional flexibility in the location of operands. More particularly, the computation engine may be configured to perform numerous multiplication operations in parallel and to accumulate results in a result memory, performing multiply-accumulate operations for each matrix/vector element in the targeted locations of the output memory. The computation engine may be both high performance and power efficient, in an embodiment, as compared to a general purpose processor (even one with vector instructions), for example.
The following detailed description makes reference to the accompanying drawings, which are now briefly described.
While embodiments described in this disclosure may be susceptible to various modifications and alternative forms, specific embodiments thereof are shown by way of example in the drawings and will herein be described in detail. It should be understood, however, that the drawings and detailed description thereto are not intended to limit the embodiments to the particular form disclosed, but on the contrary, the intention is to cover all modifications, equivalents and alternatives falling within the spirit and scope of the appended claims. The headings used herein are for organizational purposes only and are not meant to be used to limit the scope of the description. As used throughout this application, the word “may” is used in a permissive sense (i.e., meaning having the potential to), rather than the mandatory sense (i.e., meaning must). Similarly, the words “include”, “including”, and “includes” mean including, but not limited to.
Within this disclosure, different entities (which may variously be referred to as “units,” “circuits,” other components, etc.) may be described or claimed as “configured” to perform one or more tasks or operations. This formulation-[entity] configured to [perform one or more tasks]—is used herein to refer to structure (i.e., something physical, such as an electronic circuit). More specifically, this formulation is used to indicate that this structure is arranged to perform the one or more tasks during operation. A structure can be said to be “configured to” perform some task even if the structure is not currently being operated. A “clock circuit configured to generate an output clock signal” is intended to cover, for example, a circuit that performs this function during operation, even if the circuit in question is not currently being used (e.g., power is not connected to it). Thus, an entity described or recited as “configured to” perform some task refers to something physical, such as a device, circuit, memory storing program instructions executable to implement the task, etc. This phrase is not used herein to refer to something intangible. In general, the circuitry that forms the structure corresponding to “configured to” may include hardware circuits. The hardware circuits may include any combination of combinatorial logic circuitry, clocked storage devices such as flops, registers, latches, etc., finite state machines, memory such as static random access memory or embedded dynamic random access memory, custom designed circuitry, analog circuitry, programmable logic arrays, etc. Similarly, various units/circuits/components may be described as performing a task or tasks, for convenience in the description. Such descriptions should be interpreted as including the phrase “configured to.”
The term “configured to” is not intended to mean “configurable to.” An unprogrammed FPGA, for example, would not be considered to be “configured to” perform some specific function, although it may be “configurable to” perform that function. After appropriate programming, the FPGA may then be configured to perform that function.
Reciting in the appended claims a unit/circuit/component or other structure that is configured to perform one or more tasks is expressly intended not to invoke 35 U.S.C. § 112(f) interpretation for that claim element. Accordingly, none of the claims in this application as filed are intended to be interpreted as having means-plus-function elements. Should Applicant wish to invoke Section 112(f) during prosecution, it will recite claim elements using the “means for” [performing a function] construct.
In an embodiment, hardware circuits in accordance with this disclosure may be implemented by coding the description of the circuit in a hardware description language (HDL) such as Verilog or VHDL. The HDL description may be synthesized against a library of cells designed for a given integrated circuit fabrication technology, and may be modified for timing, power, and other reasons to result in a final design database that may be transmitted to a foundry to generate masks and ultimately produce the integrated circuit. Some hardware circuits or portions thereof may also be custom-designed in a schematic editor and captured into the integrated circuit design along with synthesized circuitry. The integrated circuits may include transistors and may further include other circuit elements (e.g. passive elements such as capacitors, resistors, inductors, etc.) and interconnect between the transistors and circuit elements. Some embodiments may implement multiple integrated circuits coupled together to implement the hardware circuits, and/or discrete elements may be used in some embodiments. Alternatively, the HDL design may be synthesized to a programmable logic array such as a field programmable gate array (FPGA) and may be implemented in the FPGA.
As used herein, the term “based on” or “dependent on” is used to describe one or more factors that affect a determination. This term does not foreclose the possibility that additional factors may affect the determination. That is, a determination may be solely based on specified factors or based on the specified factors as well as other, unspecified factors. Consider the phrase “determine A based on B.” This phrase specifies that B is a factor is used to determine A or that affects the determination of A. This phrase does not foreclose that the determination of A may also be based on some other factor, such as C. This phrase is also intended to cover an embodiment in which A is determined based solely on B. As used herein, the phrase “based on” is synonymous with the phrase “based at least in part on.”
This specification includes references to various embodiments, to indicate that the present disclosure is not intended to refer to one particular implementation, but rather a range of embodiments that fall within the spirit of the present disclosure, including the appended claims. Particular features, structures, or characteristics may be combined in any suitable manner consistent with this disclosure.
Turning now to
The computation engine 10 may be configured to perform one or more matrix operations (outer product operations) and one or more vector operations. Specifically, in an embodiment, the computation engine 10 may perform integer and floating point multiplications. For example, an embodiment receives input vectors (e.g. in the X memory 24 and the Y memory 26). The compute circuit 30 may include an array of compute elements (circuits). Each compute element may receive selected vector elements in the X memory 24 and the Y memory 26 and may multiply those values. Additionally, the compute element may receive the current value of a destination location in the Z memory 28 and may sum the current value with the multiplication result to generate the result for the Z memory 28 (thus accumulating the multiplication result with previous results).
In matrix mode, each vector element from the X memory 24 is multiplied by each of the vector elements from the Y memory 24 to generate the matrix elements for the output matrix. Specifically, input vectors may be loaded into the X memory 24 and the Y memory 26, and a compute instruction may be executed by the computation engine. In response to the compute instruction (and particularly the compute instruction being coded for the matrix mode), the computation engine 10 may perform the outer product operation and write the resulting outer product matrix to the Z memory 28. If the vector loaded into the X memory 24 (“X vector”) has a first number of vector elements and the vector loaded into the Y memory 26 (“Y vector”) has a second number of vector elements, the resulting matrix is a [first number] x [second number] matrix, where each entry (or element) in the matrix (element i, j) is the product of corresponding vector elements X(i) and Y(j). In an embodiment, the first number and second number are equal, and the matrix is a square matrix. Other embodiments may implement non-square matrices, or different outer product operations may produce square or non-square results based on the input vector elements.
In an embodiment, the computation engine 10 may perform outer product operations along with accumulating the result matrix with previous results in the Z memory 28 (where the accumulation may be adding or subtracting). That is, the outer product instruction may be a fused multiply-add (FMA) operation defined to multiply elements of the X vector by elements of the Y vector and add the products to corresponding elements of the Z matrix, or a fused multiply-subtract (FMS) operation defined to multiply elements of the X vector by elements of the Y vector and subtract the products from corresponding elements of the Z matrix. Alternatively, the FMS operation may include subtracting the corresponding elements of the Z matrix from the products. In an embodiment, the FMA and FMS may operate on floating point vector elements. A MAC compute instruction may also be supported for integer vector elements.
Furthermore, the compute instructions (FMA, FMS, and MAC) may be code for a vector mode. In the vector mode, a vector multiplication may be performed (e.g. each vector element in one vector may be multiplied by the corresponding vector element in the other vector). The results may be accumulated with current values in the Z memory 28, at a targeted entry of the Z memory 28. That is, in vector mode, a single entry (or row) of the Z memory 28 may be updated in vector mode, as opposed to multiple entries (rows) representing a matrix as is updated in the matrix mode.
Accordingly, each instruction may be coded for the desired mode (vector or matrix) and the instructions of different modes may be intermixed in a stream of computation instructions provided to the computation engine 10. That is, the computation engine 10 may not itself have a vector mode or matrix mode (programmed in a control register, for example), and instead may operate in either mode on an instruction-by-instruction basis. Flexibility and performance may be enhanced using an instruction-by-instruction mode selection, in some embodiments.
Additionally, the computation engine 10 may be configured to read operands from any offset within the X memory 24 and/or the Y memory 26. The operands may be selected with a register address identifying the entry in the memory 24 or 26 from which operands are to be read, and an offset into that entry. The initial operand element (vector element) may be selected from the offset, and additional vector elements may be read from adjacent locations in the entry until the end of the entry is reached. The computation engine 10 may complete the vector by reading additional vector elements/from the beginning of the next entry (the register address plus one). Thus, the data to be operated upon may be “misaligned” in the entries, and the correct data for a given operation may be read without moving data around in the memories 24 and 26. Such operation may be useful, e.g., if the operations to be performed use partially overlapping data.
In an embodiment, the vector elements may be 8 or 16 bit integers or 16, 32, or 64 bit floating point numbers. Thus, a 64 bit field in the X memory 24 or the Y memory 26 may include four 16 bit integer or eight 8 bit integers. Similarly, a 64 bit field in the X memory 24 or the Y memory 26 may include four 16 bit floating point numbers, two 32 bit floating point numbers, or one 64 bit floating point number.
As mentioned previously, the compute circuit 30 may be an array of compute elements, not only to perform the multiplications and additions that generate the elements for one result matrix element or result vector element, but also to perform multiplications for multiple matrix/vector elements in parallel. For example, if the X memory 24 and the Y memory 26 include 512 bit entries and 8 bit vector elements are implemented, 64 vector elements input matrices are stored in each entry of the X memory 24 and the Y memory 26 and may be processed in parallel in response to one compute instruction. Similarly, if 1024 bit entries are supported per entry of the memory, 128 vector elements may be processed in parallel. If 128 bit entries are supported, 16 vector elements may be processed in parallel. If 256 bit entries are supported, 32 vector elements may be processed in parallel. Alternatively, the compute circuit 30 may include a smaller number of MACs than would be used to perform all the matrix/vector element multiplications in the input operands in parallel. In such an embodiment, the computation engine 10 may use multiple passes through the compute circuit 30 for different portions of input data from the X memory 24 and the Y memory 26 to complete one array of matrix computations.
As mentioned above, the computation engine 10 may support multiple sizes of matrix/vector elements in the accumulated results, in one embodiment. For example, 16 bit result elements and 32 bit result elements may be supported for 16 bit input elements. For 32 bit input elements, 32 bit or 64 bit elements may be supported. The maximum number of result elements in the Z memory 28 may be set by the size of the Z memory 28 and the size of the accumulated element for a given operation. Smaller sizes may consume less memory in the Z memory 28. For matrix operations, the Z memory 28 may be arranged to write the smaller matrix elements in certain entries of the memory, leaving other entries unused (or unmodified). For example, if the matrix elements are ½ the size of the largest elements, every other entry in the Z memory 28 may be unused. If the matrix elements are ¼ the maximum size element, every fourth row may be used, etc. In an embodiment, the Z memory 28 may be viewed as having multiple banks, where the entries in the Z memory 28 are spread across the banks (e.g. even addressed entries may be in bank 0, and odd addressed entries may be in bank 1, for a two bank embodiment). Every fourth entry may be in a different bank if there are four banks (e.g. entries 0, 4, 8, etc. may be in bank 0, entries 1, 5, 9, etc. may be in bank 1, and so forth). Vector results may consume one row of the Z memory 28, as mentioned previously.
In an embodiment, the instructions executed by the computation engine 10 may also include memory instructions (e.g. load/store instructions). The load instructions may transfer vectors/matrices from a system memory (not shown in
In some embodiments, the computation engine 10 may include a cache 32 to store data recently accessed by the computation engine 10. The choice of whether or not to include cache 32 may be based on the effective latency experienced by the computation engine 10 and the desired level of performance for the computation engine 10. The cache 32 may have any capacity, cache line size, and configuration (e.g. set associative, direct mapped, etc.).
In the illustrated embodiment, the processor 12 is responsible for fetching the computation engine instructions (e.g. compute instructions, memory instructions, etc.) and transmitting the computation engine instructions to the computation engine 10 for execution. The overhead of the “front end” of the processor 12 fetching, decoding, etc. the computation engine instructions may be amortized over the matrix/vector computations performed by the computation engine 10. In one embodiment, the processor 12 may be configured to propagate the computation engine instruction down the pipeline (illustrated generally in
Generally, an instruction may be non-speculative if it is known that the instruction is going to complete execution without exception/interrupt. Thus, an instruction may be non-speculative once prior instructions (in program order) have been processed to the point that the prior instructions are known to not cause exceptions/speculative flushes in the processor 12 and the instruction itself is also known not to cause an exception/speculative flush. Some instructions may be known not to cause exceptions based on the instruction set architecture implemented by the processor 12 and may also not cause speculative flushes. Once the other prior instructions have been determined to be exception-free and flush-free, such instructions are also exception-free and flush-free.
In the case of memory instructions that are to be transmitted to the computation engine 10, the processing in the processor 12 may include translating the virtual address of the memory operation to a physical address (including performing any protection checks and ensuring that the memory instruction has a valid translation).
The instruction buffer 22 may be provided to allow the computation engine 10 to queue instructions while other instructions are being performed. In an embodiment, the instruction buffer 22 may be a first in, first out buffer (FIFO). That is, computation engine instructions may be processed in program order. Other embodiments may implement other types of buffers, multiple buffers for different types of instructions (e.g. load/store instructions versus compute instructions) and/or may permit out of order processing of instructions.
The X memory 24 and the Y memory 26 may each be configured to store at least one vector of matrices or vector elements defined for the computation engine instructions (e.g. 8, 16, 32, 64, etc. matrices of 8 bit matrix elements and 2×2 matrices). Similarly, the Z memory 28 may be configured to store at least one matrix computation result. The result may be an array of matrices at the result size (e.g. 16 bit matrix elements or 32 bit matrix elements). In some embodiments, the X memory 24 and the Y memory 26 may be configured to store multiple vectors of matrices and/or the Z memory 28 may be configured to store multiple result vectors of matrices. Each vector of matrices may be stored in a different bank in the memories, and operands for a given instruction may be identified by bank number. More generally, each entry in the memories 24, 26, and 28 may be addressed by a register address (e.g. register number) and thus the entries in the memories may be viewed as registers, similar to an integer or floating point register in the processor 12 (although generally significantly larger than such a register in terms of storage capacity).
The processor 12 fetches instructions from the instruction cache (ICache) 18 and processes the instructions through the various pipeline stages 20A-20N. The pipeline is generalized, and may include any level of complexity and performance enhancing features in various embodiments. For example, the processor 12 may be superscalar and one or more pipeline stages may be configured to process multiple instructions at once. The pipeline may vary in length for different types of instructions (e.g. ALU instructions may have schedule, execute, and writeback stages while memory instructions may have schedule, address generation, translation/cache access, data forwarding, and miss processing stages). Stages may include branch prediction, register renaming, prefetching, etc.
Generally, there may be a point in the processing of each instruction at which the instruction becomes non-speculative. The pipeline stage 20M may represent this stage for computation engine instructions, which are transmitted from the non-speculative stage to the computation engine 10. The retirement stage 20N may represent the state at which a given instructions results are committed to architectural state and can no longer by “undone” by flushing the instruction or reissuing the instruction. The instruction itself exits the processor at the retirement stage, in terms of the presently-executing instructions (e.g. the instruction may still be stored in the instruction cache). Thus, in the illustrated embodiment, retirement of a computation engine instruction occurs when the instruction has been successfully transmitted to the computation engine 10.
The instruction cache 18 and data cache (DCache) 16 may each be a cache having any desired capacity, cache line size, and configuration. Similarly, the lower level cache 14 may be any capacity, cache line size, and configuration. The lower level cache 14 may be any level in the cache hierarchy (e.g. the last level cache (LLC) for the processor 12, or any intermediate cache level).
In
In vector mode, the compute elements 48A-48D may perform the multiplications and accumulation with the current element in the Z memory entry 50. It is noted that, while the vector elements may be viewed as a single vector with M+1 elements, the vector elements may also be multiple vectors of fewer elements, all in one entry. The operation of the compute circuit 30 may be the same in either case.
As mentioned above in the discussion of
The compute instructions supported in the computation engine may include fused multiply add (FMA), fused multiply subtract (FMS) and multiply accumulate (MAC). FMA and FMS may operate on floating point elements (e.g. 16 bit, 32 bit, or 64 bit elements). FMA may compute Z=Z+X*Y, whereas FMS may compute Z=Z−X*Y. MAC may operate on integer operands (e.g. 8 bit or 16 bit integer operands) and may compute Z=Z+X*Y. In an embodiment, the MAC may support an optional right shift of the multiplication result before accumulating the result with Z.
As previously discussed, the Z memory 28 may be divided into banks, where the unused rows for each different size of vector elements may be allocated to the same bank, so that only ½ of the banks may be updated when twice the minimum size vector elements are used, ¼ of the banks may be updated when four times the minimum size vector elements are used, etc.
Thus, the X memory 24 in
Similarly, the Y memory 26 is accessed at the entry identified by Y RA, and the Y offset (arrow 64) may point to the initial element V0. In the example shown in
Also shown in
The memory operations may include load and store instructions. Specifically, in the illustrated embodiment, there are load and store instructions for the X, Y, and Z memories, respectively. In the case of the Z memory 28, a size parameter may indicate which matrix element size is being used (for matrix mode) and thus which rows of the Z memory are written to memory or read from memory (e.g. all rows, every other row, ever fourth row, etc.). In an embodiment, the X and Y memories may have multiple banks for storing different matrices/vectors. In such an embodiment, there may be multiple instructions to read/write the different banks or there may be an operand specifying the bank affected by the load/store X/Y instructions. In each case, an X memory bank may store a pointer to memory from/to which the load/store is performed. The pointer may be virtual, and may be translated by the processor 12 as discussed above. Alternatively, the pointer may be physical and may be provided by the processor 12 post-translation.
The compute instructions may perform a vector multiplication or matrix mode (outer product) operation, depending on the mode of the instruction (V/M in
In an embodiment, the FMA, FMS, and MAC instructions may further include variations the modify the operation being performed. The basic operation may be Z=Z+/−X*Y, but subsets of the operation may be performed in which there is no accumulation (Z=+/−X*Y), where X or Y is added to Z without multiplication (Z=Z+/−X, Z=Z+/−Y), clear (Z=0), and no-operation (NOP) (Z=Z).
In an embodiment, the compute instructions may support masking, where one or more elements may not be computed and stored in the Z memory 28. In such embodiments, the compute elements 48A-48D corresponding to the masked elements may be inactive during the operation, which may reduce power consumption. For example, there are some cases in which only even or odd numbered rows or columns of the result matrix may be updated (e.g. for complex numbers). In another case, restrict computations to the first N rows or columns may be desired (when the edges/corners of the matrices are reached, for example.). In matrix mode the intersection of the X and Y masks may define a Z mask. In vector mode, only the X mask may be used. Masked elements of the Z memory 28 may not be updated.
The peripherals 154 may include any desired circuitry, depending on the type of system 150. For example, in one embodiment, the system 150 may be a computing device (e.g., personal computer, laptop computer, etc.), a mobile device (e.g., personal digital assistant (PDA), smart phone, tablet, etc.), or an application specific computing device capable of benefitting from the computation engine 10 (e.g., neural networks, LSTM networks, other machine learning engines including devices that implement machine learning, etc.). In various embodiments of the system 150, the peripherals 154 may include devices for various types of wireless communication, such as wifi, Bluetooth, cellular, global positioning system, etc. The peripherals 154 may also include additional storage, including RAM storage, solid state storage, or disk storage. The peripherals 154 may include user interface devices such as a display screen, including touch display screens or multitouch display screens, keyboard or other input devices, microphones, speakers, etc. In other embodiments, the system 150 may be any type of computing system (e.g. desktop personal computer, laptop, workstation, net top etc.).
The external memory 158 may include any type of memory. For example, the external memory 158 may be SRAM, dynamic RAM (DRAM) such as synchronous DRAM (SDRAM), double data rate (DDR, DDR2, DDR3, etc.) SDRAM, RAMBUS DRAM, low power versions of the DDR DRAM (e.g. LPDDR, mDDR, etc.), etc. The external memory 158 may include one or more memory modules to which the memory devices are mounted, such as single inline memory modules (SIMMs), dual inline memory modules (DIMMs), etc. Alternatively, the external memory 158 may include one or more memory devices that are mounted on the IC 152 in a chip-on-chip or package-on-package implementation.
Generally, the electronic description 162 of the IC 152 stored on the computer accessible storage medium 160 may be a database which can be read by a program and used, directly or indirectly, to fabricate the hardware comprising the IC 152. For example, the description may be a behavioral-level description or register-transfer level (RTL) description of the hardware functionality in a high level design language (HDL) such as Verilog or VHDL. The description may be read by a synthesis tool which may synthesize the description to produce a netlist comprising a list of gates from a synthesis library. The netlist comprises a set of gates which also represent the functionality of the hardware comprising the IC 152. The netlist may then be placed and routed to produce a data set describing geometric shapes to be applied to masks. The masks may then be used in various semiconductor fabrication steps to produce a semiconductor circuit or circuits corresponding to the IC 152. Alternatively, the description 162 on the computer accessible storage medium 300 may be the netlist (with or without the synthesis library) or the data set, as desired.
While the computer accessible storage medium 160 stores a description 162 of the IC 152, other embodiments may store a description 162 of any portion of the IC 152, as desired (e.g. the computation engine 10 and/or the processor 12, as mentioned above).
Numerous variations and modifications will become apparent to those skilled in the art once the above disclosure is fully appreciated. It is intended that the following claims be interpreted to embrace all such variations and modifications.
Number | Name | Date | Kind |
---|---|---|---|
4719588 | Tatemichi et al. | Jan 1988 | A |
5025407 | Gulley et al. | Jun 1991 | A |
5107452 | Karmarkar | Apr 1992 | A |
5576982 | Wu | Nov 1996 | A |
5852444 | Lippincott | Dec 1998 | A |
5903769 | Arya | May 1999 | A |
6675187 | Greenberger | Jan 2004 | B1 |
6901422 | Sazegari | May 2005 | B1 |
6922716 | Desai et al. | Jul 2005 | B2 |
6959378 | Nickolls et al. | Oct 2005 | B2 |
7430577 | Buchert et al. | Sep 2008 | B2 |
7664930 | Paver et al. | Feb 2010 | B2 |
7668894 | Sazegari et al. | Feb 2010 | B2 |
7814297 | Wezelenburg | Oct 2010 | B2 |
7844352 | Vouzis | Nov 2010 | B2 |
7873812 | Mimar | Jan 2011 | B1 |
8250130 | Brokenshire et al. | Aug 2012 | B2 |
8458442 | Eichenberger | Jun 2013 | B2 |
8577950 | Eichenberger | Nov 2013 | B2 |
8650240 | Eichenberger | Feb 2014 | B2 |
8984043 | Ginzburg | Mar 2015 | B2 |
9361065 | Ajima | Jun 2016 | B2 |
9600281 | Eichenberger | Mar 2017 | B2 |
9658986 | Ge et al. | May 2017 | B2 |
10089278 | Moskovich et al. | Oct 2018 | B2 |
10338919 | Boswell | Jul 2019 | B2 |
10346163 | Bainville et al. | Jul 2019 | B2 |
10489479 | Shalev | Nov 2019 | B1 |
10489482 | Graves | Nov 2019 | B1 |
20030188127 | So | Oct 2003 | A1 |
20030221086 | Simovich et al. | Nov 2003 | A1 |
20050125631 | Symes et al. | Jun 2005 | A1 |
20050197977 | Buck | Sep 2005 | A1 |
20050203980 | Harrison et al. | Sep 2005 | A1 |
20050225562 | Higgins et al. | Oct 2005 | A1 |
20070028076 | Wezelenburg | Feb 2007 | A1 |
20090024685 | Salama et al. | Jan 2009 | A1 |
20090100247 | Moyer et al. | Apr 2009 | A1 |
20100274990 | Wilder | Oct 2010 | A1 |
20110040821 | Eichenberger | Feb 2011 | A1 |
20110040822 | Eichenberger | Feb 2011 | A1 |
20110055517 | Eichenberger et al. | Mar 2011 | A1 |
20110153707 | Ginzburg et al. | Jun 2011 | A1 |
20120254591 | Hughes et al. | Oct 2012 | A1 |
20120254592 | San Adrian et al. | Oct 2012 | A1 |
20130159665 | Kashyap | Jun 2013 | A1 |
20130198495 | Vick et al. | Aug 2013 | A1 |
20140019720 | Sprangle et al. | Jan 2014 | A1 |
20150012724 | Lutz et al. | Jan 2015 | A1 |
20160020871 | Elezabi | Jan 2016 | A1 |
20170031682 | Eapen et al. | Feb 2017 | A1 |
20170097884 | Werner et al. | Apr 2017 | A1 |
20170102892 | Pusukuri et al. | Apr 2017 | A1 |
20170192781 | Valentine et al. | Jul 2017 | A1 |
20180032312 | Hansen et al. | Feb 2018 | A1 |
20180074824 | Sazegari | Mar 2018 | A1 |
20180107630 | Zhou et al. | Apr 2018 | A1 |
20180137075 | Linderman et al. | May 2018 | A1 |
20180197067 | Mody | Jul 2018 | A1 |
20180321937 | Brown et al. | Nov 2018 | A1 |
20180357064 | Chen et al. | Dec 2018 | A1 |
20190065150 | Heddes et al. | Feb 2019 | A1 |
20190065190 | Zhang et al. | Feb 2019 | A1 |
20190079903 | Dreyer et al. | Mar 2019 | A1 |
20190102178 | Zbiciak | Apr 2019 | A1 |
20190171448 | Chen | Jun 2019 | A1 |
Number | Date | Country |
---|---|---|
106445471 | Feb 2017 | CN |
1365319 | Nov 2003 | EP |
3051412 | Aug 2016 | EP |
9110963 | Jul 1991 | WO |
2017185389 | Nov 2017 | WO |
Entry |
---|
Tang et al., “Table-Lookup Algorithms for Elementary Functions and Their Error Analysis”, 1991, pp. 232-236. |
Kantabutra, “On Hardware for Computing Exponental and Trigonometric Functions”, Mar. 1996, pp. 328-339. |
Song Han et al: EIE: Efficient Interence Engine on Compressed Deep Neural Networks, ACM Sigarch Computer Architecture News, ACM Special Interset Group on Computer Archtecture, 2 Penn Plaza, Suite 701 New York NY 10121-0701 USA, Stanford University, 2016, 12 pages. |
Shaoli Liu et al., “Cambricon,” ACM Sigarch Computer Architecture News, ACM Special Interest Group on Computer Architecture, 2 Penn Plaza, Suite 701 New York, NY 10121-0701 USA, vol. 44, No. 3, Jun. 18, 2016 (2016-06-180, pp. 393-405, XP058300636, ISSN:0163-5964, DOI: 10:1145/3007787.3001179. |
Mostafa I Soliman Ed—Anonymous: “Mat-core: A matrix core extension for general purpose processors,” Computer Engineering & Systems, 2007; IEEE, PI, Nov. 1, 2007 (Nov. 1, 2007), pp. 304-310, XP031212365, ISBN: 978-1-4244-1365-2. |
ISR/WO, PCT/US2017/048453, dated Nov. 16, 2017, 16 pages. |
ISR/WO, PCT/US2019/042744, dated Oct. 23, 2019, 15 pages. |
Number | Date | Country | |
---|---|---|---|
20200034145 A1 | Jan 2020 | US |