Applied Machine Learning (ML) is a booming field that utilizes a cascade of layers of nonlinear processing units and algorithms for feature extraction and transformation with a wide variety of usages and applications. ML typically involves two phases, training, which uses a rich set of training data to train a plurality of machine learning models, and inference, which applies the trained machine learning models to actual applications. Each of the two phases poses a distinct set of requirements for its underlying infrastructures. Various infrastructures may be used, e.g., graphics processing unit (GPU), a central processing unit (CPU), a Field Programmable Gate Array (FPGA), an Application Specific Integrated Circuit (ASIC), etc. Specifically, the training phase focuses on, as a non-limiting example, GPU or ASIC infrastructures that scale with the trained models and retraining frequency, wherein the key objective of the training phase is to achieve high performance and reduce training time. The inference phase, on the other hand, focuses on infrastructures that scale with the applications, user, and data, and the key objective of the inference phase is to achieve energy (e.g., performance per watt) and capital (e.g., return on investment) efficiency.
Inference phase of ML is usually very computationally and data intensive. Unfortunately, as the input data and model sizes grow, data movement becomes a bottleneck and data processing increases because in order to perform simple processing, three operations or instructions are performed for each data, e.g., load, processing, and store. As the amount of data grows, performing these three operations or instructions becomes burdensome. Moreover, the current computing architecture is not scalable and are not well suited for ML and its applications, since a lot of time goes in loading and storing the data in comparison to processing the data.
The foregoing examples of the related art and limitations related therewith are intended to be illustrative and not exclusive. Other limitations of the related art will become apparent upon a reading of the specification and a study of the drawings.
Aspects of the present disclosure are best understood from the following detailed description when read with the accompanying figures. It is noted that, in accordance with the standard practice in the industry, various features are not drawn to scale. In fact, the dimensions of the various features may be arbitrarily increased or reduced for clarity of discussion.
The following disclosure provides many different embodiments, or examples, for implementing different features of the subject matter. Specific examples of components and arrangements are described below to simplify the present disclosure. These are, of course, merely examples and are not intended to be limiting. In addition, the present disclosure may repeat reference numerals and/or letters in the various examples. This repetition is for the purpose of simplicity and clarity and does not in itself dictate a relationship between the various embodiments and/or configurations discussed.
Before various embodiments are described in greater detail, it should be understood that the embodiments are not limiting, as elements in such embodiments may vary. It should likewise be understood that a particular embodiment described and/or illustrated herein has elements which may be readily separated from the particular embodiment and optionally combined with any of several other embodiments or substituted for elements in any of several other embodiments described herein. It should also be understood that the terminology used herein is for the purpose of describing the certain concepts, and the terminology is not intended to be limiting. Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood in the art to which the embodiments pertain.
Each of the engines in the architecture 100 is a dedicated hardware block/component including one or more microprocessors and on-chip memory units storing software instructions programmed by a user for various machine learning operations. When the software instructions are executed by the microprocessors, each of the hardware components becomes a special purposed hardware component for practicing certain machine learning functions as discussed in detail below. In some embodiments, the architecture 100 is on a single chip, e.g., a system-on-chip (SOC).
In the example of
In some embodiments, the inference engine 160 includes a two-dimensional computing array of processing tiles, e.g., tiles 0, . . . , 63, arranged in, e.g., 8 rows by 8 columns. Each processing tile (e.g., tile 0) includes at least one on-chip memory (OCM) e.g., 210, one POD engine (or POD), e.g., 220, and one processing engine/element (PE), e.g., 230. Here, the OCMs in the processing tiles are configured to receive data from the data streaming engine 140 in a streaming fashion. The OCMs enable efficient local access to data per processing tile. The PODs are configured to perform dense or regular computations on the received data in the OCMs, e.g., matrix operations such as multiplication, matrix manipulation, tanh, sigmoid, etc., and the PEs are configured to perform sparse/irregular computations and/or complex data shape transformations of the received data in the OCMs, e.g., memory transpose, addition operation, operations on irregular data structures (such as trees, graphs, and priority queues), respectively. Both the PODs and the PEs can be programmed according to the programming instructions received from the instruction-streaming engine 150. Accordingly, the data is received and processed by each processing tile as an input data stream from the DDR memory 120 and the result is output by each processing tile as a stream of data to the DDR memory 120.
In some embodiments, a plurality of (e.g., four) processing tiles together form a processing block or quad, e.g., processing tiles 0-3 forms processing block 250, wherein the processing tiles within each processing block are coupled to one another via a routing element 240. In some embodiments, all the routing elements are connected together as a mesh interconnect to connect the processing blocks in the same row or column as a two-dimensional array. It is appreciated that the number and/or types of components within each processing tile, the formation of the processing blocks, the number of processing tiles in each processing block, and the number of processing blocks in each row and column of the inference engine 160 as shown in
In the example of
During matrix multiplication, the POD engine 220 is configured to perform a plurality of inline post processing operations immediately on output from the matrix multiplication block 602 saved in the C registers 608 without having to transmit and save the output to the OCM 210 first and then read the C matrix from the OCM 210 again for these post matrix multiplication operations. By bypassing the roundtrip to the OCM 210, the inline post processing operations following the matrix multiplication saves time and improves efficiency of the ML operation by the inference engine 160. As shown by the example of
Tanh and Sigmoid
In some embodiments, the tanh/sigmoid unit 614 of the POD engine 220 is configured to perform tanh and/or sigmoid operations/functions on each element of the output from the matrix multiplication block 602 on per-element basis before writing the output to the OCM 210. In some embodiments, the tanh/sigmoid unit 614 is configured to perform the per-element operations on the output via a lookup table, wherein values of the lookup table can be preloaded into, accepted and maintained by the tanh/sigmoid unit 614 from the memory 120 by the core 130. The tanh/sigmoid unit 614 is configured to determine a corresponding value of the tanh or sigmoid operation by looking up a value x from the lookup table.
In some embodiments, since different sections and models may be used to approximate the per-element operations based on numerical analysis, the tanh/sigmoid unit 614 is configured to utilize multiple lookup tables for the per-element operations. Specifically, the tanh/sigmoid unit 614 is configured to divide the tanh or sigmoid function into multiple sections, where each section may be represented by a curve that is extrapolated based on a specific lookup table. The tanh/sigmoid unit 614 may then determine value of the tanh and/or sigmoid function by referencing a specific lookup table corresponding to a section associated with a known x value and fetch the value of the tanh and/or sigmoid function from the corresponding lookup table accordingly.
For integer input values (e.g., Int. 8), in some embodiments, the tanh/sigmoid unit 614 is configured to utilize a single lookup table to implement the tanh and/or the sigmoid operation/function by taking advantage of the operator symmetries and relationship of the tanh and/or sigmoid function. For a non-limiting example, the tanh/sigmoid unit 614 may implement the tanh and sigmoid functions based on a single lookup table with 128 entries for both functions, wherein input to the tanh/sigmoid operations/functions is in the range of −127 to 127. In the example of the tanh operation, the output activation after the tanh operation is performed will be in the range of −127 to 127.
For floating point input values (e.g., fp16), in some embodiments, the tanh/sigmoid unit 614 is configured to implement each of the tanh and/or the sigmoid functions as a piece-wise linear approximation for a positive input range/region with a lookup table based on the exponent and mantissa values of the floating point input x. In some embodiments, the tanh/sigmoid unit 614 is configured to represent the tanh function using the sigmoid function so that only a single table representing the sigmoid function is needed. Similar to the implementation for integer numbers, the tanh/sigmoid unit 614 is configured to take advantage of the symmetry of the tanh and sigmoid functions to obtain function values for the negative input range.
In some embodiments, the tanh/sigmoid unit 614 is further configured to divide the positive region of the functions into one or more of path-through, lookup and saturation regions for floating input values. In the example of the tanh operation, tanh(x)=2*sigmoid(2x)+1 for the positive input range of x, and the tanh/sigmoid unit 614 is configured to use the symmetry of tanh function to obtain tanh(−x)=−tanh(x) for the negative input range.
In some embodiments, the tanh/sigmoid unit 614 is configured to perform both multiplication and addition of the sigmoid operation on floating point input values by utilizing a dynamic table that provides flexibility to cover multiple, e.g., up to 7 exponent regions/ranges and up to a total of 32 table entries. In some embodiments, the dynamic table is self-described with its meta data stored at the beginning of the table.
The foregoing description of various embodiments of the claimed subject matter has been provided for the purposes of illustration and description. It is not intended to be exhaustive or to limit the claimed subject matter to the precise forms disclosed. Many modifications and variations will be apparent to the practitioner skilled in the art. Embodiments were chosen and described in order to best describe the principles of the invention and its practical application, thereby enabling others skilled in the relevant art to understand the claimed subject matter, the various embodiments and the various modifications that are suited to the particular use contemplated.
This application claims the benefit of U.S. Provisional Patent Application No. 62/675,076, filed May 22, 2018, which is incorporated herein in its entirety by reference. This application is a continuation application of U.S. patent application Ser. No. 16/419,994, filed on May 22, 2019, which is a continuation-in-part of U.S. patent application Ser. No. 16/226,550, filed Dec. 19, 2018, and entitled “Architecture for dense operations in machine learning inference engine,” which are incorporated herein in its entirety by reference.
Number | Name | Date | Kind |
---|---|---|---|
4982291 | Kurahashi et al. | Jan 1991 | A |
5329611 | Pechanek et al. | Jul 1994 | A |
5481487 | Jang et al. | Jan 1996 | A |
5948098 | Leung et al. | Sep 1999 | A |
6128638 | Thomas | Oct 2000 | A |
6282583 | Pincus et al. | Aug 2001 | B1 |
6415377 | Wolf et al. | Jul 2002 | B1 |
6577312 | Deering et al. | Jun 2003 | B2 |
6640262 | Uppunda et al. | Oct 2003 | B1 |
7089380 | Schober | Aug 2006 | B1 |
7191163 | Herrera et al. | Mar 2007 | B2 |
7509363 | Clifton | Mar 2009 | B2 |
7809663 | Birch et al. | Oct 2010 | B1 |
7840914 | Agarwal et al. | Nov 2010 | B1 |
7853752 | Agarwal et al. | Dec 2010 | B1 |
7912883 | Hussain | Mar 2011 | B2 |
8200728 | Michaels | Jun 2012 | B2 |
8200940 | Lindholm | Jun 2012 | B1 |
8209703 | Yee et al. | Jun 2012 | B2 |
8504954 | Arnold | Aug 2013 | B1 |
8583896 | Cadambi et al. | Nov 2013 | B2 |
8738860 | Griffin et al. | May 2014 | B1 |
8838663 | Tang et al. | Sep 2014 | B2 |
9015217 | Arnold et al. | Apr 2015 | B2 |
9753695 | Mortensen | Sep 2017 | B2 |
9954771 | Levy et al. | Apr 2018 | B1 |
10161786 | Chang et al. | Dec 2018 | B2 |
10296556 | Zhou | May 2019 | B2 |
10305766 | Zhang et al. | May 2019 | B1 |
10614357 | Lie | Apr 2020 | B2 |
10884736 | Farooqui | Jan 2021 | B1 |
11016801 | Sodani et al. | May 2021 | B1 |
11106432 | Mangnall et al. | Aug 2021 | B2 |
11604799 | Bigdelu et al. | Mar 2023 | B1 |
20030163671 | Gschwind et al. | Aug 2003 | A1 |
20040153501 | Yamashita et al. | Aug 2004 | A1 |
20070122347 | Statnikov et al. | May 2007 | A1 |
20080040577 | Nemirovsky et al. | Feb 2008 | A1 |
20090158005 | Carmichael | Jun 2009 | A1 |
20100017420 | Archer et al. | Jan 2010 | A1 |
20110219208 | Asaad et al. | Sep 2011 | A1 |
20110307890 | Achilles et al. | Dec 2011 | A1 |
20130101035 | Wang et al. | Apr 2013 | A1 |
20130117521 | Li et al. | May 2013 | A1 |
20140007098 | Stillwell, Jr. et al. | Jan 2014 | A1 |
20150019836 | Anderson et al. | Jan 2015 | A1 |
20150046753 | Cecka et al. | Feb 2015 | A1 |
20150106568 | Feldman et al. | Apr 2015 | A1 |
20150309808 | Nandy et al. | Oct 2015 | A1 |
20150347012 | Dewitt et al. | Dec 2015 | A1 |
20160132272 | Iwashita | May 2016 | A1 |
20160162402 | Woolley, Jr. et al. | Jun 2016 | A1 |
20160170916 | Deshpande et al. | Jun 2016 | A1 |
20160224465 | Morad et al. | Aug 2016 | A1 |
20170068571 | Lu et al. | Mar 2017 | A1 |
20170083313 | Sankaralingam et al. | Mar 2017 | A1 |
20170228345 | Gupta et al. | Aug 2017 | A1 |
20170351642 | Omtzigt | Dec 2017 | A1 |
20170353397 | Che | Dec 2017 | A1 |
20170357483 | Nicol et al. | Dec 2017 | A1 |
20170364694 | Jacob et al. | Dec 2017 | A1 |
20180046458 | Kuramoto | Feb 2018 | A1 |
20180047126 | Falkenstern et al. | Feb 2018 | A1 |
20180068019 | Novikoff et al. | Mar 2018 | A1 |
20180114114 | Molchanov et al. | Apr 2018 | A1 |
20180260220 | Lacy et al. | Sep 2018 | A1 |
20180286016 | Bar-On et al. | Oct 2018 | A1 |
20180293782 | Benthin et al. | Oct 2018 | A1 |
20180307980 | Barik et al. | Oct 2018 | A1 |
20180315158 | Nurvitadhi et al. | Nov 2018 | A1 |
20180341484 | Fowers et al. | Nov 2018 | A1 |
20180349388 | Skiles et al. | Dec 2018 | A1 |
20190121641 | Knowles et al. | Apr 2019 | A1 |
20190121679 | Wilkinson et al. | Apr 2019 | A1 |
20190138210 | Lindholm | May 2019 | A1 |
20190146455 | Beylkin et al. | May 2019 | A1 |
20190147471 | McKelvey, Jr. et al. | May 2019 | A1 |
20190171604 | Brewer | Jun 2019 | A1 |
20200082198 | Yao et al. | Mar 2020 | A1 |
20200242734 | Wang et al. | Jul 2020 | A1 |
20210133911 | Yao et al. | May 2021 | A1 |
20210158155 | Zhang et al. | May 2021 | A1 |
20210216874 | Jegou et al. | Jul 2021 | A1 |
20210319317 | Power et al. | Oct 2021 | A1 |
20210390076 | Fang et al. | Dec 2021 | A1 |
20220067513 | Stevens et al. | Mar 2022 | A1 |
20220076110 | Shao et al. | Mar 2022 | A1 |
20220207783 | Kwong et al. | Jun 2022 | A1 |
20220261650 | Zhao et al. | Aug 2022 | A1 |
20230024035 | Thuerck et al. | Jan 2023 | A1 |
20230071931 | Huang et al. | Mar 2023 | A1 |
20230106651 | Xi et al. | Apr 2023 | A1 |
20230252275 | Nez et al. | Aug 2023 | A1 |
Number | Date | Country |
---|---|---|
2604142 | Aug 2022 | GB |
0245385 | Jun 2002 | WO |
2018222904 | Dec 2018 | WO |
Entry |
---|
Ceze, L., et al. Colorama: Architectural Support for Data-Centric Synchronization, 2007, IEEE, pp. 134-144 (Year: 2007). |
NanoMesh: An Asynchronous Kilo-Core System-on-Chip, Tse, et al. 2013 19th IEEE International Symposium on Asynchronous Circuits and Systems. |
Brewer, “Instructions Set Innovations for the Convey HC-1 Computer”, 2010, pp. 70-79, Year: 2010. |
Seng, et al. “Reducing Power with Dynamic Critical Path Information”, Jan. 1999, pp. 114-123; Year: 1999. |
Gelado, et al., “An Asymmetric Distributed Shared Memory Model for Heterogeneous Parallel Systems”, 2010 (Year: 2010) 12 pages. |
Number | Date | Country | |
---|---|---|---|
20210248497 A1 | Aug 2021 | US |
Number | Date | Country | |
---|---|---|---|
62675076 | May 2018 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16419994 | May 2019 | US |
Child | 17223921 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16226550 | Dec 2018 | US |
Child | 16419994 | US |