REDUCED DATA REPRESENTATION MEMORY CONTROLLER AND RELATED CHIPLET SETS

Information

  • Patent Application
  • 20240411479
  • Publication Number
    20240411479
  • Date Filed
    August 20, 2024
    4 months ago
  • Date Published
    December 12, 2024
    10 days ago
Abstract
A transform memory controller and method are described herein wherein the transform memory controller comprises logic elements configured to perform desired transform operations on data that flows to-and-from conventional computer memory elements. The transform operations are configured to perform operations on such data without the need for such data to travel to-and-from the conventional computer memory element via the processor (e.g., Central Processing Unit (CPU)) of the computer system. Several desirable transform operations are herein disclosed.
Description
BACKGROUND

Conventional computer architectures (e.g., the von Neumann architecture) are well known in the art. FIG. 1 depicts such one such architecture (100) which depicts in block diagram form how such a computer is constructed. A processing unit 111 (e.g., CPU, GPU or the like) is in electronic communication with various stages of memory—for example, random access memory (RAM 114) and read only memory (ROM 116). There may, in addition, be fast memory stores, such as cache memory, in which recent data references may be stored. Cache memory operates on the principle that recently accessed data may likely be accessed again the near future, so it is efficient to keep such memory in the fastest storage possible.



FIG. 2 depicts one such conventional hierarchy of memory—i.e., from fastest access (registers and cache) to slowest (Storage Class Memory (SCM) and Hard Disk Drive (HDD)). However, in today's computing environment, data movement now takes more energy, time, and silicon real estate than computation. FIG. 3 is a graph of energy required to move data for a 64-bit operation/computation, as performed on different computer systems. As may be seen, data movement is depicted as overtaking computation as the most dominant cost of a system in terms of energy consumption and time delay. As may be further seen in FIG. 3, the cost of moving a 64-bit operand a mere 5 mm across a chip exceeds that of a floating point operation that uses that operand (as indicated on line 302).


As seen, off chip data movement costs vastly more energy and more CPU cycles. Consequently, if data is being moved such as to-or-from main memory, it would be desirable if computations on that data could be performed with little silicon/power/time costs. It should be appreciated that 64-bit operations are merely exemplary and that other operations of other dimensions are contemplated.


In addition, some circuitries are now reaching such density that if all the circuits are on at the same time, the energy flux density may exceed that of a nuclear reactor. Hence, the notion of “dark silicon”, i.e. circuits which are on only part of the time, is acceptable. Thus, it would be desirable if a memory controller—which performs operations that are invoked only part of the time—can perform useful operations which do not add significant time delay or energy costs during essential but energy and time delay intensive data streams into and out of memory. It would also be desirable that the operations are flexible and require a common simple architecture.


BRIEF SUMMARY

Several embodiments of the present application comprising systems and methods of reduced data representation memory controllers and related chiplet sets are disclosed.


In the several embodiments, a data transformation is performed on the data as it is passing through the traditional memory controller electronics. In one embodiment, a transform may be construed as a useful, universal data transformation which can be performed by a memory side controller. The memory controller being a part of a computer system further comprising a central processor and a hierarchy of computer memory elements, the transform memory controller comprising: an input, the input receiving data signals associated with the computer memory elements; a set of logic and arithmetic elements, the set of logic elements configured to perform a transform operation on the data signals associated with the computer memory element wherein the transform operation performs a desired computation on the data signals without the need of the desired computation being performed by the CPU of the computer system; and an output, the output of the transform operation sends results of the computation to the computer memory elements.


In another set of embodiments, a method for performing transform operations on data residing in desired levels of slower memory elements, the steps of said method comprising: receiving an instruction for an operation on data in the computer system; determining the cost of the operation on the data to be performed at the central processor; and if the cost of the operation is above a desired threshold, then performing the data operation at the transform memory controller instead of at the central processor.


Other features and advantages of the present system are presented below in the Detailed Description when read in connection with the drawings presented within this application.


Features and advantages of the present system and method are presented below in the Detailed Description when read in connection with the drawings presented within this application.


All references, publications, patents, and patent applications, cited herein and/or cited in any accompanying Information Disclosure Statement (IDS), are hereby incorporated herein by reference in their entirety for all purposes.





BRIEF DESCRIPTION OF THE DRAWINGS

The accompanying figures, in which like reference numerals refer to identical or functionally similar elements throughout the separate views, and which together with the detailed description below are incorporated in and form part of the specification, serve to illustrate various examples and to explain various principles and advantages all in accordance with the present disclosure, in which:



FIG. 1 shows a conventional computer architecture, as known in the art;



FIG. 2 depicts a conventional hierarchy of computer memory, from fastest access to slowest access speeds, as is known in the art;



FIG. 3 is a graph depicting the amount of energy needed to move data across a distance on a typical chip;



FIG. 4 shows one embodiment of an architecture of a memory controller made in accordance with the principles of the present application;



FIG. 5 shows one exemplary use case and/or algorithm that employs the memory controller as made in accordance with the principles of the present application;



FIGS. 6A and 6B show two embodiments of how memory controllers of the present application can create transformations and using transforms, respectively-made in accordance with the principles of the present application;



FIGS. 7A, 7B, and 7C show different embodiments of memory controllers made in accordance with the principles of the present application;



FIG. 8 is yet another embodiment of the memory controller made in accordance with the principles of the present application;



FIGS. 9A, 9B, and 9C is a collection of transform matrices that may be applied to various data that affect a desired processing, as made in accordance with the principles of the present application;



FIG. 10 is another possible application/algorithm that uses the memory controllers of the present application; and



FIGS. 11A and 11B are views of yet another possible application/algorithm that uses the memory controllers of the present application.





DETAILED DESCRIPTION

As required, detailed embodiments are disclosed herein; however, it is to be understood that the disclosed embodiments are merely examples and that the devices, systems, and methods described herein can be embodied in various forms. Therefore, specific structural and functional details disclosed herein are not to be interpreted as limiting, but merely as a basis for the claims and as a representative basis for teaching one of ordinary skill in the art to variously employ the disclosed subject matter in virtually any proprietary detailed structure and function. Further, the terms and phrases used herein are not intended to be limiting, but rather, to provide an understandable description. Additionally, unless otherwise specifically expressed or clearly understood from the context of use, a term as used herein describes the singular and/or the plural of that term.


Overview

As mentioned, data movement now takes more energy, time, and silicon real estate than computation. Thus, a memory controller which performs operations that are invoked only part of the time can perform useful operations as the data streams into and out of memory. Various embodiments of the present application describe many useful functions that a memory controller can perform in addition to the usual functions performed by the previous memory controllers. Such usual functions of a typical memory controller may include the following functions-read by address, read by content, write by address, write by content, error correction, and more recently, encryption/decryption. In addition, there also are various addressing modes such as block transfer, streaming, or page transfer.


In one embodiment of the present application, the memory controller functionality may be expanded to perform operations in parallel with such basic memory functions—which may comprise useful functions improving the overall system performance. These operations can be programmatically invoked on the data as it is coming into and out of memory or to eliminate the transfer to the CPU all together. In some embodiments, the results of these operations may be stored in other parts of main memory or in memory controller registers. In such cases, the memory controller functions may shorten data flow paths from memory to the CPU, may process data flows from other data sources, and process data flows to and from the edge—e.g., sensors and/or actuators.


In one embodiment, the results of these operations can be stored in other parts of main memory or in memory controller registers. In the context of the present application, the term “transform” operations/algorithms/firmware and/or hardware refer to additional and/or auxiliary functions and/or hardware that may be performed to reduce the cost of moving data to-and-from storage to CPU.


In many embodiments, it suffices for the purposes of the present application that the operations on data of a desired size/dimension be costly in terms of energy consumption and time delay that it would be desirable to perform the data operations in a transform memory controller as described herein—as opposed to transferring the data signals to the processor (e.g., CPU, GPU, etc.) of the computer system to perform. In many embodiments, the transform memory controller may be configured to perform operations on data signals that would have a threshold (either actual or predicted) energy cost and/or time delay.



FIG. 4 shows one possible high-level data transform architecture 400—and the data flow therethrough—as made in the context of the present application. As may be seen, several hardware modules (e.g. 410a (GPU), 410b (CPU Side Accelerators), 410c (Memory Side Accelerators), 410d (Memory), 410e (Stream Accelerators) and 410f (IOT Data Sources)) may be either off-the-shelf (OTS) components that may be modified to perform transform memory transformation (MT) operations/algorithms, or there may additional transformation hardware that is added to standard components that affect transform memory operations. The memory controller functions can shorten data flow paths from memory to the CPU and/or can process data flows from other data sources, and process data flows to and from the edge i.e., from sensors and/or actuators.


As may be seen in FIG. 4, there are functional blocks—e.g., 402, 404 and 406, where transformation processing and/or hardware may be advantageously applied to this architecture. In addition, FIG. 4 describes some embodiments of the transform data flow. By performing the memory side functions as indicated, data movement may be minimized. In many embodiments described herein, these memory side operations are sufficiently useful such that the benefits of the additional functionality outweigh the cost. In addition, many embodiments describe functionality that is broadly useful for many types of data.


In many embodiments, the useful operations depend on the application. Because the costs of designing, fabricating, testing, and packaging these DRAM memory arrays is very expensive, the ability to customize the memory side computations could be quite useful. The various embodiments of the present application enable the system designer to locate system limiting functions in a “chiplet”—i.e., a customizable, memory side processor which optimizes various memory operations, depending on the predominant system usage cases without requiring the additional high cost of redesigning a memory chip fabrication process.


In many embodiments, such chiplets may be added to the DRAM die through the use of a silicon interposer, solder bumps or other high bandwidth chip interconnect technologies. Moreover, customizing memory through the use of chiplets permit logic optimized fabrication and testing for the chiplet and memory optimized fabrication for the DRAM memory. This may be desirable, as the optimum fabrication process for logic and memory are well known to be somewhat incompatible.


One Exemplary Transform Operation

For one example of a transform operation, FIG. 5 depicts one possible sketch transform operation that illustrates the present application. As mentioned, a useful sketch transform process proposes to use the idea of sketching or thumbnails of the data of interest. The sketch of the data may be a reduced representation of the data which summarizes one or more salient features of the blocks or flows of the data that flows through the memory controller. Simple examples might include the mean, variance, min/max, median, sub sampling, histograms, compression, encryption, neural networks, or more generally sketch transformations.



FIG. 5 depicts one aspect of sketch transformation that comes from the field of linear algebra as a consequence of the Johnson-Lindstrauss theorem is that a sketching matrix which consists of random elements result in a projection of the linear algebra problem to a lower dimensional subspace. The linear algebra projections solutions are known as random linear algebra (RandNLAProperties) such as clustering, similarity, as well as many of the above-mentioned matrix operations may be accomplished in a reduced dimensional domain by multiplying the data by appropriate sketch transforming matrices S as discussed below in the invention description. In this illustrative example the RandNLA algorithm for least-squares problems first computes sketch transform of matrices A and b, by multiplying them by S forming the reduced dimension versions SA and Sb, of the input data. These matrices solve a least-square problem in the sketched subspace to obtain a low-precision approximation. Alternatively, the sketch transformations are used to construct a traditional preconditioner for an iterative algorithm on the original input data to get high-precision approximations. For subspace-preserving embedding, if S is a random sampling matrix, then properties of a high leverage point (such as 503) will be preserved in SA. If S is a random-projection type matrix, then the information in the high leverage point may be homogenized or uniformized in SA.


In 506, another example involves high dimensional matrix multiplications. Random linear algebra of unitary matrices may employ transform matrices, SUaT and SUa, as shown. For a subspace-preserving embedding, if Ua is an orthogonal matrix, then SUa is approximately orthogonal. It should be appreciated that many of the firmware/hardware embodiments described herein may be able to process such reduced linear algebra operations as described. Thus, by multiplying data by a sketching matrix as the data leaves or enters the memory can preserve much of the properties of the data in a reduced sketched form. Subsequent linear algebra operations can be rapidly approximated on the sketched data. Only if the full details of the data are required, will it be necessary to access all of the data and incurring the energy and access time costs



FIGS. 6A and 6B are possible functional blocks of a memory controller for creating transforms and using transforms, respectively. As may be seen in FIG. 6A, transforming Memory Controller 600 creates data transforms as the memory flows to or from memory as controlled by the address. The transform matrix is prefilled with appropriate entries as discussed herein at blocks 602 and 604. The data is multiplied by the transform matrix at block 606 and the results are stored in the transform memory at block 608.



FIG. 6B depicts the architecture for using the transforms. The controller first operates on the transform memory in order to simplify and/or speed up the action required on the main memory. For merely one example, averaging can be done on the transform, or looking for elements in the matrix can first be done on the min max row/column transform, so only a few row/columns of the main memory may be accessed. The averages or cumulative sums may be computed by accessing the transforms (e.g., from block 608) rather than the main data memory. As the transforms are first accessed rather than accessing main memory, the time and/or energy to access the main memory is greatly reduced. This is the case-whether the entire computation is completed without access to the main memory, or whether the computation is eliminated and/or simplified.


In one embodiment, as the data flows to and from the main memory under the control of the address and address decoder, the data may be multiplied or other operations performed by the transform matrix. The transform matrix may be preloaded under the control of the memory controller instructions specifying the operations to be performed by the memory controller and where to store the results in the transform storage memory. Many exemplary architectures may be optimized for matrix/matrix or matrix tensor multiplication either dense or sparse and implemented as a transforming memory controller. For merely one embodiment, a FPGA block may be programmable and allow for a variant of functionality which would work well for flow processing, as a transform chiplet.


Several Exemplary Hardware Embodiments


FIGS. 7A, B and C depict several embodiments of hardware embodiments that may create and/or perform transform operations, as made in accordance with the principles of the present application. FIG. 7A is one such architecture 700. As may be seen, a memory controller chiplet 704 may be mechanically and/or electronically mated to a suitable substrate 708. Chiplet 704 may comprise logic circuits that affect transform operations as described herein. In proximity to chiplet 704, other memory elements (such as DRAM chip 702) may be so mated onto substrate 708. To affect the processing and data movement an interconnect fabric 710 may transport data to and from chiplet 704 (e.g., from DRAM 702 or onto bus 706).



FIG. 7B depicts another embodiment of a hardware implementation (720). In this embodiment, transform operations may be affected by a set of memory controller chiplets 722. Chiplets 722 may be mechanically and/or electronically mated to both DRAM chips 702 and/or bus 706. In this embodiment, memory controller chiplets are in shorter communication pathways with either DRAM memory elements and/or the bus. FIG. 7C is a top view of yet another hardware implementation as made in accordance with the principles of the present application. As may be seen, memory controller chiplet 704 may be surrounded by memory elements 702 in a one-to-many communication/processing pathway.


In another embodiment, it should be appreciated that the transform memory controller could be made such that the set of logic elements are constructed integrally with a processor to comprise a System On a Chip (SOC). In yet another embodiment, the system may comprise a set of arithmetic units, a possible set of registers, and may some program memory so it has components of a processor (e.g., CPU, GPU, etc.) but may not have an entire set of processor logic. In some embodiments, the system may be constructed similar to a digital signal processing (DSP) unit which may convolve an input data stream with a kernel located in the DSP registers.



FIG. 8 is yet another embodiment of a hardware configuration as made in accordance with the principles of the present application. As may be seen from this top view, a set of transform processors 802 may be in electronic communications with both network elements 806 and memory controllers 802. A suitable bus structure 808 may be implemented to affect efficient movement of data between the various elements.


Several Exemplary Transform Operation Embodiments


FIGS. 9A, 9B
9C, 10 and 11A and B depict several embodiments of transform processes/operations that may be affected by the architectures described herein. FIG. 9A shows one transform matrix that can perform column sum/average on the data matrix as shown. FIG. 9B shows another transform matrix that may perform cumulative sums/averages. FIG. 9C shows yet another transform matrix that may perform a subsampling operation on a data matrix as shown. In this case, a reduced size thumbnail of the data can be obtained which can be used for search. Once the search list has been reduced using the thumbnails, a search on the complete detailed data will require substantially less data I/O from memory. It will be appreciated that other transform matrices may be implemented to perform efficient computation on data.



FIG. 10 illustrates the use of transforming to create a low-density parity check (LDPC) code. Convolution against random binary matrices may give a number of relations which can be used to determine if there is an error and which element must be corrected. As may be seen, the H matrix may suffice as a parity-check matrix—and the corresponding Tanner graph is shown. In one example, the data may be multiplied by a series of random binary vectors. If a bit is in error, the data vector may no longer be in the null space of the transforming matrix and the error may be often uniquely corrected with probability guarantees.



FIGS. 11A and B depict one embodiment affecting a compressed sensing operation, as an encoder 1102 and a decoder 1110, respectively. As may be seen, encoder 1102 multiplies a signal vector, e.g., vector X 1108, which may have zero signal elements (seen as white boxes)—with a measurement array 1106 φ. This measurement or transformation array has a restricted isometry property and can, for example, consist of random numbers. The result, vector 1104 is sent. In decoder 1110, an estimate of the sparse vector 1118 (α) can be solved by minimizing the L1 norm of 1118 (α) subject to the constraint that the observation vector 1114 is equal to the product of the measurement array 1106 times dictionary matrix 1116 (Ψ) which can be a Fourier transform matrix and a sparse vector 1118 (γ=φ Ψα). An estimate of X 1108 is obtained by solving for the product the dictionary array 1116 times the sparse estimate vector 1118 (x=Ψ α). As for previous problems, the fundamental operation required is a transformation matrix times the input data to the memory which in this case may arrive from an external data source.


ENUMERATED EMBODIMENTS

EE1: In a computer system comprising a central processor and a hierarchy of computer memory elements and further comprising a transform memory controller, the transform memory controller performing operations on data residing in desired levels of slower memory elements, a method for performing transform operations on data residing in desired levels of slower memory elements, the steps of said method comprising:


receiving an instruction for an operation on data in the computer system;


determining the cost of the operation on the data to be performed at the central processor; and


if the cost of the operation is above a desired threshold, then performing the data operation at the transform memory controller instead of at the central processor.


EE 1.2: The method of EE1 wherein the step of receiving an instruction further comprises determining whether the instructions is among a set of instructions that are pre-determined to be performed at the transform memory controller.


EE 1.3: The method of EE1 the step of determining the cost of the operation on the data to be performed at the central processor is a function of the energy consumption of the received instruction to be performed at the central processor of the computer system.


EE1.4: The method of EE1 the step of determining the cost of the operation on the data to be performed at the central processor is a function of the time delay of the received instruction to be performed at the central processor of the computer system.


EE2: A transform memory controller, the controller being a part of a computer system further comprising a processor and a hierarchy of computer memory elements, the transform memory controller comprising:


an input, the input receiving data signals associated with the computer memory elements;


a set of logic elements, the set of logic elements configured to perform a transform operation on the data signals associated with the computer memory element wherein the transform operation performs a desired computation on the data signals without the need of the desired computation being performed by the processor of the computer system; and


and an output, the output of the transform operation sends results of the computation to the computer memory elements.


EE2.1: The transform memory controller of EE2 wherein:


the set of logic elements comprise a chiplet, said chiplet configured to be in electronic communications with neighboring computer memory elements.


EE2.2: The transform memory controller of EE2.1 wherein:


the chiplet is mechanically mated to a substrate, the substrate comprising computer memory elements mechanically mated to the substrate.


Now that various embodiments have been herein disclosed, it is also to be appreciated that any one or more of the particular tasks, steps, processes, methods, functions, elements and/or components described herein may suitably be implemented via hardware, software, firmware or a combination thereof. In particular, various modules, components and/or elements may be embodied by processors, electrical circuits, computers and/or other electronic data processing devices that are configured and/or otherwise provisioned to perform one or more of the tasks, steps, processes, methods and/or functions described herein. For example, a controller, a processor, computer or other electronic data processing device embodying a particular element may be provided, supplied and/or programmed with a suitable listing of code (e.g., such as source code, interpretive code, object code, directly executable code, and so forth) or other like instructions or software or firmware, such that when run and/or executed by the controller, processor, computer or other electronic data processing device one or more of the tasks, steps, processes, methods and/or functions described herein are completed or otherwise performed. Suitably, the listing of code or other like instructions or software or firmware is implemented as and/or recorded, stored, contained or included in and/or on a non-transitory computer and/or machine readable storage medium or media so as to be providable to and/or executable by the computer or other electronic data processing device. For example, suitable storage mediums and/or media can include but are not limited to: floppy disks, flexible disks, hard disks, magnetic tape, or any other magnetic storage medium or media, CD-ROM, DVD, optical disks, or any other optical medium or media, a RAM, a ROM, a PROM, an EPROM, a FLASH-EPROM, or other memory or chip or cartridge, or any other tangible medium or media from which a computer or machine or electronic data processing device can read and use. In essence, as used herein, non-transitory computer-readable and/or machine-readable mediums and/or media comprise all computer-readable and/or machine-readable mediums and/or media except for a transitory, propagating signal.


Optionally, any one or more of the particular tasks, steps, processes, methods, functions, elements and/or components described herein may be implemented on and/or embodiment in one or more general purpose computers, special purpose computer(s), a programmed microprocessor or microcontroller and peripheral integrated circuit elements, an ASIC or other integrated circuit, a digital signal processor, a hardwired electronic or logic circuit such as a discrete element circuit, a programmable logic device such as a PLD, PLA, FPGA, Graphical card CPU (GPU), or PAL, or the like. In general, any device, capable of implementing a finite state machine that is in turn capable of implementing the respective tasks, steps, processes, methods and/or functions described herein can be used.


A detailed description of one or more embodiments of the application, read along with accompanying figures, that illustrate the principles of the application has now been given. It is to be appreciated that the application is described in connection with such embodiments, but the application is not limited to any embodiment. The scope of the application is limited only by the claims and the application encompasses numerous alternatives, modifications and equivalents. Numerous specific details have been set forth in this description in order to provide a thorough understanding of the application. These details are provided for the purpose of example and the application may be practiced according to the claims without some or all of these specific details. For the purpose of clarity, technical material that is known in the technical fields related to the application has not been described in detail so that the application is not unnecessarily obscured.

Claims
  • 1. In a computer system comprising a central processor and a hierarchy of computer memory elements and further comprising a transform memory controller, the transform memory controller performing transform operations on data residing in a desired level of slower computer memory elements, a method for performing transform operations on data residing in the desired level of slower computer memory elements, the method comprising: receiving an instruction for performing a transform operation on data in the computer system;determining the cost of the transform operation on the data to be performed at the central processor; andif the cost of the transform operation is above a desired threshold, then performing the data transform operation at the transform memory controller instead of at the central processor.
  • 2. The method of claim 1, wherein the step of receiving an instruction further comprises determining whether the instruction is among a set of instructions that are pre-determined to be performed at the transform memory controller.
  • 3. The method of claim 1, wherein the performing the data transform operation at the transform memory controller comprises: receiving data signals from at least one of the computer memory elements which is accessed by the transform memory controller using read by address or content separate from access of the computer memory element by the central processor;performing at least one transform operation on the data signals received by the transform memory controller, wherein the transform operation performs a desired computation on the data signals without the desired computation being performed by the central processor of the computer system; andsending results of the desired computation to at least one of the computer memory elements which is accessed by the transform memory controller using write by address or content separate from access of the computer memory element by the processor.
  • 4. The method of claim 1, wherein the step of determining the cost of the transform operation on the data to be performed at the central processor is a function of the energy consumption of the received instruction to be performed on the data at the central processor of the computer system.
  • 5. The method of claim 1, wherein the step of determining the cost of the transform operation on the data to be performed at the central processor is a function of the time delay of the received instruction to be performed on the data at the central processor of the computer system.
  • 6. In a computer system comprising a central processor and a hierarchy of computer memory elements and further comprising a transform memory controller, the transform memory controller comprising one or more chiplets for performing operations on data residing in desired levels of slower computer memory elements, a method for performing transform operations on data residing in desired levels of slower computer memory elements, the method comprising: receiving an instruction for performing a transform operation on data in the computer system;determining the cost of the transform operation on the data to be performed at the central processor; andif the cost of the transform operation is above a desired threshold, then performing the data transform operation at the one or more chiplets of the transform memory controller instead of at the central processor.
  • 7. The method of claim 6, wherein the step of receiving an instruction further comprises determining whether the instructions is among a set of instructions that are pre-determined to be performed at the transform memory controller.
  • 8. The method of claim 6, wherein the performing the data transform operation at the one or more chiplets of the transform memory controller comprises: receiving at the one or more chiplets data signals from at least one of the computer memory elements which is accessed by the transform memory controller using read by address or content separate from access of the computer memory element by the central processor;performing at least one transform operation on the data signals received by the one or more chiplets of the transform memory controller, wherein the transform operation performs a desired computation on the data signals without the desired computation being performed by the central processor of the computer system; andsending results of the desired computation from the one or more chiplets to at least one of the computer memory elements which is accessed by the transform memory controller using write by address or content separate from access of the computer memory element by the processor.
  • 9. The method of claim 6, wherein the step of determining the cost of the transform operation on the data to be performed at the central processor is a function of the energy consumption of the received instruction to be performed on the data at the central processor of the computer system.
  • 10. The method of claim 6, wherein the step of determining the cost of the transform operation on the data to be performed at the central processor is a function of the time delay of the received instruction to be performed on the data at the central processor of the computer system.
  • 11. The method of claim 6, wherein the one or more chiplets are configured to be in electronic communications with neighboring computer memory elements in the hierarchy.
  • 12. The method of claim 11, wherein the one or more chiplets are mechanically mated to a substrate, the substrate comprising the neighboring computer memory elements mechanically mated to the substrate.
  • 13. The method of claim 11, wherein each chiplet in the one or more chiplets is in electronic communications with a neighboring computer memory element via an interconnect fabric.
  • 14. The method of claim 6, wherein the one or more chiplets comprise a plurality of chiplets mechanically mated to a substrate, the substrate comprising at least one neighboring computer memory element in the hierarchy which is mechanically mated to the substrate, and wherein each chiplet in the plurality of chiplets is in electronic communications with the neighboring computer memory element via an interconnect fabric.
  • 15. A computer system comprising: a transform memory controller;a central processor;a hierarchy of computer memory elements; and whereinthe transform memory controller comprising: one or more transform chiplets operatively coupled to at least one computer memory element in the hierarchy of computer memory elements which are all accessed by the central processor using at least one of read by address, read by content, write by address, write by content, or any combination, the at least one computer memory element being accessed by the transform memory controller using read by address or content under control of memory address and address decoder and separate from the central processor;the one or more transform chiplets operatively coupled to transform storage memory via a high bandwidth chip interconnect fabric, the one or more transform chiplets configured to perform a transform operation on data signals received by the one or more transform chiplets, and wherein the transform operation performs a desired computation on the data signals storing results in the transform storage memory without the desired computation being performed by the central processor of the computer system; andthe one or more transform chiplets operatively coupled to at least one computer memory element in the hierarchy, thereby sending results of the transform operation to the at least one computer memory element in the hierarchy which is accessed by the transform memory controller using write by address or content under control of memory address and address decoder and separate from the central processor.
  • 16. The computer system of claim 15, wherein each of the one or more transform chiplets is configured to be in electronic communications with a neighboring computer memory element in the hierarchy via an interconnect fabric.
  • 17. The computer system of claim 16, wherein the one or more transform chiplets are mechanically mated to a substrate, the substrate comprising at least one neighboring computer memory element mechanically mated to the substrate.
  • 18. The computer system of claim 16, wherein at least one of the one or more transform chiplets is electrically coupled via a first high bandwidth chip interconnect fabric to a neighboring computer memory element and electrically coupled via a second high bandwidth chip interconnect fabric, separate from the first high bandwidth chip interconnect fabric, to a neighboring bus architecture, thereby facilitating high speed transport of data to and from the transform chiplet.
  • 19. The computer system of claim 16, wherein at least one of the one or more transform chiplets is electrically coupled via a first high bandwidth chip interconnect fabric to a first neighboring computer memory element and electrically coupled via a second high bandwidth chip interconnect fabric, separate from the first high bandwidth chip interconnect fabric, to a second neighboring computer memory element, thereby facilitating high speed transport of data to and from the transform chiplet.
  • 20. The computer system of claim 16, wherein a transform chiplet is electrically coupled via a plurality of high bandwidth chip interconnect fabrics to a respective plurality of neighboring computer memory elements, in a one-to-many communication/processing pathway, thereby facilitating high speed transport of data to and from the transform chiplet.
Divisions (1)
Number Date Country
Parent 17712137 Apr 2022 US
Child 18809778 US