Multi-chip module (MCM) with scalable high bandwidth memory

Information

  • Patent Grant
  • 12182040
  • Patent Number
    12,182,040
  • Date Filed
    Thursday, August 3, 2023
    a year ago
  • Date Issued
    Tuesday, December 31, 2024
    a month ago
Abstract
Semiconductor devices, packaging architectures and associated methods are disclosed. In one embodiment, a multi-chip module (MCM) is disclosed. The MCM includes a package substrate and an integrated circuit (IC) processor chip disposed on the package substrate. The IC processor chip includes a data interface configured to support N channels. A scalable high bandwidth memory (HBM) is coupled to the IC processor chip. The scalable HBM includes a first HBM device disposed on the package substrate with a first primary data interface that supports a first set of N/2 data channels and a first data transfer rate. A second HBM device is disposed on the package substrate and supports a second set of N/2 data channels and a second data transfer rate. The first HBM device and the second HBM device are configured to collectively support the full N channels and an aggregate data rate that is a sum of the first data rate and the second data rate.
Description
TECHNICAL FIELD

The disclosure herein relates to semiconductor devices, packaging and associated methods.


BACKGROUND

As integrated circuit (IC) chips such as system on chips (SoCs) become larger, the yields realized in manufacturing the chips become smaller. Decreasing yields for larger chips increases overall costs for chip manufacturers. To address the yield problem, chiplet architectures have been proposed that favor a modular approach to SoCs. The solution employs smaller sub-processing chips, each containing a well-defined subset of functionality. Chiplets thus allow for dividing a complex design, such as a high-end processor or networking chip, into several small die instead of one large monolithic die.


One form of memory technology that is employed in certain chiplet-based SoC applications is High Bandwidth Memory (HBM), which has multiple generations that have been standardized by the Joint Electron Device Engineering Council (JEDEC). Each iteration of the standard often involves significant industry investment in packaging infrastructure to support increased channel count, bandwidth, and performance. Transitioning from a legacy HBM standard to a next generation standard may also prove problematic due to uncertain scheduling for the production of new memory technology, often resulting in device availability being relegated to lower frequency bins for indeterminate periods of time. Thus, migrating from one HBM standard to a next generation HBM standard in a chiplet-based multi-chip module (MCM) is typically a costly endeavor.


What is needed is an efficient, robust and cost-efficient way to incorporate next-generation HBM devices into a chiplet architecture.





BRIEF DESCRIPTION OF THE DRAWINGS

Embodiments of the disclosure are illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings and in which like reference numerals refer to similar elements and in which:



FIG. 1 illustrates a high-level generic embodiment of an MCM that employs a scalable HBM memory suitable for HBM device migration between device generations.



FIG. 2 illustrates one specific embodiment of the MCM of FIG. 1, employing a daisy-chained architecture.



FIG. 3 illustrates an MCM with an IC processor device, such as that employed in FIG. 2, redeployed to support a next generation HBM device.



FIG. 4 illustrates an MCM architecture similar to that of FIG. 2, and employing embedded silicon bridges to interconnect various chips.



FIG. 5 illustrates an MCM architecture similar to that of FIG. 2, and employing a passive interposer to interconnect various chips.



FIG. 6 illustrates another specific embodiment of the MCM of FIG. 1, employing a point-to-point architecture.



FIG. 7 illustrates an MCM architecture similar to that of FIG. 6, and employing an active interposer to interconnect various chips.





DETAILED DESCRIPTION

Semiconductor devices, packaging architectures and associated methods are disclosed. In one embodiment, a multi-chip module (MCM) is disclosed that includes a package substrate and an integrated circuit (IC) processor chip disposed on the package substrate. The IC processor chip includes a data interface configured to support N channels. A scalable high bandwidth memory (HBM) is coupled to the IC processor chip. The scalable HBM includes a first HBM device disposed on the package substrate with a first primary data interface that supports a first set of N/2 data channels and a first data transfer rate. A second HBM device is disposed on the package substrate and supports a second set of N/2 data channels and a second data transfer rate. The first HBM device and the second HBM device are configured to collectively support the full N channels and an aggregate data rate that is a sum of the first data rate and the second data rate. By incorporating a scalable HBM memory with individual devices that support less than the total number of desired channels, legacy devices may be employed and configured to collectively provide the total number of desired channels. This may allow for a less-costly migration between the use of legacy HBM devices and next generation HBM devices.


Throughout the disclosure provided herein, the term multi-chip module (MCM) is used to represent a semiconductor device that incorporates multiple semiconductor die or sub-packages in a single unitary package. An MCM may also be referred to as a system in a chip (SiP). With reference to FIG. 1, a block diagram of one embodiment of a multi-chip module (MCM) is shown, generally designated 100. For one embodiment, the MCM 100 includes a package substrate 102. The package substrate forms a support surface and signal routing vehicle for multiple integrated circuit (IC) chips, or chiplets, including an IC processor chip 104, and a scalable high bandwidth memory (HBM) 106. For one embodiment, the scalable HBM memory 106 includes at least a first high-bandwidth memory (HBM) device 108 and a second HBM device 110. Depending on the application, the package substrate 102 may take on one of various forms that more fully described below, such as a low-cost non-silicon substrate, a passive interposer or an active silicon interposer, a silicon bridge, or combination of forms, to name but a few.


With continued reference to FIG. 1, for one embodiment, the IC processor chip 104 takes the form of a computing resource such as a computer processing unit (CPU), graphics processing unit (GPU), or artificial intelligence processing unit. The IC processor chip 104 acts as a host device that periodically accesses the memory resources provided by the scalable HBM 106. A host memory interface 112 is provided on the IC processor chip 104 that supports a set of N independent channels. For one specific embodiment, the host memory interface 112 includes input/output (I/O) circuitry resources that are sufficient to form a set of thirty-two independent memory channels that operate at a collective data rate or bandwidth (BW). The host memory interface 112 also includes a communications interface that defines a host node for communicating with similar circuitry on the first and second HBM devices 108 and 110. For some embodiments, the IC processor chip 104 includes an HBM memory controller 114 to control transfers between core circuitry of the IC processor chip 104 and the scalable HBM memory 106. In other embodiments, the IC processor chip 104 may omit an HBM memory controller, with such functionality being incorporated into the HBM devices 108 and 110 of the scalable HBM memory 106.


Further referring to FIG. 1, for one embodiment, the first HBM device 108 takes the form of a DRAM memory device compliant with a given High Bandwidth Memory (HBM) standard, such as HBM3. The first HBM device 108 includes HBM logic 116 that may take the form of a logic base die or a portion of an active silicon substrate, such as that described more fully below with respect to FIG. 7. Depending on the topology employed by the scalable HBM memory 106, the HBM logic 116 includes a device memory interface 118 that includes device I/O circuitry that interfaces with at least a subset, such as N/2, of the set of the N memory channels provided by the host memory interface 112, and operating at a collective bandwidth of BW/2. Different topologies that incorporate different device interfaces are described more fully below. A stack of dynamic random access memory (DRAM) die 120 is vertically disposed on the HBM logic 116 and interconnected to the logic by, for example, through-silicon vias (TSVs). For embodiments where the IC processor chip 104 omits an HBM memory controller, the HBM logic 116 of the first HBM device 108 may also optionally include an on-chip HBM memory controller 122 to control transfers between the core circuitry of the IC processor chip 104 and the stack of DRAM die 120 of the first HBM device 108.


With continued reference to FIG. 1, the second HBM device 110 is formed similar to the first HBM device 108, including HBM logic 124 that may take the form of a second logic die or a second portion of an active silicon substrate, described more fully below. The HBM logic 124 includes a second device memory interface 126 that includes device I/O circuitry that interfaces with at least a second subset, such as N/2, of the set of the N memory channels provided by the host memory interface 112, and also operating at a collective bandwidth of BW/2. A second stack of dynamic random access memory (DRAM) die 128 is vertically disposed on and interconnected to the HBM logic 124. For embodiments where the IC processor chip 104 omits an HBM memory controller, the HBM logic 124 of the second HBM device 110 may also optionally include an on-chip HBM memory controller 130 to control transfers between the core circuitry of the IC processor chip 104 and the second stack of DRAM die 128 of the second HBM device 110.


Further referring to FIG. 1, the scalable HBM memory 106 may be configured in a variety of ways to suit various applications. Various embodiments for different topologies, such as daisy-chained and point-to-point architectures are described below and shown in FIGS. 2 through 7. At a high level, the first HBM device 108 and the second HBM device 110 are configured to collectively support the N independent channels and an aggregate data rate or bandwidth BW that matches the bandwidth provided by the HBM host memory interface 112. As a result, infrastructure for a next-generation HBM device that may incorporate an integer multiple number of channels and a similar expansion of memory bandwidth over a legacy HBM device may be developed and used with a unique configuration of multiple legacy HBM devices.


While FIG. 1 illustrates a generic architecture for an HBM migration scheme, FIG. 2 illustrates one specific embodiment of the generic architecture by incorporating a daisy-chained topology for an MCM 200 using HBM memory devices similar to those described above. The MCM 200 includes an IC processor chip 202 that includes a host memory interface 204 that supports N channels and exhibits a collective bandwidth BW. A scalable HBM memory 206 is coupled to the IC processor chip 202 and provides respective first and second HBM devices 208 and 210 that are configured to be accessed, and where all memory-related signals to and from the second HBM device 210 run through and are retransmitted by the first HBM device 208. Each memory transaction (read or write) from the IC processor chip 202 is bound for one memory array on one of die memory stacks. If the first HBM device 208 determines the transaction is bound for its memory stack, it need not forward the transaction to the second HBM device 210.


With continued reference to FIG. 2, the IC processor chip 202 and the first and second HBM devices 208 and 210 of the scalable HBM memory 206 may be disposed on a package substrate 207 that takes one of a variety of forms, depending on the application. As one example, the various chips may be mounted on a low-cost non-silicon substrate, a passive interposer, an active silicon interposer, a silicon bridge, or combination of the above.


Further referring to FIG. 2, for one embodiment, the first HBM device 208 includes logic 209 in the form of an HBM base die that is stacked with a plurality of DRAM die 211. The logic 209 employs a first primary device interface 212 that matches the channel count N of the host memory interface 204 of the IC processor chip 202. For one specific embodiment that uses legacy HBM devices in the scalable HBM memory 206, an integer number of legacy HBM device interface circuits, each supporting N/2 channels, may be combined or unified to form the primary device interface 212 for the first HBM device 208 to support the N channels.


With continued reference to FIG. 2, to communicate with the second HBM device 210, the first HBM device 208 employs a second port or secondary device interface 214 that may, as an example, support N or N/2 channels, depending on the application. Communications circuitry 216, forming a portion of the overall switch fabric employed throughout the MCM 200, interconnects the secondary device interface 214 to at least a portion of the primary device interface 212. For one embodiment, the communications circuitry 216 may employ multiplexing circuitry to select which portions of the N channels are passed to and from the first HBM device 208 and the second HBM device 210. In other situations, the communications circuitry 216 may take the form of in-memory processing circuitry, similar to network-on-chip (NoC) circuitry discussed in U.S. patent application Ser. No. 17/994,123, titled “MULTI-CHIP MODULE (MCM) WITH MULTI-PORT UNIFIED MEMORY, filed Nov. 25, 2022, and incorporated by reference herein in its entirety.


In some embodiments, the first HBM device 208 includes a on-chip memory controller 218 that interacts with a portion of the primary device interface 212, such as a first set of N/2 channels, to facilitate transfers between the IC processor chip 202 and the stack of DRAM die 211. A second set of the N/2 channels bypasses the first on-chip memory controller 218 and is routed on-chip to the secondary device interface 214 via the communications circuitry 216. In some circumstances, the first on-chip memory controller 218 may be omitted from the first HBM device 208, and instead a host memory controller 219 may be incorporated on the IC processor chip 202. For some embodiments, where the memory controller is omitted from the first HBM device, buffer circuitry (not shown) may be provided.


Further referring to FIG. 2, the second HBM device 210 is formed similar to the first HBM device 208, such as a replica of the first HBM device, and includes a second logic base die 228 that incorporates a second primary interface 220, and an unused second secondary interface 222. For some embodiments, the second primary and secondary interfaces 220 and 222 may include circuitry to support a same number of channels as those supported by the first primary and secondary interface circuits 212 and 214 of the first HBM device 208. In other embodiments, the second primary and secondary interfaces 220 and 222 may include circuitry to support half the number of channels as those supported by the first primary and secondary interface circuits 212 and 214 of the first HBM device 208. Depending on the application, the second HBM device 210 may or may not include an on-chip memory controller, such as at 224 to facilitate transfers between the IC processor chip 202 and a second stack of DRAM die 226 that is stacked atop the second logic base die 228.


In operation, the MCM 200 provides the infrastructure and resources to support operating N independent memory channels with a scalable HBM memory that utilizes HBM devices that separately support N/2 channels, albeit in a daisy-chained architecture. One specific example involves providing a total of thirty-two channels—with sixteen of the channels provided by the first HBM device 208, and the other sixteen channels provided in a daisy-chained manner by the second HBM device 210 via the on-chip retransmitting/repeating feature provided by the first HBM device 208. Additionally, by coordinating memory accesses to the first and second HBM devices in a concurrent manner, where both of the HBM devices 208 and 210 are accessed during respective time intervals that at least partially overlap, the aggregate memory bandwidth of the scalable HBM memory 206 for the thirty-two channels may be doubled in comparison to what the bandwidth would be for sixteen channels.


While FIG. 2 illustrates one embodiment of an MCM that utilizes a scalable HBM memory 206 that supports N channels in a legacy mode of operation using legacy devices, such as HBM3 devices, that each support N/2 channels, FIG. 3 illustrates an embodiment of an MCM 300 that supports a next generation mode of operation by reusing much of the infrastructure designed to support interoperability between, for example, the IC processor chip 202 and the scalable HBM memory 206 of FIG. 2. However, instead of employing the scalable HBM memory 206, a next generation HBM memory device 302 that individually supports N channels may be substituted as the HBM memory resource. The next generation HBM device 302 would be expected to be similar to prior generation devices, with a logic base die 304 that is coupled to a stack of memory die 306. The logic base die 304 would include a full N-channel device interface 308 to generally match the N-channel host interface 204 of the IC processor chip 202. A memory controller 1210 may optionally be included in the logic base die 304, unless it is incorporated in the IC processor chip 202. For some embodiments, additional capacity in the form of an additional next generation HBM device 314 may be provided in a daisy-chained manner, where the first HBM device 304 includes NoC circuitry to manage traffic between the two next generation HBM devices 302 and 314. The reused infrastructure and the additional capacity option may be employed in any of the MCM embodiments described herein, such as those shown in FIGS. 2, and 4-7.


Further referring to FIG. 3, the IC processor chip 202 may retain and reuse much of the infrastructure provided in supporting the legacy mode of operation, such as that shown in FIG. 2, thus reducing the costs and time typically involved in migrating from legacy memory devices to next generation devices. Programmable logic (not shown) in the IC processor chip 202 may be updated during MCM manufacture or during initialization to configure the IC processor chip 202 to operate in the legacy mode or the next generation mode. While the discussion above notes the reusability of a same IC processor chip 202 for both MCM embodiments 200 and 300, for some embodiments, the IC processor chip 202 employed in the MCM 300 may incorporate slight modifications that may be accomplished with far lower cost than a full redesign to support a next generation HBM architecture.


For some embodiments, the IC processor chip 202 employs unique interface circuitry that allows for the use of high-performance links that are compatible with a cost-efficient standard organic build-up package substrate, such as at 312. Such interface circuitry and associated links are disclosed in U.S. patent application Ser. No. 18/092,647, filed Jan. 3, 2023, titled “CHIPLET GEARBOX FOR LOW-COST MULTI-CHIP MODULE APPLICATIONS”, and incorporated by reference in its entirety.


While the use of standard non-silicon substrates may be beneficial for certain applications, other applications may utilize interface circuitry and associated links that benefit from a silicon-based primary or secondary substrate that is more suitable for fine-pitch routing. FIG. 4 illustrates one embodiment of an MCM 400 that is similar to the MCM 300 of FIG. 2, but incorporating one or more embedded silicon bridges having multiple routing layers, such as first and second embedded multi-die interconnect bridges (EMIBs) 402 and 404. The first bridge 402 spans a first distance between the first IC processor chip 202 and the first HBM device 208 to connect interfaces 204 and 212, while the second bridge 404 spans at least the distance between the first HBM device 208 and the second HBM device 210 to connect interfaces 214 and 220. The use of bridges, instead of one large silicon interposer, enables a high-level of interconnect density where it is needed—between the respective interfaces of multiple chips. For some embodiments, the bridges 402 and 404 may be embedded in a package substrate 207 to reduce costs even further.



FIG. 5 illustrates a further embodiment of an MCM 500 that is similar to the MCM 200 of FIG. 2 but incorporates a passive silicon interposer 502 that forms at least a portion of a package substrate 504. Unlike an active interposer, which includes active electronic circuitry, the passive silicon interposer 502 provides a silicon-based support structure formed with finely-pitched interconnect routing paths, and without active transistor circuitry. The use of such an interposer allows for chip contact pitch density for the chip interfaces to be on the order of approximately 30 to 70 micrometers.


While the embodiments illustrated in FIGS. 2-5 illustrate the first and second HBM devices being coupled to the IC processor chip in a daisy-chained architecture, other embodiments may connect each HBM device directly to an IC processor chip in a point-to-point architecture. FIG. 6 illustrates an MCM 600 having an IC processor chip 602 with a host interface 604 formed to support N channels. A first portion of the host interface 604 is directly connected to a first HBM device 606 via a first set of N/2 point-to-point links 607. A second portion of the host interface 604 is directly connected to a second HBM device 608 via a second set of N/2 point-to-point links 609. For one embodiment, a portion of the second set of N/2 links 609 is routed beneath the first HBM device 606. Together, the first and second HBM devices 606 and 608 form a scalable HBM memory 610. In one embodiment, the IC processor chip 602, the first HBM device 606 and the second HBM device 608 are all mounted on a package substrate 611. For some embodiments, all or a portion of the package substrate 611 may be formed of an organic material. In other embodiments, all or a portion of the package substrate 611 may be formed of an active or passive silicon-based material, or include silicon bridge structures such as in FIG. 4.


Further referring to FIG. 6, for one embodiment the first HBM device 606 includes a first logic base die 611 having a first device interface 612 that is directly coupled to the first set of N/2 channels 607. In some embodiments, the first HBM device 606 may also include an on-chip memory controller 614, while in other embodiments the memory controller 614 is disposed in the IC processor chip 602 to control memory accesses for both HBM devices 606 and 608. The first logic base die couples to a first stack of memory chips 613. The second HBM device 608 is formed similar to the first HBM device 606, with a second logic base die 615 having a second device interface 616 that is directly coupled to the second set of N/2 channels 609. The second logic base die 615 couples to a second stack of memory chips 617. For situations where the memory controller 614 is omitted from the IC processor chip 602, the second HBM memory device 608 includes an on-chip memory controller 618 that processes transactions destined for the memory stack 617 associated with it, while the memory controller 614 of the first HBM device 606 processes transactions destined for the memory stack 613 of the first HBM device 606.


In operation, the MCM 600 provides the infrastructure and resources to support operating N independent memory channels with the scalable HBM memory 610 that utilizes HBM devices that separately support N/2 channels, albeit in a point-to-point architecture. One specific example involves providing a total of thirty-two channels—with sixteen of the channels provided by the first HBM device 606 via the first set of N/2 point-to-point links 607 with the IC processor chip 602, and the other sixteen channels provided in a point-to-point manner by the second HBM device 608 via the second set of point-to-point links 609. Additionally, by coordinating memory accesses to the first and second HBM devices 606 and 608 in a concurrent manner, where both of the HBM devices are accessed during respective time intervals that at least partially overlap, the aggregate memory bandwidth of the scalable HBM memory 610 for the thirty-two channels may be doubled in comparison to what the bandwidth would be for sixteen channels.



FIG. 7 illustrates a further embodiment for an MCM 700 that employs a point-to-point architecture. The MCM includes an IC processor chip 702 that incorporates a host interface 703 to support N channels. A full set of N links 704 connects the IC processor chip 702 to a scalable HBM memory 706. For one embodiment, the scalable HBM memory 706 includes an active silicon substrate 708 formed with a device interface 710 that supports N channels and which is connected to the N links 704. For one embodiment, a memory controller 712 is formed on the active substrate 708 to control a first HBM memory stack 714 and a second HBM memory stack 716. Like other embodiments described above, the memory controller 712 may instead be disposed on the IC processor chip 702. A first set of on-substrate routing paths 718 are formed on the active substrate 708 to connect either the device interface 710 (in the event the memory controller is employed on the IC processor chip 702), or the memory controller 712 to the first HBM memory stack 714 in a point-to-point manner. A second set of on-substrate routing paths 720 are formed on the active substrate 708 to connect either the device interface 710 (in the event the memory controller is employed on the IC processor chip 702), or the memory controller 712 to the second HBM memory stack 716 in a point-to-point manner.


Further referring to FIG. 7, the active silicon substrate 708 essentially functions as a joint HBM base die for the first and second HBM memory stacks 714 and 716. For some embodiments, the active silicon substrate 708 may form a secondary substrate for mounting the HBM memory stacks 714 and 716, and formed on or embedded in a larger primary substrate 718 that may be constructed of organic material or non-organic material.


When received within a computer system via one or more computer-readable media, such data and/or instruction-based expressions of the above described circuits may be processed by a processing entity (e.g., one or more processors) within the computer system in conjunction with execution of one or more other computer programs including, without limitation, net-list generation programs, place and route programs and the like, to generate a representation or image of a physical manifestation of such circuits. Such representation or image may thereafter be used in device fabrication, for example, by enabling generation of one or more masks that are used to form various components of the circuits in a device fabrication process.


In the foregoing description and in the accompanying drawings, specific terminology and drawing symbols have been set forth to provide a thorough understanding of the present invention. In some instances, the terminology and symbols may imply specific details that are not required to practice the invention. For example, any of the specific numbers of bits, signal path widths, signaling or operating frequencies, component circuits or devices and the like may be different from those described above in alternative embodiments. Also, the interconnection between circuit elements or circuit blocks shown or described as multi-conductor signal links may alternatively be single-conductor signal links, and single conductor signal links may alternatively be multi-conductor signal links. Signals and signaling paths shown or described as being single-ended may also be differential, and vice-versa. Similarly, signals described or depicted as having active-high or active-low logic levels may have opposite logic levels in alternative embodiments. Component circuitry within integrated circuit devices may be implemented using metal oxide semiconductor (MOS) technology, bipolar technology or any other technology in which logical and analog circuits may be implemented. With respect to terminology, a signal is said to be “asserted” when the signal is driven to a low or high logic state (or charged to a high logic state or discharged to a low logic state) to indicate a particular condition. Conversely, a signal is said to be “deasserted” to indicate that the signal is driven (or charged or discharged) to a state other than the asserted state (including a high or low logic state, or the floating state that may occur when the signal driving circuit is transitioned to a high impedance condition, such as an open drain or open collector condition). A signal driving circuit is said to “output” a signal to a signal receiving circuit when the signal driving circuit asserts (or deasserts, if explicitly stated or indicated by context) the signal on a signal line coupled between the signal driving and signal receiving circuits. A signal line is said to be “activated” when a signal is asserted on the signal line, and “deactivated” when the signal is deasserted. Additionally, the prefix symbol “/” attached to signal names indicates that the signal is an active low signal (i.e., the asserted state is a logic low state). A line over a signal name (e.g., ‘<signal name>’) is also used to indicate an active low signal. The term “coupled” is used herein to express a direct connection as well as a connection through one or more intervening circuits or structures. Integrated circuit device “programming” may include, for example and without limitation, loading a control value into a register or other storage circuit within the device in response to a host instruction and thus controlling an operational aspect of the device, establishing a device configuration or controlling an operational aspect of the device through a one-time programming operation (e.g., blowing fuses within a configuration circuit during device production), and/or connecting one or more selected pins or other contact structures of the device to reference voltage lines (also referred to as strapping) to establish a particular device configuration or operation aspect of the device. The term “exemplary” is used to express an example, not a preference or requirement.


While the invention has been described with reference to specific embodiments thereof, it will be evident that various modifications and changes may be made thereto without departing from the broader spirit and scope of the invention. For example, features or aspects of any of the embodiments may be applied, at least where practicable, in combination with any other of the embodiments or in place of counterpart features or aspects thereof. Accordingly, the specification and drawings are to be regarded in an illustrative rather than a restrictive sense.

Claims
  • 1. A chiplet-based multi-chip module (MCM), comprising: a package substrate comprising a base substrate, and an active base die formed on the base substrate, the active base die comprising active circuitry;an integrated circuit (IC) processor chiplet coupled to the package substrate, the IC processor chiplet comprising a data interface configured to couple to N channels;a scalable high bandwidth memory (HBM) coupled to the IC processor chiplet and comprising a first HBM device coupled to the active base die, the active base die comprising active circuitry comprising a first primary data interface to couple to the N channels, wherein the first HBM device is dedicated to a first subset the N channels and operative at a first data transfer rate, the first HBM device comprising a first stack of memory die;a second HBM device coupled to the active base die and dedicated to a second subset of the N channels and operative at a second data transfer rate, the second HBM device comprising a second stack of memory die separate from the first stack of memory die; andwherein the IC processor chiplet, the first HBM device and the second HBM device are placed in a row and the first HBM device is located between the IC processor chiplet and the second HBM device, and the first HBM device and the second HBM device are configured to collectively couple to the N channels and an aggregate data rate that is at most a sum of the first data transfer rate and the second data transfer rate.
  • 2. The chiplet-based MCM of claim 1, further comprising: memory control circuitry coupled to the active base die.
  • 3. The chiplet-based MCM of claim 1, wherein: the active base die comprises circuitry to couple to the N channels for controlling transactions between the IC processor chiplet and the respective first stack of memory die and the second stacks of memory die.
  • 4. The chiplet-based MCM of claim 1, wherein: the package substrate comprises a base substrate, and at least one silicon bridge coupled to the base substrate to interconnect the IC processor chiplet with the first HBM device and the second HBM device.
  • 5. The chiplet-based MCM of claim 1, wherein: the package substrate comprises an organic substrate.
  • 6. The chiplet-based MCM of claim 1, further comprising: multiple bidirectional links coupling the IC processor chiplet to the scalable HBM.
  • 7. The chiplet-based MCM of claim 3, wherein: the active base die comprises memory control circuitry.
  • 8. An integrated circuit (IC) processor chiplet, comprising: a high bandwidth memory (HBM) data interface that is configurable to operate in one of a first HBM mode of operation or a second HBM mode of operation;wherein during the first HBM mode of operation, the HBM data interface is configured to couple to multiple links forming N channels and to operate at an aggregate data rate R, the multiple links to couple the IC processor chiplet to at least two DRAM stacks on an active base die, each of the at least two DRAM stacks configured to couple to N/2 channels; andwherein during the second HBM mode of operation, the HBM data interface is configured to couple to the multiple links forming the N channels and to operate at the aggregate data rate R, the multiple links to couple the IC processor chiplet to one of the at least two HBM stacks at a time, each of the at least two HBM stacks configured to couple to the N channels.
  • 9. The IC processor chiplet of claim 8, wherein: the HBM data interface is configured to couple to the at least two HBM stacks via the multiple links in a daisy-chain architecture.
  • 10. The IC processor chiplet of claim 8, wherein: the HBM data interface is configured to couple to the at least two HBM stacks via the multiple links in a point-to-point architecture.
  • 11. The IC processor chiplet of claim 8, wherein: the high bandwidth memory (HBM) data interface comprises input/output (I/O) circuitry to couple to the multiple links forming the N channels, the I/O circuitry comprising simultaneous bidirectional transceiver circuitry to transfer data along the N channels.
  • 12. A scalable high-bandwidth memory (HBM), comprising: a first HBM device coupled to an active base die that comprises active circuitry and comprising a first primary data interface to couple to an integrated circuit (IC) processor chiplet, the first primary data interface to couple to a first set of at least N/2 data channels and to operate at a first data transfer rate, the first HBM device comprising a first stack of memory die;a second HBM device coupled to the active base die and supporting a second set of at least N/2 data channels and to operate at a second data transfer rate, the second HBM device comprising a second stack of memory die that is separate from the first stack of memory die; andwherein the first HBM device and the second HBM device are configured to collectively couple to N channels and to operate at an aggregate data rate that is at most a sum of the first data transfer rate and the second data transfer rate.
  • 13. The scalable HBM memory of claim 12, further comprising: a first memory control circuit disposed in the first HBM device; anda second memory control circuit disposed in the second HBM device.
  • 14. The chiplet-based MCM of claim 1, wherein the active base die further comprises: active base die circuitry to control transactions between the IC processor chiplet and the respective first stack of memory die and the second stack of memory die, the active base die circuitry comprising a device data interface that is coupled to the data interface of the IC processor chiplet, the device data interface configured to couple to the N channels.
  • 15. The chiplet-based MCM of claim 14, wherein: the first HBM device, the second HBM device, and the device data interface are positioned on the active base die circuitry in a row.
  • 16. The chiplet-based MCM of claim 1, wherein: the package substrate comprises a base substrate and at least one silicon bridge formed on the base substrate to interconnect the IC processor chiplet with the scalable high bandwidth memory (HBM).
  • 17. The chiplet-based MCM of claim 1, wherein: the package substrate comprises a passive silicon substrate or silicon interposer.
  • 18. The IC processor chiplet of claim 8, wherein: the HBM data interface is configured to couple to the at least two HBM stacks via an active base die.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a Non-Provisional which claims priority to U.S. Provisional Application No. 63/471,234, filed Jun. 5, 2023, titled HBM3 TO HBM4 MIGRATION METHOD, which is incorporated herein by reference in its entirety.

US Referenced Citations (188)
Number Name Date Kind
4334305 Girardi Jun 1982 A
5396581 Mashiko Mar 1995 A
5677569 Choi Oct 1997 A
5892287 Hoffman Apr 1999 A
5910010 Nishizawa Jun 1999 A
6031729 Berkely Feb 2000 A
6055235 Blanc Apr 2000 A
6417737 Moloudi Jul 2002 B1
6492727 Nishizawa Dec 2002 B2
6690742 Chan Feb 2004 B2
6721313 Van Duyne Apr 2004 B1
6932618 Nelson Aug 2005 B1
7027529 Ohishi Apr 2006 B1
7248890 Raghavan Jul 2007 B1
7269212 Chau Sep 2007 B1
7477615 Oshita Jan 2009 B2
7535958 Best May 2009 B2
7593271 Ong Sep 2009 B2
7701957 Bicknell Apr 2010 B1
7907469 Sohn et al. Mar 2011 B2
7978754 Yeung Jul 2011 B2
8004330 Acimovic Aug 2011 B1
8024142 Gagnon Sep 2011 B1
8121541 Rofougaran Feb 2012 B2
8176238 Yu et al. May 2012 B2
8468381 Jones Jun 2013 B2
8483579 Fukuda Jul 2013 B2
8546955 Wu Oct 2013 B1
8704364 Banijamali et al. Apr 2014 B2
8861573 Chu Oct 2014 B2
8948203 Nolan Feb 2015 B1
8982905 Kamble Mar 2015 B2
9088334 Chakraborty Jul 2015 B2
9106229 Hutton Aug 2015 B1
9129935 Chandrasekar Sep 2015 B1
9294313 Prokop Mar 2016 B2
9349707 Sun May 2016 B1
9379878 Lugthart Jun 2016 B1
9432298 Smith Aug 2016 B1
9558143 Leidel Jan 2017 B2
9832006 Bandi Nov 2017 B1
9842784 Nasrullah Dec 2017 B2
9843538 Woodruff Dec 2017 B2
9886275 Carlson Feb 2018 B1
9934842 Mozak Apr 2018 B2
9961812 Suorsa May 2018 B2
9977731 Pyeon May 2018 B2
10171115 Shirinfar Jan 2019 B1
10402363 Long et al. Sep 2019 B2
10410694 Arbel Sep 2019 B1
10439661 Heydari Oct 2019 B1
10642767 Farjadrad May 2020 B1
10678738 Dai Jun 2020 B2
10735176 Heydari Aug 2020 B1
10748852 Sauter Aug 2020 B1
10769073 Desai Sep 2020 B2
10803548 Matam et al. Oct 2020 B2
10804204 Rubin et al. Oct 2020 B2
10825496 Murphy Nov 2020 B2
10826536 Beukema Nov 2020 B1
10855498 Farjadrad Dec 2020 B1
10935593 Goyal Mar 2021 B2
11088876 Farjadrad Aug 2021 B1
11100028 Subramaniam Aug 2021 B1
11164817 Rubin et al. Nov 2021 B2
11204863 Sheffler Dec 2021 B2
11581282 Elshirbini Feb 2023 B2
11669474 Lee Jun 2023 B1
11789649 Chatterjee et al. Oct 2023 B2
11841815 Farjadrad Dec 2023 B1
11842986 Farjadrad Dec 2023 B1
11855043 Farjadrad Dec 2023 B1
11855056 Rad Dec 2023 B1
11892242 Mao Feb 2024 B2
11893242 Farjadrad Feb 2024 B1
11983125 Soni May 2024 B2
12001355 Dreier Jun 2024 B1
20020122479 Agazzi Sep 2002 A1
20020136315 Chan Sep 2002 A1
20040088444 Baumer May 2004 A1
20040113239 Prokofiev Jun 2004 A1
20040130347 Moll Jul 2004 A1
20040156461 Agazzi Aug 2004 A1
20050041683 Kizer Feb 2005 A1
20050134306 Stojanovic Jun 2005 A1
20050157781 Ho Jul 2005 A1
20050205983 Origasa Sep 2005 A1
20060060376 Yoon Mar 2006 A1
20060103011 Andry May 2006 A1
20060158229 Hsu Jul 2006 A1
20060181283 Wajcer Aug 2006 A1
20060188043 Zerbe Aug 2006 A1
20060250985 Baumer Nov 2006 A1
20060251194 Bublil Nov 2006 A1
20070281643 Kawai Dec 2007 A1
20080063395 Royle Mar 2008 A1
20080143422 Lalithambika Jun 2008 A1
20080186987 Baumer Aug 2008 A1
20080222407 Carpenter Sep 2008 A1
20090113158 Schnell Apr 2009 A1
20090154365 Diab Jun 2009 A1
20090174448 Zabinski Jul 2009 A1
20090220240 Abhari Sep 2009 A1
20090225900 Yamaguchi Sep 2009 A1
20090304054 Tonietto Dec 2009 A1
20100177841 Yoon Jul 2010 A1
20100197231 Kenington Aug 2010 A1
20100294547 Hatanaka Nov 2010 A1
20110029803 Redman-White Feb 2011 A1
20110038286 Ta Feb 2011 A1
20110167297 Su Jul 2011 A1
20110187430 Tang Aug 2011 A1
20110204428 Erickson Aug 2011 A1
20110267073 Chengson Nov 2011 A1
20110293041 Luo Dec 2011 A1
20120082194 Tam Apr 2012 A1
20120182776 Best Jul 2012 A1
20120192023 Lee Jul 2012 A1
20120216084 Chun Aug 2012 A1
20120327818 Takatori Dec 2012 A1
20130181257 Ngai Jul 2013 A1
20130222026 Havens Aug 2013 A1
20130249290 Buonpane Sep 2013 A1
20130285584 Kim Oct 2013 A1
20140016524 Choi Jan 2014 A1
20140048947 Lee Feb 2014 A1
20140126613 Zhang May 2014 A1
20140192583 Rajan Jul 2014 A1
20140269860 Brown Sep 2014 A1
20140269983 Baeckler Sep 2014 A1
20150012677 Nagarajan Jan 2015 A1
20150172040 Pelekhaty Jun 2015 A1
20150180760 Rickard Jun 2015 A1
20150206867 Lim Jul 2015 A1
20150271074 Hirth Sep 2015 A1
20150326348 Shen Nov 2015 A1
20150358005 Chen Dec 2015 A1
20160056125 Pan Feb 2016 A1
20160071818 Wang Mar 2016 A1
20160111406 Mak Apr 2016 A1
20160217872 Hossain Jul 2016 A1
20160294585 Rahman Oct 2016 A1
20170317859 Hormati Nov 2017 A1
20170331651 Suzuki Nov 2017 A1
20180010329 Golding, Jr. Jan 2018 A1
20180082981 Gowda Mar 2018 A1
20180137005 Wu May 2018 A1
20180175001 Pyo Jun 2018 A1
20180190635 Choi Jul 2018 A1
20180210830 Malladi et al. Jul 2018 A1
20180315735 Delacruz Nov 2018 A1
20190044764 Hollis Feb 2019 A1
20190058457 Ran Feb 2019 A1
20190108111 Levin Apr 2019 A1
20190198489 Kim Jun 2019 A1
20190319626 Dabral Oct 2019 A1
20200051961 Rickard Feb 2020 A1
20200105718 Collins et al. Apr 2020 A1
20200257619 Sheffler Aug 2020 A1
20200373286 Dennis Nov 2020 A1
20210056058 Lee Feb 2021 A1
20210082875 Nelson Mar 2021 A1
20210117102 Grenier Apr 2021 A1
20210149763 Ranganathan May 2021 A1
20210181974 Ghosh Jun 2021 A1
20210183842 Fay Jun 2021 A1
20210193567 Cheah et al. Jun 2021 A1
20210225827 Lanka Jul 2021 A1
20210258078 Meade Aug 2021 A1
20210311900 Malladi Oct 2021 A1
20210365203 O Nov 2021 A1
20220051989 Agarwal Feb 2022 A1
20220121381 Brewer Apr 2022 A1
20220159860 Winzer May 2022 A1
20220179792 Banerjee Jun 2022 A1
20220222198 Lanka Jul 2022 A1
20220223522 Scearce Jul 2022 A1
20220237138 Lanka Jul 2022 A1
20220254390 Gans Aug 2022 A1
20220327276 Seshan Oct 2022 A1
20220334995 Das Sharma Oct 2022 A1
20220342840 Das Sharma Oct 2022 A1
20230039033 Zarkovsky Feb 2023 A1
20230068802 Wang Mar 2023 A1
20230090061 Zarkovsky Mar 2023 A1
20230181599 Erickson May 2023 A1
20230359579 Madhira Nov 2023 A1
20240273041 Lee Aug 2024 A1
Non-Patent Literature Citations (15)
Entry
Farjadrad et al., “A Bunch of Wires (BOW) Interface for Inter-Chiplet Communication”, 2019 IEEE Symposium on High-Performance Interconnects (HOTI), pp. 27-30, Oct. 2019.
Universal Chiplet Interconnect Express (UCle) Specification Rev. 1.0, Feb. 24, 2022.
U.S. Appl. No. 16/812,234; Mohsen F. Rad; filed Mar. 6, 2020.
Kurt Lender et al., “Questions from the Compute Express Link Exploring Coherent Memory and Innovative Cases Webinar”, Apr. 13, 2020, CXL Consortium, pp. 1-6.
Planet Analog, “The basics of SerDes (serializers/deserializers) for interfacing”, Dec. 1, 2020, Planet Analog, as preserved by the internet Archive, pp. 1-9.
Block Memory Generator v8.2 LogiCORE IP Product Guide Vivado Design Suite; Xilinx; Apr. 1, 2015.
Kurt Lender et al., “Questions from the Compute Express Link Exploring Coherent Memory and Innovative Cases Webinar”, Apr. 13, 2020, CXL consortium.
Planet Analog, “The basics of SerDes (serializers/deserializers) for interfacing”, Dec. 1, 2020, Planet Analog.
“Hot Chips 2017: Intel Deep Dives Into EMIB”, TomsHardware.com; Aug. 25, 2017.
“Using Chiplet Encapsulation Technology to Achieve Processing-In-Memory Functions”; Micromachines 2022, 13, 1790; https://www.mdpi.com/journal/micromachines; Tian et al.
“Multiport memory for high-speed interprocessor communication in MultiCom;” Scientia Iranica, vol. 8, No. 4, pp. 322-331; Sharif University of Technology, Oct. 2001; Asgari et al.
Universal Chiplet Interconnect Express (UCIe) Specification, Revision 1.1, Version 1.0, Jul. 10, 2023.
Hybrid Memory Cube Specification 2.1, Hybrid Memory Cube Consortium, HMC-30G-VSR PHY, 2014.
“Using Dual Port Memory as Interconnect”, EE Times, Apr. 26, 2005, Daniel Barry.
Quartus II Handbook Version 9.0 vol. 4: SOPC Builder; “System Interconnect Fabric for Memory-Mapped Interfaces”; Mar. 2009.
Provisional Applications (1)
Number Date Country
63471234 Jun 2023 US