Embodiments of the disclosure relate generally to memory sub-systems, and more specifically, relate to prioritized power budget arbitration for multiple concurrent memory access operations.
A memory sub-system can include one or more memory devices that store data. The memory devices can be, for example, non-volatile memory devices and volatile memory devices. In general, a host system can utilize a memory sub-system to store data at the memory devices and to retrieve data from the memory devices.
The present disclosure will be understood more fully from the detailed description given below and from the accompanying drawings of various embodiments of the disclosure.
Aspects of the present disclosure are directed to prioritized power budget arbitration for multiple concurrent memory access operations. One example of non-volatile memory devices is a negative-and (NAND) memory device. A memory device, to include each die of a multi-dice memory device, can be made up of bits arranged in a two-dimensional or a three-dimensional grid of memory cells. One or more physical blocks of memory cells can be grouped together to form a plane of the memory device in order to allow concurrent operations to take place on each plane, where these physical blocks are made up of groups of pages of memory cells.
Each memory die can include circuitry that performs concurrent memory page accesses of two or more memory planes. For example, each memory die can include multiple access line driver circuits and power circuits that can be shared by the planes of each memory die to facilitate concurrent access of pages of two or more memory planes, including different page types. For ease of description, these circuits can be generally referred to as independent plane driver circuits. Control logic on each die of the memory device includes a number of separate processing threads to perform concurrent memory access operations (e.g., read operations, program operations, and erase operations). For example, each processing thread corresponds to a respective memory plane and utilizes the associated independent plane driver circuits to perform the memory access operations on the respective memory plane. As these processing threads operate independently, the power usage and requirements associated with each processing thread also varies.
The capacitive loading of three-dimensional memory is generally large and may continue to grow as process scaling continues. Various access lines, data lines and voltage nodes can be charged or discharged very quickly during sense (e.g., read or verify), program, and erase operations so that memory array access operations can meet the performance specifications that are often required to satisfy data throughput targets as might be dictated by customer requirements or industry standards, for example. For sequential read or programming, multi-plane operations are often used to increase the system throughput. As a result, a typical memory die can have a high peak current usage, which might be four to five times the average current amplitude. Thus, with such a high average market requirement of total current usage budget, it can become challenging to operate more than four memory dice concurrently, for example.
A variety of techniques have been utilized to manage power consumption of memory sub-systems containing multiple memory dice, many of which rely on a memory sub-system controller to stagger the activity of the memory dice seeking to avoid performing high power portions of access operations concurrently in more than one memory dice. Further, as additional processing threads are utilized on each individual memory die (e.g., 4, 6, or 8 processing threads), these power management techniques are not adequate to account for the added complexity associated with budgeting current usage within each individual memory die.
Aspects of the present disclosure address the above and other deficiencies by providing prioritized power budget arbitration for multiple concurrent access operations in a memory device of a memory sub-system. In some embodiments, the memory device includes multiple dice, each die including multiple processing threads configured to perform the concurrent memory access operations, e.g., on corresponding memory planes of the memory dice. Each memory die further includes a threads manager and a peak power manager (PPM) that are together configured to perform prioritized power budget arbitration for the multiple processing threads on respective memory die of the multiple memory dice.
In these embodiments, the memory sub-system employs a token-based round robin protocol, whereby each PPM rotates (e.g., after a set number of cycles of a shared clock signal) as a holder of the token and broadcasts a quantized current budget to be consumed by its respective memory die during a given time period. The other PPMs on each other memory die receive this broadcast information, and thus, can determine an available current budget in the memory device during the time period. While holding the token, a PPM can request a certain amount of current for its respective memory die up to the available current budget in the memory device and based on an amount of current being consumed by the other memory dice of the memory device. As described in further detail below, the PPM can employ a number of different techniques to allocate the requested current among the multiple processing threads of the respective memory die, at least some of which include prioritized management of multiple concurrent processing threads.
In at least some embodiments, each die also includes a priority ring counter and a data structure (such as a lookup table) where the data structure is to store an association between a value of the priority ring counter and a subset of the multiple processing threads. Each die can further include a threads manager configured to manage the multiple processing threads presented to the PPM for power allocation. In these embodiments, the threads manager can increment the value of the priority ring counter before a power management cycle. Each new counter value changes the subset of the multiple processing threads under consideration for power allocation, and thus simplifies the number of processing threads that the PPM can manage concurrently. The threads manager can also identify one or more prioritized processing threads within the subset of the multiple processing threads, and provide identification of prioritization with the one or more prioritized processing threads to the PPM. The PPM can then prioritize allocation requests when the die of the PPM holds the token, e.g., by prioritizing allocation of power to the one or more prioritized processing threads located within the subset of the multiple processing threads during the power management cycle. The PPM of the die can also check power allocation to the one or more prioritized processing threads against a power (e.g., current) budget available and thus avoid going over budget despite prioritization of some processing threads over others.
In at least some embodiments, the PPM of a die can also manage shifts between non-prioritized management of subsets of the multiple processing threads and prioritized management of the subsets of the multiple processing threads. For example, the PPM can start a timer while a non-prioritized processing thread (e.g., an erase operation or a program operation) is running and in response to detecting allocation of the power also to a prioritized thread of the multiple processing threads (e.g., a read operation or a program operation). The timer can track a predetermined amount of time in order to ensure that prioritized allocation requests do not starve the non-prioritized processing thread of current needed to complete processing. Thus, if the timer expires while the non-prioritized processing thread is still running, the control logic can force a transition back to allocating power between subsets of the processing threads based on increments to the value of a non-priority counter, e.g., based on non-prioritized power allocation.
Advantages of this approach include, but are not limited to, an effective power management scheme for a multi-dice memory sub-system where each memory die supports multiple processing threads operating concurrently. The disclosed techniques allow support for independent parallel plane access in a memory device with significantly reduced hardware resources in the memory sub-system. This approach is highly scalable as the number of processing threads increases and does not rely on external controller intervention. Further, by prioritizing power allocation to some processing threads depending on which subset of multiple processing threads is being managed, memory operations that should be processed quickly (such as read operations and some program operations) can be prioritized over memory operations that can be performed more slowly (such as erase operations and some program operations). The allocated power to the prioritized processing threads can still be checked against a power (e.g., current) budget to ensure not exceeding the budget. Further, as mentioned, use of the timer and shifting protocols can ensure that non-prioritized processing threads are not starved of power budget. Thus, the overall performance and quality of service provided by each memory die is improved.
The memory device(s) 130 can be non-volatile memory device(s). One example of non-volatile memory devices is a negative-and (NAND) memory device. A non-volatile memory device is a package of one or more dice or logical unit (LUNs). Thus, each memory device 130 can be a die (or LUN) or can be a multi-dice package that includes multiple dice (or LUNs) on a chip, e.g., an integrated circuit package of dice. Each die can include one or more planes. Planes can be grouped into logic units (LUN). For some types of non-volatile memory devices (e.g., NAND devices), each plane includes a set of physical blocks. Each block includes a set of pages. Each page includes a set of memory cells (“cells”). A cell is an electronic circuit that stores information. Depending on the cell type, a cell can store one or more bits of binary information, and has various logic states that correlate to the number of bits being stored. The logic states can be represented by binary values, such as “0” and “1”, or combinations of such values.
Each memory device 130 can be made up of bits arranged in a two-dimensional or three-dimensional grid, also referred to as a memory array. Memory cells are etched onto a silicon wafer in an array of columns (also hereinafter referred to as bitlines) and rows (also hereinafter referred to as wordlines). A wordline can refer to one or more rows of memory cells of a memory device that are used with one or more bitlines to generate the address of each of the memory cells. The intersection of a bitline and wordline constitutes the address of the memory cell.
A memory sub-system 110 can be a storage device, a memory module, or a combination of a storage device and memory module. Examples of a storage device include a solid-state drive (SSD), a flash drive, a universal serial bus (USB) flash drive, an embedded Multi-Media Controller (eMMC) drive, a Universal Flash Storage (UFS) drive, a secure digital (SD) card, and a hard disk drive (HDD). Examples of memory modules include a dual in-line memory module (DIMM), a small outline DIMM (SO-DIMM), and various types of non-volatile dual in-line memory modules (NVDIMMs).
The computing system 100 can be a computing device such as a desktop computer, laptop computer, network server, mobile device, a vehicle (e.g., airplane, drone, train, automobile, or other conveyance), Internet of Things (IoT) enabled device, embedded computer (e.g., one included in a vehicle, industrial equipment, or a networked commercial device), or such computing device that includes memory and a processing device.
The computing system 100 can include a host system 120 that is coupled to one or more memory sub-systems 110. In some embodiments, the host system 120 is coupled to multiple memory sub-systems 110 of different types.
The host system 120 can include a processor chipset and a software stack executed by the processor chipset. The processor chipset can include one or more cores, one or more caches, a memory controller (e.g., NVDIMM controller), and a storage protocol controller (e.g., PCIe controller, SATA controller). The host system 120 uses the memory sub-system 110, for example, to write data to the memory sub-system 110 and read data from the memory sub-system 110.
The host system 120 can be coupled to the memory sub-system 110 via a physical host interface. Examples of a physical host interface include, but are not limited to, a serial advanced technology attachment (SATA) interface, a peripheral component interconnect express (PCIe) interface, universal serial bus (USB) interface, Fibre Channel, Serial Attached SCSI (SAS), a double data rate (DDR) memory bus, Small Computer System Interface (SCSI), a dual in-line memory module (DIMM) interface (e.g., DIMM socket interface that supports Double Data Rate (DDR)), etc. The physical host interface can be used to transmit data between the host system 120 and the memory sub-system 110. The host system 120 can further utilize an NVM Express (NVMe) interface to access components (e.g., the one or more memory device(s) 130) when the memory sub-system 110 is coupled with the host system 120 by the physical host interface (e.g., PCIe bus). The physical host interface can provide an interface for passing control, address, data, and other signals between the memory sub-system 110 and the host system 120.
The memory devices 130, 140 can include any combination of the different types of non-volatile memory devices and/or volatile memory devices. The volatile memory devices (e.g., memory device 140) can be, but are not limited to, random access memory (RAM), such as dynamic random access memory (DRAM) and synchronous dynamic random access memory (SDRAM).
Some examples of non-volatile memory devices (e.g., memory device(s) 130) include a negative-and (NAND) type flash memory and write-in-place memory, such as a three-dimensional cross-point (“3D cross-point”) memory device, which is a cross-point array of non-volatile memory cells. A cross-point array of non-volatile memory cells can perform bit storage based on a change of bulk resistance, in conjunction with a stackable cross-gridded data access array. Additionally, in contrast to many flash-based memories, cross-point non-volatile memory can perform a write in-place operation, where a non-volatile memory cell can be programmed without the non-volatile memory cell being previously erased. NAND type flash memory includes, for example, two-dimensional NAND (2D NAND) and three-dimensional NAND (3D NAND).
Each of the memory devices 130 can include one or more arrays of memory cells. One type of memory cell, for example, single-level cells (SLC) can store one bit per cell. Other types of memory cells, such as multi-level cells (MLCs), triple level cells (TLCs), quad-level cells (QLCs), and penta-level cells (PLCs) can store multiple bits per cell, e.g., by way of additional threshold voltage ranges. In some embodiments, each of the memory device(s) 130 can include one or more arrays of memory cells such as SLCs, MLCs, TLCs, QLCs, PLCs or any combination of such. In some embodiments, a particular memory device can include an SLC portion, and an MLC portion, a TLC portion, a QLC portion, or a PLC portion of memory cells. The memory cells of the memory devices 130 can be grouped as pages that can refer to a logical unit of the memory device used to store data. With some types of memory (e.g., NAND), pages can be grouped to form blocks.
Although non-volatile memory components such as a 3D cross-point array of non-volatile memory cells and NAND type flash memory (e.g., 2D NAND, 3D NAND) are described, the memory device 130 can be based on any other type of non-volatile memory, such as read-only memory (ROM), phase change memory (PCM), self-selecting memory, other chalcogenide based memories, ferroelectric transistor random-access memory (FeTRAM), ferroelectric random access memory (FeRAM), magneto random access memory (MRAM), Spin Transfer Torque (STT)-MRAM, conductive bridging RAM (CBRAM), resistive random access memory (RRAM), oxide based RRAM (OxRAM), negative-or (NOR) flash memory, or electrically erasable programmable read-only memory (EEPROM).
A memory sub-system controller 115 (or controller 115 for simplicity) can communicate with the memory devices 130 to perform operations such as reading data, writing data, or erasing data at the memory devices 130 and other such operations. The memory sub-system controller 115 can include hardware such as one or more integrated circuits and/or discrete components, a buffer memory, or a combination thereof. The hardware can include a digital circuitry with dedicated (i.e., hard-coded) logic to perform the operations described herein. The memory sub-system controller 115 can be a microcontroller, special purpose logic circuitry (e.g., a field programmable gate array (FPGA), an application specific integrated circuit (ASIC), etc.), or other suitable processor.
The memory sub-system controller 115 can include a processing device, which includes one or more processors (e.g., processor 117), configured to execute instructions stored in a local memory 119. In the illustrated example, the local memory 119 of the memory sub-system controller 115 includes an embedded memory configured to store instructions for performing various processes, operations, logic flows, and routines that control operation of the memory sub-system 110, including handling communications between the memory sub-system 110 and the host system 120.
In some embodiments, the local memory 119 can include memory registers storing memory pointers, fetched data, etc. The local memory 119 can also include read-only memory (ROM) for storing micro-code. While the example memory sub-system 110 in
In general, the memory sub-system controller 115 can receive commands or operations from the host system 120 and can convert the commands or operations into instructions or appropriate commands to achieve the desired access to the memory devices 130. The memory sub-system controller 115 can be responsible for other operations such as wear leveling operations, garbage collection operations, error detection and error-correcting code (ECC) operations, encryption operations, caching operations, and address translations between a logical address (e.g., a logical block address (LBA), namespace) and a physical address (e.g., physical block address) that are associated with the memory devices 130. The memory sub-system controller 115 can further include host interface circuitry to communicate with the host system 120 via the physical host interface. The host interface circuitry can convert the commands received from the host system into command instructions to access the memory devices 130 as well as convert responses associated with the memory devices 130 into information for the host system 120.
The memory sub-system 110 can also include additional circuitry or components that are not illustrated. In some embodiments, the memory sub-system 110 can include a cache or buffer (e.g., DRAM) and address circuitry (e.g., a row decoder and a column decoder) that can receive an address from the memory sub-system controller 115 and decode the address to access the memory devices 130.
In some embodiments, the memory devices 130 include local media controllers 135 that operate in conjunction with memory sub-system controller 115 to execute operations on one or more memory cells of the memory devices 130. An external controller (e.g., memory sub-system controller 115) can externally manage a memory device 130 (e.g., perform media management operations on the memory device 130). In some embodiments, memory sub-system 110 is a managed memory device, which is a raw memory device having control logic (e.g., local media controller 135) on the die and a controller (e.g., memory sub-system controller 115) for media management within the same memory device package. An example of a managed memory device is a managed NAND (MNAND) device.
In at least some embodiments, each memory device 130 includes a peak power manager (PPM) wrapper 150 that includes a threads manager 155 and a PPM 160 (peak power manager). In one embodiment, local media controller 135 of each memory device 130 includes at least a portion of the PPM wrapper 150. In such an embodiment, PPM wrapper 150 can be implemented using hardware or as firmware, stored on each memory device 130, executed by the control logic (e.g., local media controller 135) to perform the operations related to prioritized power budget arbitration for multiple concurrent access operations described herein. In some embodiments, the memory sub-system controller 115 includes at least a portion of the PPM wrapper 150. For example, the memory sub-system controller 115 can include a processor 117 (e.g., a processing device) configured to execute instructions stored in local memory 119 for performing the operations described herein.
In at least some embodiments, the PPM wrapper 150 can manage power prioritized budget arbitration for multiple concurrent access operations in the memory device(s) 130. In one embodiment, memory sub-system 110 employs a token-based protocol, where a token rotates (e.g., in round robin fashion) among multiple PPM wrappers 150 of multiple memory dice (e.g., after a set number of cycles of a shared clock signal). When the PPM wrapper 150 of a die holds the token, the PPM wrapper 150 can determine the power (e.g., current) requested by multiple processing threads (e.g., implemented by local media controller 135) of each memory device 130, select one or more prioritized processing threads of those multiple processing threads based on an available power budget in the memory sub-system, request that power from a shared current source in memory sub-system 110, and allocate the requested power to the selected processing threads. In some embodiments, if power is allocated to all prioritized processing threads and budget remains, the PPM wrapper 150 can allocate power to non-prioritized processing threads as well. The PPM wrapper 150 can further broadcast a quantized current budget to be consumed by the memory die during a given time period, so that the other PPM wrappers in memory sub-system 110 are aware of the available power budget. Further details with regards to the operations of each PPM wrapper 150 are described below.
Each memory device 130 includes an array of memory cells 104 logically arranged in rows and columns. Memory cells of a logical row are typically connected to the same access line (e.g., a word line) while memory cells of a logical column are typically selectively connected to the same data line (e.g., a bit line). A single access line can be associated with more than one logical row of memory cells and a single data line can be associated with more than one logical column. Memory cells (not shown in
Row decode circuitry 108 and column decode circuitry 111 are provided to decode address signals. Address signals are received and decoded to access the array of memory cells 104. Each memory device 130 also includes input/output (I/O) control circuitry 112 to manage input of commands, addresses and data to the memory device 130 as well as output of data and status information from each memory device 130. An address register 114 is in communication with the I/O control circuitry 112 and row decode circuitry 108 and column decode circuitry 111 to latch the address signals prior to decoding. A command register 124 is in communication with the I/O control circuitry 112 and the local media controller 135 to latch incoming commands.
A controller (e.g., the local media controller 135 internal to each memory device 130) controls access to the array of memory cells 104 in response to the commands and generates status information for the external memory sub-system controller 115, i.e., the local media controller 135 is configured to perform access operations (e.g., read operations, programming operations and/or erase operations) on the array of memory cells 104. The local media controller 135 is in communication with row decode circuitry 108 and column decode circuitry 111 to control the row decode circuitry 108 and column decode circuitry 111 in response to the addresses.
The local media controller 135 is also in communication with a cache register 118 and a data register 121. The cache register 118 latches data, either incoming or outgoing, as directed by the local media controller 135 to temporarily store data while the array of memory cells 104 is busy writing or reading, respectively, other data. During a program operation (e.g., write operation), data can be passed from the cache register 118 to the data register 121 for transfer to the array of memory cells 104; then new data can be latched in the cache register 118 from the I/O control circuitry 112. During a read operation, data can be passed from the cache register 118 to the I/O control circuitry 112 for output to the memory sub-system controller 115; then new data can be passed from the data register 121 to the cache register 118. The cache register 118 and/or the data register 121 can form (e.g., can form at least a portion of) the page buffer of each memory device 130. The page buffer can further include sensing devices such as a sense amplifier, to sense a data state of a memory cell of the array of memory cells 104, e.g., by sensing a state of a data line connected to that memory cell. A status register 122 can be in communication with I/O control circuitry 112 and the local memory controller 135 to latch the status information for output to the memory sub-system controller 115.
Each memory device 130 receives control signals at the memory sub-system controller 115 from the local media controller 135 over a control link 132. For example, the control signals can include a chip enable signal CE #, a command latch enable signal CLE, an address latch enable signal ALE, a write enable signal WE #, a read enable signal RE #, and a write protect signal WP #. Additional or alternative control signals (not shown) can be further received over control link 132 depending upon the nature of each memory device 130. In one embodiment, each memory device 130 receives command signals (which represent commands), address signals (which represent addresses), and data signals (which represent data) from the memory sub-system controller 115 over a multiplexed input/output (I/O) bus 134 and outputs data to the memory sub-system controller 115 over I/O bus 134.
For example, the commands can be received over input/output (I/O) pins [7:0] of I/O bus 134 at I/O control circuitry 112 and can then be written into a command register 124. The addresses can be received over input/output (I/O) pins [7:0] of I/O bus 134 at I/O control circuitry 112 and can then be written into address register 114. The data can be received over input/output (I/O) pins [7:0] for an 8-bit device or input/output (I/O) pins [15:0] for a 16-bit device at I/O control circuitry 112 and then can be written into cache register 118. The data can be subsequently written into data register 121 for programming the array of memory cells 104.
In an embodiment, cache register 118 can be omitted, and the data can be written directly into data register 121. Data can also be output over input/output (I/O) pins [7:0] for an 8-bit device or input/output (I/O) pins [15:0] for a 16-bit device. Although reference can be made to I/O pins, they can include any conductive node providing for electrical connection to each memory device 130 by an external device (e.g., the memory sub-system controller 115), such as conductive pads or conductive bumps as are commonly used.
It will be appreciated by those skilled in the art that additional circuitry and signals can be provided, and that each memory device 130 of
In one embodiment, a token-based protocol is used where a token cycles through each of the memory dice 230(0)-230(7) for determining and broadcasting expected peak current magnitude, even though some of the memory dice 230(0)-230(7) might be disabled in response to their respective chip enable signal. The period of time during which a given PPM wrapper 150 holds this token (e.g. a certain number of cycles of clock signal ICLK) can be referred to herein as a power management cycle of the associated memory die. At the end of the power management cycle, the token is passed to a next memory die in sequence. Eventually the token is received again by the same PPM wrapper 150 which signals the beginning oft hew power management cycle for the associated memory die. In one embodiment, the encoded value for the lowest expected peak current magnitude is configured such that each of its digits correspond to the normal logic level of the peak current magnitude indicator signal HC # where the disabled dice do not transition the peak current magnitude indicator signal HC #. In other embodiments, however, the memory dice can be configured, when otherwise disabled in response to their respective chip enable signal, to drive transitions of the peak current magnitude indicator signal HC # to indicate the encoded value for the lowest expected peak current magnitude upon being designated.
When a given PPM wrapper 150 holds the token, it can determine the peak current magnitude for the respective one of memory die 230(0)-230(7), which can be attributable to one or more processing threads on that memory die, and broadcast an indication of the same via the peak current magnitude indicator signal HC #. As described in more detail below, during a given power management cycle, the PPM wrapper 150 can arbitrate among the multiple processing threads on the respective memory die using one of a number of different arbitration schemes in order to allocate that peak current to enable concurrent memory access operations.
The memory device 130A includes a memory array 370 divided into memory planes 372(0)-372(3) that each includes a respective number of memory cells. The multi-plane memory device 130A can further include local media controller 135, including a power control circuit and access control circuit for concurrently performing memory access operations for different memory planes 372(0)-372(3). The memory cells can be non-volatile memory cells, such as NAND flash cells, or can generally be any type of memory cells.
The memory planes 372(0)-372(3) can each be divided into blocks of data, with a different relative block of data from each of the memory planes 372(0)-372(3) concurrently accessible during memory access operations. For example, during memory access operations, data block 382 of the memory plane 372(0), data block 383 of the memory plane 372(1), data block 384 of the memory plane 372(2), and data block 385 of the memory plane 372(3) can each be accessed concurrently.
Each of the memory planes 372(0)-372(3) can be coupled to a respective page buffer 376(0)-376(3). Each page buffer 376(0)-376(3) can be configured to provide data to or receive data from the respective memory plane 372(0)-372(3). The page buffers 376(0)-376(3) can be controlled by local media controller 135. Data received from the respective memory plane 372(0)-372(3) can be latched at the page buffers 376(0)-376(3), respectively, and retrieved by local media controller 135, and provided to the memory sub-system controller 115 via the NVMe interface.
Each of the memory planes 372(0)-372(3) can be further coupled to a respective access driver circuit 374(0)-374(3), such as an access line driver circuit. The driver circuits 374(0)-374(3) can be configured to condition a page of a respective block of an associated memory plane 372(0)-372(3) for a memory access operation, such as programming data (i.e., writing data), reading data, or erasing data. Each of the driver circuits 374(0)-374(3) can be coupled to a respective global access lines associated with a respective memory plane 372(0)-372(3). Each of the global access lines can be selectively coupled to respective local access lines within a block of a plane during a memory access operation associated with a page within the block. The driver circuits 374(0)-374(3) can be controlled based on signals from local media controller 135. Each of the driver circuits 374(0)-374(3) can include or be coupled to a respective power circuit, and can provide voltages to respective access lines based on voltages provided by the respective power circuit. The voltages provided by the power circuits can be based on signals received from local media controller 135.
The local media controller 135 can control the driver circuits 374(0)-374(3) and page buffers 376(0)-376(3) to concurrently perform memory access operations associated with each of a group of memory command and address pairs (e.g., received from memory sub-system controller 115). For example, local media controller 135 can control the driver circuits 374(0)-374(3) and page buffer 376(0)-376(3) to perform the concurrent memory access operations. Local media controller 135 can include a power control circuit that serially configures two or more of the driver circuits 374(0)-374(3) for the concurrent memory access operations, and an access control circuit configured to control two or more of the page buffers 376(0)-376(3) to sense and latch data from the respective memory planes 372(0)-372(3), or program data to the respective memory planes 372(0)-372(3) to perform the concurrent memory access operations.
In operation, local media controller 135 can receive a group of memory commands and address pairs via the NVMe bus, with each pair arriving in parallel or serially. In some examples, the group of memory commands and address pairs can each be associated with different respective memory planes 372(0)-372(3) of the memory array 370. The local media controller 135 can be configured to perform concurrent memory access operations (e.g., read operations or program operations) for the different memory planes 372(0)-372(3) of the memory array 370 responsive to the group of memory commands and address pairs. For example, the power control circuit of local media controller 135 can serially configure, for the concurrent memory access operations based on respective page type (e.g., UP, MP, LP, XP, SLC/MLC/TLC/QLC page), the driver circuits 374(0)-374(3) for two or more memory planes 372(0)-372(3) associated with the group of memory commands and address pairs. After the access line driver circuits 374(0)-374(3) have been configured, the access control circuit of the local media controller 135 can concurrently control the page buffers 376(0)-376(3) to access the respective pages of each of the two or more memory planes 372(0)-372(3) associated with the group of memory commands and address pairs, such as retrieving data or writing data, during the concurrent memory access operations. For example, the access control circuit can concurrently (e.g., in parallel and/or contemporaneously) control the page buffers 376(0)-376(3) to charge/discharge bitlines, sense data from the two or more memory planes 372(0)-372(3), and/or latch the data.
Based on the signals received from local media controller 135, the driver circuits 374(0)-374(3) that are coupled to the memory planes 372(0)-372(3) associated with the group of memory command and address command pairs can select blocks of memory or memory cells from the associated memory plane 372(0)-372(3), for memory operations, such as read, program, and/or erase operations. The driver circuits 374(0)-374(3) can drive different respective global access lines associated with a respective memory plane 372(0)-372(3). As an example, the driver circuit 374(0) can drive a first voltage on a first global access line associated with the memory plane 372(0), the driver circuit 374(1) can drive a second voltage on a third global access line associated with the memory plane 372(1), the driver circuit 374(2) can drive a third voltage on a seventh global access line associated with the memory plane 372(2), etc., and other voltages can be driven on each of the remaining global access lines. In some examples, pass voltages can be provided on all access lines except on an access line associated with a page of a memory plane 372(0)-372(3) to be accessed. The local media controller 135, the driver circuits 374(0)-374(3) can allow different respective pages, and the page buffers 376(0)-376(3) within different respective blocks of memory cells, to be accessed concurrently. For example, a first page of a first block of a first memory plane can be accessed concurrently with a second page of a second block of a second memory plane, regardless of page type.
The page buffers 376(0)-376(3) can provide data to or receive data from the local media controller 135 during the memory access operations responsive to signals from the local media controller 135 and the respective memory planes 372(0)-372(3). The local media controller 135 can provide the received data to memory sub-system controller 115.
It will be appreciated that the memory device 130A can include more or less than four memory planes, driver circuits, and page buffers. It will also be appreciated that the respective global access lines can include 8, 16, 32, 64, 128, etc., global access lines. The local media controller 135 and the driver circuits 374(0)-374(3) can concurrently access different respective pages within different respective blocks of different memory planes when the different respective pages are of a different page type. For example, local media controller 135 can include a number of different processing threads, such as processing threads 334(0)-334(3). Each of processing threads 334(0)-334(3) can be associated with a respective one of memory planes 372(0)-372(3) and can manage operations performed on the respective plane. For example, each of processing threads 334(0)-334(3) can provide control signals to the respective one of driver circuits 374(0)-374(3) and page buffers 376(0)-376(3) to perform those memory access operations concurrently (e.g., at least partially overlapping in time). Since the processing threads 334(0)-334(3) can perform the memory access operations, each of processing threads 334(0)-334(3) can have different current requirements at different points in time. According to the techniques described herein, the PPM wrapper 150 can determine the power budget needs of processing threads 334(0)-334(3) in a given power management cycle and identify one or more of processing threads 334(0)-334(3) using one of a number of power budget arbitration schemes described herein. The one or more processing threads 334(0)-334(3) can be determined based on an available power budget in the memory sub-system 110 during the power management cycles. For example, the PPM wrapper 150 can determine respective priorities of processing threads 334(0)-334(3), and allocate current to processing threads 334(0)-334(3) based on the respective priorities.
In some embodiments, the threads manager 155 identifies one or more processing threads, such as multiple processing threads 434(0)-434(3) in memory die 400, and requests the PPM 160 to determine whether an available current (e.g., power) budget can support running the one or more processing threads based on an amount of power associated with the one or more processing threads during a power management cycle. More specifically, because the multiple processing threads 434(0)-434(3) can generate different requests asynchronously, to manage such complexity, the threads manager 155 can manipulate and summarize these asynchronous requests in a simplified number of requests for the PPM 160. In some embodiments, a set of simplified requests sent to the PPM 160 can contain randomized thread requests to ensure equity of allocation of current to the multiple processing threads 434(0)-434(3). As will be discussed in more detail, the randomization of the requests sent by the threads manger 155 to the PPM 160 can be performed by the non-priority ring counter 444, by the priority ring counter 454, or can shift between the two ring counters as will be discussed. In some embodiments, the multiple processing threads 434(0)-434(3) correspond to the processing threads 334(0)-334(3) (
In some embodiments, the PPM 160 periodically asserts a polling window signal 460, which is received by the threads manager 155. The polling window signal 460 is asserted after the end of a previous power management cycle (e.g., when the PPM 160 gives up the token) and prior to the beginning of a subsequent power management cycle (e.g., when PPM 160 receives the token again). As the processing threads 434(0)-434(3) are regularly issuing requests for current depending on associated processing operations, during a period when the polling window signal 460 is asserted, the threads manager 155 stores or buffers the received requests in a the request registers 452. While requests are often referred to herein as requesting current allocation, this should be understood as requesting power generally, e.g., can also include requesting voltage allocation.
In some embodiments, the PPM 160 tracks the token and can determine when the token will be received (e.g., based on synchronous clock signal ICLK) and can de-assert the polling window signal 460 in advance of that time. Responsive to the polling window signal 460 being deasserted (i.e., during the subsequent the power management cycle), the threads manager 155 can stop storing additional requests in request registers 452 so that the contents of request registers 452 is static. Any new requests are not considered during this cycle, but are saved and can be considered in a subsequent power management cycle. The threads manager 155 can generate multiple current level signals, such as a full signal 462, a middle signal 464, a low signal 466, and a high-to-low signal 468 where each current level signal corresponds to the current associated with a respective set of at least one of the requests in the request registers 452. For example, the full signal 462 can represent the sum of all current requests in the request registers 452, the middle signal 464 can represent the sum of two or more, but less than all, of the current requests in the request registers 452 (e.g., the first two or more requests in the request registers 452), the low signal 466 can represent one current request from the request registers 452 (e.g., the first request in the request registers 452), and the high-to-low signal 468 can represent a low current request when a high current budget has already been allocated. The high-to-low signal 468 can be associated with requests in the request registers 452 for which the PPM 160 will immediately allocate current, without checking against a current budget, and will track such allocation as with tracking other current allocations. By polling the processing threads between power management cycles, the threads manager 155 can save significant time and processing resources compared to waiting until the token is actually received.
In these embodiments, the PPM 160 receives the full signal 462, the middle signal 464, the low signal 466, and the high-to-low signal 468 and determines whether the amount of current associated with any of these current level signals can be satisfied by an amount of current available in the memory sub-system 110 during the current power management cycle. Responsive to the amount of current available satisfying at least one of the current level signals, the PPM 160 can request that amount of current and provide an authorization signal 472, e.g., an acknowledgement, to the threads manager 155. The authorization signal 472 can indicate which of the current level signals is satisfied by the amount of available current, for example. The threads manager 155 can thus authorize one or more of processing threads 434(0)-434(3) to perform one or more memory access operations corresponding to the request in the request registers 452 based on which requests were authorized by the authorization signal 472.
As illustrated in
More specifically, Table 1 is an example of the data structure 448 of the PPM wrapper 150 used for power budget arbitration for multiple processing threads in the memory device 130 or 130A. In one embodiment, the data structure 448 is formed in or managed by the threads manager 155 using a lookup table, an array, a linked list, a record, an object, other some other data structure. In one embodiment, the data structure 448 includes a number of entries, each corresponding to one of the states of the non-priority ring counter 444. For example, for each state of the non-priority ring counter 444, the data structure 448 can identify a leading thread, and a thread combination. The leading thread can be a single processing thread having the highest priority when the non-priority ring counter 444 is in the corresponding state, and the thread combination can be a set of two or more processing threads, but less than all of the processing threads, which have a higher priority than other threads not in the set, but a lower priority than the leading thread, when the non-priority ring counter 444 is in the corresponding state.
In some embodiments, to allocate available power budget during a power management cycle, the threads manager 155 can determine a current state of the non-priority ring counter 444 and determine, from the data structure 448, a leading thread and a thread combination corresponding to the current state of the non-priority ring counter 444. The threads manager 155 can then send requests to the PPM 160, e.g., the full signal 462, the middle signal 464, and the low signal 466 based on identification of the leading thread and the thread combination. Responsive to an amount of current available in the memory sub-system during that power management cycle satisfying an amount of current associated with at least one of the leading thread or the thread combination, the PPM 160 can request that amount of current associated with the at least one of the leading thread or the thread combination and allocate that current budget accordingly.
By way of an additional example, Table 2 illustrates the data structure 448 in which a 3-bit non-priority counter 444 can hold up to eight states or values, and thus the data structure 448 can store additional combinations of possible leading threads and thread combinations. The 3-bit example of Table 2 and other Tables included below are merely exemplary for purposes of explanation, as other are envisioned, including those of 4-bit and beyond. In Table 2, the “Reg_hc_max” value corresponds to the full signal 462, the “Reg_hc_middle” corresponds to the middle signal 464, and the “Reg_hc_min” corresponds to the low signal 466.
In this example, the full signal 462 can correspond to all of the multiple processing threads requesting current, e.g., a main processing thread of the local media controller 135, and five additional co-processors (“coproc”) that can be associated with individual additional threads such as the threads0-thread3 434(0)-434(3), although more co-processors are envisioned. Further, the middle signal 464 can include a subset of the multiple processing threads (e.g., a thread combination), and the low signal 466 can include just one of the processing threads (e.g., leading thread) of the multiple processing threads. In one embodiment, the subset of the multiple processing threads is no more than half of the multiple processing threads.
At operation 605, power requests are sampled. For example, processing logic (e.g., PPM 160) can sample power requests, such as current requests or peak current magnitude requests, from one or more processing threads, such as processing threads 334(0)-334(3), of a memory device. In one embodiment, responsive to the PPM 160 receiving the token, which signals the start of a current power management cycle, PPM 160 alerts the threads manager 155 of the start of the polling window 460. In response, the threads manager 155 sends polling requests to each of the processing threads to obtain an indication of current requested during the current power management cycle. The amount of current requested can be based on a number of memory access requests pending for each processing thread and the type of memory access requests pending for each processing thread. In one embodiment, each processing thread returns a separate response to the polling request, such that threads manager 155 can determine the current request of each processing thread separately. In one embodiment, another component, or a sub-component of the PPM wrapper 150 can issue the polling requests to and receive the current requests from the processing threads.
At operation 610, an available power budget is determined. For example, the processing logic can determine an amount of current available in the memory device during the power management cycle. In one embodiment, the PPM 160 receives a signal, such as peak current magnitude indicator signal HC #, indicating the current utilized by each other PPM 160 in the multi-dice package 200 and subtracts that amount from a total amount of available current in the memory sub-system 110 or memory device 130 or 130A. In one embodiment, the processing logic compares the total current associated with all processing threads (e.g., the sum of the individual current requests) to the amount of available current during the power management cycle to determine if the available current budget satisfies the current requests of all processing threads. If the amount of current available is equal to or greater than the amount of current associated with (e.g., demanded by) all of the processing threads, the processing logic determines that the amount of current available satisfies the amount of current associated with all of the processing threads.
At operation 615, current is requested and allocated. If the processing logic determines that the amount of current available satisfies the amount of current associated with all of the processing threads, the processing logic can request the amount of current associated with all of the processing threads. For example, the PPM 160 can issue the request to a common current supply or other power source in the memory device 130 or 130A or memory sub-system 110. The PPM 160 can subsequently allocate the requested current to the processing threads, allowing all of the processing threads to complete their pending memory access operations.
If the processing logic determines that the amount of current available does not satisfy the amount of current associated with all of the processing threads, at operation 620, a thread combination is examined. For example, the processing logic can identify, from a data structure, such as data structure 448, a thread combination that corresponds to a current state of a ring counter, such as the non-priority ring counter 444. The thread combination corresponding to each state of the non-priority ring counter 444 is different ensuring that different threads are serviced in different power management cycles and no threads are ignored. In one embodiment, the processing logic compares the total current associated with the identified thread combination (e.g., the sum of the individual current requests) to the amount of available current during the power management cycle to determine if the available current budget satisfies the current requests of the thread combination. If the amount of current available is equal to or greater than the amount of current associated with the thread combination, the processing logic determines that the amount of current available satisfies the amount of current associated with the thread combination.
At operation 625, current is requested and allocated. If the processing logic determines that the amount of current available satisfies the amount of current associated with the thread combination, the processing logic can request the amount of current associated with the thread combination. For example, the PPM 160 can issue the request to a common current supply or other power source in the memory device 130 or 130A or memory sub-system 110. The PPM 160 can subsequently allocate the requested current to the processing threads, allowing the processing threads identified in the thread combination to complete their pending memory access operations.
If the processing logic determines that the amount of current available does not satisfy the amount of current associated with the thread combination, at operation 630, a leading thread is examined. For example, the processing logic can identify, from a data structure, such as data structure 448, a leading thread that corresponds to a current state of a ring counter, such as the non-priority ring counter 444. The leading thread corresponding to each state of the non-priority ring counter 444 is different ensuring that different threads are serviced in different power management cycles and no threads are ignored. In one embodiment, the processing logic compares the requested current associated with the identified leading thread to the amount of available current during the power management cycle to determine if the available current budget satisfies the current request of the leading thread. If the amount of current available is equal to or greater than the amount of current associated with the leading thread, the processing logic determines that the amount of current available satisfies the amount of current associated with the leading thread.
At operation 635, current is requested and allocated. If the processing logic determines that the amount of current available satisfies the amount of current associated with the leading thread, the processing logic can request the amount of current associated with the leading thread. For example, the PPM 160 can issue the request to a common current supply or other power source in the memory device 130 or 130A or memory sub-system 110. The PPM 160 can subsequently allocate the requested current to the leading thread, allowing the leading thread to complete its pending memory access operations.
If the processing logic determines that the amount of current available does not satisfy the amount of current associated with the leading thread, at operation 640, the current requests are paused. For example, the processing logic can pause execution of the processing threads and maintain the current requests from those processing threads until a subsequent power management cycle. In the subsequent power management cycle, there can possibly be a larger amount of available current in the memory device which can be sufficient to satisfy the request associated with at least one of the processing threads.
At operation 705, current level signals are received. For example, processing logic (e.g., PPM 160) can receive one or more current level signals, such as full signal 462, middle signal 464, and low signal 466, associated with a respective set of at least one of the requests in the request registers 452. In one embodiment, the current level signals are based on requests identified during a polling window (e.g., when polling window signal 460 is asserted) between power management cycles. In one embodiment, during the polling window, threads manager 155 receives and storing current request from the processing threads, with each request including an indication of current requested. The amount of current requested can be based on a number of memory access requests pending for each processing thread and the type of memory access requests pending for each processing thread. In one embodiment, each processing thread sends one or more separate requests, such that the threads manager 155 can determine the current request(s) of each processing thread separately, and add the corresponding request(s) to request registers 452.
At operation 710, an available power budget is determined. For example, the processing logic can determine an amount of current available in the memory device during the power management cycle (i.e., once the token is received and the polling window closes). In one embodiment, the PPM 160 receives a signal, such as peak current magnitude indicator signal HV #, indicating the current utilized by each other PPM 160 in the multi-dice package 200 and subtracts that amount from a total amount of current in the memory sub-system 110 or memory device 130 or 130A. In one embodiment, the processing logic compares the total current associated with the full signal 462 (e.g., the sum of all the individual current requests in request registers 452) to the amount of available current during the power management cycle to determine if the available current budget satisfies the full signal 462. If the amount of current available is equal to or greater than the amount of current associated with the full signal 462, the processing logic determines that the amount of current available satisfies the full signal 462.
At operation 715, current is requested and allocated. If the processing logic determines that the amount of current available satisfies the full signal 462, the processing logic can request the amount of current associated with all of the requests in request registers 452. For example, the PPM 160 can issue the request to a common current supply or other power source in the memory device 130 or 130A or memory sub-system 110. The PPM 160 can subsequently allocate the requested current to the processing threads via authorization signal 472, allowing all of the current requests in request registers 452 to be performed.
If the processing logic determines that the amount of current available does not satisfy the full signal 462 request, another current level signal is examined. For example, the processing logic compares the current associated with the middle signal 464 (e.g., the sum two or more current requests in request registers 452) to the amount of available current during the power management cycle to determine if the available current budget satisfies the middle signal 464. If the amount of current available is equal to or greater than the amount of current associated with the middle signal 464, the processing logic determines that the amount of current available satisfies the middle signal 464.
At operation 725, current is requested and allocated. If the processing logic determines that the amount of current available satisfies the amount of current associated with the middle signal 464, the processing logic can request the amount of current associated with the two or more requests from request registers 452. For example, PPM 160 can issue the request to a common current supply or other power source in the memory device 130 or 130A or memory sub-system 110. The PPM 160 can subsequently allocate the requested current to the processing threads via authorization signal 472, allowing two or more of the current requests in request registers 452 to be performed.
If the processing logic determines that the amount of current available does not satisfy the amount of current associated with the middle signal 464, at operation 730, another current level signal is examined. For example, the processing logic compares the current associated with the low signal 466 (e.g., one current request in request registers 452) to the amount of available current during the power management cycle to determine if the available current budget satisfies the low signal 466. If the amount of current available is equal to or greater than the amount of current associated with the low signal 466, the processing logic determines that the amount of current available satisfies the low signal 466.
At operation 735, current is requested and allocated. If the processing logic determines that the amount of current available satisfies the amount of current associated with the low signal 466, the processing logic can request the amount of current associated with one request from request registers 452. For example, the PPM 160 can issue the request to a common current supply or other power source in the memory device 130 or 130A or memory sub-system 110. The PPM 160 can subsequently allocate the requested current to the processing threads via authorization signal 472, allowing one current requests in request registers 452 to be performed.
If the processing logic determines that the amount of current available does not satisfy the amount of current associated with the any of the current level signals, at operation 740, the current requests are paused. For example, the processing logic can pause execution of the processing threads and maintain the current requests from those processing threads until a subsequent power management cycle. In the subsequent power management cycle, there can possibly be a larger amount of available current in the memory device which can be sufficient to satisfy at least one of the requests.
In various embodiments, because the multiple processing threads 334(0)-334(3) can asynchronously correspond to memory access operations such as a read operation or a program operation, the PPM wrapper 150 can be programmed to manage power so that snap reads and other such memory access operations can be prioritized, e.g., over non-prioritized memory operations such as some program operations and any erase operation. Thus, the memory sub-system controller 115 (or other processing device sending memory commands within the memory sub-system 110) can add the prefix value, which indicates prioritization, to different memory command packets for a memory operation sent to the memory device 130 or 130A.
In at least some embodiments, once a die, and thus an individual PPM wrapper 150, receives a memory command packet such as discussed with reference to
With additional reference to
In various embodiments, Tables 3-7 illustrate examples of the data structures that can be selected as the data structure 458 for different sets of prioritized processing threads. A prioritized processing thread is labeled with a capital “P” to indicate prioritization. Table 3 illustrates that the leading thread (“Low”) is the main processing thread and is prioritized over any other thread combination. In some cases, however, the leading thread can also be included within the thread combination (“Middle”) and thus could also be authorized along with one or more additional non-prioritized processing threads.
In one embodiment, Table 4 illustrates priority between two different processing threads, namely coproc1 and coproc4, at least one of which is the leading thread (“Min”) for each respective value of the priority ring counter 454. In this embodiment, these two prioritized processing threads are also located within the thread combination (“Middle”). Assuming the priority ring counter 454 is incremented four times (through value 011), the PPM 160 can achieve uniformity in allocation of power to the prioritized processing threads. Thus, in some embodiments, the values of “100” and “101” of the priority ring counter 454 are not included in an incrementing cycle.
In one embodiment, Table 5 illustrates priority between three different processing threads, namely main, coproc2, and coproc5. Because the data structure 458 represented by Table 5 is programmed with these three prioritized threads into the middle signal 464 and the low signal 466 requests, the PPM 160 can allocate power to these prioritized processing threads in cases where there is insufficient power budget to allocate to all of the multiple processing threads.
In one embodiment, Table 6 illustrates priority between four different processing threads, namely main, coproc1, coproc2, and coproc5. Because the data structure 458 represented by Table 5 is programmed with these three prioritized threads into the middle signal 464 and the low signal 466 requests, the PPM 160 can allocate power to these prioritized processing threads in cases where there is insufficient power budget to allocate to all of the multiple processing threads. Further, assuming the priority ring counter 454 is incremented four times (through value 011), the PPM 160 can achieve uniformity in allocation of power to the prioritized processing threads. Thus, in some embodiments, the values of “100” and “101” of the priority ring counter 454 are not included in an incrementing cycle.
In one embodiment, Table 7 illustrates priority between five different processing threads, namely main, coproc1, coproc2, coproc4, and coproc5. Because the data structure 458 represented by Table 7 is programmed with these three prioritized threads into the middle signal 464 and the low signal 466 requests, the PPM 160 can allocate power to these prioritized processing threads in cases where there is insufficient power budget to allocate to all of the multiple processing threads. Further, assuming the priority ring counter 454 is incremented five times (through value “100”), the PPM 160 can achieve uniformity in allocation of power to the prioritized processing threads. Thus, in some embodiments, the value of “101” of the priority ring counter 454 is not included in an incrementing cycle.
Thus, it can be seen that the one or more prioritized processing threads, which relates to given counter value, includes the leading thread in the low signal 466 that is a main processing thread of the memory die in one embodiment. In another embodiment, the one or more prioritized processing threads includes the leading thread and one or more processing threads of a thread combination, e.g., the middle signal 464 within the subset of the multiple processing threads. This subset, e.g., the combination of the leading thread and the thread combination, can therefore, be understood to include all prioritized processing threads.
With additional reference to
With additional reference to
In response to the PPM 160 allocating power (e.g., current) to a prioritized processing thread (e.g., “pgr3”), the threads manager 155 can, before the end of the polling period associated with the polling window signal 460, increment the value of the priority ring counter 454, e.g., which in this example is a value of “100.” Thus, the management of power allocation has passed to prioritized allocation governed by the value of the priority ring counter 454. This operation includes ensuring, during the power management cycle when the PPM 160 holds the token, that the amount of current for a new current allocation does not exceed an available budget less the total amount of currently already allocated by the PPM 160.
In these embodiments, if the amount of current for a new current allocation, e.g., to the pgr5 prioritized processing thread, does exceed an available budget less the total amount of currently already allocated, then the PPM 160 pauses allocation of power (or current) to non-prioritized processing thread, pgr0. If pausing allocation to the non-prioritized processing thread does not free up enough power budget to handle the pgr5 prioritized processing thread, then the PPM 160 may have to further pause allocating power to any new processing threads until sufficient budget is made available. In this way, the prioritized threads are prioritized over the non-prioritized processing threads. While these operations explained with reference to
Different than
With continued reference to
In some embodiments, although not specifically illustrated, upon detecting completion of the previously non-prioritized processing thread, pgr0, control logic of the threads manager 155 resets the timer 478. In other embodiments, the threads manager 155 resets the time in response to completion of all previously non-prioritized processing threads. Then, again, while the timer is running, the threads manager increments the priority ring counter 454 before each power management cycle so that the PPM 160 can prioritize allocation of the power to the one or more prioritized processing threads located within the subset of the multiple processing threads corresponding to the value of the priority ring counter 454, e.g., as illustrated in Tables 3-7. In this way, the control logic of the PPM wrapper 150 can transition back to prioritized power management until a situation occurs in which the PPM 160 allocates power to at least one non-prioritized processing thread and at least one prioritized processing threads. In response to such a situation, the threads manager 155 can again start the timer 478 as illustrated in
At operation 1110, a non-prioritized processing threads is allocated power. More specifically, the processing logic allocates power to a non-prioritized processing thread of multiple processing threads based on a value of a non-priority ring counter, the multiple processing threads to execute memory access operations on the memory array 370 (
At operation 1120, allocation of power to a prioritized processing thread is detected More specifically, the processing logic determines whether allocation of the power to a prioritized processing thread has been detected. This creates the situation just discussed in which at least one non-prioritized processing thread and at least one prioritized processing thread has been allocated power and are running concurrently.
At operation 1130, at timer is started. More specifically, in response to an affirmative detection of allocation of power to the prioritized processing thread, the control logic starts a timer, such as the timer 478 (
At operation 1140, a priority ring counter is incremented. More specifically, while the timer is running, the processing logic increments a priority ring counter before each power management cycle. The priority ring counter can be the priority ring counter 454 of
At operation 1150, prioritized processing threads are prioritized. More specifically, while the timer is running, the processing logic prioritizes allocation of the power to the one or more prioritized processing threads located within a subset of the multiple processing threads corresponding to a value of the priority ring counter.
At operation 1160, timer expiration is checked. More specifically, the processing logic, determine whether the timer expires before completion of the non-prioritized processing thread. If the answer is no, at operation 1160, then the method 1100 cycles back to operations 1140 and 1150 and continues with prioritized power management, as was done at power management cycle three with reference to
At operation 1160, power management transitions back to non-prioritized management. More specifically, in response to the timer expiring before completion of the non-prioritized processing thread, the processing logic transitions power allocation between subsets of the multiple processing threads based on increments to the value of the non-priority ring counter.
The machine can be a personal computer (PC), a tablet PC, a set-top box (STB), a Personal Digital Assistant (PDA), a cellular telephone, a web appliance, a server, a network router, a switch or bridge, or any machine capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that machine. Further, while a single machine is illustrated, the term “machine” shall also be taken to include any collection of machines that individually or jointly execute a set (or multiple sets) of instructions to perform any one or more of the methodologies discussed herein.
The example computer system 1200 includes a processing device 1202, a main memory 1204 (e.g., read-only memory (ROM), flash memory, dynamic random access memory (DRAM) such as synchronous DRAM (SDRAM) or Rambus DRAM (RDRAM), etc.), a static memory 1206 (e.g., flash memory, static random access memory (SRAM), etc.), and a data storage system 1218, which communicate with each other via a bus 1230.
Processing device 1202 represents one or more general-purpose processing devices such as a microprocessor, a central processing unit, or the like. More particularly, the processing device can be a complex instruction set computing (CISC) microprocessor, reduced instruction set computing (RISC) microprocessor, very long instruction word (VLIW) microprocessor, or a processor implementing other instruction sets, or processors implementing a combination of instruction sets. Processing device 1202 can also be one or more special-purpose processing devices such as an application specific integrated circuit (ASIC), a field programmable gate array (FPGA), a digital signal processor (DSP), network processor, or the like. The processing device 1202 is configured to execute instructions 1226 for performing the operations and steps discussed herein. The computer system 1200 can further include a network interface device 1208 to communicate over the network 1220.
The data storage system 1218 can include a machine-readable storage medium 1224 (also known as a computer-readable medium, such as a non-transitory computer-readable medium) on which is stored one or more sets of instructions 1226 or software embodying any one or more of the methodologies or functions described herein. The instructions 1226 can also reside, completely or at least partially, within the main memory 1204 and/or within the processing device 1202 during execution thereof by the computer system 1200, the main memory 1204 and the processing device 1202 also constituting machine-readable storage media. The machine-readable storage medium 1224, data storage system 1218, and/or main memory 1204 can correspond to the memory sub-system 110 of
In one embodiment, the instructions 1226 include instructions to implement functionality corresponding to the PPM wrapper 150 of
Some portions of the preceding detailed descriptions have been presented in terms of algorithms and symbolic representations of operations on data bits within a computer memory. These algorithmic descriptions and representations are the ways used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. An algorithm is here, and generally, conceived to be a self-consistent sequence of operations leading to a desired result. The operations are those requiring physical manipulations of physical quantities. Usually, though not necessarily, these quantities take the form of electrical or magnetic signals capable of being stored, combined, compared, and otherwise manipulated. It has proven convenient at times, principally for reasons of common usage, to refer to these signals as bits, values, elements, symbols, characters, terms, numbers, or the like.
It should be borne in mind, however, that all of these and similar terms are to be associated with the appropriate physical quantities and are merely convenient labels applied to these quantities. The present disclosure can refer to the action and processes of a computer system, or similar electronic computing device, that manipulates and transforms data represented as physical (electronic) quantities within the computer system's registers and memories into other data similarly represented as physical quantities within the computer system memories or registers or other such information storage systems.
The present disclosure also relates to an apparatus for performing the operations herein. This apparatus can be specially constructed for the intended purposes, or it can include a general purpose computer selectively activated or reconfigured by a computer program stored in the computer. Such a computer program can be stored in a computer readable storage medium, such as, but not limited to, any type of disk including floppy disks, optical disks, CD-ROMs, and magnetic-optical disks, read-only memories (ROMs), random access memories (RAMs), EPROMs, EEPROMs, magnetic or optical cards, or any type of media suitable for storing electronic instructions, each coupled to a computer system bus.
The algorithms and displays presented herein are not inherently related to any particular computer or other apparatus. Various general purpose systems can be used with programs in accordance with the teachings herein, or it can prove convenient to construct a more specialized apparatus to perform the method. The structure for a variety of these systems will appear as set forth in the description below. In addition, the present disclosure is not described with reference to any particular programming language. It will be appreciated that a variety of programming languages can be used to implement the teachings of the disclosure as described herein.
The present disclosure can be provided as a computer program product, or software, that can include a machine-readable medium having stored thereon instructions, which can be used to program a computer system (or other electronic devices) to perform a process according to the present disclosure. A machine-readable medium includes any mechanism for storing information in a form readable by a machine (e.g., a computer). In some embodiments, a machine-readable (e.g., computer-readable) medium includes a machine (e.g., a computer) readable storage medium such as a read only memory (“ROM”), random access memory (“RAM”), magnetic disk storage media, optical storage media, flash memory components, etc.
In the foregoing specification, embodiments of the disclosure have been described with reference to specific example embodiments thereof. It will be evident that various modifications can be made thereto without departing from the broader spirit and scope of embodiments of the disclosure as set forth in the following claims. The specification and drawings are, accordingly, to be regarded in an illustrative sense rather than a restrictive sense.
This application claims the benefit of U.S. Provisional Patent Application No. 63/243,887, filed Sep. 14, 2021, the entirety of which is incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
20150089249 | Hannon | Mar 2015 | A1 |
20190065243 | Eckert | Feb 2019 | A1 |
Entry |
---|
Li, Chongmin, et al. “Priority Based Fair Scheduling: A Memory Scheduler Design for ChipMultiprocessor Systems.” Tsinghua National Laboratory for Information Science and Technology (2012). (Year: 2012). |
Number | Date | Country | |
---|---|---|---|
20230084630 A1 | Mar 2023 | US |
Number | Date | Country | |
---|---|---|---|
63243887 | Sep 2021 | US |