This disclosure is generally related to the field of data storage. More specifically, this disclosure is related to a method and system for facilitating a converged computation and storage node in a distributed storage system.
Today, various storage systems are being used to store and access the ever-increasing amount of digital content. Some technological fields or areas may require analysis of a large amount of data, e.g., big data analysis. In such fields, a distributed storage system or distributed cluster may divide computation-heavy tasks into multiple partitions or sub-tasks, where the sub-tasks may be performed in parallel by multiple nodes in the distributed cluster. A single node in such a distributed cluster can include both computation and storage capability. A single node can be a conventional server in a distributed cluster, and can use a central processing unit (CPU) as a central hub to manage traffic among devices or components in the node or server. Because certain computation workloads may not be friendly or optimal for a CPU, various heterogeneous processors can be used as Peripheral Component Interconnect Express (PCIe) devices to handle the main computation tasks. Examples of heterogeneous processors include a general processing unit (GPU) and a field programmable gate array (FPGA). The heterogeneous processors can generally be used in a conventional server as a PCIe add-in card (AIC). By shifting the main computation tasks from the CPU to the heterogeneous processors, a distributed system can deploy the CPU with a decreased cost and power consumption.
However, the conventional server or node architecture can involve multiple layers of memory copy and bus protocols, which can result in a high consumption of resources and power as well as a sub-optimal operational efficiency. Furthermore, despite the decreased cost of CPU usage in conventional servers, the cost of using CPU cores for management instead of computation tasks still remains high.
In one embodiment, a printed circuit board comprises: a network controller; a memory controller; a heterogeneous processor; a field-programmable gate array (FPGA); and a non-volatile-media controller. The memory controller comprises: a fabric controller component configured to communicate with the network controller, the heterogeneous processor, the FPGA, and the non-volatile-media controller; and a media controller component configured to manage access relating to data stored in a volatile memory media. The FPGA is configured to perform computations relating to data stored via the non-volatile-media controller. The heterogeneous processor is configured to perform computation tasks relating to data stored via the memory controller.
In some embodiments, the fabric controller is further configured to manage data received and transmitted via an internal bus protocol to communicate with the network controller, the heterogeneous processor, the FPGA, and the non-volatile-media controller.
In some embodiments, the printed circuit board further comprises: a power module; a monitoring and operating module; and at least one network interface configured to receive data from another node or computing device and further configured to transmit data to the network controller.
In some embodiments, the printed circuit board is immersed in a liquid cooling medium.
In some embodiments, the network controller is an Ethernet controller, the non-volatile-media controller is a NAND controller, and the memory controller is a DRAM controller.
In some embodiments, the network controller further comprises: at least one static random-access memory (SRAM) or embedded dynamic random-access memory (DRAM); an Ethernet interface; at least one ARM core; a bus switch; and a bus root complex or an endpoint.
In some embodiments, the printed circuit board is configured to be plugged into a rack with a plurality of other plugged-in circuit boards. The printed circuit board comprises a node, the other plugged-in circuit boards comprise nodes, and each node is connected to one of a plurality of switches in the rack.
In some embodiments, a switch is an Ethernet switch or an access switch.
In some embodiments, the rack is immersed with the plugged-in plurality of nodes in a liquid cooling tank. A defective node of the immersed rack is identified. The defective node is removed without affecting operation of the rack or a remainder of the nodes in the rack. The defective node is replaced or repaired to obtain a new node. The new node is plugged into the rack by immersing the new node in the liquid cooling tank at a location previously occupied by the defective node.
Another system provides a method for facilitating operation of a storage system. During operation, the system receives, by a network controller of a device, a request to write first data to a non-volatile memory of the device, wherein the device comprises: the network controller; a memory controller; a heterogeneous processor; a field-programmable gate array (FPGA); and a non-volatile-media controller. The memory controller comprises: a fabric controller component configured to communicate with the network controller, the heterogeneous processor, the FPGA, and the non-volatile-media controller; and a media controller component configured to manage access relating to data stored in a volatile memory media. The system performs, by the heterogeneous processor, computation tasks relating to the first data and data stored via the memory controller. The system performs, by the FPGA, computations relating to the first data and data stored via the non-volatile-media controller. The system writes, by the non-volatile-media controller, the first data to a non-volatile media of the device.
In the figures, like reference numerals refer to the same figure elements.
The following description is presented to enable any person skilled in the art to make and use the embodiments, and is provided in the context of a particular application and its requirements. Various modifications to the disclosed embodiments will be readily apparent to those skilled in the art, and the general principles defined herein may be applied to other embodiments and applications without departing from the spirit and scope of the present disclosure. Thus, the embodiments described herein are not limited to the embodiments shown, but are to be accorded the widest scope consistent with the principles and features disclosed herein.
Overview
The embodiments described herein facilitate a distributed storage system with converged computation and storage nodes without using a general CPU, which can improve the efficiency and performance of the storage system.
As described above, in technological fields requiring analysis of large amounts of data (such as big data analysis), a distributed storage system or distributed cluster may divide computation-heavy tasks into multiple partitions or sub-tasks, where the sub-tasks may be performed in parallel by multiple nodes in the distributed cluster. A single node in such a distributed cluster can include both computation and storage capability. A single node can be a conventional server in a distributed cluster, and can use a CPU as a central hub to manage traffic among devices or components in the node or server. Because certain computation workloads may not be friendly or optimal for a CPU, various heterogeneous processors can be used as PCIe devices to handle the main computation tasks. Examples of heterogeneous processors include a GPU and an FPGA. The heterogeneous processors can generally be used in a conventional server as a PCIe AIC. By shifting the main computation tasks from the CPU to the heterogeneous processors, a distributed storage system can deploy the CPU with a decreased cost and power consumption.
However, the conventional server or node architecture can involve multiple layers of memory copy and bus protocols, which can result in a high consumption of resources and power as well a sub-optimal operational efficiency. Furthermore, despite the decreased cost of CPU usage in conventional servers, the cost of using CPU cores for management instead of computation tasks still remains high. An exemplary conventional server is described below in relation to
The embodiments described herein addresses these challenges by providing a converged computation and storage node which includes a network controller and a memory controller which is directly accessible via the network controller and other components of the node. The node can be a single printed circuit board with multiple components or integrated circuits (ICs). The node can include: the network controller, which can be an Ethernet controller; the memory controller, which can be a DRAM controller (or a “standalone DRAM controller”); a heterogeneous processor; a storage field-programmable gate array (FPGA); and a non-volatile media controller, which can be a NAND controller. The storage FPGA can be configured to perform computations relating to data stored via the NAND controller, and the heterogeneous processor can be configured to perform computation tasks. The NAND controller can be configured to manage high-capacity NAND flash. The Ethernet controller can be configured to communicate with other nodes and extend the data processing and storage functions within an internal bus. Thus, the node can provide both computation and storage functionality, and can be referred to as “a converged computation and storage node.”
In this converged computation and storage node, the standalone DRAM controller can be used and accessed directly by the heterogeneous processor, the storage FPGA, the NAND controller, and the Ethernet controller through a peer-to-peer direct memory access. This converged node can thus eliminate the need for a middle layer between an initiator and a target of the bus, which can result in a decrease in the amount of overhead traditionally associated with copying data to and from memory, as described below in relation to
Thus, by placing both computation and storage functionality in a single node (e.g., a printed circuit board), the system can eliminate the need to use a CPU, which can result in an improvement in performance and a reduction in power consumption through use of the heterogeneous processors, which can provide the computation power. In the single node, the system can further provide direct access to the node's memory controller via several other node components, including a heterogeneous processor, a storage FPGA, and a NAND controller. These improvements can result in a more efficient distributed storage system.
A “storage system infrastructure,” “storage infrastructure,” or “storage system” refers to the overall set of hardware and software components used to facilitate storage for a system. A storage system can include multiple clusters of storage servers and other servers. A “storage server” refers to a computing device which can include multiple storage devices or storage drives. A “storage device” or a “storage drive” refers to a device or a drive with a non-volatile memory which can provide persistent storage of data, e.g., a solid state drive (SSD), a hard disk drive (HDD), or a flash-based storage device.
A “computing device” refers to any server, device, node, entity, drive, or any other entity which can provide any computing capabilities.
A “node” refers to a device or printed circuit board which can be used in a distributed storage system. A “converged computation and storage node” refers to a node which includes functionality for both computation and storage, and can include the components, modules, unit, or ICs as described herein.
A “network controller” refers to a controller which can be configured to manage, handle, or otherwise process traffic received from and transmitted via, e.g., an Ethernet interface. A network controller can include the Ethernet interface, SRAM/embedded DRAM (eDRAM), a bus switch, a bus root complex, and ARM cores which provide conversion between Ethernet and an internal bus. A network controller can also communicate with or be coupled to a “memory controller” (see term below). An exemplary network controller is described below in relation to
A “heterogeneous processor” refers to a processor which can be configured to perform computation tasks, including high-cost tasks requiring high-performance, e.g., related to artificial intelligence and big data analysis.
A “non-volatile-media controller” refers to a controller, component, unit, module, IC, software, firmware, or hardware component which can be configured to manage, process, store, or other handle data stored in or to be stored in a non-volatile memory of a system. An example of a non-volatile media is NAND flash, and an example of a non-volatile-media controller is a NAND controller.
A “storage FPGA” refers to an FPGA which can be configured to perform computations relating to data stored or to be stored, e.g., via a non-volatile media controller of the same node or PCB.
A “memory controller” refers to a controller which can be configured to manage, process, store, or otherwise handle data stored in or to be stored in a volatile memory of a system. An example of a memory controller is a DRAM controller. In this disclosure, a memory controller can include a “fabric controller” and a “media controller.” The fabric controller can be used to manage data received from and transmitted via an internal bus protocol to communicate with other components or devices, including, but not limited to, a network controller, a storage FPGA, a heterogeneous processor, and a non-volatile-media controller. The media controller can be configured to handle the access of data stored in the non-volatile media and to provide reliability assurance of the stored data based on the characteristics of the non-volatile media.
Architecture of Exemplary Node and Environment in the Prior Art
At the same time, due to the increasing demand to improve the efficiency of data processing (e.g., in fields such as big data analysis), heterogeneous processors can generally be used in a conventional server as a PCIe AIC. By shifting the main computation tasks from the CPU to the heterogeneous processors, the CPU can be deployed with a decreased cost and low power consumption. However, the conventional server or node architecture can involve multiple layers of memory copy and bus protocols, which can result in high consumption of resources and power, and a sub-optimal operational efficiency. Furthermore, despite the decreased cost of CPU usage in conventional servers, the cost of using CPU cores for management instead of computation tasks still remains high.
In addition to the cost of the CPU, the conventional server is limited by several other constraints. First, CPU sources are limited. Only a few vendors fabricate or manufacture CPUs, which can result in frequent challenges in the supply chain. Second, power utilization can be difficult to control in the conventional server. Because the CPU is no longer performing the main computation tasks (at least in certain compute scenarios), the high power consumption can result in a non-trivial constraints on the deployment of the servers and the optimization of the total cost of operation (TCO). It may be sub-optimal to spend major resources (in terms of cost, components, performance, TCO, etc.) on what may be considered a secondary module (i.e., the CPU). Third, the performance of the storage system can face a bottleneck due to the usage of the CPU and the conventional node and cluster architecture. Moreover, the memory bandwidth, memory copy, and protocol overhead may all act as further constraints on the optimization of the performance of the storage system.
Thus, the conventional server which uses the CPU as the central hub can be limited by several constraints which can limit the optimization of a distributed storage system.
Exemplary Converged Computation and Storage Node
NAND controller 150 can be configured to manage the high-capacity NAND flash. Storage FPGA 140 can be configured with customized functions to perform computations or computation tasks relating to data stored via NAND controller 150. Heterogeneous processor 130 can be configured to perform computations relating to data stored via DRAM controller 120. These computations can include main computation tasks relating to, e.g., artificial intelligence and big data analysis. Ethernet controller 110 can be configured to communicate with other nodes and to extend the processing and storage of data using the internal bus. Monitoring/operating module 164 can be configured to perform monitoring and operating functions relating to operation of the node, including receiving, processing, accessing, handling, directing, and storing data.
Thus, by providing computation and storage functionality in the above described manner in converged nodes without requiring a central CPU in each node, the embodiments described herein provide converged nodes in a distributed storage system which can improve the optimization and efficiency of the distributed storage system.
Exemplary Network Controller
Exemplary Rack and Immersion in Liquid Cooling Tank
In a conventional server rack, the converged storage nodes of
Exemplary Environment for Accessing Memory in the Prior Art vs. in a Converged Computation and Storage Node
In prior art environment 500, memory module 530 is depicted as a passive data bucket because both memory controller 520 and media controller 522 are located on CPU SOC 510. Media controller 512 is designed as a slave of CPU SOC 510, which causes all DRAM usage to travel or go through CPU SOC 510. Using CPU SOC 510 as a central hub like this can result in a significant overhead in the performance of the distributed storage system.
Memory controller 650 can include two components or modules: fabric controller 652 and media controller 654. Fabric controller 652 can be configured to communicate with Ethernet controller 610, heterogeneous processor 630, storage FPGA 620, and NAND controller 640. Media controller 654 can be configured to communicate with volatile memory (e.g., DRAMs 660-666) and other types of memory (e.g., SCM 668). Each of Ethernet controller 610, heterogeneous processor 630, storage FPGA 620, and NAND controller 640 can directly access memory controller 650, and can thus access data stored in or to be stored in DRAMs 660-666 and SCM 668.
Thus, the embodiments described herein provide a distributed storage system with converged computation and storage nodes which communicate via an Ethernet connection. These converged nodes can result in an improvement in the efficiency and performance of the overall distributed storage system by removing the traditional CPU sockets, which can also result in a reduction in both cost and power consumption in scenarios in which a heterogeneous processor can provide computation power. The described memory controller can be shared by the heterogeneous processor, the storage FPGA, the non-volatile-media controller (the NAND controller), and the network controller (the Ethernet controller), but the system can allocate and recycle the memory capacity and space individually without sharing in order to ensure coherence.
Method for Facilitating Operation of a Storage System
As described above, the device can be a printed circuit board, and the rack can be immersed into a liquid cooling tank.
Exemplary Computer System and Apparatus
Content-processing system 818 can include instructions, which when executed by computer system 800, can cause computer system 800 or processor 802 to perform methods and/or processes described in this disclosure. Specifically, content-processing system 818 can include instructions for receiving and transmitting data packets, including data to be read or written and an input/output (I/O) request (e.g., a read request or a write request) (communication module 820).
Content-processing system 818 can further include instructions for receiving, by a network controller of a device, a request to write first data to a non-volatile memory of the device, wherein the device comprises: the network controller; a memory controller; a heterogeneous processor; a field-programmable gate array (FPGA); and a non-volatile-media controller (communication module 820 and network-controlling module 822). The memory controller comprises: a fabric controller component configured to communicate with the network controller, the heterogeneous processor, the FPGA, and the non-volatile-media controller; and a media controller component configured to manage access relating to data stored in a volatile memory media (memory-controlling module 824). Content-processing system 818 can include instructions for performing, by the heterogeneous processor, computation tasks relating to the first data and data stored via the memory controller (first computation-performing module 826). Content-processing system 818 can include instructions for performing, by the FPGA, computations relating to the first data and data stored via the non-volatile-media controller (second computation-performing module 828). Content-processing system 818 can include instructions for writing, by the non-volatile-media controller, the first data to a non-volatile media of the device (NAND-managing module 832).
Content-processing system 818 can include instructions for identifying a defective node of an immersed rack (defective node-managing module 830), where the rack is immersed with the plugged-in plurality of nodes into a liquid cooling tank. Content-processing system 818 can include instructions for removing the defective node without affecting operation of the rack or a remainder of the nodes in the rack, replacing or repairing the defective node to obtain a new node, and plugging the new node into the rack by immersing the new node into the liquid cooling tank in a location previously occupied by the defective node (actions which can be performed by a node-repairing module (not shown) or by a user or administrative agent.
Data 834 can include any data that is required as input or generated as output by the methods and/or processes described in this disclosure. Specifically, data 834 can store at least: data; a request; a read request; a write request; an input/output (I/O) request; data or metadata associated with a read request, a write request, or an I/O request; an indicator or identifier of a printed circuit board; an indicator or identifier of an interface or other component, circuit, IC, module, or unit on a printed circuit board; an indicator or identifier of a network controller, a processor, a heterogeneous processor, a storage FPGA, a memory controller, or a non-volatile memory controller; an indicator or identifier of a power module, a monitoring and operating module, and a network interface; an indicator or identifier of components of a network controller, including an SRAM or embedded DRAM, an Ethernet interface, at least one ARM core, a bus switch, and a bus root complex or an endpoint; a node or switch identifier; an indication or identifier of a defective node, a new node, or a rack; and any information or data related to, traveling through, stored in, or accessed from a printed circuit board or any component of a printed circuit board as described herein.
Apparatus 900 can comprise modules or units 902-914 which are configured to perform functions or operations similar to modules 820-832 of computer system 800 of
The data structures and code described in this detailed description are typically stored on a computer-readable storage medium, which may be any device or medium that can store code and/or data for use by a computer system. The computer-readable storage medium includes, but is not limited to, volatile memory, non-volatile memory, magnetic and optical storage devices such as disk drives, magnetic tape, CDs (compact discs), DVDs (digital versatile discs or digital video discs), or other media capable of storing computer-readable media now known or later developed.
The methods and processes described in the detailed description section can be embodied as code and/or data, which can be stored in a computer-readable storage medium as described above. When a computer system reads and executes the code and/or data stored on the computer-readable storage medium, the computer system performs the methods and processes embodied as data structures and code and stored within the computer-readable storage medium.
Furthermore, the methods and processes described above can be included in hardware modules. For example, the hardware modules can include, but are not limited to, application-specific integrated circuit (ASIC) chips, field-programmable gate arrays (FPGAs), and other programmable-logic devices now known or later developed. When the hardware modules are activated, the hardware modules perform the methods and processes included within the hardware modules.
The foregoing embodiments described herein have been presented for purposes of illustration and description only. They are not intended to be exhaustive or to limit the embodiments described herein to the forms disclosed. Accordingly, many modifications and variations will be apparent to practitioners skilled in the art. Additionally, the above disclosure is not intended to limit the embodiments described herein. The scope of the embodiments described herein is defined by the appended claims.
Number | Name | Date | Kind |
---|---|---|---|
3893071 | Bossen | Jul 1975 | A |
4562494 | Bond | Dec 1985 | A |
4718067 | Peters | Jan 1988 | A |
4775932 | Oxley | Oct 1988 | A |
4858040 | Hazebrouck | Aug 1989 | A |
5394382 | Hu | Feb 1995 | A |
5602693 | Brunnett | Feb 1997 | A |
5715471 | Otsuka | Feb 1998 | A |
5732093 | Huang | Mar 1998 | A |
5802551 | Komatsu | Sep 1998 | A |
5930167 | Lee | Jul 1999 | A |
6098185 | Wilson | Aug 2000 | A |
6148377 | Carter | Nov 2000 | A |
6226650 | Mahajan et al. | May 2001 | B1 |
6243795 | Yang | Jun 2001 | B1 |
6457104 | Tremaine | Sep 2002 | B1 |
6658478 | Singhal | Dec 2003 | B1 |
6795894 | Neufeld | Sep 2004 | B1 |
7351072 | Muff | Apr 2008 | B2 |
7565454 | Zuberi | Jul 2009 | B2 |
7599139 | Bombet | Oct 2009 | B1 |
7953899 | Hooper | May 2011 | B1 |
7958433 | Yoon | Jun 2011 | B1 |
8085569 | Kim | Dec 2011 | B2 |
8144512 | Huang | Mar 2012 | B2 |
8166233 | Schibilla | Apr 2012 | B2 |
8260924 | Koretz | Sep 2012 | B2 |
8281061 | Radke | Oct 2012 | B2 |
8452819 | Sorenson, III | May 2013 | B1 |
8516284 | Chan | Aug 2013 | B2 |
8527544 | Colgrove | Sep 2013 | B1 |
8751763 | Ramarao | Jun 2014 | B1 |
8819367 | Fallone | Aug 2014 | B1 |
8825937 | Atkisson | Sep 2014 | B2 |
8832688 | Tang | Sep 2014 | B2 |
8868825 | Hayes | Oct 2014 | B1 |
8904061 | O'Brien, III | Dec 2014 | B1 |
8949208 | Xu | Feb 2015 | B1 |
9015561 | Hu | Apr 2015 | B1 |
9031296 | Kaempfer | May 2015 | B2 |
9043545 | Kimmel | May 2015 | B2 |
9088300 | Chen | Jul 2015 | B1 |
9092223 | Pani | Jul 2015 | B1 |
9129628 | Fallone | Sep 2015 | B1 |
9141176 | Chen | Sep 2015 | B1 |
9208817 | Li | Dec 2015 | B1 |
9213627 | Van Acht | Dec 2015 | B2 |
9280472 | Dang | Mar 2016 | B1 |
9280487 | Candelaria | Mar 2016 | B2 |
9311939 | Malina | Apr 2016 | B1 |
9336340 | Dong | May 2016 | B1 |
9436595 | Benitez | Sep 2016 | B1 |
9495263 | Pang | Nov 2016 | B2 |
9529601 | Dharmadhikari | Dec 2016 | B1 |
9529670 | O'Connor | Dec 2016 | B2 |
9575982 | Sankara Subramanian | Feb 2017 | B1 |
9588698 | Karamcheti | Mar 2017 | B1 |
9588977 | Wang | Mar 2017 | B1 |
9607631 | Rausch | Mar 2017 | B2 |
9671971 | Trika | Jun 2017 | B2 |
9747202 | Shaharabany | Aug 2017 | B1 |
9852076 | Garg | Dec 2017 | B1 |
9875053 | Frid | Jan 2018 | B2 |
9910705 | Mak | Mar 2018 | B1 |
9912530 | Singatwaria | Mar 2018 | B2 |
9946596 | Hashimoto | Apr 2018 | B2 |
10013169 | Fisher | Jul 2018 | B2 |
10199066 | Feldman | Feb 2019 | B1 |
10229735 | Natarajan | Mar 2019 | B1 |
10235198 | Qiu | Mar 2019 | B2 |
10268390 | Warfield | Apr 2019 | B2 |
10318467 | Barzik | Jun 2019 | B2 |
10361722 | Lee | Jul 2019 | B2 |
10437670 | Koltsidas | Oct 2019 | B1 |
10459663 | Agombar | Oct 2019 | B2 |
10642522 | Li | May 2020 | B2 |
10649657 | Zaidman | May 2020 | B2 |
10678432 | Dreier | Jun 2020 | B1 |
10756816 | Dreier | Aug 2020 | B1 |
10928847 | Suresh | Feb 2021 | B2 |
20010032324 | Slaughter | Oct 2001 | A1 |
20020010783 | Primak | Jan 2002 | A1 |
20020039260 | Kilmer | Apr 2002 | A1 |
20020073358 | Atkinson | Jun 2002 | A1 |
20020095403 | Chandrasekaran | Jul 2002 | A1 |
20020112085 | Berg | Aug 2002 | A1 |
20020161890 | Chen | Oct 2002 | A1 |
20030074319 | Jaquette | Apr 2003 | A1 |
20030145274 | Hwang | Jul 2003 | A1 |
20030163594 | Aasheim | Aug 2003 | A1 |
20030163633 | Aasheim | Aug 2003 | A1 |
20030217080 | White | Nov 2003 | A1 |
20040010545 | Pandya | Jan 2004 | A1 |
20040066741 | Dinker | Apr 2004 | A1 |
20040103238 | Avraham | May 2004 | A1 |
20040143718 | Chen | Jul 2004 | A1 |
20040255171 | Zimmer | Dec 2004 | A1 |
20040267752 | Wong | Dec 2004 | A1 |
20040268278 | Hoberman | Dec 2004 | A1 |
20050038954 | Saliba | Feb 2005 | A1 |
20050097126 | Cabrera | May 2005 | A1 |
20050138325 | Hofstee | Jun 2005 | A1 |
20050144358 | Conley | Jun 2005 | A1 |
20050149827 | Lambert | Jul 2005 | A1 |
20050174670 | Dunn | Aug 2005 | A1 |
20050177672 | Rao | Aug 2005 | A1 |
20050177755 | Fung | Aug 2005 | A1 |
20050195635 | Conley | Sep 2005 | A1 |
20050235067 | Creta | Oct 2005 | A1 |
20050235171 | Igari | Oct 2005 | A1 |
20060031709 | Hiraiwa | Feb 2006 | A1 |
20060101197 | Georgis | May 2006 | A1 |
20060156012 | Beeson | Jul 2006 | A1 |
20060184813 | Bui | Aug 2006 | A1 |
20070033323 | Gorobets | Feb 2007 | A1 |
20070061502 | Lasser | Mar 2007 | A1 |
20070101096 | Gorobets | May 2007 | A1 |
20070250756 | Gower | Oct 2007 | A1 |
20070266011 | Rohrs | Nov 2007 | A1 |
20070283081 | Lasser | Dec 2007 | A1 |
20070283104 | Wellwood | Dec 2007 | A1 |
20070285980 | Shimizu | Dec 2007 | A1 |
20080034154 | Lee | Feb 2008 | A1 |
20080065805 | Wu | Mar 2008 | A1 |
20080082731 | Karamcheti | Apr 2008 | A1 |
20080104369 | Reed | May 2008 | A1 |
20080112238 | Kim | May 2008 | A1 |
20080163033 | Yim | Jul 2008 | A1 |
20080301532 | Uchikawa | Dec 2008 | A1 |
20090006667 | Lin | Jan 2009 | A1 |
20090089544 | Liu | Apr 2009 | A1 |
20090113219 | Aharonov | Apr 2009 | A1 |
20090125788 | Wheeler | May 2009 | A1 |
20090183052 | Kanno | Jul 2009 | A1 |
20090254705 | Abali | Oct 2009 | A1 |
20090282275 | Yermalayeu | Nov 2009 | A1 |
20090287956 | Flynn | Nov 2009 | A1 |
20090307249 | Koifman | Dec 2009 | A1 |
20090307426 | Galloway | Dec 2009 | A1 |
20090310412 | Jang | Dec 2009 | A1 |
20100031000 | Flynn | Feb 2010 | A1 |
20100169470 | Takashige | Jul 2010 | A1 |
20100217952 | Iyer | Aug 2010 | A1 |
20100229224 | Etchegoyen | Sep 2010 | A1 |
20100241848 | Smith | Sep 2010 | A1 |
20100321999 | Yoo | Dec 2010 | A1 |
20100325367 | Kornegay | Dec 2010 | A1 |
20100332922 | Chang | Dec 2010 | A1 |
20110031546 | Uenaka | Feb 2011 | A1 |
20110055458 | Kuehne | Mar 2011 | A1 |
20110055471 | Thatcher | Mar 2011 | A1 |
20110060722 | Li | Mar 2011 | A1 |
20110072204 | Chang | Mar 2011 | A1 |
20110099418 | Chen | Apr 2011 | A1 |
20110153903 | Hinkle | Jun 2011 | A1 |
20110161784 | Selinger | Jun 2011 | A1 |
20110191525 | Hsu | Aug 2011 | A1 |
20110218969 | Anglin | Sep 2011 | A1 |
20110231598 | Hatsuda | Sep 2011 | A1 |
20110239083 | Kanno | Sep 2011 | A1 |
20110252188 | Weingarten | Oct 2011 | A1 |
20110258514 | Lasser | Oct 2011 | A1 |
20110289263 | McWilliams | Nov 2011 | A1 |
20110289280 | Koseki | Nov 2011 | A1 |
20110292538 | Haga | Dec 2011 | A1 |
20110296411 | Tang | Dec 2011 | A1 |
20110299317 | Shaeffer | Dec 2011 | A1 |
20110302353 | Confalonieri | Dec 2011 | A1 |
20120017037 | Riddle | Jan 2012 | A1 |
20120039117 | Webb | Feb 2012 | A1 |
20120084523 | Littlefield | Apr 2012 | A1 |
20120089774 | Kelkar | Apr 2012 | A1 |
20120096330 | Przybylski | Apr 2012 | A1 |
20120117399 | Chan | May 2012 | A1 |
20120147021 | Cheng | Jun 2012 | A1 |
20120151253 | Horn | Jun 2012 | A1 |
20120159099 | Lindamood | Jun 2012 | A1 |
20120159289 | Piccirillo | Jun 2012 | A1 |
20120173792 | Lassa | Jul 2012 | A1 |
20120203958 | Jones | Aug 2012 | A1 |
20120210095 | Nellans | Aug 2012 | A1 |
20120233523 | Krishnamoorthy | Sep 2012 | A1 |
20120246392 | Cheon | Sep 2012 | A1 |
20120278579 | Goss | Nov 2012 | A1 |
20120284587 | Yu | Nov 2012 | A1 |
20120324312 | Moyer | Dec 2012 | A1 |
20120331207 | Lassa | Dec 2012 | A1 |
20130013880 | Tashiro | Jan 2013 | A1 |
20130016970 | Koka | Jan 2013 | A1 |
20130018852 | Barton | Jan 2013 | A1 |
20130024605 | Sharon | Jan 2013 | A1 |
20130054822 | Mordani | Feb 2013 | A1 |
20130061029 | Huff | Mar 2013 | A1 |
20130073798 | Kang | Mar 2013 | A1 |
20130080391 | Raichstein | Mar 2013 | A1 |
20130145085 | Yu | Jun 2013 | A1 |
20130145089 | Eleftheriou | Jun 2013 | A1 |
20130151759 | Shim | Jun 2013 | A1 |
20130159251 | Skrenta | Jun 2013 | A1 |
20130159723 | Brandt | Jun 2013 | A1 |
20130166820 | Batwara | Jun 2013 | A1 |
20130173845 | Aslam | Jul 2013 | A1 |
20130191601 | Peterson | Jul 2013 | A1 |
20130205183 | Fillingim | Aug 2013 | A1 |
20130219131 | Alexandron | Aug 2013 | A1 |
20130227347 | Cho | Aug 2013 | A1 |
20130238955 | D Abreu | Sep 2013 | A1 |
20130254622 | Kanno | Sep 2013 | A1 |
20130318283 | Small | Nov 2013 | A1 |
20130318395 | Kalavade | Nov 2013 | A1 |
20130329492 | Yang | Dec 2013 | A1 |
20140006688 | Yu | Jan 2014 | A1 |
20140019650 | Li | Jan 2014 | A1 |
20140025638 | Hu | Jan 2014 | A1 |
20140082273 | Segev | Mar 2014 | A1 |
20140082412 | Matsumura | Mar 2014 | A1 |
20140095769 | Borkenhagen | Apr 2014 | A1 |
20140095827 | Wei | Apr 2014 | A1 |
20140108414 | Stillerman | Apr 2014 | A1 |
20140108891 | Strasser | Apr 2014 | A1 |
20140164447 | Tarafdar | Jun 2014 | A1 |
20140164879 | Tam | Jun 2014 | A1 |
20140181532 | Camp | Jun 2014 | A1 |
20140195564 | Talagala | Jul 2014 | A1 |
20140215129 | Kuzmin | Jul 2014 | A1 |
20140223079 | Zhang | Aug 2014 | A1 |
20140233950 | Luo | Aug 2014 | A1 |
20140250259 | Ke | Sep 2014 | A1 |
20140279927 | Constantinescu | Sep 2014 | A1 |
20140304452 | De La Iglesia | Oct 2014 | A1 |
20140310574 | Yu | Oct 2014 | A1 |
20140359229 | Cota-Robles | Dec 2014 | A1 |
20140365707 | Talagala | Dec 2014 | A1 |
20150019798 | Huang | Jan 2015 | A1 |
20150060009 | Shelnutt | Mar 2015 | A1 |
20150062806 | Shelnutt | Mar 2015 | A1 |
20150082317 | You | Mar 2015 | A1 |
20150106556 | Yu | Apr 2015 | A1 |
20150106559 | Cho | Apr 2015 | A1 |
20150121031 | Feng | Apr 2015 | A1 |
20150142752 | Chennamsetty | May 2015 | A1 |
20150143030 | Gorobets | May 2015 | A1 |
20150199234 | Choi | Jul 2015 | A1 |
20150227316 | Warfield | Aug 2015 | A1 |
20150234845 | Moore | Aug 2015 | A1 |
20150269964 | Fallone | Sep 2015 | A1 |
20150277937 | Swanson | Oct 2015 | A1 |
20150286477 | Mathur | Oct 2015 | A1 |
20150294684 | Qjang | Oct 2015 | A1 |
20150301964 | Brinicombe | Oct 2015 | A1 |
20150304108 | Obukhov | Oct 2015 | A1 |
20150310916 | Leem | Oct 2015 | A1 |
20150317095 | Voigt | Nov 2015 | A1 |
20150341123 | Nagarajan | Nov 2015 | A1 |
20150347025 | Law | Dec 2015 | A1 |
20150363271 | Haustein | Dec 2015 | A1 |
20150363328 | Candelaria | Dec 2015 | A1 |
20150372597 | Luo | Dec 2015 | A1 |
20150382515 | James | Dec 2015 | A1 |
20160014039 | Reddy | Jan 2016 | A1 |
20160026575 | Samanta | Jan 2016 | A1 |
20160041760 | Kuang | Feb 2016 | A1 |
20160048327 | Jayasena | Feb 2016 | A1 |
20160048341 | Constantinescu | Feb 2016 | A1 |
20160054922 | Awasthi | Feb 2016 | A1 |
20160062885 | Ryu | Mar 2016 | A1 |
20160077749 | Ravimohan | Mar 2016 | A1 |
20160077764 | Ori | Mar 2016 | A1 |
20160077968 | Sela | Mar 2016 | A1 |
20160098344 | Gorobets | Apr 2016 | A1 |
20160098350 | Tang | Apr 2016 | A1 |
20160103631 | Ke | Apr 2016 | A1 |
20160110254 | Cronie | Apr 2016 | A1 |
20160132237 | Jeong | May 2016 | A1 |
20160154601 | Chen | Jun 2016 | A1 |
20160155750 | Yasuda | Jun 2016 | A1 |
20160162187 | Lee | Jun 2016 | A1 |
20160179399 | Melik-Martirosian | Jun 2016 | A1 |
20160188223 | Camp | Jun 2016 | A1 |
20160188890 | Naeimi | Jun 2016 | A1 |
20160203000 | Parmar | Jul 2016 | A1 |
20160224267 | Yang | Aug 2016 | A1 |
20160232103 | Schmisseur | Aug 2016 | A1 |
20160234297 | Ambach | Aug 2016 | A1 |
20160239074 | Lee | Aug 2016 | A1 |
20160239380 | Wideman | Aug 2016 | A1 |
20160274636 | Kim | Sep 2016 | A1 |
20160306699 | Resch | Oct 2016 | A1 |
20160306853 | Sabaa | Oct 2016 | A1 |
20160321002 | Jung | Nov 2016 | A1 |
20160335085 | Scalabrino | Nov 2016 | A1 |
20160342345 | Kankani | Nov 2016 | A1 |
20160343429 | Nieuwejaar | Nov 2016 | A1 |
20160350002 | Vergis | Dec 2016 | A1 |
20160350385 | Poder | Dec 2016 | A1 |
20160364146 | Kuttner | Dec 2016 | A1 |
20160381442 | Heanue | Dec 2016 | A1 |
20170004037 | Park | Jan 2017 | A1 |
20170010652 | Huang | Jan 2017 | A1 |
20170075583 | Alexander | Mar 2017 | A1 |
20170075594 | Badam | Mar 2017 | A1 |
20170091110 | Ash | Mar 2017 | A1 |
20170109199 | Chen | Apr 2017 | A1 |
20170109232 | Cha | Apr 2017 | A1 |
20170123655 | Sinclair | May 2017 | A1 |
20170147499 | Mohan | May 2017 | A1 |
20170161202 | Erez | Jun 2017 | A1 |
20170162235 | De | Jun 2017 | A1 |
20170168986 | Sajeepa | Jun 2017 | A1 |
20170177217 | Kanno | Jun 2017 | A1 |
20170177259 | Motwani | Jun 2017 | A1 |
20170185498 | Gao | Jun 2017 | A1 |
20170192848 | Pamies-Juarez | Jul 2017 | A1 |
20170199823 | Hayes | Jul 2017 | A1 |
20170212708 | Suhas | Jul 2017 | A1 |
20170220254 | Warfield | Aug 2017 | A1 |
20170221519 | Matsuo | Aug 2017 | A1 |
20170228157 | Yang | Aug 2017 | A1 |
20170242722 | Qiu | Aug 2017 | A1 |
20170249162 | Tsirkin | Aug 2017 | A1 |
20170262176 | Kanno | Sep 2017 | A1 |
20170262178 | Hashimoto | Sep 2017 | A1 |
20170262217 | Pradhan | Sep 2017 | A1 |
20170269998 | Sunwoo | Sep 2017 | A1 |
20170279460 | Camp | Sep 2017 | A1 |
20170285976 | Durham | Oct 2017 | A1 |
20170286311 | Juenemann | Oct 2017 | A1 |
20170322888 | Booth | Nov 2017 | A1 |
20170344470 | Yang | Nov 2017 | A1 |
20170344491 | Pandurangan | Nov 2017 | A1 |
20170353576 | Guim Bernat | Dec 2017 | A1 |
20180024772 | Madraswala | Jan 2018 | A1 |
20180024779 | Kojima | Jan 2018 | A1 |
20180033491 | Marelli | Feb 2018 | A1 |
20180052797 | Barzik | Feb 2018 | A1 |
20180067847 | Oh | Mar 2018 | A1 |
20180069658 | Benisty | Mar 2018 | A1 |
20180074730 | Inoue | Mar 2018 | A1 |
20180076828 | Kanno | Mar 2018 | A1 |
20180088867 | Kaminaga | Mar 2018 | A1 |
20180107591 | Smith | Apr 2018 | A1 |
20180113631 | Zhang | Apr 2018 | A1 |
20180143780 | Cho | May 2018 | A1 |
20180150640 | Li | May 2018 | A1 |
20180165038 | Authement | Jun 2018 | A1 |
20180165169 | Camp | Jun 2018 | A1 |
20180165340 | Agarwal | Jun 2018 | A1 |
20180167268 | Liguori | Jun 2018 | A1 |
20180173620 | Cen | Jun 2018 | A1 |
20180188970 | Liu | Jul 2018 | A1 |
20180189175 | Ji | Jul 2018 | A1 |
20180189182 | Wang | Jul 2018 | A1 |
20180212951 | Goodrum | Jul 2018 | A1 |
20180219561 | Litsyn | Aug 2018 | A1 |
20180226124 | Perner | Aug 2018 | A1 |
20180232151 | Badam | Aug 2018 | A1 |
20180260148 | Klein | Sep 2018 | A1 |
20180270110 | Chugtu | Sep 2018 | A1 |
20180293014 | Ravimohan | Oct 2018 | A1 |
20180300203 | Kathpal | Oct 2018 | A1 |
20180321864 | Benisty | Nov 2018 | A1 |
20180322024 | Nagao | Nov 2018 | A1 |
20180329776 | Lai | Nov 2018 | A1 |
20180336921 | Ryun | Nov 2018 | A1 |
20180349396 | Blagojevic | Dec 2018 | A1 |
20180356992 | Lamberts | Dec 2018 | A1 |
20180357126 | Dhuse | Dec 2018 | A1 |
20180373428 | Kan | Dec 2018 | A1 |
20180373655 | Liu | Dec 2018 | A1 |
20180373664 | Vijayrao | Dec 2018 | A1 |
20190012111 | Li | Jan 2019 | A1 |
20190050327 | Li | Feb 2019 | A1 |
20190065085 | Jean | Feb 2019 | A1 |
20190073261 | Halbert | Mar 2019 | A1 |
20190073262 | Chen | Mar 2019 | A1 |
20190087089 | Yoshida | Mar 2019 | A1 |
20190087115 | Li | Mar 2019 | A1 |
20190087328 | Kanno | Mar 2019 | A1 |
20190108145 | Raghava | Apr 2019 | A1 |
20190116127 | Pismenny | Apr 2019 | A1 |
20190166725 | Jing | May 2019 | A1 |
20190171532 | Abadi | Jun 2019 | A1 |
20190172820 | Meyers | Jun 2019 | A1 |
20190196748 | Badam | Jun 2019 | A1 |
20190196907 | Khan | Jun 2019 | A1 |
20190205206 | Hornung | Jul 2019 | A1 |
20190212949 | Pletka | Jul 2019 | A1 |
20190220392 | Lin | Jul 2019 | A1 |
20190227927 | Miao | Jul 2019 | A1 |
20190272242 | Kachare | Sep 2019 | A1 |
20190278654 | Kaynak | Sep 2019 | A1 |
20190317901 | Kachare | Oct 2019 | A1 |
20190339998 | Momchilov | Nov 2019 | A1 |
20190377632 | Oh | Dec 2019 | A1 |
20190377821 | Pleshachkov | Dec 2019 | A1 |
20190391748 | Li | Dec 2019 | A1 |
20200004456 | Williams | Jan 2020 | A1 |
20200004674 | Williams | Jan 2020 | A1 |
20200013458 | Schreck | Jan 2020 | A1 |
20200042223 | Li | Feb 2020 | A1 |
20200042387 | Shani | Feb 2020 | A1 |
20200084918 | Shen | Mar 2020 | A1 |
20200089430 | Kanno | Mar 2020 | A1 |
20200097189 | Tao | Mar 2020 | A1 |
20200143885 | Kim | May 2020 | A1 |
20200159425 | Flynn | May 2020 | A1 |
20200167091 | Haridas | May 2020 | A1 |
20200225875 | Oh | Jul 2020 | A1 |
20200242021 | Gholamipour | Jul 2020 | A1 |
20200250032 | Goyal | Aug 2020 | A1 |
20200257598 | Yazovitsky | Aug 2020 | A1 |
20200326855 | Wu | Oct 2020 | A1 |
20200328192 | Zaman | Oct 2020 | A1 |
20200348888 | Kim | Nov 2020 | A1 |
20200387327 | Hsieh | Dec 2020 | A1 |
20200401334 | Saxena | Dec 2020 | A1 |
20200409791 | Devriendt | Dec 2020 | A1 |
20210010338 | Santos | Jan 2021 | A1 |
20210089392 | Shirakawa | Mar 2021 | A1 |
20210103388 | Choi | Apr 2021 | A1 |
20210274687 | Chen | Sep 2021 | A1 |
Number | Date | Country |
---|---|---|
2003022209 | Jan 2003 | JP |
2011175422 | Sep 2011 | JP |
9418634 | Aug 1994 | WO |
1994018634 | Aug 1994 | WO |
Entry |
---|
https://web.archive.org/web/20071130235034/http://en.wikipedia.org:80/wiki/logical_block_addressing wikipedia screen shot retriefed on wayback Nov. 20, 2007 showing both physical and logical addressing used historically to access data on storage devices (Year: 2007). |
Ivan Picoli, Carla Pasco, Bjorn Jonsson, Luc Bouganim, Philippe Bonnet. “uFLIP-OC: Understanding Flash I/O Patterns on Open-Channel Solid-State Drives.” APSys'17, Sep. 2017, Mumbai, India, pp. 1-7, 2017, <10.1145/3124680.3124741 >. <hal-01654985>. |
EMC Powerpath Load Balancing and Failover Comparison with native MPIO operating system solutions. Feb. 2011. |
Tsuchiya, Yoshihiro et al. “DBLK: Deduplication for Primary Block Storage”, MSST 2011, Denver, CO, May 23-27, 2011 pp. 1-5. |
Chen Feng, et al. “CAFTL: A Content-Aware Flash Translation Layer Enhancing the Lifespan of Flash Memory based Solid State Devices”< FAST'11, San Jose, CA Feb. 15-17, 2011, pp. 1-14. |
Wu, Huijun et al. “HPDedup: A Hybrid Prioritized Data Deduplication Mechanism for Primary Storage in the Cloud”, Cornell Univ. arXiv: 1702.08153v2[cs.DC], Apr. 16, 2017, pp. 1-14https://www.syncids.com/#. |
WOW: Wise Ordering for Writes—Combining Spatial and Temporal Locality in Non-Volatile Caches by Gill (Year: 2005). |
Helen H. W. Chan et al. “HashKV: Enabling Efficient Updated in KV Storage via Hashing”, https://www.usenix.org/conference/atc18/presentation/chan, (Year: 2018). |
S. Hong and D. Shin, “NAND Flash-Based Disk Cache Using SLC/MLC Combined Flash Memory,” 2010 International Workshop on Storage Network Architecture and Parallel I/Os, Incline Village, NV, 2010, pp. 21-30. |
Arpaci-Dusseau et al. “Operating Systems: Three Easy Pieces”, Originally published 2015; Pertinent: Chapter 44; flash-based SSDs, available at http://pages.cs.wisc.edu/˜remzi/OSTEP/. |
Jimenex, X., Novo, D. and P. Ienne, “Pheonix:Reviving MLC Blocks as SLC to Extend NAND Flash Devices Lifetime, ”Design, Automation & Text in Europe Conference & Exhibition (Date), 2013. |
Yang, T. Wu, H. and W. Sun, “GD-FTL: Improving the Performance and Lifetime of TLC SSD by Downgrading Worn-out Blocks,” IEEE 37th International Performance Computing and Communications Conference (IPCCC), 2018. |
C. Wu, D. Wu, H. Chou and C. Cheng, “Rethink the Design of Flash Translation Layers in a Component-Based View”, in IEEE Acess, vol. 5, pp. 12895-12912, 2017. |
Po-Liang Wu, Yuan-Hao Chang and T. Kuo, “A file-system-aware FTL design for flash-memory storage systems,” 2009, pp. 393-398. |
S. Choudhuri and T. Givargis, “Preformance improvement of block based NAND flash translation layer”, 2007 5th IEEE/ACM/IFIP International Conference on Hardware/Software Codesign and Systems Synthesis (CODES+ISSS). Saizburg, 2007, pp. 257-262. |
A. Zuck, O. Kishon and S. Toledo. “LSDM: Improving the Preformance of Mobile Storage with a Log-Structured Address Remapping Device Driver”, 2014 Eighth International Conference on Next Generation Mobile Apps, Services and Technologies, Oxford, 2014, pp. 221-228. |
J. Jung and Y. Won, “nvramdisk: A Transactional Block Device Driver for Non-Volatile RAM”, in IEEE Transactions on Computers, vol. 65, No. 2, pp. 589-600, Feb. 1, 2016. |
Te I et al. (Pensieve: a Machine Assisted SSD Layer for Extending the Lifetime: (Year: 2018). |
ARM (“Cortex-R5 and Cortex-R5F”, Technical reference Manual, Revision r1p1) (Year:2011). |
Number | Date | Country | |
---|---|---|---|
20210357346 A1 | Nov 2021 | US |