System and method for facilitating mitigation of read/write amplification in data compression

Information

  • Patent Grant
  • 11507499
  • Patent Number
    11,507,499
  • Date Filed
    Tuesday, May 19, 2020
    4 years ago
  • Date Issued
    Tuesday, November 22, 2022
    2 years ago
Abstract
The system can receive data to be written to a non-volatile memory in the distributed storage system. The received data can include a plurality of input segments. The system can assign consecutive logical block addresses (LBAs) to the plurality of input segments. The system can then compress the plurality of input segments to generate a plurality of fixed-length compressed segments, with each fixed-length compressed segment aligned with a physical block address (PBA) in a set of PBAs. The system compresses the plurality of input segments to enable an efficient use of storage capacity in the non-volatile memory. Next, the system can write the plurality of fixed-length compressed segments to a corresponding set of PBAs in the non-volatile memory. The system can then create, in a data structure, a set of entries which map the LBAs of the input segments to the set of PBAs. This data structure can be used later by the system when processing a read request including a LBA.
Description
BACKGROUND
Field

This disclosure is generally related to the field of data storage. More specifically, this disclosure is related to a system and method for facilitating mitigation of read/write amplification when performing data compression in a data storage system.


Related Art

The proliferation of the Internet and e-commerce continues to create a vast amount of digital content. Today, various distributed storage systems have been created to access and store the ever-increasing amount of digital content. However, network bandwidth and storage capacity of physical resources are two characteristics of distributed storage systems which can greatly impact their performance, cost, and efficiency.


Even with the addition of storage capacity to a distributed storage system, the physical bandwidth can still only support a limited number of users while meeting the requirements of a Service Level Agreement (SLA). For example, when a storage system experiences a heavy load of simultaneous incoming traffic, some drives may become non-responsive due to a lack of sufficient bandwidth, even if sufficient storage capacity is available.


Data compression techniques have been used in distributed storage systems to save storage capacity and to reduce the amount of data transferred, thus enabling the efficient use of storage capacity and communication bandwidth. However, efficiency of the compression techniques has become increasingly critical with the increase in amount of digital content. The existing data compression techniques are inefficient due to the overhead from high read and write amplifications inherent in their data processing operations. Therefore, some challenges still remain in designing an efficient data compression technique that is capable of providing an improved performance of storage systems with regards to latency and an improved efficiency with respect to resource consumption, network load, read/write amplification, etc.


SUMMARY

One embodiment of the present disclosure provides a system and method for facilitating data compression in a distributed storage system. During operation, the system can receive data to be written to a non-volatile memory in the distributed storage system. The received data can include a plurality of input segments. The system can assign consecutive logical block addresses (LBAs) to the plurality of input segments. The system can then compress the plurality of input segments, e.g., by applying a data compression technique, to generate a plurality of fixed-length compressed segments, with each fixed-length compressed segment aligned with a physical block address (PBA) in a set of PBAs. The system compresses the plurality of input segments to enable an efficient use of storage capacity in the non-volatile memory. Next, the system can write the plurality of fixed-length compressed segments to a corresponding set of PBAs in the non-volatile memory. The system can then create, in a data structure, a set of entries which map the LBAs of the input segments to the set of PBAs. This data structure can be used later by the system when processing a read request including a LBA.


In some embodiments, the system can compress the plurality of input segments by: reading sequentially a subset of the plurality of input segments into a sliding window, wherein the subset includes one or more of the input segments; incrementally compressing data in the sliding window until compressed data aligns with a PBA; in response to determining that the compressed data aligns with the PBA: identifying an offset and/or a length of data input corresponding to the compressed data; and writing the compressed data to the PBA in the non-volatile memory; and moving the sliding window consecutively along the plurality of input segments based on the offset and/or the length of the data input.


In some embodiments, the data structure can include: an index field which can include a LBA as an index of the data structure; a PBA field; and a cross-bit field which can indicate whether data in one LBA is written into one PBA or more than one PBA.


In some embodiments, the system can create the set of entries in the data structure by performing the following operations: determining that compressed data associated with a LBA is written into two consecutive PBAs; and setting a flag in a cross-bit field corresponding to the LBA in the data structure.


In some embodiments, the system can receive, from a client, a data read request including the LBA. The system can identify, in the data structure, one or more PBAs corresponding to the LBA. The system can then read compressed data from the one or more PBAs. Next, the system may decompress the compressed data to generate decompressed data. The system can then provide, to the client, requested data based on the decompressed data.


In some embodiments, the system can read the compressed data from the one or more PBAs by: reading the compressed data from one PBA when a flag in a cross-bit field corresponding to the LBA in the data structure is not set; and reading the compressed data from two or more consecutive PBAs when the flag in the cross-bit field corresponding to the LBA in the data structure is set.


In some embodiments, the system can apply erasure coding to the plurality of compressed segments prior to writing the compressed segments to a journal drive.


In some embodiments, the system can apply erasure coding to the plurality of compressed segments prior to writing the compressed segments to different storage nodes in a distributed storage system.





BRIEF DESCRIPTION OF THE FIGURES


FIG. 1 illustrates an exemplary Input/Output (I/O) amplification in a data compression scheme, in accordance with the prior art.



FIG. 2 illustrates an exemplary read amplification in a data compression scheme, in accordance with the prior art.



FIG. 3 illustrates an exemplary data compression scheme with fixed-length output, in accordance with an embodiment of the present disclosure.



FIG. 4 illustrates an exemplary example of a mapping table used in a data compression scheme, in accordance with an embodiment of the present disclosure.



FIG. 5 illustrates an exemplary data compression scheme for reducing read amplification, in accordance with an embodiment of the present disclosure.



FIG. 6A illustrates an exemplary system architecture, in accordance with the prior art.



FIG. 6B illustrates an exemplary modified system architecture, in accordance with an embodiment of the present disclosure.



FIG. 7A presents a flowchart illustrating a method for facilitating a data compression scheme, in accordance with an embodiment of the present disclosure.



FIG. 7B presents a flowchart illustrating a method for facilitating a data compression scheme, in accordance with an embodiment of the present disclosure.



FIG. 7C presents a flowchart illustrating a method for facilitating a data compression scheme to process a read request, in accordance with an embodiment of the present disclosure.



FIG. 8 illustrates an exemplary computer system that facilitates a data compression scheme, in accordance with an embodiment of the present disclosure.



FIG. 9 illustrates an exemplary apparatus that facilitates a data compression scheme, in accordance with an embodiment of the present disclosure.





In the figures, like reference numerals refer to the same figure elements.


DETAILED DESCRIPTION

The following description is presented to enable any person skilled in the art to make and use the embodiments, and is provided in the context of a particular application and its requirements. Various modifications to the disclosed embodiments will be readily apparent to those skilled in the art, and the general principles defined herein may be applied to other embodiments and applications without departing from the spirit and scope of the present disclosure. Thus, the embodiments described herein are not limited to the embodiments shown, but are to be accorded the widest scope consistent with the principles and features disclosed herein.


Overview

Data-intensive operations performed by the existing data compression techniques can result in an increase in the utilization of storage resources. Specifically, existing data compression schemes generate irregular-sized compression results. Aligning these irregular-sized compression results with the LBA can involve additional data-intensive operations resulting in increased read/write amplifications and suboptimal usage of the storage resources. Furthermore, due to the inefficient data compression, decompressing the compressed data can also result in increased read amplification. Therefore, the existing data compression schemes can increase the processing burden on the storage system, increase latency, and can result in wearing out of the storage media or can decrease the life span of the storage media. Such a data compression scheme is described below in relation to FIG. 1 and FIG. 2. A system architecture including a data compression scheme is described below in relation to FIG. 6A.


Embodiments described herein address the above-mentioned drawbacks associated with the existing data compression schemes. Specifically, a system can generate a plurality of PBA-aligned fixed-length compressed data segments from LBA-aligned input data segments by applying a sliding window to these input data segments to mitigate read and write amplification of the storage system. Furthermore, the manner in which the data compression scheme is incorporated within a system architecture of a storage cluster can be changed to result in a reduction in the amount of data transferred within the storage cluster, thereby saving the communication bandwidth. Therefore, by applying the data compression scheme described in the present disclosure, the system can reduce latency, save storage system resources, and enhance the efficiency of the distributed storage system. Such a novel data compression scheme is described below in relation to FIG. 3, FIG. 4, FIG. 5, and FIG. 7. A modified system architecture with the data compression scheme, a computer system, and an apparatus facilitating the data compression scheme are described below in relation to FIG. 6B, FIG. 8, and FIG. 9, respectively.


The term “distributed storage system” refers to a set of compute nodes (or client servers) interacting through a set of storage servers (or storage nodes) via a network, such as a data center network.


The term “storage cluster” refers to a group of storage servers.


The term “storage server” refers to a server in a distributed storage system. A storage server can have multiple drives, where data may be written on to a drive for persistent storage. A storage server can also include a journal associated with the journaling file system. A drive can also include a storage, storage medium, or other storage means associated with the drive.


Data Compression Scheme



FIG. 1 illustrates an exemplary Input/Output (I/O) amplification in a data compression scheme, in accordance with the prior art. A system implementing a data compression scheme 100, receives as input a group of files, e.g., File X 102 and File Y 104, to be compressed. Files 102 and 104 are concatenated and divided into fixed-length segments, e.g., 108-114. Specifically, File X 102 is mapped to segments 108 and 110, while File Y 104 is mapped to segments 110-114. Data compression scheme 100 can compress the segments 108-114 individually to generate variable-length compressed segments C1116, C2118, C3120, and C4122. Since compressed segments 116-122 are not aligned with LBAs, they are further subject to additional processing.


For example, to write compressed segment 116 into its corresponding LBA1162, the system implementing data compression scheme 100 can split the compressed segment C1116 into two portions C11124 and C12126. Note that C11124 aligns with LBA1162 and hence can be written into a LBA1162 in a storage drive. However, since length of C12126 does not align with LBA2164, i.e., length of C12126 is less than length of LBA2164, C12126 is padded with a certain bit pattern P 128 so that length of 126 is equal to length of LBA2164. Next, when C2118 is obtained, the system implementing data compression scheme 100 can reload LBA2164 from memory, i.e., {C12130,P 128} is reloaded, and the system can drop P 128 to concatenate with a portion of C2118. Specifically, the system implementing data compression scheme 100 can split C2118 into two portions, i.e., C21132 and C22134, in a manner that a combination of the first portion of C2, i.e., C21132, and a last portion of reloaded C1, i.e., C12130, aligns with LBA2164. Then the system can write {C12130,C21132} into the same LBA2164. Note that during the process of compressing data, the compressed data, i.e., {C12130,P 128}, is first written to LBA2164 only to be read out and re-written with {C12130,C21132}. These read and write operations can increase overhead of read/write amplification during the data-intensive compression.


Next, a second portion of C2, i.e., C22134, is left to be written into subsequent LBA3166. This can be done by padding a certain bit pattern P 136 to C22134 so that a combination of 134 and 136 aligns with LBA 3166. The system can then write C22134 and P 136 into LBA3166. When C3120 is received, the system can reload {C22138, P 136} and combine it with C31140 and can drop P 136 prior to writing the new combination into LBA 3166. The system can continue to perform the reloading of compressed portions from previous LBA and dropping of the padding in this previous LBA before combining the reloaded compressed portion with a consecutive compressed portion. In other words, the system implementing such data-intensive data compression scheme 100 can perform frequent writing, reading, and dropping operations that can result in suboptimal consumption of resources.


Such suboptimal consumption of the resources is also observed when processing a read request for a portion of a file that has been compressed and stored using data compression scheme 100. Specifically, when the system receives a read request to read a part of File Y 104, e.g., portion REQ 106 in File Y 104, the system implementing data compression scheme 100 can first identify where the compressed version of File Y 104 is stored in memory. Then the system may identify that File Y 104 containing the requested portion REQ 106 is included in segment 114 and the compressed version of segment 114 is in segment C4122. Therefore, data compression scheme reads the whole segment C4122 from LBA4-LBA6, e.g., LBA4168, LBA5170, and LBA6172. Note that C32146 in LBA4168 does not correspond to File Y 104 and hence the system can drop C32146 and padding P 154 in LBA6172 at a first stage of processing the read request (operation 156).


The system can then perform decompression of compressed segments: C41148, C42150, and C43152 (operation 158). The system may then select (operation 160) the requested data and send the requested data 106 to the requesting client or source while a remaining portion of segment 114 that does not include the requested data can be dropped at a second stage of processing the read request. Note that when processing a read request, irrelevant data, e.g., C32146, C41148, C42150, and P 154, are read and dropped resulting in read amplification. The system implementing data compression scheme 100 can result in I/O amplification which can increase the processing burden, latency, and can result in a reduced lifespan of the storage media.



FIG. 2 illustrates an exemplary read amplification in a data compression scheme, in accordance with the prior art. A system implementing a data compression scheme 200 can receive as input a group of files 202-212, e.g., File A 202, File B 204, File C 206, File D 208, and File E 208, to be compressed. In practice, in order to improve a compression efficiency, it is desirable that an input data be of a large size. Therefore, to improve the compression efficiency, the system can merge or concatenate the files 202-212 to form a large file or segment 214. The system implementing data compression scheme 200 can then apply data compression to segment 214 with a high compression ratio to generate a compressed segment 216. Compressed segment 216 can then be divided to align with LBA1218, LBA2220, and LBAx 222.


However, the problems associated with such data compression become evident when processing a read request for a specific file. For example, when a request for data (REQ 210) in File D 208 is received, the system implementing data compression scheme 200 may have to read the entire compressed segment 216 to perform decompression 224. During the process of reading the requested data, the system can drop 226 a large amount of decompressed data before providing the requested data to a requesting client. Therefore, in the traditional data compression scheme 200, the entire segment 214 is read and decompressed irrespective of the size of requested data 210, e.g., size of segment 214 could be much larger than the size of requested data 210. Since a large amount of decompressed data is read and then dropped 226, the resources used for the read and decompression operations can be used inefficiently.


Embodiments disclosed herein describe an efficient data compression scheme that is capable of overcoming the disadvantages of existing data compression schemes described in relation to FIG. 1 and FIG. 2. Specifically, the data compression scheme described in the present disclosure is capable of mitigating the read and write amplifications in the storage system. Furthermore, the compression and decompression engines described in the present disclosure can be incorporated into the storage system architecture in a way that can reduce the amount of data transferred and reduce the bandwidth consumption. In the following paragraphs the implementation details of a novel data compression scheme in accordance with the present disclosure are addressed.



FIG. 3 illustrates an exemplary data compression scheme with fixed-length output, in accordance with an embodiment of the present disclosure. During operation, a data compression system implementing data compression scheme 300 can receive data to be compressed. For example, the received data could be represented as equal length data portions, i.e., A 302, B 304, C 306, D 308, E 310, F 312, G 314, H 316, I 318, and J 320. Each data portion in the group of data portions A 302-J 320 can be aligned with an LBA in LBA range 344. The system may continuously and incrementally perform compression on the received data portions until the compressed output of the system is aligned with a PBA in a PBA range 346. Data compression system implementing data compression scheme 300 can ensure that the compressed output is of fixed-length and aligned with PBA.


For example, the data compression system may identify that the data portions A 302, B 304, C 306, and a part of D 308 when compressed may result in output P1338 that aligns with a PBA. The system may use these input data portions as one input data chunk and mark them with an offset and length, e.g., {O1, L1} 326. Note that this input data chunk 326 may not align with LBA. For example, input data chunk 326 may end in the middle of a LBA, e.g., in the middle of D 308. The remaining part of data in the LBA, i.e., remaining part of data in D 308, can be grouped into a later input data chunk 328, i.e., {O2, L2} 328, to generate compressed output P2340 that is aligned with a PBA. Similarly, data portions G 314, H 316, I 318, and a part of J 320, i.e., J1322, can be grouped to form an input data chunk 330 with offset O3 and length L3. The system can then compress input data chunk 330 to generate a PBA-aligned compressed output P3342.


In one embodiment of the present disclosure, the LBAs associated with each of data portions 302-306 and part of 308 can be mapped to one PBA P1338. Similarly, LBAs of data portions including remaining part of D 308, E 310, and F 312 can be mapped to PBA P2340. The LBAs of data portions G 314, H 316, I 318, and part of J 320 (i.e., J1322) can be mapped to PBA P3342. Note that the system does not apply any data padding to the data portions 302-322 during the data compression process, except at the end of the incoming data J2324 where data padding can be applied. Since the system stores the compressed data in memory without performing any additional processing on the input data chunks 326, 328, and 330, e.g., reloading and dropping of data, the system implementing data compression scheme 300 can facilitate the mitigation of read and write amplification. The manner in which a mapping 348 between LBAs in LBA range 344 and PBAs in PBA range 346 is built is described below in relation to FIG. 4.



FIG. 4 illustrates an exemplary example of a mapping table used in a data compression scheme, in accordance with an embodiment of the present disclosure. FIG. 4 shows an implementation 400 of a mapping table which can include mappings between LBAs and PBAs. Mapping table 442 may include the following fields: a LBA 402 field, a PBA 404 field, and a cross-bit 406 field. Mapping table 442 may use LBA as an index of the table to map with a PBA. Cross-bit field 406 may indicate whether compressed data crosses a PBA boundary. In other words, cross-bit field 406 indicates whether only one PBA is to be read or two consecutive PBAs are to be read when processing a read request associated with a LBA.


The data compression system may build mapping table 442 based on the data portions used for performing data compression and the corresponding compressed data portions, e.g., mapping 428. For example, the data compression system may generate compressed data Cx 414 after applying data compression to a data portion LBAx 412. Note that compressed data Cx 414 aligns with PBAi 416 and can be stored in a non-volatile memory. The system may then include an entry 408 in mapping table 442 to denote a mapping between LBAx 412 and PBAi 416. In entry 408 of mapping table 442, cross-bit field 406 is set to “0” to indicate that LBAx 412 after compression generates Cx 414 which aligns with just one PBAi 416. However, when a consecutive data portion aligned with LBAy 418 is compressed, the compressed data portion Cy 422 can be written into both PBAi 424 and PBA i+1 426. The system can include mapping 430 into entry 410 in mapping table 442. Note that since Cy 422 is written into both PBAi 424 and PBAi+1 426, cross-bit field 406 corresponding to entry 410 in mapping table 442 can be set to “1.” With cross-bit field 406 set to “1” the system may have to read both PBAi and PBAi+1 when processing a read request for LBAy.


For example, when the system receives a read request for data associated with LBAy 418, the system may first look-up mapping table 442 to identify entry 410 for LBAy. Note that the LBA itself can also be saved in memory along with its data. Since cross-bit field in entry 410 is set to “1” the system may read out both PBAi and PBA i+1 from memory. The system may then first scan PBAi to find a data block 440 including a header 432 and endian 438 associated with LBAx 434 and Cx 436. Header 432 can mark a start position of a space where information associated with LBAx 434 and Cx 436 are stored, and endian 438 can mark an end position of this space. Next, the system may scan PBAi+1 to find another header and endian associated with LBAy and Cy. The system may identify that PBAi+1 includes the compressed form of the requested data LBAy. The system may then only decompress Cy and send the decompressed data associated with LBAy to a requesting client.



FIG. 5 illustrates an exemplary data compression scheme 500 for reducing read amplification, in accordance with an embodiment of the present disclosure. During operation, the data compression system implementing data compression scheme 500 can receive a plurality of files to be compressed. To generate fixed-length compressed outputs, the system may apply a sliding window to the received files, e.g., File A 502, File B 504, File C 506, File D 508, File E 510, and File F 512. The system can move the sliding window consecutively from left to right along the received files. When compressed data for a specific data portion within the sliding window aligns or reaches the size of a PBA, the system may temporarily halt the movement of the sliding window. The system may then mark the specific data portion within the sliding window with an offset/length and truncate this portion from the received files to represent a data input to a data compression engine. The system may then resume the movement of the sliding window from the end of the specific data portion marked with offset/length and can continue to seek data to be compressed into one or more aligned PBAs. The system can continue to move the sliding window to the end of the received files for generating PBA-aligned compressed data.


For example, sliding window 514 can include entire File A 502 and a portion of File B 504. The data compression system may start compressing in increments data within sliding window 514. In other words, instead of compressing all the data available within sliding window 514, the system may compress incremental amounts of data until a total length of compressed output aligns with a PBA1544. When the system determines that the compressed output length aligns with a PBA, the system may truncate a data chunk 516 from the files included within sliding window 514. The system may mark data chunk 516 with a length and an offset pair {O1, L1}. The system may then use data chunk 516 as an input to a compression engine 534 for generating compressed output that is aligned with PBA1544.


The system can then move sliding window 514 to a position where data chunk 516 ends. This new position of sliding window 514 is indicated by dashed block 518 in FIG. 5. The system may continue to move sliding window 514 along the set of received files 502-512 until the end of the received files is reached. This movement of sliding window 514 is denoted by blocks 518, 522, 526, and 530. As the sliding window moves consecutively along the received files 504-512, the system can provide different data chunks to a compression engine to generate multiple PBA-aligned compressed data.


Specifically, the system may apply sliding window 518 to truncate data chunk {O2,L2} 520 from the received files and apply compression 536 to data chunk 520 to obtain compressed data aligned with PBA2546. Similarly, the system may use the different positions 522, 526, and 530 of sliding window along the received files to identify data chunks {O3,L3} 524, {O4,L4} 528, and {O5,L5} 532, respectively. The system may apply compression 538, 540, and 542 to data chunks 524, 528, and 532, respectively. The output of compression 538, 540, and 542 can include fixed-length PBA-aligned compressed data, i.e., PBA3548, PBA4550, and PBA5540, respectively.


When the system receives a read request, e.g., a request to read a portion of data REQ 554 from File D 508, the system may process the read request for REQ 554 by first locating a corresponding PBA containing compressed data. For example, the system may use LBA associated with the read request to look-up a mapping table (shown in FIG. 4) to identify a PBA corresponding to the LBA in the read request. Specifically, the system may identify PBA4550 to be storing the compressed data associated with the LBA in the read request. The system may apply a decompression engine 554 to decompress compressed data in PBA4550 to obtain decompressed data chunk {O4, L4} 528. Next, the system may read requested data REQ 554 from data chunk 528 and provide data REQ 554 to a client that sent the read request.


Data compression schemes shown in FIG. 1 and FIG. 2, apply data compression to fixed-length input data to generate variable-length compressed outputs resulting in an increased read/write amplification during the data-intensive compression process. Further, when processing a read request, the existing data processing schemes can deliver a suboptimal performance because large compressed data chunks are decompressed and a significant portion of the decompressed data are dropped before actually sending the requested data. Such a suboptimal performance of the existing data compression schemes can result in high read/write amplification, high latency, and inefficient use of communication bandwidth.


In the exemplary embodiments shown in FIG. 5, the system can generate fixed-length PBA-aligned compressed data from varied-length input data chunks. Further, the system can apply a sliding window to the received files to mitigate I/O amplification during the data compression process. Moreover, when processing a read request the data compression system described in the present disclosure is capable of satisfying a small-sized read by reading just one or two PBAs, thereby reducing read amplification and the amount of data dropped when processing the read request. The reduction in read/write amplification can have an improved performance impact on the storage system in terms of latency, the amount of data transferred, and communication bandwidth consumption.


System Architecture



FIG. 6A illustrates an exemplary system architecture 600, in accordance with the prior art. In a distributed storage system, data compression can be applied at multiple places for generating different data formats. For example, system memory 602 can store original data 604 to be compressed, and original data 604 can be divided into data chunks 606-610 prior to applying compression. Specifically, compression 612 is applied to original data 604 to generate compressed original data. The system can then apply erasure coding (EC) 624 to the compressed original data prior to storing them in multiple journal drives 614. The system can also apply compression 616-620 to data chunks 606-610, respectively to generate compressed data chunks. These compressed data chunks are then subject to EC 626-630 prior to being distributed and stored in multiple storage drives 622. Note that in system architecture 600 since compression engines 612, 616-620 are placed after system memory 602, the system may require considerable amount of memory to store uncompressed original data 604. Further, uncompressed original data 604 and data chunks 606-610 are sent towards journal drives 614 and storage drives 622, which may result in a high consumption of network bandwidth.



FIG. 6B illustrates an exemplary modified system architecture 640, in accordance with an embodiment of the present disclosure. System architecture 640 can integrate data compression and decompression engines 646 at the data input in network interface card (NIC) 644. The data received via NIC 644 is compressed at 646 before transferring 648 and storing compressed data 650 in system memory 642. Compressed data 650 can then be treated as original data of the storage cluster. For example, in system memory 642 compressed data 650 can be divided into data chunks 652-656. The system may then apply erasure coding (EC) 658 to compressed data 650 prior to storing them in multiple journal drives 666. The system may also apply EC 660-664 to compressed data chunks 652-656 prior to distributing and storing them in multiple storage drives 668. Note that in system architecture 640, since compression and decompression engines 646 are placed at the entrance of the storage cluster, the amount of data transfer within system architecture 640 can be reduced. Therefore, an overall burden on the data transfer and data processing operations can be mitigated.


Exemplary Method for Facilitating a Data Compression Scheme



FIG. 7A presents a flowchart 700 illustrating a method for facilitating a data compression scheme, in accordance with an embodiment of the present disclosure. During operation, the system can receive data to be written to a non-volatile memory in the storage system (operation 702). The received data can include a plurality of input segments. The system can assign consecutive logical block addresses (LBAs) to the plurality of input segments (operation 704). The system can then compress the plurality of input segments to generate a plurality of fixed-length compressed segments, with each compressed segment aligned with a physical block address (PBA) in a set of PBAs. Specifically, the system can first sequentially read a subset of the plurality of input segments into a sliding window (operation 706). Next, the system can compress in increments data within the sliding window (operation 708). The system may determine whether the compressed data aligns with a PBA (operation 710).


In response to the system determining that the compressed data does not align with the PBA, the system may continue to incrementally compress data in the sliding window (e.g., by applying operations 708 and 710 to incrementally compress data). In response to the system determining that the compressed data aligns with the PBA, i.e., a length of compressed data is equal to a length of the PBA, the system can continue operation at Label A of FIG. 7B.



FIG. 7B presents a flowchart 720 illustrating a method for facilitating a data compression scheme, in accordance with an embodiment of the present disclosure. During operation, the system may write the compressed data to the PBA in the non-volatile memory (operation 722). The system may identify and save an offset and/or a length of data input corresponding to the compressed data (operation 724). The system can then create, in a data structure, an entry that maps the LBAs of the input segments used for generating the compressed data to one or more PBAs (operation 726). The system can then determine whether the compressed output associated with the LBA is written to at least two consecutive PBAs (operation 728). If the condition in operation 728 is true, then the system can set a flag in a cross-bit field in the data structure entry (operation 730). If the condition in operation 728 is false, then the system may continue to operation 732.


The system can proceed to move the sliding window consecutively along the plurality of input segments based on the offset and/or the length of the data input (operation 732). In other words, the system may move the sliding window to start from a position where the previous data input ended. Next, the system can determine whether the sliding window has reached the end of the plurality of input segments in the received data (operation 734). In response to the system determining that the condition in operation 734 is satisfied, the operation returns. In response to the system determining that the condition in operation 734 is not satisfied, the operation continues to operation 706 of FIG. 7A (e.g., by applying operations 706, 708, and 710 to the remaining plurality of input segments that are yet to be compressed).



FIG. 7C presents a flowchart 750 illustrating a method for facilitating a data compression scheme to process a read request, in accordance with an embodiment of the present disclosure. During operation, the system may receive, from a client, a data read request including an LBA (operation 752). The system can then identify in a data structure an entry corresponding to the LBA (operation 754). The system can then determine whether a flag in a cross-bit field in the data structure entry is set (operation 756). When the flag is set, the system may read compressed data from two or more consecutive PBAs (operation 758). When the flag is not set, then the system can read the compressed data from just one PBA (operation 760). Next, the system can decompress the compressed data to generate decompressed data (operation 762). The system can then provide to the requesting client the requested data based on the decompressed data (operation 764) and the operation returns.


Exemplary Computer System and Apparatus



FIG. 8 illustrates an exemplary computer system that facilitates a data compression scheme, in accordance with an embodiment of the present disclosure. Computer system 800 includes a processor 802, a memory 804, and a storage device 806. Computer system 800 can be coupled to a plurality of peripheral input/output devices 832, e.g., a display device 810, a keyboard 812, and a pointing device 814, and can also be coupled via one or more network interfaces to network 808. Storage device 806 can store an operating system 818 and a content processing system 820.


In one embodiment, content processing system 820 can include instructions, which when executed by processor 802 can cause computer system 800 to perform methods and/or processes described in this disclosure. During operation of computer system 800, content processing system 820 can include instructions for receiving a set of files for performing data compression (communication module 822). Content processing system 820 may further include instructions for applying a sliding window to the set of received files (sliding window module 824). Sliding window module 824 can move the sliding window consecutively along the set of received files until the end of the set of received files is reached. At each position of the sliding window, sliding window module 824 may truncate only that portion of data within the sliding window which would produce a fixed-length PBA-aligned compressed data.


Content processing system 820 can include instructions for compressing data within the sliding window to generate PBA-aligned compressed data (data compression module 826). Content processing system 820 can include instructions for storing the PBA-aligned compressed data in a PBA of a non-volatile memory (data storing module 828). Content processing system 820 can further include instructions to build a mapping table that includes a set of mappings between a set of LBAs of plurality of input segments associated with the received files to one or more PBAs of the corresponding compressed data (mapping module 830).


Content processing system 820 can further include instructions for processing a read request from a client (communication module 822). Specifically, the read request can be processed by first looking-up the mapping table to identify a LBA that was included in the read request. After the LBA has been identified in the mapping table, corresponding PBAs can be determined from the mapping table based on a flag setting in a cross-bit field. Decompression of compressed data is performed on the one or more PBAs identified in the mapping table to generate decompressed data (data decompression module 832). Content processing system 820 can include instructions to send the requested data to the client based on the decompressed data (communication module 822).



FIG. 9 illustrates an exemplary apparatus that facilitates a data compression scheme, according to one embodiment of the present disclosure. Apparatus 900 can include a plurality of units or apparatuses that may communicate with one another via a wired, wireless, quantum light, or electrical communication channel. Apparatus 900 may be realized using one or more integrated circuits, and may include fewer or more units or apparatuses than those shown in FIG. 8. Further, apparatus 900 may be integrated in a computer system, or realized as a separate device that is capable of communicating with other computer systems and/or devices. Specifically, apparatus 900 can include units 902-912, which perform functions or operations similar to modules 822-832 of computer system 800 in FIG. 8. Apparatus 900 can include: a communication unit 902, a sliding window unit 904, a data compression unit 906, a data storing unit 908, a mapping unit 910, and a data decompression unit 912.


The methods and processes described in the detailed description section can be embodied as code and/or data, which can be stored in a computer-readable storage medium as described above. When a computer system reads and executes the code and/or data stored on the computer-readable storage medium, the computer system performs the methods and processes embodied as data structures and code and stored within the computer-readable storage medium.


The data structures and code described in this detailed description are typically stored on a computer-readable storage medium, which may be any device or medium that can store code and/or data for use by a computer system. The computer-readable storage medium includes, but is not limited to, volatile memory, non-volatile memory, magnetic and optical storage devices such as disk drives, magnetic tape, CDs (compact discs), DVDs (digital versatile discs or digital video discs), or other media capable of storing computer-readable media now known or later developed.


Furthermore, the methods and processes described above can be included in hardware modules or apparatus. The hardware modules or apparatus can include, but are not limited to, application-specific integrated circuit (ASIC) chips, field-programmable gate arrays (FPGAs), dedicated or shared processors that execute a particular software module or a piece of code at a particular time, and other programmable-logic devices now known or later developed. When the hardware modules or apparatus are activated, they perform the methods and processes included within them.


The foregoing descriptions of embodiments of the present disclosure have been presented for purposes of illustration and description only. They are not intended to be exhaustive or to limit the present disclosure to the forms disclosed. Accordingly, many modifications and variations will be apparent to practitioners skilled in the art. Additionally, the above disclosure is not intended to limit the present disclosure. The scope of the present disclosure is defined by the appended claims.

Claims
  • 1. A computer-implemented method, comprising: receiving data to be written to a non-volatile memory, wherein the data includes a plurality of input segments, which are assigned with consecutive logical block addresses (LBAs);compressing the plurality of input segments to generate a plurality of fixed-length compressed segments, with each fixed-length compressed segment aligned one or more physical block addresses (PBAs) in a set of PBAs, wherein compressing the plurality of input segments comprises: reading sequentially a subset of the plurality of input segments into a sliding window, wherein the subset includes one or more of the input segments;incrementally compressing data in the sliding window until compressed data length satisfies a fixed length associated with a PBA;in response to determining that the compressed data length satisfies the fixed length: identifying an offset and/or a length of data input corresponding to the compressed data; andwriting the compressed data to the PBA in the non-volatile memory; andmoving the sliding window consecutively along the plurality of input segments based on the offset and/or the length of the data input;writing a respective compressed segment into one or more memory blocks addressed with respective PBAs in the non-volatile memory; andcreating, in a data structure, a set of entries which map the LBAs of the input segments to the set of PBAs, wherein creating the set of entries comprises: determining that the compressed segment associated with a LBA and stored in the non-volatile memory crosses a boundary between the two consecutive memory blocks addressed with respective PBAs; andsetting a field corresponding to the LBA in the data structure.
  • 2. The method of claim 1, wherein the data structure includes: an index field which includes a LBA as an index of the data structure;a PBA field; anda cross-bit field which indicates whether data in one LBA is written into one PBA or more than one PBA.
  • 3. The method of claim 1, further comprising: receiving, from a client, a data read request including the LBA;identifying, in the data structure, one or more PBAs corresponding to the LBA;reading compressed data from the one or more PBAs;decompressing the compressed data to generate decompressed data; andproviding, to the client, requested data based on the decompressed data.
  • 4. The method of claim 3, wherein reading the compressed data from the one or more PBAs further comprises: reading the compressed data from one PBA when a flag in a cross-bit field corresponding to the LBA in the data structure is not set; andreading the compressed data from two or more consecutive PBAs when the flag in the cross-bit field corresponding to the LBA in the data structure is set.
  • 5. The method of claim 1, further comprising: applying erasure coding to the plurality of compressed segments prior to writing the compressed segments to a journal drive.
  • 6. The method of claim 1, further comprising: applying erasure coding to the plurality of compressed segments prior to writing the compressed segments to different storage nodes in a distributed storage system.
  • 7. The method of claim 1, wherein compressing the plurality of input segments further comprises: continuously performing the incrementally compressing the data in the sliding window until the compressed data length satisfied the fixed length associated with the PBA.
  • 8. The method of claim 1, wherein compressing the plurality of input segments is performed by a compression/decompression engine associated with a network interface card while receiving the data to be written to the non-volatile memory, wherein the compression/decompression engine resides at an entrance to different storage nodes in a distributed storage system.
  • 9. A computer system, comprising: a co-processor; anda storage device coupled to the processor and storing instructions, which when executed by the co-processor cause the co-processor to perform a method, the method comprising: receiving data to be written to a non-volatile memory, wherein the data includes a plurality of input segments, which are assigned with consecutive logical block addresses (LBAs);compressing the plurality of input segments to generate a plurality of fixed-length compressed segments, with each fixed-length compressed segment aligned with a one or more physical block addresses (PBAs) in a set of PBAs, wherein compressing the plurality of input segments comprises: reading sequentially a subset of the plurality of input segments into a sliding window, wherein the subset includes one or more of the input segments;incrementally compressing data in the sliding window until compressed data length satisfies a fixed length associated with a PBA;in response to determining that the compressed data length satisfies the fixed length: identifying an offset and/or a length of data input corresponding to the compressed data; andwriting the compressed data to the PBA in the non-volatile memory; andmoving the sliding window consecutively along the plurality of input segments based on the offset and/or the length of the data input;writing a respective compressed segment into one or more memory blocks addressed with respective PBAs in the non-volatile memory; andcreating, in a data structure, a set of entries which map the LBAs of the input segments to the set of PBAs, wherein creating the set of entries comprises: determining that the compressed segment associated with a LBA and stored in the non-volatile memory crosses a boundary between the two consecutive memory blocks addressed with respective PBAs; andsetting a field corresponding to the LBA in the data structure.
  • 10. The computer system of claim 9, wherein the data structure includes: an index field which includes a LBA as an index of the data structure;a PBA field; anda cross-bit field which indicates whether data in one LBA is written into one PBA or more than one PBA.
  • 11. The computer system of claim 9, wherein the method further comprises: receiving, from a client, a data read request including the LBA;identifying, in the data structure, one or more PBAs corresponding to the LBA;reading compressed data from the one or more PBAs;decompressing the compressed data to generate decompressed data; andproviding, to the client, requested data based on the decompressed data.
  • 12. The computer system of claim 11, wherein reading the compressed data from the one or more PBAs further comprises: reading the compressed data from one PBA when a flag in a cross-bit field corresponding to the LBA in the data structure is not set; andreading the compressed data from two or more consecutive PBAs when the flag in the cross-bit field corresponding to the LBA in the data structure is set.
  • 13. The computer system of claim 9, wherein the method further comprises: applying erasure coding to the plurality of compressed segments prior to writing the compressed segments to a journal drive.
  • 14. The computer system of claim 9, wherein the method further comprises: applying erasure coding to the plurality of compressed segments prior to writing the compressed segments to different storage nodes in a distributed storage system.
  • 15. The computer system of claim 9, wherein compressing the plurality of input segments is performed by a compression/decompression engine associated with a network interface card while receiving the data to be written to the non-volatile memory, wherein the compression/decompression engine resides at an entrance to different storage nodes in a distributed storage system.
  • 16. An apparatus, comprising: a co-processor; anda storage medium storing instructions, which when executed by the co-processor cause the co-processor to perform a method, the method comprising: receiving data to be written to a non-volatile memory, wherein the data includes a plurality of input segments, which are assigned with consecutive logical block addresses (LBAs);compressing the plurality of input segments to generate a plurality of fixed-length compressed segments, with each fixed-length compressed segment aligned with one or more physical block address (PBA) in a set of PBAs, wherein compressing the plurality of input segments comprises: reading sequentially a subset of the plurality of input segments into a sliding window, wherein the subset includes one or more of the input segments;incrementally compressing data in the sliding window until compressed data length satisfies a fixed length associated with a PBA;in response to determining that the compressed data length satisfies the fixed length: identifying an offset and/or a length of data input corresponding to the compressed data; andwriting the compressed data to the PBA in the non-volatile memory; andmoving the sliding window consecutively along the plurality of input segments based on the offset and/or the length of the data input;writing a respective compressed segment to one or more memory blocks addressed with respective PBAs in the non-volatile memory; andcreating, in a data structure, a set of entries which map the LBAs of the input segments to the set of PBAs, wherein creating the set of entries comprises: determining that the compressed segment associated with a LBA and stored in the non-volatile memory crosses a boundary between the two consecutive memory blocks addressed with respective PBAs; andsetting a field corresponding to the LBA in the data structure.
  • 17. The apparatus of claim 16, wherein the data structure includes: an index field which includes a LBA as an index of the data structure;a PBA field; anda cross-bit field which indicates whether data in one LBA is written into one PBA or more than one PBA.
  • 18. The apparatus of claim 16, wherein the method further comprises: receiving, from a client, a data read request including the LBA;identifying, in the data structure, one or more PBAs corresponding to the LBA;reading compressed data from the one or more PBAs;decompressing the compressed data to generate decompressed data; andproviding, to the client, requested data based on the decompressed data.
  • 19. The apparatus of claim 18, wherein reading the compressed data from the one or more PBAs further comprises: reading the compressed data from one PBA when a flag in a cross-bit field corresponding to the LBA in the data structure is not set; andreading the compressed data from two or more consecutive PBAs when the flag in the cross-bit field corresponding to the LBA in the data structure is set.
  • 20. The apparatus of claim 16, wherein the method further comprises: applying erasure coding to the plurality of compressed segments prior to writing the compressed segments at least one of: a journal drive; anddifferent storage nodes in a distributed storage system.
US Referenced Citations (495)
Number Name Date Kind
3893071 Bossen Jul 1975 A
4562494 Bond Dec 1985 A
4718067 Peters Jan 1988 A
4775932 Oxley Oct 1988 A
4858040 Hazebrouck Aug 1989 A
5394382 Hu Feb 1995 A
5602693 Brunnett Feb 1997 A
5715471 Otsuka Feb 1998 A
5732093 Huang Mar 1998 A
5802551 Komatsu Sep 1998 A
5930167 Lee Jul 1999 A
6098185 Wilson Aug 2000 A
6148377 Carter Nov 2000 A
6226650 Mahajan et al. May 2001 B1
6243795 Yang Jun 2001 B1
6457104 Tremaine Sep 2002 B1
6658478 Singhal Dec 2003 B1
6795894 Neufeld Sep 2004 B1
7351072 Muff Apr 2008 B2
7565454 Zuberi Jul 2009 B2
7599139 Bombet Oct 2009 B1
7953899 Hooper May 2011 B1
7958433 Yoon Jun 2011 B1
8024719 Gorton, Jr. Sep 2011 B2
8085569 Kim Dec 2011 B2
8144512 Huang Mar 2012 B2
8166233 Schibilla Apr 2012 B2
8260924 Koretz Sep 2012 B2
8281061 Radke Oct 2012 B2
8452819 Sorenson, III May 2013 B1
8516284 Chan Aug 2013 B2
8527544 Colgrove Sep 2013 B1
8751763 Ramarao Jun 2014 B1
8819367 Fallone Aug 2014 B1
8825937 Atkisson Sep 2014 B2
8832688 Tang Sep 2014 B2
8868825 Hayes Oct 2014 B1
8904061 O'Brien, III Dec 2014 B1
8949208 Xu Feb 2015 B1
9015561 Hu Apr 2015 B1
9031296 Kaempfer May 2015 B2
9043545 Kimmel May 2015 B2
9088300 Chen Jul 2015 B1
9092223 Pani Jul 2015 B1
9129628 Fallone Sep 2015 B1
9141176 Chen Sep 2015 B1
9208817 Li Dec 2015 B1
9213627 Van Acht Dec 2015 B2
9213632 Song Dec 2015 B1
9251058 Nellans Feb 2016 B2
9258014 Anderson Feb 2016 B2
9280472 Dang Mar 2016 B1
9280487 Candelaria Mar 2016 B2
9311939 Malina Apr 2016 B1
9336340 Dong May 2016 B1
9436595 Benitez Sep 2016 B1
9495263 Pang Nov 2016 B2
9529601 Dharmadhikari Dec 2016 B1
9529670 O'Connor Dec 2016 B2
9569454 Ebsen Feb 2017 B2
9575982 Sankara Subramanian Feb 2017 B1
9588698 Karamcheti Mar 2017 B1
9588977 Wang Mar 2017 B1
9607631 Rausch Mar 2017 B2
9671971 Trika Jun 2017 B2
9722632 Anderson Aug 2017 B2
9747202 Shaharabany Aug 2017 B1
9830084 Thakkar Nov 2017 B2
9836232 Vasquez Dec 2017 B1
9852076 Garg Dec 2017 B1
9875053 Frid Jan 2018 B2
9910705 Mak Mar 2018 B1
9912530 Singatwaria Mar 2018 B2
9923562 Vinson Mar 2018 B1
9933973 Luby Apr 2018 B2
9946596 Hashimoto Apr 2018 B2
10013169 Fisher Jul 2018 B2
10199066 Feldman Feb 2019 B1
10229735 Natarajan Mar 2019 B1
10235198 Qiu Mar 2019 B2
10268390 Warfield Apr 2019 B2
10318467 Barzik Jun 2019 B2
10361722 Lee Jul 2019 B2
10417086 Lin Sep 2019 B2
10437670 Koltsidas Oct 2019 B1
10459663 Agombar Oct 2019 B2
10459794 Baek Oct 2019 B2
10466907 Gole Nov 2019 B2
10484019 Weinberg Nov 2019 B2
10530391 Galbraith Jan 2020 B2
10635529 Bolkhovitin Apr 2020 B2
10642522 Li May 2020 B2
10649657 Zaidman May 2020 B2
10649969 De May 2020 B2
10678432 Dreier Jun 2020 B1
10756816 Dreier Aug 2020 B1
10831734 Li Nov 2020 B2
10928847 Suresh Feb 2021 B2
10990526 Lam Apr 2021 B1
11016932 Qiu May 2021 B2
11023150 Pletka Jun 2021 B2
11068165 Sharon Jul 2021 B2
11068409 Li Jul 2021 B2
11126561 Li Sep 2021 B2
11138124 Tomic Oct 2021 B2
11243694 Liang Feb 2022 B2
11360863 Varadan Jun 2022 B2
20010003205 Gilbert Jun 2001 A1
20010032324 Slaughter Oct 2001 A1
20010046295 Sako Nov 2001 A1
20020010783 Primak Jan 2002 A1
20020039260 Kilmer Apr 2002 A1
20020073358 Atkinson Jun 2002 A1
20020095403 Chandrasekaran Jul 2002 A1
20020112085 Berg Aug 2002 A1
20020161890 Chen Oct 2002 A1
20030074319 Jaquette Apr 2003 A1
20030145274 Hwang Jul 2003 A1
20030163594 Aasheim Aug 2003 A1
20030163633 Aasheim Aug 2003 A1
20030217080 White Nov 2003 A1
20040010545 Pandya Jan 2004 A1
20040066741 Dinker Apr 2004 A1
20040103238 Avraham May 2004 A1
20040143718 Chen Jul 2004 A1
20040255171 Zimmer Dec 2004 A1
20040267752 Wong Dec 2004 A1
20040268278 Hoberman Dec 2004 A1
20050038954 Saliba Feb 2005 A1
20050097126 Cabrera May 2005 A1
20050138325 Hofstee Jun 2005 A1
20050144358 Conley Jun 2005 A1
20050149827 Lambert Jul 2005 A1
20050174670 Dunn Aug 2005 A1
20050177672 Rao Aug 2005 A1
20050177755 Fung Aug 2005 A1
20050195635 Conley Sep 2005 A1
20050235067 Creta Oct 2005 A1
20050235171 Igari Oct 2005 A1
20060031709 Hiraiwa Feb 2006 A1
20060101197 Georgis May 2006 A1
20060156009 Shin Jul 2006 A1
20060156012 Beeson Jul 2006 A1
20060184813 Bui Aug 2006 A1
20070033323 Gorobets Feb 2007 A1
20070061502 Lasser Mar 2007 A1
20070101096 Gorobets May 2007 A1
20070204128 Lee Aug 2007 A1
20070250756 Gower Oct 2007 A1
20070266011 Rohrs Nov 2007 A1
20070283081 Lasser Dec 2007 A1
20070283104 Wellwood Dec 2007 A1
20070285980 Shimizu Dec 2007 A1
20080028223 Rhoads Jan 2008 A1
20080034154 Lee Feb 2008 A1
20080065805 Wu Mar 2008 A1
20080082731 Karamcheti Apr 2008 A1
20080104369 Reed May 2008 A1
20080112238 Kim May 2008 A1
20080163033 Yim Jul 2008 A1
20080195829 Wilsey Aug 2008 A1
20080301532 Uchikawa Dec 2008 A1
20090006667 Lin Jan 2009 A1
20090089544 Liu Apr 2009 A1
20090110078 Crinon Apr 2009 A1
20090113219 Aharonov Apr 2009 A1
20090125788 Wheeler May 2009 A1
20090177944 Kanno Jul 2009 A1
20090183052 Kanno Jul 2009 A1
20090254705 Abali Oct 2009 A1
20090282275 Yermalayeu Nov 2009 A1
20090287956 Flynn Nov 2009 A1
20090307249 Koifman Dec 2009 A1
20090307426 Galloway Dec 2009 A1
20090310412 Jang Dec 2009 A1
20100031000 Flynn Feb 2010 A1
20100169470 Takashige Jul 2010 A1
20100217952 Iyer Aug 2010 A1
20100229224 Etchegoyen Sep 2010 A1
20100241848 Smith Sep 2010 A1
20100281254 Carro Nov 2010 A1
20100321999 Yoo Dec 2010 A1
20100325367 Kornegay Dec 2010 A1
20100332922 Chang Dec 2010 A1
20110031546 Uenaka Feb 2011 A1
20110055458 Kuehne Mar 2011 A1
20110055471 Thatcher Mar 2011 A1
20110060722 Li Mar 2011 A1
20110072204 Chang Mar 2011 A1
20110099418 Chen Apr 2011 A1
20110153903 Hinkle Jun 2011 A1
20110161621 Sinclair Jun 2011 A1
20110161784 Selinger Jun 2011 A1
20110191525 Hsu Aug 2011 A1
20110218969 Anglin Sep 2011 A1
20110231598 Hatsuda Sep 2011 A1
20110239083 Kanno Sep 2011 A1
20110252188 Weingarten Oct 2011 A1
20110258514 Lasser Oct 2011 A1
20110289263 McWilliams Nov 2011 A1
20110289280 Koseki Nov 2011 A1
20110292538 Haga Dec 2011 A1
20110296411 Tang Dec 2011 A1
20110299317 Shaeffer Dec 2011 A1
20110302353 Confalonieri Dec 2011 A1
20110302408 McDermott Dec 2011 A1
20120017037 Riddle Jan 2012 A1
20120039117 Webb Feb 2012 A1
20120084523 Littlefield Apr 2012 A1
20120089774 Kelkar Apr 2012 A1
20120096330 Przybylski Apr 2012 A1
20120117399 Chan May 2012 A1
20120147021 Cheng Jun 2012 A1
20120151253 Horn Jun 2012 A1
20120159099 Lindamood Jun 2012 A1
20120159289 Piccirillo Jun 2012 A1
20120173792 Lassa Jul 2012 A1
20120203958 Jones Aug 2012 A1
20120210095 Nellans Aug 2012 A1
20120233523 Krishnamoorthy Sep 2012 A1
20120246392 Cheon Sep 2012 A1
20120278579 Goss Nov 2012 A1
20120284587 Yu Nov 2012 A1
20120324312 Moyer Dec 2012 A1
20120331207 Lassa Dec 2012 A1
20130013880 Tashiro Jan 2013 A1
20130013887 Sugahara Jan 2013 A1
20130016970 Koka Jan 2013 A1
20130018852 Barton Jan 2013 A1
20130024605 Sharon Jan 2013 A1
20130054822 Mordani Feb 2013 A1
20130061029 Huff Mar 2013 A1
20130073798 Kang Mar 2013 A1
20130080391 Raichstein Mar 2013 A1
20130138871 Chiu May 2013 A1
20130145085 Yu Jun 2013 A1
20130145089 Eleftheriou Jun 2013 A1
20130151759 Shim Jun 2013 A1
20130159251 Skrenta Jun 2013 A1
20130159723 Brandt Jun 2013 A1
20130166820 Batwara Jun 2013 A1
20130173845 Aslam Jul 2013 A1
20130179898 Fang Jul 2013 A1
20130191601 Peterson Jul 2013 A1
20130205183 Fillingim Aug 2013 A1
20130219131 Alexandron Aug 2013 A1
20130227347 Cho Aug 2013 A1
20130238955 D Abreu Sep 2013 A1
20130254622 Kanno Sep 2013 A1
20130318283 Small Nov 2013 A1
20130318395 Kalavade Nov 2013 A1
20130329492 Yang Dec 2013 A1
20140006688 Yu Jan 2014 A1
20140019650 Li Jan 2014 A1
20140025638 Hu Jan 2014 A1
20140082273 Segev Mar 2014 A1
20140082412 Matsumura Mar 2014 A1
20140095758 Smith Apr 2014 A1
20140095769 Borkenhagen Apr 2014 A1
20140095827 Wei Apr 2014 A1
20140108414 Stillerman Apr 2014 A1
20140108891 Strasser Apr 2014 A1
20140164447 Tarafdar Jun 2014 A1
20140164879 Tam Jun 2014 A1
20140181532 Camp Jun 2014 A1
20140195564 Talagala Jul 2014 A1
20140215129 Kuzmin Jul 2014 A1
20140223079 Zhang Aug 2014 A1
20140233950 Luo Aug 2014 A1
20140250259 Ke Sep 2014 A1
20140279927 Constantinescu Sep 2014 A1
20140304452 De La Iglesia Oct 2014 A1
20140310574 Yu Oct 2014 A1
20140337457 Nowoczynski Nov 2014 A1
20140359229 Cota-Robles Dec 2014 A1
20140365707 Talagala Dec 2014 A1
20140379965 Gole Dec 2014 A1
20150006792 Lee Jan 2015 A1
20150019798 Huang Jan 2015 A1
20150039849 Lewis Feb 2015 A1
20150067436 Hu Mar 2015 A1
20150082317 You Mar 2015 A1
20150106556 Yu Apr 2015 A1
20150106559 Cho Apr 2015 A1
20150121031 Feng Apr 2015 A1
20150142752 Chennamsetty May 2015 A1
20150143030 Gorobets May 2015 A1
20150186657 Nakhjiri Jul 2015 A1
20150199234 Choi Jul 2015 A1
20150227316 Warfield Aug 2015 A1
20150234845 Moore Aug 2015 A1
20150269964 Fallone Sep 2015 A1
20150277937 Swanson Oct 2015 A1
20150286477 Mathur Oct 2015 A1
20150294684 Qjang Oct 2015 A1
20150301964 Brinicombe Oct 2015 A1
20150304108 Obukhov Oct 2015 A1
20150310916 Leem Oct 2015 A1
20150317095 Voigt Nov 2015 A1
20150341123 Nagarajan Nov 2015 A1
20150347025 Law Dec 2015 A1
20150363271 Haustein Dec 2015 A1
20150363328 Candelaria Dec 2015 A1
20150372597 Luo Dec 2015 A1
20160014039 Reddy Jan 2016 A1
20160026575 Samanta Jan 2016 A1
20160041760 Kuang Feb 2016 A1
20160048327 Jayasena Feb 2016 A1
20160048341 Constantinescu Feb 2016 A1
20160054922 Awasthi Feb 2016 A1
20160062885 Ryu Mar 2016 A1
20160077749 Ravimohan Mar 2016 A1
20160077764 Ori Mar 2016 A1
20160077968 Sela Mar 2016 A1
20160078245 Amarendran Mar 2016 A1
20160098344 Gorobets Apr 2016 A1
20160098350 Tang Apr 2016 A1
20160103631 Ke Apr 2016 A1
20160110254 Cronie Apr 2016 A1
20160124742 Rangasamy May 2016 A1
20160132237 Jeong May 2016 A1
20160141047 Sehgal May 2016 A1
20160154601 Chen Jun 2016 A1
20160155750 Yasuda Jun 2016 A1
20160162187 Lee Jun 2016 A1
20160179399 Melik-Martirosian Jun 2016 A1
20160188223 Camp Jun 2016 A1
20160188890 Naeimi Jun 2016 A1
20160203000 Parmar Jul 2016 A1
20160224267 Yang Aug 2016 A1
20160232103 Schmisseur Aug 2016 A1
20160234297 Ambach Aug 2016 A1
20160239074 Lee Aug 2016 A1
20160239380 Wideman Aug 2016 A1
20160274636 Kim Sep 2016 A1
20160283140 Kaushik Sep 2016 A1
20160306699 Resch Oct 2016 A1
20160306853 Sabaa Oct 2016 A1
20160321002 Jung Nov 2016 A1
20160335085 Scalabrino Nov 2016 A1
20160342345 Kankani Nov 2016 A1
20160343429 Nieuwejaar Nov 2016 A1
20160350002 Vergis Dec 2016 A1
20160350385 Poder Dec 2016 A1
20160364146 Kuttner Dec 2016 A1
20160381442 Heanue Dec 2016 A1
20170004037 Park Jan 2017 A1
20170010652 Huang Jan 2017 A1
20170068639 Davis Mar 2017 A1
20170075583 Alexander Mar 2017 A1
20170075594 Badam Mar 2017 A1
20170091110 Ash Mar 2017 A1
20170109199 Chen Apr 2017 A1
20170109232 Cha Apr 2017 A1
20170123655 Sinclair May 2017 A1
20170147499 Mohan May 2017 A1
20170161202 Erez Jun 2017 A1
20170162235 De Jun 2017 A1
20170168986 Sajeepa Jun 2017 A1
20170177217 Kanno Jun 2017 A1
20170177259 Motwani Jun 2017 A1
20170185316 Nieuwejaar Jun 2017 A1
20170185498 Gao Jun 2017 A1
20170192848 Pamies-Juarez Jul 2017 A1
20170199823 Hayes Jul 2017 A1
20170212680 Waghulde Jul 2017 A1
20170212708 Suhas Jul 2017 A1
20170220254 Warfield Aug 2017 A1
20170221519 Matsuo Aug 2017 A1
20170228157 Yang Aug 2017 A1
20170242722 Qiu Aug 2017 A1
20170249162 Tsirkin Aug 2017 A1
20170262176 Kanno Sep 2017 A1
20170262178 Hashimoto Sep 2017 A1
20170262217 Pradhan Sep 2017 A1
20170269998 Sunwoo Sep 2017 A1
20170277655 Das Sep 2017 A1
20170279460 Camp Sep 2017 A1
20170285976 Durham Oct 2017 A1
20170286311 Juenemann Oct 2017 A1
20170322888 Booth Nov 2017 A1
20170344470 Yang Nov 2017 A1
20170344491 Pandurangan Nov 2017 A1
20170353576 Guim Bernat Dec 2017 A1
20180024772 Madraswala Jan 2018 A1
20180024779 Kojima Jan 2018 A1
20180033491 Marelli Feb 2018 A1
20180052797 Barzik Feb 2018 A1
20180067847 Oh Mar 2018 A1
20180069658 Benisty Mar 2018 A1
20180074730 Inoue Mar 2018 A1
20180076828 Kanno Mar 2018 A1
20180088867 Kaminaga Mar 2018 A1
20180107591 Smith Apr 2018 A1
20180113631 Zhang Apr 2018 A1
20180143780 Cho May 2018 A1
20180150640 Li May 2018 A1
20180165038 Authement Jun 2018 A1
20180165169 Camp Jun 2018 A1
20180165340 Agarwal Jun 2018 A1
20180167268 Liguori Jun 2018 A1
20180173620 Cen Jun 2018 A1
20180188970 Liu Jul 2018 A1
20180189175 Ji Jul 2018 A1
20180189182 Wang Jul 2018 A1
20180212951 Goodrum Jul 2018 A1
20180219561 Litsyn Aug 2018 A1
20180226124 Perner Aug 2018 A1
20180232151 Badam Aug 2018 A1
20180260148 Klein Sep 2018 A1
20180270110 Chugtu Sep 2018 A1
20180293014 Ravimohan Oct 2018 A1
20180300203 Kathpal Oct 2018 A1
20180307620 Zhou Oct 2018 A1
20180321864 Benisty Nov 2018 A1
20180322024 Nagao Nov 2018 A1
20180329776 Lai Nov 2018 A1
20180336921 Ryun Nov 2018 A1
20180349396 Blagojevic Dec 2018 A1
20180356992 Lamberts Dec 2018 A1
20180357126 Dhuse Dec 2018 A1
20180373428 Kan Dec 2018 A1
20180373655 Liu Dec 2018 A1
20180373664 Vijayrao Dec 2018 A1
20190012111 Li Jan 2019 A1
20190034454 Gangumalla Jan 2019 A1
20190042571 Li Feb 2019 A1
20190050312 Li Feb 2019 A1
20190050327 Li Feb 2019 A1
20190065085 Jean Feb 2019 A1
20190073261 Halbert Mar 2019 A1
20190073262 Chen Mar 2019 A1
20190087089 Yoshida Mar 2019 A1
20190087115 Li Mar 2019 A1
20190087328 Kanno Mar 2019 A1
20190108145 Raghava Apr 2019 A1
20190116127 Pismenny Apr 2019 A1
20190166725 Jing May 2019 A1
20190171532 Abadi Jun 2019 A1
20190172820 Meyers Jun 2019 A1
20190196748 Badam Jun 2019 A1
20190196907 Khan Jun 2019 A1
20190205206 Hornung Jul 2019 A1
20190212949 Pletka Jul 2019 A1
20190220392 Lin Jul 2019 A1
20190227927 Miao Jul 2019 A1
20190272242 Kachare Sep 2019 A1
20190278654 Kaynak Sep 2019 A1
20190278849 Chandramouli Sep 2019 A1
20190317901 Kachare Oct 2019 A1
20190320020 Lee Oct 2019 A1
20190339998 Momchilov Nov 2019 A1
20190361611 Hosogi Nov 2019 A1
20190377632 Oh Dec 2019 A1
20190377821 Pleshachkov Dec 2019 A1
20190391748 Li Dec 2019 A1
20200004456 Williams Jan 2020 A1
20200004674 Williams Jan 2020 A1
20200013458 Schreck Jan 2020 A1
20200042223 Li Feb 2020 A1
20200042387 Shani Feb 2020 A1
20200082006 Rupp Mar 2020 A1
20200084918 Shen Mar 2020 A1
20200089430 Kanno Mar 2020 A1
20200092209 Chen Mar 2020 A1
20200097189 Tao Mar 2020 A1
20200133841 Davis Apr 2020 A1
20200143885 Kim May 2020 A1
20200159425 Flynn May 2020 A1
20200167091 Haridas May 2020 A1
20200210309 Jung Jul 2020 A1
20200218449 Leitao Jul 2020 A1
20200225875 Oh Jul 2020 A1
20200242021 Gholamipour Jul 2020 A1
20200250032 Goyal Aug 2020 A1
20200257598 Yazovitsky Aug 2020 A1
20200322287 Connor Oct 2020 A1
20200326855 Wu Oct 2020 A1
20200328192 Zaman Oct 2020 A1
20200348888 Kim Nov 2020 A1
20200364094 Kahle Nov 2020 A1
20200371955 Goodacre Nov 2020 A1
20200387327 Hsieh Dec 2020 A1
20200401334 Saxena Dec 2020 A1
20200409559 Sharon Dec 2020 A1
20200409791 Devriendt Dec 2020 A1
20210010338 Santos Jan 2021 A1
20210075633 Sen Mar 2021 A1
20210089392 Shirakawa Mar 2021 A1
20210103388 Choi Apr 2021 A1
20210124488 Stoica Apr 2021 A1
20210132999 Haywood May 2021 A1
20210191635 Hu Jun 2021 A1
20210263795 Li Aug 2021 A1
20210286555 Li Sep 2021 A1
Foreign Referenced Citations (4)
Number Date Country
2003022209 Jan 2003 JP
2011175422 Sep 2011 JP
9418634 Aug 1994 WO
1994018634 Aug 1994 WO
Non-Patent Literature Citations (19)
Entry
https://web.archive.org/web/20071130235034/http://en.wikipedia.org:80/wiki/logical_block_addressing wikipedia screen shot retriefed on wayback Nov. 20, 2007 showing both physical and logical addressing used historically to access data on storage devices (Year: 2007).
Ivan Picoli, Carla Pasco, Bjorn Jonsson, Luc Bouganim, Philippe Bonnet. “uFLIP-OC: Understanding Flash I/O Patterns on Open-Channel Solid-State Drives.” APSys'17, Sep. 2017, Mumbai, India, pp. 1-7, 2017, <10.1145/3124680.3124741>. <hal-01654985>.
EMC Powerpath Load Balancing and Failover Comparison with native MPIO operating system solutions. Feb. 2011.
Tsuchiya, Yoshihiro et al. “DBLK: Deduplication for Primary Block Storage”, MSST 2011, Denver, CO, May 23-27, 2011 pp. 1-5.
Chen Feng, et al. “CAFTL: A Content-Aware Flash Translation Layer Enhancing the Lifespan of Flash Memory based Solid State Devices”< FAST '11, San Jose, CA Feb. 15-17, 2011, pp. 1-14.
Wu, Huijun et al. “HPDedup: A Hybrid Prioritized Data Deduplication Mechanism for Primary Storage in the Cloud”, Cornell Univ. arXiv: 1702.08153v2[cs.DC], Apr. 16, 2017, pp. 1-14https://www.syncids.com/#.
WOW: Wise Ordering for Writes—Combining Spatial and Temporal Locality in Non-Volatile Caches by Gill (Year: 2005).
Helen H. W. Chan et al. “HashKV: Enabling Efficient Updated in KV Storage via Hashing”, https://www.usenix.org/conference/atc18/presentation/chan, (Year: 2018).
S. Hong and D. Shin, “NAND Flash-Based Disk Cache Using SLC/MLC Combined Flash Memory,” 2010 International Workshop on Storage Network Architecture and Parallel I/Os, Incline Village, NV, 2010, pp. 21-30.
Arpaci-Dusseau et al. “Operating Systems: Three Easy Pieces”, Originally published 2015; Pertinent: Chapter 44; flash-based SSDs, available at http://pages.cs.wisc.edu/˜remzi/OSTEP/.
Jimenex, X., Novo, D. and P. Ienne, “Pheonix:Reviving MLC Blocks as SLC to Extend NAND Flash Devices Lifetime,”Design, Automation & Text in Europe Conference & Exhibition (Date), 2013.
Yang, T. Wu, H. and W. Sun, “GD-FTL: Improving the Performance and Lifetime of TLC SSD by Downgrading Worn-out Blocks,” IEEE 37th International Performance Computing and Communications Conference (IPCCC), 2018.
C. Wu, D. Wu, H. Chou and C. Cheng, “Rethink the Design of Flash Translation Layers in a Component-Based View”, in IEEE Acess, vol. 5, pp. 12895-12912, 2017.
Po-Liang Wu, Yuan-Hao Chang and T. Kuo, “A file-system-aware FTL design for flash-memory storage systems,” 2009, pp. 393-398.
S. Choudhuri and T. Givargis, “Preformance improvement of block based NAND flash translation layer”, 2007 5th IEEE/ACM/IFIP International Conference on Hardware/Software Codesign and Systems Synthesis (CODES+ISSS). Saizburg, 2007, pp. 257-262.
A. Zuck, O. Kishon and S. Toledo. “LSDM: Improving the Preformance of Mobile Storage with a Log-Structured Address Remapping Device Driver”, 2014 Eighth International Conference on Next Generation Mobile Apps, Services and Technologies, Oxford, 2014, pp. 221-228.
J. Jung and Y. Won, “nvramdisk: A Transactional Block Device Driver for Non-Volatile RAM”, in IEEE Transactions on Computers, vol. 65, No. 2, pp. 589-600, Feb. 1, 2016.
Te I et al. (Pensieve: a Machine Assisted SSD Layer for Extending the Lifetime: (Year: 2018).
ARM (“Cortex-R5 and Cortex-R5F”, Technical reference Manual, Revision r1p1) (Year:2011).
Related Publications (1)
Number Date Country
20210365362 A1 Nov 2021 US