Distributed data storage system providing enhanced security

Information

  • Patent Grant
  • 12204660
  • Patent Number
    12,204,660
  • Date Filed
    Friday, December 22, 2023
    a year ago
  • Date Issued
    Tuesday, January 21, 2025
    6 days ago
Abstract
In computer-implemented methods and systems for secure storage and transmission of data in a distributed network environment, each piece of data is transformed in to multiple pieces of metadata for transmission and storage. Each piece of metadata is transmitted and stored on a different server, which is selected from separate pools of servers.
Description
FIELD

The present specification discloses systems and methods for storage and transmission of data in a distributed network. More particularly, the present specification relates to increasing a security level of sensitive data during transmission and storage in a network.


BACKGROUND

Multiplayer video gaming systems, such as massive multiplayer online gaming systems, use distributed computing methods that enable players to cooperate, compete and communicate with each other. Distributed computing systems involve computing capabilities that are located on different computing devices within a network. The computing devices communicate with each other to coordinate their actions to enable data transmission and storage. Cloud computing systems are commonly used for transmission and storage of data using computer system resources over networks, such as the Internet. Data transmission over such a network is often subject to security threats from hackers and can be compromised in transit. Communication of data is prone to being intercepted by elements that intend to exploit the data. Similarly, transmitted data in a network that is stored in a computing device, such as a server, can be hacked by unknown elements. A common approach to address the concern over security of data during transmission use transport layer security (TLS).


Numerous techniques for securing data storage in a distributed pool of servers have been disclosed in the prior art. U.S. Pat. No. 9,753,931 teaches improving the security of data that is stored at a data store distributed over a computer network. Source data to be protected is partitioned into multiple files, and each file is obfuscated to create multiple obfuscated data files. Information as to how each obfuscated data file was obfuscated is stored in an associated trace file. The multiple obfuscated data files are moved around a computer network via a data movement process that includes sending each of the multiple obfuscated data files to a different randomly selected computer, where the computer further obfuscates the obfuscated data the trace file and sends the further obfuscated data and trace file to a next randomly selected computer.


U.S. Pat. No. 8,024,306 discloses system that use hash values as ‘unique’ identifiers for resources distributed across a network, and each one of a set of pool servers store the hash values for a set of computers within a LAN. When a resource is required, a hash value representing the resource is retrieved and compared with hash values stored at a pool server to determine whether the pool server holds a matching hash value. Any such matching hash value found on the pool server represents an identification of a local copy of the required resource, because of the uniqueness property of secure hash values. The information within the pool server can be used to access the required resource.


U.S. Pat. No. 10,474,643 teaches a distributed file system that includes metadata servers and data servers. The metadata server includes a selecting unit to select a data server from the data servers, a chunk allocation requesting unit to request that the selected data server perform chunk allocation, a chunk list managing unit to insert a list of chunks transmitted from the data server into a chunk list pool and determine an arrangement method of the chunk list pool, and a chunk fetching unit to fetch available chunk information from the chunk list pool. The data server includes a receiving unit to receive request for chunk allocation from the metadata server, a chunk allocating unit to allocate chunks in response to the request for chunk allocation and write a list of chunks based on information about the allocated chunk, and a transmitting unit to transmit the list of chunks to the metadata server.


While conventional approaches provide security for distributed data files, they generally still suffer from excessive complexity, security vulnerabilities, and/or substantial computing resources. Accordingly, there remains a need to improve the security of data storage and transmission in a distributed computing system. In particular, there is a need to store data in a manner that limits the extent to which individual servers have knowledge about the data that they store.


SUMMARY

The following embodiments and aspects thereof are described and illustrated in conjunction with systems, tools and methods, which are meant to be exemplary and illustrative, not limiting in scope.


In some embodiments, the present specification describes a computer-implemented method for securing data storage and transmission in a distributed network, the method being implemented in a computer having a processor and a random access memory, wherein the processor is in data communication with a storage unit, the method comprising: storing data in the network, wherein storing the data comprises: selecting a first set of servers, wherein each of the first set of servers is selected from different pools of servers; selecting a first server from the first set of servers; sending a data key to each of the first set of servers except for the selected first server; receiving a first set of metadata from each of the first set of servers except for the selected first server; and decomposing the first set of metadata; sending the data key and the first set of metadata to the first server; and retrieving stored data from the network, wherein retrieving the stored data comprises: selecting a second set of servers, wherein the second set of servers is selected from each of a plurality of a pool of servers; sending the data key to the second set of servers; receiving a second set of metadata associated with the data key from the second set of servers; and recomposing the data from the second set of metadata using the data key.


Optionally, the storing data in the network further comprises synchronizing the data key and the first set of metadata within each of the plurality of pools of servers.


Optionally, the second set of metadata is obtained from the synchronization.


Optionally, the decomposing comprises using a mathematical function to decompose the first set of metadata. Still optionally, the mathematical function is the Exclusive OR (EXOR) function.


Optionally, the recomposing comprises using a mathematical function to decompose the first set of metadata. Still optionally, the mathematical function is the Exclusive OR (EXOR) function.


Optionally, the data and the metadata has a length L. Still optionally, the sending the key of data to the first set of servers comprises sending the length L. Still optionally, the length L is predefined to be sufficiently long to accommodate a longest data string scenario. Still optionally, the storing the data comprises padding the data and the metadata with randomly generated data to achieve the length L. Optionally, the retrieving the stored data comprises eliminating the padding.


Optionally, the data key comprises a data structure.


Optionally, the data key comprises a string.


Optionally, the selecting the first set of servers comprises using Domain Name System to automatically select the first set of servers that has the fastest connection speed, relative to other servers, with the computer implementing the computer-implemented method.


Optionally, the connection speed has a latency of at least less than 100 ms and a bandwidth of at least 1 Gbps.


Optionally, the selecting the first set of servers comprises receiving a list of servers available in each pool of servers and randomly choosing one server from each pool of servers.


Optionally, the sending the data key comprises sending a key adapted to identify the data.


In some other embodiments, the present specification describes a computer-implemented method for securing data storage and transmission in a distributed network, the method being implemented in a computer having a processor and a random access memory, wherein the processor is in data communication with a storage unit, the method comprising: storing data in the network, wherein storing the data comprises: selecting a first set of servers, wherein each of the first set of servers is selected from different pools of servers; selecting a first server from the first set of servers; sending a data key to each of the first set of servers except for the selected first server; receiving a first set of metadata from a second server and a third server from the first set of servers; and decomposing the first set of metadata received from the second server and the third server from the first set of servers; sending the data key and the first set of metadata to the first server; and retrieving stored data from the network, wherein retrieving the stored data comprises: selecting a second set of servers, wherein the second set of servers is selected from each of a plurality of a pool of servers; sending the data key to the second set of servers; receiving a second set of metadata associated with the data key from the second set of servers; and recomposing the data from the second set of metadata using the data key.


Optionally, storing data in the network further comprises synchronizing the data key and the first set of metadata within each of the plurality of pools of servers. The second set of metadata may be obtained from the synchronization.


Optionally, decomposing comprises using a mathematical function to decompose the first set of metadata. The mathematical function may be the Exclusive OR (EXOR) function.


Optionally, the recomposing comprises using a mathematical function to decompose the first set of metadata. The mathematical function may be the Exclusive OR (EXOR) function.


Optionally, the data and the metadata has a length L. Optionally, sending the key of data to the first set of servers comprises sending the length L.


The aforementioned and other embodiments of the present specification shall be described in greater depth in the drawings and detailed description provided below.





BRIEF DESCRIPTION OF THE DRAWINGS

These and other features and advantages of the present specification will be appreciated, as they become better understood by reference to the following detailed description when considered in connection with the accompanying drawings, wherein:



FIG. 1 illustrates an exemplary network architecture according to some implementations of the present specification;



FIG. 2A illustrates an exemplary interaction between components of the distributed network performed during storage of data by a client, in accordance with some embodiments of the present specification;



FIG. 2B is a flow diagram detailing an exemplary process used for storage of data by the client, in accordance with some embodiments of the present specification;



FIG. 3A illustrates an exemplary interaction between components of the distributed network performed during retrieval of data by the client, in accordance with some embodiments of the present specification; and



FIG. 3B is a flow diagram detailing an exemplary process used for retrieval of data by the client, in accordance with some embodiments of the present specification.





DETAILED DESCRIPTION

The present specification relates to security systems and methods in distributed computing systems. Embodiments of the present specification provide a method for storing and transmitting data to enhance or increase a security level of the data. The present specification teaches a secure and resilient data storage system and method whereby storage is distributed across a pool of servers and, apart from an index key for identifying the data and length of the data, individual servers do not have knowledge of or about the data that they store. Accordingly, each piece of data is transformed into multiple pieces of metadata, whereby each piece of metadata is then transmitted and stored on a different server. Each server is selected from a pool of servers. Therefore, individually, none of the pieces of data yield any information about the data itself during transmission, since storage is distributed across a pool of servers. Further, should there be an attack, the hacker is forced to compromise all servers where the metadata is stored and intercept communication to all servers (instead of hacking and intercepting just one server), which is difficult and nearly impossible. One of ordinary skill in the art would appreciate that a single server pool is, logically, an autonomous region that contains one or more physical servers, presenting a unified view of the storage in which the virtual machines reside. This is opposed to different server pools in which servers in the different pools present a differing view of the storage.


The present specification is directed towards multiple embodiments. The following disclosure is provided in order to enable a person having ordinary skill in the art to practice the invention. Language used in this specification should not be interpreted as a general disavowal of any one specific embodiment or used to limit the claims beyond the meaning of the terms used therein. The general principles defined herein may be applied to other embodiments and applications without departing from the spirit and scope of the invention. Also, the terminology and phraseology used is for the purpose of describing exemplary embodiments and should not be considered limiting. Thus, the present specification is to be accorded the widest scope encompassing numerous alternatives, modifications and equivalents consistent with the principles and features disclosed. For purpose of clarity, details relating to technical material that is known in the technical fields related to the invention have not been described in detail so as not to unnecessarily obscure the present invention.


In the description and claims of the application, each of the words “comprise” “include” and “have”, and forms thereof, are not necessarily limited to members in a list with which the words may be associated.


It should be noted herein that any feature or component described in association with a specific embodiment may be used and implemented with any other embodiment unless clearly indicated otherwise.


It should be appreciated that the programmatic methods described herein may be performed on any computing device, including a laptop, desktop, smartphone, tablet computer, specialized gaming console, or virtual reality system. The computing device comprises at least one processor and a nonvolatile memory that stores the programmatic instructions which, when executed by the processor, perform the methods or steps disclosed herein, including the generation of a graphical user interface that is communicated to a local or remote display. The computing device is in communication with at least one remotely located server through a network of any type.


The embodiments disclosed herein are directed to an improvement in secure transmission and storage of sensitive data including and not limited to credit card data, social security number (SSN) related data, and other data that requires secure transmission and storage. The embodiments disclosed herein are also directed to an improvement in computer-related technology (enabling computers to enable secure gaming experiences in an online multiplayer gaming environment), and thus do not recite abstract ideas or concepts. The improved secure transmission and storage of data are achieved through the use of specific rules to communicate data such as player data within the gaming environment which, when executed, enable the automation of specific content generation, transmission, and storage of data that previously were not available or could not be automated. These new secure data transmission and storage rules for communication between computing devices improve existing technological processes in distributed computing and, therefore, are not abstract or conceptual in nature. This specification therefore teaches how the disclosed inventions improve a secure communication technology using a specific set of rules, and particular solutions to the aforementioned failures in conventional communication systems to achieve the desired outcomes.


While aspects of the present specification may be described herein with reference to various gaming systems and players of online gaming systems, it should be appreciated that any such examples are for illustrative purposes only, and are not intended to be limiting.


The terminology used within this specification and detailed description of the various embodiments is for the purpose of describing particular embodiments only and is not intended to limit the invention.


The term “a multiplayer game environment” or “massively multiplayer online game” may be construed to mean a specific hardware architecture in which one or more servers electronically communicate with, and concurrently support game interactions with, a plurality of client devices, thereby enabling each of the client devices to simultaneously play in the same instance of the same game. Preferably, the plurality of client devices number in the dozens, preferably hundreds, preferably thousands. In one embodiment, the number of concurrently supported client devices ranges from 10 to 5,000,000 and every whole number increment or range therein. Accordingly, a multiplayer game environment or massively multiplayer online game is a computer-related technology, a non-generic technological environment, and should not be abstractly considered a generic method of organizing human activity divorced from its specific technology environment.


In embodiments, the term “metadata” may refer to any set of randomly generated data (such as, but not limited to numbers or binary numbers).


Exemplary System Architecture



FIG. 1 illustrates an exemplary network architecture 100 according to some implementations of the present specification. Network architecture 100 may be used to implement various embodiments described in subsequent sections of the present specification. For example, network architecture 100 may implement various programs that result in improved security of data transmission and storage in the network.



FIG. 1 illustrates a client device C 102 that is configured to transmit, store, and retrieve data through communication to multiple servers. Each server is assigned to one of several pools of servers. A first pool X 104 comprises servers X1 104a, X2 104b, and X3 104c. A second pool Y 106 comprises servers Y1 106a, Y2 106b, and Y3 106c. A third pool Z 108 comprises servers Z1 108a, Z2 108b, and Z3 108c. FIG. 1 illustrates a representative set of server pools 104, 106, and 108. The actual pools of servers may vary in numbers and the numbers of servers in each pool may also vary. In embodiments, at least two pools of n servers are employed. In a preferred embodiment, at least three pools of n servers are employed for increased security. In embodiments, each pool of servers has at least three individual servers which may increase resilience. In some embodiments, each pool of servers may operate using a different operating system or network/communication protocol. A single pool of servers may be defined as a logically autonomous processing or computing system that is made of two or more physical servers. A pool of servers provides a unified view of the data present throughout the pool of servers and are typically defined by a software program that is configured to integrate, associate, or otherwise link an IP address associated with one of the physical servers in the pool with IP addresses of other physical servers in the pool.


In an embodiment, client device C 102 is a computer system that is configured to host gameplay between (or with) other devices. Client device C 102 may be configured as a gaming console, a handheld gaming device, a personal computer (e.g., a desktop computer, a laptop computer, etc.), a smartphone, a tablet computing device, a smart television, and/or other device that can be used to communicate data. Client device C 102 may include one or more processors, one or more storage devices (which may store one or more applications), one or more peripherals, and/or other components. Processors may be programmed by one or more computer program instructions. For example, processors may be programmed by the application and/or other instructions. The other devices may include other computing systems which may, similar to client device C 102, perform data transmission and request for storage of data within the network architecture 100.


Depending on the system configuration, an application (or portions thereof) may be part of a game application, which generates a game instance to facilitate gameplay. Alternatively or additionally, the application may run on a device such as a server to perform its designated function(s) for users in an “online” game hosted by the server. In embodiments, portions or all of the application 192 may run on computer system of client C 102 and/or at least one server among the pools 104, 106, 108 of servers.


The application may include programmatic instructions to be implemented by a computing system in client C 102 and in servers in the pools 104, 106 and 108, each of which are described in greater detail herein. Servers in pools 104, 106, and 108 may include one or more computing devices. The servers may include one or more physical processors programmed by computer program instructions, one or more storage devices (which may store the application), and/or other components. The processors may be programmed by one or more computer program instructions. As used herein, for convenience, the various instructions will be described as performing an operation, when, in fact, the various instructions program the processors in the computer system of client C 102 and servers in pools 104, 106 and 108, to perform the operation.


At least one peripheral may be used to obtain an input (for example, direct input, measured input, among other types of input) from a user at client C 102. Data input is securely transmitted and stored by the application executed within the distributed network architecture 100. The at least one peripherals may include, without limitation, a game controller, a gamepad, a keyboard, a mouse, an imaging device such as a camera, a motion sensing device, a light sensor, a biometric sensor, and/or other peripheral device that can obtain an input from and/or relating to a user. In some embodiments, the user is a player of the online game. Peripherals may be coupled to a corresponding client C 102 via a wired and/or wireless connection.


Although illustrated in FIG. 1 as a single component, client device C 102 and servers in pools 104, 106, and 108 may each include a plurality of individual processing components, each programmed with at least some of the functions described herein. In this manner, some components of client device C 102 and servers in pools 104, 106, and 108 may perform some functions while other components may perform other functions, as would be appreciated. The one or more processors may each include one or more physical processors that are programmed by computer program instructions. Thus, any one or all of client device C 102 and servers in pools 104, 106, and 108 may function as a host computer programmed by the application. The various instructions described herein are exemplary only. Other configurations and numbers of instructions may be used, so long as the processor(s) are programmed to perform the functions described herein.


The various programmatic instructions described herein may be stored in a storage device which may comprise random access memory (RAM), read only memory (ROM), and/or other memory. The storage device may store the computer program instructions (e.g., the aforementioned instructions) to be executed by the processor as well as data that may be manipulated by processor. The storage device may comprise floppy disks, hard disks, optical disks, tapes, or other storage media for storing computer-executable instructions and/or data. Furthermore, it should be appreciated that although the various storage devices housing the instructions are illustrated in FIG. 1 as being co-located within a single processing unit, in implementations in which the processor(s) include multiple processing units, one or more instructions may be executed remotely and discretely from the other instructions.


The description of the functionality provided by the different instructions described herein is for illustrative purposes, and is not intended to be limiting, as any of the programmatic instructions may provide more or less functionality than is described. For example, one or more of the instructions may be eliminated, and some or all of its functionality may be provided by any of the other instructions. As another example, the processor(s) may be programmed by one or more additional instructions that may perform some or all of the functionality attributed herein to one of the instructions.


The various components illustrated in FIG. 1 may be coupled to at least one other component via a network, which may include any one or more of, for instance, the Internet, an intranet, a PAN (Personal Area Network), a LAN (Local Area Network), a WAN (Wide Area Network), a SAN (Storage Area Network), a MAN (Metropolitan Area Network), a wireless network, a cellular communications network, a Public Switched Telephone Network, and/or other network. In FIG. 1, as well as in the other Figures, different numbers of entities than those depicted may be used. Furthermore, according to various implementations, the components described herein may be implemented in hardware and/or software that configure hardware.


The various databases described herein may be, include, or interface to, for example, an Oracle™ relational database sold commercially by Oracle Corporation. Other databases, such as Informix™, DB2 (Database 2) or other data storage, including file-based, or query formats, platforms, or resources such as OLAP (On Line Analytical Processing), SQL (Structured Query Language), a SAN (storage area network), Microsoft Access™ or others may also be used, incorporated, or accessed. The database may comprise one or more such databases that reside in one or more physical devices and in one or more physical locations. The database may store a plurality of types of data and/or files and associated data or file descriptions, administrative information, or any other data.



FIG. 2A illustrates an exemplary interaction between components of the distributed network architecture 100 performed during storage of data (D) by client device C 102, in accordance with some embodiments of the present specification. FIG. 2B is a flow diagram showing an exemplary process used for storage of data (D) by client device C 102, in accordance with some embodiments of the present specification. Referring simultaneously to FIGS. 2A and 2B, at step 202, client device C 102 selects at least one server from each of the plurality of pools of servers. In one embodiment, step 202 is performed implicitly, where Domain Name System (DNS) is used to automatically select the server in a pool that will have a faster connection to client C 102, relative to other servers. In embodiments, connection speed is measured based on the latency for pinging a server from the client and/or based on the available bandwidth for communication between the client and server. Connection speed for a certain server is relative to connection speed for the other services and, when step 202 is performed implicitly, the fastest connection is chosen. In some embodiments, when step 202 is performed implicitly, the connection speed has a latency of at least less than 100 ms and a bandwidth of at least 1 Gbps. In another embodiment, step 202 is performed explicitly, where client C 102 receives a list of the servers available in each pool and randomly chooses one. In embodiments, the selection is performed randomly by client device C 102. In the example, servers X2 104b, Y1 106a, and Z3 108c, are selected from pools 104, 106 and 108, respectively. At step 204, client C device 102 randomly selects one server from the servers X2 104b, Y1 106a, and Z3 108c selected at step 202, for storage of data. In the illustrated example, server Y1 106a is selected at this step 204.


At step 206, client device C 102 sends a key (K) having a length (L) of data to the servers X2 104b and Z3 108c, which were not selected at step 204. The same K and L are sent to the servers. The key (K) represents a data structure adapted to identify the data (D) that is being stored and will subsequently be retrieved. In embodiments, key (K) is a string, or any other format that can enable subsequent retrieval of the original data. The length (L) represents the length of the data (D). In embodiments, different lengths of data may be used by client device C 102. In some embodiments, client device C 102 uses a specific length L to represent the data (D). In embodiments, length (L) is predefined so that the server does not need to know the length since it is consistent throughout. In case of predefined length (L), such as a credit card number, the length does not need to be communicated to the servers. In alternate embodiments, if length (L) is variable, the length is required to be communicated to the servers. In embodiments, client device C 102 determines the type of data (D) that is being processed for storage and subsequent retrieval. The type of data (D) is known to client C 102 from the application storing the data, such as a SSN or credit card number. The known type of data (D) enables client (C) 102 to use appropriate storage with the server, since the length of data is fixed and there is no need to communicate L. Alternatively, client (C) 102 uses generic storage and communicates the length (L) to the servers. In one embodiment, a first type of data is a Social Security Number (SSN) which has a first length and a second type of data is a credit card (CC) number which has a second, different length. Client device C 102 may pad or augment the original data with randomly generated data to achieve a predefined length L. In embodiments, while uniform length L of data is not required in all scenarios, there are advantages to predefine length such as speed and predictability of storage. In embodiments, length L is predefined to be sufficiently long to accommodate the longest data string scenario. Once the data is processed for storage and subsequently recomposed, client device C 102 eliminates the padding, since the type of data is known to client device C 102, and therefore client device C 102 can determine the amount of padding.


At step 208, client device C 102 receives metadata returned by the servers X2 104b and Z3 108c. In the illustrated example, server X2 104b returns to the client device C 102 DX2 metadata which is randomly generated and has a length L. Similarly, server Z3 108c returns to the client device C 102 DZ3 metadata which is randomly generated and has a length L. Key (K) is communicated back to the servers by client device C 102 alongside, or implied, similarly to an answer to a web API call. Additionally, server X2 104b now stores key K that is associated with metadata DX2, and server Z3 108c stores key K that is associated with metadata DZ3.


At step 210, client C 102 decomposes the data to obtain metadata associated with the server Y1 106a selected at step 204. In one embodiment, the Exclusive OR (EXOR/EOR, denoted by symbol A) algorithmic function is used to decompose, and therefore subsequently recompose, the data. While some embodiments of the present specification are described with reference to the use of EXOR, it should be appreciated that EXOR is only one of the possible methods to decompose/recompose the data. Other mathematical operations and/or functions may be used in place of EXOR to achieve the objectives prescribed by the present specification. For example, addition or subtraction functions may be used in place of EXOR. Using the EXOR function enables storing an arbitrary data length and having a predictable computational time. In embodiments, any length L may be selected as long as that length is uniform across all servers and data. However, in order to use EXOR it should be noted that binary data of equal lengths (L) must be compared. When the client decomposes the data, the data is represented by:

DY1=D{circumflex over ( )}DX2{circumflex over ( )}DZ3


At step 212, client device C 102 sends key K and metadata DY1 to server Y1 106a. At this point, server X2 104b stores K and metadata DX2, server Y1 106a stores K and metadata DY1, and server Z3 stores K and metadata DZ3. Additionally, data D required to be stored by C 102 is now represented as:

D=DX2{circumflex over ( )}DY1{circumflex over ( )}DZ3


Further, at step 214, the metadata (DX2, DY1, DZ3) stored within each server (X2 104b, Y1 106a, Z3 108c) is synchronized with the pool corresponding to each server. In some embodiments, the synchronization is performed to address availability of the overall system and make interception of the communication more difficult. If a server is not available or communication to that server is not possible, any other server in the same pool can be used to successfully complete the transaction. Intercepting communication to all the servers participating in a transaction is harder as communication will have to either be intercepted at all the servers or at all the clients. In the illustrated example, server X2 104b sends K and its associated metadata DX2 to servers X1 104a and X3 104c. Similarly, server Y1 106a sends K and its associated metadata DY1 to servers Y2 106b and Y3 106c. Server Z3 108c sends K and its associated metadata DZ3 to servers Z1 108a and Z2 108b. The data is synchronized between servers in a pool so that subsequently, the retrieval of data can be managed without the knowledge of the original server that stored the metadata.



FIG. 3A illustrates an exemplary interaction between components of the distributed network architecture 100 performed during retrieval of data (D) by client device C 102, in accordance with some embodiments of the present specification. FIG. 3B is a flow diagram showing an exemplary process used for retrieval of data (D) by client device C 102, in accordance with some embodiments of the present specification. Referring simultaneously to FIGS. 3A and 3B, at step 304 client device C 102 randomly selects at least one server from each of the plurality of server pools within the network 100. For example, client device C 102 selects server X1 104a from pool 104, server Y3 106c from pool 106, and server Z2 108b from pool 108. At step 304, client C 102 sends a key K to each server (X1 104a, Y3 106c and Z2 108b) selected at step 302. At step 306, each selected server (X1 104a, Y3 106c and Z2 108b), in response to the key K sent by client device C 102, returns to client device C 102 metadata that was previously provided while synchronizing the servers within each pool. Therefore, in the example shown, server X1 104a returns to client C 102 metadata DX1, server Y3 106c returns to client device C 102 metadata DY3, and server Z2 108b returns to client device C 102 metadata DZ2. At step 308, after receiving metadata from the selected servers (X1 104a, Y3 106c, Z2 108b) from each of the plurality of the pool of servers (104, 106, and 108), client C 102 recomposes the stored data D using the designated mathematical function. In the illustrated example, the EXOR function is used. Therefore, client C 102 recomposes the data D using EXOR:

D=DX1{circumflex over ( )}DY3{circumflex over ( )}DZ2


Embodiments of the present specification offer several advantages. The sensitive data (D) itself is never transmitted over the network 100. A partial metadata (DX, DY, DZ) is used to recreate the data (D). Additionally, the sensitive data (D) itself is never stored anywhere on the network 100 since partial metadata (DX, DY, DZ) are used. Further, security of storing the sensitive data D is improved by the embodiments of the present specification since an attacker is faced with the difficult task of compromising at least one server in multiple pools of servers in order to acquire the data D. In some embodiments, the servers are operated using different technologies and are secured with different security measures, making it even harder to acquire the data D.


Also, data transmission security is improved by embodiments of the present specification, since in order to intercept the data D, an attacker is required to intercept all the transmissions from all the servers in the multiple pools of servers. It may be difficult for the attacker to intercept data because the attacker does not likely know, in advance, which individual servers located within the pool of servers will be selected by the client for storage and retrieval of data D. Additionally, if the transmission of the data is based on different technologies and with different security measures for each server, depending on the server pool, the task of intercepting the sensitive data D is even harder.


The above examples are merely illustrative of the many applications of the system of present invention. Although only a few embodiments of the present invention have been described herein, it should be understood that the present invention might be embodied in many other specific forms without departing from the spirit or scope of the invention. Therefore, the present examples and embodiments are to be considered as illustrative and not restrictive, and the invention may be modified within the scope of the appended claims.

Claims
  • 1. A system for securing data storage and transmission in a network comprising at least a first server pool and a second server pool, wherein first server pool comprises a first plurality of servers and wherein the second server pool comprises a second plurality of servers, the system comprising: first programmatic instructions stored in a non-transient computer medium wherein, when executed by a processor, the first programmatic instructions cause data to be stored in the network by selecting a first server of the first plurality of servers from the first server pool and selecting a second server of the second plurality of servers from the second server pool, wherein the first server has a faster connection speed than other servers of the first plurality of servers and wherein the second server has a faster connection speed than other servers of the second plurality of servers;second programmatic instructions stored in the non-transient computer medium wherein, when executed by the processor, the second programmatic instructions select the first server;third programmatic instructions stored in the non-transient computer medium wherein, when executed by the processor, the third programmatic instructions send a data key to the second server and not the first server;fourth programmatic instructions stored in the non-transient computer medium wherein, when executed by the processor, the fourth programmatic instructions receive a first set of metadata from the second server;fifth programmatic instructions stored in the non-transient computer medium wherein, when executed by the processor, the fifth programmatic instructions decompose the first set of metadata using a mathematical function; andsixth programmatic instructions stored in the non-transient computer medium wherein, when executed by the processor, the sixth programmatic instructions send the data key and the first set of metadata to the first server;seventh programmatic instructions stored in the non-transient computer medium wherein, when executed by the processor, the seventh programmatic instructions select a third server of the first plurality of servers, select a fourth server of the second plurality of servers, send the data key to at least one of the third server and fourth server, receive a second set of metadata associated with the data key from at least one of the third server and fourth server, and recompose the received second set of metadata using the data key.
  • 2. The system of claim 1, wherein the seventh programmatic instructions sending the data key to at least one of the third server and fourth server comprises sending a key adapted to identify the data.
  • 3. The system of claim 1, wherein the seventh programmatic instructions recomposing the second set of metadata comprises using the same mathematical function used to decompose the first set of metadata.
  • 4. The system of claim 1, wherein the mathematical function is the Exclusive OR (EXOR) function.
  • 5. The system of claim 1, further comprising eighth programmatic instructions stored in the non-transient computer medium wherein, when executed by the processor, the eighth programmatic instructions synchronize the data key and the first set of metadata.
  • 6. The system of claim 5, further comprising ninth programmatic instructions stored in the non-transient computer medium wherein, when executed by the processor, the ninth programmatic instructions obtain a second set of metadata from the synchronization.
  • 7. The system of claim 1, wherein the data and the metadata has a length L.
  • 8. The system of claim 7, wherein the third programmatic instructions sending a data key to the second server and not the first server comprises sending the length L.
  • 9. The system of claim 7, wherein the length L is predefined to be sufficiently long to accommodate a longest data string scenario.
  • 10. The system of claim 7, wherein the first programmatic instructions causing the data to be stored in the network comprises padding the data and the metadata with randomly generated data to achieve the length L.
  • 11. The system of claim 10, further comprising eighth programmatic instructions stored in the non-transient computer medium wherein, when executed by the processor, the eighth programmatic instructions retrieve stored data from the network and wherein the retrieving the stored data comprises eliminating the padding.
  • 12. The system of claim 1 wherein the data key comprises a data structure.
  • 13. The system of claim 1 wherein the data key comprises a string.
  • 14. The system of claim 1, wherein the connection speed has a latency of at least less than 100 ms and a bandwidth of at least 1 Gbps.
  • 15. The system of claim 1, wherein the first plurality of servers is chosen randomly from a list of servers available in the first pool of servers.
  • 16. The system of claim 1, wherein the second plurality of servers is chosen randomly from a list of servers available in the second pool of servers.
CROSS-REFERENCE

The present application is a continuation application of U.S. patent application Ser. No. 17/563,155, titled “Distributed Data Storage System Providing Enhanced Security” and filed on Dec. 28, 2021, which relies on, for priority, U.S. Patent Provisional Application No. 63/131,944, titled “Distributed Data Storage System Providing Enhanced Security” and filed on Dec. 30, 2020. The above-referenced applications are herein incorporated by reference in their entirety.

US Referenced Citations (279)
Number Name Date Kind
5530796 Wang Jun 1996 A
5561736 Moore Oct 1996 A
5563946 Cooper Oct 1996 A
5685775 Bakoglu Nov 1997 A
5706507 Schloss Jan 1998 A
5708764 Borrel Jan 1998 A
5736985 Lection Apr 1998 A
5737416 Cooper Apr 1998 A
5745678 Herzberg Apr 1998 A
5768511 Galvin Jun 1998 A
5825877 Dan Oct 1998 A
5835692 Cragun Nov 1998 A
5878233 Schloss Mar 1999 A
5883628 Mullaly Mar 1999 A
5900879 Berry May 1999 A
5903266 Berstis May 1999 A
5903271 Bardon May 1999 A
5911045 Leyba Jun 1999 A
5920325 Morgan Jul 1999 A
5923324 Berry Jul 1999 A
5969724 Berry Oct 1999 A
5977979 Clough Nov 1999 A
5990888 Blades Nov 1999 A
6014145 Bardon Jan 2000 A
6025839 Schell Feb 2000 A
6059842 Dumarot May 2000 A
6069632 Mullaly May 2000 A
6081270 Berry Jun 2000 A
6081271 Bardon Jun 2000 A
6091410 Lection Jul 2000 A
6094196 Berry Jul 2000 A
6098056 Rusnak Aug 2000 A
6104406 Berry Aug 2000 A
6111581 Berry Aug 2000 A
6134588 Guenthner Oct 2000 A
6144381 Lection Nov 2000 A
6148328 Cuomo Nov 2000 A
6185614 Cuomo Feb 2001 B1
6201881 Masuda Mar 2001 B1
6222551 Schneider Apr 2001 B1
6271842 Bardon Aug 2001 B1
6271843 Lection Aug 2001 B1
6282547 Hirsch Aug 2001 B1
6311206 Malkin Oct 2001 B1
6334141 Varma Dec 2001 B1
6336134 Varma Jan 2002 B1
6337700 Kinoe Jan 2002 B1
6353449 Gregg Mar 2002 B1
6356297 Cheng Mar 2002 B1
6411312 Sheppard Jun 2002 B1
6426757 Smith Jul 2002 B1
6445389 Bossen Sep 2002 B1
6452593 Challener Sep 2002 B1
6462760 Cox, Jr. Oct 2002 B1
6469712 Hilpert, Jr. Oct 2002 B1
6473085 Brock Oct 2002 B1
6499053 Marquette Dec 2002 B1
6505208 Kanevsky Jan 2003 B1
6525731 Suits Feb 2003 B1
6549933 Barrett Apr 2003 B1
6567109 Todd May 2003 B1
6618751 Challenger Sep 2003 B1
RE38375 Herzberg Dec 2003 E
6657617 Paolini Dec 2003 B2
6657642 Bardon Dec 2003 B1
6684255 Martin Jan 2004 B1
6717600 Dutta Apr 2004 B2
6734884 Berry May 2004 B1
6765596 Lection Jul 2004 B2
6781607 Benham Aug 2004 B1
6819669 Rooney Nov 2004 B2
6832239 Kraft Dec 2004 B1
6836480 Basso Dec 2004 B2
6886026 Hanson Apr 2005 B1
6948168 Kuprionas Sep 2005 B1
RE38865 Dumarot Nov 2005 E
6993596 Hinton Jan 2006 B2
7028296 Irfan Apr 2006 B2
7062533 Brown Jun 2006 B2
7143409 Herrero Nov 2006 B2
7209137 Brokenshire Apr 2007 B2
7230616 Taubin Jun 2007 B2
7249123 Elder Jul 2007 B2
7263511 Bodin Aug 2007 B2
7287053 Bodin Oct 2007 B2
7305438 Christensen Dec 2007 B2
7308476 Mannaru Dec 2007 B2
7404149 Fox Jul 2008 B2
7426538 Bodin Sep 2008 B2
7427980 Partridge Sep 2008 B1
7428588 Berstis Sep 2008 B2
7429987 Leah Sep 2008 B2
7436407 Doi Oct 2008 B2
7439975 Hsu Oct 2008 B2
7443393 Shen Oct 2008 B2
7447996 Cox Nov 2008 B1
7467181 McGowan Dec 2008 B2
7475354 Guido Jan 2009 B2
7478127 Creamer Jan 2009 B2
7484012 Hinton Jan 2009 B2
7503007 Goodman Mar 2009 B2
7506264 Polan Mar 2009 B2
7515136 Kanevsky Apr 2009 B1
7525964 Astley Apr 2009 B2
7552177 Kessen Jun 2009 B2
7565650 Bhogal Jul 2009 B2
7571224 Childress Aug 2009 B2
7571389 Broussard Aug 2009 B2
7580888 Ur Aug 2009 B2
7596596 Chen Sep 2009 B2
7640587 Fox Dec 2009 B2
7667701 Leah Feb 2010 B2
7698656 Srivastava Apr 2010 B2
7702784 Berstis Apr 2010 B2
7714867 Doi May 2010 B2
7719532 Schardt May 2010 B2
7719535 Tadokoro May 2010 B2
7734691 Creamer Jun 2010 B2
7737969 Shen Jun 2010 B2
7743095 Goldberg Jun 2010 B2
7747679 Galvin Jun 2010 B2
7765478 Reed Jul 2010 B2
7768514 Pagan Aug 2010 B2
7773087 Fowler Aug 2010 B2
7774407 Daly Aug 2010 B2
7782318 Shearer Aug 2010 B2
7792263 D Amora Sep 2010 B2
7792801 Hamilton, II Sep 2010 B2
7796128 Radzikowski Sep 2010 B2
7808500 Shearer Oct 2010 B2
7814152 McGowan Oct 2010 B2
7827318 Hinton Nov 2010 B2
7843471 Doan Nov 2010 B2
7844663 Boutboul Nov 2010 B2
7847799 Taubin Dec 2010 B2
7856469 Chen Dec 2010 B2
7873485 Castelli Jan 2011 B2
7882222 Dolbier Feb 2011 B2
7882243 Ivory Feb 2011 B2
7884819 Kuesel Feb 2011 B2
7886045 Bates Feb 2011 B2
7890623 Bates Feb 2011 B2
7893936 Shearer Feb 2011 B2
7904829 Fox Mar 2011 B2
7921128 Hamilton, II Apr 2011 B2
7940265 Brown May 2011 B2
7945620 Bou-Ghannam May 2011 B2
7945802 Hamilton, II May 2011 B2
7970837 Lyle Jun 2011 B2
7970840 Cannon Jun 2011 B2
7985138 Acharya Jul 2011 B2
7990387 Hamilton, II Aug 2011 B2
7996164 Hamilton, II Aug 2011 B2
8001161 Finn Aug 2011 B2
8004518 Fowler Aug 2011 B2
8005025 Bodin Aug 2011 B2
8006182 Bates Aug 2011 B2
8013861 Hamilton, II Sep 2011 B2
8018453 Fowler Sep 2011 B2
8018462 Bhogal Sep 2011 B2
8019797 Hamilton, II Sep 2011 B2
8019858 Bauchot Sep 2011 B2
8022948 Garbow Sep 2011 B2
8022950 Brown Sep 2011 B2
8026913 Garbow Sep 2011 B2
8028021 Reisinger Sep 2011 B2
8028022 Brownholtz Sep 2011 B2
8037416 Bates Oct 2011 B2
8041614 Bhogal Oct 2011 B2
8046700 Bates Oct 2011 B2
8051462 Hamilton, II Nov 2011 B2
8055656 Cradick Nov 2011 B2
8056121 Hamilton, II Nov 2011 B2
8057307 Berstis Nov 2011 B2
8062130 Smith Nov 2011 B2
8063905 Brown Nov 2011 B2
8070601 Acharya Dec 2011 B2
8082245 Bates Dec 2011 B2
8085267 Brown Dec 2011 B2
8089481 Shearer Jan 2012 B2
8092288 Theis Jan 2012 B2
8095881 Reisinger Jan 2012 B2
8099338 Betzler Jan 2012 B2
8099668 Garbow Jan 2012 B2
8102334 Brown Jan 2012 B2
8103640 Lo Jan 2012 B2
8103959 Cannon Jan 2012 B2
8105165 Karstens Jan 2012 B2
8108774 Finn Jan 2012 B2
8113959 De Judicibus Feb 2012 B2
8117551 Cheng Feb 2012 B2
8125485 Brown Feb 2012 B2
8127235 Haggar Feb 2012 B2
8127236 Hamilton, II Feb 2012 B2
8128487 Hamilton, II Mar 2012 B2
8131740 Cradick Mar 2012 B2
8132235 Bussani Mar 2012 B2
8134560 Bates Mar 2012 B2
8139060 Brown Mar 2012 B2
8139780 Shearer Mar 2012 B2
8140340 Bhogal Mar 2012 B2
8140620 Creamer Mar 2012 B2
8140978 Betzler Mar 2012 B2
8140982 Hamilton, II Mar 2012 B2
8145676 Bhogal Mar 2012 B2
8145725 Dawson Mar 2012 B2
8149241 Do Apr 2012 B2
8151191 Nicol, II Apr 2012 B2
8156184 Kurata Apr 2012 B2
8165350 Fuhrmann Apr 2012 B2
8171407 Huang May 2012 B2
8171408 Dawson May 2012 B2
8171559 Hamilton, II May 2012 B2
8174541 Greene May 2012 B2
8176421 Dawson May 2012 B2
8176422 Bergman May 2012 B2
8184092 Cox May 2012 B2
8184116 Finn May 2012 B2
8185450 McVey May 2012 B2
8185829 Cannon May 2012 B2
8187067 Hamilton, II May 2012 B2
8199145 Hamilton, II Jun 2012 B2
8203561 Carter Jun 2012 B2
8214335 Hamilton, II Jul 2012 B2
8214433 Dawson Jul 2012 B2
8214750 Hamilton, II Jul 2012 B2
8214751 Dawson Jul 2012 B2
8217953 Comparan Jul 2012 B2
8219616 Dawson Jul 2012 B2
8230045 Kawachiya Jul 2012 B2
8230338 Dugan Jul 2012 B2
8233005 Finn Jul 2012 B2
8234234 Shearer Jul 2012 B2
8234579 Do Jul 2012 B2
8239775 Beverland Aug 2012 B2
8241131 Bhogal Aug 2012 B2
8245241 Hamilton, II Aug 2012 B2
8245283 Dawson Aug 2012 B2
8261339 Aldridge Sep 2012 B2
8265253 D Amora et al. Sep 2012 B2
8310497 Comparan Nov 2012 B2
8334871 Hamilton, II Dec 2012 B2
8360886 Karstens Jan 2013 B2
8364804 Childress Jan 2013 B2
8425326 Chudley Apr 2013 B2
8442946 Hamilton, II May 2013 B2
8506372 Chudley Aug 2013 B2
8514249 Hamilton, II Aug 2013 B2
8554841 Kurata Oct 2013 B2
8607142 Bergman Dec 2013 B2
8607356 Hamilton, II Dec 2013 B2
8624903 Hamilton, II Jan 2014 B2
8626836 Dawson Jan 2014 B2
8692835 Hamilton, II Apr 2014 B2
8721412 Chudley May 2014 B2
8827816 Bhogal Sep 2014 B2
8838640 Bates Sep 2014 B2
8849917 Dawson Sep 2014 B2
8911296 Chudley Dec 2014 B2
8992316 Smith Mar 2015 B2
9083654 Dawson Jul 2015 B2
9152914 Haggar Oct 2015 B2
9205328 Bansi Dec 2015 B2
9286731 Hamilton, II Mar 2016 B2
9299080 Dawson Mar 2016 B2
9364746 Chudley Jun 2016 B2
9525746 Bates Dec 2016 B2
9583109 Kurata Feb 2017 B2
9682324 Bansi Jun 2017 B2
9764244 Bansi Sep 2017 B2
9789406 Marr Oct 2017 B2
9808722 Kawachiya Nov 2017 B2
9934273 MacCarthaigh Apr 2018 B1
11640474 Song May 2023 B2
20090113448 Smith Apr 2009 A1
20140344725 Bates Nov 2014 A1
20160191671 Dawson Jun 2016 A1
20200076892 Surcouf Mar 2020 A1
20220029989 Mathur Jan 2022 A1
Foreign Referenced Citations (71)
Number Date Country
768367 Mar 2004 AU
2005215048 Oct 2011 AU
2143874 Jun 2000 CA
2292678 Jul 2005 CA
2552135 Jul 2013 CA
1334650 Feb 2002 CN
1202652 Oct 2002 CN
1141641 Mar 2004 CN
1494679 May 2004 CN
1219384 Sep 2005 CN
1307544 Mar 2007 CN
100407675 Jul 2008 CN
100423016 Oct 2008 CN
100557637 Nov 2009 CN
101001678 May 2010 CN
101436242 Dec 2010 CN
101801482 Dec 2014 CN
668583 Aug 1995 EP
0627728 Sep 2000 EP
0717337 Aug 2001 EP
0679977 Oct 2002 EP
0679978 Mar 2003 EP
0890924 Sep 2003 EP
1377902 Aug 2004 EP
0813132 Jan 2005 EP
1380133 Mar 2005 EP
1021021 Sep 2005 EP
0930584 Oct 2005 EP
0883087 Aug 2007 EP
1176828 Oct 2007 EP
2076888 Jul 2015 EP
2339938 Oct 2002 GB
2352154 Jul 2003 GB
3033956 Apr 2000 JP
3124916 Jan 2001 JP
3177221 Jun 2001 JP
3199231 Aug 2001 JP
3210558 Sep 2001 JP
3275935 Feb 2002 JP
3361745 Jan 2003 JP
3368188 Jan 2003 JP
3470955 Sep 2003 JP
3503774 Dec 2003 JP
3575598 Jul 2004 JP
3579823 Jul 2004 JP
3579154 Oct 2004 JP
3701773 Oct 2005 JP
3777161 Mar 2006 JP
3914430 Feb 2007 JP
3942090 Apr 2007 JP
3962361 May 2007 JP
4009235 Sep 2007 JP
4225376 Dec 2008 JP
4653075 Dec 2010 JP
5063698 Aug 2012 JP
5159375 Mar 2013 JP
5352200 Nov 2013 JP
5734566 Jun 2015 JP
117864 Aug 2004 MY
55396 Dec 1998 SG
2002073457 Sep 2002 WO
20020087156 Oct 2002 WO
2004086212 Oct 2004 WO
2005079538 Sep 2005 WO
2007101785 Sep 2007 WO
2008037599 Apr 2008 WO
2008074627 Jun 2008 WO
2008095767 Aug 2008 WO
2009037257 Mar 2009 WO
2009104564 Aug 2009 WO
2010096738 Aug 2010 WO
Related Publications (1)
Number Date Country
20240126898 A1 Apr 2024 US
Provisional Applications (1)
Number Date Country
63131944 Dec 2020 US
Continuations (1)
Number Date Country
Parent 17563155 Dec 2021 US
Child 18394678 US