Data protection hub

Information

  • Patent Grant
  • 9547769
  • Patent Number
    9,547,769
  • Date Filed
    Wednesday, July 3, 2013
    11 years ago
  • Date Issued
    Tuesday, January 17, 2017
    7 years ago
Abstract
Embodiments of the invention broadly described, introduce systems and methods for protecting data at a data protection hub using a data protection policy. One embodiment of the invention discloses a method for protecting unprotected data. The method comprises receiving a data protection request message comprising unprotected data and one or more policy parameters, determining a data protection transformation using the policy parameters, performing the data protection transformation on the unprotected data to generate protected data, and sending the protected data.
Description
BACKGROUND

Ensuring security of data at rest and in transit continues to be a necessity. Implementing data security typically requires a high degree of investment in hardware, software, and the appropriate level of expertise to assess the security needs of the data that needs protection. Government or industry compliance standards may further increase the complexity and cost to implement data security.


There can be many different organizations that have different data protection requirements. Different organizations may include different divisions, projects and groups. Typically, each organization implements data security independently based on their needs. However, implementing independent solutions for each organization leads to a number of problems.


First, the differences in the security implementations inhibit flow of data securely between the different organizations. This is especially true if the flow of data occurs between organizations within a single corporation. Second, each implementation of data security may require expensive investment on the part of the organization. Third, varying implementations between the different organizations with varying level of expertise may lead to implementations that do not adhere to industry, national, and international standards and may expose the data to potential compromise.


Furthermore, some small merchants may not have the ability to access cryptographic services or other security solutions.


Embodiments of the invention address these and other problems.


SUMMARY

Embodiments of the invention broadly described, introduce systems and methods for protecting data at a data protection hub using a data protection policy.


One embodiment of the invention discloses a computer-implemented method for protecting data. The method comprises receiving a data protection request message comprising unprotected data and one or more policy parameters, determining a data protection transformation using the policy parameters, performing the data protection transformation on the unprotected data to generate protected data, and sending the protected data.


One embodiment of the invention discloses a data protection hub. The data protection hub comprises a processor and a non-transitory computer-readable storage medium, comprising code executable by the processor for implementing a method comprising receiving a data protection request message comprising unprotected data and one or more policy parameters, determining a data protection transformation using the policy parameters, performing the data protection transformation on the unprotected data to generate protected data, and sending the protected data.


One embodiment of the invention discloses a computer-implemented method for protecting unprotected data. The method comprises sending, by a processor, a data protection request message comprising unprotected data and one or more policy parameters, and receiving, by the processor, protected data, wherein the protected data is generated by a data protection transformation, wherein the data protection transformation is determined using the policy parameters.


Further details regarding embodiments of the invention can be found in the Detailed Description and the Figures.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 shows a system according to an embodiment of the invention.



FIG. 2 shows one potential embodiment of a data protection hub.



FIG. 3 shows one potential embodiment of a federated data protection hub.



FIG. 4 shows a method for a data protection hub to receive and process a data protection policy.



FIG. 5 shows a flow chart illustrating an exemplary data protection policy.



FIG. 6 shows a method 600 for generating protected data at a data protection hub 200.



FIG. 7 shows a method for generating protected data by hashing the unprotected data.



FIG. 8 shows a method for generating protected data by masking the unprotected data.



FIG. 9 shows a method for generating protected data by tokenizing the unprotected data.



FIG. 10 shows a method for generating protected data by encrypting the unprotected data.



FIG. 11 shows a method for deprotecting data using a data protection hub.



FIG. 12 shows a method for retrieving unprotected data associated with a token using a data protection hub.



FIG. 13 shows a method for decrypting protected data using a data protection hub.



FIG. 14 shows a method for migrating protected data using a data protection hub.



FIG. 15 shows an exemplary system for using a data protection hub according to an embodiment of the invention.



FIG. 16 shows an example of a payment device in the form of a card.



FIG. 17 is a high level block diagram of a computer system that may be used to implement any of the entities or components described for embodiments of the invention.





DETAILED DESCRIPTION

Prior to discussing embodiments of the invention, descriptions of some terms may be helpful in understanding embodiments of the invention.


The term “server computer” may include a powerful computer or cluster of computers. For example, the server computer can be a large mainframe, a minicomputer cluster, or a group of servers functioning as a unit. In one example, the server computer may be a database server coupled to a Web server. The server computer may be coupled to a database and may include any hardware, software, other logic, or combination of the preceding for servicing the requests from one or more client computers. The server computer may comprise one or more computational apparatuses and may use any of a variety of computing structures, arrangements, and compilations for servicing the requests from one or more client computers.


The term “unprotected data” may include any data that has not been protected, such that unauthorized use by an unauthorized person could cause harm to a person associated with the unprotected data. For example, unprotected data may include a primary account number for a credit card. If an unauthorized person uses the credit card number, potential harm could be done to the authorized person (e.g., the credit card holder) that is associated with the primary account number. Unprotected data may not have been transformed, obfuscated, encrypted, or otherwise protected by a data protection hub. Typically, unprotected data may be stored in cleartext, so that the information represented by the unprotected data is readily apparent. Examples of unprotected data may include: a numeric value representing an age of a customer, a string literal comprising a user's password or primary account number associated with a payment account, or a serialized object value (e.g., in JSON) comprising program state. In some cases, unprotected data may store sensitive information, so transmission of unprotected data over insecure communication networks and storage of unprotected data in insecure databases should be avoided.


The term “protected data” may include any data that has been protected by a data protection hub, typically by using a data protection transformation. In some embodiments, protected data may be considered safe to send over insecure communications networks or store in insecure databases.


A “data protection transformation” may include any transformation of unprotected data to protected data, such that information represented by the unprotected data is not readily apparent from the protected data. A data protection transformation may be a cryptographic operation such as encryption or cryptographic hashing. Alternately, the data protection algorithm may use other operations such as tokenization or data masking. In some embodiments, a data protection transformation may be a reversible transformation (e.g., tokenization or encryption), so that the original unprotected data may be derived using the protected data. In other embodiments, the data protection transformation may be an irreversible or substantially irreversible process (e.g., hashing or data masking), so that the original unprotected data may not be derived from the protected data.


The term “deprotected data” may include any data that is substantially unprotected but was previously protected. For example, a primary account number for a payment card may be “unprotected data,” while the subsequent encryption of that primary account number may product an encrypted primary account number which is an example of “protected data” and the sequent decryption of the encrypted primary account number may be an example of “deprotected data.” In this example, the “deprotected data” may also be “unprotected data.”


The term “transformation data parameters” may include any parameters (e.g., secrets, or other data) used to perform a data protection transformation. For example, transformation data parameters for a hashing transformation may include the hashing algorithm (e.g., SHA-256 or SHA-512), a number of rounds of hashing to perform, and an output format of the hash. In some embodiments, transformation data parameters may also include information used to perform a data deprotection transformation. For example, transformation data parameters for an encryption algorithm may include a symmetric encryption algorithm (e.g., AES), and the encryption key, so that unprotected data may be determined by decrypting the protected data using the encryption key.


The term “data deprotection transformation” may include any transformation of protected data to unprotected data, such that the unprotected data was used to generate the protected data. A data deprotection transformation may typically be the inverse operation for the data protection transformation used to generate the protected data. For example, if the protected data was generated by symmetric encryption using a secret key, then the data deprotection transformation may be decryption using the secret key. The term “deprotection” may include any operation or transformation wherein unprotected data is determined from protected data.


The term “data protection hub” may include any suitable server computer or other computing device operable to receive requests for data protection and provide protected data. An exemplary data protection hub may determine protected data from received unprotected data and policy parameters. An exemplary data protection hub according to some embodiments of the invention is shown in FIG. 2. An exemplary federated data protection hub according to some embodiments of the invention is shown in FIG. 3.


The term “hub management computer” may include any suitable computing device used to generate or send a data protection policy file to a data protection hub. The hub management computer may connect to a data protection hub using a console, communications link, or any other suitable medium.


The term “data protection policy” may include any suitable procedure that is used to determine a data protection transformation for unprotected data. For example, the data protection policy may be used to determine that a user's password should be hashed, a user's social security number should be encrypted, and a user's credit card number should be tokenized. The data protection policy may also indicate the algorithms used to perform the data protection transformation (e.g, AES, SHA-256, RSA, etc.).


In some embodiments, the data protection policy may be determined using “policy parameters.” In various embodiments of the invention, policy parameters may indicate a sensitivity of the unprotected data, a use case for the unprotected data, or a size of the unprotected data. For example, highly sensitive data such as a user's social security number may be encrypted with a 2048-bit key, whereas less sensitive data such as the user's first name may be encrypted with a 512-bit key. In another example, unprotected data which is intended to be sent over untrusted networks may be tokenized and stored at the data protection hub for recovery.


In some embodiments of the invention, a data protection policy may include some or all aspects of a compliance standard or data protection law, such as the Payment Card Industry Data Security Standard (PCI DSS), Health Insurance Portability and Accountability Act (HIPAA), Control Objectives for Information and Related Technology (COBIT), and the European Data Protection Directive.


The term “data protection policy file” may include a data file used to store a data protection policy. A data protection policy file may be in any suitable format, such as a markup language (e.g., XML), a logical language (e.g., Prolog), or an imperative language (e.g., C++).


The term “policy schema” may include any specification or model used to indicate the policy parameters associated with a data protection policy. For example, a policy schema may indicate three levels of data sensitivity for unprotected data: personal account information (PAI), personally identifiable data (PII), and sensitive authentication information (SAD. Similarly, the policy schema may indicate two values indicating the reversibility of the transformation: a first value indicating that the data protection transformation should be reversible, and a second value indicating that the transformation should be irreversible. Accordingly, policy parameters would comprise a level of sensitivity and a value for the reversibility of the transformation.


In some embodiments of the invention, the policy schema may be defined as an XML schema. In such embodiments, policy parameters may be defined using an XML document conforming to the XML schema.


The term “client credentials” may include any data provided which allows a data protection hub to verify that a client computer is authorized to deprotect the protected data. Client credentials may comprise a user name and password, a digital signature, or any other suitable data.


The term “data migration” may include any process in which protected data is migrated from a first computing device to a second computing device, such that the second computing device may operate the data protection hub to perform data deprotection transformations on the protected data.


Embodiments of the invention provide for many technical advantages. For example, embodiments of the invention reduce the cost of developing, deploying, and maintaining multiple data protection services by providing a centralized data protection hub that may be utilized by any number of client computers for a variety of applications. The data protection hub may receive unprotected data and policy parameters in a generic format from a client computer, and appropriately perform a data protection transformation in order to generate protected data. Thus, each application may protect sensitive data without needing to implement or deploy independent instances of data protection services. This reduces code complexity (e.g., because only a single codebase needs to be maintained to implement data protection features), deployment costs (e.g., because a single physical or virtual server may provide services to several applications), and time to deploy new applications (e.g., because an existing data protection hub may be used to provide data protection to a new application with minimal changes).


Embodiments of the invention provide for the additional technical advantage of generating protected data in compliance with data protection policies specified by a number of heterogenous users. It may be desirable to maintain a consistent data protection policy encompassing multiple applications for several reasons. For example, industry standards or governmental regulations may place constraints on protected data or data protection transformations used to generate protected data. These constraints may depend on various characteristics of the data, such as the nature of sensitive information stored or the potential for compromise associated with the protected data. The data protection hub may receive these characteristics as policy parameters from a client computer. The data protection hub may then determine a data protection transformation using the policy parameters. This enables applications to comply with a data protection policy by describing the characteristics of the data to be protected, without needing to independently develop software logic implementing the data protection policy. This further reduces software complexity and duplication of functionality between multiple applications conforming to a data protection policy. In addition, if a data protection policy is to be changed, the change may only need to be performed on the data protection hub; the change may be transparent to applications using the data protection hub as long as the policy parameters remain the same. This improves software reliability and longevity, since a data protection policy change at the data protection hub may not require any changes to applications using the data protection hub.


Embodiments of the invention provide for the additional technical advantage of providing an interface whereby a data protection policy may be reconfigured on-demand by a hub management computer. This allows the hub management computer to modify the data protection transformations to comply with new standards or regulations established after deployment of the data protection hub. In one example, a collision vulnerability may be found in a hashing algorithm used in a data protection transformation. Accordingly, the data protection policy may be modified to use a more secure hashing algorithm. In another example, a governmental organization may publish data protection standards specifying a longer minimum key length for data protection transformations relating to encryption. Accordingly, the data protection policy may be modified to use the longer key length. In a third example, the data protection policy may be modified to use higher-complexity data protection transformations (e.g., SHA-3 instead of SHA-2) as time passes and higher performance computing devices become available. Thus, allowing the data protection policy to be reconfigured allows data protection to be dynamically updated according to current needs, improving data security without requiring any changes to the data protection hub software.


Embodiments of the invention provide for the additional technical advantage of facilitating migration of protected data between client computers and applications. If data protection for each application is managed separately, then in order to migrate protected data from a first application to a second application, the data may be transformed to unprotected data by the first application, transmitted to the second application, and re-protected by the second application. This is inefficient, because it requires de-protecting and re-protecting the data and is insecure because the data is transmitted in an unprotected state. Alternately, the format of the protected data and any secret information (e.g., encryption or decryption keys, mappings between tokens and data values, etc.) may be transmitted with the protected data to the second application. However, this is also inefficient because a mapping, and insecure because transmitting the secret information may be a security liability.


For example, a first application may encrypt authentication data using the Advanced Encryption Standard (AES), and a second application may encrypt authentication data using the Data Encryption Standard (DES). If authentication data is to be shared between the applications, both applications may be configured to parse data encrypted using either algorithm. This increases software complexity and data storage requirements.


In contrast, embodiments of the invention enable secure and efficient migration of protected data. In some embodiments, a first application may send a data migration request message to a data protection hub indicating the data to migrate. Then, the first application may send the protected data to the second application. Subsequently, the second application may use the protected data, for example by sending the protected data to the data protection hub for deprotection. Thus, security is maintained because any secret information remains at rest at the data protection hub. Further, the protected data does not need to deprotected and re-protected, improving the efficiency of the process.


The above examples highlight only a few of the advantages of protecting data at a data protection hub using a data protection policy.


I. Exemplary Data Protection Hub Systems

A system 100 for use of a data protection hub can be seen in FIG. 1. The system 100 includes a client computer A 101, client computer B 102, client computer C 103, client computer D 104, communications network 105, data protection hub 200, and hub management computer 107. As shown, data protection hub 200 may be in communication with each of the client computers 101, 102, 103, 104 and may service the requests to and from these client computers. Communications network 105 may facilitate data transmissions between client computers 101-104 and data protection hub 200. Client computers 101-104 may run one or more applications, which may connect to data protection hub 200 in order to perform various data protection services, such as protection of unprotected data, deprotection of protected data, and data migration.


A suitable communications network 105 may be any one and/or the combination of the following; a direct interconnection; the Internet; a Local Area Network (LAN); a Metropolitan Area Network (MAN); an Operating Missions as Nodes on the Internet (OMNI); a secured custom connection; a Wide Area Network (WAN); a wireless network (e.g., employing protocols such as, but not limited to a Wireless Application Protocol MAP), I-mode, and/or the like); and/or the like.



FIG. 2 shows one potential embodiment of a data protection hub 200. As depicted in FIG. 2, the data protection hub 200 may comprise several implantation layers, such as data protection abstraction interface 210, policy decision layer 220, service layer 230, data layer 240, and infrastructure layer 250. Data protection hub 200 may also comprise a management interface 201, a policy management engine 202, and a resource management engine 260. The various layers, interfaces, and engines may be embodied by computer code, residing on computer readable media.


Data protection abstraction interface 210 may comprise one or more APIs, protocols, or schemas used by client computers to send data protection requests, data deprotection requests, data migration requests, or other communication. Data protection abstraction interface 210 may be configured to parse and process such communication. In addition, data protection abstraction interface 210 may be used to send protected data, unprotected data, or other data to the various client computers such as client computer 101. Data protection abstraction interface 210 may typically comprise one or more networking or communication apparatuses which enable the data protection hub 200 to communicate with communications network 105. Typically, data protection requests and data deprotection requests from a client computer 101 may be forwarded to policy decision layer 220 for further processing.


Policy decision layer 220 may comprise program logic or other software configured to determine a data protection in accordance with a data protection policy. Policy decision layer 220 may determine policy parameters for a data protection request and communicate with policy management engine 202 to determine the corresponding data protection transformation. In some embodiments, policy decision layer 220 may determine a module in service layer 230 to be used.


Service layer 230 may comprise a plurality of service modules configured to provide various data protection, deprotection, or migration services.


For example, hashing as a service module 231 may be configured to hash unprotected data. Hashing as a service module 231 interface with hashing engine 251. Hashing may refer to any algorithm or function that takes as input data and returns a fixed-size bit string (i.e., a hash), or as otherwise known in the art. In various embodiments, hashing may be performed using SHA-2, MD5, bcrypt, Blowfish, or any other suitable method. For example, a user's password (e.g., “hunter2”) may be hashed to generate a trusted hash (e.g., “a3b8919fcb9d2”). Subsequently, a password entered by the user may be compared to the trusted hash in order authenticate the user. In some embodiments, hashing as a service module 231 may be implemented in accordance with the method described in FIG. 7.


Data migration as a service module 232 may be configured to perform data migration operations on unprotected data. Data migration may include any process in which protected data is migrated from a first computing device to a second computing device, such that the second computing device may operate the data protection hub to perform data deprotection transformations on the protected data. For example, data migration as a service module 232 may be used to migrate a sensitive user account database between a first client computer and a second client computer. In some embodiments, data migration as a service module 232 may be implemented in accordance with FIG. 14.


Data masking as a service module 233 may be configured to perform masking of unprotected data. Data masking may refer to any algorithm or method for de-contexting or removing sensitive information from input data, or as otherwise known in the art. In some embodiments, sensitive data may be masked by removing parts or elements of the data. For example, a standard 16-digit credit card personal account number (PAN) (e.g., “4061-7240-6172-4061”) may be masked by removing the seventh digit through the twelfth digit (e.g., “4061-72XX-XXXX-4061”). Thus, an attacker would not be able to cause harm with the remaining digits. In addition, in some embodiments, sensitive data may be protected by moving the data to a different context. For example, a PAN, account expiration date, and card verification value (e.g., CVV2) may be decontexted by appending the CVV2 to an address field associated with the card holder. Thus, an attacker not knowing the new context of the data would not be able to understand the de-contexted data.


In some embodiments, data masking as a service module 233 may communicate with policy decision layer 220 to determine sensitive data fields in the unprotected data. In addition, data masking as a service module 233 may use scrubbing/masking engine 252 to mask the unprotected data. In some embodiments, data masking as a service module 23 may be implemented in accordance with FIG. 8.


Tokenization as a service module 234 may be configured to perform tokenization of unprotected data and detokenization of protected data. Tokenization may include substituting a sensitive data element with a non-sensitive element and saving a mapping between the elements, or as otherwise known in the art. An authorized user may subsequently perform detokenization to retrieve the sensitive data using the token.


Any suitable data may be tokenized. For example, a sensitive 16-digit PAN (e.g., “4061-7240-6172-4061”) may be tokenized to produce a token (e.g., “a349ba0530a41169083”). In some circumstances, the token may be used as a substitute for the PAN. For example, a user may use the token in order to pay for a transaction instead of providing the PAN. A trusted party, such as the issuer associated with the PAN, may subsequently detokenize the data to retrieve the PAN and process the payment transaction.


In some embodiments, tokenization as a service module 233 may communicate with policy decision layer 220 in order to determine a tokenization format for the protected data. In addition, tokenization as a service module 234 may use token generator 253 to generate cryptographically secure tokens. A mapping between tokens and the unprotected data may be stored in data storage 241. In some embodiments, tokenization as a service module 234 may be implemented in accordance with FIGS. 9 and 12.


Encryption as a service module 235 may be configured to perform encryption of unprotected data and decryption of protected data. Encryption may include any encoding of information so that unauthorized users cannot read it, but authorized users can, or as otherwise known in the art. In various embodiments of the invention, encryption may be perform using symmetric keys, or public/private key pairs. For example, a user's date of birth (e.g., “Jan. 2, 1934”) may be encrypted using an encryption key to produce encrypted data (e.g., 0x34A69F), so that the encrypted data may be unreadable. An authorized user may then decrypt the data, so that the date of birth is readable.


Encryption as a service module 235 may communicate with policy decision layer 220 to determine an encryption format for the protected data and generate an encryption key. In addition, encryption as a service module 235 may encrypt the unprotected data using crypto provider 254. Encryption as a service module 235 may store the encryption key using key manager 242. In some embodiments, encryption as a service module 235 may be implemented in accordance with FIGS. 10 and 13.


Data layer 240 may be configured to store any data relating to the operation of data protection hub. Data layer 240 may comprise a data storage module 241 and a key manager module 242.


Data storage module 241 may store any suitable data stored by data protection hub 200. For example, in various embodiments of the invention, data storage module 241 may store data relating to the data protection policy, transformation data parameters, mappings between tokens and unprotected data, protected data, or any other suitable data.


Key manager 242 may be configured to store and retrieve any suitable data relating to encryption keys. For example, key manager 242 may store symmetric keys, public/private key pairs, or keys used to generate a keyed hash. In some embodiments, key manager 242 may manage one or more keys stored in secure key store 256. In such embodiments, key manager 242 may retrieve a key from a secure key store 256 when requested by the service layer 230.


Infrastructure layer 250 may comprise any software or hardware elements used to perform data protection operations. For example, infrastructure layer 250 may comprise a hashing engine 251, scrubbing/masking engine 252, token generator module 253, crypto provider module 254, true random number generator (T-RNG) module 255, or secure key store 256.


Hashing engine 251 may comprise any software or hardware element used to perform hashing operations. For example, hashing engine 251 may comprise a general-purpose graphics processing unit (GPGPU) or a CPU running hashing software, or specialized hardware configured to perform hashing operations. Typically, hashing as a service module 231 may communicate with hashing engine 251 in order to compute a hash value in a given format.


Scrubbing/masking engine 252 may comprise any software or hardware element used to perform scrubbing or masking of unprotected data. Typically, scrubbing/masking engine 252 may mask sensitive data fields specified by data masking as a service module 233.


Token generator module 253 may comprise any software or hardware element used to tokenize unprotected data. For example, token generator 253 may comprise a physical device configured to generate a sequence of cryptographic tokens. Alternately, tokens may be generated by software running in the data protection hub 200. Typically, token generator module 253 may generate a token used by assigned to an unprotected data value by tokenization as a service module 234.


Crypto provider module 254 may comprise any software or hardware element used to perform cryptographic operations. For example, crypto provider module 254 may comprise a secure cryptoprocessor such as a trusted platform module (TPM). Crypto provider module 254 may be operable to generate keys, encrypt or decrypt data, cryptographically hash data, or any other suitable operation. In some embodiments, crypto provider module 254 may be in communication with various other elements in the infrastructure layer 250, such as hashing engine 251 (e.g., when computing an HMAC), T-RNG provider 255 (e.g., when generating a random key), or secure key store 256 (e.g., when decrypting data using a key stored in the secure key store). Typically, crypto provider module 254 may be used by encryption as a service module 235 to encrypt unprotected data or decrypt protected data, or by hashing as a service 231 to compute hash-based message authentication codes (HMACs).


True random number generator (T-RNG) provider 255 may comprise any software or hardware element used to generate random or pseudorandom numbers. For example, T-RNG provider may comprise a hardware random number generator that generates random numbers from a physical phenomenon. In some embodiments of the invention, T-RNG provider may be used by crypto provider module 254 or secure key store 256 in order to ensure that a generated key is statistically random.


Secure key store 256 may comprise any software or hardware element to store cryptographic keys. For example, secure key store 256 may comprise a hardware security module (HSM) operable to manage and store cryptographic keys. In some embodiments of the invention, secure key store 256 and crypto provider 254 may be integrated, so that cryptographic operations using keys stored in secure key store 256 may be performed using crypto provider 254. Typically, secure key store 256 may be used by encryption as a service module 235 to store encryption keys, or by tokenization as a service module 234 to store tokenization root keys.


In some embodiments of the invention, the various elements in infrastructure layer 250 may be duplicated. For example, the number of crypto providers 254 and secure key stores 256 may be scaled depending on the number of keys stored by data protection hub 200.


Data protection hub 200 may also comprise a management interface 201, a policy management engine 202.


Management interface 201 may comprise any suitable API, protocol, or schema used by a hub management computer 106 to interface with data protection hub 200. For example, management interface 201 may comprise a web interface such as a web site, a console, or a management API.


Policy management engine 202 may be configured to manage data protection policy settings associated with the data protection hub 200. Data protection policies may include any suitable procedures that are used to determine a data protection transformation for unprotected data. For example, the data protection policy may be used to determine an appropriate data protection transformation for a user's password (e.g., hashing), a user's social security number (e.g., encryption), or a user's credit card number (e.g., tokenization). The data protection policy may also indicate the algorithms used to perform the data protection transformation (e.g, AES, SHA-256, RSA, etc.).


In some embodiments, the data protection policy may be determined using “policy parameters.” In various embodiments of the invention, policy parameters may indicate a sensitivity of the unprotected data, a use case for the unprotected data, or a size of the unprotected data. For example, highly sensitive data such as a user's social security number may be encrypted with a 2048-bit key, whereas less sensitive data such as the user's first name may be encrypted with a 512-bit key. In another example, unprotected data which is intended to be sent over untrusted networks may be tokenized and stored at the data protection hub for recovery.


Policy management engine 202 may be parse a data protection policy file and publish a policy schema comprising policy parameters to be used by client computers.


Resource management engine 260 may be configured to manage access to computing resources for various layers, services, modules, and other elements in the data protection hub 200. In some embodiments, the data protection hub may be deployed to a cloud or virtualized environment, so that layers and/or modules may be implemented using virtual machines (VMs). In such embodiments, resource management engine 260 may be used to ensure each element has sufficient resources to operate. For example, if a VM comprising the encryption as a service module 235 is being heavily utilized, resource management engine 260 may increase the number of cores or amount main memory allotted to the VM. Alternately, resource management engine 260 may spawn another instance of the VM, so that utilization of the encryption as a service module 235 may be split among the VMs.


Resource management engine 260 may also be responsible for managing physical hardware or other infrastructure associated with the data protection hub 200. For example, if the HSMs implementing the secure key store 256 are approaching maximum capacity, resource management engine 260 may move VMs to physical servers with empty HSMs or may indicate to IT or other personnel that additional HSMs need to be procured.


II. Exemplary Federated Hub Systems

In some embodiments of the invention, the data protection hub 200 may be distributed across multiple network environments. For example, data protection hub 200 may service client computers in different network tiers, subnets, or contained inside different firewalls. In such embodiments, the same data protection hub 200 may be federated; one or more instantiations of the data protection hub 200 may be run locally to a network environment, so that the instantiation services local client computers. The federated instantiations may maintain communication with each other, so that data may be synchronized between the instantiations. Thus, the federated instantiations of the data protection hub 200 may behave like a single data protection hub 200 while providing local access to client computers in different network environments.



FIG. 3 shows one potential embodiment of a federated data protection hub. The federated data protection hub comprises data protection hub A 313, data protection hub B 323, data protection hub C 333, and data protection hub D 343. In some embodiments, each data protection hub instantiation 313, 323, 333, and 343 may have the structure described for FIG. 2, and the descriptions are herein incorporated. Each data protection hub instantiation in communication with a corresponding client computer and is enclosed in a separate network firewall. For example, firewall A 311 includes client computer 312 and data protection hub 313. In addition, each data protection hub instantiation is in communication with the other instantiations.


III. Exemplary Data Protection Policy Methods


FIG. 4 shows a method for a data protection hub 200 to receive and process a data protection policy. The method of FIG. 4 may typically be performed any time the data protection policy is to be changed.


At step 401, hub management computer 106 defines a data protection policy file. A data protection policy file may include a data file used to store a data protection policy. The hub management computer 106 may define the data protection policy file using any suitable means. For example, the data protection policy file may be defined in accordance with one or more industry standards or governmental regulations. In some embodiments, a data protection policy file may indicate a conditional relationship between various policy parameters and a corresponding data protection transformation. An exemplary data protection policy is shown in FIG. 5.



FIG. 5 shows a flow chart illustrating an exemplary data protection policy 500. The data protection policy 500 comprises a plurality of conditional expressions 501-507 used to determine a plurality of data protection transformations 508-513. The data protection policy may be used to determine a data protection transformation based on policy parameters associated with data to be protected. An exemplary evaluation of the data protection policy 500 is shown below.


At conditional step 501, data protection hub 200 determines if the data to be protected is personal account information. As used herein, “personal account information” may include data relating to an account operated by a user. For example, for a credit card account, personal account information may include a primary account number (PAN), an expiration date of a credit card, and a name printed on the card. If the data to be protected is personal account information, the flow proceeds to step 504. Otherwise, the flow proceeds to step 502.


At conditional step 504, data protection hub 200 determines if the data needs to be re-presented. As used herein, data needs to be “re-presented” if the original, unprotected data must be derivable from the protected data. For example, an application may store a user's social security number in order to verify that the social security number on file matches a number provided by a user. In this case, the user's social security number may not need to be re-presented, since the actual value of the SSN does not need to be known to perform verification. In contrast, a card's expiration date may need to be re-presentable so that a new card can be printed and mailed in advance of the expiration date of the existing card. If the data to be protected does not need to be re-presented, at step 508 the data protection transformation is determined to be hashing using the SHA-2 hashing algorithm. Otherwise, the flow proceeds to step 507.


At conditional step 507, data protection hub 200 determines if the data will be transferred over an untrusted network. If the data will not be transferred over an untrusted network, then at step 509 the data protection transformation is determined to be encryption using a symmetric key. Otherwise, at step 510 the data protection transformation is determined to be tokenization.


At conditional step 502, data protection hub 200 determines if the data is personally identifiable information. As used herein, “personally identifiable information” (PII) may refer to information that may be used to identify a user but may not be unique to the account. For example, a home address or phone number associated with an account may be PII, since multiple accounts may share the same home address or phone number. If the data to be protected is personally identifiable information, the flow proceeds to step 505. Otherwise, the flow proceeds to step 503.


At conditional step 505, data protection computer 200 determines if the data to be protected can be de-contexted. As used herein, data may be “de-contexted” if parts of the data may be removed so that the data is no longer sensitive. For example, a 9-digit social security number may be de-contexted by keeping only the last four digits. If the data to be protected can be decontexted, at step 511 the data protection transformation is determined to be data masking. Otherwise, the flow proceeds to step 507, as described above.


At conditional step 503, data protection computer 200 determines if the data is sensitive authentication data. As used herein, “sensitive authentication data” may include any data which is used to verify the authenticity of a user. Examples of sensitive authentication data may include passwords, password recovery questions and answers, and secret keys. If the data to be protected is sensitive authentication data, the flow proceeds to step 506. Otherwise, at step 514, the data is kept in cleartext, because it is not sensitive.


At conditional step 506, data protection computer 200 determines if the data protection transformation should be reversible. As used herein, a “reversible transformation” generates protected data from unprotected data, wherein the protected data may be used to derive the unprotected data. Examples of irreversible or substantially irreversible data protection transformations may include one-way (i.e., cryptographic) hashing and data masking. Examples of reversible data protection transformations may include encryption and tokenization. In one embodiment, a data protection policy may specify that a data protection transformation should be irreversible if a compromise of the unprotected data would be considered a secret only known to the user. For example, it is generally accepted practice for a user's password to be secret to a user; thus, the password can be protected with an irreversible transformation. However, a user's password recovery security question may not be considered secret to the user; as such, the recovery question may be protected with a reversible transformation. If the transformation should be irreversible, at step 513 the data protection transformation is determined to be hashing using the bcrypt algorithm. Otherwise, at step 514, the data protection transformation is determined to be encryption using a public key associated with the user.


It should be noted that the exemplary data protection policy 500 shown in FIG. 5 is intended to be descriptive but not limiting. For example, data protection policies may comprise any suitable number and or type of conditions and data protection transformations for processing data. The types and numbers of conditions and transformations may be arbitrarily chosen by the user. In addition, data protection policies do not need to be represented as flow charts; they may be structured as tables, decision trees, using program code, or any in any other suitable format.


Returning to FIG. 4, at step 402 hub management computer 106 sends the data protection policy file to data protection hub 200 using a hub management interface 201. For example, in some embodiments, management interface 201 may comprise a web interface such as a web site connected to by client computer 101. Client computer 101 may use the website to define the data protection policy file, or upload the data protection policy file to the website.


At step 403, data protection hub 200 parses the data protection policy file to determine a data protection policy. In some embodiments, the data protection policy file may comprise an XML schema. In such embodiments, the XML schema may be parsed using any suitable means, such as a jQuery or another XML parsing library. The data protection policy may be integrated into policy management engine 202 and/or policy decision layer 220, so that any subsequent data protection and deprotection may be in accordance with the data protection policy.


At step 404, data protection hub 200 retransforms protected data using the determined data protection policy. Retransformation of the protected data may include ensuring that existing protected data is protected in accordance with the new data protection policy. For example, in some embodiments, data protection hub 200 may retrieve data stored in data storage module 241 and or key manager module 242. In one example, a new data protection policy may specify that encrypted data should be protected using 1024-bit length keys instead of 512-bit length keys as specified by the previous data protection policy. Accordingly, at step 404, data protection hub 200 may retrieve and decrypt protected data stored in data storage module 241 using 512-bit length keys stored in key manager 242. Data protection hub 200 may then generate new 1024-bit length encryption keys, re-encrypt the decrypted data, and store the encrypted data in data storage 241. Data protection hub 200 may also replace each 512-bit key with a corresponding 1024-bit key in key manager 242.


At step 405, data protection hub 200 publishes a policy schema comprising policy parameters. The policy schema may include any specification, model, or format used to indicate the policy parameters associated with a data protection policy. For example, a policy schema may indicate in an XML schema the policy parameters that may be used in accordance with the data protection policy 500. The XML schema may be published (e.g., made available to client computers such as client computers 101-104) using any suitable method. For example, data protection hub 200 may host the XML schema using a web interface (e.g., a web server). Alternately, wrapper source files, client stubs (e.g., using the SOAP framework), APIs, or libraries may be provided to client computers 101-104 in order to indicate policy parameters associated with the data protection policy.


Policy parameters for the exemplary data protection policy 500 may include logical values (e.g., yes or no) for some or all of the conditionals 501-507. For example, policy parameters may be used to indicate whether unprotected data sent to the data protection hub 200 is personal account information, whether the unprotected data needs to be re-presented, or whether the unprotected data will be transferred over an untrusted network. Policy parameters specified by a client computer 101 may be sufficient for data protection hub 200 to determine a data protection transformation for the unprotected data.


In some embodiments, the policy schema may specify one or more subsets of conditionals, wherein each subset comprises enough conditionals to determine a data protection transformation. For example, if client computer 101 indicates in policy parameters that unprotected data is not personal account information, is personally identifiable information, and can be de-contexted, then the client computer 101 may not also need to indicate whether the data will be transferred over an untrusted network (e.g., because the latter indication would not affect the determined data protection transformation). Accordingly, client computer 101 may use the policy schema to determine the necessary policy parameters to send to data protection hub 200 for some unprotected data.


IV. Exemplary Data Protection Methods


FIG. 6 shows a method 600 for generating protected data at a data protection hub 200.


At step 601, a client computer 101 sends a data protection request message to the data protection hub 200 including unprotected data and policy parameters. In one example, a client computer 101 may transmit the data protection request message for a user's credit card number. In some embodiments, the credit card number may be present with a plurality of other payment card numbers of other users. The request message may include policy parameters indicating that the unprotected data is personal account information, needs to be represented, and will be transferred over a personal network.


At step 602, data protection hub 200 determines a data protection transformation using the policy parameters. In some embodiments, policy decision layer 220 may be used to determine the data protection transformation. For the example policy parameters described above for step 601, and exemplary data protection policy 500, data protection hub 200 may determine the data protection transformation to be tokenization.


At step 603, data protection hub 200 performs the determined data protection transformation on the unprotected data to generate protected data. In some embodiments, the data protection transformation may be performed using an appropriate component of service layer 230 of data protection hub 200. For example, if the data protection transformation is tokenization on a credit card number, at step 603 tokenization as a service module 234 may use token generator module 253 to generate a secure token.


At step 604, data protection hub 200 stores transformation data parameters associated with the data protection transformation. In some embodiments, the transformation data parameters may be stored in data layer 240. For example, if the data protection transformation is tokenization on a credit card number, at step 604 tokenization as a service module 234 may store a mapping between the credit card number and the secure token in data storage module 241.


At step 605, data protection hub 200 sends the protected data to the client computer 101. In some embodiments, the protected data may be transmitted using data protection abstraction interface 210. For example, if the protected data is a secure token, data protection hub 200 may send the secure token to client computer 101.


It should be noted that method 600 is intended to be descriptive and non-limiting. For example, instead of or in addition to sending the protected data to client computer 101 at step 605, data protection hub 200 may store the protected data. Client computer 101 may subsequently request the protected data.


In addition, a data protection request message may comprise multiple elements of unprotected data, each of which may be associated with different policy parameters. For example, a data protection request message may comprise three data fields: a user's credit card number, email address, and password, and separate policy parameters for each of the data fields.



FIGS. 7-10 described below show exemplary methods for transforming unprotected data to protected data in the form of hashes, masked data, tokens, and encrypted data, respectively. In some embodiments, the method of FIGS. 7-10 may be performed after data protection hub 200 determines a data protection transformation to perform on the unprotected data (e.g., after step 602 in FIG. 6).



FIG. 7 shows a method for generating protected data by hashing the unprotected data.


At step 701, data protection hub 200 determines the data protection transformation for the unprotected data to be hashing. As used herein, “hashing” may refer to any algorithm or function that takes as input data and returns a fixed-size bit string (i.e., a hash), or as otherwise known in the art.


At step 702, data protection hub 200 determines a hashing algorithm and salt using the policy parameters. In some embodiments of the invention, a data protection policy may specify a hashing algorithm or salting methodology to use for the unprotected data. For example, data protection policy 500 shown in FIG. 5 specifies hashing using the SHA-2 or bcyrpt hashing algorithms, depending on the policy parameters associated with the unprotected data.


At step 703, data protection hub 200 determines the hash of the unprotected data using the hashing algorithm and salt. In some embodiments, data protection hub 200 may use hashing as a service module 231 and hashing engine 251 to perform hashing operations.


At step 704, data protection hub 200 stores the hashing algorithm and salt used to generate the hash. In some embodiments, the hashing algorithm and salt may be stored in data layer 240, such as in data storage module 241. The hashing algorithm and salt may be considered transformation data parameters associated with the protected data.


At step 705, data protection hub 200 sends the hash to client computer.


The method of FIG. 7 may performed for any suitable unprotected data, data protection policy, and hashing algorithm. In one example, data protection hub 200 may be used to determine the data protection transformation for a user's password of “abc123”. At step 702, in accordance with the data protection policy 500, the hashing algorithm may be determined to be bcrypt, since password is sensitive authentication data and the transformation should be irreversible. At step 703, data protection hub 200 computes the hash of “abc123” (e.g., “$2a$10$5W8SOHEeTJGb”). At step 704, data protection hub 200 stores transformation data parameters including an indication that bcrypt was used as the hashing algorithm in data storage module 241. At step 705, data protection hub 200 sends the hash to client computer 101. Client computer 101 may then, for example, store the hash in a password database.



FIG. 8 shows a method for generating protected data by masking the unprotected data.


At step 801, data protection hub 200 determines the data protection transformation for the unprotected data to be data masking. As used herein, “data masking” may refer to any algorithm or method for de-contexting or removing sensitive information from input data, or as otherwise known in the art. In some embodiments, sensitive data may be masked by removing parts or elements of the data. For example, a standard 16-digit credit card personal account number (PAN) (e.g., “4061-7240-6172-4061”) may be masked by removing the seventh digit through the twelfth digit (e.g., “4061-72XX-XXXX-4061”). Thus, an attacker would not be able to cause harm with the remaining digits. In addition, in some embodiments, sensitive data may be protected by moving the data to a different context. For example, a PAN, account expiration date, and card verification value (e.g., CVV2) may be de-contexted by appending the CVV2 to an address field associated with the card holder. Thus, an attacker not knowing the new context of the data would not be able to understand the de-contexted data.


At step 802, data protection hub 200 determines sensitive data fields to be masked using the policy parameters. For example, policy parameters may indicate that the unprotected data is a social security number. Accordingly, data protection hub 200 may determine that the first five digits of the social security number are sensitive, but the last four digits are less sensitive.


At step 803, data protection hub 200 masks the sensitive data fields of the unprotected data to generate masked data. The masked data may comprise a shortened form of the unprotected data or may keep the same length but have placeholder data inserted. For example, for the social security number “123-45-6789”, the masked, protected data may comprise “XXX-XX-6789”, or simply “6789”.


At step 804, data protection hub 200 sends the masked data to the client computer 101.



FIG. 9 shows a method for generating protected data by tokenizing the unprotected data.


At step 901, data protection hub 200 determines the data protection transformation to be tokenization. As used herein, “tokenization” may include substituting a sensitive data element with a non-sensitive element and saving a mapping between the elements, or as otherwise known in the art. An authorized user may subsequently detokenize the token to retrieve the sensitive data.


At step 902, data protection hub 200 determines a tokenization algorithm and format using policy parameters associated with the unprotected data. The tokenization algorithm may include an algorithm to generate statistically random data sequences, or an indication of token generator hardware 253 to use to generate a token. The tokenization format may comprise any information used to determine the format of the token, such as the length of the token, the encoding of the token, the character set used to represent the token, etc.


At step 903, data protection hub 200 generates a token using the tokenization algorithm. The token may be generated using software resident on the data protection hub 200, a hardware token generator 253, a combination of the two, or using any other suitable method.


At step 904, data protection hub 200 stores a mapping between the token and the unprotected data. In some embodiments, the mapping may be stored in data storage module 241. The mapping may be used to retrieve a token associated with some unprotected data, or to retrieve unprotected data associated with a token.


At step 905, data protection hub 200 sends the token to the client computer 101.


Any suitable data may be tokenized. For example, a sensitive 16-digit PAN (e.g., “4061-7240-6172-4061”) may be tokenized to produce a token (e.g., “a349ba0530a41169083”). In some circumstances, the token may be used as a substitute for the PAN. For example, a user may use the token in order to pay for a transaction instead of providing the PAN.



FIG. 10 shows a method for generating protected data by encrypting the unprotected data.


At step 1001, data protection hub 200 determines the data protection transformation for the unprotected to be encryption. As used herein, “encryption” may include any encoding of information so that unauthorized users cannot read it, but authorized users can, or as otherwise known in the art. In various embodiments of the invention, encryption may be perform using symmetric keys, or public/private key pairs.


At step 1002, data protection hub 200 determines an encryption algorithm and encryption key using policy parameters associated with the unprotected data. In some embodiments of the invention, a data protection policy may specify an encryption algorithm associated with policy parameters. For example, data protection policy 500 shown in FIG. 5 specifies encryption using a symmetric key or a public key, depending on the policy parameters associated with the unprotected data. The determined encryption algorithm may comprise a key length, a method of key generation, a method of encrypting data, or any other information used to generate a key and encrypt data. In some embodiments, data protection hub 200 may generate an encryption key using encryption as a service module 235 and/or crypto provider module 254.


At step 1003, data protection hub 200 encrypts the unprotected data using the encryption key to generate protected data. The unprotected data may be encrypted using software resident on the data protection hub 200, a hardware crypto provider 254, a combination of the two, or using any other suitable method.


At step 1004, data protection hub 200 stores the encryption algorithm and encryption key used to generate the encrypted data. In some embodiments, the encryption algorithm may be stored using data storage module 241, and the encryption key may be stored using key manager 242.


At step 1005, data protection hub 200 sends the encrypted data to the client computer 101.


V. Exemplary Data Deprotection Methods


FIG. 11 shows a method 1100 for deprotecting data using a data protection hub 200. In some embodiments, the method of FIG. 11 may be performed after data protection hub 200 protects unprotected data (e.g., after the method of FIG. 6).


At step 1101, a client computer 101 sends a data deprotection request to data protection hub 200 including protected data and client credentials. The protected data may include any data that has been protected by a data protection hub, typically by using a data protection transformation. Client credentials may include any data provided by client computer 101 which allows data protection hub 200 to verify that client computer 101 is authorized to deprotect the protected data, such as a user name and password, a digital signature, etc.


At step 1102, data protection hub 200 validates the client credentials.


At step 1103, data protection hub 200 retrieves transformation data parameters associated with the protected data. The transformation data parameters may typically include any data used to deprotect the protected data. For example, if the protected data is encrypted, then transformation data parameters may include a decryption key. If the protected data is a token, the transformation data parameters may include a mapping associating the token with the unprotected data. In some embodiments of the invention, the transformation data parameters may be retrieved from data storage module 241.


At step 1104, data protection hub 200 determines a data deprotection transformation using transformation data parameters. The data deprotection transformation may include any transformation of protected data to unprotected data, such that the unprotected data was used to generate the protected data. For the example, if the data was protected using encryption, the data deprotection transformation may be a decryption operation.


At step 1105, data protection hub 200 performs the data deprotection transformation using the transformation data parameters. In some embodiments, the data deprotection transformation may be performed using an appropriate component of service layer 230 of data protection hub 200. For example, if the data deprotection transformation is decryption, at step 1105 encryption as a service module 235 may be used to decrypt the protected data.


At step 1106, data protection hub 200 sends the unprotected data to the client computer 101.



FIGS. 12 and 13 described below show exemplary methods for transforming protected data in the form of tokens or encrypted data to deprotected data. In some embodiments, the method of FIGS. 12 and 13 may be performed after data protection hub 200 determines a data deprotection transformation for the protected data (e.g., after step 1104).



FIG. 12 shows a method for retrieving unprotected data associated with a token using a data protection hub 200.


At step 1201, data protection hub 200 determines the protected data is a token.


At step 1202, data protection hub 200 retrieves unprotected data associated with the token. In some embodiments, data protection hub 200 may retrieve the unprotected data using a mapping associating the token to the unprotected data that may be retrieved using data storage module 241.


At step 1203, data protection hub 200 sends the unprotected data to client computer 101.



FIG. 13 shows a method for decrypting protected data using a data protection hub 200.


At step 1301, data protection hub 200 determines the protected data is encrypted.


At step 1302, data protection hub 200 retrieves a decryption algorithm and key associated with the protected data. In some embodiments, data protection hub 200 may retrieve the decryption key using key manager 242. At step 1303, data protection hub 200 decrypts the protected data using the decryption key to determine the unprotected data.


At step 1304, data protection hub 200 sends the unprotected data to client computer 101.


VI. Exemplary Data Migration Methods


FIG. 14 shows a method for migrating protected data using a data protection hub 200. In some embodiments of the invention, the method described for FIG. 14 may be used with the system described in FIG. 3. In some embodiments, each data protection hub instantiation 313, 323, 333, and 343 may have the structure described for FIG. 2, and the descriptions are herein incorporated.


At step 1401, a client computer A 312 sends a data migration request message to a data protection hub A 313 including an indication of the data to migrate. A data migration request message may be sent by client computer A 312 to request accessibility of protected data associated with the client computer A 312 from a client computer B 323, such that the client computer B 323 may operate a data protection hub to perform data deprotection transformations on the protected data.


In some embodiments of the invention, the indication of data to migrate may comprise protected or unprotected data. In other embodiments, the indication may include an index, tag, or other information sufficient to identify protected data. Thus, the entirety of the protected data need not be transmitted to data protection hub A 313.


At step 1402, data protection hub A 313 determines transformation data parameters associated with the indicated data. In some embodiments of the invention, the transformation data parameters may be retrieved using data storage module 241.


At step 1403, data protection hub A 313 sends the determined transformation data parameters to data protection hub B 323. Data protection hub A 313 may send the transformation data parameters to data protection hub B 323, if, for example, data protection hub B 323 is closer in proximity to client computer B 322. The transformation data parameters may be associated with client computer B 322, so that credentials used by client computer B 322 may serve to authenticate and authorize deprotection of the migrated data. At step 1404, data protection hub B 323 stores the transformation data parameters.


At step 1405, client computer A 312 sends the protected data to client computer B 322. In some embodiments of the invention, the protected data may be sent from client computer A 312 to client computer B 322 without passing through a data protection hub, such as through an untrusted channel. This does not compromise the security of the data, because it is protected, and any secret information (e.g., decryption keys) required to deprotect the data are not transmitted at step 1405. Thus, the efficiency of data migration is improved.


After the method completes, client computer B 322 may access the protected data by sending a data deprotection request to data protection hub B 323. Thus, client computer B 322 now has full access to the migrated data.


The method of FIG. 14 may performed for any suitable protected data and data protection policy. In one example, client computer A 312 may want to send a database of sensitive consumer information to client computer B 322. The sensitive consumer database may have been previously protected using data protection hub A 312. In order to initiate the migration, at step 1401, client computer A 312 sends a data migration request message to data protection hub 313 indicating that the consumer database is to be migrated to client computer B. Then, at step 1402, data protection hub A 313 retrieves transformation data parameters associated with the consumer database. At step 1403, data protection hub A 313 determines that data protection hub B 323 is the most desirable data protection hub (e.g., because client computer B 322 has the fastest connection to data protection hub B 323). Accordingly, data protection hub A 313 sends the transformation data parameters to data protection hub B 323. At step 1404, data protection hub B stores the transformation data parameters. Finally, at step 1405, client computer A 312 sends the consumer database to client computer B 322 over a separate channel (e.g., by mailing a Blu-Ray disc of the database). After the method completes, client computer B 322 may access the consumer database by sending data deprotection requests to data protection hub B 323.


It should be noted that the method of FIG. 14 is intended to be descriptive and non-limiting. In some embodiments, data migration may be performed for two client computers 101 and 102 connecting to a single data protection hub 200. In such embodiments, step 1403 may not need to be performed. Instead, data protection hub 200 may be configured to accept credentials of client computer B 102 to access the protected data.


VII. Exemplary Payment Processing Systems


FIG. 15 shows an exemplary system for using a data protection hub according to an embodiment of the invention. The system comprises consumer (not shown) who may operate a portable consumer device 1501. The consumer may use portable device 1501 to conduct purchase transactions at an access device (not shown) connected to a merchant computer 1502. Merchant computer 1502 may be connected to acquirer computer 1503. Acquirer computer 1503 may be connected to issuer computer 1505 via payment processing network 1504. Payment processing network 1504 may also be connected to cloud service provider 1506.


As used herein, a “cloud service provider” may typically refer to a server or service that provides a distributed, cloud-based, parallel or otherwise separate environment from payment processing network 1504. Cloud service provider 1504 may be used to execute virtual machines (VMs) defined by a payment processing network 1504 or any other suitable party, such as a merchant 1502, acquirer 1503, merchant processor, issuer processor, or any suitable third-party.


As used herein, an “issuer” may typically refer to a business entity (e.g., a bank) that maintains financial accounts for a consumer and often issues a portable consumer device 1501 such as a credit or debit card to the consumer. A “merchant” is typically an entity that engages in transactions and can sell goods or services. An “acquirer” is typically a business entity (e.g., a commercial bank) that has a business relationship with a particular merchant or other entity. Some entities can perform both issuer and acquirer functions. Some embodiments may encompass such single entity issuer-acquirers. Each of the entities (e.g., merchant computer 1502, acquirer computer 1503, payment processing network 1504, and issuer computer 1505) may comprise one or more computer apparatuses to enable communications, or to perform one or more of the functions described herein.


The payment processing network 1504 may include data processing subsystems, networks, and operations used to support and deliver certificate authority services, authorization services, exception file services, and clearing and settlement services. An exemplary payment processing network may include VisaNet™. Payment processing networks such as VisaNet™ are able to process credit card transactions, debit card transactions, and other types of commercial transactions. VisaNet™, in particular, includes a VIP system (Visa Integrated Payments system) which processes authorization requests and a Base II system which performs clearing and settlement services.


The payment processing network 1504 may include one or more server computers. A server computer is typically a powerful computer or cluster of computers. For example, the server computer can be a large mainframe, a minicomputer cluster, or a group of servers functioning as a unit. In one example, the server computer may be a database server coupled to a Web server. The payment processing network 1504 may use any suitable wired or wireless network, including the Internet.


In a typical payment transaction, the consumer purchases a good or service at merchant 1502 using a portable consumer device 1501. The user's portable consumer device 1501 can interact with an access device at a merchant associated with merchant computer 1502. For example, the consumer may tap the portable consumer device 1501 against an NFC reader in the access device 102. Alternately, the consumer may indicate payment details to the merchant electronically, such as in an online transaction.


An authorization request message is generated by the access device and is then forwarded to the acquirer computer 1503. After receiving the authorization request message, the authorization request message is then sent to the payment processing network 1504. The payment processing network 1504 then forwards the authorization request message to the corresponding issuer computer 1505 associated with the issuer of the portable consumer device 1501.


An “authorization request message” may be an electronic message that is sent to a payment processing network and/or an issuer of a payment card to request authorization for a transaction. An authorization request message according to some embodiments may comply with ISO 8583, which is a standard for systems that exchange electronic transaction information associated with a payment made by a consumer using a payment device or payment account. The authorization request message may include an issuer account identifier that may be associated with a payment device or payment account. An authorization request message may also comprise additional data elements corresponding to “identification information” including, by way of example only: a service code, a CVV (card verification value), a dCVV (dynamic card verification value), an expiration date, etc. An authorization request message may also comprise “transaction information,” such as any information associated with a current transaction, such as the transaction amount, merchant identifier, merchant location, etc., as well as any other information that may be utilized in determining whether to identify and/or authorize a transaction. The authorization request message may also include other information such as information that identifies the access device that generated the authorization request message, information about the location of the access device, etc.


After the issuer computer 1505 receives the authorization request message, the issuer computer 1505 sends an authorization response message back to the payment processing network 1504 to indicate whether the current transaction is authorized (or not authorized). In some embodiments, issuer computer 1505 may examine the testing transaction score to determine whether the current transaction should be authorized. The payment processing network 1504 then forwards the authorization response message back to the acquirer 1503. In some embodiments, payment processing network 1504 may decline the transaction even if issuer computer 1505 has authorized the transaction, for example depending on the value of the fraud risk score or testing transaction score. The acquirer 1503 then sends the response message back to the merchant computer 1502.


An “authorization response message” may be an electronic message reply to an authorization request message generated by an issuing financial institution or a payment processing network. The authorization response message may include, by way of example, one or more of the following status indicators: “Approval”—indicating a transaction was approved; “Decline”—indicating a transaction was not approved; or “Call Center”—indicating a response is pending more information and the merchant must call the toll-free authorization phone number. The authorization response message may also include an authorization code, which may be a code that a credit card issuing bank returns in response to an authorization request message in an electronic message (either directly or through the payment processing network) to the merchant's access device (e.g. POS equipment) that indicates approval of the transaction. The code may serve as proof of authorization. As noted above, in some embodiments, a payment processing network may generate or forward the authorization response message to the merchant.


After the merchant computer 1502 receives the authorization response message, the access device coupled to the merchant computer 1502 may then provide the authorization response message for the consumer. The response message may be displayed on the contactless access device, or may be printed out on a receipt. Alternately, if the transaction is an online transaction, the merchant may provide a web page or other indication of the authorization response message.


At the end of the day, a normal clearing and settlement process can be conducted by the payment processing network 1504. A clearing process is a process of exchanging financial details between an acquirer and an issuer to facilitate posting of transactions to a customer's payment account and reconciliation of the user's settlement position.


In some embodiments of the invention, portable consumer device 1501, merchant computer 1502, acquirer computer 1503, payment processing network 1504, and issuer computer 1505 may make use of data protection hub 200.


For example, some embodiments of the invention may be used to enable token-based payment authentication. In one such embodiment, payment processing network 1504 may define a data protection policy at data protection hub 200 specifying an industry standard tokenization method. Issuer computer 1505 may then generate a plurality of one time use tokens by submitting a data protection request comprising the personal account number (PAN) to the data protection hub 200. Issuer computer 1505 may then load consumer device 1501 (e.g., a smart card) with the plurality of one time use tokens. When a consumer conducts a payment transaction, portable consumer device may provide one of the one time use tokens to merchant computer 1502. Merchant computer 1502 may then send an authorization request message to issuer computer 1505 including the token. Issuer computer 1505 may then send a data deprotection request message to the data protection hub 200 indicating the token. Data protection hub 200 may then retrieve the PAN associated with the token and send the PAN to the issuer computer 1505. Issuer computer 1505 may subsequently retrieve a consumer's account based on the PAN and determine whether the transaction should be approved or declined. If approved, the issuer may return an authorization response message completing the transaction.


In another example, payment processing network 1504 may send protected data to cloud service provider 1506. For example, payment processing network 1504 may maintain some or all transaction processing logic at cloud service provider 1506. In order for the cloud service provider 1506 to deprotect the data, the cloud service provider 1506 may send a data deprotection request message to data protection hub 200. Cloud service provider 1506 may then use the deprotected data for any suitable method.


In yet other examples, multiple systems within the payment processing network 1504 may use data protection hub 200 to provide conformance to payment card industry standards or other data protection standards.


VIII. Exemplary Computer Apparatuses


FIG. 16 shows an example of a payment device 1501″ in the form of a card. As shown, the payment device 1501″ comprises a plastic substrate 1501(m). In some embodiments, a contactless element 1501(o) for interfacing with an access device may be present on, or embedded within, the plastic substrate 1501(m). User information 1501(p) such as an account number, expiration date, and/or a user name may be printed or embossed on the card. A magnetic stripe 1501(n) may also be on the plastic substrate 1501(m). In some embodiments, the payment device 1501″ may comprise a microprocessor and/or memory chips with user data stored in them.


As noted above and shown in FIG. 16, the payment device 1501″ may include both a magnetic stripe 1501(n) and a contactless element 1501(o). In some embodiments, both the magnetic stripe 1501(n) and the contactless element 1501(o) may be in the payment device 1501″. In some embodiments, either the magnetic stripe 1501(n) or the contactless element 1501(o) may be present in the payment device 1501″.



FIG. 17 is a high level block diagram of a computer system that may be used to implement any of the entities or components described above. The subsystems shown in FIG. 17 are interconnected via a system bus 1775. Additional subsystems include a printer 1703, keyboard 1706, fixed disk 1707, and monitor 1709, which is coupled to display adapter 1704. Peripherals and input/output (I/O) devices, which couple to I/O controller 1700, can be connected to the computer system by any number of means known in the art, such as a serial port. For example, serial port 1705 or external interface 1708 can be used to connect the computer apparatus to a wide area network such as the Internet, a mouse input device, or a scanner. The interconnection via system bus 1775 allows the central processor 1702 to communicate with each subsystem and to control the execution of instructions from system memory 1701 or the fixed disk 1707, as well as the exchange of information between subsystems. The system memory 1701 and/or the fixed disk may embody a computer-readable medium.


IX. Additional Embodiments

As described, the inventive service may involve implementing one or more functions, processes, operations or method steps. In some embodiments, the functions, processes, operations or method steps may be implemented as a result of the execution of a set of instructions or software code by a suitably-programmed computing device, microprocessor, data processor, or the like. The set of instructions or software code may be stored in a memory or other form of data storage element which is accessed by the computing device, microprocessor, etc. In other embodiments, the functions, processes, operations or method steps may be implemented by firmware or a dedicated processor, integrated circuit, etc.


It should be understood that the present invention as described above can be implemented in the form of control logic using computer software in a modular or integrated manner. Based on the disclosure and teachings provided herein, a person of ordinary skill in the art will know and appreciate other ways and/or methods to implement the present invention using hardware and a combination of hardware and software.


Any of the software components or functions described in this application may be implemented as software code to be executed by a processor using any suitable computer language such as, for example, Java, C++ or Perl using, for example, conventional or object-oriented techniques. The software code may be stored as a series of instructions, or commands on a computer-readable medium, such as a random access memory (RAM), a read-only memory (ROM), a magnetic medium such as a hard-drive or a floppy disk, or an optical medium such as a CD-ROM. Any such computer-readable medium may reside on or within a single computational apparatus, and may be present on or within different computational apparatuses within a system or network.


While certain exemplary embodiments have been described in detail and shown in the accompanying drawings, it is to be understood that such embodiments are merely illustrative of and not intended to be restrictive of the broad invention, and that this invention is not to be limited to the specific arrangements and constructions shown and described, since various other modifications may occur to those with ordinary skill in the art.


As used herein, the use of “a”, “an” or “the” is intended to mean “at least one”, unless specifically indicated to the contrary.

Claims
  • 1. A data protection hub, comprising: a hardware random number generator configured to generate random numbers;a cryptoprocessor configured to perform cryptographic operations;a processor coupled to the hardware random number generator and the cryptoprocessor and configured for executing code;networking apparatus configured to receive from a client computer, a data protection request message comprising unprotected data, a data protection policy file and one or more policy parameters;a non-transitory computer-readable storage medium, comprising code executable by the processor for: determining a data protection transformation by parsing the data protection policy file, wherein the data protection policy file is used to determine the data protection transformation based on the one or more policy parameters;selecting one or more of the hardware random number generator or the cryptoprocessor based on the determined data protection transformation;performing, using the selected one or more of the hardware random number generator or the cryptoprocessor, the data protection transformation on the unprotected data to generate protected data, wherein the data protection transformation comprises: hashing the unprotected data using one of a plurality of hashing algorithms, based on the determined data protection transformation; andmasking sensitive data fields of the unprotected data based on further determining that the sensitive data fields includes unprotected data; andthe networking apparatus further configured to send the protected data, to the client computer.
  • 2. The data protection hub of claim 1, the method further comprising: determining, using the one or more policy parameters an encryption algorithm and encryption key, wherein the data protection transformation encrypts the unprotected data using the encryption algorithm and encryption key.
  • 3. The data protection hub of claim 1, the method further comprising: determining, using the one or more policy parameters, a tokenization algorithm and format, wherein the data protection transformation tokenizes the unprotected data using the tokenization algorithm and format.
  • 4. The data protection hub of claim 1, the method further comprising: receiving the protected data;performing a data deprotection transformation on the protected data to determine the unprotected data; andsending the unprotected data.
  • 5. The data protection hub of claim 1, the method further comprising: receiving a data migration request message comprising an indication of the protected data to migrate;determining transformation data parameters associated with the protected data; andsending the transformation data parameters to a second data protection hub.
  • 6. A computer-implemented method comprising: receiving, by a processor, a data protection request message from a client computer, the message comprising unprotected data, a data protection file and one or more policy parameters;determining, by the processor, a data protection transformation by parsing the data protection policy file, wherein the data protection policy file is used to determine the data protection transformation based on the one or more policy parameters;selecting one or more of a hardware random number generator or a cryptoprocessor based on the determined data protection transformation, wherein: the hardware random number generator generates random numbers, andthe cryptoprocessor performs cryptographic operations;performing, using the selected one or more of the hardware random number generator or the cryptoprocessor, the data protection transformation on the unprotected data to generate protected data, wherein the data protection transformation comprises: hashing the unprotected data using one of a plurality of hashing algorithms, based on the determined data protection transformation; andmasking sensitive data fields of the unprotected data based on further determining that the sensitive data fields includes unprotected data; andsending, by the processor, the protected data to the client computer.
  • 7. The method of claim 6, the method further comprising: determining, using the one or more policy parameters, an encryption algorithm and encryption key, wherein the data protection transformation encrypts the unprotected data using the encryption algorithm and encryption key.
  • 8. The method of claim 6, the method further comprising: determining, using the one or more policy parameters, a tokenization algorithm and format, wherein the data protection transformation tokenizes the unprotected data using the tokenization algorithm and format.
  • 9. The method of claim 6 further comprising: receiving, by the processor, the protected data;performing, by the processor, a data deprotection transformation on the protected data to determine the unprotected data; andsending, by the processor, the unprotected data.
  • 10. The method of claim 6 further comprising: receiving a data migration request message comprising an indication of the protected data to migrate;determining transformation data parameters associated with the protected data; andsending the transformation data parameters to a second data protection hub.
CROSS-REFERENCES TO RELATED APPLICATIONS

The present application is a non-provisional application of and claims priority to U.S. Provisional Application No. 61/667,822, filed on Jul. 3, 2012, the entire contents of which are herein incorporated by reference in their entirety for all purposes.

US Referenced Citations (516)
Number Name Date Kind
5613012 Hoffman Mar 1997 A
5781438 Lee Jul 1998 A
5883810 Franklin Mar 1999 A
5953710 Fleming Sep 1999 A
5956699 Wong Sep 1999 A
6000832 Franklin Dec 1999 A
6014635 Harris Jan 2000 A
6044360 Picciallo Mar 2000 A
6163771 Walker Dec 2000 A
6227447 Campisano May 2001 B1
6236981 Hill May 2001 B1
6267292 Walker Jul 2001 B1
6327578 Linehan Dec 2001 B1
6341724 Campisano Jan 2002 B2
6385596 Wiser May 2002 B1
6422462 Cohen Jul 2002 B1
6425523 Shem Ur Jul 2002 B1
6592044 Wong Jul 2003 B1
6636833 Flitcroft Oct 2003 B1
6748367 Lee Jun 2004 B1
6805287 Bishop Oct 2004 B2
6879965 Fung Apr 2005 B2
6891953 DeMello May 2005 B1
6901387 Wells May 2005 B2
6931382 Laage Aug 2005 B2
6938019 Uzo Aug 2005 B1
6941285 Sarcanin Sep 2005 B2
6980670 Hoffman Dec 2005 B1
6990470 Hogan Jan 2006 B2
6991157 Bishop Jan 2006 B2
7051929 Li May 2006 B2
7069249 Stolfo Jun 2006 B2
7103576 Mann, III Sep 2006 B2
7113930 Eccles Sep 2006 B2
7136835 Flitcroft Nov 2006 B1
7177835 Walker Feb 2007 B1
7177848 Hogan Feb 2007 B2
7194437 Britto Mar 2007 B1
7209561 Shankar et al. Apr 2007 B1
7264154 Harris Sep 2007 B2
7287692 Patel Oct 2007 B1
7292999 Hobson Nov 2007 B2
7350230 Forrest Mar 2008 B2
7353382 Labrou Apr 2008 B2
7379919 Hogan May 2008 B2
RE40444 Linehan Jul 2008 E
7415443 Hobson Aug 2008 B2
7444676 Asghari-Kamrani Oct 2008 B1
7469151 Khan Dec 2008 B2
7548889 Bhambri Jun 2009 B2
7567934 Flitcroft Jul 2009 B2
7567936 Peckover Jul 2009 B1
7571139 Giordano Aug 2009 B1
7571142 Flitcroft Aug 2009 B1
7580898 Brown Aug 2009 B2
7584153 Brown Sep 2009 B2
7593896 Flitcroft Sep 2009 B1
7606560 Labrou Oct 2009 B2
7627531 Breck Dec 2009 B2
7627895 Gifford Dec 2009 B2
7650314 Saunders Jan 2010 B1
7685037 Reiners Mar 2010 B2
7702578 Fung Apr 2010 B2
7707120 Dominguez Apr 2010 B2
7712655 Wong May 2010 B2
7734527 Uzo Jun 2010 B2
7753265 Harris Jul 2010 B2
7770789 Oder, II Aug 2010 B2
7784685 Hopkins, III Aug 2010 B1
7793851 Mullen Sep 2010 B2
7801826 Labrou Sep 2010 B2
7805376 Smith Sep 2010 B2
7805378 Berardi Sep 2010 B2
7818264 Hammad Oct 2010 B2
7828220 Mullen Nov 2010 B2
7835960 Breck Nov 2010 B2
7840595 Blitzer Nov 2010 B1
7841523 Oder, II Nov 2010 B2
7841539 Hewton Nov 2010 B2
7844550 Walker Nov 2010 B2
7848980 Carlson Dec 2010 B2
7849020 Johnson Dec 2010 B2
7853529 Walker Dec 2010 B1
7853995 Chow Dec 2010 B2
7865414 Fung Jan 2011 B2
7873579 Hobson Jan 2011 B2
7873580 Hobson Jan 2011 B2
7890393 Talbert Feb 2011 B2
7891563 Oder, II Feb 2011 B2
7896238 Fein Mar 2011 B2
7908216 Davis et al. Mar 2011 B1
7922082 Muscato Apr 2011 B2
7931195 Mullen Apr 2011 B2
7938318 Fein May 2011 B2
7954705 Mullen Jun 2011 B2
7959076 Hopkins, III Jun 2011 B1
7996288 Stolfo Aug 2011 B1
8025223 Saunders Sep 2011 B2
8046256 Chien Oct 2011 B2
8060448 Jones Nov 2011 B2
8060449 Zhu Nov 2011 B1
8074877 Mullen Dec 2011 B2
8074879 Harris Dec 2011 B2
8082210 Hansen Dec 2011 B2
8095113 Kean Jan 2012 B2
8104679 Brown Jan 2012 B2
RE43157 Bishop Feb 2012 E
8109436 Hopkins, III Feb 2012 B1
8121942 Carlson Feb 2012 B2
8121956 Carlson Feb 2012 B2
8126449 Beenau Feb 2012 B2
8171525 Pelly May 2012 B1
8196813 Vadhri Jun 2012 B2
8205791 Randazza Jun 2012 B2
8219489 Patterson Jul 2012 B2
8224702 Mengerink Jul 2012 B2
8225385 Chow Jul 2012 B2
8229852 Carlson Jul 2012 B2
8265993 Chien Sep 2012 B2
8280777 Mengerink Oct 2012 B2
8281991 Wentker et al. Oct 2012 B2
8328095 Oder, II Dec 2012 B2
8336088 Raj et al. Dec 2012 B2
8346666 Lindelsee et al. Jan 2013 B2
8376225 Hopkins, III Feb 2013 B1
8380177 Laracey Feb 2013 B2
8387873 Saunders Mar 2013 B2
8401539 Beenau Mar 2013 B2
8401898 Chien Mar 2013 B2
8402555 Grecia Mar 2013 B2
8403211 Brooks Mar 2013 B2
8412623 Moon Apr 2013 B2
8412837 Emigh Apr 2013 B1
8417642 Oren Apr 2013 B2
8447699 Batada May 2013 B2
8453223 Svigals May 2013 B2
8453925 Fisher Jun 2013 B2
8458487 Palgon Jun 2013 B1
8484134 Hobson Jul 2013 B2
8485437 Mullen Jul 2013 B2
8494959 Hathaway Jul 2013 B2
8498908 Mengerink Jul 2013 B2
8504475 Brand et al. Aug 2013 B2
8504478 Saunders Aug 2013 B2
8510816 Quach Aug 2013 B2
8533860 Grecia Sep 2013 B1
8538845 Liberty Sep 2013 B2
8555079 Shablygin Oct 2013 B2
8566168 Bierbaum Oct 2013 B1
8567670 Stanfield Oct 2013 B2
8571939 Lindsey Oct 2013 B2
8577336 Mechaley, Jr. Nov 2013 B2
8577803 Chatterjee Nov 2013 B2
8578176 Mattsson Nov 2013 B2
8583494 Fisher Nov 2013 B2
8584251 Mcguire Nov 2013 B2
8589237 Fisher Nov 2013 B2
8589271 Evans Nov 2013 B2
8589291 Carlson Nov 2013 B2
8595098 Starai Nov 2013 B2
8595812 Bomar Nov 2013 B2
8595850 Spies Nov 2013 B2
8606638 Dragt Dec 2013 B2
8606700 Carlson Dec 2013 B2
8606720 Baker Dec 2013 B1
8615468 Varadarajan Dec 2013 B2
8620754 Fisher Dec 2013 B2
8635157 Smith Jan 2014 B2
8646059 Von Behren Feb 2014 B1
8651374 Brabson Feb 2014 B2
8656180 Shablygin Feb 2014 B2
8751391 Freund Jun 2014 B2
8762263 Gauthier et al. Jun 2014 B2
8838982 Carlson et al. Sep 2014 B2
8887308 Grecia Nov 2014 B2
9065643 Hurry et al. Jun 2015 B2
9070129 Sheets et al. Jun 2015 B2
9160741 Wentker et al. Oct 2015 B2
9229964 Stevelinck Jan 2016 B2
9245267 Singh Jan 2016 B2
9249241 Dai et al. Feb 2016 B2
9256871 Anderson et al. Feb 2016 B2
9280765 Hammad Mar 2016 B2
20010029485 Brody Oct 2001 A1
20010034720 Armes Oct 2001 A1
20010054003 Chien Dec 2001 A1
20020007320 Hogan Jan 2002 A1
20020016749 Borecki Feb 2002 A1
20020029193 Ranjan Mar 2002 A1
20020035548 Hogan Mar 2002 A1
20020073045 Rubin Jun 2002 A1
20020116341 Hogan Aug 2002 A1
20020133467 Hobson Sep 2002 A1
20020147913 Lun Yip Oct 2002 A1
20030028481 Flitcroft Feb 2003 A1
20030130955 Hawthorne Jul 2003 A1
20030191709 Elston Oct 2003 A1
20030191945 Keech Oct 2003 A1
20040010462 Moon Jan 2004 A1
20040050928 Bishop Mar 2004 A1
20040059682 Hasumi Mar 2004 A1
20040093281 Silverstein May 2004 A1
20040139008 Mascavage Jul 2004 A1
20040143532 Lee Jul 2004 A1
20040158532 Breck Aug 2004 A1
20040210449 Breck Oct 2004 A1
20040210498 Freund Oct 2004 A1
20040232225 Bishop Nov 2004 A1
20040260646 Berardi Dec 2004 A1
20050037735 Coutts Feb 2005 A1
20050080730 Sorrentino Apr 2005 A1
20050108178 York May 2005 A1
20050199709 Linlor Sep 2005 A1
20050246293 Ong Nov 2005 A1
20050269401 Spitzer Dec 2005 A1
20050269402 Spitzer Dec 2005 A1
20060123461 Lunt Jun 2006 A1
20060235795 Johnson Oct 2006 A1
20060237528 Bishop Oct 2006 A1
20060278704 Saunders Dec 2006 A1
20070091790 Passey Apr 2007 A1
20070107044 Yuen May 2007 A1
20070129955 Dalmia Jun 2007 A1
20070136193 Starr Jun 2007 A1
20070136211 Brown Jun 2007 A1
20070170247 Friedman Jul 2007 A1
20070179885 Bird Aug 2007 A1
20070208671 Brown Sep 2007 A1
20070245414 Chan Oct 2007 A1
20070288377 Shaked Dec 2007 A1
20070291995 Rivera Dec 2007 A1
20080015988 Brown Jan 2008 A1
20080029607 Mullen Feb 2008 A1
20080035738 Mullen Feb 2008 A1
20080052226 Agarwal Feb 2008 A1
20080054068 Mullen Mar 2008 A1
20080054079 Mullen Mar 2008 A1
20080054081 Mullen Mar 2008 A1
20080065554 Hogan Mar 2008 A1
20080065555 Mullen Mar 2008 A1
20080201264 Brown Aug 2008 A1
20080201265 Hewton Aug 2008 A1
20080228646 Myers Sep 2008 A1
20080243702 Hart Oct 2008 A1
20080245855 Fein Oct 2008 A1
20080245861 Fein Oct 2008 A1
20080283591 Oder, II Nov 2008 A1
20080302869 Mullen Dec 2008 A1
20080302876 Mullen Dec 2008 A1
20080313264 Pestoni Dec 2008 A1
20090006262 Brown Jan 2009 A1
20090010488 Matsuoka Jan 2009 A1
20090037333 Flitcroft Feb 2009 A1
20090037388 Cooper Feb 2009 A1
20090043702 Bennett Feb 2009 A1
20090048971 Hathaway Feb 2009 A1
20090106112 Dalmia Apr 2009 A1
20090106160 Skowronek Apr 2009 A1
20090134217 Flitcroft May 2009 A1
20090157555 Biffle Jun 2009 A1
20090159673 Mullen Jun 2009 A1
20090159700 Mullen Jun 2009 A1
20090159707 Mullen Jun 2009 A1
20090173782 Muscato Jul 2009 A1
20090200371 Kean Aug 2009 A1
20090248583 Chhabra Oct 2009 A1
20090276347 Kargman Nov 2009 A1
20090281948 Carlson Nov 2009 A1
20090294527 Brabson Dec 2009 A1
20090307139 Mardikar Dec 2009 A1
20090308921 Mullen Dec 2009 A1
20090319773 Frenkel et al. Dec 2009 A1
20090319786 Viscomi Dec 2009 A1
20090327131 Beenau Dec 2009 A1
20100008535 Abulafia Jan 2010 A1
20100088237 Wankmueller Apr 2010 A1
20100094755 Kloster Apr 2010 A1
20100106644 Annan Apr 2010 A1
20100120408 Beenau May 2010 A1
20100133334 Vadhri Jun 2010 A1
20100138347 Chen Jun 2010 A1
20100145860 Pelegero Jun 2010 A1
20100161433 White Jun 2010 A1
20100185545 Royyuru Jul 2010 A1
20100211505 Saunders Aug 2010 A1
20100223186 Hogan Sep 2010 A1
20100228668 Hogan Sep 2010 A1
20100235284 Moore Sep 2010 A1
20100258620 Torreyson Oct 2010 A1
20100291904 Musfeldt Nov 2010 A1
20100299267 Faith et al. Nov 2010 A1
20100306076 Taveau Dec 2010 A1
20100325041 Berardi Dec 2010 A1
20110010292 Giordano Jan 2011 A1
20110016047 Wu Jan 2011 A1
20110016320 Bergsten Jan 2011 A1
20110040640 Erikson Feb 2011 A1
20110047076 Carlson et al. Feb 2011 A1
20110083018 Kesanupalli Apr 2011 A1
20110087596 Dorsey Apr 2011 A1
20110093397 Carlson Apr 2011 A1
20110125597 Oder, II May 2011 A1
20110153437 Archer Jun 2011 A1
20110153498 Makhotin et al. Jun 2011 A1
20110154466 Harper Jun 2011 A1
20110161233 Tieken Jun 2011 A1
20110178926 Lindelsee et al. Jul 2011 A1
20110191244 Dai Aug 2011 A1
20110238511 Park Sep 2011 A1
20110238573 Varadarajan Sep 2011 A1
20110246317 Coppinger Oct 2011 A1
20110258111 Raj et al. Oct 2011 A1
20110272471 Mullen Nov 2011 A1
20110272478 Mullen Nov 2011 A1
20110276380 Mullen Nov 2011 A1
20110276381 Mullen Nov 2011 A1
20110276424 Mullen Nov 2011 A1
20110276425 Mullen Nov 2011 A1
20110289565 Resch Nov 2011 A1
20110295745 White Dec 2011 A1
20110302081 Saunders Dec 2011 A1
20120028609 Hruska Feb 2012 A1
20120030047 Fuentes et al. Feb 2012 A1
20120035998 Chien Feb 2012 A1
20120041881 Basu Feb 2012 A1
20120047237 Arvidsson Feb 2012 A1
20120066078 Kingston Mar 2012 A1
20120066769 Latchem Mar 2012 A1
20120072350 Goldthwaite Mar 2012 A1
20120078735 Bauer Mar 2012 A1
20120078798 Downing Mar 2012 A1
20120078799 Jackson Mar 2012 A1
20120095852 Bauer Apr 2012 A1
20120095865 Doherty Apr 2012 A1
20120116902 Cardina May 2012 A1
20120123882 Carlson May 2012 A1
20120123940 Killian May 2012 A1
20120129514 Beenau May 2012 A1
20120143767 Abadir Jun 2012 A1
20120143772 Abadir Jun 2012 A1
20120158580 Eram Jun 2012 A1
20120158593 Garfinkle Jun 2012 A1
20120173431 Ritchie Jul 2012 A1
20120185386 Salama Jul 2012 A1
20120197807 Schlesser Aug 2012 A1
20120203664 Torossian Aug 2012 A1
20120203666 Torossian Aug 2012 A1
20120215688 Musser Aug 2012 A1
20120215696 Salonen Aug 2012 A1
20120221421 Hammad Aug 2012 A1
20120226582 Hammad Sep 2012 A1
20120231844 Coppinger Sep 2012 A1
20120233004 Bercaw Sep 2012 A1
20120246070 Vadhri Sep 2012 A1
20120246071 Jain Sep 2012 A1
20120246079 Wilson et al. Sep 2012 A1
20120265631 Cronic Oct 2012 A1
20120271770 Harris Oct 2012 A1
20120297446 Webb Nov 2012 A1
20120300932 Cambridge Nov 2012 A1
20120303503 Cambridge Nov 2012 A1
20120303961 Kean Nov 2012 A1
20120304273 Bailey Nov 2012 A1
20120310725 Chien Dec 2012 A1
20120310831 Harris Dec 2012 A1
20120311345 Dhuse Dec 2012 A1
20120316992 Oborne Dec 2012 A1
20120317035 Royyuru Dec 2012 A1
20120317036 Bower Dec 2012 A1
20130017784 Fisher Jan 2013 A1
20130018757 Anderson et al. Jan 2013 A1
20130019098 Gupta Jan 2013 A1
20130031006 Mccullagh et al. Jan 2013 A1
20130054337 Brendell Feb 2013 A1
20130054466 Muscato Feb 2013 A1
20130054474 Yeager Feb 2013 A1
20130081122 Svigals Mar 2013 A1
20130091028 Oder, II Apr 2013 A1
20130110658 Lyman May 2013 A1
20130111599 Gargiulo May 2013 A1
20130117185 Collison May 2013 A1
20130124290 Fisher May 2013 A1
20130124291 Fisher May 2013 A1
20130124364 Mittal May 2013 A1
20130138525 Bercaw May 2013 A1
20130144888 Faith Jun 2013 A1
20130145148 Shablygin Jun 2013 A1
20130145172 Shablygin Jun 2013 A1
20130159178 Colon Jun 2013 A1
20130159184 Thaw Jun 2013 A1
20130166402 Parento Jun 2013 A1
20130166456 Zhang Jun 2013 A1
20130173736 Krzeminski Jul 2013 A1
20130185202 Goldthwaite Jul 2013 A1
20130191286 Cronic Jul 2013 A1
20130191289 Cronic Jul 2013 A1
20130198071 Jurss Aug 2013 A1
20130198080 Anderson et al. Aug 2013 A1
20130200146 Moghadam Aug 2013 A1
20130204787 Dubois Aug 2013 A1
20130204793 Kerridge Aug 2013 A1
20130212007 Mattsson Aug 2013 A1
20130212017 Bangia Aug 2013 A1
20130212019 Mattsson Aug 2013 A1
20130212024 Mattsson Aug 2013 A1
20130212026 Powell et al. Aug 2013 A1
20130212666 Mattsson Aug 2013 A1
20130218698 Moon Aug 2013 A1
20130218769 Pourfallah et al. Aug 2013 A1
20130226799 Raj Aug 2013 A1
20130226813 Voltz Aug 2013 A1
20130246199 Carlson Sep 2013 A1
20130246202 Tobin Sep 2013 A1
20130246203 Laracey Sep 2013 A1
20130246258 Dessert Sep 2013 A1
20130246259 Dessert Sep 2013 A1
20130246261 Purves et al. Sep 2013 A1
20130246267 Tobin Sep 2013 A1
20130254028 Salci Sep 2013 A1
20130254052 Royyuru Sep 2013 A1
20130254102 Royyuru Sep 2013 A1
20130254117 Von Mueller Sep 2013 A1
20130262296 Thomas Oct 2013 A1
20130262302 Lettow Oct 2013 A1
20130262315 Hruska Oct 2013 A1
20130262316 Hruska Oct 2013 A1
20130262317 Collinge Oct 2013 A1
20130275300 Killian Oct 2013 A1
20130275307 Khan Oct 2013 A1
20130275308 Paraskeva Oct 2013 A1
20130282502 Jooste Oct 2013 A1
20130282575 Mullen Oct 2013 A1
20130282588 Hruska Oct 2013 A1
20130297501 Monk et al. Nov 2013 A1
20130297504 Nwokolo Nov 2013 A1
20130297508 Belamant Nov 2013 A1
20130304649 Cronic Nov 2013 A1
20130308778 Fosmark Nov 2013 A1
20130311382 Fosmark Nov 2013 A1
20130317982 Mengerink Nov 2013 A1
20130332344 Weber Dec 2013 A1
20130339253 Sincai Dec 2013 A1
20130346314 Mogollon Dec 2013 A1
20140007213 Sanin Jan 2014 A1
20140013106 Redpath Jan 2014 A1
20140013114 Redpath Jan 2014 A1
20140019352 Shrivastava Jan 2014 A1
20140025581 Calman Jan 2014 A1
20140025585 Calman Jan 2014 A1
20140025958 Calman Jan 2014 A1
20140032417 Mattsson Jan 2014 A1
20140032418 Weber Jan 2014 A1
20140040137 Carlson Feb 2014 A1
20140040139 Brudnicki Feb 2014 A1
20140040144 Plomske Feb 2014 A1
20140040145 Ozvat Feb 2014 A1
20140040148 Ozvat Feb 2014 A1
20140040628 Fort Feb 2014 A1
20140041018 Bomar Feb 2014 A1
20140046853 Spies Feb 2014 A1
20140047551 Nagasundaram et al. Feb 2014 A1
20140052532 Tsai Feb 2014 A1
20140052620 Rogers Feb 2014 A1
20140052637 Jooste Feb 2014 A1
20140068706 Aissi Mar 2014 A1
20140074637 Hammad Mar 2014 A1
20140108172 Weber et al. Apr 2014 A1
20140114857 Griggs et al. Apr 2014 A1
20140143137 Carlson May 2014 A1
20140164243 Aabye et al. Jun 2014 A1
20140188586 Carpenter et al. Jul 2014 A1
20140294701 Dai et al. Oct 2014 A1
20140310183 Weber Oct 2014 A1
20140330721 Wang Nov 2014 A1
20140330722 Laxminarayanan et al. Nov 2014 A1
20140331265 Mozell et al. Nov 2014 A1
20140337236 Wong et al. Nov 2014 A1
20140344153 Raj et al. Nov 2014 A1
20140372308 Sheets Dec 2014 A1
20150019443 Sheets et al. Jan 2015 A1
20150032625 Dill Jan 2015 A1
20150032626 Dill Jan 2015 A1
20150032627 Dill Jan 2015 A1
20150046338 Laxminarayanan Feb 2015 A1
20150046339 Wong et al. Feb 2015 A1
20150052064 Karpenko et al. Feb 2015 A1
20150088756 Makhotin et al. Mar 2015 A1
20150106239 Gaddam et al. Apr 2015 A1
20150112870 Nagasundaram et al. Apr 2015 A1
20150112871 Kumnick Apr 2015 A1
20150120472 Aabye et al. Apr 2015 A1
20150127529 Makhotin et al. May 2015 A1
20150127547 Powell et al. May 2015 A1
20150140960 Powell et al. May 2015 A1
20150142673 Nelsen et al. May 2015 A1
20150161597 Subramanian et al. Jun 2015 A1
20150178724 Ngo et al. Jun 2015 A1
20150180836 Wong et al. Jun 2015 A1
20150186864 Jones et al. Jul 2015 A1
20150193222 Pirzadeh et al. Jul 2015 A1
20150195133 Sheets et al. Jul 2015 A1
20150199679 Palanisamy et al. Jul 2015 A1
20150199689 Kumnick et al. Jul 2015 A1
20150220917 Aabye et al. Aug 2015 A1
20150269566 Gaddam et al. Sep 2015 A1
20150312038 Palanisamy Oct 2015 A1
20150319158 Kumnick Nov 2015 A1
20150332262 Lingappa Nov 2015 A1
20150356560 Shastry et al. Dec 2015 A1
20160028550 Gaddam et al. Jan 2016 A1
20160042263 Gaddam et al. Feb 2016 A1
20160065370 Le Saint et al. Mar 2016 A1
20160092696 Guglani et al. Mar 2016 A1
20160092872 Prakash et al. Mar 2016 A1
20160103675 Aabye et al. Apr 2016 A1
20160119296 Laxminarayanan et al. Apr 2016 A1
Foreign Referenced Citations (19)
Number Date Country
2156397 Feb 2010 EP
11-068730 Mar 1999 JP
2001-223729 Aug 2001 JP
10-1996-0015309 May 1996 KR
10-2009-0061253 Jun 2009 KR
0135304 May 2001 WO
0135304 May 2001 WO
2004042536 May 2004 WO
2006113834 Oct 2006 WO
2009032523 Mar 2009 WO
2010078522 Jul 2010 WO
2012068078 May 2012 WO
2012098556 Jul 2012 WO
2012142370 Oct 2012 WO
2012167941 Dec 2012 WO
2013048538 Apr 2013 WO
2013056104 Apr 2013 WO
2013119914 Aug 2013 WO
2013179271 Dec 2013 WO
Non-Patent Literature Citations (30)
Entry
International Search Report and Written Opinion mailed Oct. 18, 2013 in PCT/US2013/049332, 9 pages.
Wang, U.S. Appl. No. 62/000,288, Payment System Canonical Address Format filed May 19, 2014, unpublished.
Sharma et al., U.S. Appl. No. 62/003,717, Mobile Merchant Application filed May 28, 2014, unpublished.
Kalgi et al., U.S. Appl. No. 62/024,426, Secure Transactions Using Mobile Devices filed Jul. 14, 2014, unpublished.
Prakash et al., U.S. Appl. No. 62/037,033, Sharing Payment Token filed Aug. 13, 2014, unpublished.
Hoverson et al., U.S. Appl. No. 62/038,174, Customized Payment Gateway filed Aug. 15, 2014, unpublished.
Wang, U.S. Appl. No. 62/042,050, Payment Device Authentication and Authorization System filed Aug. 26, 2014, unpublished.
Gaddam et al., U.S. Appl. No. 62/053,736, Completing Transactions Without a User Payment Device filed Sep. 22, 2014, unpublished.
Patterson, U.S. Appl. No. 62/054,346, Mirrored Token Vault filed Sep. 23, 2014, unpublished.
Dimmick, U.S. Appl. No. 14/952,514, Systems Communications With Non-Sensitive Identifiers filed Nov. 25, 2015, unpublished.
Dimmick, U.S. Appl. No. 14/952,444, Tokenization Request Via Access Device filed Nov. 25, 2015, unpublished.
Prakash et al., U.S. Appl. No. 14/955,716, Provisioning Platform for Machine-To-Machine Devices filed Dec. 1, 2015, unpublished.
Wong et al., U.S. Appl. No. 14/966,948, Automated Access Data Provisioning filed Dec. 11, 2015, unpublished.
Stubbs et al., U.S. Appl. No. 62/103,522, Methods and Systems for Wallet Provider Provisioning filed Jan. 14, 2015, unpublished.
McGuire, U.S. Appl. No. 14/600,523, Secure Payment Processing Using Authorization Request filed Jan. 20, 2015, unpublished.
Flurscheim et al., U.S. Appl. No. 15/004,705, Cloud-Based Transactions With Magnetic Secure Transmission filed Jan. 22, 2016, unpublished.
Flurscheim et al., U.S. Appl. No. 62/108,403, Wearables With NFC HCE filed Jan. 27, 2015, unpublished.
Sabba et al., U.S. Appl. No. 15/011,366, Token Check Offline filed Jan. 29, 2016, unpublished.
Patterson, U.S. Appl. No. 15/019,157, Token Processing Utilizing Multiple Authorizations filed Feb. 9, 2016, unpublished.
Cash et al., U.S. Appl. No. 15/041,495, Peer Forward Authorization of Digital Requests filed Feb. 11, 2016, unpublished.
Le Saint et al., U.S. Appl. No. 15/008,388, Methods for Secure Credential Provisioning filed Jan. 27, 2016, unpublished.
Kinagi, U.S. Appl. No. 62/117,291, Token and Cryptogram Using Transaction Specific Information filed Feb. 17, 2015, unpublished.
Galland et al. U.S. Appl. No. 62/128,709, Tokenizing Transaction Amounts filed Mar. 5, 2015, unpublished.
Rangarajan et al., U.S. Appl. No. 61/751,763, Payments Bridge filed Jan. 11, 2013, unpublished.
Li, U.S. Appl. No. 61/894,749, Methods and Systems for Authentication and Issuance of Tokens in a Secure Environment filed Oct. 23, 2013, unpublished.
Aissi et al., U.S. Appl. No. 61/738,832, Management of Sensitive Data filed Dec. 18, 2012, unpublished.
Wong et al., U.S. Appl. No. 61/879,362, Systems and Methods for Managing Mobile Cardholder Verification Methods filed Sep. 18, 2013, unpublished.
Powell, U.S. Appl. No. 61/892,407, Issuer Over-The-Air Update Method and System filed Oct. 17, 2013, unpublished.
Powell, U.S. Appl. No. 61/926,236, Methods and Systems for Provisioning Mobile Devices With Payment Credentials and Payment Token Identifiers filed Jan. 10, 2014, unpublished.
Petition for Inter Partes Review of U.S. Pat. No. 8,533,860 Challenging Claims 1-30 Under 35 U.S.C. § 312 and 37 C.F.R. § 42.104, filed Feb. 17, 2016, Before the USPTO Patent Trial and Appeal Board, IPR 2016-00600, 65 pages.
Related Publications (1)
Number Date Country
20140013452 A1 Jan 2014 US
Provisional Applications (1)
Number Date Country
61667822 Jul 2012 US