The present invention generally relates to a database system, and more specifically to responding to a database query by executing a differentially private version of the query on the database.
Personally identifiable information, such as health data, financial records, telecom data, and confidential business intelligence, such as proprietary data or data restricted by contractual obligations, is valuable for analysis and collaboration. Yet, only a fraction of such sensitive information is used by organizations or analysts for statistical or predictive analysis. Privacy regulations, security concerns, and technological challenges suppress the full value of data, especially personally identifiable information and confidential and proprietary records.
Methods that attempt to solve this problem, such as access controls, data masking, hashing, anonymization, aggregation, and tokenization, are invasive and resource intensive, compromise analytical utility, or do not ensure privacy of the records. For example, data masking may remove or distort data, compromising the statistical properties of the data. As another example, many of the above mentioned methods are not effective when information is stored in disparate data sources. Technology which enables organizations or analysts to execute advanced statistical and predictive analysis on sensitive information across disparate data sources without revealing record-level information is needed.
A hardware database privacy device is communicatively coupled to a private database system. The hardware database privacy device receives a request from a client device to perform a query of the private database system and identifies a level of differential privacy corresponding to the request. The identified level of differential privacy includes a privacy parameter ε indicating the degree of information released about data in the private database system.
The differentially private hardware database privacy device identifies a set of data stored in the private database system and a set of operations to be performed on the set of data that corresponds to the requested query. After the set of data is accessed, the set of operations is modified based on the identified level of differential privacy such that a performance of the modified set of operations produces a result set that is differentially private. The modified set of operations is performed on the accessed set of data to produce the differentially private result set. The differentially private result set is provided to the client device for display on a hardware display of the client device.
The Figures (FIGS.) and the following description describe certain embodiments by way of illustration only. One skilled in the art will readily recognize from the following description that alternative embodiments of the structures and methods illustrated herein may be employed without departing from the principles described herein. Reference will now be made in detail to several embodiments, examples of which are illustrated in the accompanying figures. It is noted that wherever practicable similar or like reference numbers may be used in the figures and may indicate similar or like functionality.
The database 106 is one or more private databases managed by one or more entities that can only be accessed by authorized or trusted users. For example, the database 106 may contain health data of patients, financial records, telecom data, and confidential business intelligence, such as proprietary data or data restricted by contractual obligations. The information stored in the database 106 is of interest to one or more clients 104, but clients 104 may not have the necessary authorization to access to information contained in the databases 106.
The feature values may be numerical in nature, e.g., Features 1 and 10, or categorical in nature, e.g., Features 2 and 11. In the case of categorical feature values, each category may be denoted as an integer. For example, in Feature 11 of
Returning to
The query 108 submitted by the client 104 may be simple queries, such as count queries that request the number of entries in the databases 106 that satisfy a condition specified by the client 104, or complicated queries, such as predictive analytics queries that request a data analytics model trained on the databases 106. Upon submitting a query 108 to the differentially private security system 102, the client 104 receives a DP response 112 to a differentially private version of the submitted query 114.
The client 104 specifies a set of privacy parameters each time the client 104 submits query 108. The privacy parameters indicate an amount of decrease in the privacy budget of the client 104 in return for a response to the query 108. As described below in more detail with reference to the privacy system 160 in
The differentially private security system 102 receives an analytical query 108 from the client 104 and applies a differentially private version of the query 114 on the database 106, such that it releases a degree of information about the database 106 indicated by the privacy parameters specified by the client 104, but also protects a degree of privacy of the databases 106 specified by the entities managing the database 106. For example, the entities managing the database 106 may also set a maximum threshold on the degree of information released about the database 106 for a given query 108 that the client 104 may not exceed. Thus, the differentially private security system balances privacy protection of the database 106 while releasing useful information on the database 106 to the client 104. The differentially private security system 102 may have complete or partial access to the databases 106.
Upon receiving a query 108, the differentially private security system 102 applies DP query 114 to the database 106 and returns a DP response 112 to the client 104. The DP query 114 is a differentially private version of the query 108 that satisfies a definition of differential privacy described in more detail with reference to the privacy system 160 in
Differentially Private Security System
The differentially private security system 102 includes a user interface 150, a library 152, an account management system 154, a query handling engine 156, a data integration module 158, and a privacy system 160. Some embodiments of the differentially private security system 102 have different or additional modules than the ones described here. Similarly, the functions can be distributed among the modules in a different manner than is described here. Certain modules and functions can be incorporated into other modules of the differentially private security system 102.
The user interface 150 can generate a graphical user interface on a dedicated hardware device of the differentially private security system 102 or the client 104 in which the client 104 can submit an analytical query 108 and the desired privacy parameters, and view DP response 112 in the form of numerical values or images. The client 104 may also inspect database 106 schemata, view an associated privacy budget, or cache the DP response 112 to view the response later. The user interface 150 submits properly formatted query commands to other modules of the differentially private security system 102.
The library 152 contains software components that can be included in external programs that allow the client 104 to submit the analytical query 108, receive the DP response 112, and other functions within a script or program. For example, the client 104 may use the software components of the library 152 to construct custom data analytic programs. Each of the software components in the library 152 submits properly formatted query commands to other modules of the differentially private security system 102.
The account management system 154 receives properly formatted query commands (herein “query commands” or “QC”), parses the received query commands, and updates the account of the client 104 according to the received query command. For example, the account management system 154 may check the query commands for syntactic correctness, or check whether a client 104 has access to a requested resource. As another example, the account management system 154 may check whether the privacy parameters specified by the client 104 for a given analytical query 108 can be accommodated, and if so, decrement the privacy budget of the client 104 by the amount specified in the query 108. Query commands verified by the account management system 154 are provided to the query handling engine 156. Examples of query commands accommodated by the differentially private security system 102 are listed below.
The query handling engine 156 transforms the received query commands into appropriate function calls and database access commands by parsing the query command string. The function calls are specific to the query 108 requested by the client 104, and the access commands allow access to the required database 106. Different databases 106 require different access commands. The access commands are provided to the database integrator 158.
The database integrator 158 receives the access commands to one or more databases 106 and collects the required databases 106 and merges them into a single data object. The data object has a structure similar to that of a database structure described in reference to
The privacy system 160 receives the data object from the database integrator 158, appropriate function calls from the query handling engine 156 indicating the type of query 108 submitted by the client 104, privacy parameters specified for the query 108, and produces a DP response 112 to a differentially private version of the query 108 with respect to the databases 106. The privacy system 160 will be described in further detail in reference to
Privacy System
Definition of Differential Privacy
For a given query 108, the privacy system 160 receives a data object X, function calls indicating the type of query 108, privacy parameters specified by the client 104, and outputs a DP response 112 to a differentially private version of the query 108 with respect to X. Each data object X is a collection of row vectors xi=1, 2, . . . n, in which each row vector x; has a series of p elements xij=2, . . . , p.
A query M satisfies the definition of e-differential privacy if for all:
where is the space of all possible data objects, X, X′ neighboring data objects, S is an output space of query M, and neighboring databases are defined as two data objects X, X′ that have at most one different entry from one another. That is, given two neighboring data objects X, X′ in which one has an individual's data entry, and the other does not, there is no output of query M that an adversary can use to distinguish between X, X′. That is, an output of such a query M that is differentially private reveals no information about the data object X. The privacy parameter ε controls the amount of information that the query M reveals about any individual data entry in X, and represents the degree of information released about the entries in X. For example, in the definition given above, a small value of ε indicates that the probability an output of query M will disclose information on a specific data entry is small, while a large value of ε indicates the opposite.
As another definition of differential privacy, a query M is (ε,δ)-differentially private if for neighboring data objects X, X′:
The privacy parameter δ measures the improbability of the output of query M satisfying ε-differential privacy. As discussed in reference to
There are three important definitions for discussing the privacy system 160: global sensitivity, local sensitivity, and smooth sensitivity. Global sensitivity of a query M is defined as
where X, X′ are any neighboring data objects, such that d(X, X′)=1. This states that the global sensitivity is the most the output of query M could change by computing M on X and X′.
The local sensitivity of a query M on the data object X is given by:
where the set {X′: d(X, X′)=1} denotes all data objects that have at most one entry that is different from X. That is, the local sensitivity LSM(X) is the sensitivity of the output of the query M on data objects X′ that have at most one different entry from X, measured by a norm function.
Related to the local sensitivity LSM(X), the smooth sensitivity given a parameter/p is given by:
where d(X, X′) denotes the number of entries that differ between X and X′.
Notation for Random Variables
The notation in this section is used for the remainder of the application to denote the following random variables.
1) G(σ2), denotes a zero-centered Gaussian random variable with the probability density function
2) L(b) denotes a zero-centered Laplacian random variable with the probability density function
3) C(γ) denotes a zero-centered Cauchy random variable with the probability density function
Further, a vector populated with random variables R as its elements is denoted by ν(R). A matrix populated with random variables R as its elements is denoted by M(R).
Count Engine 302
The count engine 302 produces a DP response 112 responsive to the differentially private security system 102 receiving a query 108 for counting the number of entries in a column of the data object X that satisfy a condition specified by the client 104, given privacy parameters (ε,δ). An example query command for accessing the count engine 302 is given in QC1 above. For the example data object X shown in
The count engine 302 retrieves the count q from X. If privacy parameter δ is equal to zero, the count engine 302 returns
as the DP response 112 for display on the user interface 150, where c1 is a constant. An example value for c1 may be 1. If the privacy parameter δ is non-zero, the count engine 302 returns
as the DP response 112 for display on the user interface 150, where c1 is a constant. An example value for c1 may be 1.
The client 104 may request visualization of entries in the data object X for analysis of trends or patterns that depend on the features of the entries. In one embodiment, the privacy system 160 generates a differentially private visualization of the requested data entries from X.
The privacy system 160 first maps the requested entries from X for the selected features specified by the client 104. For example, as shown in the visualization 410 of
For each disjoint region, the privacy system 160 submits a differentially private count query to the count engine 302, and randomly plots a number of entries determined by the DP response 112 of the count engine 302 for that region. The resulting DP visualization plot is returned to the client 104 for display to a user by the user interface 150. For example, square 440 in visualization 410 contains 3 entries, while the same square in DP visualization 420 contains 4 randomly plotted entries determined by the DP response 112 of the count engine 302.
Median Engine 304
The median engine 304 produces a DP response 112 responsive to the differentially private security system 102 receiving a query 108 for generating the median of entries in a column of the data object X that satisfy a condition specified by the client 104, given privacy parameters (ε,δ). An example query command for accessing the median engine 304 is given in QC2 above. For the example data object X shown in
The median engine 304 aggregates the values of entries satisfying the condition specified by the client 104 into a list U, and retrieves the median q from U. If privacy parameter δ is equal to zero, the median engine 304 returns
as the DP response 112 for display on the user interface 150, in which c1, c2 are constant factors. Example values for c1, c2 may be 6 and ⅙, respectively. If δ is non-zero, the median engine 304 returns
as the DP response 112 for display on the user interface 150. Example values for c1, c2 may be 2 and 1, respectively.
Mean Engine 306
The mean engine 306 produces a DP response 112 responsive the differentially private security system 102 receiving a query 108 for generating the mean of entries in a column of the data object X that satisfy a condition specified by the client 104, given privacy parameters (ε,δ). An example query command for accessing the mean engine 306 is given in QC3 above. For the example data object X shown in
The mean engine 306 aggregates the values of entries satisfying the condition specified by the client 104 into a list U. Assuming there are n values in U, the mean engine 306 further divides U into m sub-lists Vj=1, 2, . . . , m each with n m values. The mean engine 306 aggregates each mean rj of sub-list Vj into a list R. The mean engine 306 requests a differentially private median query of the values in R to the median engine 304. The resulting output from the median engine 304 is returned as the DP response 112 for display on the user interface 150.
Variance Engine 308
The variance engine 308 produces a DP response 112 responsive to the differentially private security system 102 receiving a query 108 for generating the variance of entries in a column of the data object X that satisfy a condition specified by the client 104, given privacy parameters (ε,δ). An example query command for accessing the variance engine 308 is given in QC4 above. For the example data object X shown in
The variance engine 308 aggregates the values of entries satisfying the condition specified by the client 104 into a list U. Assuming there are n values in U, the variance engine 308 further divides U into m sub-lists Vj=1, 2, . . . , m each with n m values. The variance engine 308 aggregates each variance r; of sub-list Vj into a list R. The variance engine 308 requests a differentially private median query of the values in R to the median engine 304. The resulting output from the median engine 304 is returned as the DP response 112 for display on the user interface 150.
IQR Engine 310
The IQR engine 310 produces a DP response 112 responsive to the differentially private security system 102 receiving a query 108 for generating the interquartile range (IQR) of entries in a column of the data object X that satisfy a condition specified by the client 104, given privacy parameters (ε,δ). An example query command for accessing the IQR engine 310 is given in QC5 above. For the example data object X shown in
In one embodiment, the IQR engine 310 aggregates the values of entries satisfying the condition specified by the client 104 into a list U. Assuming there are n values in U, the sample IQR of U is denoted as IQR(U), and a log transform of IQR(U) is denoted as:
The IQR engine 310 further maps the quantity Hn(U) to an integer k0 such that Hn(U)∈[k0, k0+1). The IQR engine 310 extracts a value A0(U) indicating the number of entries in U required to change in order for the new list U to satisfy Hn(Ũ) ∉[k0, k0+1).
The IQR engine 310 then generates a value R0(U) given by:
in which c1 is a constant factor. If R0(U) is greater than a predetermined threshold, the IQR engine 310 returns
as the DP response 112 for display on the user interface 150. If R0(U) is equal to or less than the predetermined threshold, the IQR engine 310 returns “No Answer” as the DP response 112 for display on the user interface 150.
In another embodiment, the IQR engine 310 aggregates the values of entries satisfying the condition specified by the client 104 into an ordered list U. The IQR engine 310 retrieves the first quartile and the third quartile from U, given by q and q′, respectively. If δ is zero, the IQR engine 310 returns:
as the DP response 112 for display on the user interface 150, in which c1, c2 are constant factors.
If δ is non-zero, the IQR engine 310 returns:
as the DP response 112 for display on the user interface 150, in which c1, c2 are constant factors.
Batch Gradient Engine 312
The batch gradient engine 312 produces a DP response 112 responsive to the differentially private security system 102 receiving a valid query 108 for generating a set of parameters θ for a general linear model that captures the correlation between a series of observable features and a dependent feature, given privacy parameters (ε,δ). The general linear model is trained on the selected columns of X. An example query command for accessing the batch gradient engine 312 is given in QC6 above.
Given a row vector x that contains a series of observable features and a label feature y, the correlation between the observable features and the label feature in a general linear model may be given as:
y=xθT,
where θ is a row vector containing parameters of the model. That is, the label feature is modeled as a weighted sum of the observable features, where each value in θ is the weight given to a corresponding observable feature.
For the example data object X shown in
Examples of general linear models supported by the batch gradient engine 312 are, but not limited to, linear regression, logistic regression, and support vector machine (SVM) classifiers.
The optimal values for the set of parameters θ is found by training the general linear model on training data (Xtrain, ytrain) consisting of selected columns of data object X. Specifically, Xtrain is a matrix database in which each column corresponds to a selected observable feature, and y is a column vector of the selected label feature values. Each entry in Xtrain has a one-to-one correspondence with an entry in y. The optimal θ is generally found by minimizing a loss function on (Xtrain, ytrain) over possible values of θ. Mathematically, the minimization is given by:
The batch gradient engine 312 returns a DP response 112 θDP of a differentially private batch gradient query by perturbing the loss function to be minimized. Specifically, the perturbed minimization is given by:
in which K is the Lipschitz constant for loss function (⋅). If j is the index of the columns in Xtrain, xij denotes the value of entry i and column j in Xtrain, and it is publicly known that for each column j, aj≤xij≤bj, R2 may be given by:
R2=max(∥u∥2|aj≤uj≤bj)
where a is a vector having elements uj. The DP response 112 θDP may be provided for display on the user interface 150.
Stochastic Gradient Engine 314
Similarly to the batch gradient engine 312, the stochastic gradient engine 314 produces a DP response 112 responsive to the differentially private security system 102 receiving a valid query 108 for generating a set of parameters θ for a general linear model that captures the correlation between a series of observable features and a label feature, given privacy parameters (ε,δ). An example query command for accessing the stochastic gradient engine 314 is given in QC7 above.
Similar to the batch gradient engine 312, examples of general linear models supported by the stochastic gradient engine 314 are, but not limited to, linear regression, logistic regression, and support vector machine (SVM) classifiers.
The stochastic gradient engine 314 also minimizes a loss function on training data (Xtrain, ytrain) over possible values of 0 to find the optimal vales of parameter vector θ. However, the stochastic gradient engine 314 may minimize the loss function based on individual points or a subset of the training data, instead of the entire training data.
As discussed in reference to the batch gradient engine 312, a general minimization problem for finding the optimal values for θ over training data (Xtrain, ytrain) is given by:
where (⋅) is a loss function. The minimization is solved by applying stochastic gradient descent on the loss function (⋅) with respect to θ. This involves the steps of identifying an initial set of values for θ, calculating the gradient of the loss function with respect to θ, and updating θ based on the calculated gradient. The steps are repeated until the algorithm reaches convergence, and an optimal set of values for θ that minimize the loss function are identified.
Specifically, given the estimate for the parameter θt at time t, stochastic gradient descent generates a new estimate θt+1 at the next time step t+1 by the following equation:
θt+1θt−ηt·n·∇θ
in which ∇θ
The stochastic gradient engine 314 returns a DP response 112 θDP of a differentially private stochastic gradient query by perturbing the update of θ at one or more time steps of the stochastic gradient descent algorithm. Specifically, a perturbed update at time t to t+1 is given by:
The stochastic gradient engine 314 may output the perturbed update at each time step as the DP response 112 for display on the user interface 150, or the converged parameter vector θDP as the DP response 112 for display on the user interface 150.
Random Forest Engine 316
The random forest engine 316 produces a DP response 112 responsive to the differentially private security system 102 receiving a valid query 108 for generating a trained random forest classifier that bins a series of feature values into one among multiple categories, given privacy parameters (ε,δ). The random forest classifier is trained on the selected columns of X. An example query command for accessing the random forest engine 316 is given in QC8 above. For the example data object X shown in
The random forest classifier, is trained on training data (Xtrain, ytrain) to learn the correlation between the selected features of an entry and the category the entry belongs to. Specifically, Xtrain is a matrix database in which each column corresponds to a selected feature of interest to the client 104, and y is a column vector of already known labels indicating the category of a corresponding entry. Each entry in Xtrain has a one-to-one correspondence with a label entry in y. Upon being trained, the random forest classifier, or a classifier in general, receives a new data entry with selected feature values and generates an estimate of the category for the new entry.
The random forest classifier is an ensemble of individual binary decision tree classifiers, in which each binary decision tree generates an estimate for the category of an entry. Given a new data entry, the random forest classifier aggregates the category estimates from each binary decision tree and produces a final estimate for the category of the data entry.
For each trained binary decision tree, each node except the root node corresponds to a partition of training data entries formed by a split s at a parent node. The split s at the parent node is based on a test condition of a feature of the training data (Xtrain, ytrain) that compares the feature value of an entry to a reference value, and verifies whether the feature value meets that condition or not. Returning to the example shown in
At the end of the training process, each leaf node is associated with a category that has a dominant proportion in the corresponding partition at the leaf node. In
The random forest engine 316 returns a DP response 112 of a differentially private random forest query by perturbing the proportion of training data entries at leaf nodes of each trained binary decision tree. Specifically, the random forest engine 316 trains a random forest classifier T with an ensemble of Ntrees binary decision trees Bj=1, 2, . . . , Ntrees using training data (Xtrain, ytrain) from the data object X. Assuming a binary classification problem with two labels “0” and “1,” the random forest engine 316 perturbs the proportion of data entries associated with each category for each leaf node tL. The perturbed proportion pDP(⋅|tL) is given by:
The random forest engine 316 returns the random forest classifier TDP containing an ensemble of perturbed binary decision trees BDPj=1, 2, . . . , Ntrees as the DP response 112. Moreover, the random forest engine 316 may display the perturbed proportion of data entries for leaf nodes of each binary decision tree BDPj=1, 2, . . . , Ntrees for display on the user interface 150.
Histogram Engine 318
The histogram engine 318 produces a DP response 112 responsive to the differentially private security system 102 receiving a query 108 for generating a histogram of a selected column in X, given privacy parameters (ε,δ). The histogram engine 318 creates one or more bins corresponding to one or more disjoint ranges of the selected feature values, and indicates the number or proportion of entries that belong to each bin. An example query command for accessing the histogram engine 318 is given in QC9 above. For the example data object X shown in
The histogram engine 318 returns a DP response 112 of a differentially private histogram query by perturbing the counts for each bin.
In one embodiment, the histogram engine 318 generates the requested histogram from the selected column of X, and perturbs the counts of each bin by submitting a request to the count engine 302.
In another embodiment, the histogram engine 318 generates the requested histogram from the selected column of X, and perturbs the counts of each bin by decomposing the counts using a private wavelet decomposition algorithm. In such an embodiment, the histogram engine 318 aggregates the counts qi=1, 2, . . . , B for each bin bi=1, 2, . . . , B into a matrix (or vector) Q. The histogram engine 318 decomposes Q into a tree structure that is representative of a wavelet decomposition. Each leaf node of the tree corresponds to a count qi, and each parent node of the tree corresponds to one of multiple wavelet coefficients Cj=1, 2, . . . , m. The value of a wavelet coefficient cj is calculated based on the counts qt incorporated in the leaf nodes of the tree. This allows a count qi to be reconstructed as a function ƒi of the wavelet coefficients cj=1, 2, . . . , m. That is, for each count qi:
qi=ƒi(c0,C1, . . . ,Cm).
The histogram engine 318 generates a perturbed histogram by perturbing the wavelet coefficients, and reconstructing the counts using the perturbed wavelet coefficients. Specifically, the perturbed wavelet coefficients cDPi=1, 2, . . . , m are given by:
The reconstructed counts from the perturbed wavelet coefficients is now given by:
qiDP=ƒi(C0DP,C1DP, . . . ,CmDP.
The histogram engine 318 outputs the perturbed histogram as the DP response 112 for display on the user interface 150.
In one embodiment, the histogram engine 318 may also be used to generate a differentially private visualization of data entries as described above in reference to the count engine 302 and
Model Testing Engine 320
The model testing engine 320 produces a DP response 112 responsive to the differentially private security system 102 receiving a query 108 for testing the performance of a classification model, given privacy parameters (ε,δ). The classification model is trained and tested on selected columns of X. As such, the model testing engine 320 may be appended to any other module that trains a classifier on X, such as the batch gradient engine 312, the stochastic gradient engine 314, or the random forest engine 316. For the example data object X shown in
As discussed in reference to the random forest engine 316, classification models in general is trained on training data (Xtrain, ytrain) to learn the correlation between selected features of an entry and the category the entry belongs to. The training data (Xtrain, ytrain) may be extracted from a subset of entries contained in the data object X. Upon being trained, the classifier is able to receive a new data entry containing values for the selected features and generate an estimate of the category for the new entry.
Often times, the estimate of the category for an entry is determined by applying a cutoff threshold to a numerical, not categorical, output of a classifier. For example, in the random forest classifier described in reference to the random forest engine 316, the category associated with a leaf node tL is determined by the proportion of training data entries associated with each category, which is a numerical value. The random forest engine 316 may determine that a leaf node is associated with category “0” if the proportion of entries associated with label “0” is above a cutoff threshold of 0.5, 0.6, or 0.7. As another example, logistic regression classifiers output a numerical value in the range of [0, 1] given an entry of feature values. The entry may be classified into category “0” if the associated output is below a cutoff threshold of 0.5, 0.4, or 0.3. Regardless of the example, the cutoff threshold for determining the boundary between each category is a critical parameter depending on the context the classifier is applied to.
The model testing engine 320 receives a trained classifier and tests the performance of the trained classifier a series of cutoff thresholds, and generates a confusion matrix for each threshold indicating the performance of the classifier. The model testing engine 320 may test the performance of the classifier on testing data (Xtest, ytest). Similarly to training data, Xtest contains a set of entries with selected feature values, and ytest contains a vector of already known labels for each corresponding entry in Xtest. However, in contrast to training data, testing data (Xtest, ytest) comprises entries that are not present in training data (Xtrain, ytrain). That is, testing data comprises entries that the classifier has not “seen” yet.
The model testing engine 320 generates a series of cutoff thresholds based on the numerical values of p.
For example, as shown in
For each threshold mi, the model testing engine 320 generates corresponding category label estimates from p, and compares the estimates to the vector of known labels ytest. Given the comparisons, the model testing engine 320, constructs a confusion matrix that evaluates the performance of the classifier.
For each threshold mi, the model testing engine 320 generates a perturbed confusion matrix by using the histogram engine 318. This is because each entry contributes to only one among the 4 disjoint categories, and thus, the entries in the confusion matrix 700 can be viewed as a histogram. The model testing engine 320 outputs each threshold mi, and the corresponding perturbed confusion matrix as the DP response 112 for display on the user interface 150.
Synthetic Database Engine 322
The synthetic database engine 322 produces a DP response 112 responsive to the differentially private security system 102 receiving a query 108 for transforming X into a synthetic database S, given privacy parameters (ε,δ). The resulting synthetic database S has a number of entries corresponding to that in X, but a fewer number of columns or features than X. Moreover, the spatial relationship between a pair of entries in X is retained in S. The transformation of X to S is (ε,δ)-differentially private with respect to a neighboring data object X′ with a 1-element difference from X.
The synthetic database engine 322 produces a DP response 112 of a differentially private synthetic database query by projecting the elements of X to S using a projection matrix. Assuming that data object X is a n×p matrix having n rows and p columns, the transformation by the synthetic database engine 322 is given by:
where J is a p×k projection matrix, with k<p. The resulting synthetic database matrix S is a n×k matrix containing equal number of entries or rows as data object matrix X, but containing a smaller number of features or columns than the original data object X.
As discussed above, the transformation using projection matrix J is (ε,δ)-differentially private. Moreover, the spatial relationship between a pair of entries in X is retained in S. That is, the distance between a pair of entries (xi,xj) in the p-dimensional feature space of X is approximately equal to the distance between a pair of entries (si,sj) in the k-dimensional feature space of S. The synthetic database engine 322 outputs S as the DP response 112 for display on the user interface 150.
In the example of
Validation Engine 324
The validation engine 324 produces a DP response 112 responsive to the differentially private security system 102 receiving a request for whether a query 108 satisfies the definition of (ε,δ)-differential privacy for privacy parameters (ε,δ). In one embodiment, the validation engine 324 may receive a function call from the client 104 that points to the query 108. The query 108 may be, for example, an analytical model or an algorithm that can be applied to a data object X.
The validation engine 324 certifies whether the received query 108 satisfies the definition of (ε,δ)-differential privacy by applying the query 108 to example pairs of neighboring data objects (Z, Z′). Specifically, the validation engine 324 generates pairs of neighboring data objects (Z, Z′), having at most 1 entry different from each other. The validation engine 324 applies the received query 108 to each example pair of neighboring data objects (Z, Z′) and determines whether an estimate of the quantity Pr[M(X)∈S]/Pr[M(X′)∈S] satisfies the definition of (ε,δ)-differential privacy a sampling of outputs from S of the query M and over the randomness of the query M.
In one embodiment, the validation engine 324 may output a binary value to the client 104 as the DP response 112 for display on the user interface 150 that indicates whether or not the query 108 is (ε,δ)-differentially private. In some embodiments, the validation engine 324, in response to a determination that the query 108 is not (ε,δ)-differentially private, can reject or deny the query.
A request from a client device to perform a query is received 1010 and a level of differential privacy corresponding to the request is identified. A set of data in the private database system and a set of operations to be performed based on the received request is identified 1012. The set of identified data in the private database system is accessed 1014. The set of operations is modified 1016 based on the received level of differential privacy. The set of modified operations is performed 1018 on the set of data to produce a differentially private result set. The differentially private result set is provided 1020 to the client device for display on the client device.
The machine may be a server computer, a client computer, a personal computer (PC), a tablet PC, a set-top box (STB), a smartphone, an internet of things (IoT) appliance, a network router, switch or bridge, or any machine capable of executing instructions 1124 (sequential or otherwise) that specify actions to be taken by that machine. Further, while only a single machine is illustrated, the term “machine” shall also be taken to include any collection of machines that individually or jointly execute instructions 1124 to perform any one or more of the methodologies discussed herein.
The example computer system 1100 includes one or more processing units (generally processor 1102). The processor 1102 is, for example, a central processing unit (CPU), a graphics processing unit (GPU), a digital signal processor (DSP), a controller, a state machine, one or more application specific integrated circuits (ASICs), one or more radio-frequency integrated circuits (RFICs), or any combination of these. The computer system 1100 also includes a main memory 1104. The computer system may include a storage unit 1116. The processor 1102, memory 1104 and the storage unit 1116 communicate via a bus 1108.
In addition, the computer system 1106 can include a static memory 1106, a display driver 1110 (e.g., to drive a plasma display panel (PDP), a liquid crystal display (LCD), or a projector). The computer system 1100 may also include alphanumeric input device 1112 (e.g., a keyboard), a cursor control device 1114 (e.g., a mouse, a trackball, a joystick, a motion sensor, or other pointing instrument), a signal generation device 1118 (e.g., a speaker), and a network interface device 1120, which also are configured to communicate via the bus 1108.
The storage unit 1116 includes a machine-readable medium 1122 on which is stored instructions 1124 (e.g., software) embodying any one or more of the methodologies or functions described herein. The instructions 1124 may also reside, completely or at least partially, within the main memory 1104 or within the processor 1102 (e.g., within a processor's cache memory) during execution thereof by the computer system 1100, the main memory 1104 and the processor 1102 also constituting machine-readable media. The instructions 1124 may be transmitted or received over a network 1126 via the network interface device 1120.
While machine-readable medium 1122 is shown in an example embodiment to be a single medium, the term “machine-readable medium” should be taken to include a single medium or multiple media (e.g., a centralized or distributed database, or associated caches and servers) able to store the instructions 1124. The term “machine-readable medium” shall also be taken to include any medium that is capable of storing instructions 1124 for execution by the machine and that cause the machine to perform any one or more of the methodologies disclosed herein. The term “machine-readable medium” includes, but not be limited to, data repositories in the form of solid-state memories, optical media, and magnetic media.
Some portions of the above description describe the embodiments in terms of algorithmic processes or operations. These algorithmic descriptions and representations are commonly used by those skilled in the data processing arts to convey the substance of their work effectively to others skilled in the art. These operations, while described functionally, computationally, or logically, are understood to be implemented by computer programs comprising instructions for execution by a processor or equivalent electrical circuits, microcode, or the like. Furthermore, it has also proven convenient at times, to refer to these arrangements of functional operations as modules, without loss of generality. The described operations and their associated modules may be embodied in software, firmware, hardware, or any combinations thereof.
As used herein any reference to “one embodiment” or “an embodiment” means that a particular element, feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment. The appearances of the phrase “in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment.
Some embodiments may be described using the expression “coupled” and “connected” along with their derivatives. It should be understood that these terms are not intended as synonyms for each other. For example, some embodiments may be described using the term “connected” to indicate that two or more elements are in direct physical or electrical contact with each other. In another example, some embodiments may be described using the term “coupled” to indicate that two or more elements are in direct physical or electrical contact. The term “coupled,” however, may also mean that two or more elements are not in direct contact with each other, but yet still co-operate or interact with each other. The embodiments are not limited in this context.
As used herein, the terms “comprises,” “comprising,” “includes,” “including,” “has,” “having” or any other variation thereof, are intended to cover a non-exclusive inclusion. For example, a process, method, article, or apparatus that comprises a list of elements is not necessarily limited to only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Further, unless expressly stated to the contrary, “or” refers to an inclusive or and not to an exclusive or. For example, a condition A or B is satisfied by any one of the following: A is true (or present) and B is false (or not present), A is false (or not present) and B is true (or present), and both A and B are true (or present).
In addition, use of the “a” or “an” are employed to describe elements and components of the embodiments herein. This is done merely for convenience and to give a general sense of the disclosure. This description should be read to include one or at least one and the singular also includes the plural unless it is obvious that it is meant otherwise.
Upon reading this disclosure, those of skill in the art will appreciate still additional alternative structural and functional designs for a system and a process for receiving a query for a private database, and responding to the query by executing a differentially private version of the query on the private database. Thus, while particular embodiments and applications have been illustrated and described, it is to be understood that the described subject matter is not limited to the precise construction and components disclosed herein and that various modifications, changes and variations which will be apparent to those skilled in the art may be made in the arrangement, operation and details of the method and apparatus disclosed herein.
This application is a continuation of U.S. patent application Ser. No. 16/810,708, filed Mar. 5, 2020, continuation of U.S. patent application Ser. No. 16/238,439, filed Jan. 2, 2019, which is a continuation of U.S. patent application Ser. No. 15/793,907, filed Oct. 25, 2017, which is a continuation of U.S. patent application Ser. No. 15/203,797, filed Jul. 7, 2016, which claims the benefit of U.S. Patent Application No. 62/249,938, filed Nov. 2, 2015, all of which are incorporated by reference herein.
Number | Name | Date | Kind |
---|---|---|---|
6038563 | Bapat et al. | Mar 2000 | A |
6438549 | Aldred et al. | Aug 2002 | B1 |
6546389 | Agrawal et al. | Apr 2003 | B1 |
6618721 | Lee | Sep 2003 | B1 |
6823338 | Byrne et al. | Nov 2004 | B1 |
7219237 | Trimberger | May 2007 | B1 |
7356840 | Bedell et al. | Apr 2008 | B1 |
7698250 | Dwork et al. | Apr 2010 | B2 |
7801967 | Bedell et al. | Sep 2010 | B1 |
9002803 | Qayyum et al. | Apr 2015 | B2 |
9094378 | Yung et al. | Jul 2015 | B1 |
9244976 | Zhang et al. | Jan 2016 | B1 |
9384226 | Goel et al. | Jul 2016 | B1 |
10192069 | Nerurkar et al. | Jan 2019 | B2 |
10229287 | Nerurkar et al. | Mar 2019 | B2 |
10242224 | Nerurkar et al. | Mar 2019 | B2 |
10467234 | Nerurkar et al. | Nov 2019 | B2 |
10489605 | Nerurkar et al. | Nov 2019 | B2 |
10586068 | Nerurkar et al. | Mar 2020 | B2 |
10642847 | Nerurkar et al. | May 2020 | B1 |
10726153 | Nerurkar et al. | Jul 2020 | B2 |
10733320 | Nerurkar et al. | Aug 2020 | B2 |
11055432 | Hockenbrocht et al. | Jul 2021 | B2 |
11100247 | Nerurkar et al. | Aug 2021 | B2 |
20010034847 | Gaul, Jr. | Oct 2001 | A1 |
20030110467 | Balakrishnan | Jun 2003 | A1 |
20030177118 | Moon et al. | Sep 2003 | A1 |
20040225896 | Ng | Nov 2004 | A1 |
20040250120 | Ng | Dec 2004 | A1 |
20050278786 | Tippett et al. | Dec 2005 | A1 |
20060053112 | Chitkara et al. | Mar 2006 | A1 |
20060161527 | Dwork et al. | Jul 2006 | A1 |
20060200431 | Dwork et al. | Sep 2006 | A1 |
20060224597 | Fitzpatrick et al. | Oct 2006 | A1 |
20060238503 | Smith et al. | Oct 2006 | A1 |
20060265396 | Raman et al. | Nov 2006 | A1 |
20060282433 | Dutta et al. | Dec 2006 | A1 |
20070047558 | Ayers et al. | Mar 2007 | A1 |
20070136027 | Dwork et al. | Jun 2007 | A1 |
20070143289 | Dwork et al. | Jun 2007 | A1 |
20070239982 | Aggarwal et al. | Oct 2007 | A1 |
20080033960 | Banks et al. | Feb 2008 | A1 |
20080133935 | Elovici et al. | Jun 2008 | A1 |
20090119298 | Faitelson et al. | May 2009 | A1 |
20090177685 | Tanyeri et al. | Jul 2009 | A1 |
20090249436 | Coles et al. | Oct 2009 | A1 |
20090254971 | Herz et al. | Oct 2009 | A1 |
20090265354 | Machak et al. | Oct 2009 | A1 |
20090327228 | Krause et al. | Dec 2009 | A1 |
20110064221 | McSherry | Mar 2011 | A1 |
20110078143 | Aggarwal | Mar 2011 | A1 |
20110125730 | Bordawekar et al. | May 2011 | A1 |
20110131222 | Dicrescenzo | Jun 2011 | A1 |
20110208763 | Mcsherry et al. | Aug 2011 | A1 |
20110238611 | Mcsherry et al. | Sep 2011 | A1 |
20110282865 | Talwar et al. | Nov 2011 | A1 |
20120109830 | Vogel | May 2012 | A1 |
20120143922 | Rane et al. | Jun 2012 | A1 |
20120166483 | Choudhary et al. | Jun 2012 | A1 |
20120197864 | Bourdoncle et al. | Aug 2012 | A1 |
20120226492 | Tsuboi et al. | Sep 2012 | A1 |
20130031136 | Shah | Jan 2013 | A1 |
20130145473 | Cormode et al. | Jun 2013 | A1 |
20130332891 | Schmitlin et al. | Dec 2013 | A1 |
20140013400 | Warshavsky et al. | Jan 2014 | A1 |
20140028291 | Lee et al. | Jan 2014 | A1 |
20140088989 | Krishnapuram et al. | Mar 2014 | A1 |
20140214735 | Harik | Jul 2014 | A1 |
20140281572 | Wang et al. | Sep 2014 | A1 |
20140283091 | Zhang | Sep 2014 | A1 |
20150235051 | Fawaz et al. | Aug 2015 | A1 |
20150286827 | Fawaz et al. | Oct 2015 | A1 |
20150293923 | Eide et al. | Oct 2015 | A1 |
20160036827 | Kling et al. | Feb 2016 | A1 |
20160105409 | Torman et al. | Apr 2016 | A1 |
20160218738 | Kim | Jul 2016 | A1 |
20160283738 | Wang et al. | Sep 2016 | A1 |
20160283938 | Streuter et al. | Sep 2016 | A1 |
20160306709 | Shaull | Oct 2016 | A1 |
20160335455 | Mohan et al. | Nov 2016 | A1 |
20170124152 | Nerurkar et al. | May 2017 | A1 |
20170126694 | Nerurkar et al. | May 2017 | A1 |
20170169253 | Curcio et al. | Jun 2017 | A1 |
20170235974 | Zhang et al. | Aug 2017 | A1 |
20170316391 | Peikert et al. | Nov 2017 | A1 |
20170359364 | Thakurta et al. | Dec 2017 | A1 |
20180039674 | Seyvet et al. | Feb 2018 | A1 |
20180048653 | Nerurkar et al. | Feb 2018 | A1 |
20180048654 | Nerurkar et al. | Feb 2018 | A1 |
20180233925 | Boehmer et al. | Aug 2018 | A1 |
20180239924 | Rickard, Jr. et al. | Aug 2018 | A1 |
20180329952 | Ramachandra et al. | Nov 2018 | A1 |
20180349384 | Nerurkar et al. | Dec 2018 | A1 |
20190026489 | Nerurkar et al. | Jan 2019 | A1 |
20190138743 | Nerurkar et al. | May 2019 | A1 |
20190141052 | Nerurkar et al. | May 2019 | A1 |
20190147188 | Benaloh et al. | May 2019 | A1 |
20190318121 | Hockenbrocht et al. | Oct 2019 | A1 |
20200210610 | Nerurkar et al. | Jul 2020 | A1 |
20240012928 | Nerurkar et al. | Jan 2024 | A1 |
Number | Date | Country |
---|---|---|
2998839 | May 2017 | CA |
2998839 | Mar 2021 | CA |
108537055 | Sep 2018 | CN |
110198302 | Sep 2019 | CN |
3353734 | Aug 2018 | EP |
WO-2015090445 | Jun 2015 | WO |
WO-2015157020 | Oct 2015 | WO |
WO-2017078808 | May 2017 | WO |
WO-2017187207 | Nov 2017 | WO |
Entry |
---|
Dankar, F. et al. “Practicing Differential Privacy in Health Care: A Review,” 2013. (Year: 2013). |
Li, C. et al. “Optimizing linear counting queries under differential privacy,” In Proceedings of the twenty-ninth ACM SIGMOD-SIGACT-SIGART symposium on Principles of database systems. Jun. 6, 2010 (pp. 123-134). (Year: 2010). |
Rastogi, V. et al. “Differentially private aggregation of distributed Time-series with transformation and encryption”, Proc. ACM SIGMOD Int. Conf. Manage. Data, pp. 735-746, 2010 (Year: 2010). |
Goryczka S. et al., “A comprehensive comparison of multiparty secure additions with differential privacy,” IEEE transactions on dependable and secure computing. Oct. 1, 2015;14(5):463-77. (Year: 2015). |
Chaudhuri, S. et al. “Database Access Control & Privacy: Is There a Common Ground?” CIDR 2011: Fifth Biennial Conference on Innovative Data Systems Research, Jan. 9-12, 2011, pp. 96-103. |
European Patent Office, Extended European Search Report, European Patent Application No. 19785548.9, dated Dec. 10, 2021, 10 pages. |
European Patent Office, Extended European Search Report, European Patent Application No. 19889515.3, dated Jun. 24, 2022, seven pages. |
Gaboardi, M. et al. “PSI: A Private Data Sharing Interface.” arXiv Preprint arXiv:1609.04340v3, Aug. 4, 2018, pp. 1-35. |
Metoui, N. et al. “Differential Privacy Based Access Control.” OTM 2016: On the Move to Meaningful Internet Systems, Oct. 18, 2016, pp. 962-974. |
Mironov, I. “Renyi Differential Privacy.” arXiv Preprint arXiv: 1702.07476v3, Aug. 25, 2017, pp. 1-13. |
Saranya, R. et al. “Precision-Constrained Privacy Preserving Role-Based Access Control.” International Journal of Emerging Technology in Computer Science & Electronics, vol. 13, No. 1, Mar. 2015, pp. 405-408. |
“U.S. Appl. No. 15/166,035, Non Final Office Action mailed May 10, 2018”, 18 pgs. |
“U.S. Appl. No. 15/203,797, Examiner Interview Summary mailed Mar. 26, 2018”, 3 pgs. |
“U.S. Appl. No. 15/203,797, Final Office Action mailed Jun. 8, 2018”, 13 pgs. |
“U.S. Appl. No. 15/203,797, Non Final Office Action mailed Jan. 17, 2018”, 13 pgs. |
“U.S. Appl. No. 15/203,797, Notice of Allowance mailed Sep. 11, 2019”, 8 pgs. |
“U.S. Appl. No. 15/203,797, Response filed Apr. 6, 2018 to Non Final Office Action mailed Jan. 17, 2018”, 19 pgs. |
“U.S. Appl. No. 15/793,898, Examiner Interview Summary mailed Feb. 27, 2018”, 2 pgs. |
“U.S. Appl. No. 15/793,898, Examiner Interview Summary mailed Jun. 18, 2018”, 3 pgs. |
“U.S. Appl. No. 15/793,898, Final Office Action mailed Feb. May 15, 2018”, 14 pgs. |
“U.S. Appl. No. 15/793,898, Non Final Office Action mailed Feb. 7, 2018”, 11 pgs. |
“U.S. Appl. No. 15/793,898, Notice of Allowance mailed Nov. 11, 2018”, 12 pgs. |
“U.S. Appl. No. 15/793,898, Response filed Apr. 6, 2018, to Non Final Office Action mailed Feb. 7, 2018”, 18 pgs. |
“U.S. Appl. No. 15/793,898, Response filed Jul. 10, 2018, to Final Office Action mailed May 15, 2018”, 12 pgs. |
“U.S. Appl. No. 15/793,907, Examiner Interview Summary mailed Feb. 27, 2018”, 2 pgs. |
“U.S. Appl. No. 15/793,907, Examiner Interview Summary mailed Jun. 18, 2018”, 3 pgs. |
“U.S. Appl. No. 15/793,907, Final Office Action mailed May 15, 2018”, 14 pgs. |
“U.S. Appl. No. 15/793,907, Non Final Office Action mailed Jan. 31, 2018”, 11 pgs. |
“U.S. Appl. No. 15/793,907, Notice of Allowance mailed Oct. 26, 2018”, 11 pgs. |
“U.S. Appl. No. 15/793,907, Response filed Apr. 6, 2018 to Non Final Office Action mailed Jan. 31, 2018”, 17 pgs. |
“U.S. Appl. No. 15/793,907, Response filed Jul. 10, 2018 to Final Office Action mailed May 5, 2018”, 11 pgs. |
“U.S. Appl. No. 15/960,486, Non Final Office Action mailed May 10, 2019”, 9 pgs. |
“U.S. Appl. No. 15/960,486, Notice of Allowance mailed Aug. 28, 2019”, 9 pgs. |
“U.S. Appl. No. 15/960,486, Response filed Jul. 18, 2019 to Non Final Office Action mailed May 10, 2019”, 11 pgs. |
“U.S. Appl. No. 16/040,478, Notice of Allowance mailed Aug. 6, 2019”, 11 pgs. |
“U.S. Appl. No. 16/144,790, Non Final Office Action mailed Oct. 17, 2019”, 12 pgs. |
“U.S. Appl. No. 16/144,790, Notice of Allowance mailed Mar. 25, 2020”, 10 pgs. |
“U.S. Appl. No. 16/144,790, Response filed Feb. 12, 2020 to Non Final Office Action mailed Oct. 17, 2019”, 16 pgs. |
“U.S. Appl. No. 16/238,437, Corrected Notice of Allowability mailed Jan. 28, 2020”. |
“U.S. Appl. No. 16/238,437, Non Final Office Action mailed Oct. 31, 2019”, 13 pgs. |
“U.S. Appl. No. 16/238,437, Notice of Allowability mailed Dec. 30, 2019”, 9 pgs. |
“U.S. Appl. No. 16/238,437, Response filed Nov. 21, 2019 to Non Final Office Action mailed Oct. 31, 2019”, 12 pgs. |
“U.S. Appl. No. 16/238,437, Corrected Notice of Allowability mailed Jun. 26, 2020”, 2 pgs. |
“U.S. Appl. No. 16/238,439, Non Final Office Action mailed Oct. 28, 2019”, 15 pgs. |
“U.S. Appl. No. 16/238,437, Notice of Allowance mailed Apr. 10, 2020”, 9 pgs. |
“U.S. Appl. No. 16/238,437, Response filed Feb. 27, 2020 to Non Final Office Action mailed Oct. 28, 2019”, 12 pgs. |
“U.S. Appl. No. 16/810708, Non Final Office Action mailed Nov. 10, 2020”, 18 pgs. |
“U.S. Appl. No. 16/810,708, Notice of Allowance mailed Apr. 20, 2021”, 10 pgs. |
“U.S. Appl. No. 16/810,708, Response filed Mar. 9, 2021 to Non Final Office Action mailed Nov. 10, 2020”, 14 pgs. |
“U.S. Appl. No. 16/225,573, Preliminary Amendment filed Sep. 27, 2023”, 11 pgs. |
“Canadian Application Serial No. 2,998,839, Office Action mailed Jan. 7, 2020”, 3 pgs. |
“Canadian Application Serial No. 2,998,839, Office Action mailed Feb. 12, 2019”, 3 pgs. |
“Canadian Application Serial No. 2,998,839, Response filed May. 5, 2020 to Office Action mailed Jan. 7, 2020”, 16 pgs. |
“Canadian Application Serial No. 2,998,839, Response filed Aug. 7, 2019 to Office Action mailed Feb. 12, 2019”, 17 pgs. |
“European Application Serial No. 16862625.7, Amendment filed Sep. 16, 2022”, 48 pgs. |
“European Application Serial No. 16862625.7, Communication Pursuant to Article 94(3) EPC mailed May 4, 2020”, 9 pgs. |
“European Application Serial No. 16862625.7, decision to refuse mailed May 11, 2022”, 18 pgs. |
“European Application Serial No. 16862625.7, Extended European Search Report mailed Mar. 27, 2019”, 9 pgs. |
“European Application Serial No. 16862625.7, Office Action mailed Oct. 24, 2022”, 2 pgs. |
“European Application Serial No. 16862625.7, Response filed Feb. 25, 2022 to Summons to attend oral proceedings mailed Oct. 22, 2021”, 22 pgs. |
“European Application Serial No. 16862625.7, Response filed Aug. 22, 2019 to Extended European Search Report mailed Mar. 27, 2019”, 12 pgs. |
“European Application Serial No. 16862625.7, Response filed Oct. 30, 2020 to Communication Pursuant to Article 94(3) EPC mailed May 4, 2020”, 9 pgs. |
“European Application Serial No. 16862625.7, Response ro Communication pursuant to Rules 161 and 162 EPC filed Oct. 29, 2018”, 8 pgs. |
“European Application Serial No. 16862625.7, Summoned to attend oral proceedings mailed Oct. 22, 2021”, 9 pgs. |
“European Application Serial No. 20153847.7, Extended European Search Report mailed Apr. 30, 2020”, 11 pgs. |
“European Application Serial No. 20173244.3, Extended European Search Report mailed Sep. 14, 2020”, 11 pgs. |
“International Application Serial No. PCT/US2016/044178, International Preliminary Report on Patentability mailed May 17, 2018”, 9 pgs. |
“International Application Serial No. PCT/US2016/044178, International Search Report mailed Oct. 18, 2016”. 3 pgs. |
“International Application Serial No. PCT/US2016/044178, Written Opinion mailed Oct. 18, 2016”, 7 pgs. |
“International Application Serial No. PCT/US2019/015035, International Search Report mailed Jun. 20, 2019”, 2 pgs. |
“International Application Serial No. PCT/US2019/015035, Written Opinion mailed Jun. 20, 2019”, 3 pgs. |
Agrawal, R., et al., “Privacy-Preserving Data Mining”, ACM SIGMOD, (May 2000), pp. 439-450. |
Amirebekyan, A., et al., “Privacy-Preserving Regression Algorithms”, Proceedings of the 7th WSEAS International Conference on Simulation, Modeling, and Optimization, (2007), 37-45. |
Beigi, G, et al., “Privacy in Social Media: Identification, Mitigation and Applications”, ACM Trans. Web, Vol. 9, No. 4, Article 39, (Jul. 2018), 1-36. |
Bost, R., et al., “Machine Learning Classification over Encrypted Data”, NDSS '15, (Feb. 8-11, 2015), 1-14. |
Chaudhuri, K., et al., “Privacy-preserving logistic regression”, Advances in Neural Information Processing Systems, (2009), 8 pages. |
Cock, M. D, et al., “Fast, Privacy Preserving Linear Regression over Distributed Datasets based on Pre-Distributed Data”, Proceedings ofthe 8th ACM Workshop on Artificial Intelligence and Security, (2015), 3-14. |
Du, Wenliang, et al., “Privacy-Preserving Multivariate Statistical Analysis: Linear Regression and Classification”, Electrical Engineering and Computer Science. Paper 12, Proceedinqs of the 2004 SIAM International Conference on Data Mininq, [Online]. Retrieved from the Internet: <URL: https://surface.syr.edu/cgi/viewcontent.cgi?referer=&httpsredir=1&article=1011&context=eecs<, (Jan. 1, 2004), 222-233. |
Dwork, C., et al., “A Firm Foundation for Private Data Analysis”, Proceedings of the ACM, Vol. 54, Issue 1, (Jan. 2011), 8 pages. |
Dwork, C., et al., “Calibrating noise to sensitivity in private data analysis”, In Proceedings of the Third Conference on Theory of Cryptography, ser. TCC'06. Berlin, Heidelberg: Springer-Verlag, [Online] Retrieved from the internet: <http://dx.doi.org/10.1007/1168187814>, (2006), 265-284. |
Dwork, C., et al., “Differential Privacy and Robust Statistics”, Proceedings ofthe Forty- First Annual ACM Symposium on Theory of Computing, [Online] Retrieved from the internet: <http://www.stat.cmu.edu/˜jingle/dprs_stoc09.pdf>. (Nov. 14, 2008). 41 pages. |
Dwork, Cynthia, et al., “Differential Privacy: A Survey of Results”, TAMC, LNCS 4978, Aqrawal, M. et al, (eds.), (2008), 1-19. |
Fang, W., et al., “Privacy preserving linear regression modeling ofdistributed databases”, Optimization Letters, vol. 7, (2013), pp. 807-818. |
Fletcher, S, et al., “A Differentially Private Decision Forest”, Proceedings ofthe 13th Australasian Data Mining Conference (AusDM), Sydney, Australia, vol. 168, (2015), 99-108. |
Frades, M R, “Overview on Techniques in Cluster Analysis”, Bioinformaticsin Clinical Research, Methods in Molecular Biology (Methods and Protocols), vol. 593, (2010), pp. 81-107. |
Fraley, C., et al., “How Many Clusters? Which Clustering Method? Answers Via Model-Based Cluster Analvsis”. The Computer Journal, Vol. 41, No. 8. (1998). 578-588. |
Friedman, A., et al., “Data Mining with Differential Privacy”, Proceedings of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, [Online] Retrieved from the internet: <http://users.cis.flu.edu/˜Izhen001/activities/KDD_USB_key_2010/docs/p493.pdf>, (Dec. 2010), 10 pages. |
Geumlek, J, et al., “Renyi Differential Privacy Mechanisms for Posterior Sampling”, NIPS 2017: Advances in Neural Information Processing Systems 30, (Oct. 2, 2017), 1-34. |
Han, S., et al., “Privacy-Preserving Gradient-Descent Methods”, IEEE Transactions on Knowledge and Data Engineering, 22(6), (Jun. 29, 2009), pp. 884-899. |
Huang, Yiqing, et al., “Telco Churn Prediction with Big Data”, Proceedings ofthe 2015 ACM SIGMOD International Conference on Management of Data, [Online] Retrieved from the internet: <http://users.wpi.edu/˜yli15/Includes/SIGMOD15Telco.pdf>, (Jun. 4, 2015), 13 pgs. |
Jagannathan, G., et al., “A Practical Differentially Private Random Decision Tree Classifier”, International Conference on Data Mining Workshops, Proceedings ofthe ICDM International Workshop on the Privacv Aspects of Data Mining. (2009). 8 pages. |
Jayaraman, B., et al., “Evaluating Differentially Private Machine Learning in Practice”, 28th USENIX Security Symposium, (Feb. 2019). 1-18. |
Ji, Z., et al., “Differeniial rPriVacy andMachine Learning: a Survey and Review”, Cornell University Library—arXiv preprint, [Online] Retrieved from the internet: <http://arxiv.org/pdf/1412.7584.pdf>, (Dec. 24, 2014), 30 pages. |
Kellaris, G, et al., “Practical differential privacy via grouping and smoothing”, Proceedings of the VLDB Endowment vol. 6, No. 5, (Mar. 1, 2013), 301-312. |
Koufogiannis, F., et al., “Gradual Release of Sensitive Data under Differential Privacy”, Cornel University,CrthoaraDhv and Security,, (Oct. 15. 2018). 1-22. |
Liu, H, et al., “Privacy-Presenting Monotonicity of Differential Privacy Mechanisms”, Applied Sciences, vol. 8,No. 11, (Oct. 28, 2018), 1-32. |
Nissim, K., et al., “Smooth Sensitivity and Sampling in Private Data Analysis”, Proceedings of the Thirty-Ninth Annual ACM Symposium on Theory of Computing, [Online]. Retrieved from the Internet: <http://www.cse.psu.edu/˜sxr48/pubs/smooth-sensitivity-stoc.pdf>, (Jun. 11-13, 2007), 75-84. |
Patil, A., et al., “Differential Private Random Forest”, International Conference on Advances in Computing, Communications and Informatics, [Online] Retrieved from the internet: <http://ieeexplore.ieee.org/stamp/stamp.jsp?tp-&arnumber=6968348&isnumber=6968191>, (Sep. 27, 2014), 9 pages. |
Peng, S., et al., “Query Optimization for Differentially Private Data Management Systems”, ICDE Conference, (2013), pp. 1093-1104. |
Sanil, Ashish P, et al.; “Privacy Preserving Regression Modelling Via Distributed Computation”, Proceedings ofthe Tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, (2004), 677-682. |
Shang, S, et al., “The Application of Differential Privacy for Rank Aggregation: Privacy and Accuracy”, 17th International Conference on Information Fusion, (Jul. 7, 2014). 1-7. |
Xiao, X., et al., “Differential privacy via wavelet transforms”, IEEE Transactions on Knowledge and Data Engineering, vol. 23, No. 8, (Aug. 2011), pp. 1200-1214. |
Xiao, X., et al., “iReduct: Differential Privacy with Reduced Relative Errors”, SIGMOD' 11, (Jun. 12-16, 2011), 229-240. |
Xu, J., et al., “Differentially Private Histogram Publication”, IEEE 28th International Conference on Data Engineering, (Apr. 2012), 32-43. |
Zhang, J., et al., “Functional Mechanism: Regression Analysis under Differential Privacy”, Proceedings of the VLDB Endowment, vol. 5, No. 11, (2012), pp. 1364-1375. |
Zhang, N., et al., “Distributed Data Mining with Differential Privacy”, IEEE ICC proceedings (2011), 5 pages. |
Number | Date | Country | |
---|---|---|---|
20210357523 A1 | Nov 2021 | US |
Number | Date | Country | |
---|---|---|---|
62249938 | Nov 2015 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16810708 | Mar 2020 | US |
Child | 17389100 | US | |
Parent | 16238439 | Jan 2019 | US |
Child | 16810708 | US | |
Parent | 15793907 | Oct 2017 | US |
Child | 16238439 | US | |
Parent | 15203797 | Jul 2016 | US |
Child | 15793907 | US |