The disclosure relates to a method for determining a probability of a kidney stone in a subject being a uric-acid (UA) stone, and more particularly to a machine-learning-based method for determining a probability of a kidney stone in a subject being a UA stone.
Nephrolithiasis is a crystallopathy where a patient has a kidney stone developing in his/her urinary tract. According to composition of the kidney stone, nephrolithiasis could be roughly classified as uric-acid (UA) stone disease or non-UA stone disease. Patients with UA stone disease account for 10% to 15% of patients with nephrolithiasis. Many studies have indicated that UA stone disease is related to features of metabolic syndrome and nutrient partitioning disorders (e.g., diabetes mellitus and obesity). Unlike non-UA stone disease, most UA stone diseases do not require surgery to treat.
Dual-energy computed tomography (CT) is considered an accurate test that can be conducted before treatment for nephrolithiasis to differentiate UA stone disease from non-UA stone disease. However, equipment of dual-energy CT is prohibitively expensive for most clinics around the world, and thus is not widely available. In addition, concerns about safety arise due to the radiation exposure associated with using dual-energy CT.
Therefore, an object of the disclosure is to provide a method for determining a probability of a kidney stone in a subject being a uric-acid (UA) stone that can alleviate at least one of the drawbacks of the prior art.
According to the disclosure, the method includes steps of: establishing, by using a machine learning algorithm, a prediction model based on a plurality of training data sets that are related to a plurality of patients, each of the plurality of training data sets at least including an estimated glomerular filtration rate (eGFR) and a value of urine pH; and feeding an input variable set into the prediction model so as to obtain the probability of the kidney stone in the subject being a UA stone, the input variable set being related to the subject and including an eGFR and a value of urine pH of the subject.
Other features and advantages of the disclosure will become apparent in the following detailed description of the embodiment(s) with reference to the accompanying drawings. It is noted that various features may not be drawn to scale.
Before the disclosure is described in greater detail, it should be noted that where considered appropriate, reference numerals or terminal portions of reference numerals have been repeated among the figures to indicate corresponding or analogous elements, which may optionally have similar characteristics.
Referring to
The method at least includes steps S91 and S93 delineated below.
In step S91, the prediction system 2 establishes, by using a machine learning algorithm, a prediction model 22 based on a plurality of training data sets that are related to a plurality of patients. Each of the plurality of training data sets at least includes an estimated glomerular filtration rate (eGFR) and a value of urine pH of a patient. The prediction model 22 is mathematically expressed as
where y represents output of the prediction model 22 (i.e., a probability of a kidney stone being a UA stone) and has a value ranging from zero to one, x represents input to the prediction model 22 and is in the form of a vector, and ƒ(x) is a function of the input to the prediction model 22 and will be further explained later. When the output of the prediction model 22 has a value of one, it means that the probability of the kidney stone being a UA stone is 100%; when the output of the prediction model 22 has a value of zero, it means that the probability of the kidney stone being a UA stone is 0%. However, representation of output of the prediction model 22 is not limited to the disclosure herein and may vary in other embodiments. The prediction model 22 is built using the Python programming language and built-in libraries of Python, and is trained by applying a built-in optimization engine of Python with a weighted binary cross-entropy function serving as the objective function. Model parameters of the prediction model 22 are trained based on the training data sets. Since implementing a prediction model by using a machine learning algorithm has been well known to one skilled in the relevant art (e.g., reference may be made to an MIT Press book “Deep Learning” authored by Goodfellow, Bengio and Courville), detailed explanation of the same is omitted herein for the sake of brevity.
In step S93, the prediction system 2 feeds an input variable set into the prediction model 22 so as to obtain the probability of the kidney stone in the subject being a UA stone. The input variable set is related to the subject and at least includes an eGFR and a value of urine pH of the subject.
In some embodiments, each of the plurality of training data sets further includes an age of a patient, and a gender indicator that indicates gender of the patient (e.g., one for female and zero for male), and a creatinine concentration that is related to the patient. The prediction system 2 includes an eGFR computing unit 21. In such embodiments, the input variable set further includes an age of the subject, a gender indicator that indicates gender of the subject (e.g., one for female and zero for male), and a creatinine concentration that is related to the subject. However, in some embodiments (e.g., the embodiments shown in
Accordingly, in these embodiments, prior to step S93, the method further includes step S92. In step S92, the eGFR computing unit 21 of the prediction system 2 calculates the eGFR based on the age, the gender indicator and the creatinine concentration. Specifically, the eGFR computing unit 21 calculates the eGFR by using the isotope dilution mass spectrometry traceable Modification of Diet in Renal Disease formula that is mathematically expressed as
where Scr represents the creatinine concentration, Age represents the age, and G represents the gender indicator which has a value of one indicating that the subject is a female, or a value of zero indicating that the subject is a male. For information about the isotope dilution mass spectrometry traceable Modification of Diet in Renal Disease formula, reference may be made to “A new equation to estimate glomerular filtration rate” authored by Levey, Stevens, Schmid, Zhang, Castro, Feldman, Kusek, Egger, Van Lente, Greene, et al. in Annals of Internal Medicine, 2009, and detailed explanation thereof is omitted herein for the sake of brevity.
In some embodiments where each of the plurality of training data sets includes the age of a patient, the gender indicator that indicates gender of the patient, a creatinine concentration that is related to the patient, or a combination thereof, the prediction system 2 may omit the eGFR computing unit 21, in which case the input variable set is directly received by the prediction system 2 (e.g., inputted by a user using an input device of the prediction system 2) and then fed into the prediction model 22.
In order to validate performance of the method according to the disclosure, data collected from two groups of patients who have been diagnosed with nephrolithiasis was used. A first group of patients (hereinafter also referred to as “Cohort A”) includes 1098 patients, and 146 patients thereamong (i.e., 13.3% of the patients of Cohort A) have UA stone disease. A second group of patients (hereinafter also referred to as “Cohort B”) includes 71 patients, 3 patients thereamong (i.e., 4.23% of the patients of Cohort B) have UA stone disease. It is worth to note that determination as to whether a patient of Cohort A or Cohort B has UA stone disease or not is made based on analysis of infrared spectroscopy performed on a kidney stone that is taken out from the patient by surgery. For Cohort A, data from sixty percent of the patients in Cohort A was used to serve as the training data sets for training the prediction model 22, and data from the remaining forty percent of the patients in Cohort A was used to serve as validation data sets. It is worth to note that the aforesaid division of the data from Cohort A into the training data sets and the validation data sets is devised such that the proportion of the patients whose data is used as the training data sets and who have UA stone disease to all the patients whose data is used as the training data sets is approximate to the proportion of the patients whose data is used as the validation data sets and who have UA stone disease to all the patients whose data is used as the validation data sets, so as to ensure that there is only a minor statistical difference between the training data sets and the validation data sets. For Cohort B, data related to all the patients in Cohort B was used to serve as validation data sets. Predictive power (i.e., accuracy of prediction) of the prediction model 22 applied to the validation data (which is collected from Cohort A and Cohort B) is evaluated by using receiver operating characteristic (ROC) analysis based on an area under curve (AUC) that is related to an ROC curve. Specifically, in ROC analysis, an optimal cutoff point on the ROC curve is determined based on Youden's index. Then, the optimal cutoff point is used to calculate sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV). Since ROC analysis has been well known to one skilled in the relevant art, detailed explanation of the same is omitted herein for the sake of brevity.
In a first embodiment of the method, each of the plurality of training data sets includes the eGFR and the value of urine pH of a patient, the prediction system 2 includes the eGFR computing unit 21, and the input variable set includes the eGFR and the value of urine pH of the subject. Particularly, the prediction system 2 receives a variable x0 that represents the creatinine concentration, a variable x1 that represents the gender indicator, a variable x2 that represents the age, and a variable x4 that represents the value of urine pH. The eGFR computing unit 21 calculates, based on the variables x0,x1 and x2, a variable x3 that represents the eGFR, and then the variable x3 thus calculated is fed into the prediction model 22 together with the variable x4 as the input variable set. In response to receipt of the input variable set, the prediction model 22 obtains the probability of a kidney stone being a UA stone.
In the mathematical expression of the prediction model 22 according to the first embodiment, ƒ(x) is a function that is mathematically expressed as ƒ(x)=w0+w3(x)ƒ3(x)+w4(x)ƒ4(x), where w0 is a constant, each of w3(x), w4(x), ƒ3(x) and ƒ4(x) is a function, and w0, w3(x), w4(x), ƒ3(x) and ƒ4(x) are determined by using a fully connected neural network. In particular, the function ƒ(x) can be simplified as ƒ(x)=w0+w3ƒ3(x3)+w4ƒ4(x4), where each of w0, w3 and w4 is a scalar trainable parameter, and each of ƒ3(x3) and ƒ4(x4) is a nonlinear function determined by using a fully connected two-layer neural network. For ƒ3(x3), a first layer of the fully connected two-layer neural network has a dimension m3 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ4(x4), a first layer of the fully connected two-layer neural network has a dimension m4 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. Each of ƒ3(x3) and ƒ4(x4) is mathematically expressed as
where i is equal to 3 or 4, each of Ai,1 and Bi,1 is an mi-by-1 vector containing mi number of trainable model parameters, Ai,2 is a 1-by-mi vector containing mi number of trainable model parameters, Bi,2 is a scalar trainable parameter, and each of σi,1 and σi,2 is a standard nonlinear activation function commonly used in building a neural network. In total, the function ƒ(x) contains Σi=343mi+5 number of trainable model parameters. In this embodiment, each of m3 and ma is equal to 20, each of B3,1 and B4,1 is manually set as a zero vector, and the function ƒ(x) contains only 85 trainable model parameters, accordingly.
For the prediction model 22 according to the first embodiment, an ROC curve for the prediction model 22 fed with the validation data sets collected from Cohort A is illustrated in
Referring to
In the mathematical expression of the prediction model 22 according to the second embodiment, ƒ(x) is a function that is mathematically expressed as ƒ(x)=w0+w3(x)ƒ3(x)+w4(x)ƒ4(x)+w5(x)ƒ5(x), where w0 is a constant, each of w3(x), w4(x), w5(x), ƒ3(x), ƒ4(x) and ƒ5(x) is a function, and w0, w3(x), w4(x), w5(x), ƒ3(x), ƒ4(x) and ƒ5(x) are determined by using a fully connected neural network. In particular, the function ƒ(x) can be simplified as ƒ(x)=w0+w3ƒ3(x3)+w4ƒ4(x4)+w5ƒ5(x5), where each of w0, w3, w4 and w5 is a scalar trainable parameter, and each of ƒ3(x3), ƒ4(x4) and ƒ5(x5) is a nonlinear function determined by using a fully connected two-layer neural network. For ƒ3(x3), a first layer of the fully connected two-layer neural network has a dimension m3 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ4(x4), a first layer of the fully connected two-layer neural network has a dimension m4 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ5(x5), a first layer of the fully connected two-layer neural network has a dimension m5 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. Each of ƒ3(x3), ƒ4(x4) and ƒ5(x5) is mathematically expressed as
where i is equal to 3, 4 or 5, each of Ai,1 and Bi,1 is an mi-by-1 vector containing mi number of trainable model parameters, Ai,2 is a 1-by-mi vector containing mi number of trainable model parameters, Bi,2 is a scalar trainable parameter, and each of σi,1 and σi,2 is a standard nonlinear activation function commonly used in building a neural network. In total, the function ƒ(x) contains Σi=353mi+7 number of trainable model parameters. In this embodiment, each of m3, m4 and m5 is equal to 20, each of B3,1, B4,1 and B5,1 is manually set as a zero vector, and the function ƒ(x) contains only 127 trainable model parameters, accordingly.
For the prediction model 22 according to the second embodiment, an ROC curve for the prediction model 22 fed with the validation data sets collected from Cohort A is illustrated in
In a third embodiment of the method, each of the plurality of training data sets includes the eGFR, the value of urine pH, the BMI and the age of a patient. The prediction system 2 includes the eGFR computing unit 21, and the input variable set includes the eGFR, the value of urine pH, the BMI and the age of the subject. Particularly, the prediction system 2 receives a variable x0 that represents the creatinine concentration, a variable x1 that represents the gender indicator, a variable x2 that represents the age, a variable x4 that represents the value of urine pH, and a variable x5 that represents the BMI. The eGFR computing unit 21 calculates, based on the variables x0,x1 and x2, a variable x3 that represents the eGFR, and then the variable x3 thus calculated is fed into the prediction model 22 together with the variables x2,x4 and x5 as the input variable set. In response to receipt of the input variable set, the prediction model 22 obtains the probability of a kidney stone being a UA stone.
In the mathematical expression of the prediction model 22 according to the third embodiment, ƒ(x) is a function that is mathematically expressed as ƒ(x)=w0+w2(x)ƒ2(x)+w3(x)ƒ3(x)+w4(x)ƒ4(x)+w5(x)ƒ5(x), where w0 is a constant, each of w2(x), w3(x), w4(x), w5(x), ƒ2(x), ƒ3(x), ƒ4(x) and ƒ5(x) is a function, and w0, w2(x), w3(x), w4(x), w5(x), ƒ2(x), ƒ3(x), ƒ4(x) and ƒ5(x) are determined by using a fully connected neural network. In particular, the function ƒ(x) can be simplified as ƒ(x)=w0+w2ƒ2(x2)+w3ƒ3(x3)+w4ƒ4(x4)+w5ƒ5(x5), where each of w0, w2, w3, w4 and w5 is a scalar trainable parameter, and each of ƒ2(x2), ƒ3(x3), ƒ4(x4) and ƒ5(x5) is a nonlinear function determined by using a fully connected two-layer neural network. For ƒ2(x2), a first layer of the fully connected two-layer neural network has a dimension m2 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ3(x3), a first layer of the fully connected two-layer neural network has a dimension m3 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For fa (x4), a first layer of the fully connected two-layer neural network has a dimension m4 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ5(x5), a first layer of the fully connected two-layer neural network has a dimension m5 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. Each of ƒ2(x2), ƒ3(x3), ƒ4(x4) and ƒ5(x5) is mathematically expressed as
where i is equal to 2, 3, 4 or 5, each of Ai,1 and Bi,1 is an mi-by-1 vector containing mi number of trainable model parameters, Ai,2 is a 1-by-mi vector containing mi number of trainable model parameters, Bi,2 is a scalar trainable parameter, and each of σi,1 and σi,2 is a standard nonlinear activation function commonly used in building a neural network. In total, the function ƒ(x) contains Σi=253mi+9 number of trainable model parameters. In this embodiment, each of m2, m3, m4 and m5 is equal to 20, each of B2,1, B3,1, B4,1 and B5,1 is manually set as a zero vector, and the function ƒ(x) contains only 169 trainable model parameters, accordingly.
For the prediction model 22 according to the third embodiment, an ROC curve for the prediction model 22 fed with the validation data sets collected from Cohort A is illustrated in
In a fourth embodiment of the method, each of the plurality of training data sets includes the eGFR, the value of urine pH, the BMI and the age of a patient, and the gender indicator indicating gender of the patient. The prediction system 2 includes the eGFR computing unit 21, and the input variable set includes the eGFR, the value of urine pH, the BMI and the age of the subject, and the gender indicator indicating gender of the subject. Particularly, the prediction system 2 receives a variable x0 that represents the creatinine concentration, a variable x1 that represents the gender indicator, a variable x2 that represents the age, a variable x4 that represents the value of urine pH, and a variable x5 that represents the BMI. The eGFR computing unit 21 calculates, based on the variables x0,x1 and x2, a variable x3 that represents the eGFR, and then the variable x3 thus calculated is fed into the prediction model 22 together with the variables x1,x2,x4 and x5 as the input variable set for the prediction model 22 to obtain the probability of a kidney stone being a UA stone.
In the mathematical expression of the prediction model 22, ƒ(x) is a function that is mathematically expressed as ƒ(x)=w0+w1(x)ƒ1(x)+w2(x)ƒ2(x)+w3(x)ƒ3(x)+w4(x)ƒ4(x)+w5(x)ƒ5(x), where w0 is a constant, each of w1(x), w2(x), w3(x), w4(x), w5(x), ƒ1(x), ƒ2(x), ƒ3(x), ƒ4(x) and ƒ5(x) is a function, and w0, w1(x), w2(x), w3(x), w4(x), w5(x), ƒ1(x), ƒ2(x), ƒ3(x), ƒ4(x) and ƒ5(x) are determined by using a fully connected neural network. In particular, the function ƒ(x) can be simplified as ƒ(x)=w0+w1ƒ1(x1)+w2ƒ2(x2)+w3ƒ3(x3)+w4ƒ4(x4)+w5ƒ5(x5), where each of w0, w1, w2, w3, w4 and w5 is a scalar trainable parameter, and each of ƒ1(x1), ƒ2(x2), ƒ3(x3), ƒ4(x4) and ƒ5(x5) is a nonlinear function determined by using a fully connected two-layer neural network. For ƒ1(x1), a first layer of the fully connected two-layer neural network has a dimension m1 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ2(x2), a first layer of the fully connected two-layer neural network has a dimension m2 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ3(x3), a first layer of the fully connected two-layer neural network has a dimension m3 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ4(x4), a first layer of the fully connected two-layer neural network has a dimension m4 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ5(x5), a first layer of the fully connected two-layer neural network has a dimension m5 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. Each of ƒ1(x1), ƒ2(x2), ƒ3(x3), ƒ4(x4) and ƒ5(x5) is mathematically expressed as
where i is equal to 1, 2, 3, 4 or 5, each of Ai,1 and Bi,1 is an mi-by-1 vector containing mi number of trainable model parameters, Ai,2 is a 1-by-mi vector containing mi number of trainable model parameters, Bi,2 is a scalar trainable parameter, and each of σi,1 and σi,2 is a standard nonlinear activation function commonly used in building a neural network. In total, the function ƒ(x) contains Σi=153mi+11 number of trainable model parameters. In this embodiment, each of m1, m2, m3, m4 and m5 is equal to 20, each of B1,1, B2,1, B3,1, B4,1 and B5,1 is manually set as a zero vector, and the function ƒ(x) contains only 211 trainable model parameters, accordingly.
For the prediction model 22 according to the fourth embodiment, an ROC curve for the prediction model 22 fed with the validation data sets collected from Cohort A is illustrated in
In a fifth embodiment of the method, each of the plurality of training data sets includes the eGFR, the value of urine pH, the BMI and the age of a patient, the gender indicator indicating gender of the patient, and a diabetes mellitus (DM) indicator indicating whether the patient was ever diagnosed with DM. For example, the DM indicator may have a value of one to indicate that the patient had once been diagnosed with DM, or a value of zero to indicate that the patient was never diagnosed with DM. The prediction system 2 includes the eGFR computing unit 21.
In the fifth embodiment, the input variable set includes the eGFR, the value of urine pH, the BMI and the age of the subject, the gender indicator indicating gender of the subject, and a DM indicator indicating whether the subject was ever diagnosed with DM. For example, the DM indicator may have a value of one to indicate that the subject had once been diagnosed with DM, or a value of zero to indicate that the subject was never diagnosed with DM. Particularly, the prediction system 2 receives a variable x0 that represents the creatinine concentration, a variable x1 that represents the gender indicator, a variable x2 that represents the age, a variable x4 that represents the value of urine pH, a variable x5 that represents the BMI, and a variable x6 that represents the DM indicator. The eGFR computing unit 21 calculates, based on the variables x0,x1 and x2, a variable x3 that represents the eGFR, and then the variable x3 thus calculated is fed into the prediction model 22 together with the variables x1,x2,x4,x5 and x6 as the input variable set for the prediction model 22 to obtain the probability of a kidney stone being a UA stone.
In the mathematical expression of the prediction model 22, ƒ(x) is a function that is mathematically expressed as ƒ(x)=w0+w1(x)ƒ1(x)+w2(x)ƒ2(x)+w3(x)ƒ3(x)+w4(x)ƒ4(x)+w5(x)ƒ5(x)+w6(x)ƒ6(x), where w0 is a constant, each of w1(x), w2(x), w3(x), w4(x), w5(x), w6(x), ƒ1(x), ƒ2(x), ƒ3(x), ƒ4(x), ƒ5(x) and ƒ6(x) is a function, and w0, w1(x), w2(x), w3(x), w4(x), w5(x), w6(x), ƒ1(x), ƒ2(x), ƒ3(x), ƒ4(x), ƒ5(x) and ƒ6(x) are determined by using a fully connected neural network. In particular, the function ƒ(x) can be simplified as ƒ(x)=w0+w1ƒ1(x1)+w2ƒ2(x2)+w3ƒ3(x3)+w4ƒ4(x4)+w5ƒ5(x5)+w6ƒ6(x6), where each of w0, w1, w2, w3, w4, w5 and w6 is a scalar trainable parameter, and each of ƒ1(x1), ƒ2(x2), ƒ3(x3), ƒ4(x4), ƒ5(x5) and ƒ6(x6) is a nonlinear function determined by using a fully connected two-layer neural network. For ƒ1(x1), a first layer of the fully connected two-layer neural network has a dimension m1 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ2(x2), a first layer of the fully connected two-layer neural network has a dimension m2 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ3(x3), a first layer of the fully connected two-layer neural network has a dimension m3 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ4(x4), a first layer of the fully connected two-layer neural network has a dimension m4 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ5(x5), a first layer of the fully connected two-layer neural network has a dimension m5 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ6(x6), a first layer of the fully connected two-layer neural network has a dimension m6 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. Each of ƒ1(x1), ƒ2(x2), ƒ3(x3), ƒ4(x4), ƒ5(x5) and ƒ6(x6) is mathematically expressed as
where i is equal to 1, 2, 3, 4, 5 or 6, each of Ai,1 and Bi,1 is an mi-by-1 vector containing mi number of trainable model parameters, Ai,2 is a 1-by-mi vector containing mi number of trainable model parameters, Bi,2 is a scalar trainable parameter, and each of σi,1 and σ1,2 is a standard nonlinear activation function commonly used in building a neural network. In total, the function ƒ(x) contains Σi=163mi+13 number of trainable model parameters. In this embodiment, each of m1, m2, m3, m4, m5 and m6 is equal to 20, each of B1,1, B2,1, B3,1, B4,1, B5,1 and B6,1 is manually set as a zero vector, and the function ƒ(x) contains only 253 trainable model parameters, accordingly.
For the prediction model 22 according to the fifth embodiment, an ROC curve for the prediction model 22 fed with the validation data sets collected from Cohort A is illustrated in
In a sixth embodiment of the method, each of the plurality of training data sets includes the eGFR, the value of urine pH, the BMI and the age of a patient, the gender indicator indicating gender of the patient, the DM indicator indicating whether the patient was ever diagnosed with DM, a gout indicator indicating whether the patient was ever diagnosed with gout, and a bacteriuria indicator indicating whether the patient was ever diagnosed with bacteriuria. For example, the gout indicator may have a value of one to indicate that the patient had once been diagnosed with gout, or a value of zero to indicate that the patient was never diagnosed with gout. Similarly, the bacteriuria indicator may have a value of one to indicate that the patient had once been diagnosed with bacteriuria, or a value of zero to indicate that the patient was never diagnosed with bacteriuria. The prediction system 2 includes the eGFR computing unit 21.
In the sixth embodiment, the input variable set includes the eGFR, the value of urine pH, the BMI and the age of the subject, the gender indicator indicating gender of the subject, the DM indicator indicating whether the subject was ever diagnosed with DM, a gout indicator indicating whether the subject was ever diagnosed with gout, and a bacteriuria indicator indicating whether the subject was ever diagnosed with bacteriuria. For example, the gout indicator may have a value of one to indicate that the subject had once been diagnosed with gout, or a value of zero to indicate that the subject was never diagnosed with gout. Similarly, the bacteriuria indicator may have a value of one to indicate that the subject had once been diagnosed with bacteriuria, or a value of zero to indicate that the subject was never diagnosed with bacteriuria. Particularly, the prediction system 2 receives a variable x0 that represents the creatinine concentration, a variable x1 that represents the gender indicator, a variable x2 that represents the age, a variable x4 that represents the value of urine pH, a variable x5 that represents the BMI, a variable x6 that represents the DM indicator, a variable x7 that represents the gout indicator and a variable x8 that represents the bacteriuria indicator. The eGFR computing unit 21 calculates, based on the variables x0,x1 and x2, a variable x3 that represents the eGFR, and then the variable x3 thus calculated is fed into the prediction model 22 together with the variables x1,x2,x4,x5,x6,x7 and x8 as the input variable set for the prediction model 22 to obtain the probability of a kidney stone being a UA stone.
In the mathematical expression of the prediction model 22, ƒ(x) is a function that is mathematically expressed as ƒ(x)=w0+w1(x)ƒ1(x)+w2(x)ƒ2(x)+w3(x)ƒ3(x)+w4(x)ƒ4(x)+w5(x)ƒ5(x)+w6(x)ƒ6(x), where w0 is a constant, each of w1(x), w2(x), w3(x), w4(x), w5(x), w6(x), ƒ1(x), ƒ2(x), ƒ3(x), ƒ4(x), ƒ5(x) and ƒ6(x) is a function, and w0, w1(x), w2(x), w3(x), w4(x), w5(x), w6(x), ƒ1(x), ƒ2(x), ƒ3(x), ƒ4(x), ƒ5(x) and ƒ6(x) are determined by using a fully connected neural network. In particular, the function ƒ(x) can be simplified as ƒ(x)=w0+w1ƒ1(x1)+w2ƒ2(x2)+w3ƒ3(x3)+w4ƒ4(x4)+w5ƒ5(x5)+w6ƒ6(x6,x7,x8), where each of w0, w1, w2, w3, w4, w5 and w6 is a scalar trainable parameter, and each of ƒ1(x1), ƒ2(x2), ƒ3(x3), ƒ4(x4), ƒ5(x5) and ƒ6(x6,x7,x8) is a nonlinear function determined by using a fully connected two-layer neural network. For ƒ1(x1), a first layer of the fully connected two-layer neural network has a dimension m1 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ2(x2), a first layer of the fully connected two-layer neural network has a dimension m2 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ3(x3), a first layer of the fully connected two-layer neural network has a dimension m3 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ4(x4), a first layer of the fully connected two-layer neural network has a dimension m4 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ5(x5), a first layer of the fully connected two-layer neural network has a dimension m5 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ6(x6,x1,x8), a first layer of the fully connected two-layer neural network has a dimension m6 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. Each of ƒ1(x1), ƒ2(x2), ƒ3(x3), ƒ4(x4) and ƒ5(x5) is mathematically expressed as
and
ƒ6(x6,x7,x8) is mathematically expressed as
where i is equal to 1, 2, 3, 4 or 5, each of Ai,1 and Bi,1 is an mi-by-1 vector containing mi number of trainable model parameters, Ai,2 is a 1-by-mi vector containing mi number of trainable model parameters, Bi,2 is a scalar trainable parameter, and each of σi,1 and σ1,2 is a standard nonlinear activation function commonly used in building a neural network, each of A6,1, A7,1, A8,1 and B6,1 is an m6-by-1 vector containing m6 number of trainable model parameters, A6,2 is a 1-by-m6 vector containing m6 number of trainable model parameters, B6,2 is a scalar trainable parameter, and each of σ6,1 and σ6,2 is a standard nonlinear activation function commonly used in building a neural network. In total, the function ƒ(x) contains Σi=163mi+2m6+13 number of trainable model parameters. In this embodiment, each of m1, m2, m3, m4, m5 and m6 is equal to 20, each of B1,1, B2,1, B3,1, B4,1, B5,1 and B6,1 is manually set as a zero vector, and the function ƒ(x) contains only 293 trainable model parameters, accordingly.
For the prediction model 22 according to the sixth embodiment, an ROC curve for the prediction model 22 fed with the validation data sets collected from Cohort A is illustrated in
Referring to
In the seventh embodiment, the input variable set includes the eGFR, the value of urine pH, the BMI and the age of the subject, the gender indicator indicating gender of the subject, the DM indicator indicating whether the subject was ever diagnosed with DM, the gout indicator indicating whether the subject was ever diagnosed with gout, the bacteriuria indicator indicating whether the subject was ever diagnosed with bacteriuria, and a hypertension indicator indicating whether the subject was ever diagnosed with hypertension. For example, the hypertension indicator may have a value of one to indicate that the subject had once been diagnosed with hypertension, or a value of zero to indicate that the subject was never diagnosed with hypertension. Particularly, the prediction system 2 receives a variable x0 that represents the creatinine concentration, a variable x1 that represents the gender indicator, a variable x2 that represents the age, a variable x4 that represents the value of urine pH, a variable x5 that represents the BMI, a variable x6 that represents the DM indicator, a variable x7 that represents the gout indicator, a variable x8 that represents the bacteriuria indicator, and a variable x9 that represents the hypertension indicator. The eGFR computing unit 21 calculates, based on the variables x0,x1 and x2, a variable x3 that represents the eGFR, and then the variable x3 thus calculated is fed into the prediction model 22 together with the variables x1,x2,x4,x5,x6,x7,x8 and x9 as the input variable set for the prediction model 22 to obtain the probability of a kidney stone being a UA stone.
In the mathematical expression of the prediction model 22, ƒ(x) is a function that is mathematically expressed as ƒ(x)=w0+w1(x)ƒ1(x)+w2(x)ƒ2(x)+w3(x)ƒ3(x)+w4(x)ƒ4(x)+w5(x)ƒ5(x)+w6(x)ƒ6(x), where w0 is a constant, each of w1(x), w2(x), w3(x), w4(x), w5(x), w6(x), ƒ1(x), ƒ2(x), ƒ3(x), ƒ4(x), ƒ5(x) and ƒ6(x) is a function, and w0, w1(x), w2(x), w3(x), w4(x), w5(x), w6(x), ƒ1(x), ƒ2(x), ƒ3(x), ƒ4(x), ƒ5(x) and ƒ6(x) are determined by using a fully connected neural network. In particular, the function ƒ(x) can be simplified as ƒ(x)=w0+w1ƒ1(x1)+w2ƒ2(x2)+w3ƒ3(x3)+w4ƒ4(x4)+w5ƒ5(x5)+w6(x9) ƒ6(x6,x7,x8), where each of w0, w1, w2, w3, w4 and w5 is a scalar trainable parameter, and each of ƒ1(x1), ƒ2(x2), ƒ3(x3), ƒ4(x4), ƒ5(x5), w6(x9) and ƒ6(x6,x1,x8) is a nonlinear function determined by using a fully connected two-layer neural network. For ƒ1(x1), a first layer of the fully connected two-layer neural network has a dimension m1 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ2(x2), a first layer of the fully connected two-layer neural network has a dimension m2 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ3(x3), a first layer of the fully connected two-layer neural network has a dimension m3 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ4(x4), a first layer of the fully connected two-layer neural network has a dimension m4 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ5(x5), a first layer of the fully connected two-layer neural network has a dimension m5 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For ƒ6(x6,x7,x8), a first layer of the fully connected two-layer neural network has a dimension m6 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. For w6(x9), a first layer of the fully connected two-layer neural network has a dimension m7 that is an integer ranging from 1 to 100, and a second layer of the fully connected two-layer neural network has a dimension of 1. Each of ƒ1(x1), ƒ2(x2), ƒ3(x3), ƒ4(x4) and ƒ5(x5) is mathematically expressed as
ƒ6(x6,x7,x8) is mathematically expressed as
and
w6(x9) is mathematically expressed as
where i is equal to 1, 2, 3, 4, or 5, each of Ai,1 and Bi,1 is an mi-by-1 vector containing mi number of trainable model parameters, Ai,2 is a 1-by-mi vector containing m; number of trainable model parameters, Biz is a scalar trainable parameter, and each of σi,1 and σ1,2 is a standard nonlinear activation function commonly used in building a neural network, each of A6,1, A7,1, A8,1 and B6,1 is a m6-by-1 vector containing m6 number of trainable model parameters, A6,2 is a 1-by-m6 vector containing m6 number of trainable model parameters, B6,2 is a scalar trainable parameter, each of σ6,1 and σ6,2 is a standard nonlinear activation function commonly used in building a neural network, each of A9,1, and B7,1 is an m7-by-1 vector containing my number of trainable model parameters, A7,2 is a 1-by-m7 vector containing my number of trainable model parameters, each of B7,2 and w6,0 is a scalar trainable parameter, and each of σ7,1 and σ7,2 is a standard nonlinear activation function commonly used in building a neural network. In total, the function ƒ(x) contains Σi=173mi+2m6+14 number of trainable model parameters. In this embodiment, each of m1, m2, m3, m4, m5 and m6 is equal to 20, m7 is equal to 5, each of B1,1, B2,1, B3,1, B4,1, B5,1, B6,1 and B7,1 is manually set as a zero vector, and the function ƒ(x) contains only 304 trainable model parameters, accordingly.
For the prediction model 22 according to the fifth embodiment, an ROC curve for the prediction model 22 fed with the validation data sets collected from Cohort A is illustrated in
Referring to
In one embodiment, the prediction model is mathematically expressed as
where y represents output of the prediction model 22, x represents input to the prediction model 22, and ƒ(x) is a function that is mathematically expressed as
where n is a positive integer greater than one, i is a positive integer ranging between one and n, w0 is a constant, each of w1(x), wi(x), wn(x), ƒ1(x), ƒi(x) and ƒn(x) is a function, and w0, w1(x), wi(x), wn(x), ƒ1(x), ƒi(x) and ƒn(x) are determined by using a fully connected neural network.
To sum up, for the method for determining a probability of a kidney stone in a subject being a UA stone according to the disclosure, a machine-learning-based prediction model is established and is utilized to obtain the probability based on an input variable set that is related to the subject and that includes at least an eGFR and a value of urine pH, which are easily available at an initial clinical visit of the subject. It is worth to note that results of validation exemplarily shown by AUCs of ROC curves illustrated in
In the description above, for the purposes of explanation, numerous specific details have been set forth in order to provide a thorough understanding of the embodiment(s). It will be apparent, however, to one skilled in the art, that one or more other embodiments may be practiced without some of these specific details. It should also be appreciated that reference throughout this specification to “one embodiment,” “an embodiment,” an embodiment with an indication of an ordinal number and so forth means that a particular feature, structure, or characteristic may be included in the practice of the disclosure. It should be further appreciated that in the description, various features are sometimes grouped together in a single embodiment, figure, or description thereof for the purpose of streamlining the disclosure and aiding in the understanding of various inventive aspects; such does not mean that every one of these features needs to be practiced with the presence of all the other features. In other words, in any described embodiment, when implementation of one or more features or specific details does not affect implementation of another one or more features or specific details, said one or more features may be singled out and practiced alone without said another one or more features or specific details. It should be further noted that one or more features or specific details from one embodiment may be practiced together with one or more features or specific details from another embodiment, where appropriate, in the practice of the disclosure.
While the disclosure has been described in connection with what is (are) considered the exemplary embodiment(s), it is understood that this disclosure is not limited to the disclosed embodiment(s) but is intended to cover various arrangements included within the spirit and scope of the broadest interpretation so as to encompass all such modifications and equivalent arrangements.