LINEAR MACHINE LEARNING METHOD BASED ON DNA HYBRIDIZATION REACTION TECHNOLOGY

Information

  • Patent Application
  • 20250029016
  • Publication Number
    20250029016
  • Date Filed
    June 25, 2024
    7 months ago
  • Date Published
    January 23, 2025
    11 days ago
Abstract
A new linear machine learning method based on DNA hybridization reaction technology, includes a machine learning training part, an algorithm part, and a testing part. This machine learning method has the ability to learn linear functions. Unlike silicon circuits, the learning algorithm is implemented through the synchronization of DNA hybridization reactions. Therefore, the calculation mode of this machine learning method is a parallel computing model, and the weights of this machine learning are obtained through training without the involvement of electronic computers. Through the method, it is possible to learn multivariable linear functions without any limitation on the number of input terms. Due to the non-negative DNA concentration, the method used a dual track model for negative data processing operations.
Description
TECHNICAL FIELD

The disclosure relates to the fields of biocomputing and machine learning, and particularly to a linear machine learning method based on deoxyribonucleic acid (DNA) hybridization reaction technology.


BACKGROUND

Since the invention of electronic computer in 1946, computer technology has penetrated into every aspect of people's lives and work, making a huge contribution to the development of society. Due to the rapid development of science and technology in recent years, the calculation time for solving technological problems using traditional electronic computers has increased exponentially with the increase of problem-solving scale. In order to meet the growing demand for large-scale and ultra-large-scale computing, the characteristics of high-performance computing and accelerated computing performance are needed. Breaking through the constraints of silicon semiconductor devices and developing non-traditional computers are important ways of future computing technology. DNA computing based on DNA molecules includes DNA computing models based on DNA hybridization reactions, DNA computing models based on DNA enzyme, DNA computing models based on DNA tile, and DNA computing models based on nanoparticle. DNA hybridization technology is an important research method in the DNA computing. The driving force of DNA hybridization comes from the intermolecular forces under the condition of complementary Waston Crick bases. The DNA hybridization reaction process has features of parallelism, programmability, autonomy, dynamic cascading, high information storage, and low power consumption, and spontaneously occurs at room temperature. By utilizing single-molecule self-assembly and fluorescence labeling, the DNA hybridization technology has been applied in biosensors, molecular detection DNA nanorobots, drug delivery and diagnosis and treatment. In addition, the DNA hybridization technology can combine with nanoparticles, quantum dots, and proteins to promote the development of parallel computing models, integrated cryptography, and nanoelectronics.


Learning ability is an important indicator of human intelligence, and the purpose of machine learning is to enable machines to have learning ability. The machine learning includes supervised learning (SL), unsupervised learning (UL), semi-supervised learning (SSL), deep learning (DL), reinforcement learning (RL), and transfer learning (TR). Methods of the machine learning include mechanical learning, inductive learning, explanation-based learning, genetic-algorithm-based learning, and neural-network-based learning. By utilizing the controllability and the ability to be compiled of DNA molecules, various algorithms of machine learning are implemented to achieve a part of functions of the human brain.


At present, the learning process of most artificial intelligence systems based on DNA molecular computing requires the assistance of electronic computers. The weight update in artificial intelligence systems is not completed by DNA molecular circuits, but by the assistance of electronic computing or existing databases (such as large handwritten digit databases). The weight update process is the learning and training process of artificial intelligence systems, and it is also the core content of artificial intelligence systems. Therefore, most artificial intelligence systems based on DNA molecular computing currently do not have true learning ability, only have certain recognition and classification functions.


Given the above, the present disclosure provides a new linear machine learning method based on DNA hybridization reaction technology.


SUMMARY

At present, the learning process of most artificial intelligence systems based on DNA molecular computing requires the assistance of electronic computers. In order to overcome the shortcomings of the prior art, the present disclosure provides a new linear machine learning method based on DNA hybridization reaction technology (also referred to as linear machine learning method based on DNA hybridization reaction technology). A machine learning system based on DNA molecular circuits provided by the present disclosure can complete the training and testing process without the assistance of electronic computers, achieving complete biological machine learning.


The system includes three parts, including a machine learning training part, an algorithm part, and a testing part. The linear machine learning method based on DNA hybridization reaction technology has the ability to learn linear functions. Unlike silicon circuits, the learning algorithm is implemented through the synchronization of DNA hybridization reactions. Therefore, the computing mode of the machine learning method is a parallel computing model, and the weights of this machine learning are obtained through training without the involvement of electronic computers. Through the method, it is possible to learn multivariable linear functions without any limitation on the number of input terms. Due to the non-negative DNA concentration, the method uses a dual track model for negative data processing operations.


The linear machine learning method based on DNA hybridization reaction technology, includes: a training part, an algorithm part and a testing part.


(1) Reaction expressions of the training part are as follows:










X
1
+

+


W
1
+




k
1



I
+


+

X
1
+

+

W
1
+





(

1

a

)













X
1
-

+


W
1
-




k
1



I
+


+

X
1
-

+

W
1
-





(

1

b

)













X
1
+

+


W
1
-




k
1



I
-


+

X
1
+

+

W
1
-





(

1

c

)













X
1
-

+


W
1
+




k
1



I
-


+

X
1
-

+

W
1
+





(

1

d

)













X
2
+

+


W
2
+




k
1



I
+


+

X
2
+

+

W
2
+





(

2

a

)













X
2
-

+


W
2
-




k
1



I
+


+

X
2
-

+

W
2
-





(

2

b

)













X
2
+

+


W
2
-




k
1



I
-


+

X
2
+

+

W
2
-





(

2

c

)













X
2
-

+


W
2
+




k
1



I
-


+

X
2
-

+

W
2
+





(

2

d

)


















X
N
+

+


W
N
+




k
1



I
+


+

X
N
+

+

W
N
+






(

3

a

)














X
N
-

+


W
N
-




k
1



I
+


+

X
N
-

+

W
N
-





(

3

d

)













X
N
+

+


W
N
-




k
1



I
-


+

X
N
+

+

W
N
-





(

3

c

)













X
N
-

+


W
N
+




k
1



I
-


+

X
N
-

+

W
N
+





(

3

d

)













Y
+

+


H
+




k
1



I
-


+

H
+





(

4

a

)













Y
-

+


H
-




k
1



I
+


+

H
-





(

4

b

)













I
+

+


C
+




k
1



I
+


+

C
+

+

Y
+





(

4

c

)













I
-

+


C
+




k
1



I
-


+

C
-

+

Y
+





(

4

d

)













X
1
+

+


X
1
-




k
2


Φ





(

5

a

)













W
1
+

+


W
1
-





k
2


Φ





(

5

b

)













X
2
+

+


X
2
-




k
2


Φ





(

5

c

)













W
2
+

+


W
2
-




k
2


Φ





(

5

d

)


















X
i
+

+


X
i
-




k
2


Φ





(

6

a

)













W
i
+

+


W
i
-





k
2


Φ





(

6

b

)













Y
+

+


Y
-





k
2


Φ





(

6

c

)













C
+

+


C
-





k
2


Φ





(

6

d

)













H
+

+


H
-





k
2


Φ





(

6

e

)













I
+

+


I
-





k
2


Φ





(

6

f

)







The reactions (1a) to (3d) belong to a catalytic reaction module 1 (also referred to as first catalytic reaction module). The reaction (4a) and the reaction (4b) belong to a catalytic reaction module 2 (also referred to as second catalytic reaction module). The reaction (4c) and the reaction (4d) belong to the catalytic reaction module 1 and reactions (5a) to (6f) belong to an annihilation reaction.


Where [Xi]t=[Wi+]t−[Xi]t, i=1, 2 . . . , N, N represents the number of input items, [*]t represents a concentration of a substance * at a time t, and x, represents an i-th input value. [Wi]t=[Wi+]t−[Wi]t, Wi represents an i-th weight, and differential equations of concentrations varying over time of substances I+ and I are:









{






d
[

I
+

]

dt

=


k
1

(





[

W
1
+

]

t

[

X
1
+

]

t

+




[

W
2
+

]

t

[

X
2
+

]

t

+

+












[

W
N
+

]

t

[

X
N
+

]

t

+




[

W
1
-

]

t

[

X
1
-

]

t

+




[

W
2
-

]

t

[

X
2
-

]

t

+

+











[

W
N
-

]

t

[

X
N
-

]

t

+




[

Y
-

]

t

[

H
-

]

t


)








d
[

I
-

]

dt

=


k
1

(





[

W
1
-

]

t

[

X
1
-

]

t

+




[

W
2
+

]

t

[

X
2
-

]

t

+

+












[

W
N
+

]

t

[

X
N
-

]

t

+




[

W
1
-

]

t

[

X
1
+

]

t

+




[

W
2
-

]

t

[

X
2
+

]

t

+

+











[

W
N
-

]

t

[

X
N
+

]

t

+




[

Y
+

]

t

[

H
+

]

t


)








(
7
)







The differential equations (7) are simplified as follows:






{







d
[

I
+

]

t

dt

=


k
1

(





i
=
1

N


(





[

W
i
+

]

t

[

X
i
+

]

t

+




[

W
i
-

]

t

[

X
i
-

]

t


)


+




[

Y
-

]

t

[

H
-

]

t


)










d
[

I
-

]

t

dt

=


k
1

(





i
=
1

N


(





[

W
i
+

]

t

[

X
i
-

]

t

+




[

W
i
-

]

t

[

X
i
+

]

t


)


+




[

Y
+

]

t

[

H
+

]

t


)









And:












d
[
I
]

t


d

t


=





d
[

I
+

]

t


d

t


-



d
[

I
-

]

t


d

t









=






k
1



(




i
=
1

N


(





[

W
i
+

]

t

[

X
i
+

]

t

+




[

W
i
-

]

t

[

X
i
-

]

t

-




[

W
i
+

]

t

[

X
i
-

]

t

-
















[

W
i
-

]

t

[

X
i
+

]

t

+




[

Y
-

]

t

[

H
-

]

t

-




[

Y
+

]

t

[

H
+

]

t


)

)











Supposing [H+]t=[H]t=1 nanomoles per liter (nM), and









d


I

(
t
)



d

t


=


k
1

(





i
=
1

N




W
i

(
t
)




X
i

(
t
)



-

Y

(
t
)


)


,




when a DNA hybridization reaction network reaches dynamic equilibrium, i.e.,









d


I

(
t
)



d

t


=
0

,




then:






Y
=




i
=
1

N



W
i



X
i







where Y(t)=[Y+]t−[Y]t, Y represents an output value of a system.


(2) Reaction expressions of the algorithm part are as follows:











D
+

+

X
1
+





k
3




D
+

+

X
1
+

+

W
1
+






(

8

a

)














D
-

+

X
1
-





k
3




D
-

+

X
1
-

+

W
1
+






(

8

b

)














Y
+

+

X
1
-





k
3




Y
+

+

X
1
-

+

W
1
+






(

8

c

)














Y
-

+

X
1
+





k
3




D
-

+

X
1
+

+

W
1
+






(

8

d

)














D
+

+

X
1
-





k
3




D
+

+

X
1
-

+

W
1
-






(

8

e

)














D
-

+

X
1
+





k
3




D
-

+

X
1
+

+

W
1
-






(

8

f

)














Y
+

+

X
1
+





k
3




Y
+

+

X
1
+

+

W
1
-






(

8

g

)














Y
-

+

X
1
-





k
3



D
+

X
1
-

+

W
1
-






(

8

h

)














D
+

+

X
2
+





k
3




D
+

+

X
2
+

+

W
2
+






(

9

a

)














D
-

+

X
2
-





k
3




D
-

+

X
2
-

+

W
2
+






(

9

b

)














Y
+

+

X
2
-





k
3




Y
+

+

X
2
-

+

W
2
+






(

9

c

)














Y
-

+

X
2
+





k
3




D
-

+

X
2
+

+

W
2
+






(

9

d

)














D
+

+

X
2
-





k
3




D
+

+

X
2
-

+

W
2
-






(

9

e

)














D
-

+

X
2
+





k
3




D
-

+

X
2
+

+

W
2
-






(

9

f

)














Y
+

+

X
2
+





k
3




Y
+

+

X
2
+

+

W
2
-






(

9

g

)














Y
-

+

X
2
-





k
3




D
-

+

X
2
-

+

W
2
-






(

9

h

)



















D
+

+

X
N
+





k
3




D
+

+

X
N
+

+

W
N
+







(

10

a

)















D
-

+

X
N
-





k
3




D
-

+

X
N
-

+

W
N
+






(

10

b

)














Y
+

+

X
N
-





k
3




Y
+

+

X
N
-

+

W
N
+






(

10

c

)














Y
-

+

X
N
+





k
3




D
-

+

X
N
+

+

W
N
+






(

10

d

)














D
+

+

X
N
-





k
3




D
+

+

X
N
-

+

W
N
-






(

10

e

)














D
-

+

X
N
+





k
3




D
-

+

X
N
+

+

W
N
-






(

10

f

)














Y
+

+

X
N
+





k
3




Y
+

+

X
N
+

+

W
N
-






(

10

g

)














Y
-

+

X
N
+





k
3




D
-

+

X
N
-

+

W
N
-






(

10

h

)







Where reactions (8)-(10) belong to the catalytic reaction module 1.









{






d
[

W
1
+

]


d

t


=


k
3

(





[

D
+

]

t

[

X
1
+

]

t

+




[

D
-

]

t

[

X
1
-

]

t

+




[

Y
+

]

t

[

X
1
-

]

t

+




[

Y
-

]

t

[

X
1
+

]

t


)









d
[

W
1
-

]


d

t


=


k
3

(





[

D
+

]

t

[

X
1
-

]

t

+




[

D
-

]

t

[

X
1
+

]

t

+




[

Y
+

]

t

[

X
1
+

]

t

+




[

Y
-

]

t

[

X
1
-

]

t


)









(
11
)












{






d
[

W
2
+

]


d

t


=


k
3

(





[

D
+

]

t

[

X
2
+

]

t

+




[

D
-

]

t

[

X
2
-

]

t

+




[

Y
+

]

t

[

X
2
-

]

t

+




[

Y
-

]

t

[

X
2
+

]

t


)









d
[

W
2
-

]


d

t


=


k
3

(





[

D
+

]

t

[

X
2
-

]

t

+




[

D
-

]

t

[

X
2
+

]

t

+




[

Y
+

]

t

[

X
2
+

]

t

+




[

Y
-

]

t

[

X
2
-

]

t


)









(
12
)

















{






d
[

W
N
+

]


d

t


=


k
3

(





[

D
+

]

t

[

X
N
+

]

t

+




[

D
-

]

t

[

X
N
-

]

t

+




[

Y
+

]

t

[

X
N
-

]

t

+




[

Y
-

]

t

[

X
N
+

]

t


)









d
[

W
N
-

]


d

t


=


k
3

(





[

D
+

]

t

[

X
N
-

]

t

+




[

D
-

]

t

[

X
N
+

]

t

+




[

Y
+

]

t

[

X
N
+

]

t

+




[

Y
-

]

t

[

X
N
-

]

t


)










(
13
)








where D(t)=[D+]t−[D]t, D represents an expectation value.


Concluding from differential equations (11)-(13), what is obtained is:











d



W
1

(
t
)



d

t


=





d
[

W
1
+

]

t


d

t


-



d
[

W
1
-

]

t


d

t









=







k
3



(





[

D
+

]

t

[

X
1
+

]

t

+




[

D
-

]

t

[

X
1
-

]

t

-




[

D
+

]

t

[

X
1
-

]

t

-




[
D
]

t

[

X
1
+

]

t


)


+







k
2



(





[

Y
+

]

t

[

X
1
-

]

t

+




[

Y
-

]

t

[

X
1
+

]

t

-




[

Y
+

]

t

[

X
1
+

]

t

+




[

Y
-

]

t

[

X
1
-

]

t


)











=




k
3

[


(



[

D
+

]

t

-


[

D
-

]

t


)

-

(



[

Y
+

]

t

-


[

Y
-

]

t


)


]

×

(



[

X
1
+

]

t

-


[

X
1
-

]

t


)








=





k
3

(



[
D
]

t

-


[
Y
]

t


)

[

X
1

]

t















d



W
2

(
t
)



d

t


=





d
[

W
2
+

]

t


d

t


-



d
[

W
2
-

]

t


d

t









=







k
3



(





[

D
+

]

t

[

X
2
+

]

t

+




[

D
-

]

t

[

X
2
-

]

t

-




[

D
+

]

t

[

X
2
-

]

t

-




[

D
-

]

t

[

X
2
*

]

t


)


+







k
2

(





[

Y
+

]

t

[

X
2
-

]

t

+




[

Y
-

]

t

[

X
2
+

]

t

-




[

Y
+

]

t

[

X
2
+

]

t

+




[

Y
-

]

t

[

X
2
-

]

t












=




k
3

[


(



[

D
+

]

t

-


[

D
-

]

t


)

-

(



[

Y
+

]

t

-


[

Y
-

]

t


)


]

×

(



[

X
2
+

]

t

-


[

X
2
-

]

t


)








=





k
3

(



[
D
]

t

-


[
Y
]

t


)

[

X
2

]

t




















d



W
N

(
t
)



d

t


=





d
[

W
N
+

]

t


d

t


-



d
[

W
N
-

]

t


d

t









=






k
3



(





[

D
+

]

t

[

X
N
+

]

t

+




[

D
-

]

t

[

X
N
-

]

t

-




[

D
+

]

t

[

X
N
-

]

t

-













[

D
-

]

t

[

X
N
*

]

t

)

+


k
2

(





[

Y
+

]

t

[

X
N
-

]

t

+




[

Y
-

]

t

[

X
N
*

]

t













=




k
3

[


(



[

D
+

]

t

-


[

D
-

]

t


)

-

(



[

Y
+

]

t

-


[

Y
-

]

t


)


]

×

(



[

X
N
+

]

t

-


[

X
N
-

]

t


)








=






k
3

(



[
D
]

t

-


[
Y
]

t


)

[

X
N

]

t

.








(3) Reaction expressions of the testing part are as follows:











X
1
+

+


W
ˆ

1
+





k
1





I
^

+

+

X
1
+

+


W
ˆ

1
+






(

14

a

)














X
1
-

+


W
ˆ

1
-





k
1





I
^

+

+

X
1
-

+


W
ˆ

1
+






(

14

b

)














X
1
+

+


W
ˆ

1
-





k
1





I
^

-

+

X
1
+

+


W
ˆ

1
-






(

14

c

)














X
1
-

+


W
ˆ

1
+





k
1





I
^

-

+

X
1
-

+


W
ˆ

1
+






(

14

d

)














X
2
+

+


W
ˆ

2
+





k
1





I
^

+

+

X
2
+

+


W
ˆ

2
+






(

15

a

)














X
2
+

+


W
ˆ

2
+





k
1





I
^

+

+

X
2
+

+


W
ˆ

2
+






(

15

b

)














X
2
+

+


W
ˆ

2
-





k
1





I
^

-

+

X
2
+

+


W
ˆ

2
-






(

15

c

)














X
2
-

+


W
ˆ

2
+





k
1





I
^

-

+

X
2
-

+


W
ˆ

2
+






(

15

d

)



















X
N
+

+


W
ˆ

N
+





k
1





I
^

+

+

X
N
+

+


W
ˆ

N
+







(

16

a

)















X
N
-

+


W
ˆ

N
-





k
1





I
^

+

+

X
N
-

+


W
ˆ

N
-






(

16

b

)














X
N
+

+


W
ˆ

N
-





k
1





I
^

-

+

X
N
+

+


W
ˆ

N






(

16

c

)














X
N
-

+


W
ˆ

N
+





k
1





I
^

-

+

X
N
-

+


W
ˆ

N
+






(

16

d

)















Y
ˆ

+

+


H
ˆ

+





k
1





I
^

-

+


H
ˆ

+






(

17

a

)















Y
ˆ

-

+


H
ˆ

-





k
1





I
^

+

+


H
^

-






(

17

b

)















I
^

+

+


C
^

+





k
1





I
^

+

+


C
^

+

+


Y
^

+






(

17

c

)















I
^

-

+


C
^

-





k
1





I
^

-

+


C
^

-

+


Y
^

+






(

17

d

)














X
i
+

+

X
i
-





k
2


Φ




(

18

a

)















W
ˆ

i
+

+


W
i
-

ˆ






k
2


Φ




(

18

b

)















Y
ˆ

+

+


Y
-

ˆ





k
2



Φ




(

18

c

)















C
ˆ

+

+


C
-

ˆ





k
2



Φ




(

18

d

)















H
ˆ

+

+


H
-

ˆ





k
2



Φ




(

18

e

)















I
ˆ

+

+


I
-

ˆ





k
2



Φ




(

18

f

)







where reactions (14d) to (16d) belong to the catalytic reaction module 1. Reactions (17a) and (17b) belong to the catalytic reaction module 2; reactions (17c) and (17d) belong to the first catalytic module and reactions (18a)-(18f) belong to the annihilation reaction module.


The beneficial effects of above technical solutions are as follows.


(1) The machine learning system based on DNA molecular circuits provided by the present disclosure has the ability to autonomously learn and can complete the training and testing process without the assistance of electronic computers (the weights of the machine learning are obtained through training), achieving complete biological machine learning without the involvement of electronic computers.


(2) The multivariate linear function relationship is f(x1, x2, . . . , xN)=w1x1+w2x2+ . . . +wNxN. The number of independent variables is N, and artificial neural networks of existing research results based on DNA molecule computing can only handle the linear function relationship of two independent variables. Through the method, multivariate linear functions can be learned without any limit on the number of input items, which can be any positive integer.


(3) Due to the fact that the input and output values of DNA molecular circuits are represented by the concentration of DNA molecules, which is non-negative, data with negative numbers cannot be processed. The present disclosure adopts a bimolecular concentration difference to represent the input and output values, as the concentration difference can be positive or negative and can represent all real numbers.


(4) The present disclosure can be used for fitting and predicting the relationship between total current and voltage in parallel circuits, and can also predict the value of partial resistance.





BRIEF DESCRIPTION OF DRAWINGS


FIG. 1 illustrates a flowchart of the present disclosure.



FIG. 2 illustrates a schematic diagram of main DNA strands displacement reaction of a first submodule of a catalytic reaction module 1 of the present disclosure.



FIG. 3 illustrates a schematic diagram of main DNA strands displacement reaction of a second submodule of the catalytic reaction module 1 of the present disclosure.



FIG. 4 illustrates a schematic diagram of main DNA strands displacement reaction of a third submodule of the catalytic reaction module 1 of the present disclosure.



FIG. 5 illustrates a schematic diagram of main DNA strands displacement reaction of a fourth submodule of the catalytic reaction module 1 of the present disclosure.



FIG. 6 illustrates a schematic diagram of main DNA strands displacement reaction of a catalytic reaction module 2 of the present disclosure.



FIG. 7 illustrates a schematic diagram of main DNA strands displacement reaction of an annihilation reaction module of the present disclosure.



FIG. 8 illustrated a circuit diagram of a parallel circuit of the present disclosure.



FIG. 9 illustrated a schematic diagram of weight update trajectory of the present disclosure.



FIG. 10 illustrated a schematic diagram of evolution of average relative error with training times of the present disclosure.



FIG. 11 illustrated a schematic diagram of variation of training times with the number of training rounds of the present disclosure.



FIG. 12 illustrated a schematic diagram of evolution of relative error with test data of the present disclosure.





DETAILED DESCRIPTION OF EMBODIMENTS

In order to clarify the purpose, technical solutions, and advantages of the embodiments of the present disclosure, a clear and complete description of the technical solutions in the embodiments of the present disclosure will be provided below. Apparently, the described embodiments are a part of the embodiments of the present disclosure, not all embodiments. The following is a detailed description of the present disclosure in conjunction with the accompanying drawings. However, it should be understood that the provision of the drawings is only for a better understanding of the present disclosure, and they should not be understood as limitations to the present disclosure.


Specific steps are as follows.


1. Design of Linear Machine Learning Based on Idealized Reaction.

(1) The training part of machine learning:











X
1
+

+

W
1
+





k
1




I
+

+

X
1
+

+

W
1
+






(

1

a

)














X
1
-

+

W
1
-





k
1




I
+

+

X
1
-

+

W
1
-






(

1

b

)














X
1
+

+

W
1
-





k
1



I
+

X
1
+

+

W
1
-






(

1

c

)














X
1
-

+

W
1
+





k
1



I
+

X
1
-

+

W
1
+






(

1

d

)














X
2
+

+

W
2
+





k
1




I
+

+

X
2
+

+

W
2
+






(

2

a

)














X
2
-

+

W
2
-





k
1




I
+

+

X
2
-

+

W
2
-






(

2

b

)














X
2
+

+

W
2
-





k
1




I
-

+

X
2
+

+

W
2
-






(

2

c

)














X
2
-

+

W
2
+





k
1




I
-

+

X
2
-

+

W
2
+






(

2

d

)



















X
N
+

+

W
N
+





k
1




I
+

+

X
N
+

+

W
N
+






(

3

a

)














X
N
-

+

W
N
-





k
1




I
+

+

X
N
-

+

W
N
-






(

3

b

)














X
N
+

+

W
N
-





k
1




I
-

+

X
N
+

+

W
N
-






(

3

c

)














X
N
-

+

W
N
+





k
1




I
-

+

X
N
-

+

W
N
+






(

3

d

)














Y
+

+

H
+





k
1




I
-

+

H
+






(

4

a

)














Y
-

+

H
-





k
1




I
+

+

H
-






(

4

b

)














I
+

+

C
+





k
1




I
+

+

C
+

+

Y
+






(

4

c

)














I
-

+

C
-





k
1




I
-

+

C
-

+

Y
+






(

4

d

)














X
1
+

+

X
1
-





k
2


Φ




(

5

a

)














W
1
+

+

W
1
-





k
2


Φ




(

5

b

)














X
2
+

+

X
2
-





k
2


Φ




(

5

c

)














W
2
+

+

W
2
-





k
2


Φ




(

5

d

)



















X
i
+

+

X
i
-





k
2


Φ




(

6

a

)














W
i
+

+

W
i
-





k
2


Φ




(

6

b

)














Y
+

+

Y
-





k
2


Φ




(

6

c

)














C
+

+

C
-





k
2


Φ




(

6

d

)














H
+

+

H
-





k
2


Φ




(

6

e

)














I
+

+

I
-





k
2


Φ




(

6

f

)







Where Xi(t)=[Xi+]t−[Xi]t, i=1, 2, . . . , N, N represents the number of input items, [*]t represents a concentration of a substance * at a time t, and Xi represents an i-th input item, that is, X1, X2, . . . , XN respectively represent the first input item, the second input item, . . . , the N-th input item of the machine learning (i.e., learning machine). Wi(t)=[Wi+]t−[Wi]t, Wi represents an i-th weight, that is, W1, W2, . . . , WN respectively represent the weight corresponding to the first input item, the weight corresponding to the second input item, . . . , the weight corresponding to the N-th input item. Xi+ and Xi respectively represent a positive part and a negative part of Xi, specifically, X1+, X2+, . . . , XN+ respectively represent positive parts of X1, X2, . . . , XN; and X1, X2, . . . , XN respectively represent negative parts of X1, X2, . . . , XN. Apparently, [Xi+]t and [Xi]t respectively represent concentrations of the positive part and the negative part of Xi at the time t. Wi+ and Wi respectively represent a positive part and a negative part of Wi, specifically, W1+, W2+, . . . , WN+ respectively represent positive parts of W1, W2, . . . , WN; and W1, W2, . . . , WN respectively represent negative parts of W1, W2, . . . , WN. Apparently, [Wi+]t and [Wi]t respectively represent concentrations of the positive part and the negative part of Wi at the time t. Y+ and Y respectively represent the positive part and the negative part of Y. H+ and H respectively represent the positive part and the negative part of H. C+ and C respectively represent the positive part and the negative part of C. I+ and I respectively represent the positive part and the negative part of I. k1 and k2 respectively represent different reaction rates. Φ represents a waste, which is meaningless to the machine learning. Supposing [H+]t=[H]t≡1 nM, differential equations of concentrations varying over time of substances I+ and I are as follows:









{







d
[

I
+

]

t


d

t


=





k
1

(





[

W
1
+

]

t

[

X
1
+

]

t

+





[

W
2
+

]



t

[

X
2
+

]

t


+

+




[

W
N
+

]

t

[

X
N
+

]

t

+














[

W
1
-

]

t

[

X
1
-

]

t

+




[

W
2
-

]

t

[

X
2
-

]

t

+

+




[

W
N
-

]

t

[

X
N
-

]

t

+













[

Y
-

]

t

[

H
-

]

t

)









d
[

I
-

]

t


d

t


=





k
1

(





[

W
1
+

]

t

[

X
1
-

]

t

+




[

W
2
+

]

t

[

X
2
-

]

t

+

+




[

W
N
+

]

t

[

X
N
-

]

t

+














[

W
1
-

]

t

[

X
1
+

]

t


+




[

W
2
-

]

t

[

X
2
+

]

t

+

+




[

W
N
-

]

t

[

X
N
+

]

t

+













[

Y
+

]

t

[

H
+

]

t

)








(
7
)







Where [Y+] and [Y] respectively represent concentrations of the positive part and the negative part of Y at the time t; [I+]t and [I]t respectively represent concentrations of the positive part and the negative part of I at the time t; [H+]t and [H]t respectively represent concentrations of the positive part and the negative part of H at the time t.


The differential equations (7) are simplified as follows:






{







d
[

I
+

]

t


d

t


=


k
1

(





i
=
1

N


(





[

W
i
+

]

t

[

X
i
+

]

t

+




[

W
i
-

]

t

[

X
i
-

]

t


)


+




[

Y
-

]

t

[

H
-

]

t


)










d
[

I
-

]

t


d

t


=


k
1

(





i
=
1

N


(





[

W
i
+

]

t

[

X
i
-

]

t

+




[

W
i
-

]

t

[

X
i
+

]

t


)


+




[

Y
+

]

t

[

H
+

]

t


)










and








d
[
I
]

t


d

t


=





d
[

I
+

]

t


d

t


-



d
[

I
-

]

t


d

t



=


k
1

(




i
=
1

N


(





[

W
i
+

]

t

[

X
i
+

]

t

+




[

W
i
-

]

t

[

X
i
-

]

t

-





[

W
i
+

]

t

[

X
i
-

]

t

-




[

W
i
-

]

t

[

X
i
+

]

t

+




[

Y
-

]

t

[

H
-

]

t

-




[

Y
+

]

t

[

H
+

]

t


)


)






Due to supposing [H+]t=[H]t≡1 (nM) and








d


I

(
t
)



d

t


=



k
1

(





i
=
1

N




W
i

(
t
)




X
i

(
t
)



-

Y

(
t
)


)

.





When DNA hybridization reaction network reach dynamic equilibrium, and it can be concluded that:







Y

(
t
)

=




i
=
1

N




W
i

(
t
)





X
i

(
t
)

.







Apparently. Y(t)=[Y+]t−[Y]t, representing an output value of the system.


Reaction expressions of the algorithm part are as follows:











D
+

+

X
1
+





k
3




D
+

+

X
1
+

+

W
1
+






(

8

a

)














D
-

+

X
1
-





k
3




D
-

+

X
1
-

+

W
1
+






(

8

b

)














Y
+

+

X
1
-





k
3




Y
+

+

X
1
-

+

W
1
+






(

8

c

)














Y
-

+

X
1
+





k
3




D
-

+

X
1
+

+

W
1
+






(

8

d

)














D
+

+

X
1
-





k
3




D
+

+

X
1
-

+

W
1
-






(

8

e

)














D
-

+

X
1
+





k
3




D
-

+

X
1
+

+

W
1
-






(

8

f

)














Y
+

+

X
1
+





k
3




Y
+

+

X
1
+

+

W
1
-






(

8

g

)














Y
-

+

X
1
-





k
3




D
-

+

X
1
-

+

W
1
-






(

8

h

)














D
+

+

X
2
+





k
3




D
+

+

X
2
+

+

W
2
+






(

9

a

)














D
-

+

X
2
-





k
3




D
-

+

X
2
-

+

W
2
+






(

9

b

)














Y
+

+

X
2
-





k
3




Y
+

+

X
2
-

+

W
2
+






(

9

c

)














Y
-

+

X
2
+





k
3




D
-

+

X
2
+

+

W
2
+






(

9

d

)














D
+

+

X
2
-





k
3




D
+

+

X
2
-

+

W
2
-






(

9

e

)














D
-

+

X
2
+





k
3




D
-

+

X
2
+

+

W
2
-






(

9

f

)














Y
+

+

X
2
+





k
3




Y
+

+

X
2
+

+

W
2
-






(

9

g

)














Y
-

+

X
2
-





k
3




D
-

+

X
2
-

+

W
2
-






(

9

h

)



















D
+

+

X
N
+





k
3




D
+

+

X
N
+

+

W
N
+







(

10

a

)















D
-

+

X
N
-





k
3




D
-

+

X
N
-

+

W
N
+






(

10

b

)














Y
+

+

X
N
-





k
3




Y
+

+

X
N
-

+

W
N
+






(

10

c

)














Y
-

+

X
N
+





k
3




D
-

+

X
N
+

+

W
N
+






(

10

d

)














D
+

+

X
N
-





k
3




D
+

+

X
N
-

+

W
N
-






(

10

e

)














D
-

+

X
N
+





k
3




D
-

+

X
N
+

+

W
N
-






(

10

f

)














Y
+

+

X
N
+





k
3




Y
+

+

X
N
+

+

W
N
-






(

10

g

)














Y
-

+

X
N
+





k
3




D
-

+

X
N
-

+

W
N
-






(

10

h

)







Where D+ and D respectively represent a positive part and a negative part of D; and k3 represent a reaction rate which is different from k1 and k2.


Differential equations of concentrations varying over time of substances Wi+ and Wi are:









{






d
[

W
1
+

]


d

t


=


k
3

(





[

D
+

]

t

[

X
1
+

]

t

+




[

D
-

]

t

[

X
1
-

]

t

+




[

Y
+

]

t

[

X
1
-

]

t

+




[

Y
-

]

t

[

X
1
+

]

t



)









d
[

W
1
-

]


d

t


=


k
3

(





[

D
+

]

t

[

X
1
-

]

t

+




[

D
-

]

t

[

X
1
+

]

t

+




[

Y
+

]

t

[

X
1
+

]

t

+




[

Y
-

]

t

[

X
1
-

]

t



)









(
11
)












{






d
[

W
2
+

]


d

t


=


k
3

(





[

D
+

]

t

[

X
2
+

]

t

+




[

D
-

]

t

[

X
2
-

]

t

+




[

Y
+

]

t

[

X
2
-

]

t

+




[

Y
-

]

t

[

X
2
+

]

t



)









d
[

W
2
-

]


d

t


=


k
3

(





[

D
+

]

t

[

X
2
-

]

t

+




[

D
-

]

t

[

X
2
+

]

t

+




[

Y
+

]

t

[

X
2
+

]

t

+




[

Y
-

]

t

[

X
2
-

]

t



)









(
12
)

















{






d
[

W
N
+

]


d

t


=


k
3

(





[

D
+

]

t

[

X
N
+

]

t

+




[

D
-

]

t

[

X
N
-

]

t

+




[

Y
+

]

t

[

X
N
-

]

t

+




[

Y
-

]

t

[

X
N
+

]

t



)









d
[

W
N
-

]


d

t


=


k
3

(





[

D
+

]

t

[

X
N
-

]

t

+




[

D
-

]

t

[

X
N
+

]

t

+




[

Y
+

]

t

[

X
N
+

]

t

+




[

Y
-

]

t

[

X
N
-

]

t



)










(
13
)








Where D(t)=[D+]t−[D]t represents an expectation value; and [D+]t and [D]t respectively represent concentrations of the positive part and the negative part of D at the time t.


Concluding from the differential equations (11) to (13), what is obtained is:











d



W
1

(
t
)


dt

=





d
[

W
1
+

]

t

dt

-



d
[

W
1
-

]

t

dt


=



k
3

(





[

D
+

]

t

[

X
1
+

]

t

+




[

D
-

]

t

[

X
1
-

]

t

-
















[

D
+

]

t

[

X
1
-

]

t

-




[

D
-

]

t

[

X
1
+

]

t


)

+









k
2

(





[

Y
+

]

t

[

X
1
-

]

t

+




[

Y
-

]

t

[

X
1
+

]

t

-














[

Y
+

]

t

[

X
1
+

]

t

+




[

Y
-

]

t

[

X
1
-

]

t


)






=




k
3

[


(



[

D
+

]

t

-


[

D
-

]

t


)

-

(



[

Y
+

]

t

-


[

Y
-

]

t


)


]

×









(



[

X
1
+

]

t

-


[

X
1
-

]

t


)







=





k
3

(



[
D
]

t

-


[
Y
]

t


)

[

X
1

]

t















d



W
2

(
t
)


dt

=





d
[

W
2
+

]

t

dt

-



d
[

W
2
-

]

t

dt


=



k
3

(





[

D
+

]

t

[

X
2
+

]

t

+




[

D
-

]

t

[

X
2
-

]

t

-
















[

D
+

]

t

[

X
2
-

]

t

-




[

D
-

]

t

[

X
2
*

]

t


)

+









k
2

(





[

Y
+

]

t

[

X
2
-

]

t

+




[

Y
-

]

t

[

X
2
+

]

t

-














[

Y
+

]

t

[

X
2
+

]

t

+




[

Y
-

]

t

[

X
2
-

]

t


)






=




k
3

[


(



[

D
+

]

t

-


[

D
-

]

t


)

-

(



[

Y
+

]

t

-


[

Y
-

]

t


)


]

×









(



[

X
2
+

]

t

-


[

X
2
-

]

t


)







=





k
3

(



[
D
]

t

-


[
Y
]

t


)

[

X
2

]

t






















dW
N

(
t
)

dt

=





d
[

W
N
+

]

t

dt

-



d
[

W
N
-

]

t

dt


=



k
3

(





[

D
+

]

t

[

X
N
+

]

t

+




[

D
-

]

t

[

X
N
-

]

t

-
















[

D
+

]

t

[

X
N
-

]

t

-




[

D
-

]

t

[

X
N
*

]

t


)

+









k
2

(





[

Y
+

]

t

[

X
N
-

]

t

+




[

Y
-

]

t

[

X
N
*

]

t









=




k
3

[


(



[

D
+

]

t

-


[

D
-

]

t


)

-

(



[

Y
+

]

t

-


[

Y
-

]

t


)


]

×









(



[

X
N
+

]

t

-


[

X
N
-

]

t


)







=





k
3

(



[
D
]

t

-


[
Y
]

t


)

[

X
N

]

t





.




Reaction expressions of the testing part of machine learning are as follows:










X
1
+

+



W
^

1
+




k
1




I
^

+


+

X
1
+

+


W
^

1
+





(

14

a

)













X
1
-

+



W
^

1
-




k
1




I
^

+


+

X
1
-

+


W
^

1
+





(

14

b

)













X
1
+

+



W
^

1
-




k
1




I
^

-


+

X
1
+

+


W
^

1
-





(

14

c

)













X
1
-

+



W
^

1
+




k
1




I
^

-


+

X
1
-

+


W
^

1
+





(

14

d

)













X
2
+

+



W
^

2
+




k
1




I
^

+


+

X
2
+

+


W
^

2
+





(

15

a

)













X
2
+

+



W
^

2
+




k
1




I
^

+


+

X
2
+

+


W
^

2
+





(

15

b

)













X
2
+

+



W
^

2
-




k
1




I
^

-


+

X
2
+

+


W
^

2
-





(

15

c

)













X
2
-

+



W
^

2
+




k
1




I
^

-


+

X
2
-

+


W
^

2
+





(

15

d

)


















X
N
+

+



W
^

N
+




k
1




I
^

+


+

X
N
+

+


W
^

N
+





(

16

a

)













X
N
-

+



W
^

N
-




k
1




I
^

+


+

X
N
-

+


W
^

N
-





(

16

b

)













X
N
+

+



W
^

N
-




k
1




I
^

-


+

X
N
+

+


W
^

N
-





(

16

c

)













X
N
-

+



W
^

N
+




k
1




I
^

-


+

X
N
-

+


W
^

N
+





(

16

d

)














Y
^

+

+



H
^

+




k
1




I
^

-


+


H
^

+





(

17

a

)














Y
^

-

+



H
^

-




k
1




I
^

+


+


H
^

-





(

17

b

)














I
^

+

+



C
^

+




k
1




I
^

+


+


C
^

+

+


Y
^

+





(

17

c

)














I
^

-

+



C
^

-




k
1




I
^

-


+


C
^

-

+


Y
^

+





(

17

d

)













X
i
+

+


X
i
-




k
2


Φ





(

18

a

)














W
^

i
+

+



W
^

i
-




k
2


Φ





(

18

b

)














Y
^

+

+



Y
^

-




k
2


Φ





(

18

c

)














C
^

+

+



C
^

-




k
2


Φ





(

18

d

)














H
^

+

+



H
^

-




k
2


Φ





(

18

e

)














I
^

+

+



I
^

-




k
2


Φ





(

18

f

)







Where Ŵi(t)=[Ŵi+]t−[Ŵi]t Ŵi represents a weight obtained after the rounds of training, that is, Ŵ1, Ŵ2, . . . , ŴN represent weights obtained after the rounds of learning, which are also weights used in the testing part. Ŵi+ and Ŵi respectively represent the positive part and the negative part of Ŵi. Ŷ represents an output of the testing part; Ŷ+ and Ŷ respectively represent the positive part and the negative part of Ŷ; Î, Ĥ and Ĉ represent substances involved in chemical reaction in the testing part; Ĥ+ and Ĥ respectively represent the positive part and the negative part of Ĥ; Ĉ+ and Ĉ respectively represent the positive part and the negative part of Ĉ; Î+ and Î respectively represent the positive part and the negative part of Î.


Concluded from the reaction equations (14a) to (18f), differential equations of concentrations varying over time of substances Î+ and Î are:






{







d
[


I
^

+

]

t

dt

=


k
1

(





i
=
1

N


(





[


W
^

i
+

]

t

[

X
i
+

]

t

+




[


W
^

i
-

]

t

[

X
i
-

]

t


)


+




[


Y
^

-

]

t

[


H
ˆ

-

]

t


)










d
[


I
^

-

]

t

dt

=


k
1

(





i
=
1

N


(





[


W
ˆ

i
+

]

t

[

X
i
-

]

t

+




[


W
^

i
-

]

t

[

X
i
+

]

t


)


+




[


Y
^

+

]

t

[


H
ˆ

+

]

t


)









What can be concluded from above differential equations is:








d


I
^


dt

=





d
[


I
^

+

]

t

dt

-



d
[


I
^

-

]

t

dt


=


k
1

(





i
=
1

N


(





[


W
ˆ

i
+

]

t

[

X
i
+

]

t

+




[


W
ˆ

i
-

]

t

[

X
i
-

]

t

-




[


W
ˆ

i
+

]

t

[

X
i
-

]

t

+




[


W
ˆ

i
-

]

t

[

X
i
+

]

t


)


-

(





[


γ
^

+

]

t

[


H
ˆ

+

]

t

-




[

γ
^

]

t

[


H
ˆ

-

]

t


)


)






When [Ĥ+]t=[Ĥ]t≡1 nM, what can be obtained is:








d


I
^


dt

=



k
1

(





i
=
1

N




W
ˆ

i



X
i



-

Y
ˆ


)

.





When Î+ and Î reach dynamic equilibrium,








df
dt

=
0

,




then:







Y
ˆ

=




i
=
1

N




W
~

i




X
i

.







Where Ŷ(t)=[Ŷ+]t−[Ŷ]t, Ŷ represents the output of machine learning.


It should be noted that the catalytic reaction module 1, the catalytic reaction module 2, and the annihilation reaction module in the training part, the algorithm part, and the testing part belong to a same type of reaction module, but do not belong to a same reaction module. For example, from differential equations of concentrations varying over time of î+ and Î, it can be seen that the symbol has a subscript i below it. When i changes, the DNA molecules undergo changes accordingly, which means that all the reaction modules provided in the present disclosure do not refer to any particular one reaction, but describe one type of reaction.


The catalytic reaction module 1 (catalytic reaction module 2) in the training and testing parts belong to the same type of the catalytic reaction module 1 (catalytic reaction module 2). The identification symbols of the catalytic reaction module 1 (catalytic reaction module 2) in the training and testing parts are different, suggesting different DNA molecules are used, and the identification symbols in the testing part have pointed caps (i.e., {circumflex over ( )}), while those in the training part does not.


2. Implementing Linear Machine Learning by Using DNA Molecular Circuits.

Due to the fact that reactants and products in idealized reactions are abstract substances rather than specific biochemical substances, while DNA hybridization reactions can achieve any type of idealized reaction. The DNA hybridization reactions are utilized to implement the linear learning machine.


The reaction equations (1) to (18) can be classified into several types of reactions: as shown in FIG. 2, the reaction equations (1a) to (3d) and (13a) to (15d) belong to the submodule 1 of the catalytic reaction module 1. As illustrated in FIG. 3, the reaction equations (4c), (4d), (17c), and (17d) belong to a first type of catalytic reaction, which can be achieved by the submodule 2 of the catalytic reaction module 1. As illustrated in FIG. 4, the reaction equations (8c), (8d), (8g), (8h), (9c), (9d), (9g), (9h), (10c), (10d), (10g), and (10h) belong to the first type of catalytic reaction, which can be achieved by the submodule 3 of the catalytic reaction module 1. As illustrated in FIG. 5, the reaction equations (8a), (8b), (8e), (8f), (9a), (9b), (9e), (9f), (10a), (10b), (10e), and (10f) belong to the first type of catalytic reaction, which can be achieved by the submodule 4 of the catalytic reaction module 1. As illustrated in FIG. 6, the reaction equations (4a), (4b), (17a) and (17b) belong to a second type of catalytic reaction. The reaction equations (5a) to (6f) and (18a) to (18f) belong to the annihilation reaction type and can be achieved through the annihilation reaction module. Due to the homogeneity and cascading nature of these reaction modules, they can be cascaded into DNA molecular circuits to achieve the machine learning system. The three types of DNA reaction modules are described as follows:


(1) The Catalytic Reaction Module 1

The idealized reaction equation for the catalytic reaction module 1 is








X
i
+

+


W
i
+




k
i



I
+


+

X
i
+

+

W
i
+


,




which is obtained by following DNA strand displacement reactions:









{






W
i
+

+

Ap
i
+





q
m


q
i




Ad
i
+

+

Aq
i
+









Ad
i
+

+


X
i
+




q
m



Ac
i
+


+

waste







Ac
i
+

+


Ab
i
+




q
m



I
+


+

X
i
+

+

W
i
+

+

waste








(
19
)







Where I+ is catalyzed, Xi represents an input signal DNA molecule, and Wi+ represents a weight reporting strand. Api+, Aqi+ and Abi+ are auxiliary DNA strands and initial concentrations of the auxiliary DNA strands are Cm, meeting Cm≥[Xi30]0, [Wi+]0, [I+]0. A reaction rate qi and ki meet qi≤qm, ki=qi, and qm represents a maximum reaction rate. The DNA implementation of the submodule 1 of the catalytic reaction module 1 is illustrated in FIG. 2. As illustrated in FIGS. 3, 4, 5, the DNA implementation process of the submodules 2-4 of the catalytic reaction module 1 is similar to the submodule 1.


(2) The Catalytic Reaction Module 2

The idealized reaction equation of the catalytic reaction module 2 is:








Y
+

+


H
+




k
i



I
-


+

H
+


,




which is obtained by following DNA strand displacement reactions:









{






Y
+

+

Am
+





q
m


q
i




An
+

+

Ae
+









Ae
+

+


H
+




q
m



As
+


+

waste







As
+

+


Ah
+




q
m



H
+


+

I
-

+

waste








(
20
)







Where I is catalyzed, Am+, An+, Ah+ and H+ are auxiliary DNA strands and initial concentrations of the auxiliary DNA strands Am+, An+ and Ah+ are respectively set as [Am+]0=[An+]0=[Ah+]0=Cm, meeting Cm□[Y+]0,[I]0. An initial concentration of H+ is 1 nM. The reaction rate qi meets qi≤qm, ki=qi. The DNA implementation of the catalytic reaction module 2 is illustrated in FIG. 6.


(3) The Annihilation Reaction Module

The idealized reaction equation of the annihilation reaction module is:








W
i
+

+


W
i
-




k
i


Φ


,




which is obtained by following DNA strand displacement reactions:









{






W
i
+

+

Wa
i
+





q
m


q
i




Wb
i
+

+

Wt
i
+









Wt
i
+

+


W
i
-




q
m


Φ









(
21
)







Where Wi+ and Wi are annihilated, Wai+ and Wbi+ are auxiliary DNA strands and initial concentrations of the auxiliary DNA strands are Cm, meeting Cm□[Wi+]0, [Wi]0. The reaction rate qi meets qi≤qm, ki=qi. The DNA implementation of the annihilation reaction module is illustrated in FIG. 7.


3. Training of linear machine learning.


This section belongs to the application of the machine learning system. The molecular learning machine has the ability to predict the relationship between the total current and voltage of parallel circuits, which is obtained through data training. Therefore, a third part is the training of the machine learning system. In order to test the learning ability of the molecular learning machine, it is necessary to test the molecular learning machine.


As illustrated in FIG. 8, the resistance value of the sliding rheostat is adjusted to measure the voltages U1, U2 and the total current I. The two voltage values and one total current are used as a set of training data. By adjusting the sliding rheostat, another set of training data can be obtained. These obtained training data are inputted into the DNA molecular machine learning system. Through the processing of the DNA molecular learning machine algorithm, weights are updated and the relative error between the output of the DNA molecular learning machine and the expected value is calculated. When the relative error reaches or is lower than the set value of 0.2, the training goal is achieved and the training is stopped. The weights obtained through training are the value of w1, w2, . . . wN of the linear function relationship I(U1, U2, . . . . UN)=w1U1+w2U2+ . . . +wNUN. The functional relationship between them can be fitted by using the voltage and current values when the value of w1, w2, . . . wN are obtained. These parameter values correspond to the reciprocals of the fixed resistance (the relationship between partial voltage and total current in a parallel circuit is






I
=



U
1


R
1


+


U
2


R
2


+

+



U
3


R
3


.






Therefore, the DNA molecule learning machine can predict the partial resistance value R1, R2, . . . RN.


The present disclosure utilizes a DNA linear learning machine to learn the relationship between the total current and voltages of a parallel circuit I(U1, U2, . . . UN)=w1U1+w2U2+ . . . +wNUN. The weights wi and inputs Ui(i=1, 2, . . . N) are both real numbers. As the weights and inputs are represented by the concentration of DNA strands, wi, Ui, I≥0.


(1) Training of Linear Machine Learning.

The training of machine learning includes multiple rounds of training. One round of training includes M groups of training data, each group of which is composed of N data, that is, the machine learning has N inputs and i=1, 2, . . . , N. The training data set is shuffled to obtain another round of training data. In the first round of training data, Ψi=[αi(1,1), αi(2,1), . . . , αi(M, 1)] and Γ={{tilde over (d)}1(1), {tilde over (d)}1(2), . . . , {tilde over (d)}1 (M)} respectively represent partial voltage and total current. The data can be normalized as follows:









{







x
ˆ

i

(

k
,
l

)

=

ρ



α
i

(

k
,
l

)

/
η










d
ˆ

l

(
k
)

=

ρ




d
˜

l

(
k
)

/
η









(
22
)








where





{






η
i

=


max

(

Ψ
i

)

-

min

(

Ψ
i

)








η
=

max

(

[


η
1

,

η
2

,





η
N



]

)










d
˜

l

(
k
)

=







i
=
1

N



w
i




x
i

(

l
,
k

)






.





αi(k,l) represents the i-th data of the k-th group of data in the l-th round of training, k=1, 2, . . . , M, l=1, 2, . . . , Λ and ρ are positive adjustment parameters. Ψi=[xi(1), xi(2), . . . , xi(K)]. {circumflex over (x)}i(k,l) and {circumflex over (d)}l (k) are the initial concentration setting of the input signal DNA strand for machine learning. The first round of training data meets {circumflex over (x)}1 (k,l)=[Xik+]0−[Xik−]0 and d1(k)=[Dk+]0−[Dk]0, shuffling the order can obtain training data for other rounds.


(2) Assessment of Training of Linear Machine Learning.

In the l-th training, a relative error el(k) is defined as follows.











e
l

(
k
)

=




D
l

(
k
)

-



d
ˆ

l

(
k
)





d
ˆ

l




(
k
)







(
23
)








where





{






D
l

(
k
)

=

ψ

(


Y
1

(

k
,
l

)

)









Y
1

(

k
,
l

)

=






n
=
1


L





V
ˆ


n

1


(
l
)



y
n




(

k
,
l

)











y
n

(

k
,
l

)

=

ψ

(


S
n




(

k
,
l

)


)









S
n

(

k
,
l

)

=





i
=
1


N





W
ˆ



in


(
l
)



x
i




(

k
,
l

)











Ŵin (l) and {circumflex over (V)}n1 (l) represent the weights of the input layer and hidden layer obtained after the l-th training.


To assess the training results, the average relative error is defined as follows:











e
¯

l

=


1
K







k
=
1


K




"\[LeftBracketingBar]"



e
l

(
k
)



"\[RightBracketingBar]"








(
24
)







After multiple trainings, when the average relative error reaches the target value, the training will be stopped.


As illustrated in FIG. 8, a nonlinear neural network with two input nodes is taken as an example to illustrate the training and assessment of the relationship between voltage and total current in a parallel circuit containing two fixed resistors based on DNA strand displacement reaction neural network.


Raw training data is U1∈{1.2V, 1.4V, 1.6V, . . . , 5.0V}, U2∈{1.3V, 1.5V, 1.7V, . . . , 5.1V} and I∈{0.37A, 0.43A, 0.49A, . . . , 1.51A} and there are 20 sets of data in total. The value of ρ is that ρ=3. Initial concentrations of DNA strands X1+, X1, X2+ and X2 are set as [X1+]0∈{2.2632 nM, 2.4211 nM, 2.5789 nM, . . . , 5.2632 nM} [X1]0∈{1.6316 nM, 1.6842 nM, 1.7368 nM, . . . , 2.6316 nM}, [X2+]0∈{2.3158 nM, 2.4737 nM, 2.6316 nM, . . . , 5.3158 nM}, [X2]0∈{1.6316 nM, 1.6842 nM, 1.7368 nM, . . . , 2.6316 nM} Initial concentration of auxiliary DNA molecules and reaction rates are set in Table 1.



FIG. 9 illustrated the update trajectory of weights during 20 rounds of training. It is evident that after training, the endpoint value of weights is very close to the target value, indicating that the linear learning machine has good learning ability.


As illustrated in FIG. 10, in 20 rounds of training, the average relative error is all around 0.2, and the training goal is basically achieved.



FIG. 11 illustrated a total number of training times required to achieve the training objectives in 20 rounds of training, which is clearly 4.









TABLE 1







concentrations of DNA strands and reaction rate settings












Reaction rate
Value
Concentration
Value

















q1
0.01/nM/s
[I+]0
2
nM



q2
0.001/nM/s
[I]0
2
nM



q3
0.001/nM/s
[Y+]0
20
nM



qm
1.0/nM/s
[Y]0
20
nM





[C+]0
20
nM





[C]0
20
nM





[H+]0
1
nM





[H]0
1
nM





+]0
2
nM





]0
2
nM





+]0
100
nM





]0
100
nM





+]0
200
nM





]0
200
nM





Cm
2000
nM










(3) Testing and Assessment of Linear Machine Learning.

The testing of machine learning includes multiple rounds of testing. One round of testing includes P sets of training data. The test data sets are shuffled to obtain another round of test data. In the first round of testing, the partial resistance and total voltage values are represented by Φi=[βi(1,1), βi(2,1), . . . , Bi(P,1)] and Γ={{tilde over (d)}′1(1), {tilde over (d)}′1(2), . . . , {tilde over (d)}′1(M)} respectively. The data can be normalized as follows.









{







x
ˆ

i


(

p
,
g

)

=

ρ




β
i

(

p
,
g

)

/
θ











d
ˆ

g


(
p
)

=

ρ





d
˜

g


(
p
)

/
θ










(
25
)







Where βi(p,g) represents an i-th data of a p-th set of data in a g-th round of training. p=1, 2, . . . , P, g=1, 2, . . . , G, θi=max(Φi)−min(Φi), θ=max([θ1, θ2, . . . θN]) and {tilde over (d)}′1(p)=Σi=1Nwixi(g,p).


In the g-th round of testing, the relative error e′g(p) is defined as follows.











e
g


(
p
)

=





y
˜

g


(
p
)

-



d
ˆ

g


(
p
)





d
ˆ

l


(
p
)






(
26
)








where








y
ˆ

g

(
p
)

=




i
=
1

N




W
ˆ

i






x
ˆ

i


(
k
)

.







{tilde over (w)}i represents the weight obtained after this round of training.


To assess the results of this round of training, the average relative error during the testing phase is defined as follows:











e
_

g


=


1
P






p
=
1

P




"\[LeftBracketingBar]"



e
g

(
k
)



"\[RightBracketingBar]"








(
27
)







Still taking the nonlinear neural network with two input nodes as an example to illustrate the test results of the neural network based on DNA strand replacement reaction. The raw data for the test are as follows.


U1∈{1.27V, 1.31V, 1.34V, . . . , 1.90V}, U2∈{1.24V, 1.28V, 1.31V, . . . , 1.86V} and I∈{0.77A, 0.80A, 0.83A, . . . , 1.64A} and there are 30 sets of data in total. As illustrated in FIG. 12, average relative error of 30 sets of data in each round of 20 rounds of testing are all about 0.1, indicating the neural network based on DNA strand displacement reaction basically meets the testing requirements.


The above embodiments are only used to illustrate the technical solution of the present disclosure, and not to limit it. Although the present disclosure has been described in detail with reference to the aforementioned embodiments, those skilled in the art should understand that they can still modify the technical solutions recorded in the aforementioned embodiments, or equivalently replace some or all of the technical features. And these modifications or replacements do not make the essence of the corresponding technical solutions deviate from the scope of the technical solutions of the various embodiments of the present disclosure.

Claims
  • 1. A linear machine learning method based on deoxyribonucleic acid (DNA) hybridization reaction technology, comprising: a training part, an algorithm part and a testing part; wherein, (1) reaction expressions of the training part are as follows:
  • 2. The linear machine learning method as claimed in claim 1, wherein a reaction expression of the first catalytic reaction module is:
  • 3. The linear machine learning method as claimed in claim 1, wherein the first catalytic reaction module, the second catalytic reaction module and the annihilation reaction module in the training part, the algorithm part and the testing part belong to a same type of reaction module but are not a same reaction module.
Priority Claims (1)
Number Date Country Kind
2023108023956 Jul 2023 CN national