The present application relates to the data processing field, and in particular, to a system and a method for training a parameter set in a neural network in the data processing field.
A neural network is a mathematical model in which information is processed by simulating a cerebral neural synaptic structure, is abstraction, simplification, and simulation of a human brain, and may reflect a basic property of the human brain. The neural network includes a large quantity of nodes (which are also referred as neurons) and weighted connections between the nodes. Each node represents a specific output function, called an excitation function, and a connection between every two nodes represents a weighted value for a signal passing through the connection. The neural network may be expressed by using a mathematical function:
Y=f(X,W),
where X represents an input of a network, Y represents an output of the network, and W represents a parameter set of the network.
The following describes an issue about training of the neural network briefly by using supervised learning as an example. The training of the neural network is to seek for the parameter set W of the foregoing function. A training process of the neural network is to offer a data set:
D={(X1,Y1), (X2,Y2), . . . , (XN,YN)}
for training, define a value function of each training data (Xi,Yi) as:
E
i=½(Yi−f(Xi,W))2,
and determine W, so as to minimize a value of:
Deep learning is one of training methods for the neural network. Currently, deep learning can have been well used for resolving actual application problems such as speech recognition, image recognition, and text processing. Generally, in the neural network, training needs to be performed by using a great deal of training data, so as to ensure that an operation result of the neural network reaches a certain degree of accuracy. Correspondingly, a larger training data scale indicates a larger calculation amount and a longer time required for training. To improve a training speed of the neural network, coprocessors such as a graphic processing unit (GPU) are widely applied to calculation of training of deep learning. However, these coprocessors have relatively small memory, and cannot accommodate a weighted parameter set of a large-scale neural network.
Moreover, according to the existing technical solution, a main control node sends copies of a neural network to operation nodes and instructs the operation nodes to perform training. Each operation node is equipped with at least a GPU to perform operation processing. The main control node regularly queries statuses of the operation nodes when the operation nodes perform the training, and updates weighted parameter of the copies of the neural network on the main control node and the operation nodes after the operation nodes are in a stop state. In the prior art, numerous operation nodes are used for cooperatively training a large-scale neural network, and in a conventional synchronous updating manner, all calculation nodes in a system can only simultaneously perform training based on a same parameter set W, so that overall performance of the system is limited by a slowest node and a network bandwidth of the system. When one or some nodes are disabled, an entire training process is severely affected.
Therefore, an existing training system of a neural network has poor reliability and supports only one main control node, and when the main control node is disabled, entire training fails. Moreover, operation nodes of the existing training system can simultaneously perform training only based on a same parameter set W, and a scale and overall performance of the system are limited by memory sizes of the main control node and the operation nodes.
Embodiments of the present application provides a system and a method for training a parameter set in a neural network, which can improve reliability of a training process of a neural network and training efficiency.
According to a first aspect, a system for training a parameter set in a neural network is provided, where the system includes a main-control-node set, where the main-control-node set includes M main control nodes, the main-control-node set is used for controlling a process of training the parameter set in the neural network and storing a data set and a parameter set that are used in the process of training the parameter set, the data set includes multiple data subsets, the parameter set includes multiple parameter subsets, the multiple parameter subsets are stored separately in different main control nodes, a set of parameter subsets stored in all main control nodes in the main-control-node set is the parameter set, every two of the M main control nodes are in a communication connection, and at least one main control node of the M main control nodes is configured to back up the parameter set, where M is a positive integer greater than 1. The system also includes N training-node sets, where each of the N training-node sets is in a communication connection with the main-control-node set, the training-node set includes multiple training nodes, the training node is configured to receive the data subset and the parameter set that are delivered by the main-control-node set, train, according to the received data subset and parameter set, a parameter subset for which the training node is responsible, and send a training result to a main control node storing the parameter subset, where N is a positive integer greater than 1, data subsets used any two of the N training-node sets for training are different, and a set of parameter subsets trained by all training nodes in each training-node set is the parameter set.
With reference to the first aspect, in a first possible implementation manner of the first aspect, the training result is a parameter variation, obtained by the training node by training, according to the received data subset and parameter set, the parameter subset for which the training node is responsible, of the parameter subset for which the training node is responsible, and the main control node in the main-control-node set is further configured to: receive the parameter variation sent by the training node; and update, according to the parameter variation, the parameter subset stored in the main control node.
With reference to the first aspect or the first possible implementation manner of the first aspect, in a second possible implementation manner of the first aspect, the main-control-node set is specifically used for: dividing the parameter set into multiple parameter subsets; storing the multiple parameter subsets separately in different main control nodes, where the set of the parameter subsets stored in all of the main control nodes in the main-control-node set is the parameter set; and determining each training node in the N training-node sets according to sizes of the multiple parameter subsets.
With reference to any one of the first aspect and the first and second possible implementation manners of the first aspect, in a third possible implementation manner of the first aspect, the main control node is specifically configured to: update, at a first time point according to a parameter variation sent by a first training node of a first-training-node set, the parameter subset stored in the main control node; and update, at a second time point according to a parameter variation sent by a second training node of a second-training-node set, the parameter subset stored in the main control node.
With reference to any one of the first aspect and the first to third possible implementation manners of the first aspect, in a fourth possible implementation manner of the first aspect, the main-control-node set is specifically used for: determining, according to an accuracy of the training result, whether to stop the process of training the parameter set.
With reference to any one of the first aspect and the first to fourth possible implementation manners of the first aspect, in a fifth possible implementation manner of the first aspect, the training node is further configured to: receive an instruction sent by the main-control-node set and stop the process of training the parameter set.
With reference to any one of the first aspect and the first to fifth possible implementation manners of the first aspect, in a sixth possible implementation manner of the first aspect, every two training nodes in a same training-node set are in a communication connection.
According to a second aspect, a method for training a parameter set in a neural network is provided, where the method is performed by the main-control-node set in the system for training a parameter set in a neural network according to any one of the first aspect and the first to sixth possible implementation manners of the first aspect, where the system further includes N training-node sets, where the main-control-node set includes M main control nodes, and every two of the M main control nodes are in a communication connection, where M is a positive integer greater than 1, and N is a positive integer greater than 1. The method includes storing, by the main-control-node set, a data set and a parameter set that are used for training, where the data set includes multiple data subsets, the parameter set includes multiple parameter subsets, the multiple parameter subsets are stored separately in different main control nodes, a set of parameter subsets stored in all main control nodes in the main-control-node set is the parameter set, and at least one main control node of the M main control nodes is configured to back up the parameter set. The method also includes delivering, by a main control node in the main-control-node set, a data subset and a parameter subset to a training node that is responsible for training the parameter subset stored in the main control node; and receiving, by the main control node in the main-control-node set, a training result sent by the training node, where the training node belongs to a training-node set, the training-node set is in a communication connection with the main-control-node set, the training-node set includes multiple training nodes, and the training result is obtained by performing, according to the received data subset and parameter set that are delivered by the main-control-node set, training.
With reference to the second aspect, in a first possible implementation manner of the second aspect, the training result is a parameter variation, obtained by the training node by training, according to the received data subset and parameter set that are delivered by the main-control-node set, the parameter subset for which the training node is responsible, of the parameter subset, and the method further includes: receiving, by the main control node in the main-control-node set, the parameter variation sent by the training node; and updating, by the main control node in the main-control-node set according to the parameter variation, the parameter subset stored in the main control node.
With reference to the second aspect or the first possible implementation manner of the second aspect, in a second possible implementation manner of the second aspect, the storing, by the main-control-node set, a data set and a parameter set that are used for training includes: dividing, by the main-control-node set, the parameter set into multiple parameter subsets; and storing the multiple parameter subsets separately in different main control nodes, where the set of the parameter subsets stored in all of the main control nodes in the main-control-node set is the parameter set; and the method further includes: determining, by the main-control-node set, each training node in the N training-node sets according to sizes of the multiple parameter subsets.
With reference to any one of the second aspect and the first and second possible implementation manners of the second aspect, in a third possible implementation manner of the second aspect, the updating, by the main control node in the main-control-node set according to the parameter variation, the parameter subset stored in the main control node includes: updating, by the main control node in the main-control-node set at a first time point according to a parameter variation sent by a first training node of a first-training-node set, the parameter subset stored in the main control node; and updating, by the main control node in the main-control-node set at a second time point according to a parameter variation sent by a second training node of a second-training-node set, the parameter subset stored in the main control node.
With reference to any one of the second aspect and the first to third possible implementation manners of the second aspect, in a fourth possible implementation manner of the second aspect, the method further includes: determining, by the main-control-node set according to an accuracy of the training result, whether to stop the process of training the parameter set.
With reference to any one of the second aspect and the first to fourth possible implementation manners of the second aspect, in a fifth possible implementation manner of the second aspect, at least one main control node stores and is responsible for one of the parameter subsets, correspondingly, at least two training nodes are responsible for one of the parameter subsets, the at least two training nodes belong to different training-node sets, data subsets used any two of the multiple training-node sets for training are different, and a set of parameter subsets trained by all training nodes in each training-node set is the parameter set.
With reference to any one of the second aspect and the first to fifth possible implementation manners of the second aspect, in a sixth possible implementation manner of the second aspect, every two training nodes in a same training-node set are in a communication connection.
Based on the foregoing technical solutions, according to the system and the method for training a parameter set in a neural network provided in the embodiments of the present application, a training process is controlled by a main-control-node set including multiple main control nodes, every two of which are in a communication connection, which can avoid a case in which entire training fails when a main control node is disabled, and can improve reliability of the training process; and a parameter set is trained in parallel by configuring multiple training-node sets, which can improve training efficiency.
To describe the technical solutions in the embodiments of the present application more clearly, the following briefly introduces the accompanying drawings required for describing the embodiments of the present application. Apparently, the accompanying drawings in the following description show merely some embodiments of the present application, and a person of ordinary skill in the art may still derive other drawings from these accompanying drawings without creative efforts.
The following clearly describes the technical solutions in the embodiments of the present application with reference to the accompanying drawings in the embodiments of the present application. Apparently, the described embodiments are some but not all of the embodiments of the present application. All other embodiments obtained by a person of ordinary skill in the art based on the embodiments of the present application without creative efforts shall fall within the protection scope of the present application.
Therefore, according to the system for training a parameter set in a neural network provided in this embodiment of the present application, a training process is controlled by a main-control-node set including multiple main control nodes, every two of which are in a communication connection, which can avoid a case in which entire training fails when a main control node is disabled, and can improve reliability of the training process; and a parameter set is trained in parallel by configuring multiple training-node sets, which can improve training efficiency.
Specifically, the system 100 for training a parameter set includes a main-control-node set 110 and at least two training-node sets 120. The main-control-node set 110 includes at least two main control nodes, every two of the main control nodes are in a communication connection, and at least one main control node is configured to back up the parameter set, which can improve reliability of a training process. The training-node set 120 may be obtained by means of division performed by the main-control-node set 110 according to a data processing scale and performance (such as a memory size) of a training node for forming the training-node set 120.
The system 100 for training a parameter set in this embodiment of the present application may be applied to a training process of a neural network. Inputs of the training process of the neural network are a neural network function:
Y=f(X,W)
an initial parameter set, and a data set D for training, and an output is a parameter set W of a neural network obtained after training. The main-control-node set 110 is used for controlling a training process, for example, the main-control-node set 110 controls the training process to start or end, controls a data subset used each training-node set, and determines each training node in a training-node set. The main-control-node set 110 is further used for storing a data set D and a parameter set W that are used in the training process. The parameter set W includes multiple parameter subsets, the multiple parameter subsets are stored separately in different main control nodes, and a set of parameter subsets stored in all main control nodes in the main-control-node set 110 is the parameter set W.
The training node in the training-node set 120 is configured to receive a data subset delivered by the main-control-node set 110 and a current parameter set W, and train a parameter subset for which the training node is responsible according to the received data subset and current parameter set W, and send a parameter variation ΔW, which may be obtained by performing training according to the data subset and the current parameter set W and is used for updating, to the main control node. In the training process, data subsets used any two of the N training-node sets 120 for training are different, and a set of parameter subsets trained by all training nodes in each training-node set 120 is the parameter set. That is, multiple training-node sets 120 process different data subsets in parallel. For a same parameter subset, multiple training nodes train the parameter subset at a same time point, which can improve efficiency of the training process.
Therefore, according to the system for training a parameter set in a neural network provided in this embodiment of the present application, a training process is controlled by a main-control-node set including multiple main control nodes, every two of which are in a communication connection, which can avoid a case in which entire training fails when a main control node is disabled, and can improve reliability of the training process; and a parameter set is trained in parallel by configuring multiple training-node sets, which can improve training efficiency.
In this embodiment of the present application, a data set includes multiple data subsets, and a parameter set includes multiple parameter subsets. Data subsets used any two of the N training-node sets 120 for training are different. At least two training nodes train a same parameter subset, and the two training nodes belong to different training-node sets 120.
Specifically, the system 100 for training a parameter set includes more than one training-node set 120. In this case, a data set stored in the main-control-node set 110 includes multiple data subsets, and during training, the main-control-node set 110 delivers different data subsets to different training-node sets 120 A parameter set stored in the main-control-node set no includes multiple parameter subsets, and main control nodes in the main-control-node set no separately store and are responsible for maintaining different parameter subsets. A training node, responsible for a parameter subset, in the training-node set 120 receives, from a corresponding main control node, the parameter subset which the main control node stores and is responsible for maintaining, and a set of parameter subsets received from multiple main control nodes is the parameter set. According to the data subset and the parameter set, a training node trains a parameter subset for which the training node is responsible. At least two training nodes train a same parameter subset, and these two training nodes belong to different training-node sets 120. That is, when there are multiple training-node sets 120, the multiple training-node sets 120 process different data subsets in parallel. For a same parameter subset, multiple training nodes train the parameter subset at a same time point, which can improve efficiency of the training process.
It should be understood that, a quantity of the main control nodes in the main-control-node set 110 in the system 100 shown in
Therefore, according to the system for training a parameter set in a neural network provided in this embodiment of the present application, a training process is controlled by a main-control-node set including multiple main control nodes, every two of which are in a communication connection, which can avoid a case in which entire training fails when a main control node is disabled, and can improve reliability of the training process. Moreover, a parameter set is trained in parallel by configuring multiple training-node sets, which can improve training efficiency.
The main control nodes in the main-control-node set 110 and the training nodes in the training-node set 120 are all calculation devices.
Optionally, as an embodiment, at a same time point, a parameter set used at least one of the N training-node sets 120 for training is different from the parameter set currently stored in the main-control-node set 110.
Alternatively, optionally, as an embodiment, the main control node is specifically configured to: update, at a first time point according to a parameter variation sent by a first training node of a first-training-node set, the parameter subset stored in the main control node; and update, at a second time point according to a parameter variation sent by a second training node of a second-training-node set, the parameter subset stored in the main control node.
Specifically, all training-node sets 120 in the system 100 run independently and in parallel and do not affect each other. Disability of any training-node set 120 does not affect continuous training of the entire system 100. At a time point of the training process, at least one of the N training-node sets 120 calculates a difference between a used parameter set and the parameter set currently stored in the main-control-node set 110. In other words, at a time point of the training process, a parameter set used at least one of the N training-node sets 120 for training is different from a parameter set used another training-node set 120 for training. That is, the main-control-node set 110 updates parameter sets W asynchronously. The main control node updates, at the first time point according to the parameter variation sent by the first training node of the first-training-node set, the parameter subset stored in the main control node; and updates, at the second time point according to the parameter variation sent by the second training node of the second-training-node set, the parameter subset stored in the main control node. At a time point, a current parameter set W of the main-control-node set 110 may be already different from a parameter set W currently used the training-node set 120 for the training.
Optionally, as an embodiment, the main-control-node set 110 may be specifically used for: dividing the parameter set into multiple parameter subsets; storing the multiple parameter subsets separately in different main control nodes, where the set of the parameter subsets stored in all of the main control nodes in the main-control-node set 110 is the parameter set; and determining each training node in the N training-node sets 120 according to sizes of the multiple parameter subsets.
Specifically, at the start of the training, the main-control-node set 110 performs initialization work, for example, the main-control-node set 110 obtains the training-node sets 120 by means of division, configures the data set and the parameter set for training, and initializes an original model. The configuring a parameter set W for training is specifically dividing the parameter set W into multiple parameter subsets W1,W2, . . . , WK. Each main control node is responsible for maintaining one or more parameter subsets. If a main control node Mj is responsible for storing, updating, and maintaining a parameter subset Wi, Mj is referred to as a sink main node of Wi.
According to a size of the parameter set W and a memory (memory of a coprocessor of a training node) size of each training node for forming the training-node set 120, the main-control-node set 110 divides all training nodes configured to form the training-node set 120. Generally, a larger size of the parameter subset indicates a stronger capability of a training node that needs to be allocated to the parameter subset. It is assumed that there are totally P training-node sets 120 that are recorded as C1,C2, . . . , CP. Each training node is responsible for at least one parameter subset, and the training-node sets 120 cooperatively store and process an entire copy of the parameter set W.
Optionally, as an embodiment, the main-control-node set 110 backs up the parameter set by using a disk array redundant array of independent disks (RAID) 0/1/5/6 or erasure coding.
Specifically, to ensure reliability of the system 100, the main-control-node set 110 may back up the parameter set by using an encoding method of RAID 0/1/5/6 or erasure coding (Erasure Coding). In this way, in a case in which some main control nodes are disabled, the system 100 can recover the disabled parameter subset by using a corresponding decoding operation, so as to maintain normal running. It should be understood that, the reliability of the system 100 may be further ensured by using another encoding method, which is not limited in this embodiment of the present application.
Optionally, as an embodiment, the training node may be specifically configured to: receive an instruction sent by the main-control-node set 110 and stop the process of training the parameter set.
Specifically, for a training node in a training-node set Ck, it is required to access a sink main node of a parameter subset for which the training node is responsible, and download a copy of a latest parameter subset. A set of all latest parameter subsets acquired by all training nodes of the training-node set Ck by using a communications network is a latest parameter set, which is recorded as Wk. Different training-node sets may acquire latest parameter sets W from the main-control-node set 110 at different time points. However, the parameter set W constantly changes. Therefore, at a same time point, copies, used different training-node sets for calculation, of the parameter set W may be different.
If training needs to be performed, the training node in the training-node set Ck further needs to acquire some data of the data set from the main-control-node set 110o, that is, a data subset. Data subsets acquired by training nodes in a same training-node set are the same. Further, the training node performs training according to the parameter set Wk and the data subset, so as to obtain a parameter variation ΔWik corresponding to the parameter subset Wi for which the training node is responsible. The training node sends the parameter variation ΔWik, obtained by training, to a main control node responsible for a corresponding parameter subset Wi, that is, a sink main node. A set of parameter variations ΔWik obtained by calculation by all training nodes in the training-node set Ck is recorded as ΔWk. A manner in which the training node acquires the parameter subset and data from the main-control-node set no is not limited in this embodiment of the present application.
In the training process, the training node performs a model training by using a constantly received parameter set and data subset as input and until receiving from the main-control-node set 110 a training stop instruction sent by the main-control-node set 110, the training node stops the process of training the parameter set.
Optionally, as an embodiment, if in the training process, parameters of the training nodes in the training-node set are correlated, it is necessary for the training nodes to exchange data with each other. In this case, every two training nodes in a same training-node set may be in a communication connection.
Optionally, as an embodiment, the training result is a parameter variation, obtained by the training node by training, according to the received data subset and parameter set, the parameter subset for which the training node is responsible, of the parameter subset for which the training node is responsible, and the main control node in the main-control-node set 110 is further configured to: receive the parameter variation sent by the training node; and update, according to the parameter variation, the parameter subset stored in the main control node.
Specifically, the main control node in the main-control-node set 110 receives a parameter variation ΔWik, which is obtained by the training node by performing training according to the data set and the parameter set and used for updating, from a training node in a training-node set Ck, so as to update the parameter subset Wi for which the main control node in the main-control-node set is responsible. That is, after receiving an entire parameter set variation ΔWk from a training-node set Ck, the main-control-node set updates the parameter set W of the neural network. The main-control-node set updates the parameter set W asynchronously, that is, at a same time point, a current parameter set W of the main-control-node set may be already different from a parameter set Wk used a training-node set Ck in a training process. Such an asynchronous updating manner may make full use of training capabilities of all training-node sets. In addition, a specific method for updating the parameter set W by the main-control-node set is not limited in this embodiment of the present application.
Optionally, as an embodiment, the main-control-node set is specifically used for: determining, according to an accuracy of the training result, whether to stop the process of training the parameter set.
Specifically, the main-control-node set 110 determines, according to whether the training result is accurate, whether the current training should be stopped. For example, the main-control-node set 110 may determine, when a variation ΔWk of a parameter set W is less than a threshold, to stop the training process; or determine, when an updated parameter set W makes a change value of a result Y, obtained by calculation according to a mathematical function:
Y=f(X,W)
of a parameter set W and a neural network, less than a threshold, to stop the training process, which is not limited in this embodiment of the present application.
The following describes a working process of the system 100 provided in this embodiment of the present application in detail with reference to a specific example.
The system 100 provided in this embodiment of the present application is applied to an image classification system based on a deep convolutional neural network, and performs training by using an optimization algorithm based on mini-batch stochastic gradient descent (Mini-batch Stochastic Gradient Descent). An input X of the deep convolutional neural network is an image, an output Y is an image category, and a data set of a training process is:
D={(Xi,Yi)}.
A parameter set of the convolutional neural network is W, and parameters included in a parameter set trained by the system are a mini-batch size m and a learning rate α.
210: A system 100 includes P training-node sets Ck (k=1, 2, . . . , p). Training nodes Ck1 and Ck2 download respectively from main control nodes M1 and M2 latest parameter subsets W1 and W2, which are recorded as W1k and W2k. If the main control node M1 is disabled, the training node Ck1 may download the parameter subset W2 from the main control node M2, download W3=W1⊕W2 from a main control node M3, and then obtain a parameter subset W1k by training W2⊕W3. If the main control node M2 s disabled, the training node Ck2 may download the parameter subset W1 from the main control node M1, download W3=W1⊕W2 from the main control node M3, and then obtain a parameter subset W2k by training W1⊕W3.
220: Both of the training nodes Ck1 and Ck2 receive a same batch of training data:
{Dik=(Xik,Yik)|i=1,2, . . . , m},
from a main-control-node set, and perform forward transfer training based on the parameter subsets W1k and W2k respectively. In a training process, the training nodes Ck1 and Ck2 may communicate with each other, so as to perform necessary data exchange.
230: For each training data:
D
i
k=(Xik,Yik),
where i=1,2, . . . m, the training nodes Ck1 and Ck2 separately obtain errors:
E
i
k=½(Yik−f(Xik,Wk))2,
corresponding to the training nodes by training, where i=1, 2, . . . m , and then perform back transfer training by using an error back propagation (Error Back Propagation, EBP) algorithm, so as to obtain:
by training respectively. In the training process, the training nodes Ck1 and Ck2 may communicate with each other, so as to perform necessary data exchange.
240: The training nodes Ck1 and Ck2 obtain parameter variations:
respectively.
250: The training nodes Ck1 and Ck2 upload ΔW1k and ΔW2k to the main control nodes M1 and M2 respectively. The training nodes Ck1 and Ck2 repeat steps 210 to 250 until receiving a training stop instruction from the main-control-node set.
260: The main-control-node set includes the main control nodes M1 and M2. Step 260 is performed in parallel with steps 210 to 250. The main control nodes M1 and M2 receive the parameter variations ΔW1k and ΔW2k from the training nodes Ck1 and Ck2 of training-node sets respectively. According to the parameter variations ΔW1k and ΔW2k, the main control nodes M1 and M2 update the parameter subsets W1 and W2 according to the following formulas:
W
1
=W
1
αΔW
1
k, and
W
2
=W
2
−αΔW
2
k
respectively. The main control nodes M1 and M2 transmit updated parameter subsets W1 and W2 to the main control node M3. The main control node M3 updates W3 according to the following formula:
W
3
=W
1
⊕W
2,
270: The main-control-node set determines, according to an accuracy of a training result, whether to stop the training process. If a training stop condition is not met, steps 210 to 270 are repeated; or if a training stop condition is met, step 280 is performed.
280: The main-control-node set sends the training stop instruction to the training-node sets.
Therefore, according to the system for training a parameter set in a neural network provided in this embodiment of the present application, a training process is controlled by a main-control-node set including multiple main control nodes, every two of which are in a communication connection, which can avoid a case in which entire training fails when a main control node is disabled, and can improve reliability of the training process; and moreover, a parameter set is trained in parallel by configuring multiple training-node sets, which can improve training efficiency.
The following describes a method 300 for training a parameter set in a neural network corresponding to an embodiment of the present application in detail.
S310: The main-control-node set stores a data set and a parameter set that are used for training, where the data set includes multiple data subsets, the parameter set includes multiple parameter subsets, the multiple parameter subsets are stored separately in different main control nodes, a set of parameter subsets stored in all main control nodes in the main-control-node set is the parameter set, and at least one main control node of the M main control nodes is configured to back up the parameter set.
S320: A main control node in the main-control-node set delivers a data subset and a parameter subset to a training node that is responsible for training the parameter subset stored in the main control node.
S330: The main control node in the main-control-node set receives a training result sent by the training node, where the training node belongs to a training-node set, the training-node set is in a communication connection with the main-control-node set, the training-node set includes multiple training nodes, and the training result is obtained by performing, according to the received data subset and parameter set that are delivered by the main-control-node set, training.
Therefore, according to the method for training a parameter set in a neural network provided in this embodiment of the present application, a training process is controlled by a main-control-node set including multiple main control nodes, every two of which are in a communication connection, which can avoid a case in which entire training fails when a main control node is disabled, and can improve reliability of the training process; and a parameter set is trained in parallel by multiple training-node sets, which can improve training efficiency.
Optionally, as an embodiment, the training result is a parameter variation, obtained by the training node by training, according to the received data subset and parameter set that are delivered by the main-control-node set, the parameter subset for which the training node is responsible, of the parameter subset, and the method 300 further includes: receiving, by the main control node in the main-control-node set, the parameter variation sent by the training node; and updating, by the main control node in the main-control-node set according to the parameter variation, the parameter subset stored in the main control node.
Optionally, as an embodiment, the storing, by the main-control-node set, a data set and a parameter set that are used for training includes: dividing, by the main-control-node set, the parameter set into multiple parameter subsets; and storing the multiple parameter subsets separately in different main control nodes, where the set of the parameter subsets stored in all of the main control nodes in the main-control-node set is the parameter set; and the method 300 further includes: determining, by the main-control-node set, each training node in the N training-node sets according to sizes of the multiple parameter subsets.
Optionally, as an embodiment, the updating, by the main control node in the main-control-node set according to the parameter variation, the parameter subset stored in the main control node includes: updating, by the main control node in the main-control-node set at a first time point according to a parameter variation sent by a first training node of a first-training-node set, the parameter subset stored in the main control node; and updating, by the main control node in the main-control-node set at a second time point according to a parameter variation sent by a second training node of a second-training-node set, the parameter subset stored in the main control node.
Optionally, as an embodiment, the method 300 further includes: determining, by the main-control-node set according to an accuracy of the training result, whether to stop the process of training the parameter set.
Optionally, as an embodiment, at least one main control node stores and is responsible for one of the parameter subsets, correspondingly, at least two training nodes are responsible for one of the parameter subsets, the at least two training nodes belong to different training-node sets, data subsets used any two of the multiple training-node sets for training are different, and a set of parameter subsets trained by all training nodes in each training-node set is the parameter set.
Optionally, as an embodiment, every two training nodes in a same training-node set are in a communication connection.
Therefore, according to the method for training a parameter set in a neural network provided in this embodiment of the present application, a training process is controlled by a main-control-node set including multiple main control nodes, every two of which are in a communication connection, which can avoid a case in which entire training fails when a main control node is disabled, and can improve reliability of the training process; and moreover, training is performed in parallel by configuring multiple training-node sets, which can improve training efficiency.
It should be understood that, in the embodiments of the present application, “Y corresponding to X” represents that Y and X are correlated, and Y may be determined according to X. It should be further understood that, determining Y according to X does not mean that Y is determined only according to X, but means that Y may be further determined according to X and/or other information.
A person of ordinary skill in the art may be aware that, in combination with the examples described in the embodiments disclosed in this specification, units and algorithm steps may be implemented by electronic hardware, computer software, or a combination thereof. To clearly describe the interchangeability between the hardware and the software, the foregoing has generally described compositions and steps of each example according to functions. Whether the functions are performed by hardware or software depends on particular applications and design constraint conditions of the technical solutions. A person skilled in the art may use different methods to implement the described functions for each particular application, but it should not be considered that the implementation goes beyond the scope of the present application.
It may be clearly understood by a person skilled in the art that, for the purpose of convenient and brief description, for a detailed working process of the foregoing system, apparatus, and unit, reference may be made to a corresponding process in the foregoing method embodiments, and details are not described herein again.
In the several embodiments provided in the present application, it should be understood that the disclosed system, apparatus, and method may be implemented in other manners. For example, the described apparatus embodiment is merely exemplary. For example, the unit division is merely logical function division and may be other division in actual implementation. For example, a plurality of units or components may be combined or integrated into another system, or some features may be ignored or not performed. In addition, the displayed or discussed mutual couplings or direct couplings or communication connections may be implemented by using some interfaces. The indirect couplings or communication connections between the apparatuses or units may be implemented in electronic, mechanical, or other forms.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one position, or may be distributed on a plurality of network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments of the present application.
In addition, functional units in the embodiments of the present application may be integrated into one processing unit, or each of the units may exist alone physically, or two or more units are integrated into one unit. The integrated unit may be implemented in a form of hardware, or may be implemented in a form of a software functional unit.
When the integrated unit is implemented in the form of a software functional unit and sold or used as an independent product, the integrated unit may be stored in a computer-readable storage medium. Based on such an understanding, the technical solutions of the present application essentially, or the part contributing to the prior art, or all or some of the technical solutions may be implemented in a form of a software product. The computer software product is stored in a storage medium, and includes several instructions for instructing a computer device (which may be a personal computer, a server, or a network device) to perform all or some of the steps of the methods described in the embodiments of the present application. The foregoing storage medium includes: any medium that can store program code, such as a universal serial bus (USB) flash drive, a removable hard disk, a read-only memory (ROM), a random access memory (RAM), a magnetic disk, or an optical disc.
The foregoing descriptions are merely specific embodiments of the present application, but are not intended to limit the protection scope of the present application. Any modification or replacement readily figured out by a person skilled in the art within the technical scope disclosed in the present application shall fall within the protection scope of the present application. Therefore, the protection scope of the present application shall be subject to the protection scope of the claims.
Number | Date | Country | Kind |
---|---|---|---|
201510036813.0 | Jan 2015 | CN | national |
This application is a continuation of International Application No. PCT/CN2015/086011, filed on Aug. 4, 2015, which claims priority to Chinese Patent Application No. 201510036813.0 filed on Jan. 26, 2015. The disclosures of the aforementioned applications are hereby incorporated by reference in their entireties.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/CN2015/086011 | Aug 2015 | US |
Child | 15455259 | US |