This disclosure relates to the field of wireless communication technologies, and in particular, to a model scheduling method and an apparatus.
To cope with the vision of inclusive intelligence in the future, intelligence is further evolved at a wireless network architecture layer, and artificial intelligence (AI) is to be deeply integrated with a wireless network. A concept of federated learning (FL) is proposed to effectively resolve difficulties in current development of the artificial intelligence. While user data privacy and security are fully ensured, the FL facilitates edge devices and a central-end server to collaborate to efficiently complete learning tasks of a model.
However, in a federated learning system, a premise that a plurality of local models can be fused in a weighted averaging manner is that the local models have a same structure and a same quantity of parameters. However, in a network, distributed nodes have different device capabilities, but using a same local model for the distributed nodes typically limits flexibility of the distributed nodes, and different requirements of different distributed nodes cannot be met. In addition, each distributed node needs to report a local model in each training process, resulting in high air interface transmission overheads.
This disclosure provides a model scheduling method and an apparatus, to meet requirements of different distributed nodes and reduce air interface transmission overheads.
According to a first aspect, a model scheduling method is provided. The method may be performed by a first node. Optionally, the first node may be a communication device or a chip/chip system. In the method, the first node receives first information from a second node, where the first information indicates a first model. The first node processes a second model based on the first model, to obtain a third model. A structure of the third model is the same as a structure of the first model, and the second model and the first model are heterogeneous. The first node sends the second model to the second node. The third model meets a preset condition.
Based on the foregoing solution, each node may process the second model based on the first model, to obtain the third model. Because structures of third models of all nodes are the same, impact on output results of the third models of the nodes due to different structures of models is reduced. Each node selects, based on evaluation of the third model, whether to send the second model to a central node, so that air interface transmission overheads can be reduced.
In some embodiments, the first information includes a model index and/or a parameter of the first model. In a possible case, the first information may include the model index. The first model may be preset, and each first model may correspond to a model index. In another possible case, the first information may include the parameter of the first model. For example, the first information may include a parameter used to describe the first model. Based on the foregoing solution, the first model is indicated by using the model index, so that the air interface transmission overheads can be greatly reduced. In addition, because the first model is a lightweight model and has a small quantity of parameters, the first model is indicated by using the parameter of the first model. In comparison with some approaches, overheads needed for performing transmission of a complete model are small, and air interface utilization can be improved.
In some embodiments, the first node receives information about a fourth model from the second node. The fourth model is obtained by using a second model of at least one first node. The first node processes the fourth model based on the second model of the first node, to obtain a fifth model. A structure of the fifth model is the same as a structure of the second model of the first node. The first node trains the fifth model based on a preset parameter, to obtain an updated second model. Based on the foregoing solution, model accuracy may be improved through iterative training.
In some embodiments, the first node processes the updated second model based on a sixth model, to obtain an updated third model. The sixth model is indicated by the second node. The first node sends the updated second model to the second node. The updated third model meets the preset condition. Optionally, the sixth model and the first model may be a same model, or may be different models. The sixth model may be indicated by the second node. For example, the first node may receive indication information of the sixth model from the second node, and the indication information of the sixth model may include a model index and/or a parameter of the sixth model.
Based on the foregoing solution, the first node may select, in each training process based on a requirement, whether to update the first model, so that a fourth model obtained through training can better meet a requirement of the second node. In addition, the first node may evaluate performance of the second model based on the first model, and determine, based on an evaluation result, whether to send the second model to the second node. Therefore, a second model with poor performance may not be sent to the second node, so that accuracy of the fourth model obtained through the training can be improved.
In some embodiments, the preset condition includes a first threshold, and the first threshold indicates performance of the third model or the first threshold indicates a performance variation of the third model. Based on the foregoing solution, by using the first threshold, the third model is evaluated, in other words, a training status of a model of the first node is evaluated. When the performance or the performance variation of the third model meets an evaluation parameter, the second model is sent to the second node, so that the air interface transmission overheads can be reduced.
In some embodiments, the preset condition is indicated by the second node. Optionally, the preset condition may include the following condition 1 and condition 2. Condition 1: The preset condition includes the evaluation parameter. The evaluation parameter may indicate the performance of the third model, or indicate the performance variation of the third model. Optionally, the evaluation parameter may include the first threshold. Condition 2: The preset condition includes indication information. The first node may receive the indication information from the second node. The indication information may indicate the first node to send the second model. Based on the foregoing solution, the second node may flexibly indicate the preset condition based on the training status of the model.
In some embodiments, the first node inputs the preset parameter into the third model, to obtain an output result. The first node sends second information to the second node. The second information indicates that model training is unfulfilled, and the output result does not meet the first threshold. Based on the foregoing solution, when the third model does not meet the preset condition, the first node may send the second information to the second node, to notify the second node that the model training is unfulfilled.
In some embodiments, the first node inputs a preset parameter into the third model, to obtain an output result. The first node sends the output result to the second node. The first node receives the indication information from the second node, where the indication information indicates the first node to send the second model.
Based on the foregoing solution, depending on the output result sent by the first node, the second node makes a decision and schedules the first node to send the second model. In comparison with a case in which each first node reports the second model, the air interface transmission overheads can be reduced, and a model obtained by the second node through training can be more accurate. This improves model training efficiency.
In some embodiments, the first model is preset based on at least one of the following: a geographical location of the first node and a task executed by the first node.
Based on the foregoing solution, different first models are preset for first nodes in different geographical environments, so that the first model can be related to the geographical environment, and an output result obtained by using the first model can also be consistent with the geographical environment in which the first node is located. This may improve accuracy of the first model.
In some embodiments, the first node receives the first information from the second node through a sidelink. Alternatively, the first node receives the first information from the second node through a common channel. Alternatively, the first node receives the first information from the second node through a dedicated channel of the first node.
According to a second aspect, a model scheduling method is provided. The method may be performed by a second node. Optionally, the second node may be a communication device or a chip/chip system. In the method, the second node sends first information to at least one first node, where the first information indicates a first model. The second node receives a second model from the at least one first node, where the second model and the first model are heterogeneous. The first model is used to process the second model, to obtain a third model that has a same structure as the first model. The second node processes the second model to obtain a fourth model. The second node sends the fourth model to the at least one first node.
Based on the foregoing solution, each node may process the second model based on the first model, to obtain the third model. Because structures of third models of all nodes are the same, impact on output results of the third models of the nodes due to different structures of models is reduced. Each node selects, based on evaluation of the third model, whether to send the second model to a central node, so that air interface transmission overheads can be reduced.
In some embodiments, the first information includes a model index, and the model index indicates the first model. Alternatively, the first information includes a parameter of the first model. In a possible case, the first information may include the model index. The first model may be preset, and each first model may correspond to a model index. In another possible case, the first information may include the parameter of the first model. For example, the first information may include a parameter used to describe the first model. Based on the foregoing solution, the first model is indicated by using the model index, so that the air interface transmission overheads can be greatly reduced. In addition, because the first model is a lightweight model and has a small quantity of parameters, the first model is indicated by using the parameter of the first model. In comparison with some approaches, overheads needed for performing transmission of a complete model are small, and air interface utilization can be improved.
In some embodiments, the second node receives an updated second model from the at least one first node, where the updated second model is obtained based on the fourth model. Based on the foregoing solution, model accuracy may be improved through iterative training.
In some embodiments, the second node sends a first threshold to the at least one first node, where the first threshold indicates performance of the third model or the first threshold indicates a performance variation of the third model. Based on the foregoing solution, by using the first threshold, the first node evaluates the third model, in other words, evaluates a training status of a model of the first node. When the performance or the performance variation of the third model meets an evaluation parameter, the first node sends the second model to the second node, so that the air interface transmission overheads can be reduced.
In some embodiments, the second node receives second information from a third node, where the second information indicates that model training is unfulfilled, and the third node is one or more nodes in the at least one first node. Based on the foregoing solution, when the third model does not meet a preset condition, the third node may send the second information to the second node, to notify the second node that the model training is unfulfilled. Optionally, the third node and the first node may be a same node. In other words, the first node may send the second information to the second node when the third model does not meet the preset condition, and the second node may receive the second information from the first node. Alternatively, the third node and the first node may be different nodes. The third node may send the second information to the second node when a third model obtained by processing a second model of the third node based on the first model does not meet the preset condition, and the second node may receive the second information from the third node.
In some embodiments, the second node receives an output result from the at least one first node, where the output result is an output result of the third model. The second node sends indication information to a fourth node, where the indication information indicates the fourth node to send the second model. The fourth node is one or more nodes that are in the at least one first node and whose output results meet a second threshold. Optionally, the fourth node and the first node may be a same node. In other words, the first node may send the output result to the second node. Alternatively, the fourth node and the first node may be different nodes. In other words, the fourth node may send the output result to the second node.
Based on the foregoing solution, depending on the output result sent by the first node, the second node makes a decision and schedules the first node to send the second model. In comparison with a case in which each first node reports the second model, the air interface transmission overheads can be reduced, and a model obtained by the second node through training can be more accurate. This improves model training efficiency.
In some embodiments, the first model is preset based on at least one of the following: a geographical location of the first node and a task executed by the first node. Based on the foregoing solution, different first models are preset for first nodes in different geographical environments, so that the first model can be related to the geographical environment, and an output result obtained by using the first model can also be consistent with the geographical environment in which the first node is located. This may improve accuracy of the first model.
In some embodiments, the second node sends the first information to the at least one first node through a sidelink. Alternatively, the second node sends the first information to the at least one first node through a common channel. Alternatively, the second node separately sends the first information to the at least one first node through a dedicated channel of the at least one first node.
According to a third aspect, a model scheduling method is provided. The method may be performed by a first apparatus. Optionally, the first apparatus may be a communication device or a chip/chip system. In the method, the first apparatus receives I intermediate results from a second apparatus. I is an integer greater than or equal to 1. The I intermediate results are outputs of I first models. Each of the I first models may be obtained by fusing at least one third model. The first apparatus inputs an ith intermediate result into a second model, to obtain an ith first output result. I is an integer greater than or equal to 1, and i is less than or equal to I. The second model matches at least one of a task type of the first apparatus and location information of the first apparatus. The first apparatus determines a reference model in the I first models based on I first output results.
Based on the foregoing solution, the second apparatus may fuse the at least one third model to obtain the first model. The second apparatus may send the I intermediate results of the I first models to the first apparatus, and the first apparatus may determine the reference model based on the I intermediate results. In comparison with sending intermediate results of all third models to the first apparatus, in the foregoing solution, air interface transmission overheads can be reduced.
In some embodiments, the first apparatus sends first information to the second apparatus, where the first information indicates the reference model. The first apparatus receives second information from the second apparatus, where the second information indicates an intermediate result of at least one target third model. The reference model is obtained by fusing the at least one target third model. The first apparatus separately inputs the intermediate result of the at least one target third model into the second model, to obtain at least one second output result. The first apparatus determines a target model in the at least one target third model based on the at least one second output result. It may be understood that a quantity of target third models is less than a quantity of all the third models. For example, there are Q third models, and Q may be an integer greater than or equal to 2. In this case, the quantity of target third models is less than Q.
Based on the foregoing solution, when determining the reference model, the first apparatus may determine the at least one target third model that is fused to obtain the reference model, and may determine, based on the intermediate result of the at least one target third model from the second apparatus, a target model for executing a task. In comparison with a technical solution in which the second apparatus sends the intermediate results of all the third models to the first apparatus, the air interface transmission overheads can be reduced.
In some embodiments, the I intermediate results are carried in radio resource control (RRC) signaling. Alternatively, the I intermediate results are carried in a media access control (MAC) physical data unit (PDU). Alternatively, the I intermediate results are carried in downlink control information (DCI).
According to a fourth aspect, a model scheduling method is provided. The method may be performed by a second apparatus. Optionally, the second apparatus may be a communication device or a chip/chip system. In the method, the second apparatus groups a third model included in a third model set into I groups. Each third model in the third model set matches at least one of a task type of a first apparatus and location information of the first apparatus. I is an integer greater than or equal to 1. The second apparatus fuses a third model in an ith group, to obtain an ith first model. I is an integer greater than or equal to 1, and i is less than or equal to I. The second apparatus inputs a preset parameter into the ith first model, to obtain an ith intermediate result. The second apparatus sends I intermediate results to the first apparatus. The I intermediate results may be used to determine a reference model in I first models.
Based on the foregoing solution, the second apparatus may group the third model, and fuse a third model in a group to obtain a first model. The second apparatus may send the I intermediate results of the I first models to the first apparatus, to determine the reference model. In comparison with sending intermediate results of all third models to the first apparatus, in the foregoing solution, air interface transmission overheads can be reduced.
In some embodiments, the second apparatus receives first information from the first apparatus. The first information indicates the reference model, and the reference model is one of the I first models. The reference model may be obtained by fusing at least one target third model. It may be understood that the at least one target third model may be one or more third models in the third model set, and a quantity of the at least one target third model is less than a quantity of all third models included in the third model set. The second apparatus inputs the preset parameter into the at least one target third model, to obtain an intermediate result of the at least one target third model. The second apparatus sends second information to the first apparatus, where the second information includes the intermediate result of the at least one target third model. The intermediate result of the at least one target third model may be used to determine a target model.
Based on the foregoing solution, the second apparatus may send the intermediate result of the at least one target third model to the first apparatus, so that the first apparatus determines, based on the intermediate result of the at least one target third model, a target model for executing a task. In comparison with a technical solution in which the second apparatus sends the intermediate results of all the third models to the first apparatus, the air interface transmission overheads can be reduced.
In some embodiments, the I intermediate results are carried in RRC signaling. Alternatively, the I intermediate results are carried in a MAC PDU. Alternatively, the I intermediate results are carried in DCI.
According to a fifth aspect, a communication apparatus is provided, and includes a processing unit and a transceiver unit.
The transceiver unit is configured to receive first information from a second node, where the first information indicates a first model. The processing unit is configured to process a second model based on the first model, to obtain a third model. A structure of the third model is the same as a structure of the first model, and the second model and the first model are heterogeneous. The transceiver unit is further configured to send the second model to the second node. The third model meets a preset condition.
In some embodiments, the first information includes a model index and/or a parameter of the first model.
In some embodiments, the transceiver unit is further configured to receive information about a fourth model from the second node. The fourth model is obtained by using a second model of at least one first node. The processing unit is further configured to process the fourth model based on the second model of the first node, to obtain a fifth model. A structure of the fifth model is the same as a structure of the second model of the first node. The processing unit is further configured to train the fifth model based on a preset parameter, to obtain an updated second model.
In some embodiments, the processing unit is further configured to process the updated second model based on a sixth model, to obtain an updated third model. The transceiver unit is further configured to send the updated second model to the second node. The updated third model meets the preset condition. Optionally, the sixth model and the first model may be a same model, or may be different models. The sixth model is indicated by the second node. For example, the transceiver unit may receive indication information of the sixth model from the second node, and the indication information of the sixth model may include a model index and/or a parameter of the sixth model.
In some embodiments, the preset condition includes a first threshold, and the first threshold indicates performance of the third model or the first threshold indicates a performance variation of the third model.
In some embodiments, the preset condition is indicated by the second node. Optionally, the preset condition may include the following condition 1 and condition 2. Condition 1: The preset condition includes an evaluation parameter. The evaluation parameter may indicate the performance of the third model, or indicate the performance variation of the third model. Optionally, the evaluation parameter may include the first threshold. Condition 2: The preset condition includes indication information. The first node may receive the indication information from the second node. The indication information may indicate the first node to send the second model. Based on the foregoing solution, the second node may flexibly indicate the preset condition based on a training status of a model.
In some embodiments, the processing unit is further configured to input the preset parameter into the third model, to obtain an output result. The transceiver unit is further configured to send second information to the second node. The second information indicates that model training is unfulfilled, and the output result does not meet the first threshold.
In some embodiments, the processing unit is further configured to input a preset parameter into the third model, to obtain an output result. The transceiver unit is further configured to send the output result to the second node. The first node receives the indication information from the second node, where the indication information indicates the first node to send the second model.
In some embodiments, the first model is preset based on at least one of the following: a geographical location of the first node and a task executed by the first node.
In some embodiments, the transceiver unit is specifically configured to receive the first information from the second node through a sidelink. Alternatively, the transceiver unit is specifically configured to receive the first information from the second node through a common channel. Alternatively, the transceiver unit is specifically configured to receive the first information from the second node through a dedicated channel of the first node.
According to a sixth aspect, a communication apparatus is provided, and includes a processing unit and a transceiver unit.
The transceiver unit is configured to send first information to at least one first node, where the first information indicates a first model. The transceiver unit is further configured to receive a second model from the at least one first node, where the second model and the first model are heterogeneous. The first model is used to process the second model, to obtain a third model that has a same structure as the first model. The processing unit is configured to process the second model to obtain a fourth model. The second node sends the fourth model to the at least one first node.
In some embodiments, the first information includes a model index, and the model index indicates the first model. Alternatively, the first information includes a parameter of the first model. In a possible case, the first information may include the model index. The first model may be preset, and each first model may correspond to a model index. In another possible case, the first information may include the parameter of the first model. For example, the first information may include a parameter used to describe the first model.
In some embodiments, the transceiver unit is further configured to receive an updated second model from the at least one first node, where the updated second model is obtained based on the fourth model.
In some embodiments, the transceiver unit is further configured to send a first threshold to the at least one first node, where the first threshold indicates performance of the third model or the first threshold indicates a performance variation of the third model.
In some embodiments, the transceiver unit is further configured to receive second information from a third node, where the second information indicates that model training is unfulfilled, and the third node is one or more nodes in the at least one first node. Optionally, the third node and the first node may be a same node, or the third node and the first node may be different nodes.
In some embodiments, the transceiver unit is further configured to receive an output result from the at least one first node, where the output result is an output result of the third model. The transceiver unit is further configured to send indication information to a fourth node, where the indication information indicates the fourth node to send the second model. The fourth node is one or more nodes that are in the at least one first node and whose output results meet a second threshold. Optionally, the fourth node and the first node may be a same node, or the fourth node and the first node may be different nodes.
In some embodiments, the first model is preset based on at least one of the following: a geographical location of the first node and a task executed by the first node.
In some embodiments, the transceiver unit is specifically configured to send the first information to the at least one first node through a sidelink. Alternatively, the transceiver unit is specifically configured to send the first information to the at least one first node through a common channel. Alternatively, the transceiver unit is specifically configured to separately send the first information to the at least one first node through a dedicated channel of the at least one first node.
According to a seventh aspect, a communication apparatus is provided, and includes a processing unit and a transceiver unit.
The transceiver unit is configured to receive I intermediate results from a second apparatus. I is an integer greater than or equal to 1. The I intermediate results are outputs of I first models. Each of the I first models may be obtained by fusing at least one third model. The processing unit is configured to input an ith intermediate result into a second model, to obtain an ith first output result. I is an integer greater than or equal to 1, and i is less than or equal to I. The second model matches at least one of a task type of a first apparatus and location information of the first apparatus. The processing unit is further configured to determine a reference model in the I first models based on I first output results.
In some embodiments, the transceiver unit is further configured to send first information to the second apparatus, where the first information indicates the reference model. The reference model is obtained by fusing at least one target third model. The transceiver unit is further configured to receive second information from the second apparatus, where the second information includes an intermediate result of the at least one third model. The processing unit is further configured to separately input the intermediate result of the at least one target third model into the second model, to obtain at least one second output result. The processing unit is further configured to determine a target model in the at least one target third model based on the at least one second output result. It may be understood that a quantity of target third models is less than a quantity of all third models.
In some embodiments, the I intermediate results are carried in RRC signaling. Alternatively, the I intermediate results are carried in a MAC PDU. Alternatively, the I intermediate results are carried in DCI.
According to an eighth aspect, a communication apparatus is provided, and includes a processing unit and a transceiver unit.
The processing unit is configured to group a third model included in a third model set into I groups. Each third model in the third model set matches at least one of a task type of a first apparatus and location information of the first apparatus. I is an integer greater than or equal to 1. The processing unit is further configured to fuse a third model in an ith group, to obtain an ith first model. I is an integer greater than or equal to 1, and i is less than or equal to I. The processing unit is further configured to input a preset parameter into the ith first model, to obtain an ith intermediate result. The transceiver unit is configured to send I intermediate results to the first apparatus. The I intermediate results may be used to determine a reference model in I first models.
In some embodiments, the transceiver unit is further configured to receive first information from the first apparatus. The first information indicates the reference model, and the reference model is one of the I first models. The reference model may be obtained by fusing at least one target third model. It may be understood that the at least one target third model may be one or more third models in the third model set, and a quantity of the at least one target third model is less than a quantity of all third models included in the third model set. The processing unit is further configured to input the preset parameter into the at least one target third model, to obtain an intermediate result of the at least one target third model. The transceiver unit is further configured to send second information to the first apparatus, where the second information includes the intermediate result of the at least one target third model. The intermediate result of the at least one target third model may be used to determine a target model.
In some embodiments, the I intermediate results are carried in RRC signaling. Alternatively, the I intermediate results are carried in a MAC PDU. Alternatively, the I intermediate results are carried in DCI.
The communication apparatus according to any one of the fifth aspect to the eighth aspect may be a communication device or a chip/chip system in a communication device.
According to a ninth aspect, an embodiment of this disclosure provides a communication apparatus. The communication apparatus may be the communication apparatus according to any one of the fifth aspect to the eighth aspect, and may be a communication device or a chip/chip system in a communication device. The communication apparatus includes a communication interface and a processor, and optionally, further includes a memory. The memory is configured to store a computer program, instructions, or data. The processor is coupled to the memory and the communication interface. When the processor reads the computer program, the instructions, or the data, the communication apparatus is enabled to perform the method performed by the first node in any embodiment of the first aspect, the communication apparatus is enabled to perform the method performed by the second node in any embodiment of the second aspect, the communication apparatus is enabled to perform the method performed by the first apparatus in any embodiment of the third aspect, or the communication apparatus is enabled to perform the method performed by the second apparatus in any embodiment of the fourth aspect.
It should be understood that the communication interface may be implemented by using an antenna, a feeder, a codec, or the like in the communication apparatus. Alternatively, if the communication apparatus is a chip disposed in a communication device at a transmit end or a communication device at a receive end, the communication interface may be an input/output interface of the chip, for example, an input/output pin. The communication apparatus may further include a transceiver, configured to perform communication between the communication apparatus and another device. For example, when the communication apparatus is the first node, the other device is the second node, or when the communication apparatus is the second node, the other device is the first node. For example, when the communication apparatus is the first apparatus, the other device is the second apparatus; or when the communication apparatus is the second apparatus, the other device is the first apparatus.
According to a tenth aspect, this disclosure provides a communication apparatus. The communication apparatus may be the communication apparatus according to any one of the fifth aspect to the eighth aspect, and may be a communication device or a chip/chip system in a communication device. The communication apparatus includes a logic circuit and an input/output interface. The logic circuit may be configured to perform the method performed by the first node in any embodiment of the first aspect, configured to perform the method performed by the second node in any embodiment of the second aspect, configured to perform the method performed by the first apparatus in any embodiment of the third aspect, or configured to perform the method performed by the second apparatus in any embodiment of the fourth aspect. The input/output interface is configured to perform communication between the communication apparatus and another device. For example, when the communication apparatus is the first node, the other device is the second node, or when the communication apparatus is the second node, the other device is the first node. For example, when the communication apparatus is the first apparatus, the other device is the second apparatus; or when the communication apparatus is the second apparatus, the other device is the first apparatus.
According to an eleventh aspect, an embodiment of this disclosure provides a chip system. The chip system includes a processor, configured to implement the embodiments of any one of the first aspect to the fourth aspect. In some embodiments, the chip system further includes a memory, configured to store program instructions and/or data. The chip system may include a chip, or may include a chip and another discrete component.
According to a twelfth aspect, an embodiment of this disclosure provides a communication system. The communication system includes the communication apparatus according to any embodiment of the fifth aspect and the communication apparatus according to any embodiment of the sixth aspect.
According to a thirteenth aspect, an embodiment of this disclosure provides a communication system. The communication system includes the communication apparatus according to any embodiment of the seventh aspect and the communication apparatus according to any embodiment of the eighth aspect.
According to a fourteenth aspect, this disclosure provides a computer-readable storage medium. The computer-readable storage medium stores a computer program or instructions. When the computer program or the instructions are run, the method performed by the first node, the second node, the first apparatus, or the second apparatus in the foregoing aspects is performed.
According to a fifteenth aspect, a computer program product is provided. The computer program product includes computer program code or instructions. When the computer program code or the instructions are run, the method performed by the first node, the second node, the first apparatus, or the second apparatus in the foregoing aspects is performed.
For beneficial effects of the fifth aspect to the fifteenth aspect and implementations thereof, refer to the descriptions of beneficial effects of the methods according to the first aspect to the fourth aspect and implementations thereof.
The technical solutions provided in embodiments of this disclosure are described below with reference to the accompanying drawings.
The technical solutions provided in embodiments of this disclosure may be applied to a 4th generation (4G) communication system, for example, a Long-Term Evolution (LTE) communication system, may be applied to a 5th generation (5G) communication system, for example, a 5G new radio (NR) communication system, or may be applied to various communication systems in the future, for example, a 6th generation (6G) communication system. The method provided in embodiments of this disclosure may be further applied to a device-to-device (D2D) communication system, a vehicle-to-everything (V2X) communication system, a machine-to-machine (M2M) communication system, a machine type communication (MTC) system, an internet of things (IoT) communication system, a BLUETOOTH system, a WI-FI system, a long-range radio (LoRa) system, or an internet of vehicles system. The technical solutions provided in embodiments of this disclosure may be further applied to a satellite communication system, and the satellite communication system may be integrated with the foregoing communication system.
The terminal device in this disclosure includes a device that provides a voice and/or data signal connectivity for a user. Specifically, the terminal device includes a device that provides the voice for the user, includes a device that provides the data signal connectivity for the user, or includes the device that provides the voice and the data signal connectivity for the user. For example, the terminal device may include a hand-held device with a wireless connection function or a processing device connected to a wireless modem. The terminal device may include user equipment (UE), a wireless terminal device, a mobile terminal device, a D2D communication terminal device, V2X terminal device, an M2M/MTC terminal device, an IoT terminal device, a subscriber unit, a subscriber station, a mobile station, a remote station, an access point (AP), a remote terminal device, an access terminal device, a user terminal device, a user agent, a user device, a satellite, an uncrewed aerial vehicle, a balloon, an aircraft, or the like. For example, the terminal device may include a mobile phone (or referred to as a “cellular” phone), a computer having a mobile terminal device, or a portable, pocket-sized, hand-held, or a computer-embedded mobile apparatus. For example, the terminal device may be a device such as a personal communication service (PCS) phone, a cordless phone, a Session Initiation Protocol (SIP) phone, a wireless local loop (WLL) station, or a personal digital assistant (PDA). The terminal device further includes a limited device, for example, a device with low power consumption, a device with a limited storage capability, or a device with a limited computing capability. For example, the terminal device includes an information sensing device such as a barcode, radio frequency identification (RFID), a sensor, a Global Positioning System (GPS), or a laser scanner. By way of example and not limitation, in embodiments of this disclosure, the terminal device may alternatively be a wearable device. The wearable device may also be referred to as a wearable intelligent device, an intelligent wearable device, or the like, and is a general term of wearable devices that are intelligently designed and developed for daily wear by using a wearable technology. If the various terminal devices described above are located in a vehicle (for example, placed in the vehicle or installed in the vehicle), the terminal devices may be all considered as vehicle-mounted terminal devices. For example, the vehicle-mounted terminal devices are also referred to as on-board units (OBUs).
Terminal devices may be distributed across an entire wireless communication system, and may be stationary or mobile. The terminal device may include a mobile device, a mobile station, a mobile unit, a radio unit, a remote unit, a user agent, a mobile client, or the like.
The network device in this disclosure includes, for example, an access network (AN) device, for example, a base station (for example, an access point), and may be a device that is in an access network and that communicates with a wireless terminal device through an air interface in one or more cells. Alternatively, the network device is, for example, a road side unit (RSU) in a V2X technology. The network device may include an evolved NodeB (NodeB or eNB or e-NodeB, evolved NodeB) in an LTE or long term evolution-advanced (LTE-A) system, may include a next generation NodeB (gNB) in an evolved packet core (EPC), a 5th generation (5G) mobile communication technology, or a new radio (NR) system, or may include a central unit (CU) and a distributed unit (DU) in a cloud access network (Cloud RAN) system, a satellite, an uncrewed aerial vehicle, a balloon, an aircraft, or the like.
The network device may be configured to communicate with the terminal device under control of a network device controller (not shown in
The network device may include a network device transceiver station (BTS), a wireless transceiver, a basic service set (BSS), an extended service set (ESS), a NodeB, an eNodeB, a gNodeB, and the like.
To cope with the vision of inclusive intelligence in the future, intelligence is further evolved at a wireless network architecture layer, and AI is to be deeply integrated with a wireless network. A concept of FL is proposed to effectively resolve difficulties in current development of the artificial intelligence. While user data privacy and security are fully ensured, the FL facilitates edge devices and a central-end server to collaborate to efficiently complete learning tasks of a model.
It should be noted that in this disclosure, “indication” may include a direct indication, an indirect indication, an explicit indication, and an implicit indication. When a piece of indication information is described as indicating A, it may be understood as that the indication information carries A, directly indicates A, or indirectly indicates A.
In this disclosure, information indicated by the indication information is referred to as to-be-indicated information. In a specific implementation process, the to-be-indicated information is indicated in a plurality of manners, for example, but not limited to, the following manners: The to-be-indicated information is directly indicated, for example, the to-be-indicated information or an index of the to-be-indicated information is indicated. Alternatively, the to-be-indicated information may be indirectly indicated by indicating other information, and there is an association relationship between the other information and the to-be-indicated information. Alternatively, only a part of the to-be-indicated information may be indicated, and the other part of the to-be-indicated information is known or pre-agreed on. For example, specific information may alternatively be indicated by using an arrangement sequence that is of a plurality of pieces of information and that is pre-agreed on (for example, specified in a protocol), to reduce indication overheads to some extent.
The to-be-indicated information may be sent as a whole, or may be divided into a plurality of pieces of sub-information for separate sending. In addition, sending periodicities and/or sending occasions of these pieces of sub-information may be the same or may be different. A specific sending method is not limited in this disclosure. The sending periodicities and/or the sending occasions of these pieces of sub-information may be predefined, for example, predefined according to a protocol, or may be configured by a transmitter device by sending configuration information to a receiver device. The configuration information may include, for example, but not limited to, one or a combination of at least two of radio resource control signaling, MAC layer signaling, and physical layer signaling. The radio resource control signaling includes, for example, RRC signaling. The MAC layer signaling includes, for example, a MAC control element (CE). The physical layer signaling includes, for example, DCI.
Then, the central node sends a global model wgt of a latest version to all distributed nodes by broadcast, to perform a new round of training.
In addition to the local model wkt, the distributed node may further report a local gradient gkt for training. The central node averages local gradients, and updates the global model based on a direction of an average gradient.
It may be learned that, in an FL framework, a dataset exists on the distributed node. To be specific, the distributed node collects a local dataset, performs local training, and reports, to the central node, a local result obtained through training. The central node does not have a dataset, is only responsible for fusing training results of the distributed nodes to obtain a global model, and delivers the global model to the distributed nodes.
However, in a federated learning system, a premise that a plurality of local models can be fused in a weighted averaging manner is that the local models have a same structure and a same quantity of parameters. However, in a network, the distributed nodes have different device capabilities, but using a same local model for the distributed nodes typically limits flexibility of the distributed nodes, and different requirements of different distributed nodes cannot be met.
To meet the requirements of the different distributed nodes, local models obtained through training by the distributed nodes may be heterogeneous. The central node may fuse these heterogeneous local models. A typical method for heterogeneous model fusion is to convert models of different structures into a same structure in a knowledge distillation manner, and then perform fusion. Through knowledge distillation, knowledge can be transferred from one network to another, and the two networks may be homogeneous or heterogeneous. The knowledge distillation may be used to convert a large network into a small network and retain performance close to that of the large network. In the knowledge distillation, knowledge learned from a plurality of networks may alternatively be transferred to one network. A specific method of heterogeneous fusion is as follows: The central node first collects heterogeneous models of the distributed nodes, distills the heterogeneous models into homogeneous models, and fuses the homogeneous models. Then, fused homogeneous models are fed back to the distributed nodes. The distributed nodes restore the fused homogeneous models to network structures of the distributed nodes, and start a next round of fusion.
However, in each round of heterogeneous model fusion, each distributed node needs to report a local model. This results in high air interface transmission overheads.
In view of this, an embodiment of this disclosure provides a model scheduling method. In the method, each distributed node may select, through training of a first model, a distributed node that sends a second model to the central node, so that frequency of sending an assisted model by each distributed node can be reduced, the air interface transmission overheads can be reduced, and air interface utilization can be improved.
To facilitate understanding of the technical solutions provided in embodiments of this disclosure, the following models are described.
It should be noted that the first model may be a convolutional neural network model, a deep learning network model, a recursive neural network model, or the like. A type of the first model is not specifically limited in this disclosure.
Optionally, the second model may be related to a task of the distributed node. For example, when the task of the distributed node includes a channel assessment task, an input of the second model may be a moving speed or location information, or the like, and an output of the second model may be a channel parameter, for example, a quantity of spatial flows, coherence bandwidth, or a multipath delay.
It may be understood that a task that can be executed by the first model is the same as a task that can be executed by the second model.
It should be noted that the second model may be a convolutional neural network model, a deep learning network model, a recursive neural network model, or the like. A type of the second model is not specifically limited in this disclosure.
It may be understood that a task that can be executed by the third model is the same as the task that can be executed by the second model.
It should be noted that the third model may be a convolutional neural network model, a deep learning network model, a recursive neural network model, or the like. A type of the third model is not specifically limited in this disclosure.
It may be understood that a task that can be executed by the fourth model is the same as the task that can be executed by the second model.
It should be noted that the fourth model may be a convolutional neural network model, a deep learning network model, a recursive neural network model, or the like. A type of the fourth model is not specifically limited in this disclosure.
S301: The second node sends first information to the first node.
Correspondingly, the first node receives the first information from the second node. Optionally, the second node may send the first information to at least one first node. It may be understood that first information of all first nodes may be the same or different.
In some embodiments, the first information may indicate a first model. For example, the first information may include a model index and/or a parameter of the first model.
In a possible case, the first information may include the model index. The first model may be preset, and each first model may correspond to a model index. For example, a correspondence between the first model and the model index may be preset. In this way, the first model may be determined by using the model index included in the first information.
For example, a plurality of first models may be preset. For example, a structure of each first model may be preset. For example, a type of each first model, a quantity of layers of each first model, a quantity of neurons included in each layer, a connection manner between neurons, and a weight of a connection between neurons may be set. In addition, a corresponding model index is set for each first model. The first node may determine the first model by using the model index included in the first information, for example, may determine a structure of the first model. Optionally, the first information may further include the parameter of the first model, and the parameter of the first model may be used to describe the structure of the first model. For example, the parameter of the first model may include the type of each first model, the quantity of layers of each first model, the quantity of neurons included in each layer, the connection manner between the neurons, the weight of the connection between the neurons, and the like.
Optionally, the first model may be preset based on a geographical location of the first node. For example, the first model may vary with the geographical location of the first node. For example, the geographical location may include a region formed by a geographical environment like a mountainous region or a city. Alternatively, the geographical location may include a region formed by space in which the first node is located, like a room, a basement, or an outdoor region. In this way, different first models are preset for first nodes in different geographical environments, so that the first model can be related to the geographical environment, and an output result obtained by using the first model can also be consistent with the geographical environment in which the first node is located. This may improve accuracy of the first model.
In another possible case, the first information may include the parameter of the first model. For example, the first information may include a parameter used to describe the first model, for example, a type of the first model, a quantity of layers of the first model, a quantity of neurons included in each layer, a connection manner between neurons, and a weight of a connection between neurons. The first node may determine the first model by using the parameter that is of the first model and that is included in the first information, for example, may determine the structure of the first model.
Based on the foregoing solution, the first model is indicated by using the model index, so that air interface transmission overheads can be greatly reduced. In addition, because the first model is a lightweight model and has a small quantity of parameters, the first model is indicated by using the parameter of the first model. In comparison with some approaches, overheads needed for performing transmission of a complete model are small, and air interface utilization can be improved.
It may be understood that the first model may be a model related to a task of the first node. For example, the first node may send task information to the second node, and the second node may determine the task of the first node based on the task information, and determine the first model.
In an example, the second node may send the first information through a downlink common channel. The first node may receive the first information through the common channel. The common channel may be a channel shared by a plurality of first nodes. Optionally, the second node may scramble the first information by using a radio network temporary identity (RNTI). In this way, the first node may attempt to descramble the first information by using an RNTI of the first node. When the first node can descramble the first information by using the RNTI of the first node, the first node may determine the first model based on the first information. Alternatively, the second node may include an identifier of the first node in the first information. In this way, the first node may determine the first information based on the identifier of the first node, and determine the first model based on the first information.
In another example, the second node may send the first information to the first node through a dedicated channel of the first node on a downlink. The dedicated channel is dedicated to the first node. In other words, another node like a third node cannot perform transmission of data with the second node through the dedicated channel of the first node.
In still another example, the second node may send the first information to the first node through a sidelink. For example, the second node is a distributed node, and the first node is also a distributed node. The second node may send the first information to the first node through a sidelink between the second node and the first node.
S302: The first node processes a second model based on the first model, to obtain a third model.
Optionally, the at least one first node may process the second model based on the first model, to obtain the third model.
It may be understood that the first model and the second model are heterogeneous, and a structure of the third model is the same as the structure of the first model. The first node may process the second model to obtain the third model that is homogeneous with the first model. For example, the first node may perform knowledge distillation on the second model based on the first model, to obtain the third model. In other words, the first node may distill knowledge of the second model to the first model. For another example, the first model may be a model of a same type as the second model, the parameter of the first model may include a first indication, and the first indication may indicate a specific connection on the second model that can be deleted. For example, the first indication is 1-bit information. On a connection or a parameter of the first model, if a value of the first indication is 0, it may indicate that the connection or the parameter in a corresponding second model can be deleted. Optionally, on a connection or a parameter of the first model, if a value of the first indication is 1, it may indicate that the connection or the parameter in a corresponding second model cannot be deleted. The second model may be cut into a third model with a smaller quantity of parameters by using the parameter of the first model.
S303: The first node sends the second model to the second node.
Optionally, the at least one first node may send a model to the second node. A third model of the at least one first node meets a preset condition.
For example, the first node may send the second model to the second node when the third model meets the preset condition. Optionally, the preset condition may include the following condition 1 and condition 2.
The evaluation parameter may indicate performance of the third model, or indicate a performance variation of the third model. Optionally, the evaluation parameter may include a first threshold. In other words, the first node may send the second model to the second node when the performance of the third model meets the first threshold. Alternatively, the first node may send the second model to the second node when the performance variation of the third model meets the first threshold. It may be understood that the evaluation parameter, for example, the first threshold, may be preset or predefined in a protocol.
In an example, the performance of the third model may be understood as accuracy of an output result of the third model. For example, the first node may input a preset parameter into the third model, to obtain the output result. The preset parameter may be related to the task of the first node. The preset parameter may be a parameter predefined in a protocol. When the accuracy of the output result meets the evaluation parameter, for example, the accuracy is not less than the first threshold, it may be considered that the third model meets the preset condition, and the first node may send the second model to the second node.
A channel assessment task is used as an example for description. In the channel assessment task, an input of the third model may be a parameter predefined in a protocol, for example, predefined location information or a predefined user moving speed. An output of the third model is a channel parameter, including a quantity of spatial flows, coherence bandwidth, a multipath delay, and the like. In the foregoing manner, the output of the third model may be compared with a predefined output, to determine accuracy of the output of the third model, so that the performance of the third model can be determined. Because the third model is obtained by processing the second model, the performance of the third model may reflect performance of the second model. In this way, the performance of the second model may be determined by using a small quantity of parameters and a small calculation amount.
For example, in an image recognition task, the preset parameter may include a group of marked pictures, and the output of the third model may be a picture recognition result. Accuracy of the third model may be obtained by comparing the output of the third model with a mark. For another example, in a channel feedback result compression task, the preset parameter may be a group of channel parameters, and may include channel parameters such as a delay, Doppler, and a quantity of spatial flows. The preset parameter is input into the third model, and then passes through a decompression model. A restored preset parameter is compared with the input preset parameter, to obtain compression accuracy. For another example, in the foregoing channel feedback result compression task, the preset parameter is input into a compression model, and is then decompressed by using the third model. A restored preset parameter is compared with the input preset parameter, to obtain decompression accuracy. It may be understood that, in this embodiment of this disclosure, the second node may perform iterative training on a model. In other words, the first node may send the second model to the second node for a plurality of times. The iterative training is described subsequently. In another example, the performance variation of the third model may be understood as a difference between accuracy of a current output result of the third model and accuracy of a previous output result of the third model. When the difference meets the evaluation parameter, for example, the difference is not less than the first threshold, it may be considered that the third model meets the condition, and the first node may send the second model to the second node.
Optionally, the first node may send second information to the second node. The second information may indicate that model training is unfulfilled, for example, may indicate that training of the third model is unfulfilled. For example, when the output result of the third model does not meet the first threshold, the first node may send the second information to the second node. For example, the first node may send the second information to the second node when the accuracy of the current output result of the third model does not meet the evaluation parameter, for example, the accuracy is less than the first threshold. For another example, the first node may send the second information to the second node when the difference between the accuracy of the current output result of the third model and the accuracy of the previous output result of the third model does not meet the evaluation parameter, for example, the difference is less than the first threshold.
In a possible case, the second information may be a negative acknowledgement (NACK). In another possible case, the second information may be a bit sequence. The bit sequence may be 1 bit, 2 bits, 3 bits, or the like. For example, the second information is 1 bit. When a value of the second information is “0”, it may indicate that the model training is unfulfilled, or when a value of the second information is “1”, it may be considered that the model training is unfulfilled.
Optionally, the condition 1 may be indicated by the second node, may be predefined in a protocol, or may be preset.
Based on the foregoing solution, the third model is evaluated by using the evaluation parameter, in other words, a training status of a model of the first node is evaluated. When the performance or the performance variation of the third model meets the evaluation parameter, the second model is sent to the second node, so that air interface transmission overheads can be reduced. In addition, because the structure of the third model is the same as the structure of the first model, a quantity of parameters that needs to be input into the third model is small and the structure is simple. A calculation process of obtaining the output result based on the third model is simpler than a calculation process of obtaining an output result based on the second model, and requires a shorter time.
The first node may receive the indication information from the second node. The indication information may indicate the first node to send the second model. In other words, the first node may send the second model to the second node when receiving the indication information of the second node.
In an example, the first node may input a preset parameter into the third model, to obtain an output result. The first node may send third information to the second node. The third information may include the output result, or the third information may include quantization of the output result.
It may be understood that the quantization of the output result may be understood as that the first node processes the output result to obtain the quantization of the output result. For example, the first node may determine a level of the output result, and use the level as the quantization of the output result. The level of the output result may be predefined or preset. For example, the level of the output result may be classified into a level 1, a level 2, and a level 3, and each level corresponds to a range of one output result. The first node may determine, based on the output result, a specific range to which the output result belongs, and determine the level corresponding to the output result. The second node may determine, based on the third information, whether the first node needs to be scheduled to send the second model. For example, the second node may send the indication information to the first node when the output result indicated by the third information or the quantization of the output result meets a second threshold. It should be noted that the second threshold may be preset or predefined in a protocol.
Based on the foregoing solution, depending on the third information sent by the first node, the second node makes a decision and schedules the first node to send the second model. In comparison with a case in which each first node reports the second model, the air interface transmission overheads can be reduced, and a model obtained by the second node through training can be more accurate. This improves model training efficiency.
S304: The second node processes the second model to obtain a fourth model.
In a possible case, the second node may fuse the second model to obtain a fourth model. Optionally, the second node may receive a second model from the at least one first node. In a possible case, capabilities of the first nodes are different. Therefore, structures of second models of the first nodes are different. The second node may separately distill knowledge of the second models of the first nodes to models of a same structure, and fuse the models of the same structure to obtain the fourth model. For example, the second node may determine a weight of the models of the same structure, and perform weighted averaging on the models of the same structure to obtain the fourth model.
It should be noted that the foregoing model fusion manner is merely shown as an example, and does not constitute a limitation on a model fusion method in embodiments of this disclosure. In this embodiment of this disclosure, the second models may alternatively be fused in another model fusion manner.
In another possible case, the second node may select one second model from second models as the fourth model, and send the fourth model to the first node. It may be understood that the second node may select, from the second models, a second model with best performance or a best performance variation amount as the fourth model.
S305: The second node sends the fourth model to the first node.
For example, the second node may send information about the fourth model to the first node, where the information about the fourth model may indicate the fourth model. For example, the information about the fourth model may include a parameter of the fourth model. Optionally, the second node may send the fourth model to the at least one first node.
In some embodiments, the first node may execute a corresponding task based on the fourth model. For example, the first node may input the preset parameter into the fourth model, to obtain an output result.
In another possible implementation, the first node and the second node may perform iterative training on the fourth model. For example, the first node may process the fourth model based on the second model of the first node, to obtain a fifth model. A structure of the fifth model may be the same as a structure of the second model of the first node. A structure of the fourth model may be the same as or different from the structure of the second model. For example, the first node may distill knowledge of the fourth model to the second model of the first node, to obtain the fifth model. For another example, the first node may cut the second model based on the parameter of the fourth model, to obtain the fifth model. It may be understood that, for a manner in which the first node processes the fourth model based on the second model of the first node, refer to the foregoing manner in which the first node processes the second model of the first node based on the first model. Details are not described herein again.
The first node may train the fifth model based on the preset parameter, to obtain an updated second model. The first node may process the updated second model based on a sixth model. For example, the first node may distill knowledge of the updated second model to the sixth model, to obtain an updated third model. For another example, the first node may cut the fifth model into an updated third model with a smaller quantity of parameters based on a parameter of the sixth model. It may be understood that, for a manner of processing the updated second model based on the sixth model, refer to the manner of processing the second model based on the first model. Details are not described herein again.
Optionally, the sixth model and the first model may be a same model, or may be different models. The sixth model may be indicated by the second node. For example, after S305, the second node may send information about the sixth model to the first node. In other words, in an iterative training process, first models used in all training processes may be the same or may be different.
In a possible case, when the first models used in all the training processes are the same, after S305, the second node may send the information about the sixth model (the same as information about the first model) to the first node or may not send the information about the sixth model. When the first node does not receive the information about the sixth model, the first node may process the fifth model by using the first model, to obtain the updated third model.
In another possible case, when the first models used in all the training processes are different, after S305, the second node may send the information about the sixth model (different from information about the first model) to the first node, and the first node may process the fifth model by using the sixth model, to obtain the updated third model.
When obtaining the updated third model, the first node may return to perform S302, in other words, the first node may send the updated second model to the second node. For example, the first node may send the updated second model to the second node when the updated third model meets the preset condition. In this case, the second node may process the updated second model to obtain an updated fourth model. It may be understood that, for a manner in which the second node obtains the updated fourth model, refer to the foregoing manner in which the second node obtains the fourth model. Details are not described herein again.
Based on the foregoing solution, the first node may select, in each training process based on a requirement, whether to update the first model, so that a fourth model obtained through training can better meet a requirement of the second node. In addition, the first node may evaluate the performance of the second model based on the first model, and determine, based on an evaluation result, whether to send the second model to the second node. Therefore, a second model with poor performance may not be sent to the second node, so that accuracy of the fourth model obtained through the training can be improved.
In some embodiments, the first node may stop sending the second model to the second node in at least one of the following case 1 and case 2.
The specified quantity of times may be set based on an empirical value, may be predefined in a protocol, may be preconfigured, or may be indicated by the second node.
For example, in an iterative training process, when the quantity of times that the first node sends the second model to the second node reaches 10 times, the first node may not send the second model to the second node in a subsequent training process. Optionally, in an 11th training process, when the first node receives the indication information of the second node, the first node may not send the second model to the second node either. Optionally, the first node may select a fourth model received from the second node in a 10th training process, to execute a corresponding task.
The specified threshold may be set based on an empirical value, may be predefined in a protocol, may be preconfigured, or may be indicated by the second node.
For example, in an iterative training process, when the traffic occupied by the first node for sending the second model to the second node reaches 1G, the first node may not send the second model to the second node in a subsequent training process. Optionally, in the subsequent training process, when the first node receives the indication information of the second node, the first node may not send the second model to the second node either. Optionally, the first node may select a fourth model received from the second node in a latest training process, to execute a corresponding task.
Based on the foregoing solution, when at least one of the foregoing case 1 and case 2 is met, the first node may stop sending the second model to the second node, so that the air interface transmission overheads can be reduced.
The following describes, with reference to
S401: The node N sends first information to the node K1.
Correspondingly, the node K1 receives the first information from the node N.
S401 may be a specific implementation of S301, and may be implemented with reference to S301.
Optionally, the node N may send first information to the node K2 and the node K3. It may be understood that, for a manner in which the node N sends the first information to the node K2 and the node K3, refer to an implementation in which the node N sends the first information to the node K1.
S402: The node K1 processes a second model of the node K1 based on a first model, to obtain a third model of the node K1.
S402 may be a specific implementation of S302, and may be implemented with reference to S302. Optionally, the node K2 processes a second model of the node K2 based on a first model, to obtain a third model of the node K2. Optionally, the node K3 processes a second model of the node K3 based on a first model, to obtain a third model of the node K3.
It may be understood that a manner in which a node Km processes a second model of the node Km based on the first model may be implemented in the manner in which the first node processes the second model based on the first model in the embodiment shown in
It may be understood that first models of all nodes are the same. For example, first models of the node K1, the node K2, and the node K3 are the same.
S403: The node K1 inputs a preset parameter into the third model of the node K1, to obtain a first output result.
The preset parameter may be related to a task of the node K1. For details, refer to related descriptions of the preset parameter in the embodiment shown in
Optionally, the node K2 inputs a preset parameter into the third model of the node K2, to obtain a second output result. Optionally, the node K3 inputs a preset parameter into the third model of the node K3, to obtain a third output result.
In a possible case, the embodiment shown in
S404A: The node K1 determines, based on the first output result, whether the third model of the node K1 meets an evaluation parameter.
In the embodiment shown in
For example, when the node K1 determines that accuracy of the first output result is not less than a first threshold, or a difference between accuracy of the first output result and accuracy of a first output result in previous training is not less than a first threshold, S407A may be performed. Similarly, when the node K2 determines that accuracy of the second output result is not less than the first threshold, or a difference between accuracy of the second output result and accuracy of a second output result in previous training is not less than the first threshold, S407A may be performed.
Because the third model of the node K3 does not meet the evaluation parameter, the node K3 may continue to train the second model of the node K3. For example, when the node K3 determines that accuracy of the third output result is less than the first threshold, or a difference between accuracy of the third output result and accuracy of a third output result in previous training is less than the first threshold, the node K3 may perform S407B.
In another possible case, the embodiment shown in
S404B: The node K1 obtains third information of the node K1 based on the first output result.
For example, the third information of the node K1 may include the first output result or quantization of the first output result. For the third information, refer to related descriptions of the third information in the embodiment shown in
Optionally, the node K2 may obtain third information of the node K2 based on the second output result. Optionally, the node K3 may obtain third information of the node K3 based on the third output result.
S405B: The node K1 sends the third information of the node K1 to the node N.
Optionally, the node K2 sends the third information of the node K2 to the node N. Optionally, the node K3 may send the third information of the node K3 to the node N.
S406B: The node N sends indication information to the node K1.
The indication information may indicate the node Km to send the second model of the node Km. The node N may determine, based on the third information of the node K1, whether performance of the third model of the node K1 meets a requirement. For example, the node N may determine whether the first output result or the quantization of the first output result is not less than a second threshold. When the first output result or the quantization of the first output result is not less than the second threshold, it may be considered that the performance of the third model of the node K1 meets the requirement.
Optionally, the node N may determine, based on the third information of the node K2, whether performance of the third model of the node K2 meets the requirement. Optionally, the node N may determine, based on the third information of the node K3, whether performance of the third model of the node K3 meets the requirement.
In this embodiment of this disclosure, an example in which the performance of the third model of the node K1 and the performance of the third model of the node K2 meet the requirement is used for description. The node N may send the indication information to the node K1 and the node K2.
S407A: The node K1 sends the second model of the node K1 to the node N.
Optionally, the node K2 may send the second model of the node K2 to the node N.
S407B: Train the second model.
For example, the node K3 may continue to train the second model of the node K3 based on a predefined parameter.
S408: The node N processes the second model of the node K1 and the second model of the node K2, to obtain a fourth model.
S408 may be a specific implementation of S304. For implementation, refer to S304. Details are not described herein again.
For example, the node N may distill knowledge of the second model of the node K1 to a model A. Similarly, the node N may distill knowledge of the second model of the node K2 to a model A. The node N may fuse the two models A, for example, may perform weighted averaging on the two models A, to obtain the fourth model.
For another example, the node N may select one of the second model of the node K1 and the second model of the node K2 as the fourth model.
S409: The node N sends the fourth model to the node K1.
S409 may be a specific implementation of S305. For implementation, refer to S305. Details are not described herein again.
Optionally, the node N may send the fourth model to the node K2. Optionally, the node N may send the fourth model to the node K3.
Based on S401 to S409, each node completes 1st training. In this embodiment of this disclosure, two times of training are used as an example for description. The embodiment shown in
S410: The node K1 processes the fourth model based on a sixth model, to obtain an updated third model of the node K1.
In a possible case, the sixth model may be indicated by the node N. It may be understood that, for an indication manner of the sixth model, refer to S401 for implementation. For example, before S410, the node N may send information about the sixth model to the node K1. In another possible case, the sixth model and the first model may be a same model. In other words, in S410, the node K1 may process the fourth model based on the first model, to obtain the updated third model of the node K1.
Optionally, the node K2 processes the fourth model based on a sixth model, to obtain an updated third model of the node K2. Optionally, the node K3 processes the fourth model based on a sixth model, to obtain an updated third model of the node K3.
It may be understood that sixth models of all the nodes are the same. For example, sixth models of the node K1, the node K2, and the node K3 are the same.
In a 2nd training process, after the node K1, the node K2, and the node K3 each obtain the updated third model, each node may return to perform S403 to S409.
Based on the embodiment shown in
An embodiment of this disclosure further provides another model scheduling method.
It should be noted that information mentioned below and the information mentioned above are different information. Different information is not necessarily understood as that content of the information is different, but is understood as that the information is reordered. Similarly, a model mentioned below and the model mentioned above are different models. Different models are not necessarily understood as that parameters and structures of the models are different, but are understood as that the models are reordered.
S501: The second apparatus groups a third model included in a third model set into I groups.
I is an integer greater than or equal to 1. The third model included in the third model set may match a parameter like a task type of the first apparatus and/or location information of the first apparatus. Optionally, the second apparatus may receive the parameter like the task type of the first apparatus and/or the location information of the first apparatus from the first apparatus.
For example, the third model set may include eight models: a third model 1 to a third model 8. The second apparatus may divide the eight models into two groups. A first group includes the third model 1 to the third model 4, and a second group includes the third model 5 to the third model 8.
It may be understood that the second apparatus may randomly group the third model included in the third model set into the I groups.
S502: The second apparatus fuses a third model in an ith group, to obtain an ith first model.
I is an integer greater than or equal to 1 and less than or equal to I. Optionally, i is an integer from 1 to I. For example, the second apparatus may perform weighted averaging on the third model in the ith group, to obtain the ith first model. It may be understood that the foregoing weighted averaging is merely shown as an example of fusing a model, and does not constitute a limitation on a manner in which the second apparatus fuses the third model in this embodiment of this disclosure.
It may be understood that structures of third models in the ith group may be the same, or at least one third model in the third models in the ith group is heterogeneous. When the at least one third model is heterogeneous, the second apparatus may separately distill knowledge of each third model to a model, and fuse each model.
For example, the second apparatus may fuse the third model 1 to the third model 4, to obtain a first model 1. The second apparatus may fuse the third model 5 to the third model 8, to obtain a first model 2.
S503: The second apparatus inputs a preset parameter into the ith first model, to obtain an ith intermediate result.
The preset parameter may match the parameter like the task type of the first apparatus and/or the location information of the first apparatus. The preset parameter may be predefined in a protocol. For details, refer to related descriptions of the preset parameter in the embodiment shown in
For example, the second apparatus may input the preset parameter into the first model 1, to obtain an intermediate result 1. The second apparatus may input the preset parameter into the first model 2, to obtain an intermediate result 2.
S504: The second apparatus sends I intermediate results to the first apparatus.
Correspondingly, the first apparatus may receive the I intermediate results from the second apparatus.
For example, the second apparatus may send the intermediate result 1 to the first apparatus, and send the intermediate result 2 to the first apparatus.
In some embodiments, the I intermediate results may be sent to the first apparatus by using an RRC message, a MAC PDU, or DCI.
S505: The first apparatus inputs the ith intermediate result into a second model, to obtain an ith first output result.
The second model may be a model trained by the first apparatus. The second model matches the parameter like the task type of the first apparatus and/or the location information of the first apparatus.
For example, the first apparatus may input the intermediate result 1 into the second model, to obtain a first output result 1. The first apparatus may input the intermediate result 2 into the second model, to obtain a first output result 2.
S506: The first apparatus determines a reference model in I first models based on I first output results.
The first apparatus may select a first output result with highest accuracy from the I first output results. A first model corresponding to the first output result is used as the reference model. It may be understood that the first output result with the highest accuracy may be a first output result with a largest value. It may be understood that for a manner of determining accuracy of a model, refer to related descriptions in the embodiment shown in
For example, the first apparatus may select a first output result with highest accuracy from the first output result 1 and the first output result 2. It is assumed that accuracy of the first output result 1 is greater than accuracy of the first output result 2, that is, the first apparatus may determine that the first model 1 is the reference model.
Based on the foregoing solution, the second apparatus may group a plurality of third models, and fuse a third model in a group to obtain a first model. The second apparatus may send the I intermediate results of the I first models to the first apparatus, to determine the reference model. In comparison with sending intermediate results of the plurality of third models to the first apparatus, in the foregoing solution, air interface transmission overheads can be reduced.
Optionally, the embodiment shown in
S507: The first apparatus sends first information to the second apparatus.
Correspondingly, the second apparatus receives the first information from the first apparatus.
The first information indicates the reference model. For example, the first information may indicate information about a group of the reference model. For example, the first information may indicate the first model 1. The reference model may be obtained by fusing at least one target third model. In this specification, the reference model may be the first model 1, and the at least one target third model may include the third model 1, the third model 2, the third model 3, and the third model 4.
S508: The second apparatus inputs the preset parameter into the at least one target third model, to obtain an intermediate result of the at least one target third model.
For example, the second apparatus may separately input the preset parameter into the third model 1 to the third model 4, to obtain an intermediate result of the third model 1, an intermediate result of the third model 2, an intermediate result of the third model 3, and an intermediate result of the third model 4.
S509: The second apparatus sends the intermediate result of the at least one target third model to the first apparatus.
Correspondingly, the first apparatus may receive the intermediate result of the at least one target third model from the second apparatus.
For example, the second apparatus may send the intermediate result of the third model 1, the intermediate result of the third model 2, the intermediate result of the third model 3, and the intermediate result of the third model 4 to the first apparatus.
Optionally, the second apparatus may send the intermediate result of the at least one third model to the first apparatus by using an RRC message, a MAC PDU, or DCI.
S510: The first apparatus inputs the intermediate result of the at least one target third model into the second model, to obtain at least one second output result.
For example, the first apparatus may input the intermediate result of the third model 1 into the second model, to obtain a second output result 1; input the intermediate result of the third model 2 into the second model, to obtain a second output result 2; input the intermediate result of the third model 3 into the second model, to obtain a second output result 3; and input the intermediate result of the third model 4 into the second model, to obtain a second output result 4.
S511: The first apparatus determines a target model in the at least one target third model based on the at least one second output result.
For example, the first apparatus may determine a second output result with highest accuracy from the at least one second output result, and use, as the target model, a target third model corresponding to the second output result with the highest accuracy.
For example, the first apparatus may determine a second output result with highest accuracy from the second output result 1, the second output result 2, the second output result 3, and the second output result 4. Assuming that accuracy of the second output result 2 is the highest, the first apparatus may use the third model 2 as the target model.
Optionally, the target model may be an initial status model used for training, or may be used only as an initial model structure used for training. The first apparatus may train the target model based on the preset parameter, to obtain a model for executing a corresponding task.
In some embodiments, when the second apparatus receives the first information, the second apparatus may use, as a new third model set, a group of third models corresponding to the reference model indicated by the first information, and return to perform S501, until a quantity of third models included in the group corresponding to the reference model is less than a preset threshold. It may be understood that the preset threshold may be set based on an empirical value, for example, may be 2, 3, 4, or 5.
In a possible case, the second apparatus may not group the third model in the third model set, but send an intermediate result of the third model to the first apparatus. The first apparatus may input the intermediate result of the third model into the second model, to obtain a plurality of output results. The first apparatus may determine the target model based on the plurality of output results.
However, in the foregoing solution, the second apparatus needs to send a large quantity of intermediate results. Consequently, the air interface transmission overheads are high. Based on the solution shown in
The following describes a communication apparatus for implementing the foregoing method in embodiments of this disclosure with reference to the accompanying drawings. Therefore, all the foregoing content may be used in the following embodiments. Repeated content is not described again.
In some possible implementations, the communication apparatus 600 can correspondingly implement behavior and functions of the first node in the foregoing method embodiments. For example, the communication apparatus 600 may be the first node, or may be a component (for example, a chip or a circuit) used in the first node. The transceiver unit 620 may be configured to perform all receiving or sending operations performed by the first node in the embodiment shown in
The transceiver unit 620 is configured to receive first information from a second node, where the first information indicates a first model. The processing unit 610 is configured to process a second model based on the first model, to obtain a third model. A structure of the third model is the same as a structure of the first model, and the second model and the first model are heterogeneous. The transceiver unit 620 is further configured to send the second model to the second node. The third model meets a preset condition.
In some embodiments, the transceiver unit 620 is further configured to receive information about a fourth model from the second node. The fourth model is obtained by using a second model of at least one first node. The processing unit 610 is further configured to process the fourth model based on the second model of the first node, to obtain a fifth model. A structure of the fifth model is the same as a structure of the second model of the first node. The processing unit 610 is further configured to train the fifth model based on a preset parameter, to obtain an updated second model.
In some embodiments, the processing unit 610 is further configured to process the updated second model based on a sixth model, to obtain an updated third model. The sixth model is indicated by the second node. The transceiver unit 620 is further configured to send the updated second model to the second node. The updated third model meets the preset condition.
In some embodiments, the processing unit 610 is further configured to input the preset parameter into the third model, to obtain an output result. The transceiver unit 620 is further configured to send second information to the second node. The second information indicates that model training is unfulfilled, and the output result does not meet a first threshold.
In some embodiments, the processing unit 610 is further configured to input a preset parameter into the third model, to obtain an output result. The transceiver unit 620 is further configured to send the output result to the second node. The first node receives indication information from the second node, where the indication information indicates the first node to send the second model.
In some embodiments, the transceiver unit 620 is specifically configured to receive the first information from the second node through a sidelink. Alternatively, the transceiver unit 620 is specifically configured to receive the first information from the second node through a common channel. Alternatively, the transceiver unit 620 is specifically configured to receive the first information from the second node through a dedicated channel of the first node.
In some possible implementations, the communication apparatus 600 can correspondingly implement behavior and functions of the second node in the foregoing method embodiments. For example, the communication apparatus 600 may be the second node, or may be a component (for example, a chip or a circuit) used in the second node. The transceiver unit 620 may be configured to perform all receiving or sending operations performed by the second node in the embodiment shown in
The transceiver unit 620 is configured to send first information to at least one first node, where the first information indicates a first model. The transceiver unit 620 is further configured to receive a second model from the at least one first node, where the second model and the first model are heterogeneous. The first model is used to process the second model, to obtain a third model that has a same structure as the first model. The processing unit 610 is configured to process the second model to obtain a fourth model. The second node sends the fourth model to the at least one first node.
In some embodiments, the transceiver unit 620 is further configured to receive an updated second model from the at least one first node, where the updated second model is obtained based on the fourth model.
In some embodiments, the transceiver unit 620 is further configured to send a first threshold to the at least one first node, where the first threshold indicates performance of the third model or the first threshold indicates a performance variation of the third model.
In some embodiments, the transceiver unit 620 is further configured to receive second information from a third node, where the second information indicates that model training is unfulfilled, and the third node is one or more nodes in the at least one first node.
In some embodiments, the transceiver unit 620 is further configured to receive an output result from the at least one first node, where the output result is an output result of the third model. The transceiver unit 620 is further configured to send indication information to a fourth node, where the indication information indicates the fourth node to send the second model. The fourth node is one or more nodes that are in the at least one first node and whose output results meet a second threshold.
In some embodiments, the transceiver unit 620 is specifically configured to send the first information to the at least one first node through a sidelink. Alternatively, the transceiver unit 620 is specifically configured to send the first information to the at least one first node through a common channel. Alternatively, the transceiver unit 620 is specifically configured to separately send the first information to the at least one first node through a dedicated channel of the at least one first node.
In some possible implementations, the communication apparatus 600 can correspondingly implement behavior and functions of the first apparatus in the foregoing method embodiments. For example, the communication apparatus 600 may be the first apparatus, or may be a component (for example, a chip or a circuit) used in the first apparatus. The transceiver unit 620 may be configured to perform all receiving or sending operations performed by the first apparatus in the embodiment shown in
The transceiver unit 620 is configured to receive I intermediate results from a second apparatus. I is an integer greater than or equal to 1. The I intermediate results are outputs of I first models. Each of the I first models may be obtained by fusing at least one third model. The processing unit 610 is configured to input an ith intermediate result into a second model, to obtain an ith first output result. I is an integer greater than or equal to 1, and i is less than or equal to I. The second model matches at least one of a task type of the first apparatus and location information of the first apparatus. The processing unit 610 is further configured to determine a reference model in the I first models based on I first output results.
In some embodiments, the transceiver unit 620 is further configured to send first information to the second apparatus, where the first information indicates the reference model. The reference model is obtained by fusing at least one target third model. The transceiver unit 620 is further configured to receive second information from the second apparatus, where the second information includes an intermediate result of the at least one target third model. The processing unit 610 is further configured to separately input the intermediate result of the at least one target third model into the second model, to obtain at least one second output result. The processing unit 610 is further configured to determine a target model in the at least one target third model based on the at least one second output result.
In some embodiments, the communication apparatus 600 can correspondingly implement behavior and functions of the second apparatus in the foregoing method embodiments. For example, the communication apparatus 600 may be the second apparatus, or may be a component (for example, a chip or a circuit) used in the second apparatus. The transceiver unit 620 may be configured to perform all receiving or sending operations performed by the second apparatus in the embodiment shown in
The processing unit 610 is configured to group a third model included in a third model set into I groups. Each third model in the third model set matches at least one of a task type of a first apparatus and location information of the first apparatus. I is an integer greater than or equal to 1. The processing unit 610 is further configured to fuse a third model in an ith group, to obtain an ith first model. I is an integer greater than or equal to 1, and i is less than or equal to I. The processing unit 610 is further configured to input a preset parameter into the ith first model, to obtain an ith intermediate result. The transceiver unit 620 is configured to send I intermediate results to the first apparatus. The I intermediate results may be used to determine a reference model in I first models.
In some embodiments, the transceiver unit 620 is further configured to receive first information from the first apparatus. The first information indicates the reference model, and the reference model is one of the I first models. The reference model may be obtained by fusing at least one target third model. The processing unit 610 is further configured to input the preset parameter into the at least one target third model, to obtain an intermediate result of the at least one target third model. The transceiver unit 620 is further configured to send second information to the first apparatus, where the second information includes the intermediate result of the at least one target third model. The intermediate result of the at least one target third model may be used to determine a target model.
For operations performed by the processing unit 610 and the transceiver unit 620, refer to the related descriptions in the foregoing method embodiments.
It should be understood that the processing unit 610 in this embodiment of this disclosure may be implemented by a processor or a processor-related circuit component, and the transceiver unit 620 may be implemented by a transceiver, a transceiver-related circuit component, or a communication interface.
Based on a same concept, as shown in
Based on a same concept, as shown in
The communication apparatus 800 may include at least one processor 810. The processor 810 is coupled to a memory. Optionally, the memory may be located inside the apparatus, or may be located outside the apparatus. For example, the communication apparatus 800 may further include at least one memory 820. The memory 820 stores a computer program, configuration information, a computer program or instructions, and/or data necessary for implementing any one of the foregoing embodiments. The processor 810 may execute the computer program stored in the memory 820, to complete the method in any one of the foregoing embodiments.
A coupling in this embodiment of this disclosure may be an indirect coupling or a communication connection between apparatuses, units, or modules in an electrical form, a mechanical form, or another form, and is used for information exchange between the apparatuses, the units, or the modules. The processor 810 may cooperate with the memory 820. A specific connection medium between a transceiver 830, the processor 810, and the memory 820 is not limited in this embodiment of this disclosure.
The communication apparatus 800 may further include the transceiver 830, and the communication apparatus 800 may exchange information with another device via the transceiver 830. The transceiver 830 may be a circuit, a bus, a transceiver, or any other apparatus that can be configured to exchange information, or is referred to as a signal transceiver unit. As shown in
In some embodiments, the communication apparatus 800 may be used in a first node. Specifically, the communication apparatus 800 may be the first node, or may be an apparatus that can support the first node and implement a function of the first node in any one of the foregoing embodiments. The memory 820 stores a computer program, a computer program or instructions, and/or data necessary for implementing a function of the first node in any one of the foregoing embodiments. The processor 810 may execute the computer program stored in the memory 820, to complete the method performed by the first node in any one of the foregoing embodiments. When used in the first node, the transmitter 831 in the communication apparatus 800 may be configured to send a second model through the antenna 833.
In some embodiments, the communication apparatus 800 may be used in a second node. Specifically, the communication apparatus 800 may be the second node, or may be an apparatus that can support the second node and implement a function of the second node in any one of the foregoing embodiments. The memory 820 stores a computer program, a computer program or instructions, and/or data necessary for implementing a function of the second node in any one of the foregoing embodiments. The processor 810 may execute the computer program stored in the memory 820, to complete the method performed by the second node in any one of the foregoing embodiments. When used in the second node, the transmitter 831 in the communication apparatus 800 may be configured to send first information through the antenna 833.
In some embodiments, the communication apparatus 800 may be used in a first apparatus. Specifically, the communication apparatus 800 may be the first apparatus, or may be an apparatus that can support the first apparatus and implement a function of the first apparatus in any one of the foregoing embodiments. The memory 820 stores a computer program, a computer program or instructions, and/or data necessary for implementing a function of the first apparatus in any one of the foregoing embodiments. The processor 810 may execute the computer program stored in the memory 820, to complete the method performed by the first apparatus in any one of the foregoing embodiments. When used in the first apparatus, the receiver 832 in the communication apparatus 800 may be configured to receive I intermediate results through the antenna 833.
In some embodiments, the communication apparatus 800 may be used in a second apparatus. Specifically, the communication apparatus 800 may be the second apparatus, or may be an apparatus that can support the second apparatus and implement a function of the second apparatus in any one of the foregoing embodiments. The memory 820 stores a computer program, a computer program or instructions, and/or data necessary for implementing a function of the second apparatus in any one of the foregoing embodiments. The processor 810 may execute the computer program stored in the memory 820, to complete the method performed by the second apparatus in any one of the foregoing embodiments. When used in the second apparatus, the transmitter 831 in the communication apparatus 800 may be configured to send I intermediate results through the antenna 833.
The communication apparatus 800 provided in this embodiment may be used in the first node to complete the method performed by the first node, used in the second node to complete the method performed by the second node, used in the first apparatus to complete the method performed by the first apparatus, or used in the second apparatus to complete the method performed by the second apparatus. Therefore, for technical effects that can be achieved by this embodiment, refer to the foregoing method embodiments. Details are not described herein again.
In embodiments of this disclosure, the processor may be a general-purpose processor, a digital signal processor, an application-specific integrated circuit, a field programmable gate array or another programmable logic device, a discrete gate or transistor logic device, or a discrete hardware component, and may implement or execute the methods, steps, and logical block diagrams disclosed in embodiments of this disclosure. The general-purpose processor may be a microprocessor, any other processor, or the like. The steps of the method disclosed with reference to embodiments of this disclosure may be directly performed by a hardware processor, or may be performed by using a combination of hardware in the processor and a software module.
In embodiments of this disclosure, the memory may be a non-volatile memory, for example, a hard disk drive (HDD) or a solid-state drive (SSD), or may be a volatile memory, for example, a random-access memory (RAM). Alternatively, the memory may be any other medium that can be configured to carry or store expected program code in a form of an instruction or a data structure and that can be accessed by a computer, but is not limited thereto. The memory in this embodiment of this disclosure may alternatively be a circuit or any other apparatus that can implement a storage function, and is configured to store a computer program, a computer program or instructions, and/or data.
Based on the foregoing embodiments, refer to
The following describes in detail operations performed by the communication apparatus when the communication apparatus is used in the first node, the second node, the first apparatus, or the second apparatus.
In an optional implementation, the communication apparatus 900 may be used in the first node, to perform the method performed by the first node, specifically, for example, the method performed by the first node in the embodiment shown in
For example, the input/output interface 910 is configured to input first information from a second node, where the first information indicates a first model. The logic circuit 920 is configured to process a second model based on the first model, to obtain a third model. A structure of the third model is the same as a structure of the first model, and the second model and the first model are heterogeneous. The input/output interface 910 is further configured to output the second model to the second node. The third model meets a preset condition.
In another optional implementation, the communication apparatus 900 may be used in the second node, to perform the method performed by the second node, specifically, for example, the method performed by the second node in the embodiment shown in
For example, the input/output interface 910 is configured to output first information to at least one first node, where the first information indicates a first model. The input/output interface 910 is further configured to input a second model from the at least one first node, where the second model and the first model are heterogeneous. The first model is used to process the second model, to obtain a third model that has a same structure as the first model. The logic circuit 920 is configured to process the second model to obtain a fourth model. The second node outputs the fourth model to the at least one first node.
In another optional implementation, the communication apparatus 900 may be used in the first apparatus, to perform the method performed by the first apparatus, specifically, for example, the method performed by the first apparatus in the embodiment shown in
For example, the input/output interface 910 is configured to input I intermediate results from a second apparatus. I is an integer greater than or equal to 1. The I intermediate results are outputs of I first models. Each of the I first models may be obtained by fusing at least one third model. The logic circuit 920 is configured to input an ith intermediate result into a second model, to obtain an ith first output result. I is an integer greater than or equal to 1, and i is less than or equal to I. The second model matches at least one of a task type of the first apparatus and location information of the first apparatus. The logic circuit 920 is further configured to determine a reference model in the I first models based on I first output results.
In another optional implementation, the communication apparatus 900 may be used in the second apparatus, to perform the method performed by the second apparatus, specifically, for example, the method performed by the second apparatus in the embodiment shown in
For example, the logic circuit 920 is configured to group a third model included in a third model set into I groups. Each third model in the third model set matches at least one of a task type of a first apparatus and location information of the first apparatus. I is an integer greater than or equal to 1. The logic circuit 920 is further configured to fuse a third model in an ith group, to obtain an ith first model. I is an integer greater than or equal to 1, and i is less than or equal to I. The logic circuit 920 is further configured to input a preset parameter into the ith first model, to obtain an ith intermediate result. The input/output interface 910 is configured to output I intermediate results to the first apparatus. The I intermediate results may be used to determine a reference model in I first models.
The communication apparatus 900 provided in this embodiment may be used in the first node to complete the method performed by the first node, used in the second node to complete the method performed by the second node, used in the first apparatus to complete the method performed by the first apparatus, or used in the second apparatus to complete the method performed by the second apparatus. Therefore, for technical effects that can be achieved by this embodiment, refer to the foregoing method embodiments. Details are not described herein again.
Based on the foregoing embodiments, an embodiment of this disclosure further provides a communication system. The communication system includes at least one communication apparatus used in a first node and at least one communication apparatus used in a second node. For technical effect that can be achieved by this embodiment, refer to the foregoing method embodiments. Details are not described herein again.
Based on the foregoing embodiments, an embodiment of this disclosure further provides a communication system. The communication system includes at least one communication apparatus used in a first apparatus and at least one communication apparatus used in a second apparatus. For technical effect that can be achieved by this embodiment, refer to the foregoing method embodiments. Details are not described herein again.
Based on the foregoing embodiments, an embodiment of this disclosure further provides a computer-readable storage medium. The computer-readable storage medium stores a computer program or instructions. When the instructions are executed, the method performed by the first node in any one of the foregoing embodiments is implemented, the method performed by the second node in any one of the foregoing embodiments is implemented, the method performed by the first apparatus in any one of the foregoing embodiments is implemented, or the method performed by the second apparatus in any one of the foregoing embodiments is implemented. The computer-readable storage medium may include any medium that can store program code, for example, a USB flash drive, a removable hard disk drive, a read-only memory, a random access memory, a magnetic disk, or an optical disc.
To implement functions of the communication apparatuses in
A person skilled in the art should understand that embodiments of this disclosure may be provided as a method, a system, or a computer program product. Therefore, this disclosure may use a form of hardware only embodiments, software only embodiments, or embodiments with a combination of software and hardware. In addition, this disclosure may use a form of a computer program product that is implemented on one or more computer-usable storage media (including but not limited to a disk memory, a compact disc ROM (CD-ROM), an optical memory, and the like) that include computer-usable program code.
This disclosure is described with reference to the flowcharts and/or block diagrams of the method, the device (system), and the computer program product according to embodiments of this disclosure. It should be understood that a computer program or instructions may be used to implement each procedure and/or each block in the flowcharts and/or the block diagrams and a combination of a procedure and/or a block in the flowcharts and/or the block diagrams. The computer program or instructions may be provided for a general-purpose computer, a dedicated computer, an embedded processor, or a processor of another programmable data processing device to generate a machine, so that the instructions executed by the computer or the processor of another programmable data processing device generate an apparatus for implementing a specific function in one or more procedures in the flowcharts and/or in one or more blocks in the block diagrams.
The computer program or instructions may alternatively be stored in a computer-readable memory that can instruct the computer or other programmable data processing device to work in a specific manner, so that the instructions stored in the computer-readable memory generate an artifact that includes an instruction apparatus. The instruction apparatus implements a specific function in one or more procedures in the flowcharts and/or in one or more blocks in the block diagrams.
The computer program or instructions may alternatively be loaded onto the computer or other programmable data processing device, so that a series of operation steps are performed on the computer or other programmable device to generate computer-implemented processing. Therefore, the instructions executed on the computer or other programmable device provide steps for implementing a specific function in one or more procedures in the flowcharts and/or in one or more blocks in the block diagrams.
It is clear that a person skilled in the art may make various modifications and variations to embodiments of this disclosure without departing from the scope of embodiments of this disclosure. In this case, this disclosure is intended to cover these modifications and variations of embodiments of this disclosure provided that they fall within the scope of protection defined by the following claims and their equivalent technologies.
This is a continuation of Int'l Patent App. No. PCT/CN2022/113137, filed on Aug. 17, 2022, which is incorporated by reference.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/CN2022/113137 | Aug 2022 | WO |
Child | 19054317 | US |