This application relates to the field of communication technologies, and in particular, to a neural network training method and a related apparatus.
A wireless communication system may include three parts: a transmitter, a channel, and a receiver. The channel is used for transmission of signals exchanged between the transmitter and the receiver. For example, the transmitter may be an access network device, for example, a base station (BS), and the receiver may be a terminal device. For another example, the transmitter may be a terminal device, and the receiver may be an access network device.
To optimize performance of the communication system, the transmitter and the receiver may be optimized. The transmitter and the receiver each may have an independent mathematical model. Therefore, the transmitter and the receiver are independently optimized based on respective mathematical models. For example, a mathematical channel model may be used to generate a channel sample to optimize the transmitter and the receiver.
However, because the mathematical channel model is non-ideal and non-linear, the channel sample generated by the mathematical channel model defined in a protocol can hardly reflect an actual channel environment. Transmission of a large quantity of actual channel samples between the transmitter and the receiver occupies excessive air interface resources, affecting data transmission efficiency.
According to a first aspect, an embodiment of this application provides a neural network training method, including the following operations.
A first device receives first channel sample information from a second device. The first device determines a first neural network. The first neural network is obtained through training based on the first channel sample information, and is used to perform inference to obtain second channel sample information.
Optionally, an example in which the first device is an access network device and the second device is a terminal device is used for description. It may be understood that the first device may be the access network device, a chip used in the access network device, a circuit used in the access network device, or the like; and the second device may be the terminal device, a chip used in the terminal device, a circuit used in the terminal device, or the like.
In a possible design, the method includes: the first device obtains the first neural network through training based on the first channel sample information. The first neural network is used to generate new channel sample information, for example, the second channel sample information.
In a possible design, the method includes: the first device receives information about the first neural network from a third device, and determines the first neural network based on the information about the first neural network. The first neural network is obtained by the third device through training based on the first channel sample information.
According to this method, the first neural network may be obtained through training based on the first channel sample information. The first neural network is used to perform inference to obtain the second channel sample information. According to this method, air interface signaling overheads during channel sample information transmission can be effectively reduced.
In a possible design, the second channel sample information is used to train a second neural network and/or a third neural network, and the second neural network and/or the third neural network are/is used for transmission of target information between the first device and the second device.
In a possible design, the method includes: the first device trains the second neural network and/or the third neural network based on the second channel sample information or based on the second channel sample information and the first channel sample information. The second neural network and/or the third neural network are/is used for transmission of target information between the first device and the second device.
In a possible design, the method includes: the first device receives information about the second neural network and/or information about the third neural network from the third device. The second neural network and/or the third neural network are/is used for transmission of target information between the first device and the second device. The second neural network and/or the third neural network are/is obtained by the third device through training based on the second channel sample information or based on the second channel sample information and the first channel sample information.
According to this method, air interface signaling overheads can be effectively reduced, and a channel environment in which a trained neural network is located can be adapted. The second neural network and the third neural network obtained through training are closer to an actual channel environment, and communication performance is improved. A speed of training the second neural network and the third neural network is also greatly improved.
In a possible design, the method further includes: the first device sends a first reference signal to the second device. Optionally, the first reference signal includes a demodulation reference signal DMRS or a channel state information reference signal CSI-RS. Optionally, a sequence type of the first reference signal includes a ZC sequence or a gold sequence.
In a possible design, the first channel sample information includes but is not limited to a second reference signal and/or channel state information (CSI). The second reference signal is the first reference signal propagated through a channel. Alternatively, it is described as that the second reference signal is the first reference signal received by the second device from the first device.
In a possible design, the information about the first neural network includes a model variation of the first neural network relative to a reference neural network.
In a possible design, the information about the first neural network includes one or more of the following: a weight of a neural network, an activation function of a neuron, a quantity of neurons at each layer of the neural network, an inter-layer cascading relationship of the neural network, and a network type of each layer of the neural network.
In a possible design, the first neural network is a generative neural network. Optionally, the first neural network is a generative adversarial network (GAN) or a variational autoencoder (VAE).
According to this method, the second channel sample information that has a same distribution as or has a similar distribution to the first channel sample information may be obtained, so that the second channel sample information is closer to an actual channel environment.
In a possible design, the method further includes: the first device receives capability information of the second device from the second device. The capability information indicates one or more of the following information:
According to this method, the first device may receive the capability information sent by the second device. The capability information notifies the first device of related information about the second device. The first device may perform an operation related to the third neural network based on the capability information, to ensure that the second device can normally use the third neural network.
In a possible design, the method further includes: the first device sends the information about the third neural network to the second device.
In this embodiment, after completing the training of the third neural network, the first device sends the information about the third neural network to the second device. The information about the third neural network includes but is not limited to: a weight of a neural network, an activation function of a neuron, a quantity of neurons at each layer of the neural network, an inter-layer cascading relationship of the neural network, and/or a network type of each layer of the neural network. For example, the information about the third neural network may indicate different activation functions for different neurons.
In a possible design, when the third neural network is preconfigured (or predefined) in the second device, the information about the third neural network may alternatively be a model variation of the third neural network. The model variation includes but is not limited to: a weight of a changed neural network, a changed activation function, a quantity of neurons at one or more layers of the changed neural network, an inter-layer cascading relationship of the changed neural network, and/or a network type of one or more layers of the changed neural network. For example, the pre-configuration may be performed by the access network device using signaling, and the pre-definition may be performed in a protocol. For example, the third neural network in the terminal device is predefined as a neural network A in the protocol.
In embodiments of this application, there may be a plurality of implementation solutions for the information about the third neural network, so that implementation flexibility of the solution is improved.
According to a second aspect, an embodiment of this application provides a neural network training method, including: a second device performs channel estimation based on a first reference signal received from a first device, to determine first channel sample information. The second device sends the first channel sample information to the first device. The second device receives information about a third neural network from the first device. The third neural network is used for transmission of target information between the first device and the second device.
In a possible design, the method further includes: the second device sends capability information of the second device to the first device.
For descriptions of the first channel sample information, the information about the third neural network, the capability information of the second device, and the like, refer to the first aspect. Details are not described herein again.
According to a third aspect, an embodiment of this application provides a neural network training method, including the following operations.
A first device sends a first reference signal to a second device. The first device receives information about a first neural network from the second device. The first neural network is used to perform inference to obtain second channel sample information.
In the method, the second device (for example, a terminal device) obtains the first neural network through training based on the first channel sample information. The first neural network is used to perform inference to obtain the second channel sample information. According to this method, air interface signaling overheads during channel sample information transmission can be effectively reduced.
In a possible design, the second channel sample information is used to train a second neural network and/or a third neural network, and the second neural network and/or the third neural network are/is used for transmission of target information between the first device and the second device.
In a possible design, the method includes: the first device trains the second neural network and/or the third neural network based on the second channel sample information or based on the second channel sample information and the first channel sample information. The second neural network and/or the third neural network are/is used for transmission of target information between the first device and the second device.
In a possible design, the method includes: the first device receives information about the second neural network and/or information about the third neural network from a third device. The second neural network and/or the third neural network are/is used for transmission of target information between the first device and the second device. The second neural network and/or the third neural network are/is obtained by the third device through training based on the second channel sample information or based on the second channel sample information and the first channel sample information.
According to this method, air interface signaling overheads can be effectively reduced, and a channel environment in which a trained neural network is located can be adapted. The second neural network and the third neural network obtained through training are closer to an actual channel environment, and communication performance is improved. A speed of training the second neural network and the third neural network is also greatly improved.
Specifically, for descriptions of the first reference signal, the first neural network, the information about the first neural network, the second neural network, and/or the third neural network, refer to the first aspect. Details are not described again.
In a possible design, the method further includes: the first device sends the information about the third neural network to the second device.
Specifically, for descriptions of the information about the third neural network, refer to the first aspect. Details are not described again.
In a possible design, the method further includes the following operations.
The first device receives capability information from the second device. The capability information indicates one or more of the following information about the second device:
According to a fourth aspect, an embodiment of this application provides a neural network training method, including the following operations.
A second device performs channel estimation based on a first reference signal received from a first device, to determine first channel sample information. The second device determines a first neural network. The first neural network is obtained through training based on first channel sample information. The second device sends information about the first neural network to the first device.
Specifically, for descriptions of the first reference signal, the first neural network, the first channel sample information, the information about the first neural network, and the like, refer to the third aspect. Details are not described again.
In a possible design, the method further includes the following operations.
The second device receives information about a third neural network from the first device. For the information about the third neural network, refer to the third aspect. Details are not described herein again.
In a possible design, the method further includes: the second device sends capability information to the first device. For the capability information, refer to the third aspect. Details are not described herein again.
According to a fifth aspect, an embodiment of this application provides a neural network training method, including: a third device receives first channel sample information from a first device. The third device obtains a first neural network through training based on the first channel sample information. The first neural network is used to perform inference to obtain second channel sample information.
In a possible design, the method further includes: training a second neural network and/or a third neural network based on the second channel sample information, and sending information about the second neural network and/or information about the third neural network to the first device. The second neural network and/or the third neural network are/is used for transmission of target information between the first device and a second device.
According to a sixth aspect, an apparatus is provided. The apparatus may be an access network device, an apparatus in the access network device, or an apparatus that can be used with the access network device.
In a possible design, the apparatus may include modules for performing the method/operations/steps/actions described in the first aspect. The modules may be hardware circuits, may be software, or may be implemented by using a combination of a hardware circuit and software. In a design, the apparatus may include a processing module and a transceiver module.
For example, the transceiver module is configured to receive first channel sample information from a second device.
The processing module is configured to determine a first neural network. The first neural network is obtained through training based on the first channel sample information, and is used to perform inference to obtain second channel sample information.
For descriptions of the first neural network, the first channel sample information, the second channel sample information, and other operations, refer to the first aspect. Details are not described herein again.
In a possible design, the apparatus may include modules for performing the method/operations/steps/actions described in the third aspect. The modules may be hardware circuits, may be software, or may be implemented by using a combination of a hardware circuit and software. In a design, the apparatus may include a processing module and a transceiver module.
For example, the transceiver module is configured to send a first reference signal to the second device, and receive information about a first neural network from the second device. The first neural network is used to perform inference to obtain the second channel sample information.
For descriptions of the first neural network, the first channel sample information, the second channel sample information, and other operations, refer to the third aspect. Details are not described herein again.
According to a seventh aspect, an apparatus is provided. The apparatus may be a terminal device, an apparatus in the terminal device, or an apparatus that can be used with the terminal device.
In a possible design, the apparatus may include modules for performing the method/operations/steps/actions described in the second aspect. The modules may be hardware circuits, may be software, or may be implemented by using a combination of a hardware circuit and software. In a design, the apparatus may include a processing module and a transceiver module.
For example, the processing module is configured to perform channel estimation based on a first reference signal received from a first device, to determine first channel sample information.
The transceiver module is configured to send the first channel sample information to the first device.
The transceiver module is further configured to receive information about a third neural network from the first device. The third neural network is used for transmission of target information between the first device and a second device.
For descriptions of the first reference signal, the first channel sample information, the third neural network, and other operations, refer to the second aspect. Details are not described herein again.
In a possible design, the apparatus may include modules for performing the method/operations/steps/actions described in the fourth aspect. The modules may be hardware circuits, may be software, or may be implemented by using a combination of a hardware circuit and software. In a design, the apparatus may include a processing module and a transceiver module.
For example, the processing module is configured to perform channel estimation based on the first reference signal received from the first device, to determine the first channel sample information.
The processing module is further configured to determine a first neural network. The first neural network is obtained through training based on the first channel sample information.
The transceiver module is configured to send information about the first neural network to the first device.
For descriptions of the first reference signal, the first channel sample information, the first neural network, and other operations, refer to the fourth aspect. Details are not described herein again.
According to an eighth aspect, an apparatus is provided. The apparatus may be an AI node, an apparatus in the AI node, or an apparatus that can be used with the AI node.
In a possible design, the apparatus may include modules for performing the method/operations/steps/actions described in the fifth aspect. The modules may be hardware circuits, may be software, or may be implemented by using a combination of a hardware circuit and software. In a design, the apparatus may include a processing module and a transceiver module.
For example, the transceiver module is configured to receive first channel sample information from a first device.
The processing module is configured to obtain a first neural network through training based on the first channel sample information. The first neural network is used to perform inference to obtain second channel sample information.
In a possible design, the processing module is further configured to train a second neural network and/or a third neural network based on the second channel sample information. The transceiver module is further configured to send information about the second neural network and/or information about the third neural network to the first device. The second neural network and/or the third neural network are/is used for transmission of target information between the first device and a second device.
According to a ninth aspect, an embodiment of this application provides an apparatus.
In a possible design, the apparatus includes a processor, configured to implement the method described in the first aspect. The apparatus may further include a memory, configured to store instructions and data. The memory is coupled to the processor, and the processor may implement the method described in the first aspect when executing the instructions stored in the memory. The apparatus may further include a communication interface. The communication interface is used by the apparatus to communicate with another device. For example, the communication interface may be a transceiver, a circuit, a bus, a module, or another type of communication interface.
In a possible design, the apparatus includes:
The processor is further configured to determine a first neural network. The first neural network is obtained through training based on the first channel sample information, and is used to perform inference to obtain second channel sample information.
For descriptions of the first neural network, the first channel sample information, the second channel sample information, and other operations, refer to the first aspect. Details are not described herein again.
In a possible design, the apparatus includes a processor, configured to implement the method described in the third aspect. The apparatus may further include a memory, configured to store instructions and data. The memory is coupled to the processor, and the processor may implement the method described in the third aspect when executing the instructions stored in the memory. The apparatus may further include a communication interface. The communication interface is used by the apparatus to communicate with another device. For example, the communication interface may be a transceiver, a circuit, a bus, a module, or another type of communication interface.
In a possible design, the apparatus includes:
For descriptions of the first neural network, the first channel sample information, the second channel sample information, and other operations, refer to the third aspect. Details are not described herein again.
According to a tenth aspect, an embodiment of this application provides an apparatus.
In a possible design, the apparatus includes a processor, configured to implement the method described in the second aspect. The apparatus may further include a memory, configured to store instructions and data. The memory is coupled to the processor, and the processor may implement the method described in the second aspect when executing the instructions stored in the memory. The apparatus may further include a communication interface. The communication interface is used by the apparatus to communicate with another device. For example, the communication interface may be a transceiver, a circuit, a bus, a module, or another type of communication interface.
In a possible design, the apparatus includes:
For descriptions of the first reference signal, the first channel sample information, the third neural network, and other operations, refer to the second aspect. Details are not described herein again.
In a possible design, the apparatus includes a processor, configured to implement the method described in the fourth aspect. The apparatus may further include a memory, configured to store instructions and data. The memory is coupled to the processor, and the processor may implement the method described in the fourth aspect when executing the instructions stored in the memory. The apparatus may further include a communication interface. The communication interface is used by the apparatus to communicate with another device. For example, the communication interface may be a transceiver, a circuit, a bus, a module, or another type of communication interface.
In a possible design, the apparatus includes:
For descriptions of the first reference signal, the first channel sample information, the first neural network, and other operations, refer to the fourth aspect. Details are not described herein again.
According to an eleventh aspect, an embodiment of this application provides an apparatus. The apparatus includes a processor, configured to implement the method described in the fifth aspect. The apparatus may further include a memory, configured to store instructions and data. The memory is coupled to the processor, and the processor may implement the method described in the fifth aspect when executing the instructions stored in the memory. The apparatus may further include a communication interface. The communication interface is used by the apparatus to communicate with another device. For example, the communication interface may be a transceiver, a circuit, a bus, a module, or another type of communication interface.
In a possible design, the apparatus includes:
In a possible design, the processor is further configured to train a second neural network and/or a third neural network based on the second channel sample information. The processor further sends information about the second neural network and/or information about the third neural network to the first device through the communication interface. The second neural network and/or the third neural network are/is used for transmission of target information between the first device and a second device.
According to a twelfth aspect, an embodiment of this application further provides a computer-readable storage medium, including instructions. When the instructions are run on a computer, the computer is enabled to perform the method in any one of the first aspect to the fifth aspect.
According to a thirteenth aspect, an embodiment of this application further provides a computer program product, including instructions. When the instructions are run on a computer, the computer is enabled to perform the method in any one of the first aspect to the fifth aspect.
According to a fourteenth aspect, an embodiment of this application provides a chip system. The chip system includes a processor, and may further include a memory, configured to implement the method in any one of the first aspect to the fifth aspect. The chip system may include a chip, or may include a chip and another discrete component.
According to a fifteenth aspect, an embodiment of this application further provides a communication system, and the communication system includes:
A wireless communication system includes communication devices, and the communication devices may perform wireless communication by using a radio resource. The communication devices may include an access network device and a terminal device, and the access network device may also be referred to as an access side device. The radio resource may include a link resource and/or an air interface resource. The air interface resource may include at least one of a time domain resource, a frequency domain resource, a code resource, and a space resource. In embodiments of this application, “at least one (type)” may alternatively be described as “one (type) or more (types)”, and “a plurality of (types)” may be two (types), three (types), four (types), or more (types). This is not limited in embodiments of this application.
In embodiments of this application, “/” may represent an “or” relationship between associated objects. For example, A/B may represent A or B. “And/or” may be used to indicate that three relationships exist between associated objects. For example, A and/or B may represent the following three cases: Only A exists, both A and B exist, and only B exists. A and B may be singular or plural. To facilitate descriptions of the technical solutions in embodiments of this application, terms such as “first” and “second” may be used to distinguish between technical features with same or similar functions. The terms such as “first” and “second” do not limit a quantity and an execution sequence, and the terms such as “first” and “second” do not indicate a definite difference. In embodiments of this application, the terms such as “example” or “for example” are used to represent an example, an illustration, or a description. Any embodiment or design scheme described with “example” or “for example” should not be explained as being more preferred or having more advantages than another embodiment or design scheme. Use of the terms such as “example” or “for example” is intended to present a related concept in a specific manner for ease of understanding.
A specific quantity of network elements in the communication system is not limited in embodiments of this application.
The terminal device in embodiments of this application may also be referred to as a terminal or an access terminal, and may be a device having a wireless transceiver function. The terminal device may communicate with one or more core networks (CNs) via the access network device. The terminal device may be a subscriber unit, a subscriber station, a mobile station, a mobile console, a remote station, a remote terminal, a mobile device, a user terminal, user equipment (UE), a user agent, a user apparatus, or the like. The terminal device may be deployed on land, including indoor or outdoor, in a handheld manner or vehicle-mounted manner, may be deployed on water (for example, on a ship), or may be deployed in air (for example, on a plane, a balloon, or a satellite). The terminal device may be a cellular phone, a cordless phone, a session initiation protocol (SIP) phone, a smartphone, a mobile phone, a wireless local loop (WLL) station, or a personal digital assistant (PDA). Alternatively, the terminal device may be a handheld device, a computing device, or another device having a wireless communication function, a vehicle-mounted device, a wearable device, an unmanned aerial vehicle device, a terminal in internet of things or internet of vehicles, a terminal in a fifth generation (5G) mobile communication network, relay user equipment, or a terminal in a future evolved mobile communication network, or the like. The relay user equipment may be, for example, a 5G residential gateway (RG). For another example, the terminal device may be a virtual reality (virtual reality, VR) terminal, an augmented reality (AR) terminal, a wireless terminal in industrial control, a wireless terminal in self-driving, a wireless terminal in remote medical, a wireless terminal in a smart grid, a wireless terminal in transportation safety, a wireless terminal in a smart city, a wireless terminal in a smart home, or the like. This is not limited in embodiments of this application. In embodiments of this application, an apparatus configured to implement a function of the terminal may be a terminal, or may be an apparatus that can support the terminal in implementing the function, for example, a chip system. The apparatus may be installed in the terminal or used with the terminal. In embodiments of this application, the chip system may include a chip, or may include a chip and another discrete component.
The access network device may be considered as a sub-network of a carrier network, and is an implementation system between a service node in the carrier network and the terminal device. To access the carrier network, the terminal device may first pass through the access network device, and then may be connected to the service node in the carrier network via the access network device. The access network device in embodiments of this application is a device that is located in a (radio) access network ((R)AN) and that can provide the wireless communication function for the terminal device. The access network device includes a base station, and includes but is not limited to, for example, a next generation NodeB (gNB) in a 5G system and an evolved NodeB (eNB) in a long term evolution (LTE) system, a radio network controller (RNC), a NodeB (NB), a base station controller (BSC), a base transceiver station (BTS), a home base station (for example, home evolved NodeB, or home NodeB, HNB), a base band unit (BBU), a transmitting and receiving point (TRP), a transmitting point (TP), a pico base station device, a mobile switching center, and an access network device in a future network. In systems using different radio access technologies, devices having a function of the access network device may have different names. In embodiments of this application, an apparatus configured to implement the function of the access network device may be an access network device, or may be an apparatus that can support the access network device in implementing the function, for example, a chip system. The apparatus may be installed in the access network device or used with the access network device.
The technical solutions provided in embodiments of this application may be applied to various communication systems, for example, an LTE system, a 5G system, a wireless-fidelity (Wi-Fi) system, a future sixth generation mobile communication system, or a system integrating a plurality of communication systems. This is not limited in embodiments of this application. 5G may also be referred to as new radio (NR).
The technical solutions provided in embodiments of this application may be applied to various communication scenarios, for example, may be applied to one or more of the following communication scenarios: enhanced mobile broadband (eMBB) communication, ultra-reliable low-latency communication (URLLC), machine type communication (MTC), massive machine type communication (mMTC), device-to-device (D2D) communication, vehicle to everything (V2X) communication, vehicle to vehicle (V2V) communication, internet of things (IoT), and the like. In embodiments of this application, the term “communication” may also be described as “transmission”, “information transmission”, “data transmission”, “signal transmission”, or the like. The transmission may include sending and/or receiving. In embodiments of this application, communication between the access network device and the terminal device is used as an example to describe the technical solutions. A person skilled in the art may also apply the technical solutions to other communication between a scheduling entity and a subordinate entity, for example, communication between a macro base station and a micro base station, and/or communication between a terminal device 1 and a terminal device 2.
In addition, network architectures and service scenarios described in this application are intended to describe the technical solutions in this application more clearly, and do not constitute a limitation on the technical solutions provided in this application. A person skilled in the art may know that as the network architectures evolve and a new service scenario emerges, the technical solutions provided in this application are also applicable to similar technical problems.
As shown in
The processor 204 may be a general-purpose processor, a digital signal processor, an application-specific integrated circuit, a field programmable gate array and/or another programmable logic device, for example, a discrete gate or a transistor logic device, and/or a discrete hardware component. The general-purpose processor may be a microprocessor, any conventional processor, or the like. The processor 204 may implement or perform the methods, steps, and logical block diagrams disclosed in embodiments of this application. The steps in the methods disclosed with reference to embodiments of this application may be performed and completed by a hardware processor, or may be performed and completed by a combination of hardware and a software module in the processor. For example, the processor 204 may be a central processing unit (CPU), or may be a dedicated processor, for example, but not limited to, a digital signal processor (DSP), an application-specific integrated circuit (ASIC), and/or a field programmable gate array (FPGA). The processor 204 may alternatively be a neural processing unit (NPU). In addition, the processor 204 may alternatively be a combination of a plurality of processors. In the technical solutions provided in embodiments of this application, the processor 204 may be configured to perform related steps in subsequent method embodiments. The processor 204 may be a processor that is specially designed to perform the foregoing steps and/or operations, or may be a processor that performs the foregoing steps and/or operations by reading and executing the instructions 2031 stored in the memory 203. The processor 204 may need to use the data 2032 in a process of performing the foregoing steps and/or operations.
The transceiver 202 includes the transmitter 2021 and the receiver 2022. In an optional implementation, the transmitter 2021 is configured to send a signal through at least one of the antennas 206. The receiver 2022 is configured to receive a second reference signal through at least one of the antennas 206.
In embodiments of this application, the transceiver 202 is configured to support the communication apparatus in performing a receiving function and a sending function. A processor having a processing function is considered as the processor 204. The receiver 2022 may also be referred to as an input port, a receiver circuit, a receive bus, or another apparatus for implementing the receiving function, or the like. The transmitter 2021 may be referred to as a transmit port, a transmitter circuit, a transmit bus, or another apparatus for implementing the transmitting function, or the like. The transceiver 202 may also be referred to as a communication interface.
The processor 204 may be configured to execute the instructions stored in the memory 203, for example, to control the transceiver 202 to receive a message and/or send a message, to complete functions of the communication apparatus in the method embodiments of this application. In an implementation, it may be considered that a function of the transceiver 202 is implemented by using a transceiver circuit or a dedicated transceiver chip. In embodiments of this application, that the transceiver 202 receives a message may be understood as that the transceiver 202 inputs a message, and that the transceiver 202 sends a message may be understood as that the transceiver 202 outputs a message.
The memory 203 may be a nonvolatile memory, for example, a hard disk drive (HDD) or a solid-state drive (SSD), or may be a volatile memory, for example, a random access memory (RAM). The memory is any other medium that can carry or store expected program code in a form of instructions or a data structure and that can be accessed by a computer, but is not limited thereto. The memory in embodiments of this application may alternatively be a circuit or any other apparatus that can implement a storage function, and is configured to store the program instructions and/or the data. The memory 203 is specifically configured to store the instructions 2031 and the data 2032. The processor 204 may perform the steps and/or the operations in the method embodiments of this application by reading and executing the instructions 2031 stored in the memory 203. The data 2032 may need to be used in a process of performing the operations and/or the steps in the method embodiments of this application.
Optionally, the communication apparatus may further include the I/O interface 210. The I/O interface 210 is configured to receive instructions and/or data from a peripheral device, and output instructions and/or data to the peripheral device.
The following describes some concepts in embodiments of this application.
Machine learning (ML) has attracted wide attention in academia and industry in recent years. Due to great advantages of machine learning in addressing structured information and massive data, many researchers in the communication field also focus on machine learning. Machine learning-based communication technologies have great potential in signal classification, channel estimation, and/or performance optimization, and the like. Most communication systems are designed block by block, which means that these communication systems are formed by a plurality of modules. For a communication architecture based on a module design, many technologies may be developed to optimize performance of each module. However, the optimal performance of each module does not mean the optimal performance of the entire communication system. End-to-end optimization (namely, optimization of the entire communication system) is better than optimization of a single model. Machine learning provides an advanced and powerful tool for maximizing end-to-end performance as much as possible. In a wireless communication system, in a complex and large-scale communication scenario, a channel status changes rapidly. A plurality of conventional communication models, for example, a massive multiple-input multiple-output (MIMO) model, heavily rely on channel state information, and performance of the communication models deteriorates in nonlinear time-varying channels. Therefore, accurate acquisition of channel state information (CSI) of the time-varying channels is very important for system performance. By using a machine learning technology, a communication system may learn a sudden change channel model and feedback a channel state in time.
Based on the foregoing considerations, the use of the machine learning technology in wireless communication can adapt to a new requirement in a future wireless communication scenario.
Machine learning is an important technical way to implement artificial intelligence. Machine learning may include supervised learning, unsupervised learning, and reinforcement learning.
In terms of supervised learning, based on collected sample values and sample labels, a mapping relationship between the sample values and the sample labels is learned by using a machine learning algorithm, and the learned mapping relationship is expressed by using a machine learning model. The sample label may also be referred to as a label for short. A process of training a machine learning model is a process of learning the mapping relationship. For example, during signal detection, a received signal including noise is a sample, and a real constellation point corresponding to the signal is a label. Machine learning expects to learn a mapping relationship between the sample and the label through training. In other words, a machine learning model is enabled to learn a signal detector. During training, a model parameter is optimized by calculating an error between a predicted value of a model and an actual label. After learning the mapping relationship, a sample label of a new sample may be predicted based on the learned mapping. The mapping relationship learned through supervised learning may include linear mapping or non-linear mapping. A learning task may be classified into a classification task and a regression task based on a type of a label.
In terms of unsupervised learning, an internal pattern of a sample is explored autonomously by using an algorithm based on a collected sample value. In unsupervised learning, a sample is used as a supervised signal in a type of algorithm, in other words, a model learns a mapping relationship between a sample and a sample, which is also referred to as self-supervised learning. During training, a model parameter is optimized by calculating an error between a predicted value of a model and a sample. For example, self-supervised learning may be used for signal compression and decompression restoration. Common algorithms include an autoencoder, a generative adversarial network, and the like.
Reinforcement learning is different from supervised learning, and is an algorithm that learns a policy of resolving problems by interacting with an environment. Different from supervised learning and unsupervised learning, reinforcement learning does not have clear “correct” action label data. The algorithm needs to interact with the environment to obtain a reward signal fed back by the environment and adjust a decision action to obtain a larger reward signal value. For example, in downlink power control, a reinforcement learning model adjusts a downlink transmit power of each user based on a total system throughput fed back by a wireless network, to expect to obtain a higher system throughput. The goal of reinforcement learning is also to learn a mapping relationship between an environment status and an optimal decision action. However, a label of “correct action” cannot be obtained in advance. Therefore, a network cannot be optimized by calculating an error between an action and the “correct action”. Reinforcement learning training is implemented through iterative interaction with the environment.
A deep neural network (DNN) is a specific implementation form of machine learning. According to a universal approximation theorem, the DNN may theoretically approximate to any continuous function, so that the DNN can learn any mapping. A conventional communication system needs to design a communication module with rich expert knowledge. However, a DNN-based deep learning communication system may automatically discover an implicit pattern structure from a large quantity of data sets, establish a mapping relationship between data, and obtain performance better than that of a conventional modeling method.
The idea of the DNN comes from a neuron structure of a brain tissue. Each neuron performs a weighted summation operation on an input value of the neuron, and outputs a result of the weighted summation through an activation function.
The term wixi represents a product of wi and xi, a data type of b is an integer or a floating point number, and a data type of wi is an integer or a floating point number. The DNN usually has a multi-layer structure. Each layer of the DNN may include one or more neurons. An input layer of the DNN processes a received value by using a neuron, and then transfers the processed value to an intermediate hidden layer. Similarly, the hidden layer transfers a calculation result to an adjacent next hidden layer or an adjacent output layer, to generate a final output of the DNN.
The DNN usually includes one or more hidden layers, and the hidden layer may affect capabilities of extracting information and fitting a function. Increasing a quantity of hidden layers of the DNN or increasing a quantity of neurons at each layer can improve a function fitting capability of the DNN. A parameter of each neuron includes a weight, an offset, and an activation function. A set including parameters of all neurons in the DNN is referred to as a DNN parameter (or referred to as a neural network parameter). The weight and the offset of the neuron may be optimized in a training process, so that the DNN has capabilities of extracting a data feature and expressing a mapping relationship. The DNN usually optimizes the neural network parameter by using a supervised learning or non-supervised learning policy.
Based on a network construction manner, the DNN may include a feedforward neural network (FNN), a convolutional neural network (CNN), and a recurrent neural network (recurrent neural network, RNN).
A generative neural network (generative neural network) is a special type of deep learning neural network. Different from a classification task and a prediction task that are mainly performed by a common neural network, the generative neural network may learn a probability distribution function that a group of training samples obey. Therefore, the generative neural network may be used to model a random variable, and may be used to establish a conditional probability distribution between variables. A common generative neural network includes a generative adversarial network (GAN) and a variational autoencoder (VAE).
The foregoing FNN, CNN, RNN, GAN, and VAE are neural network structures, and these network structures are constructed based on neurons.
Thanks to advantages of machine learning in modeling and information feature extraction, machine learning-based communication solutions may be designed, and good performance can be achieved. The solutions include but are not limited to CSI compression feedback, adaptive constellation point design, and/or robust precoding, and the like. In these solutions, a sending or receiving module in an original communication system is replaced with a neural network model, to optimize transmission performance or reduce processing complexity. To support different application scenarios, different neural network model information may be predefined or preconfigured, so that the neural network model can adapt to requirements of different scenarios.
In embodiments of this application, in a neural network-based communication system, some communication modules of an access network device and/or a terminal device may use the neural network model.
It should be noted that the neural network in embodiments of this application is not limited to any specific application scenario, but may be used in any communication scenario, such as CSI compressed feedback, adaptive constellation point design, and/or robust precoding.
Adaptive training is needed for the neural network to ensure communication performance. For example, for each different wireless system parameter (including one or more of a radio channel type, a bandwidth, a quantity of antennas at the receiving end, a quantity of antennas at the transmitting end, a modulation order, a quantity of paired users, a channel encoding method, and a code rate), a set of neural network model information (including neural network structure information and a neural network parameter) corresponding to the wireless system parameter is defined. An artificial intelligence (AI) --based adaptive modulation constellation design is used as an example. When there is one antenna at the transmitting end and two antennas at the receiving end, a set of neural network model information is needed to generate a corresponding modulation constellation. When there is one antenna at the transmitting end and four antennas at the receiving end, another set of neural network model information is needed to generate a corresponding modulation constellation. Similarly, different radio channel types, bandwidths, modulation orders, quantities of paired users, channel encoding methods, and/or code rates may correspond to different neural network model information.
To enable the neural networks (at the receiving end and the transmitting end) implement good performance in an actual channel condition, actual channel sample information needs to be used for joint training when the neural networks are trained. Usually, a larger quantity of channel sample information used during training indicates a better training effect. Sources of the channel sample information include: 1. obtained by the receiving end based on actual measurement; and 2. generated by using a mathematical (channel) model. Details are as follows.
1. A signal receiver (for example, UE) (for example, a reference signal or a synchronization signal) measures a real channel to obtain the channel sample information. The channel sample information can accurately reflect a channel environment. If a network element that is used to train a neural network by using the channel sample information is a signal transmitter, the receiver feeds back the channel sample information to the transmitter (for example, a base station) after obtaining the channel sample information through measurement. The transmitter trains the neural network based on the channel sample information. To improve a training effect and performance of a neural network obtained through training, the signal receiver needs to feed back a large quantity of channel sample information to the signal transmitter. However, feeding back a large quantity of channel sample information occupies a large quantity of air interface resources, and affects data transmission efficiency between the transmitter and the receiver.
2. In a possible implementation, for different channel types, a channel model may be modeled by using a mathematical expression. In other words, a mathematical (channel) model is used to simulate a real channel. For example, mathematical channel models such as a tapped delay line (TDL) and a cluster delay line (CDL) may be defined in a protocol. Each channel type may be further subdivided into a plurality of subtypes. For example, the TDL channel and the CDL channel each have five subtypes: A, B, C, D, and E. Each subtype is further subdivided into a plurality of typical channel scenarios based on specific parameters. For example, a subtype is subdivided into a plurality of typical scenarios such as 10 nanoseconds (ns), 30 ns, 100 ns, 300 ns, or 1000 ns based on a multipath delay interval. Therefore, when training the neural networks at the receiving end and the transmitting end, a mathematical channel model similar to an actual environment may be selected to generate a large quantity of channel sample information similar to an actual channel, and the channel sample information is used for training.
Although signaling overheads for obtaining a channel sample may be greatly reduced via the mathematical channel model, there is also a disadvantage that the mathematical channel model is mismatched with the actual channel model. For example, the TDL channel model assumes a limited quantity of reflection paths, and a channel coefficient of each path obey a simple Rayleigh distribution. However, the quantity of reflection paths of the actual channel varies in different environments, and the Rayleigh distribution cannot accurately describe a distribution of the channel coefficient of each path. In addition, a multipath delay interval varies with scenarios, and a rough division into several typical values inevitably leads to modeling errors. Therefore, it is difficult to accurately describe the actual channel model by using a mathematical channel modeling method, and a training effect of the neural network is affected. In other words, there is a problem of data model mismatch.
In conclusion, how to generate channel sample information close to an actual channel scenario on a premise of saving air interface resources becomes an urgent problem to be resolved currently.
The following describes the technical solutions provided in this application with reference to the accompanying drawings.
Optionally, 1001: A first device sends a first reference signal to a second device.
In this embodiment, an example in which the first device is an access network device and the second device is a terminal device is used for description. It may be understood that the first device may be the access network device or a chip in the access network device, and the second device may be the terminal device or a chip in the terminal device. Alternatively, the first device may be the terminal device or a chip in the terminal device, and the second device may be the access network device or a chip in the access network device. This is not limited herein.
The first device sends the first reference signal to the second device. The first reference signal may be a synchronization signal, a synchronization signal and PBCH block (SSB), a demodulation reference signal (DMRS), a channel state information reference signal (CSI-RS), or the like. The first reference signal may also be referred to as a first signal. Alternatively, the first reference signal may be a newly defined reference signal. This is not limited herein. A sequence type of the first reference signal includes but is not limited to a ZC (zadoff-chu) sequence or a gold sequence. Optionally, the DMRS may be a physical downlink control channel (PDCCH) DMRS or a physical downlink shared channel (PDSCH) DMRS.
Optionally, before step 1001, when the second device needs to use a neural network (where the neural network used by the second device is referred to as a third neural network), the second device sends capability information to the first device. The capability information indicates one or more of the following information about the second device.
(1) Whether to support using the neural network to replace or implement a function of a communication module.
The communication module includes but is not limited to an OFDM modulation module, an OFDM demodulation module, a constellation mapping module, a constellation demapping module, a channel encoding module, a channel decoding module, a precoding module, an equalization module, an interleaving module, and/or a de-interleaving module.
(2) Whether to support a network type of the third neural network, or indicate a network type of a supported third neural network.
The network type of the third neural network includes one or more of the following: a fully connected neural network, a radial basis function neural network, a convolutional neural network, a recurrent neural network, a Hopfield neural network, a restricted Boltzmann machine, a deep belief network, or the like; and the third neural network may be any one of the foregoing neural networks, or the third neural network may be a combination of the foregoing neural networks, and this is not limited herein.
(3) Whether to support receiving information about the third neural network using signaling.
(4) Stored preconfigured third neural network.
For example, if the predefined or preconfigured third neural network already exists in the second device, the capability information may carry an identifier or an index of the third neural network.
Optionally, the capability information may further carry an identifier or an index of another predefined or preconfigured neural network that already exists in the second device.
(5) Memory space for storing the third neural network.
(6) Computing power information that may be used to run the neural network.
The computing power information is computing power information for running the neural network, for example, including information such as a computing speed of a processor and/or an amount of data that can be processed by the processor.
Optionally, 1002: The second device performs channel estimation based on the first reference signal received from the first device, to determine first channel sample information.
In this embodiment, after receiving the first reference signal, the second device performs channel estimation on the first reference signal, to determine the first channel sample information.
Specifically, for ease of description, a signal received at the second device is referred to as a second reference signal, where the signal is the first reference signal that undergoes transmission over a channel. It is assumed that the first reference signal is x and the second reference signal is y, and the channel for propagating the first reference signal may be understood as a function having a transition probability P(y|x).
The second device may preconfigure information about the first reference signal for the first device using signaling, or may agree on information about the first reference signal in a protocol. This is not limited. In this embodiment of this application, a type of the signaling is not limited, for example, may be a broadcast message, system information, radio resource control (RRC) signaling, a media access control (MAC) control element (CE), or downlink control information (DCI).
Because the second device already knows in advance the information about the first reference signal x sent by the first device, the second device performs channel estimation based on the second reference signal y and the sent first reference signal x after receiving the second reference signal y, to determine the first channel sample information. In this case, the channel experienced by the first reference signal from being sent to being received may be estimated, for example, an amplitude change and a phase rotation experienced by the first reference signal. The first channel sample information includes but is not limited to the second reference signal and/or channel state information (CSI).
A channel estimation algorithm used when the second device performs channel estimation includes but is not limited to a least square (LS) estimation method or a linear minimum mean square error (LMMSE) estimation method.
1003: The second device sends the first channel sample information to the first device.
In this embodiment, the second device sends the first channel sample information to the first device.
1004: The first device obtains a first neural network through training based on the first channel sample information.
In this embodiment, the first device obtains the first neural network through training based on the first channel sample information. The first neural network is used to generate new channel sample information.
Optionally, the first neural network is a generative neural network. In this embodiment of this application, an example in which the first neural network is a GAN is used for description. It may be understood that the first neural network may be another type of neural network, for example, a VAE. This is not limited herein. The first neural network includes a generator network and a discriminator network based on function division. The generator network is used to generate second channel sample information, and the discriminator network is used to determine whether the newly generated second channel sample information and the first channel sample information from the second device obey a similar probability statistical feature. The generator network and the discriminator network of the first neural network are jointly trained by using a machine learning method, so that the second channel sample information output by the generator network may be converged to a probability distribution of the first channel sample information from the second device. The following separately provides examples for description.
For example, when the first channel sample information is CSI (denoted CSI as h), for a schematic structural diagram of the first neural network, refer to
For example, when the first channel sample information is the second reference signal (y) described above, for a schematic structural diagram of the first neural network, refer to
In the first neural network shown in
It is assumed that the first reference signal sent by the first device is x, and after the first reference signal is propagated through a channel, the second reference signal received by the second device is y. In this case, the channel may be understood as a function of a transition probability P(y|x). The first neural network expects to learn a probability transition feature of the channel, and for each input x, an output y can be generated with a probability of P(y|x). In other words, the first neural network simulates a process in which a signal is propagated through the channel.
Specifically, the first device uses first channel sample information to train the reference neural network. A neural network obtained after the training is completed is referred to as the first neural network. In embodiments of this application, for ease of description, a neural network used before the first channel sample information is used for training may be referred to as the reference neural network. For example, the first device trains the reference neural network by using first channel sample information that is from the second device and that includes 5000 samples, to obtain the first neural network through training.
In an optional implementation, the reference neural network may preconfigure parameters of some neurons or a structure of the neural network. The reference neural network may further preconfigure information about another neural network. This is not limited herein. For example, the reference neural network may be a neural network that is obtained through training in a predefined channel environment and that is used to generate channel sample information (for example, the CSI or the second reference signal). The first device further trains the reference neural network by using the first channel sample information, to obtain the first neural network through training.
In another optional implementation, the reference neural network may be an initialized neural network. For example, a parameter of each neuron in the reference neural network is zero, a random value, or another agreed value, and/or a structure of the neural network is a random or an agreed initial structure. This is not limited herein. The first device starts to train the reference neural network by using the first channel sample information, to obtain the first neural network through training.
1005: The first device performs inference to obtain the second channel sample information by using the first neural network.
In this embodiment, after obtaining the trained first neural network, the first device performs inference to obtain the second channel sample information by using the first neural network.
For example, the first device inputs random noise into the first neural network, and the second channel sample information may be obtained through inference by using the first neural network. For example, when the first neural network is a GAN, the random noise is input into the generator network of the first neural network, to obtain the second channel sample information. For another example, when the first neural network is a VAE, the random noise is input into a decoder network of the first neural network, to obtain the second channel sample information.
A form of the second channel sample information is the same as a form of the first channel sample information. For example, when the first channel sample information is the CSI (h), the second channel sample information is CSI (h̃). When the first channel sample information is the second reference signal (y), the second channel sample information is a second reference signal (ỹ).
Optionally, 1006: The first device trains a second neural network and/or the third neural network based on the second channel sample information.
For example, when the first channel sample information is the CSI (denoted the CSI as h), for a schematic structural diagram of the first device training the second neural network and/or the third neural network based on the second channel sample information, refer to
For example, when the first channel sample information is the second reference signal (y) described above, for a schematic structural diagram of the first device training the second neural network and/or the third neural network based on the second channel sample information, refer to
In this embodiment, after generating the second channel sample information based on the first neural network, the first device trains the second neural network and/or the third neural network based on the second channel sample information. The second neural network and/or the third neural network are/is used for transmission of target information between the first device and the second device.
Optionally, the first device trains the second neural network and/or the third neural network based on the first channel sample information and the second channel sample information.
The following describes the second neural network and the third neural network separately.
In this embodiment of this application, that the second neural network is used in the transmitter may be understood as that the second neural network performs data sending processing, for example, the constellation mapping neural network shown in
A network type of the second neural network includes but is not limited to: a fully connected neural network, a radial basis function (RBF) neural network, a CNN, a recurrent neural network, a Hopfield neural network, a restricted Boltzmann machine, a deep belief network, or the like. The second neural network may be any one of the foregoing neural networks, or the second neural network may be a combination of the foregoing neural networks. This is not limited herein.
In this embodiment of this application, for ease of description, the first device is used as the transmitter for description, in other words, the second neural network is used in the first device. It may be understood that the second device may also be used as the transmitter, in other words, the second neural network is used in the second device.
In this embodiment of this application, that the third neural network is used in the receiver may be understood as that the third neural network performs receiving processing on data from the transmitter, for example, the constellation demapping neural network shown in
A network type of the third neural network includes but is not limited to: a fully connected neural network, a radial basis function neural network, a convolutional neural network, a recurrent neural network, a Hopfield neural network, a restricted Boltzmann machine, a deep belief network, or the like. The third neural network may be any one of the foregoing neural networks, or the third neural network may be a combination of the foregoing neural networks. This is not limited herein.
It should be noted that the second neural network and the third neural network may use a same neural network, may use different neural networks, or may use partially same neural networks. This is not limited herein.
In this embodiment of this application, for ease of description, the second device is used as the receiver for description, in other words, the third neural network is used in the second device. It may be understood that the first device may alternatively be used as the receiver, in other words, the third neural network is used in the first device.
The second neural network and/or the third neural network are/is used for transmission of target information between the first device and the second device. Based on the foregoing descriptions, it may be understood that when the second neural network and the third neural network perform different processing, the target information is different.
For example, when the second neural network is a constellation mapping neural network, and the third neural network is a constellation demapping neural network, the target information is a modulation symbol sequence.
For example, when the second neural network is a CSI compression neural network, and the third neural network is a CSI decompression neural network, the target information is compressed CSI or CSI.
For example, when the second neural network is a channel encoding neural network, and the third neural network is a channel decoding neural network, the target information is to-be-sent bit information before encoding or to-be-sent bit information after encoding.
For example, when the second neural network is an OFDM modulation neural network, and the third neural network is an OFDM demodulation neural network, the target information is an OFDM symbol.
For example, when the second neural network is a precoding neural network, and the third neural network is an equalizer neural network, the target information is a sending signal, and the sending signal arrives at the receiver after being transmitted through a channel.
The first device may independently train the second neural network, independently train the third neural network, or jointly train the second neural network and the third neural network.
For example, when the second neural network is a precoding neural network at the transmitter (for example, the first device), the first device may independently train the second neural network. A function of the precoding neural network is to generate a transmitting weighting coefficient of each transmitting antenna based on a spatial feature of a channel, so that a plurality of transmitting data streams may be spatially isolated, and a receiving signal to interference plus noise ratio of a signal may be improved. An input of the precoding neural network is CSI of a channel, and an output of the precoding neural network is a precoding weight on each transmitting antenna. In other words, the target information of the second neural network is the precoding weight on each transmitting antenna.
For example, when the third neural network is an equalizer neural network at the receiver (for example, the second device), the first device may independently train the third neural network. A function of the equalizer neural network is to cancel a distortion effect caused by a channel to a propagated signal, and recover a signal transmitted by the transmitter. An input of the equalizer neural network is CSI of a channel and a second reference signal at the receiver, and an output of the equalizer neural network is a signal recovered after equalization. In other words, the target information of the third neural network is the signal recovered after equalization.
The first device may independently train the second neural network by using the second channel sample information generated by the first neural network, or may independently train the third neural network by using the second channel sample information and locally generated sending data, and send information about a trained third neural network to the second device for use. The first device may further jointly train the second neural network and the third neural network, to implement optimal end-to-end communication performance.
After completing the training of the third neural network, the first device may perform step 1007. The completion of the training of the third neural network includes but is not limited to: A quantity of training samples for training reaches a preset threshold. For example, if a quantity of second channel sample information for training the third neural network reaches 45000, it is considered that the training of the third neural network is completed. Alternatively, if a quantity of times of training reaches a preset threshold. For example, if a quantity of times of training the third neural network reaches 30000, it is considered that the training of the third neural network is completed.
Optionally, 1007: The first device sends the information about the third neural network to the second device.
In this embodiment, after completing the training of the third neural network, the first device sends the information about the third neural network to the second device. The information about the third neural network includes but is not limited to: a weight of a neural network, an activation function of a neuron, a quantity of neurons at each layer of the neural network, an inter-layer cascading relationship of the neural network, and/or a network type of each layer of the neural network. For example, the information about the third neural network may indicate a same activation function or different activation functions for different neurons. This is not limited.
Optionally, when the third neural network is preconfigured (or predefined) in the second device, the information about the third neural network may alternatively be a model variation of the third neural network. The model variation includes but is not limited to: a weight of a changed neural network, a changed activation function, a quantity of neurons at each layer of the changed neural network, an inter-layer cascading relationship of the changed neural network, and/or a network type of each layer of the changed neural network. For example, the pre-configuration may be performed on the terminal device by the access network device using signaling, and the pre-definition may be performed in a protocol. For example, the third neural network in the terminal device is predefined as a neural network A in the protocol.
The first device sends the information about the third neural network to the second device using signaling.
In this embodiment of this application, the signaling includes but is not limited to a broadcast message (for example, a master information block (MIB)), a system message (for example, a system information block (SIB)), radio resource control (RRC) signaling, a medium access control control element (MAC CE), and/or downlink control information (DCI). The MAC CE and/or the DCI may be a common message/common messages of a plurality of devices, or may be a specific message/specific messages specific to a device (for example, a second device).
The second device uses the third neural network in the second device based on the information about the third neural network.
After completing the training of the second neural network, the first device uses the second neural network.
When the second neural network is preconfigured (or predefined) in the first device, the first device updates the local second neural network based on the training result in step 1006. When the third neural network is preconfigured in the second device, the second device updates the local third neural network based on the information about the third neural network in step 1007. When there is no third neural network in the second device, the information about the third neural network in step 1007 may include a complete third neural network. The second device locally configures the trained third neural network based on the information about the third neural network.
Then, if the target information is downlink information, the first device may process the target information (namely, an input of the second neural network) or obtain the target information (namely, an output of the second neural network) by using the second neural network. The second device may process the target information (namely, an input of the third neural network) or restore the target information (namely, an output of the third neural network) by using the third neural network. For example, the first device obtains a modulation symbol by using the second neural network, and the second device recovers a bit from the received modulation symbol by using the third neural network.
If the target is uplink information, the second device may process the target information (namely, an input of the third neural network) or obtain the target information (namely, an output of the third neural network) by using the third neural network. The first device may process the target information (namely, an input of the second neural network) or restore the target information (namely, an output of the second neural network) by using the second neural network. For example, the second device compresses CSI by using the third neural network, to obtain the compressed CSI. The first device recovers the CSI from the received compressed CSI by using the second neural network.
In this embodiment of this application, the second device (for example, the terminal device) may send a small quantity of first channel sample information to the first device (for example, the access network device). The first device obtains the first neural network through training based on the first channel sample information. The first neural network is used to perform inference to obtain the second channel sample information. The first device trains the second neural network and/or the third neural network based on the second channel sample information or based on the second channel sample information and the first channel sample information. The second neural network and/or the third neural network are/is used for transmission of target information between the first device and the second device. According to this method, air interface signaling overheads can be effectively reduced, and a channel environment in which a trained neural network is located can be adapted. The second neural network and the third neural network obtained through training are closer to an actual channel environment, and communication performance is improved. A speed of training the second neural network and the third neural network is also greatly improved.
Similar to
Optionally, 1101: A first device sends a first reference signal to a second device.
1102: The second device performs channel estimation based on the first reference signal received from the first device, to determine first channel sample information.
Optionally, 1103: The second device sends the first channel sample information to the first device.
Steps 1101 to 1103 are the same as the foregoing steps 1001 to 1003, and details are not described herein again.
1104. The first device sends the first channel sample information to a third device.
In this embodiment, after receiving the first channel sample information from the second device, the first device sends the first channel sample information to the third device.
1105: The third device obtains a first neural network through training based on the first channel sample information.
1106: The third device performs inference to obtain second channel sample information by using the first neural network.
Optionally, 1106 may be replaced with: The third device sends information about the first neural network to the first device. The first device performs inference to obtain the second channel sample information by using the first neural network. Optionally, the first device sends the second channel sample information to the third device.
1107: The third device trains a second neural network and/or a third neural network based on the second channel sample information.
Steps 1105 to 1107 are similar to the foregoing steps 1004 to 1006, and details are not described herein again.
1108: The third device sends information about the second neural network and/or information about the third neural network to the first device.
In this embodiment, after obtaining the third neural network through training, the third device may send the information about the third neural network to the second device by using the first device. Optionally, the third device may send the information about the third neural network to the first device by using another device (for example, another access network device). Optionally, if a direct communication link exists between the third device and the first device, the third device may further directly send the information about the third neural network to the first device.
After obtaining the second neural network through training, the third device may send the information about the second neural network to the second device.
1109: The first device sends the information about the third neural network to the second device.
Step 1109 is the same as step 1007, and details are not described herein again.
Optionally, the embodiment shown in
In this embodiment of this application, training a neural network in the foregoing manner, air interface signaling overheads are effectively reduced, and a channel environment in which a trained neural network is located can be adapted. The second neural network and the third neural network obtained through training are closer to an actual channel environment, and communication performance is improved. A speed of training the second neural network and the third neural network is also greatly improved. Steps related to neural network training may be performed by another device (the third device), to effectively reduce computing load and power consumption of the first device (for example, an access network device).
Similar to
Optionally, 1201: The second device sends capability information to a first device.
In this embodiment, similar to the embodiments shown in
In embodiments of this application, for ease of description, a first neural network that is not trained by using first channel sample information or a first neural network that is not updated may be referred to as a reference neural network. After the reference neural network is trained by using the first channel sample information, the first neural network is obtained.
The reference neural network is similar to the reference neural network in step 1004. The second device starts to train the reference neural network by using the first channel sample information, to obtain the first neural network through training.
In step 1201, the second device sends the capability information to the first device. The capability information indicates one or more of the following information:
(1) Whether to support using the neural network to replace or implement a function of a communication module.
The communication module includes but is not limited to an OFDM modulation module, an OFDM demodulation module, a constellation mapping module, a constellation demapping module, a channel encoding module, a channel decoding module, a precoding module, an equalization module, an interleaving module, and/or a de-interleaving module.
(2) Whether to support a network type of the first neural network, or indicate a network type of a supported first neural network.
For example, the capability information indicates whether the second device supports a VAE or whether the second device supports a GAN. For another example, the capability information indicates that the second device supports the GAN, supports the VAE, supports the GAN and the VAE, or does not support the GAN or the VAE.
(3) Whether to support a network type of a third neural network, or indicate a network type of a supported third neural network.
The network type of the third neural network includes one or more of the following: a fully connected neural network, a radial basis function neural network, a convolutional neural network, a recurrent neural network, a Hopfield neural network, a restricted Boltzmann machine, a deep belief network, or the like; and the third neural network may be any one of the foregoing neural networks, or the third neural network may be a combination of the foregoing neural networks, and this is not limited herein.
(4) Whether to support receiving information about the reference neural network using signaling, where the reference neural network is the first neural network not trained by using the first channel sample information or is used to train the first neural network.
(5) Whether to support receiving information about the third neural network using signaling.
(6) Stored preconfigured reference neural network.
For example, if the predefined or preconfigured reference neural network already exists in the second device, the capability information may carry an identifier or an index of the reference neural network.
Optionally, the capability information may further carry an identifier or an index of another predefined or preconfigured neural network that already exists in the second device.
(7) Memory space for storing the first neural network and/or the third neural network.
(8) Computing power information that may be used to run the neural network.
The computing power information means computing power information for running the neural network, for example, includes information such as a computing speed of a processor and/or an amount of data that can be processed by the processor.
(9) Location information of the second device.
For example, the longitude, the latitude, and/or the altitude of the second device are reported, and for another example, a channel environment of the second device is reported, for example, an office, a subway station, a park, or a street.
The capability information includes but is not limited to the foregoing information. For example, the capability information may further include a type of channel sample information that is supported by the second device and that is used to obtain the first neural network through training. For example, the channel sample information that is supported by the second device (the terminal device) and that is used to obtain the first neural network through training is CSI, or the channel sample information that is supported by the second device and that is used to obtain the first neural network through training is a second reference signal (y).
Step 1201 is an optional step. When step 1201 is performed, the second device may send the capability information when accessing the first device for the first time, or the second device may periodically send the capability information to the first device. This is not limited herein.
Optionally, after the first device receives the capability information from the second device, and when the capability information indicates that no reference neural network is configured (stored) in the second device, the first device may send the information about the reference neural network to the second device. The information about the reference neural network includes: a weight of a neural network, an activation function of a neuron, a quantity of neurons at each layer of the neural network, an inter-layer cascading relationship of the neural network, and/or a network type of each layer of the neural network. The information about the reference neural network may be referred to as specific information about the reference neural network. Activation functions of different neurons may be the same, or may be different. This is not limited.
Optionally, 1202: The first device sends a first reference signal to the second device.
This is the same as the foregoing step 1001, and details are not described herein again.
Optionally, 1203: The second device performs channel estimation based on the first reference signal received from the first device, to determine the first channel sample information.
Same as step 1002, details are not described herein again.
1204: The second device obtains the first neural network through training based on the first channel sample information.
The method in which the second device obtains the first neural network through training based on the first channel sample information is similar to the method in 1004 in which the first device obtains the first neural network through training based on the first channel sample information. Details are not described herein.
For example, the second device obtains the first neural network through training based on the first channel sample information and the reference neural network.
Optionally, the second device may determine the reference neural network in any one of the following manners 1 to 3.
Manner 1: The information about the reference neural network is agreed in a protocol, or the second device determines the information about the reference neural network.
In a possible implementation, the information about a reference neural network is agreed (predefined) in the protocol, in other words, the specific information about the reference neural network is agreed.
In embodiments of this application, the information about the reference neural network includes: a weight of a neural network, an activation function of a neuron, a quantity of neurons at each layer of the neural network, an inter-layer cascading relationship of the neural network, and/or a network type of each layer of the neural network. The information about the reference neural network may further be referred to as the specific information about the reference neural network. Activation functions of different neurons may be the same, or may be different. This is not limited.
In a possible implementation, information about a plurality of reference neural networks is agreed in the protocol. Each reference neural network corresponds to a channel environment. The channel environment includes one or more of the following: an office, an indoor site of a shopping mall, a subway station, a city street, a square, a suburb, and the like. As described above, the second device reports the location information of the second device to the first device, and the first device may determine the channel environment in which the second device is located, to obtain a reference neural network corresponding to the channel environment. For example, the access network device may maintain one piece of map information, and corresponding channel environments corresponding to different location information are marked in the map information. For example, the map information is shown in Table 1:
Optionally, if the second device reports the location information of the second device, namely, the channel environment of the second device, to the first device by using the capability information, the first device may determine the information about the reference neural network corresponding to the channel environment.
In a possible implementation, the second device determines the information about the reference neural network, in other words, determines the specific information about the reference neural network. As described above, the second device may report the information about the reference neural network to the first device by using the capability information.
Manner 2: The reference neural network is indicated by the first device for the second device.
In a possible implementation, the first device sends the information about the reference neural network to the second device, in other words, the first device sends the specific information about the reference neural network to the second device.
In a possible implementation, information about a plurality of candidate reference neural networks is agreed in the protocol, and each candidate reference neural network corresponds to a unique index. The first device sends an identifier to the second device using signaling. The identifier indicates an index of a reference neural network that is of the plurality of candidate reference neural networks and that is configured for the second device.
In a possible implementation, the first device configures the information about the plurality of candidate reference neural networks for the second device by using first signaling. In other words, the first device configures specific information about the plurality of candidate reference neural networks. Each candidate reference neural network corresponds to a unique index. The first device sends an identifier to the second device by using second signaling. The identifier indicates an index of a reference neural network that is of the plurality of candidate reference neural networks and that is configured for the second device.
In a possible implementation, as described above, the second device may send, to the first device by using the capability information, information about the plurality of candidate reference neural networks stored or owned by the second device. Each candidate reference neural network corresponds to a unique index. The first device sends an identifier to the second device using signaling. The identifier indicates an index of a reference neural network that is of the plurality of candidate reference neural networks and that is configured for the second device.
In a possible implementation, the first device sends a type of the reference neural network to the second device using signaling. Information about various types of reference neural networks is agreed in the protocol, or is determined by the second device independently.
1205: The second device sends information about the first neural network to the first device.
After obtaining the first neural network through training, the second device sends the information about the first neural network to the first device.
Optionally, the information about the first neural network may be specific information about the first neural network. The specific information may include: a weight of a neural network, an activation function of a neuron, a quantity of neurons at each layer of the neural network, an inter-layer cascading relationship of the neural network, and/or a network type of each layer of the neural network. Activation functions of different neurons may be the same, or may be different. This is not limited.
Optionally, the information about the first neural network may alternatively be a model variation of the first neural network relative to the reference neural network. The model variation includes but is not limited to: a weight of a changed neural network, a changed activation function, a quantity of neurons at one or more layers of the changed neural network, an inter-layer cascading relationship of the changed neural network, and/or a network type of one or more layers of the changed neural network. When the information about the first neural network is the model variation, air interface signaling overheads can be effectively reduced.
1206: The first device determines the first neural network based on the information about the first neural network.
The first device determines the first neural network based on the information about the first neural network from the second device.
Optionally, when the information about the first neural network is the specific information about the first neural network, the first device determines the first neural network based on the specific information.
Optionally, when the information about the first neural network is the model variation, the first device determines the first neural network based on the model variation and the reference neural network.
1207: The first device performs inference to obtain second channel sample information based on the first neural network.
1208: The first device trains a second neural network and/or the third neural network based on the second channel sample information.
1209: The first device sends the information about the third neural network to the second device.
Steps 1207 to 1209 are the same as the foregoing steps 1005 to 1007, and details are not described herein again.
In this embodiment, the second device obtains the first neural network through training based on the first channel sample information and the reference neural network.
In this embodiment, the second device (for example, the terminal device) obtains the first neural network through training based on the first channel sample information. The first neural network is used to perform inference to obtain the second channel sample information. The first device trains the second neural network and/or the third neural network based on the second channel sample information. The second neural network and/or the third neural network are/is used for transmission of target information between the first device and the second device. Air interface signaling overheads are effectively reduced, and a channel environment in which a trained neural network is located can be adapted. The second neural network and the third neural network obtained through training are closer to an actual channel environment, and communication performance is improved. A speed of training the second neural network and the third neural network is also greatly improved. A computing task for training the first neural network is allocated to the second device, to effectively reduce computing burden of the first device (for example, the access network device), and improve communication performance.
In the foregoing embodiments provided in this application, the methods provided in embodiments of this application are separately described from perspectives of the first device, the second device, the third device, and interaction between the first device, the second device, and the third device. To implement functions in the methods provided in embodiments of this application, the first device, the second device, and the third device may include a hardware structure and/or a software module, and implement the foregoing functions in a form of a hardware structure, a software module, or a hardware structure plus a software module. Whether a function in the foregoing functions is performed by using a hardware structure, a software module, or a hardware structure plus a software module depends on a specific application and a design constraint condition of the technical solution.
Optionally, the communication apparatus 1700 is configured to implement the function of the first device in the foregoing method embodiments.
In a possible implementation, the transceiver module 1710 is configured to send a first reference signal to a second device.
The transceiver module 1710 is further configured to receive first channel sample information from the second device.
The processing module 1720 is configured to determine a first neural network. The first neural network is obtained through training based on the first channel sample information, and is used to perform inference to obtain second channel sample information. The second channel sample information is used to train a second neural network and/or a third neural network, and the second neural network and/or the third neural network are/is used for transmission of target information between the first device and the second device.
In a possible implementation, the transceiver module 1710 is configured to send the first reference signal to the second device.
The transceiver module 1710 is further configured to receive information about the first neural network from the second device.
The processing module 1720 is configured to determine the second neural network and/or the third neural network. The second neural network and/or the third neural network are/is used for transmission of target information between the first device and the second device, and are/is obtained through training based on the second channel sample information; and the second channel sample information is obtained through inference by using the first neural network.
Optionally, the communication apparatus 1700 is configured to implement the function of the second device in the foregoing method embodiments.
In a possible implementation, the processing module 1720 is configured to perform channel estimation based on the first reference signal received from the first device, to determine the first channel sample information.
The transceiver module 1710 is configured to send the first channel sample information to the first device.
The transceiver module 1710 is further configured to receive information about the third neural network from the first device. The third neural network is used for transmission of target information between the first device and the second device.
In a possible implementation, the processing module 1720 is configured to perform channel estimation based on the first reference signal received from the first device, to determine the first channel sample information.
The processing module 1720 is further configured to determine the first neural network. The first neural network is obtained through training based on the first channel sample information.
The transceiver module 1710 is configured to send the information about the first neural network to the first device.
Optionally, the communication apparatus 1700 is configured to implement the function of the third device in the foregoing method embodiments.
In a possible implementation, the transceiver module 1710 is configured to receive the first channel sample information from the first device.
The processing module 1720 is configured to determine the first neural network. The first neural network is obtained through training based on the first channel sample information, and is used to perform inference to obtain the second channel sample information. The second channel sample information is used to train the second neural network and/or the third neural network, and the second neural network and/or the third neural network are/is used for transmission of target information between the first device and the second device.
Optionally, the communication apparatus may further include a storage unit. The storage unit may store data and/or instructions (or may be referred to as code or a program). The foregoing units may interact with or be coupled to the storage unit, to implement a corresponding method or function. For example, the processing unit 1720 may read data or instructions in the storage unit, so that the communication apparatus implements the methods in the foregoing embodiments. The coupling in embodiments of this application may be an indirect coupling or a communication connection between apparatus, units, or modules in an electrical form, a mechanical form, or another form, and is used for information exchange between the apparatus, the units, or the modules.
Division into the modules in embodiments of this application is an example, is merely division into logical functions, and may be other division during actual implementation. In addition, functional modules in embodiments of this application may be integrated into one processor, or each of the modules may exist alone physically, or two or more modules may be integrated into one module. The integrated module may be implemented in a form of hardware, or may be implemented in a form of a software functional module.
In an example, the units in the foregoing communication apparatus may be one or more integrated circuits configured to implement the foregoing methods, for example, one or more application-specific integrated circuits (ASICs), or one or more microprocessors (DSPs), or one or more field programmable gate arrays (FPGAs), or a combination of at least two of these integrated circuit forms. For another example, when the units in the communication apparatus may be implemented in a form of scheduling a program by a processing element, the processing element may be a general-purpose processor, for example, a central processing unit (CPU) or another processor that may invoke the program. For still another example, the units may be integrated and implemented in a form of a system-on-a-chip (SOC).
All or a part of the technical solutions provided in embodiments of this application may be implemented by using software, hardware, firmware, or any combination thereof. When software is used to implement the embodiments, all or a part of the embodiments may be implemented in a form of a computer program product. The computer program product includes one or more computer instructions. When the computer program instructions are loaded and executed on the computer, the procedure or functions according to embodiments of the present invention are all or partially generated. The computer may be a general-purpose computer, a dedicated computer, a computer network, an AI node, an access network device, a terminal device, or another programmable apparatus. The computer instructions may be stored in a computer-readable storage medium or may be transmitted from a computer-readable storage medium to another computer-readable storage medium. For example, the computer instructions may be transmitted from a website, computer, server, or data center to another website, computer, server, or data center in a wired (for example, a coaxial cable, an optical fiber, or a digital subscriber line (DSL)) or wireless (for example, infrared, radio, or microwave) manner. The computer-readable storage medium may be any usable medium accessible by the computer, or a data storage device such as a server or a data center, integrating one or more usable media. The usable medium may be a magnetic medium (for example, a floppy disk, a hard disk, or a magnetic tape), an optical medium (for example, a digital video disc (DVD)), a semiconductor medium, or the like.
In embodiments of this application, when there is no logical conflict, embodiments may be mutually referenced. For example, methods and/or terms in the method embodiments may be mutually referenced, and functions and/or terms in the apparatus embodiments may be mutually referenced. For example, functions and/or terms between the apparatus embodiments and the method embodiments may be mutually referenced.
It is clearly that, a person skilled in the art may make various modifications and variations to this application without departing from the scope of this application. This application is intended to cover these modifications and variations of this application provided that they fall within the scope of protection defined by the following claims and their equivalent technologies.
This application is a continuation of International Application No. PCT/CN2020/142103, filed on Dec. 31, 2020, the disclosure of which is hereby incorporated by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/CN2020/142103 | Dec 2020 | WO |
Child | 18345904 | US |