WO2023123429A1 - Method and device for training channel information feedback model, apparatus, and storage medium - Google Patents

Method and device for training channel information feedback model, apparatus, and storage medium Download PDF

Info

Publication number
WO2023123429A1
WO2023123429A1 PCT/CN2021/143874 CN2021143874W WO2023123429A1 WO 2023123429 A1 WO2023123429 A1 WO 2023123429A1 CN 2021143874 W CN2021143874 W CN 2021143874W WO 2023123429 A1 WO2023123429 A1 WO 2023123429A1
Authority
WO
WIPO (PCT)
Prior art keywords
information
encoder
channel information
transfer learning
decoder
Prior art date
Application number
PCT/CN2021/143874
Other languages
French (fr)
Chinese (zh)
Inventor
李德新
田文强
刘文东
肖寒
Original Assignee
Oppo广东移动通信有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Oppo广东移动通信有限公司 filed Critical Oppo广东移动通信有限公司
Priority to CN202180103608.5A priority Critical patent/CN118120156A/en
Priority to PCT/CN2021/143874 priority patent/WO2023123429A1/en
Publication of WO2023123429A1 publication Critical patent/WO2023123429A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04BTRANSMISSION
    • H04B7/00Radio transmission systems, i.e. using radiation field
    • H04B7/02Diversity systems; Multi-antenna system, i.e. transmission or reception using multiple antennas
    • H04B7/04Diversity systems; Multi-antenna system, i.e. transmission or reception using multiple antennas using two or more spaced independent antennas
    • H04B7/06Diversity systems; Multi-antenna system, i.e. transmission or reception using multiple antennas using two or more spaced independent antennas at the transmitting station

Definitions

  • the present application relates to the field of communication technology, and in particular to a training method, device, equipment and storage medium of a channel information feedback model.
  • the terminal device generally generates channel information through Channel State Information (CSI) measurement, and feeds the channel information back to the network device.
  • CSI Channel State Information
  • the channel information is regarded as an image to be compressed
  • the encoder is used to compress and feed back the channel information
  • the decoder is used at the sending end to reconstruct the compressed channel information , the channel information can be preserved to a greater extent.
  • Embodiments of the present application provide a training method, device, equipment, and storage medium for a channel information feedback model. Described technical scheme is as follows:
  • a method for training a channel information feedback model is provided, which is applied to a source-side terminal, and the method includes:
  • the channel information feedback model is trained based on an error between the restored channel information and the initial channel information.
  • a method for training a channel information feedback model is provided, which is applied to a target-side terminal, where the channel information feedback model includes: a second encoder and a second decoder, and the method includes:
  • the second transfer learning information is used to assist in transfer learning
  • the second transfer learning information includes: matrix size information corresponding to the second encoder and mask operation, The second encoder is obtained by training based on the mask operation;
  • a method for training a channel information feedback model is provided, which is applied to a network device, and the method includes:
  • the second transfer learning information is used to assist transfer learning
  • the second transfer learning information includes: matrix size information corresponding to the second encoder and mask operation, the The second encoder is obtained by training based on the mask operation;
  • a training device for a channel information feedback model comprising: a mask module, a model processing module and a training module;
  • the masking module is configured to perform a masking operation on initial channel information to obtain masked channel information
  • the model processing module is configured to input the masked channel information into the channel information feedback model, and output restored channel information;
  • the training module is configured to train the channel information feedback model based on the error between the recovered channel information and the initial channel information.
  • a training device for a channel information feedback model includes: a second encoder and a second decoder, and the device includes: a decoder generating module and an information receiving module , a training module and a decoder sending module;
  • the decoder generating module is configured to generate the second decoder
  • the information receiving module is configured to receive second transfer learning information sent by a network device, the second transfer learning information is used to assist transfer learning, and the second transfer learning information includes: the second encoder and mask Matrix size information corresponding to the code operation, the second encoder is obtained by training based on the mask operation;
  • the training module is configured to jointly train the second encoder and the second decoder based on the matrix size information
  • the decoder sending module is configured to send the trained second decoder to the network device.
  • a training device for a channel information feedback model comprising: an information sending module and a decoder receiving module;
  • the information sending module is configured to send second transfer learning information to the target terminal, the second transfer learning information is used to assist transfer learning, and the second transfer learning information includes: a second encoder and a mask operation Corresponding matrix size information, the second encoder is obtained by training based on the mask operation;
  • the decoder receiving module is configured to receive the second decoder sent by the target terminal, the second decoder is obtained by training after the target terminal performs transfer learning based on the second transfer learning information .
  • a terminal device includes: a processor; wherein,
  • the processor is configured to perform a masking operation on initial channel information to obtain masked channel information
  • the processor is configured to input the masked channel information into a channel information feedback model, and output restored channel information;
  • the processor is configured to train the channel information feedback model based on an error between the recovered channel information and the initial channel information.
  • a terminal device includes: a processor and a transceiver connected to the processor; wherein,
  • the processor configured to generate a second decoder
  • the transceiver is configured to receive second transfer learning information sent by a network device, the second transfer learning information is used to assist transfer learning, and the second transfer learning information includes: the second encoder and a mask Operating the corresponding matrix size information, the second encoder is obtained by training based on the mask operation;
  • the processor is configured to jointly train the second encoder and the second decoder based on the matrix size information
  • the transceiver is configured to send the trained second decoder to the network device.
  • a network device includes: a transceiver; wherein,
  • the transceiver is configured to send second transfer learning information to the target-side terminal, the second transfer learning information is used to assist in transfer learning, and the second transfer learning information includes: a second encoder corresponding to a mask operation The matrix size information of the second encoder is obtained by training based on the mask operation;
  • the transceiver is configured to receive the second decoder sent by the target-side terminal, where the second decoder is trained by the target-side terminal after performing transfer learning based on the second transfer learning information.
  • a computer-readable storage medium is provided, and executable instructions are stored in the readable storage medium, and the executable instructions are loaded and executed by a processor to implement the channel described in the above aspect Training methods for information feedback models.
  • a chip is provided, the chip includes a programmable logic circuit and/or program instructions, and when the chip is run on a computer device, it is used to realize the channel information described in the above aspect Feedback model training method.
  • a computer program product is provided.
  • the computer program product When the computer program product is run on a processor of a computer device, the computer device executes the method for training a channel information feedback model described in the above aspect.
  • model training use the mask operation to shield part of the initial channel information, reduce the redundant information input during the channel information feedback model training, and reduce the resources for model training Overhead, speed up the training speed of the model, and improve the generalization ability of the training model.
  • FIG. 1 is a schematic diagram of a channel information feedback system provided by an exemplary embodiment of the present application
  • FIG. 2 is a schematic diagram of transfer learning based on a pre-training-fine-tuning mode provided by an exemplary embodiment of the present application;
  • Fig. 3 is a block diagram of a communication system provided by an exemplary embodiment of the present application.
  • FIG. 4 is a flow chart of a method for training a channel information feedback model provided in an exemplary embodiment of the present application
  • Fig. 5 is a schematic diagram of a mask operation provided by an exemplary embodiment of the present application.
  • Fig. 6 is a schematic diagram of a channel information feedback model in the form of an encoder-decoder provided by an exemplary embodiment of the present application;
  • FIG. 7 is a flow chart of a method for training a channel information feedback model provided in an exemplary embodiment of the present application.
  • Fig. 8 is a schematic diagram of a mask operation provided by an exemplary embodiment of the present application.
  • FIG. 9 is a schematic diagram of a channel information feedback system provided by an exemplary embodiment of the present application.
  • FIG. 10 is a flowchart of a method for training a channel information feedback model provided in an exemplary embodiment of the present application.
  • Fig. 11 is a flowchart of a training method of a channel information feedback model provided by an exemplary embodiment of the present application.
  • FIG. 12 is a flow chart of a method for training a channel information feedback model provided in an exemplary embodiment of the present application.
  • FIG. 13 is a schematic diagram of a training process of a channel information feedback model provided by an exemplary embodiment of the present application.
  • Fig. 14 is a structural block diagram of a training device for a channel information feedback model provided by an exemplary embodiment of the present application.
  • FIG. 15 is a structural block diagram of a training device for a channel information feedback model provided by an exemplary embodiment of the present application.
  • Fig. 16 is a structural block diagram of a training device for a channel information feedback model provided by an exemplary embodiment of the present application.
  • Fig. 17 is a schematic structural diagram of a communication device provided by an exemplary embodiment of the present application.
  • a codebook-based eigenvector feedback scheme is usually used to enable the base station to obtain downlink CSI.
  • the base station sends a downlink channel state information reference signal (Channel State Information-Reference Signal, CSI-RS) to the terminal, and the terminal uses the CSI-RS to estimate the CSI of the downlink channel, and performs eigenvalue decomposition on the estimated downlink channel, The eigenvector corresponding to the downlink channel is obtained.
  • the terminal calculates the corresponding matching codeword coefficients of the feature vector in the preset codebook and performs quantization feedback, and the terminal restores the feature vector according to the quantized CSI fed back by the user.
  • the neural network architecture commonly used in deep learning is nonlinear and data-driven. It can extract features from the actual channel matrix data, and restore the channel matrix information compressed and fed back by the terminal side as much as possible on the base station side. While ensuring the restoration of channel information, It also provides a possibility for the terminal side to reduce the CSI feedback overhead.
  • the CSI feedback based on deep learning regards the channel information as the image to be compressed, uses the deep learning self-encoder to compress the channel information, and reconstructs the compressed channel image at the sending end, which can preserve the channel information to a greater extent .
  • FIG. 1 A typical channel information feedback system is shown in FIG. 1 .
  • the entire feedback system is divided into encoder and decoder parts, which are deployed at the sending end and receiving end respectively.
  • the transmitting end obtains the channel information through channel estimation
  • the channel information matrix is compressed and encoded through the neural network of the encoder, and the compressed bit stream is fed back to the receiving end through the air interface feedback link, and the receiving end passes the decoder according to the feedback bit stream
  • the channel information is restored to obtain complete feedback channel information.
  • the encoder shown in Figure 1 uses the superposition of multiple layers of fully connected layers, and the design of the convolutional layer and residual structure is used in the decoder.
  • the information is input into the encoder, and the information is firstly convoluted through the convolution (conv) layer, and then the dimensions of the information are changed through the reshape (Reshape) layer, and then through the full connection (dense ) layer to complete the encoding of the information;
  • the input information is first processed through the fully connected (dense) layer, and then the information is input into the semantic segmentation network RefineNet for processing.
  • RefineNet includes: reshaping ( Reshape) layer, at least one convolution (conv) layer and the design of the residual structure, and then perform convolution (conv) on the information to complete the decoding of the information.
  • Reshape reshaping
  • conv convolution
  • the network model structure inside the encoder and decoder can be flexibly designed.
  • Migration learning can be understood as using existing knowledge, models, and structures to help achieve learning goals on target data.
  • Transfer learning based on the pre-training-fine-tuning mode refers to: training a network in the source domain, directly using it for the data of the target domain, and fine-tuning on the data of the target domain, as shown in Figure 2. Therefore, transfer learning based on the pre-training-fine-tuning mode can make better use of limited computing resources, and can also deal with the problem of insufficient data in new scenarios.
  • the channel information feedback in the related art is a codebook-based eigenvector feedback scheme.
  • this scheme only selects the optimal feedback matrix and corresponding feedback coefficients from the codebook according to the estimated channel, but the codebook itself is The preset finite sequence, that is, the mapping process from the estimated channel to the channel in the codebook is quantized and lossy.
  • the fixed codebook design cannot be dynamically adjusted according to channel changes, which reduces the accuracy of the feedback channel information, thereby reducing the performance of precoding.
  • the existing deep learning-based channel information feedback schemes use deep neural networks (Deep Neural Networks, DNN), convolutional neural networks (Convolution Neural Networks, CNN) to directly encode the channel information obtained after channel estimation.
  • Compressed feedback compared with the traditional codebook-based channel information feedback, significantly improves the feedback accuracy.
  • the model performance of the channel information feedback scheme based on deep learning is strongly related to the diversity of data, which requires a large amount of real channel data to provide support, and the cost of real channel data collection is high.
  • the training process also brings a lot of computing overhead.
  • the embodiment of the present application proposes a training method of a channel information feedback model.
  • the mask operation is used to shield part of the initial channel information, Reduce the redundant information input during channel information feedback model training, reduce the resource overhead of model training, accelerate the training speed of the model, and improve the generalization ability of the training model.
  • FIG. 3 shows a block diagram of a communication system provided by an exemplary embodiment of the present application.
  • the communication system may include: an access network 12 and a terminal device 14 .
  • the access network 12 includes several network devices 120 .
  • the network device 120 may be a base station, and the base station is a device deployed in an access network to provide a wireless communication function for a terminal.
  • the base station may include various forms of macro base stations, micro base stations, relay stations, access points and so on.
  • the names of devices with base station functions may be different. For example, in LTE systems, they are called eNodeB or eNB; in 5G NR-U systems, they are called gNodeB or gNB. .
  • the description "base station" may change.
  • the above-mentioned devices that provide the wireless communication function for the terminal device 14 are collectively referred to as network devices.
  • the terminal device 14 may include various handheld devices with wireless communication functions, vehicle-mounted devices, wearable devices, computing devices or other processing devices connected to wireless modems, as well as various forms of user equipment, mobile stations (Mobile Station, MS) , terminal (terminal device) and so on.
  • the network device 120 and the terminal device 14 communicate with each other through a certain air interface technology, such as a Uu interface.
  • the terminal device 14 includes: a source-side terminal and a target-side terminal.
  • the source-side terminal is a device for performing the pre-training phase of the model in the transfer learning
  • the target-side terminal is a device for performing the fine-tuning phase of the model in the transfer learning.
  • GSM Global System of Mobile Communication
  • CDMA Code Division Multiple Access
  • WCDMA Wideband Code Division Multiple Access
  • GPRS General Packet Radio Service
  • LTE Long Term Evolution
  • FDD Frequency Division Duplex
  • FDD Frequency Division Duplex
  • TDD Time Division Duplex
  • LTE-A Long Term Evolution
  • NR New Radio
  • NR evolution system of NR system
  • LTE on unlicensed frequency band LTE-based access to Unlicensed spectrum, LTE-U
  • NR-U Universal Mobile Telecommunication System
  • UMTS Universal Mobile Telecommunication System
  • WiMAX Worldwide Interoperability for Microwave Access
  • D2D Device to Device
  • M2M Machine to Machine
  • MTC Machine Type Communication
  • V2V Vehicle to Vehicle
  • V2X Vehicle to Everything
  • Fig. 4 shows a flowchart of a method for training a channel information feedback model provided by an exemplary embodiment of the present application.
  • the method can be applied to a terminal device in a communication system as shown in FIG. 3, and the method includes:
  • Step 410 Perform a masking operation on the initial channel information to obtain masked channel information.
  • the initial channel information is channel information determined after the terminal device performs channel estimation.
  • the masking operation refers to an operation of masking part of information to reduce redundant information.
  • channel information is characterized by high redundancy.
  • some channel information can be shielded by adding a mask to reduce redundant information.
  • visual images are also characterized by high redundancy, for example, missing pixel information can be recovered from adjacent pixel blocks.
  • a method of using a mask operation to hide initial channel information is proposed, thereby reducing redundant information.
  • FIG. 5 a schematic diagram of a masking operation is shown in FIG. 5 .
  • the masked channel information H' is obtained.
  • the initial channel information H has more redundant information. It can be understood that FIG. 5 is only an exemplary illustration, and in practice, channel information may not be presented in the same form as the image shown in FIG. 5 .
  • Step 420 Input the masked channel information into the channel information feedback model, and output the restored channel information.
  • the channel information feedback model is a model for compressing and feeding back the input channel information, and reconstructing and recovering the compressed channel information.
  • the masked channel information is used as the input of the channel information feedback model, and the channel information feedback model is used to predict the masked channel information, so that the output recovery channel information.
  • the channel information feedback model is in the form of an encoder-decoder.
  • FIG. 6 shows a schematic diagram of processing channel information using an encoder-decoder channel information feedback model.
  • the encoder compresses and encodes the estimated channel information H, and feeds it back to the receiving end through the feedback link of the air interface.
  • the feedback link of the air interface actually transmits a feedback vector, which is obtained from the output of the neural network of the encoder at the transmitting end, and is used as part of the input of the neural network at the receiving end for channel information at the receiving end. recover.
  • Step 430 Based on the error between the recovered channel information and the initial channel information, train the channel information feedback model.
  • the channel information feedback model After obtaining the restored channel information output by the channel information feedback model, compare the restored channel information with the corresponding initial channel information to judge the accuracy of the masked content in the initial channel information predicted by the channel information feedback model, and then When there is an error between the restored channel information and the corresponding initial channel information, the channel information feedback model is corrected according to the existing error, so that the generated channel information feedback model has the ability to reconstruct and restore the channel information.
  • the technical solution provided by this embodiment in the case of implementing the channel information feedback scheme based on deep learning, uses the mask operation to shield part of the initial channel information during model training, reducing the channel information feedback model training.
  • the redundant information input at the time reduces the resource overhead of model training, accelerates the training speed of the model, and improves the generalization ability of the training model.
  • Fig. 7 shows a flowchart of a method for training a channel information feedback model provided by an exemplary embodiment of the present application.
  • the method can be applied to a terminal device in a communication system as shown in FIG. 3, and the method includes:
  • Step 710 Divide the channel matrix used to represent the initial channel information into multiple non-overlapping matrix blocks.
  • the matrix size information corresponding to each divided matrix block is the same.
  • the channel matrix used to represent the initial channel information is a 25*25 matrix, and the matrix is divided into 25 5*5 matrix blocks.
  • Step 720 Generate position indices for the matrix blocks to form a sequence of matrix blocks.
  • the position index is an index used to characterize the position of each matrix block in the matrix block sequence.
  • 25 matrix blocks correspond to position indices of 0, 1, .
  • Step 730 Sampling the matrix block sequence, and masking unsampled matrix blocks in the matrix block sequence to obtain masked channel information.
  • the sampled matrix blocks in the matrix block sequence are retained, and the unsampled matrix blocks are deleted, so as to obtain masked channel information.
  • the sampling manner corresponding to the sampling includes: random sampling; or grid sampling. That is, the selection scheme of the masking operation includes a random masking strategy and a grid masking strategy. It can be understood that the selection scheme of the masking operation is not limited to the above two masking strategies, for example, using other prior knowledge to set the mask distribution is within the protection scope of the present application.
  • the grid sampling may be grid sampling at equal intervals.
  • the sampling corresponding to the mask operation shown in FIG. 5 is random sampling
  • the sampling corresponding to the mask operation shown in FIG. 8 is grid sampling.
  • the terminal device randomly samples the matrix block sequence according to uniform distribution, and the sampling rate is 50%.
  • the terminal device performs grid sampling on the matrix block sequence according to uniform distribution, and the sampling rate is 25%.
  • sampling rate is only an exemplary description, and this embodiment of the present application does not limit the numerical value of the sampling rate.
  • a smaller sampling rate is adopted; when the amount of channel information on the local side is small, a larger sampling rate is adopted.
  • Step 740 Input the masked channel information into the channel information feedback model, and output the restored channel information.
  • step 420 For the specific implementation manner of this step, refer to the above-mentioned step 420, which will not be repeated here.
  • Step 750 Based on the error between the restored channel information and the initial channel information, train the channel information feedback model.
  • step 430 For the specific implementation manner of this step, refer to the above-mentioned step 430, which will not be repeated here.
  • the technical solution provided by this embodiment provides different masking strategies such as a random masking strategy and a grid masking strategy to perform a masking operation to ensure the rationality of the masking operation.
  • the overall architecture of this embodiment may be shown in FIG. 9 .
  • FIG. 9 the following operation flow is mainly shown: mask operation, encoder, codeword stuffing and decoder.
  • Encoder The masked channel information H' is input to the encoder, and compression coding is performed to obtain compressed coding information.
  • Codeword stuffing performing codeword stuffing on the compressed coded information to obtain the filled compressed coded information, that is, to obtain complete compressed coded information.
  • Decoder The filled compressed coded information is input to the decoder for decompression to obtain the restored channel information H".
  • the channel information feedback model on the local side includes: a first encoder and a first decoder; inputting the masked channel information into the channel information feedback model, the step of obtaining the restored channel information includes:
  • the masked channel information is input as a model of the first coder, and the masked channel information is compressed through the first coder to obtain compressed coded information.
  • Codeword filling refers to filling codewords at masked positions.
  • the compressed coding information obtained by the first encoder is the coding of the visible matrix blocks in the matrix block sequence corresponding to the channel matrix, and based on the position index, the codeword is filled in the corresponding position of the mask.
  • the training of the channel information feedback model by the terminal device corresponds to the pre-training stage in the transfer learning of the pre-training-fine-tuning mode.
  • the terminal device is a source-side terminal, and the source-side terminal also needs to
  • the trained encoder is uploaded, and the target terminal performs the fine-tuning stage, and the trained decoder is uploaded to the network device.
  • Fig. 10 shows a flowchart of a method for training a channel information feedback model provided by an exemplary embodiment of the present application.
  • the method can be applied to a communication system as shown in FIG. 3, and the method includes:
  • Step 1010 After the training of the channel information feedback model is completed, the terminal at the source side sends the first transfer learning information of the channel information feedback model to the network device.
  • the network device receives the first transfer learning information of the channel information feedback model sent by the source-side terminal.
  • the first transfer learning information is used to perform transfer learning on the channel information feedback model.
  • the channel information feedback model at the terminal at the source side includes: a first encoder, and the first transfer learning information includes:
  • the first transfer learning information carries model parameters of the first encoder.
  • the first transfer learning information carries matrix size information corresponding to the mask operation performed by the terminal at the source side.
  • the matrix size information is used to indicate the size of each matrix block in the matrix block sequence of the input channel information feedback model.
  • Step 1020 The network device sends second transfer learning information to the target terminal.
  • the target-side terminal receives the second migration learning information sent by the network device.
  • the second transfer learning information is used to assist transfer learning.
  • the second transfer learning information includes:
  • the second transfer learning information carries the model parameters of the second encoder.
  • the second encoder is obtained by training based on the mask operation.
  • the second transfer learning information carries matrix size information corresponding to the mask operation performed by the terminal at the source side.
  • the matrix size information is used to indicate the size of each matrix block in the matrix block sequence of the input channel information feedback model.
  • the matrix size information corresponding to the mask operation in the second transfer learning information is the matrix size information corresponding to the mask operation in the first transfer learning information
  • the second encoder in the second transfer learning information is It is obtained based on the first encoder in the first transfer learning information.
  • Step 1030 The target-side terminal generates a second decoder.
  • the target side terminal generates a new decoder.
  • Step 1040 The target terminal performs joint training on the second encoder and the second decoder based on the matrix size information.
  • the target terminal uses the pre-trained second encoder to jointly train the second encoder and the new second decoder under the new data set, so as to complete transfer learning.
  • the transfer learning of the pre-training-fine-tuning mode refers to the pre-training of a network, directly using it for the data of the target scene, and fine-tuning on the target scene data.
  • the model can help other scenarios to achieve the same function.
  • the second encoder is pre-trained, and the pre-trained second encoder is used for retraining together with the new second decoder, thereby saving computing resources of the target-side terminal.
  • Step 1050 The target-side terminal sends the trained second decoder to the network device.
  • the network device receives the second decoder sent by the target-side terminal, and the second decoder is trained by the target-side terminal after performing transfer learning based on the second transfer learning information.
  • the target-side terminal uses the trained second encoder, and the network device side uses the received second decoder, with the target-side terminal as the sender of the channel information, and the network device as the source of the channel information
  • the receiving end implements a channel information feedback scheme based on deep learning by using the second encoder of the terminal on the target side and the second decoder on the network device side.
  • the technical solution provided by this embodiment is to enhance the design of the channel information feedback model in the form of encoder-decoder in the migration scenario of pre-training-fine-tuning mode, and use the mask operation to reduce the input in the pre-training stage. Redundant information accelerates the pre-training speed of the model, improves the generalization ability of the pre-trained model, and improves model performance.
  • the matrix size information is used to indicate the size of each matrix block in the matrix block sequence of the input channel information feedback model.
  • the matrix size information indicates that the size of the matrix block is 5*5, and then the terminal on the target side divides the channel matrix corresponding to the channel information on its own side into multiple matrix blocks of 5*5.
  • the matrix block sequence is input as a model of the second encoder, and the matrix block sequence is compressed through the second encoder to obtain compressed coding information.
  • the compressed coded information is input as a model of the second decoder, and the compressed coded information is decompressed via the second decoder to obtain restored channel information.
  • the second encoder in the target domain supports variable-length sequence input, and the input of the second encoder is a matrix block sequence of a complete channel, do not perform mask processing, and correspondingly , after the second encoder outputs the compressed encoding information, it does not need to perform codeword padding, so as to make full use of the limited data of the current scene.
  • the second encoder is an encoder indicated to the network device by a source-side terminal.
  • the network device after receiving the first transfer learning information sent by the source-side terminal, the network device directly sends the first transfer-learning information as the second transfer-learning information to the target-side terminal device.
  • the second encoder in the second transfer learning information in the above embodiment is equivalent to the first encoder in the first transfer learning information.
  • an encoder is pre-trained by a source-side terminal, and the encoder is migrated to a target-side terminal, and the redundant information input in the pre-training stage is reduced by using a mask operation. Accelerated the pre-training speed of the model.
  • Step 1101 The source-side terminal obtains channel data and executes a masking strategy.
  • the terminal at the source side divides the channel data into regular non-overlapping N small block matrices (patches).
  • a position index 0, 1, 2, ..., N-1 is generated for each matrix block, forming a sequence of matrix blocks.
  • the sequence of matrix blocks is then sampled and the remaining matrix blocks are masked (ie deleted).
  • Step 1102 The source-side terminal jointly trains the encoder and the decoder.
  • the masked channel information is used as the input of the encoder, and a codeword filling operation is added after the encoder accordingly, and the input of the decoder is the filled codeword, including the visible matrix block codeword and The padding codeword at the corresponding position of the mask.
  • the decoder and encoder can adopt an asymmetric design. Compared with the parameter scale of the encoder, the decoder can appropriately reduce the number of network layers and parameters, thereby reducing the pre-training time.
  • Step 1103 The terminal at the source side sends the matrix size information corresponding to the encoder and the mask operation to the network device.
  • Step 1104 The network device sends the matrix size information corresponding to the encoder and the mask operation to the target terminal.
  • Step 1105 The target terminal generates a new decoder.
  • Step 1106 The terminal on the target side processes the channel information into a matrix block sequence adapted by the encoder according to the matrix size information.
  • Step 1107 The target side terminal uses the pre-trained encoder to combine with the new decoder, and retrains under the new data set to complete model migration.
  • Step 1108 the target terminal synchronizes the decoder to the network device.
  • the second encoder is a global encoder obtained by aggregate calculation of model parameters of multiple encoders by the network device, and the multiple encoders come from multiple source-side terminals respectively.
  • Fig. 12 shows a flowchart of a method for training a channel information feedback model provided by an exemplary embodiment of the present application.
  • the method can be applied to a communication system as shown in FIG. 3, and the method includes:
  • Step 1210 After the training of the channel information feedback model is completed, multiple source-side terminals respectively send the first transfer learning information of the channel information feedback model to the network device.
  • the network device receives the first transfer learning information of the channel information feedback model respectively sent by multiple source-side terminals.
  • the first transfer learning information is used to perform transfer learning on the channel information feedback model.
  • the channel information feedback model at the terminal at the source side includes: a first encoder, and the first transfer learning information includes:
  • the first transfer learning information carries model parameters of the first encoder.
  • the first transfer learning information carries matrix size information corresponding to the mask operation performed by the terminal at the source side.
  • the matrix size information is used to indicate the size of each matrix block in the matrix block sequence of the input channel information feedback model.
  • the network device delivers the same masking policy parameter to multiple source-side terminals, where the masking policy parameter is a parameter related to the masking operation.
  • the mask policy parameters include at least one of the following:
  • the matrix size information is used to indicate the size of each matrix block in the matrix block sequence of the input channel information feedback model.
  • sampling information is used to indicate the execution mode of sampling in the mask operation.
  • the sampling information includes at least one of the following: sampling mode; sampling rate.
  • Step 1220 The network device aggregates and calculates the model parameters of multiple trained first encoders to obtain a global encoder.
  • aggregation calculation refers to a calculation method that calculates a set of values and returns a single value.
  • the model parameters of multiple first encoders are aggregated and calculated, and a final model parameter of a global encoder is returned.
  • the embodiment of the present application does not limit the specific implementation form of the aggregated calculation.
  • Step 1230 The network device sends the second transfer learning information to the target terminal.
  • the target-side terminal receives the second migration learning information sent by the network device.
  • the second transfer learning information is used to assist transfer learning.
  • the second transfer learning information includes: the global encoder and matrix size information corresponding to the mask operation, and the global encoder is obtained by training based on the mask operation.
  • the matrix size information corresponding to the mask operation in the second transfer learning information is the matrix size information corresponding to the mask operation in the first transfer learning information
  • the global encoder in the second transfer learning information is based on It is obtained by aggregate calculation of multiple first encoders in the multiple first transfer learning information.
  • Step 1240 The target-side terminal generates a second decoder.
  • the target side terminal generates a new decoder.
  • Step 1250 The target-side terminal performs joint training on the global encoder and the second decoder.
  • the target terminal uses the pre-trained global encoder to jointly train the global encoder and the new second decoder under the new data set, so as to complete transfer learning.
  • Step 1260 The target-side terminal sends the trained second decoder to the network device.
  • multiple source-side terminals cooperate to train to obtain a shared global encoder, and the data redundancy under multiple terminal devices is higher, and the mask operation can be used to significantly reduce Small data redundancy is conducive to enhancing the representation ability of the model to extract potential features, and at the same time speeding up the pre-training speed of the model.
  • Step 1301 unify the masking strategy: the network device uniformly configures masking strategy parameters, and then distributes them to n candidate source-side terminals: source-side terminal 1, source-side terminal 2, . . . , source-side terminal n.
  • Step 1302 pre-training the encoder: the source-side terminals each perform a masking operation, and use the masked channel information as input to train the encoder-decoder.
  • the masking strategy-based autoencoder network architecture is consistent.
  • the components include: mask operation, encoder, codeword filling, and decoder.
  • Each terminal device requires the above four components, and the working structure and flow of each terminal device can refer to the embodiment shown in FIG. 9 , which will not be described in detail here.
  • the decoder and encoder can adopt an asymmetric design. Compared with the parameter scale of the encoder, the decoder can appropriately reduce the number of network layers and parameters, thereby reducing the pre-training time.
  • Step 1303 uploading the encoder: each source-side terminal deletes the decoder part to save device memory resources, retains only the encoder part, and uploads the encoder to the network device for synchronization.
  • Step 1304 aggregation calculation: the base station server or the over-the-air computing node performs aggregation calculation on the encoder model parameters of each coordinated source-side terminal to obtain a global encoder.
  • Step 1305 delivering global encoder and matrix size information: network devices, such as base station servers or air computing nodes, deliver the global encoder and matrix size information corresponding to mask operations to the target terminal.
  • network devices such as base station servers or air computing nodes, deliver the global encoder and matrix size information corresponding to mask operations to the target terminal.
  • target-side terminals there may be multiple target-side terminals, and it is not limited to source-side terminals. All terminals under the network device can be used as candidate target-side terminals, depending on system policies.
  • Step 1306, fine-tuning stage the terminal on the target side uses the matrix size information to process the existing channel information data into a matrix block sequence, without masking, and directly inputs the complete matrix block sequence to the encoder-decoder.
  • the encoder here is a global encoder, but the target-side terminal needs to regenerate an initialized decoder.
  • the size of the encoder model here can appropriately increase the parameter scale in order to obtain better decoding performance.
  • Step 1307 upload the encoder: the encoder is a model that needs to be deployed at the receiving end, so the target terminal must also send the trained decoder to the network device to ensure that the network device can send the encoder to the target terminal The codewords are correctly analyzed and restored to complete channel information.
  • each participant does not need to share the data in the local device, which fully guarantees the data privacy and security of the participants.
  • the steps performed by the source-side terminal can independently realize the training method of the channel information feedback model on the side of the source-side terminal
  • the steps performed by the target-side terminal can independently realize the channel information on the side of the target-side terminal.
  • the steps performed by the network device can be independently implemented as the training method of the channel information feedback model on the network device side.
  • Fig. 14 shows a structural block diagram of an apparatus for training a channel information feedback model provided by an exemplary embodiment of the present application.
  • the apparatus can be implemented as a source-side terminal, or can be implemented as a part of a source-side terminal.
  • the apparatus includes: code module 1402, model processing module 1404 and training module 1406;
  • the masking module 1402 is configured to perform a masking operation on initial channel information to obtain masked channel information
  • the model processing module 1404 is configured to input the masked channel information into the channel information feedback model, and output restored channel information;
  • the training module 1406 is configured to train the channel information feedback model based on the error between the recovered channel information and the initial channel information.
  • the masking module 1402 is configured to:
  • the sampling method corresponding to the sampling includes:
  • the channel information feedback model includes: a first encoder and a first decoder
  • the model processing module 1404 is configured to:
  • the device further includes: an information reporting module;
  • the information reporting module is configured to send the first migration learning information of the channel information feedback model to the network device after the training of the channel information feedback model is completed, and the first migration learning information is used for the channel Information feedback model for transfer learning.
  • the channel information feedback model includes: a first encoder, and the first transfer learning information includes:
  • the matrix size information corresponding to the mask operation is the matrix size information corresponding to the mask operation.
  • the device further includes: a parameter receiving module
  • the parameter receiving module is configured to receive a masking policy parameter issued by a network device, and the masking policy parameter is a parameter related to the masking operation.
  • the mask policy parameters include at least one of the following:
  • the sampling information corresponding to the mask operation is the sampling information corresponding to the mask operation.
  • Fig. 15 shows a structural block diagram of a training device for a channel information feedback model provided by an exemplary embodiment of the present application.
  • the device can be implemented as a target terminal, or can be implemented as a part of the target terminal.
  • the device includes: decoding A decoder generating module 1502, an information receiving module 1504, a training module 1506 and a decoder sending module 1508;
  • the decoder generation module 1502 configured to generate the second decoder
  • the information receiving module 1504 is configured to receive second transfer learning information sent by a network device, the second transfer learning information is used to assist transfer learning, and the second transfer learning information includes: the second encoder and Matrix size information corresponding to the mask operation, the second encoder is obtained by training based on the mask operation;
  • the training module 1506 is configured to jointly train the second encoder and the second decoder based on the matrix size information
  • the decoder sending module 1508 is configured to send the trained second decoder to the network device.
  • the training module 1506 is used for:
  • the matrix block sequence is input as a model of the second encoder, and the matrix block sequence is compressed through the second encoder to obtain compressed encoding information;
  • the second encoder and the second decoder are jointly trained based on an error between the recovered channel information and the initial channel information.
  • the second encoder is an encoder indicated to the network device by a source-side terminal.
  • the second encoder is a global encoder obtained by aggregate calculation of model parameters of multiple encoders by the network device, and the multiple encoders come from multiple one of the source-side terminals.
  • Fig. 16 shows a structural block diagram of a training device for a channel information feedback model provided by an exemplary embodiment of the present application.
  • the device can be implemented as a network device, or can be implemented as a part of the network device.
  • the device includes: an information sending module 1602 and a decoder receiving module 1604;
  • the information sending module 1602 is configured to send second transfer learning information to the target terminal, the second transfer learning information is used to assist in transfer learning, and the second transfer learning information includes: a second encoder and a mask Operating the corresponding matrix size information, the second encoder is obtained by training based on the mask operation;
  • the decoder receiving module 1604 is configured to receive the second decoder sent by the target terminal, where the second decoder is obtained by training after the target terminal performs transfer learning based on the second transfer learning information. of.
  • the second encoder is an encoder indicated to the network device by a source-side terminal
  • the device also includes: an information receiving module;
  • the information receiving module is configured to receive a piece of first transfer learning information sent by the source-side terminal, the first transfer learning information is used to assist transfer learning, and the first transfer learning information includes: a first encoder Matrix size information corresponding to the mask operation.
  • the second encoder is a global encoder obtained by aggregate calculation of model parameters of multiple encoders by the network device;
  • the device also includes: an information receiving module and an aggregation calculation module;
  • the information receiving module is configured to receive a plurality of first transfer learning information respectively sent by a plurality of the source-side terminals, the first transfer learning information is used to assist transfer learning, and the first transfer learning information includes: Matrix size information corresponding to the first encoder and the mask operation;
  • the aggregation calculation module is configured to perform aggregation calculation on model parameters of multiple trained first encoders to obtain the global encoder.
  • the device further includes: a parameter configuration module
  • the parameter configuration module is configured to deliver the same masking policy parameter to multiple terminals at the source side, where the masking policy parameter is a parameter related to the masking operation.
  • the mask policy parameters include at least one of the following:
  • the sampling information corresponding to the mask operation is the sampling information corresponding to the mask operation.
  • FIG. 17 shows a schematic structural diagram of a communication device (terminal device or network device) provided by an exemplary embodiment of the present application.
  • the communication device 1700 includes: a processor 1701 , a transceiver 1702 and a memory 1703 .
  • the processor 1701 includes one or more processing cores, and the processor 1701 executes various functional applications by running software programs and modules.
  • the transceiver 1702 can be used to receive and send information, and the transceiver 1702 can be a communication chip.
  • the memory 1703 may be used to store a computer program, and the processor 1701 is used to execute the computer program, so as to implement various steps performed by the communication device in the foregoing method embodiments.
  • the memory 1703 can be realized by any type of volatile or non-volatile storage device or their combination, and the volatile or non-volatile storage device includes but not limited to: random access memory (Random-Access Memory, RAM) And read-only memory (Read-Only Memory, ROM), erasable programmable read-only memory (Erasable Programmable Read-Only Memory, EPROM), electrically erasable programmable read-only memory (Electrically Erasable Programmable Read-Only Memory, EEPROM), flash memory or other solid-state storage technologies, compact disc read-only memory (CD-ROM), high-density digital video disc (Digital Video Disc, DVD) or other optical storage, tape cartridges, tapes, disks storage or other magnetic storage devices.
  • RAM Random-Access Memory
  • ROM read-only memory
  • EPROM erasable programmable read-only memory
  • EPROM erasable programmable Read-Only Memory
  • EEPROM Electrically erasable programmable read-only memory
  • the processor 1701 and the transceiver 1702 involved in the embodiment of the present application can perform the steps performed by the source-side terminal in any of the methods shown in the above-mentioned embodiments, where I won't repeat them here.
  • the processor 1701 is configured to perform a masking operation on initial channel information to obtain masked channel information
  • the processor 1701 is configured to input the masked channel information into a channel information feedback model, and output restored channel information;
  • the processor 1701 is configured to train the channel information feedback model based on an error between the recovered channel information and the initial channel information.
  • the processor 1701 and the transceiver 1702 involved in the embodiment of the present application may perform the steps performed by the target-side terminal in any of the methods shown in the above-mentioned embodiments, where I won't repeat them here.
  • the processor 1701 is configured to generate a second decoder
  • the transceiver 1702 is configured to receive second transfer learning information sent by a network device, the second transfer learning information is used to assist transfer learning, and the second transfer learning information includes: the second encoder and mask Matrix size information corresponding to the code operation, the second encoder is obtained by training based on the mask operation;
  • the processor 1701 is configured to jointly train the second encoder and the second decoder based on the matrix size information
  • the transceiver 1702 is configured to send the trained second decoder to the network device.
  • the processor 1701 and the transceiver 1702 involved in the embodiment of the present application can execute the steps performed by the network device in any of the methods shown in the above embodiments, which are not described here Let me repeat.
  • the communication device when the communication device is implemented as a network device,
  • the transceiver 1702 is configured to send second transfer learning information to the target terminal, the second transfer learning information is used to assist transfer learning, and the second transfer learning information includes: a second encoder and a mask operation Corresponding matrix size information, the second encoder is obtained by training based on the mask operation;
  • the transceiver 1702 is configured to receive the second decoder sent by the target-side terminal, where the second decoder is trained by the target-side terminal after performing transfer learning based on the second transfer learning information.
  • a computer-readable storage medium stores at least one instruction, at least one program, a code set or an instruction set, the at least one instruction, the At least one section of program, the code set or instruction set is loaded and executed by the processor to implement the training method of the channel information feedback model executed by the communication device provided in the above method embodiments.
  • a chip is also provided, the chip includes a programmable logic circuit and/or program instructions, and when the chip is run on a computer device, it is used to realize the channel information feedback described in the above aspect The training method of the model.
  • a computer program product is also provided.
  • the computer program product runs on a processor of a computer device, the computer device executes the method for training a channel information feedback model described in the above aspects.
  • the program can be stored in a computer-readable storage medium.
  • the above-mentioned The storage medium mentioned may be a read-only memory, a magnetic disk or an optical disk, and the like.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Cable Transmission Systems, Equalization Of Radio And Reduction Of Echo (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)

Abstract

The present application relates to the technical field of communications. Disclosed are a method and device for training a channel information feedback model, an apparatus, and a storage medium. The method comprises: performing a mask operation on initial channel information to obtain masked channel information; inputting the masked channel information into a channel information feedback model, and outputting restored channel information; and training the channel information feedback model on the basis of a discrepancy between the restored channel information and the initial channel information.

Description

信道信息反馈模型的训练方法、装置、设备及存储介质Training method, device, equipment and storage medium of channel information feedback model 技术领域technical field
本申请涉及通信技术领域,特别涉及一种信道信息反馈模型的训练方法、装置、设备及存储介质。The present application relates to the field of communication technology, and in particular to a training method, device, equipment and storage medium of a channel information feedback model.
背景技术Background technique
终端设备一般通过信道状态信息(Channel State Information,CSI)测量,生成信道信息,并将信道信息反馈给网络设备。The terminal device generally generates channel information through Channel State Information (CSI) measurement, and feeds the channel information back to the network device.
相关技术中,已引入基于深度学习的信道信息反馈方案:将信道信息视作待压缩图像,利用编码器对信道信息进行压缩反馈,并在发送端利用解码器对压缩后的信道信息进行重构,可以更大程度地保留信道信息。In related technologies, a channel information feedback scheme based on deep learning has been introduced: the channel information is regarded as an image to be compressed, the encoder is used to compress and feed back the channel information, and the decoder is used at the sending end to reconstruct the compressed channel information , the channel information can be preserved to a greater extent.
发明内容Contents of the invention
本申请实施例提供了一种信道信息反馈模型的训练方法、装置、设备及存储介质。所述技术方案如下:Embodiments of the present application provide a training method, device, equipment, and storage medium for a channel information feedback model. Described technical scheme is as follows:
根据本申请的一个方面,提供了一种信道信息反馈模型的训练方法,应用于源侧终端中,所述方法包括:According to one aspect of the present application, a method for training a channel information feedback model is provided, which is applied to a source-side terminal, and the method includes:
对初始信道信息进行掩码操作,得到掩码信道信息;Perform a masking operation on the initial channel information to obtain masked channel information;
将所述掩码信道信息输入所述信道信息反馈模型,输出恢复信道信息;Input the masked channel information into the channel information feedback model, and output the restored channel information;
基于所述恢复信道信息与所述初始信道信息之间的误差,对所述信道信息反馈模型进行训练。The channel information feedback model is trained based on an error between the restored channel information and the initial channel information.
根据本申请的一个方面,提供了一种信道信息反馈模型的训练方法,应用于目标侧终端中,所述信道信息反馈模型包括:第二编码器和第二解码器,所述方法包括:According to one aspect of the present application, a method for training a channel information feedback model is provided, which is applied to a target-side terminal, where the channel information feedback model includes: a second encoder and a second decoder, and the method includes:
生成所述第二解码器;generating said second decoder;
接收网络设备发送的第二迁移学习信息,所述第二迁移学习信息用于辅助进行迁移学习,所述第二迁移学习信息包括:所述第二编码器和掩码操作对应的矩阵尺寸信息,所述第二编码器是基于所述掩码操作进行训练得到的;receiving second transfer learning information sent by the network device, where the second transfer learning information is used to assist in transfer learning, where the second transfer learning information includes: matrix size information corresponding to the second encoder and mask operation, The second encoder is obtained by training based on the mask operation;
基于所述矩阵尺寸信息,对所述第二编码器和所述第二解码器进行联合训练;jointly training the second encoder and the second decoder based on the matrix size information;
向所述网络设备发送训练好的所述第二解码器。sending the trained second decoder to the network device.
根据本申请的一个方面,提供了一种信道信息反馈模型的训练方法,应用于网络设备中,所述方法包括:According to one aspect of the present application, a method for training a channel information feedback model is provided, which is applied to a network device, and the method includes:
向目标侧终端发送第二迁移学习信息,所述第二迁移学习信息用于辅助进行迁移学习,所述第二迁移学习信息包括:第二编码器和掩码操作对应的矩阵尺寸信息,所述第二编码器是基于所述掩码操作进行训练得到的;Sending second transfer learning information to the target side terminal, where the second transfer learning information is used to assist transfer learning, where the second transfer learning information includes: matrix size information corresponding to the second encoder and mask operation, the The second encoder is obtained by training based on the mask operation;
接收所述目标侧终端发送的第二解码器,所述第二解码器是所述目标侧终端基于所述第二迁移学习信息进行迁移学习后,训练得到的。receiving the second decoder sent by the target-side terminal, where the second decoder is trained by the target-side terminal after performing transfer learning based on the second transfer learning information.
根据本申请的一个方面,提供了一种信道信息反馈模型的训练装置,所述装置包括:掩码模块、模型处理模块和训练模块;According to one aspect of the present application, a training device for a channel information feedback model is provided, the device comprising: a mask module, a model processing module and a training module;
所述掩码模块,用于对初始信道信息进行掩码操作,得到掩码信道信息;The masking module is configured to perform a masking operation on initial channel information to obtain masked channel information;
所述模型处理模块,用于将所述掩码信道信息输入所述信道信息反馈模型,输出恢复信道信息;The model processing module is configured to input the masked channel information into the channel information feedback model, and output restored channel information;
所述训练模块,用于基于所述恢复信道信息与所述初始信道信息之间的误差,对所述信道信息反馈模型进行训练。The training module is configured to train the channel information feedback model based on the error between the recovered channel information and the initial channel information.
根据本申请的一个方面,提供了一种信道信息反馈模型的训练装置,所述信道信息反馈模型包括:第二编码器和第二解码器,所述装置包括:解码器生成模块、信息接收模块、训练模块和解码器发送模块;According to one aspect of the present application, a training device for a channel information feedback model is provided, the channel information feedback model includes: a second encoder and a second decoder, and the device includes: a decoder generating module and an information receiving module , a training module and a decoder sending module;
所述解码器生成模块,用于生成所述第二解码器;The decoder generating module is configured to generate the second decoder;
所述信息接收模块,用于接收网络设备发送的第二迁移学习信息,所述第二迁移学习信息用于辅助进行迁移学习,所述第二迁移学习信息包括:所述第二编码器和掩码操作对应的矩阵尺寸信息,所述第二编码器是基于所述掩码操作进行训练得到的;The information receiving module is configured to receive second transfer learning information sent by a network device, the second transfer learning information is used to assist transfer learning, and the second transfer learning information includes: the second encoder and mask Matrix size information corresponding to the code operation, the second encoder is obtained by training based on the mask operation;
所述训练模块,用于基于所述矩阵尺寸信息,对所述第二编码器和所述第二解码器进行联合训练;The training module is configured to jointly train the second encoder and the second decoder based on the matrix size information;
所述解码器发送模块,用于向所述网络设备发送训练好的所述第二解码器。The decoder sending module is configured to send the trained second decoder to the network device.
根据本申请的一个方面,提供了一种信道信息反馈模型的训练装置,所述装置包括:信息发送模块和解码器接收模块;According to one aspect of the present application, a training device for a channel information feedback model is provided, the device comprising: an information sending module and a decoder receiving module;
所述信息发送模块,用于向目标侧终端发送第二迁移学习信息,所述第二迁移学习信息用于辅助进行迁移学习,所述第二迁移学习信息包括:第二编码器和掩码操作对应的矩阵尺寸信息,所述第二编码器是基于所述掩码操作进行训练得到的;The information sending module is configured to send second transfer learning information to the target terminal, the second transfer learning information is used to assist transfer learning, and the second transfer learning information includes: a second encoder and a mask operation Corresponding matrix size information, the second encoder is obtained by training based on the mask operation;
所述解码器接收模块,用于接收所述目标侧终端发送的第二解码器,所述第二解码器是所述目标侧终端基于所述第二迁移学习信息进行迁移学习后,训练得到的。The decoder receiving module is configured to receive the second decoder sent by the target terminal, the second decoder is obtained by training after the target terminal performs transfer learning based on the second transfer learning information .
根据本申请的一个方面,提供了一种终端设备,所述终端设备包括:处理器;其中,According to one aspect of the present application, a terminal device is provided, and the terminal device includes: a processor; wherein,
所述处理器,用于对初始信道信息进行掩码操作,得到掩码信道信息;The processor is configured to perform a masking operation on initial channel information to obtain masked channel information;
所述处理器,用于将所述掩码信道信息输入信道信息反馈模型,输出恢复信道信息;The processor is configured to input the masked channel information into a channel information feedback model, and output restored channel information;
所述处理器,用于基于所述恢复信道信息与所述初始信道信息之间的误差,对所述信道信息反馈模型进行训练。The processor is configured to train the channel information feedback model based on an error between the recovered channel information and the initial channel information.
根据本申请的一个方面,提供了一种终端设备,所述终端设备包括:处理器和与所述处理器相连的收发器;其中,According to one aspect of the present application, a terminal device is provided, and the terminal device includes: a processor and a transceiver connected to the processor; wherein,
所述处理器,用于生成第二解码器;the processor, configured to generate a second decoder;
所述收发器,用于接收网络设备发送的第二迁移学习信息,所述第二迁移学习信息用于辅助进行迁移学习,所述第二迁移学习信息包括:所述第二编码器和掩码操作对应的矩阵尺寸信息,所述第二编码器是基于所述掩码操作进行训练得到的;The transceiver is configured to receive second transfer learning information sent by a network device, the second transfer learning information is used to assist transfer learning, and the second transfer learning information includes: the second encoder and a mask Operating the corresponding matrix size information, the second encoder is obtained by training based on the mask operation;
所述处理器,用于基于所述矩阵尺寸信息,对所述第二编码器和所述第二解码器进行联合训练;The processor is configured to jointly train the second encoder and the second decoder based on the matrix size information;
所述收发器,用于向所述网络设备发送训练好的所述第二解码器。The transceiver is configured to send the trained second decoder to the network device.
根据本申请的一个方面,提供了一种网络设备,所述网络设备包括:收发器;其中,According to one aspect of the present application, a network device is provided, and the network device includes: a transceiver; wherein,
所述收发器,用于向目标侧终端发送第二迁移学习信息,所述第二迁移学习信息用于辅助进行迁移学习,所述第二迁移学习信息包括:第二编码器和掩码操作对应的矩阵尺寸信息,所述第二编码器是基于所述掩码操作进行训练得到的;The transceiver is configured to send second transfer learning information to the target-side terminal, the second transfer learning information is used to assist in transfer learning, and the second transfer learning information includes: a second encoder corresponding to a mask operation The matrix size information of the second encoder is obtained by training based on the mask operation;
所述收发器,用于接收所述目标侧终端发送的第二解码器,所述第二解码器是所述目标侧终端基于所述第二迁移学习信息进行迁移学习后,训练得到的。The transceiver is configured to receive the second decoder sent by the target-side terminal, where the second decoder is trained by the target-side terminal after performing transfer learning based on the second transfer learning information.
根据本申请的一个方面,提供了一种计算机可读存储介质,所述可读存储介质中存储有可执行指令,所述可执行指令由处理器加载并执行以实现如上述方面所述的信道信息反馈模型的训练方法。According to one aspect of the present application, a computer-readable storage medium is provided, and executable instructions are stored in the readable storage medium, and the executable instructions are loaded and executed by a processor to implement the channel described in the above aspect Training methods for information feedback models.
根据本申请实施例的一个方面,提供了一种芯片,所述芯片包括可编程逻辑电路和/或程序指令,当所述芯片在计算机设备上运行时,用于实现上述方面所述的信道信息反馈模型的训练方法。According to an aspect of an embodiment of the present application, a chip is provided, the chip includes a programmable logic circuit and/or program instructions, and when the chip is run on a computer device, it is used to realize the channel information described in the above aspect Feedback model training method.
根据本申请的一个方面,提供了一种计算机程序产品,该计算机程序产品在计算机设备的处理器上运行时,使得计算机设备执行上述方面所述的信道信息反馈模型的训练方法。According to one aspect of the present application, a computer program product is provided. When the computer program product is run on a processor of a computer device, the computer device executes the method for training a channel information feedback model described in the above aspect.
本申请实施例提供的技术方案至少包括如下有益效果:The technical solutions provided by the embodiments of the present application at least include the following beneficial effects:
在执行基于深度学习的信道信息反馈方案的情况下,在进行模型训练时,利用掩码操作屏蔽部分的初始信道信息,降低信道信息反馈模型训练时输入的冗余信息,减小模型训练的资源开销,加速模型的训练速度,提高训练模型的泛化能力。In the case of implementing the channel information feedback scheme based on deep learning, during model training, use the mask operation to shield part of the initial channel information, reduce the redundant information input during the channel information feedback model training, and reduce the resources for model training Overhead, speed up the training speed of the model, and improve the generalization ability of the training model.
附图说明Description of drawings
为了更清楚地说明本申请实施例中的技术方案,下面将对实施例描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the technical solutions in the embodiments of the present application, the drawings that need to be used in the description of the embodiments will be briefly introduced below. Obviously, the drawings in the following description are only some embodiments of the present application. For those skilled in the art, other drawings can also be obtained based on these drawings without creative effort.
图1是本申请一个示例性实施例提供的信道信息反馈系统的示意图;FIG. 1 is a schematic diagram of a channel information feedback system provided by an exemplary embodiment of the present application;
图2是本申请一个示例性实施例提供的基于预训练-微调模式的迁移学习的示意图;FIG. 2 is a schematic diagram of transfer learning based on a pre-training-fine-tuning mode provided by an exemplary embodiment of the present application;
图3是本申请一个示例性实施例提供的通信系统的框图;Fig. 3 is a block diagram of a communication system provided by an exemplary embodiment of the present application;
图4是本申请一个示例性实施例提供的信道信息反馈模型的训练方法的流程图;FIG. 4 is a flow chart of a method for training a channel information feedback model provided in an exemplary embodiment of the present application;
图5是本申请一个示例性实施例提供的掩码操作的示意图;Fig. 5 is a schematic diagram of a mask operation provided by an exemplary embodiment of the present application;
图6是本申请一个示例性实施例提供的编码器-解码器的形式的信道信息反馈模型的示意图;Fig. 6 is a schematic diagram of a channel information feedback model in the form of an encoder-decoder provided by an exemplary embodiment of the present application;
图7是本申请一个示例性实施例提供的信道信息反馈模型的训练方法的流程图;FIG. 7 is a flow chart of a method for training a channel information feedback model provided in an exemplary embodiment of the present application;
图8是本申请一个示例性实施例提供的掩码操作的示意图;Fig. 8 is a schematic diagram of a mask operation provided by an exemplary embodiment of the present application;
图9是本申请一个示例性实施例提供的信道信息反馈系统的示意图;FIG. 9 is a schematic diagram of a channel information feedback system provided by an exemplary embodiment of the present application;
图10是本申请一个示例性实施例提供的信道信息反馈模型的训练方法的流程图;FIG. 10 is a flowchart of a method for training a channel information feedback model provided in an exemplary embodiment of the present application;
图11是本申请一个示例性实施例提供的信道信息反馈模型的训练方法的流程图;Fig. 11 is a flowchart of a training method of a channel information feedback model provided by an exemplary embodiment of the present application;
图12是本申请一个示例性实施例提供的信道信息反馈模型的训练方法的流程图;FIG. 12 is a flow chart of a method for training a channel information feedback model provided in an exemplary embodiment of the present application;
图13是本申请一个示例性实施例提供的信道信息反馈模型的训练过程的示意图;FIG. 13 is a schematic diagram of a training process of a channel information feedback model provided by an exemplary embodiment of the present application;
图14是本申请一个示例性实施例提供的信道信息反馈模型的训练装置的结构框图;Fig. 14 is a structural block diagram of a training device for a channel information feedback model provided by an exemplary embodiment of the present application;
图15是本申请一个示例性实施例提供的信道信息反馈模型的训练装置的结构框图;FIG. 15 is a structural block diagram of a training device for a channel information feedback model provided by an exemplary embodiment of the present application;
图16是本申请一个示例性实施例提供的信道信息反馈模型的训练装置的结构框图;Fig. 16 is a structural block diagram of a training device for a channel information feedback model provided by an exemplary embodiment of the present application;
图17是本申请一个示例性实施例提供的通信设备的结构示意图。Fig. 17 is a schematic structural diagram of a communication device provided by an exemplary embodiment of the present application.
具体实施方式Detailed ways
为使本申请的目的、技术方案和优点更加清楚,下面将结合附图对本申请实施方式作进一步地详细描述。In order to make the purpose, technical solution and advantages of the present application clearer, the implementation manners of the present application will be further described in detail below in conjunction with the accompanying drawings.
首先,对本申请实施例中涉及的技术知识进行简单介绍:First, a brief introduction to the technical knowledge involved in the embodiments of this application:
基于码本的特征向量反馈方案Codebook-Based Eigenvector Feedback Scheme
在目前的新空口(New Radio,NR)系统中,通常采用基于码本的特征向量反馈方案使得基站获取下行CSI。具体地,基站向终端发送下行信道状态信息参考信号(Channel State Information-Reference Signal,CSI-RS),终端利用CSI-RS估计得到下行信道的CSI,并对估计得到的下行信道进行特征值分解,得到该下行信道对应的特征向量。终端按照一定规则,计算该特征向量在预先设定的码本中对应匹配的码字系数并进行量化反馈,终端根据用户反馈的量化后的CSI恢复特征向量。In the current New Radio (NR) system, a codebook-based eigenvector feedback scheme is usually used to enable the base station to obtain downlink CSI. Specifically, the base station sends a downlink channel state information reference signal (Channel State Information-Reference Signal, CSI-RS) to the terminal, and the terminal uses the CSI-RS to estimate the CSI of the downlink channel, and performs eigenvalue decomposition on the estimated downlink channel, The eigenvector corresponding to the downlink channel is obtained. According to certain rules, the terminal calculates the corresponding matching codeword coefficients of the feature vector in the preset codebook and performs quantization feedback, and the terminal restores the feature vector according to the quantized CSI fed back by the user.
基于深度学习的信道信息反馈方案Channel information feedback scheme based on deep learning
鉴于人工智能(Artificial Intelligence,AI)技术,尤其是深度学习在计算机视觉、自然语言处理等方面取得了巨大的成功,通信领域开始尝试利用深度学习来解决传统通信方法难以解决的技术难题,例如深度学习。深度学习中常用的神经网络架构是非线性且是数据驱动的,可以对实际信道矩阵数据进行特征提取,并在基站侧尽可能还原终端侧压缩反馈的信道矩阵信息,在保证还原信道信息的同时,也为终端侧降低CSI反馈开销提供了可能性。基于深度学习的CSI反馈将信道信息视作待压缩图像,利用深度学习自编码器对信道信息进行压缩反馈,并在发送端对压缩后的信道图像进行重构,可以更大程度地保留信道信息。In view of artificial intelligence (AI) technology, especially deep learning has achieved great success in computer vision, natural language processing, etc., the field of communication has begun to try to use deep learning to solve technical problems that are difficult to solve by traditional communication methods, such as deep learning. study. The neural network architecture commonly used in deep learning is nonlinear and data-driven. It can extract features from the actual channel matrix data, and restore the channel matrix information compressed and fed back by the terminal side as much as possible on the base station side. While ensuring the restoration of channel information, It also provides a possibility for the terminal side to reduce the CSI feedback overhead. The CSI feedback based on deep learning regards the channel information as the image to be compressed, uses the deep learning self-encoder to compress the channel information, and reconstructs the compressed channel image at the sending end, which can preserve the channel information to a greater extent .
一种典型的信道信息反馈系统如图1所示。整个反馈系统分为编码器及解码器部分,分别部署在发送端与接收端。发送端通过信道估计得到信道信息后,通过编码器的神经网络对信道信息矩阵进行压缩编码,并将压缩后的比特流通过空口反馈链路反馈给接收端,接收端通过解码器根据反馈比特流对信道信息进行恢复,以获得完整的反馈信道信息。A typical channel information feedback system is shown in FIG. 1 . The entire feedback system is divided into encoder and decoder parts, which are deployed at the sending end and receiving end respectively. After the transmitting end obtains the channel information through channel estimation, the channel information matrix is compressed and encoded through the neural network of the encoder, and the compressed bit stream is fed back to the receiving end through the air interface feedback link, and the receiving end passes the decoder according to the feedback bit stream The channel information is restored to obtain complete feedback channel information.
图1中所示的编码器采用了多层全连接层的叠加,解码器中采用了卷积层与残差结构的设计。示例性 的,在编码器一侧,将信息输入编码器,先通过卷积(conv)层对信息进行卷积操作,再通过重塑(Reshape)层改变信息的维度,再通过全连接(dense)层进行处理,完成对信息的编码;在解码器一侧,对输入的信息先通过全连接(dense)层进行处理,再将信息输入语义分割网络RefineNet进行处理,RefineNet中包括:重塑(Reshape)层,至少一个卷积(conv)层以及残差结构的设计,再将信息进行卷积(conv),完成对信息的解码。在该编解码框架不变的情况下,编码器和解码器内部的网络模型结构可进行灵活设计。The encoder shown in Figure 1 uses the superposition of multiple layers of fully connected layers, and the design of the convolutional layer and residual structure is used in the decoder. Exemplarily, on the side of the encoder, the information is input into the encoder, and the information is firstly convoluted through the convolution (conv) layer, and then the dimensions of the information are changed through the reshape (Reshape) layer, and then through the full connection (dense ) layer to complete the encoding of the information; on the decoder side, the input information is first processed through the fully connected (dense) layer, and then the information is input into the semantic segmentation network RefineNet for processing. RefineNet includes: reshaping ( Reshape) layer, at least one convolution (conv) layer and the design of the residual structure, and then perform convolution (conv) on the information to complete the decoding of the information. Under the condition that the encoding and decoding framework remains unchanged, the network model structure inside the encoder and decoder can be flexibly designed.
基于预训练-微调模式的迁移学习Transfer learning based on pre-training-fine-tuning model
迁移学习可以理解为利用了已有的知识、模型、结构来帮助达成在目标数据上的学习目标。基于预训练-微调模式的迁移学习指的是:在源领域训练好一个网络,直接将其用于目标域的数据,并在目标域数据上进行微调,具体如图2所示。所以,基于预训练-微调模式的迁移学习可以更好地利用有限的计算资源,也可以应对新场景数据量不足问题。Migration learning can be understood as using existing knowledge, models, and structures to help achieve learning goals on target data. Transfer learning based on the pre-training-fine-tuning mode refers to: training a network in the source domain, directly using it for the data of the target domain, and fine-tuning on the data of the target domain, as shown in Figure 2. Therefore, transfer learning based on the pre-training-fine-tuning mode can make better use of limited computing resources, and can also deal with the problem of insufficient data in new scenarios.
相关技术中的信道信息反馈为基于码本的特征向量反馈方案,然而,该方案仅是根据估计出的信道从码本中挑选最优的反馈矩阵和对应的反馈系数,但其码本本身是预先设定的有限序列,即从估计出的信道到码本中的信道的映射过程是量化有损的。同时,固定的码本设计无法根据信道的变化而进行动态的调整,这使得反馈的信道信息精确度下降,进而降低了预编码的性能。The channel information feedback in the related art is a codebook-based eigenvector feedback scheme. However, this scheme only selects the optimal feedback matrix and corresponding feedback coefficients from the codebook according to the estimated channel, but the codebook itself is The preset finite sequence, that is, the mapping process from the estimated channel to the channel in the codebook is quantized and lossy. At the same time, the fixed codebook design cannot be dynamically adjusted according to channel changes, which reduces the accuracy of the feedback channel information, thereby reducing the performance of precoding.
进一步地,目前已有的基于深度学习的信道信息反馈方案利用深度神经网络(Deep Neural Networks,DNN)、卷积神经网络(Convolution Neural Networks,CNN)等对信道估计后得到的信道信息进行直接编码压缩反馈,相比传统的基于码本的信道信息反馈,显著提升了反馈精度。然而,基于深度学习的信道信息反馈方案的模型性能与数据多样性强相关,需要大量真实的信道数据提供支撑,真实的信道数据采集成本高,同时,训练过程也带来大量的计算开销。Furthermore, the existing deep learning-based channel information feedback schemes use deep neural networks (Deep Neural Networks, DNN), convolutional neural networks (Convolution Neural Networks, CNN) to directly encode the channel information obtained after channel estimation. Compressed feedback, compared with the traditional codebook-based channel information feedback, significantly improves the feedback accuracy. However, the model performance of the channel information feedback scheme based on deep learning is strongly related to the diversity of data, which requires a large amount of real channel data to provide support, and the cost of real channel data collection is high. At the same time, the training process also brings a lot of computing overhead.
此外,由于无线环境不够稳定,数据分布也会随着时间推移会有变化。有限数据集下,即使充分模型训练,但随着时间推移导致数据分布产生变化后,模型性能难以保证。In addition, due to the unstable wireless environment, the data distribution will change over time. Under the limited data set, even if the model is fully trained, it is difficult to guarantee the performance of the model after the data distribution changes over time.
因此,如何在不同的信道场景下,应对时间推移带来的数据分布变化,同时保证信道向量压缩反馈与恢复的精度,是一项亟待解决的模型泛化问题。Therefore, how to cope with the changes in data distribution brought about by time lapse under different channel scenarios, while ensuring the accuracy of channel vector compression feedback and recovery, is an urgent model generalization problem to be solved.
针对上述问题,本申请实施例提出一种信道信息反馈模型的训练方法,在执行基于深度学习的信道信息反馈方案的情况下,在进行模型训练时,利用掩码操作屏蔽部分的初始信道信息,降低信道信息反馈模型训练时输入的冗余信息,减小模型训练的资源开销,加速模型的训练速度,提高训练模型的泛化能力。In view of the above problems, the embodiment of the present application proposes a training method of a channel information feedback model. In the case of implementing a channel information feedback scheme based on deep learning, when performing model training, the mask operation is used to shield part of the initial channel information, Reduce the redundant information input during channel information feedback model training, reduce the resource overhead of model training, accelerate the training speed of the model, and improve the generalization ability of the training model.
图3示出了本申请一个示例性实施例提供的通信系统的框图,该通信系统可以包括:接入网12和终端设备14。FIG. 3 shows a block diagram of a communication system provided by an exemplary embodiment of the present application. The communication system may include: an access network 12 and a terminal device 14 .
接入网12中包括若干个网络设备120。网络设备120可以是基站,所述基站是一种部署在接入网中用以为终端提供无线通信功能的装置。基站可以包括各种形式的宏基站,微基站,中继站,接入点等等。在采用不同的无线接入技术的系统中,具备基站功能的设备的名称可能会有所不同,例如在LTE系统中,称为eNodeB或者eNB;在5G NR-U系统中,称为gNodeB或者gNB。随着通信技术的演进,“基站”这一描述可能会变化。为方便本申请实施例中,上述为终端设备14提供无线通信功能的装置统称为网络设备。The access network 12 includes several network devices 120 . The network device 120 may be a base station, and the base station is a device deployed in an access network to provide a wireless communication function for a terminal. The base station may include various forms of macro base stations, micro base stations, relay stations, access points and so on. In systems using different wireless access technologies, the names of devices with base station functions may be different. For example, in LTE systems, they are called eNodeB or eNB; in 5G NR-U systems, they are called gNodeB or gNB. . As communications technology evolves, the description "base station" may change. For convenience in this embodiment of the present application, the above-mentioned devices that provide the wireless communication function for the terminal device 14 are collectively referred to as network devices.
终端设备14可以包括各种具有无线通信功能的手持设备、车载设备、可穿戴设备、计算设备或连接到无线调制解调器的其他处理设备,以及各种形式的用户设备,移动台(Mobile Station,MS),终端(terminal device)等等。为方便描述,上面提到的设备统称为终端。网络设备120与终端设备14之间通过某种空口技术互相通信,例如Uu接口。The terminal device 14 may include various handheld devices with wireless communication functions, vehicle-mounted devices, wearable devices, computing devices or other processing devices connected to wireless modems, as well as various forms of user equipment, mobile stations (Mobile Station, MS) , terminal (terminal device) and so on. For convenience of description, the devices mentioned above are collectively referred to as terminals. The network device 120 and the terminal device 14 communicate with each other through a certain air interface technology, such as a Uu interface.
可选的,终端设备14包括:源侧终端和目标侧终端。其中,源侧终端是用于执行迁移学习中的模型的预训练阶段的设备,目标侧终端是用于执行迁移学习中的模型的微调阶段的设备。Optionally, the terminal device 14 includes: a source-side terminal and a target-side terminal. Wherein, the source-side terminal is a device for performing the pre-training phase of the model in the transfer learning, and the target-side terminal is a device for performing the fine-tuning phase of the model in the transfer learning.
本申请实施例的技术方案可以应用于各种通信系统,例如:全球移动通讯(Global System of Mobile Communication,GSM)系统、码分多址(Code Division Multiple Access,CDMA)系统、宽带码分多址(Wideband Code Division Multiple Access,WCDMA)系统、通用分组无线业务(General Packet Radio Service, GPRS)、长期演进(Long Term Evolution,LTE)系统、LTE频分双工(Frequency Division Duplex,FDD)系统、LTE时分双工(Time Division Duplex,TDD)系统、先进的长期演进(Advanced Long Term Evolution,LTE-A)系统、新无线(New Radio,NR)系统、NR系统的演进系统、非授权频段上的LTE(LTE-based access to Unlicensed spectrum,LTE-U)系统、NR-U系统、通用移动通信系统(Universal Mobile Telecommunication System,UMTS)、全球互联微波接入(Worldwide Interoperability for Microwave Access,WiMAX)通信系统、无线局域网(Wireless Local Area Networks,WLAN)、无线保真(Wireless Fidelity,WiFi)、第6代移动通信技术(6-Generation,6G)系统、下一代通信系统或其他通信系统等。The technical solutions of the embodiments of the present application can be applied to various communication systems, such as: Global System of Mobile Communication (GSM) system, Code Division Multiple Access (CDMA) system, wideband code division multiple access (Wideband Code Division Multiple Access, WCDMA) system, General Packet Radio Service (General Packet Radio Service, GPRS), Long Term Evolution (Long Term Evolution, LTE) system, LTE Frequency Division Duplex (Frequency Division Duplex, FDD) system, LTE Time Division Duplex (Time Division Duplex, TDD) system, Advanced Long Term Evolution (LTE-A) system, New Radio (NR) system, evolution system of NR system, LTE on unlicensed frequency band (LTE-based access to Unlicensed spectrum, LTE-U) system, NR-U system, Universal Mobile Telecommunication System (UMTS), Worldwide Interoperability for Microwave Access (WiMAX) communication system, Wireless local area network (Wireless Local Area Networks, WLAN), wireless fidelity (Wireless Fidelity, WiFi), 6th generation mobile communication technology (6-Generation, 6G) system, next generation communication system or other communication systems, etc.
通常来说,传统的通信系统支持的连接数有限,也易于实现,然而,随着通信技术的发展,移动通信系统将不仅支持传统的通信,还将支持例如,设备到设备(Device to Device,D2D)通信,机器到机器(Machine to Machine,M2M)通信,机器类型通信(Machine Type Communication,MTC),车辆间(Vehicle to Vehicle,V2V)通信以及车联网(Vehicle to Everything,V2X)系统等。本申请实施例也可以应用于这些通信系统。Generally speaking, the number of connections supported by traditional communication systems is limited and easy to implement. However, with the development of communication technology, mobile communication systems will not only support traditional communication, but also support, for example, Device to Device (Device to Device, D2D) communication, Machine to Machine (M2M) communication, Machine Type Communication (MTC), Vehicle to Vehicle (V2V) communication and Vehicle to Everything (V2X) system, etc. The embodiments of the present application may also be applied to these communication systems.
图4示出了本申请一个示例性实施例提供的信道信息反馈模型的训练方法的流程图。该方法可以应用于如图3示出的通信系统中的终端设备中,该方法包括:Fig. 4 shows a flowchart of a method for training a channel information feedback model provided by an exemplary embodiment of the present application. The method can be applied to a terminal device in a communication system as shown in FIG. 3, and the method includes:
步骤410:对初始信道信息进行掩码操作,得到掩码信道信息。Step 410: Perform a masking operation on the initial channel information to obtain masked channel information.
其中,初始信道信息是终端设备进行信道估计后,确定出的信道信息。Wherein, the initial channel information is channel information determined after the terminal device performs channel estimation.
其中,掩码操作指的是对部分信息进行屏蔽,从而降低冗余信息的操作。Wherein, the masking operation refers to an operation of masking part of information to reduce redundant information.
可以理解的是,信道信息具有高度冗余的特点,为解决这种情况,可以通过加掩码的方式,屏蔽部分信道信息来降低冗余信息。图像处理领域中,视觉图像也具有高度冗余的特点,例如,缺失的像素信息可以从相邻的像素块中恢复。在本申请实施例中,提出一种利用掩码操作的方式掩藏初始信道信息,从而降低冗余信息。It can be understood that channel information is characterized by high redundancy. In order to solve this situation, some channel information can be shielded by adding a mask to reduce redundant information. In the field of image processing, visual images are also characterized by high redundancy, for example, missing pixel information can be recovered from adjacent pixel blocks. In the embodiment of the present application, a method of using a mask operation to hide initial channel information is proposed, thereby reducing redundant information.
示例性的,掩码操作示意如图5所示。对初始信道信息H执行掩码操作后,得到掩码信道信息H’,初始信道信息H相较于掩码信道信息H’而言,冗余信息更多。可以理解的是,图5中仅为示例性的说明,在实际中,信道信息可以并不如同于图5所示的图像的呈现形式。Exemplarily, a schematic diagram of a masking operation is shown in FIG. 5 . After performing the masking operation on the initial channel information H, the masked channel information H' is obtained. Compared with the masked channel information H', the initial channel information H has more redundant information. It can be understood that FIG. 5 is only an exemplary illustration, and in practice, channel information may not be presented in the same form as the image shown in FIG. 5 .
步骤420:将掩码信道信息输入信道信息反馈模型,输出恢复信道信息。Step 420: Input the masked channel information into the channel information feedback model, and output the restored channel information.
其中,信道信息反馈模型是用于对输入的信道信息进行压缩反馈,并对压缩后的信道信息进行重构恢复的模型。Wherein, the channel information feedback model is a model for compressing and feeding back the input channel information, and reconstructing and recovering the compressed channel information.
在本申请实施例中,在对初始信道信息进行掩码操作得到掩码信道信息后,将掩码信道信息作为信道信息反馈模型的输入,利用信道信息反馈模型预测屏蔽的信道信息,从而输出恢复信道信息。In the embodiment of the present application, after the initial channel information is masked to obtain the masked channel information, the masked channel information is used as the input of the channel information feedback model, and the channel information feedback model is used to predict the masked channel information, so that the output recovery channel information.
可选的,信道信息反馈模型是编码器-解码器的形式。Optionally, the channel information feedback model is in the form of an encoder-decoder.
示例性的,结合参考图6,其示出了使用编码器-解码器形式的信道信息反馈模型对信道信息进行处理的示意图。在当前反馈周期中,发送端进行信道估计后,通过编码器对估计后的信道信息H进行压缩编码,并通过空口的反馈链路反馈给接收端。更详细地说,空口的反馈链路实际传输的是一个反馈向量,这个反馈向量是发送端编码器的神经网络输出得到的,并作为接收端神经网络的部分输入,用于接收端进行信道信息恢复。Exemplarily, refer to FIG. 6 , which shows a schematic diagram of processing channel information using an encoder-decoder channel information feedback model. In the current feedback cycle, after the transmitting end performs channel estimation, the encoder compresses and encodes the estimated channel information H, and feeds it back to the receiving end through the feedback link of the air interface. In more detail, the feedback link of the air interface actually transmits a feedback vector, which is obtained from the output of the neural network of the encoder at the transmitting end, and is used as part of the input of the neural network at the receiving end for channel information at the receiving end. recover.
步骤430:基于恢复信道信息与初始信道信息之间的误差,对信道信息反馈模型进行训练。Step 430: Based on the error between the recovered channel information and the initial channel information, train the channel information feedback model.
在得到信道信息反馈模型输出的恢复信道信息之后,将恢复信道信息与对应的初始信道信息进行比对,以判断信道信息反馈模型预测的初始信道信息中被掩码的内容的准确度,并在恢复信道信息与对应的初始信道信息存在误差时,根据存在的误差对信道信息反馈模型进行相应的修正,以使生成的信道信息反馈模型具备对信道信息的重构恢复能力。After obtaining the restored channel information output by the channel information feedback model, compare the restored channel information with the corresponding initial channel information to judge the accuracy of the masked content in the initial channel information predicted by the channel information feedback model, and then When there is an error between the restored channel information and the corresponding initial channel information, the channel information feedback model is corrected according to the existing error, so that the generated channel information feedback model has the ability to reconstruct and restore the channel information.
综上所述,本实施例提供的技术方案,在执行基于深度学习的信道信息反馈方案的情况下,在进行模型训练时,利用掩码操作屏蔽部分的初始信道信息,降低信道信息反馈模型训练时输入的冗余信息,减小模型训练的资源开销,加速模型的训练速度,提高训练模型的泛化能力。To sum up, the technical solution provided by this embodiment, in the case of implementing the channel information feedback scheme based on deep learning, uses the mask operation to shield part of the initial channel information during model training, reducing the channel information feedback model training. The redundant information input at the time reduces the resource overhead of model training, accelerates the training speed of the model, and improves the generalization ability of the training model.
下面,对掩码操作进行进一步说明。Next, the masking operation will be further described.
图7示出了本申请一个示例性实施例提供的信道信息反馈模型的训练方法的流程图。该方法可以应用于如图3示出的通信系统中的终端设备中,该方法包括:Fig. 7 shows a flowchart of a method for training a channel information feedback model provided by an exemplary embodiment of the present application. The method can be applied to a terminal device in a communication system as shown in FIG. 3, and the method includes:
步骤710:将用于表示初始信道信息的信道矩阵划分为非重叠的多个矩阵块。Step 710: Divide the channel matrix used to represent the initial channel information into multiple non-overlapping matrix blocks.
其中,划分后的每个矩阵块对应的矩阵尺寸信息相同。Wherein, the matrix size information corresponding to each divided matrix block is the same.
示例性的,用于表示初始信道信息的信道矩阵为25*25的矩阵,将该矩阵划分为25个5*5的矩阵块。Exemplarily, the channel matrix used to represent the initial channel information is a 25*25 matrix, and the matrix is divided into 25 5*5 matrix blocks.
步骤720:为矩阵块生成位置索引,组成矩阵块序列。Step 720: Generate position indices for the matrix blocks to form a sequence of matrix blocks.
其中,位置索引是用于表征每个矩阵块在矩阵块序列中的位置的索引。Wherein, the position index is an index used to characterize the position of each matrix block in the matrix block sequence.
示例性的,25个矩阵块分别对应0,1,…,24的位置索引,进而组成矩阵块序列。Exemplarily, 25 matrix blocks correspond to position indices of 0, 1, .
步骤730:对矩阵块序列进行采样,并屏蔽矩阵块序列中未被采样的矩阵块,得到掩码信道信息。Step 730: Sampling the matrix block sequence, and masking unsampled matrix blocks in the matrix block sequence to obtain masked channel information.
也即,对矩阵块序列中被采样的矩阵块进行保留,对未被采样的矩阵块进行删除,从而得到掩码信道信息。That is, the sampled matrix blocks in the matrix block sequence are retained, and the unsampled matrix blocks are deleted, so as to obtain masked channel information.
可选的,采样对应的采样方式包括:随机采样;或,栅格采样。也即,掩码操作的选择方案包括随机掩码策略和栅格掩码策略。可以理解的是,掩码操作的选择方案不局限于上述两种掩码策略,例如利用其他先验知识来设置掩码分布,都在本申请的保护范围之内。其中,栅格采样可以是等间隔的栅格采样。Optionally, the sampling manner corresponding to the sampling includes: random sampling; or grid sampling. That is, the selection scheme of the masking operation includes a random masking strategy and a grid masking strategy. It can be understood that the selection scheme of the masking operation is not limited to the above two masking strategies, for example, using other prior knowledge to set the mask distribution is within the protection scope of the present application. Wherein, the grid sampling may be grid sampling at equal intervals.
示例性的,图5所示的掩码操作对应的采样为随机采样,图8所示的掩码操作对应的采样为栅格采样。Exemplarily, the sampling corresponding to the mask operation shown in FIG. 5 is random sampling, and the sampling corresponding to the mask operation shown in FIG. 8 is grid sampling.
示例性的,终端设备按照均匀分布对矩阵块序列进行随机采样,采样率取50%。示例性的,终端设备照均匀分布对矩阵块序列进行栅格采样,采样率取25%。Exemplarily, the terminal device randomly samples the matrix block sequence according to uniform distribution, and the sampling rate is 50%. Exemplarily, the terminal device performs grid sampling on the matrix block sequence according to uniform distribution, and the sampling rate is 25%.
可以理解的是,上述采样率仅为示例性的说明,本申请实施例对采样率的数值不加以限制。示例性的,在本侧的信道信息的数量较多的情况下,则采取较小的采样率;在本侧的信道信息的数量较少的情况下,则采取较大的采样率。It can be understood that the foregoing sampling rate is only an exemplary description, and this embodiment of the present application does not limit the numerical value of the sampling rate. Exemplarily, when the amount of channel information on the local side is large, a smaller sampling rate is adopted; when the amount of channel information on the local side is small, a larger sampling rate is adopted.
步骤740:将掩码信道信息输入信道信息反馈模型,输出恢复信道信息。Step 740: Input the masked channel information into the channel information feedback model, and output the restored channel information.
本步骤的具体实施方式参见上述步骤420,在此不再赘述。For the specific implementation manner of this step, refer to the above-mentioned step 420, which will not be repeated here.
步骤750:基于恢复信道信息与初始信道信息之间的误差,对信道信息反馈模型进行训练。Step 750: Based on the error between the restored channel information and the initial channel information, train the channel information feedback model.
本步骤的具体实施方式参见上述步骤430,在此不再赘述。For the specific implementation manner of this step, refer to the above-mentioned step 430, which will not be repeated here.
综上所述,本实施例提供的技术方案,提供了随机掩码策略和栅格掩码策略等不同的掩码策略来执行掩码操作,保障掩码操作的合理性。To sum up, the technical solution provided by this embodiment provides different masking strategies such as a random masking strategy and a grid masking strategy to perform a masking operation to ensure the rationality of the masking operation.
示例性的,结合上述的掩码操作以及编码器-解码器的模型结构,本实施例的整体架构可以如图9所示。Exemplarily, in combination with the above mask operation and the encoder-decoder model structure, the overall architecture of this embodiment may be shown in FIG. 9 .
在图9中,主要示出了如下操作流程:掩码操作、编码器、码字填充和解码器。In FIG. 9 , the following operation flow is mainly shown: mask operation, encoder, codeword stuffing and decoder.
掩码操作:初始信道信息对应的信道矩阵H在进行掩码操作后,得到掩码信道信息H’。Masking operation: After the channel matrix H corresponding to the initial channel information is masked, the masked channel information H' is obtained.
编码器:掩码信道信息H’输入编码器,进行压缩编码得到压缩编码信息。Encoder: The masked channel information H' is input to the encoder, and compression coding is performed to obtain compressed coding information.
码字填充:对压缩编码信息进行码字填充,得到填充后的压缩编码信息,也即,得到完整的压缩编码信息。Codeword stuffing: performing codeword stuffing on the compressed coded information to obtain the filled compressed coded information, that is, to obtain complete compressed coded information.
解码器:填充后的压缩编码信息输入解码器进行解压缩,得到恢复信道信息H”。Decoder: The filled compressed coded information is input to the decoder for decompression to obtain the restored channel information H".
相应的,若本侧的信道信息反馈模型包括:第一编码器,第一解码器;将掩码信道信息输入信道信息反馈模型,得到恢复信道信息这一步骤包括:Correspondingly, if the channel information feedback model on the local side includes: a first encoder and a first decoder; inputting the masked channel information into the channel information feedback model, the step of obtaining the restored channel information includes:
(1)将掩码信道信息作为第一编码器的模型输入,经由第一编码器对掩码信道信息进行压缩处理,得到压缩编码信息。(1) The masked channel information is input as a model of the first coder, and the masked channel information is compressed through the first coder to obtain compressed coded information.
(2)对压缩编码信息进行码字填充,得到填充后的压缩编码信息。(2) Filling the compressed coded information with codewords to obtain the filled compressed coded information.
码字填充指的是对屏蔽位置处的码字进行填充。Codeword filling refers to filling codewords at masked positions.
示例性的,第一编码器得到的压缩编码信息是对信道矩阵相应的矩阵块序列中的可见矩阵块的编码,则基于位置索引,在屏蔽的对应位置对码字做填充。Exemplarily, the compressed coding information obtained by the first encoder is the coding of the visible matrix blocks in the matrix block sequence corresponding to the channel matrix, and based on the position index, the codeword is filled in the corresponding position of the mask.
(3)将填充后的压缩编码信息作为第一解码器的模型输入,经由第一解码器对填充后的压缩编码信 息进行解压缩处理,得到恢复信道信息。(3) The compressed coded information after filling is input as the model of the first decoder, and the compressed coded information after filling is decompressed through the first decoder to obtain the restored channel information.
在示例性实施例中,终端设备对信道信息反馈模型的训练对应于预训练-微调模式的迁移学习中的预训练阶段,终端设备为源侧终端,该源侧终端还需要将在本侧预训练好的编码器进行上传,由目标侧终端执行微调阶段,训练好解码器并上传给网络设备。In an exemplary embodiment, the training of the channel information feedback model by the terminal device corresponds to the pre-training stage in the transfer learning of the pre-training-fine-tuning mode. The terminal device is a source-side terminal, and the source-side terminal also needs to The trained encoder is uploaded, and the target terminal performs the fine-tuning stage, and the trained decoder is uploaded to the network device.
图10示出了本申请一个示例性实施例提供的信道信息反馈模型的训练方法的流程图。该方法可以应用于如图3示出的通信系统中,该方法包括:Fig. 10 shows a flowchart of a method for training a channel information feedback model provided by an exemplary embodiment of the present application. The method can be applied to a communication system as shown in FIG. 3, and the method includes:
步骤1010:在信道信息反馈模型的训练完成后,源侧终端向网络设备发送信道信息反馈模型的第一迁移学习信息。Step 1010: After the training of the channel information feedback model is completed, the terminal at the source side sends the first transfer learning information of the channel information feedback model to the network device.
相应的,网络设备接收源侧终端发送的信道信息反馈模型的第一迁移学习信息。第一迁移学习信息用于对信道信息反馈模型进行迁移学习。Correspondingly, the network device receives the first transfer learning information of the channel information feedback model sent by the source-side terminal. The first transfer learning information is used to perform transfer learning on the channel information feedback model.
其中,源侧终端处的信道信息反馈模型包括:第一编码器,第一迁移学习信息包括:Wherein, the channel information feedback model at the terminal at the source side includes: a first encoder, and the first transfer learning information includes:
·第一编码器。• A first encoder.
也即,第一迁移学习信息中携带第一编码器的模型参数。That is, the first transfer learning information carries model parameters of the first encoder.
·掩码操作对应的矩阵尺寸信息。• The matrix size information corresponding to the mask operation.
也即,第一迁移学习信息中携带源侧终端执行掩码操作对应的矩阵尺寸信息。其中,矩阵尺寸信息用于指示输入信道信息反馈模型的矩阵块序列中的每个矩阵块的尺寸。That is, the first transfer learning information carries matrix size information corresponding to the mask operation performed by the terminal at the source side. Wherein, the matrix size information is used to indicate the size of each matrix block in the matrix block sequence of the input channel information feedback model.
步骤1020:网络设备向目标侧终端发送第二迁移学习信息。Step 1020: The network device sends second transfer learning information to the target terminal.
相应的,目标侧终端接收网络设备发送的第二迁移学习信息。第二迁移学习信息用于辅助进行迁移学习。Correspondingly, the target-side terminal receives the second migration learning information sent by the network device. The second transfer learning information is used to assist transfer learning.
其中,第二迁移学习信息包括:Wherein, the second transfer learning information includes:
·第二编码器。• Second encoder.
也即,第二迁移学习信息中携带第二编码器的模型参数。其中,第二编码器是基于掩码操作进行训练得到的。That is, the second transfer learning information carries the model parameters of the second encoder. Wherein, the second encoder is obtained by training based on the mask operation.
·掩码操作对应的矩阵尺寸信息。• The matrix size information corresponding to the mask operation.
也即,第二迁移学习信息中携带源侧终端执行掩码操作对应的矩阵尺寸信息。其中,矩阵尺寸信息用于指示输入信道信息反馈模型的矩阵块序列中的每个矩阵块的尺寸。That is, the second transfer learning information carries matrix size information corresponding to the mask operation performed by the terminal at the source side. Wherein, the matrix size information is used to indicate the size of each matrix block in the matrix block sequence of the input channel information feedback model.
可以理解的是,第二迁移学习信息中的掩码操作对应的矩阵尺寸信息即为第一迁移学习信息中的掩码操作对应的矩阵尺寸信息,第二迁移学习信息中的第二编码器是基于第一迁移学习信息中的第一编码器得到的。It can be understood that the matrix size information corresponding to the mask operation in the second transfer learning information is the matrix size information corresponding to the mask operation in the first transfer learning information, and the second encoder in the second transfer learning information is It is obtained based on the first encoder in the first transfer learning information.
步骤1030:目标侧终端生成第二解码器。Step 1030: The target-side terminal generates a second decoder.
也即,目标侧终端生成一个新的解码器。That is, the target side terminal generates a new decoder.
步骤1040:目标侧终端基于矩阵尺寸信息,对第二编码器和第二解码器进行联合训练。Step 1040: The target terminal performs joint training on the second encoder and the second decoder based on the matrix size information.
也即,目标终端利用预训练好的第二编码器,将第二编码器和新的第二解码器在新的数据集下一起联合训练,以完成迁移学习。That is, the target terminal uses the pre-trained second encoder to jointly train the second encoder and the new second decoder under the new data set, so as to complete transfer learning.
可以理解的是,预训练-微调模式的迁移学习,是指把预训练好一个网络,直接将其用于目标场景的数据,并在目标场景数据上进行微调,可以实现某一场景下已有的模型可以帮助其他场景实现同样的功能。在本申请实施例中,预训练好第二编码器,并利用预训练好的第二编码器与新的第二解码器一起再训练,从而节省目标侧终端的计算资源。It is understandable that the transfer learning of the pre-training-fine-tuning mode refers to the pre-training of a network, directly using it for the data of the target scene, and fine-tuning on the target scene data. The model can help other scenarios to achieve the same function. In the embodiment of the present application, the second encoder is pre-trained, and the pre-trained second encoder is used for retraining together with the new second decoder, thereby saving computing resources of the target-side terminal.
步骤1050:目标侧终端向网络设备发送训练好的第二解码器。Step 1050: The target-side terminal sends the trained second decoder to the network device.
相应的,网络设备接收目标侧终端发送的第二解码器,第二解码器是目标侧终端基于第二迁移学习信息进行迁移学习后,训练得到的。Correspondingly, the network device receives the second decoder sent by the target-side terminal, and the second decoder is trained by the target-side terminal after performing transfer learning based on the second transfer learning information.
可选的,在步骤1050之后,目标侧终端使用训练好的第二编码器,网络设备侧使用接收到的第二解码器,由目标侧终端作为信道信息的发送端,网络设备作为信道信息的接收端,使用目标侧终端的第二编 码器和网络设备侧的第二解码器实现基于深度学习的信道信息反馈方案。Optionally, after step 1050, the target-side terminal uses the trained second encoder, and the network device side uses the received second decoder, with the target-side terminal as the sender of the channel information, and the network device as the source of the channel information The receiving end implements a channel information feedback scheme based on deep learning by using the second encoder of the terminal on the target side and the second decoder on the network device side.
综上所述,本实施例提供的技术方案,对编码器-解码器形式的信道信息反馈模型在预训练-微调模式的迁移场景下进行增强设计,利用掩码操作来降低预训练阶段输入的冗余信息,加速模型的预训练速度,同时提高预训练模型的泛化能力,提升模型性能。To sum up, the technical solution provided by this embodiment is to enhance the design of the channel information feedback model in the form of encoder-decoder in the migration scenario of pre-training-fine-tuning mode, and use the mask operation to reduce the input in the pre-training stage. Redundant information accelerates the pre-training speed of the model, improves the generalization ability of the pre-trained model, and improves model performance.
下面,对目标侧终端基于矩阵尺寸信息,对第二编码器和第二解码器进行联合训练的方式进行说明。Next, a manner in which the target terminal performs joint training on the second encoder and the second decoder based on the matrix size information will be described.
(1)按照矩阵尺寸信息将用于表示初始信道信息的信道矩阵划分为非重叠的多个矩阵块,多个矩阵块组成矩阵块序列。(1) Divide the channel matrix used to represent the initial channel information into multiple non-overlapping matrix blocks according to the matrix size information, and the multiple matrix blocks form a matrix block sequence.
其中,矩阵尺寸信息用于指示输入信道信息反馈模型的矩阵块序列中的每个矩阵块的尺寸。Wherein, the matrix size information is used to indicate the size of each matrix block in the matrix block sequence of the input channel information feedback model.
示例性的,矩阵尺寸信息指示矩阵块的尺寸为5*5,则目标侧终端相应将本侧的信道信息对应的信道矩阵划分为5*5的多个矩阵块。Exemplarily, the matrix size information indicates that the size of the matrix block is 5*5, and then the terminal on the target side divides the channel matrix corresponding to the channel information on its own side into multiple matrix blocks of 5*5.
(2)将矩阵块序列作为第二编码器的模型输入,经由第二编码器对矩阵块序列进行压缩处理,得到压缩编码信息。(2) The matrix block sequence is input as a model of the second encoder, and the matrix block sequence is compressed through the second encoder to obtain compressed coding information.
(3)将压缩编码信息作为第二解码器的模型输入,经由第二解码器对压缩编码信息进行解压缩处理,得到恢复信道信息。(3) The compressed coded information is input as a model of the second decoder, and the compressed coded information is decompressed via the second decoder to obtain restored channel information.
(4)基于恢复信道信息与初始信道信息之间的误差,对第二编码器和第二解码器进行联合训练。(4) Jointly train the second encoder and the second decoder based on the error between the recovered channel information and the initial channel information.
如上述步骤所示,迁移到目标域以后,目标域的第二编码器支持可变长度的序列输入,第二编码器输入的是一个完整信道的矩阵块序列,不要做掩码处理,相应的,第二编码器输出压缩编码信息后,不需要再做码字填充,以便充分利用当前场景的有限数据。As shown in the above steps, after migrating to the target domain, the second encoder in the target domain supports variable-length sequence input, and the input of the second encoder is a matrix block sequence of a complete channel, do not perform mask processing, and correspondingly , after the second encoder outputs the compressed encoding information, it does not need to perform codeword padding, so as to make full use of the limited data of the current scene.
在一种可能的实现方式中,第二编码器是由一个源侧终端指示给网络设备的编码器。In a possible implementation manner, the second encoder is an encoder indicated to the network device by a source-side terminal.
也即,网络设备在接收到源侧终端发送的第一迁移学习信息后,直接将第一迁移学习信息作为第二迁移学习信息发送给目标侧终端设备。上述实施例中的第二迁移学习信息中的第二编码器等同于第一迁移学习信息中的第一编码器。That is, after receiving the first transfer learning information sent by the source-side terminal, the network device directly sends the first transfer-learning information as the second transfer-learning information to the target-side terminal device. The second encoder in the second transfer learning information in the above embodiment is equivalent to the first encoder in the first transfer learning information.
综上所述,本实施例提供的技术方案,由一个源侧终端预训练得到编码器,并将该编码器迁移到目标侧终端,利用掩码操作来降低预训练阶段输入的冗余信息,加速了模型的预训练速度。To sum up, in the technical solution provided by this embodiment, an encoder is pre-trained by a source-side terminal, and the encoder is migrated to a target-side terminal, and the redundant information input in the pre-training stage is reduced by using a mask operation. Accelerated the pre-training speed of the model.
示例性的,结合图11对上述实现方式进行示例性的说明,如图11所示,执行如下步骤:Exemplarily, the above-mentioned implementation manner is exemplarily described in conjunction with FIG. 11 , as shown in FIG. 11 , the following steps are performed:
步骤1101:源侧终端获取信道数据并执行掩码策略。Step 1101: The source-side terminal obtains channel data and executes a masking strategy.
示例性的,源侧终端将信道数据划分为规则的非重叠N个小块矩阵(补丁)。为每个矩阵块生成一个位置索引0,1,2,...,N-1,组成矩阵块序列。然后对矩阵块序列进行采样,并屏蔽(即删除)剩余的矩阵块。Exemplarily, the terminal at the source side divides the channel data into regular non-overlapping N small block matrices (patches). A position index 0, 1, 2, ..., N-1 is generated for each matrix block, forming a sequence of matrix blocks. The sequence of matrix blocks is then sampled and the remaining matrix blocks are masked (ie deleted).
步骤1102:源侧终端联合训练编码器和解码器。Step 1102: The source-side terminal jointly trains the encoder and the decoder.
示例性的,将经过掩码后的信道信息作为编码器的输入,并相应在编码器后增加一个码字填充操作,解码器的输入是填充后的码字,包括可见的矩阵块码字和掩码对应位置的填充码字。Exemplarily, the masked channel information is used as the input of the encoder, and a codeword filling operation is added after the encoder accordingly, and the input of the decoder is the filled codeword, including the visible matrix block codeword and The padding codeword at the corresponding position of the mask.
在本实施例中,解码器和编码器可以采用非对称设计,相较于编码器的参数规模,解码器可以适当地减小网络层数和参数量,从而减少预训练时间。In this embodiment, the decoder and encoder can adopt an asymmetric design. Compared with the parameter scale of the encoder, the decoder can appropriately reduce the number of network layers and parameters, thereby reducing the pre-training time.
步骤1103:源侧终端向网络设备发送编码器和掩码操作对应的矩阵尺寸信息。Step 1103: The terminal at the source side sends the matrix size information corresponding to the encoder and the mask operation to the network device.
步骤1104:网络设备向目标侧终端发送编码器和掩码操作对应的矩阵尺寸信息。Step 1104: The network device sends the matrix size information corresponding to the encoder and the mask operation to the target terminal.
步骤1105:目标侧终端生成新的解码器。Step 1105: The target terminal generates a new decoder.
步骤1106:目标侧终端按照矩阵尺寸信息将信道信息处理为编码器适配的矩阵块序列。Step 1106: The terminal on the target side processes the channel information into a matrix block sequence adapted by the encoder according to the matrix size information.
步骤1107:目标侧终端利用预训练的编码器联合新的解码器,在新的数据集下再训练以完成模型迁移。Step 1107: The target side terminal uses the pre-trained encoder to combine with the new decoder, and retrains under the new data set to complete model migration.
步骤1108:目标侧终端将解码器同步给网络设备。Step 1108: the target terminal synchronizes the decoder to the network device.
在另一种可能的实现方式中,第二编码器是由网络设备对多个编码器的模型参数进行聚合计算后,得到的全局编码器,多个编码器分别来自于多个源侧终端。In another possible implementation manner, the second encoder is a global encoder obtained by aggregate calculation of model parameters of multiple encoders by the network device, and the multiple encoders come from multiple source-side terminals respectively.
图12示出了本申请一个示例性实施例提供的信道信息反馈模型的训练方法的流程图。该方法可以应用于如图3示出的通信系统中,该方法包括:Fig. 12 shows a flowchart of a method for training a channel information feedback model provided by an exemplary embodiment of the present application. The method can be applied to a communication system as shown in FIG. 3, and the method includes:
步骤1210:在信道信息反馈模型的训练完成后,多个源侧终端分别向网络设备发送信道信息反馈模型的第一迁移学习信息。Step 1210: After the training of the channel information feedback model is completed, multiple source-side terminals respectively send the first transfer learning information of the channel information feedback model to the network device.
相应的,网络设备接收多个源侧终端分别发送的信道信息反馈模型的第一迁移学习信息。第一迁移学习信息用于对信道信息反馈模型进行迁移学习。Correspondingly, the network device receives the first transfer learning information of the channel information feedback model respectively sent by multiple source-side terminals. The first transfer learning information is used to perform transfer learning on the channel information feedback model.
其中,源侧终端处的信道信息反馈模型包括:第一编码器,第一迁移学习信息包括:Wherein, the channel information feedback model at the terminal at the source side includes: a first encoder, and the first transfer learning information includes:
·第一编码器。• A first encoder.
也即,第一迁移学习信息中携带第一编码器的模型参数。That is, the first transfer learning information carries model parameters of the first encoder.
·掩码操作对应的矩阵尺寸信息。• The matrix size information corresponding to the mask operation.
也即,第一迁移学习信息中携带源侧终端执行掩码操作对应的矩阵尺寸信息。其中,矩阵尺寸信息用于指示输入信道信息反馈模型的矩阵块序列中的每个矩阵块的尺寸。That is, the first transfer learning information carries matrix size information corresponding to the mask operation performed by the terminal at the source side. Wherein, the matrix size information is used to indicate the size of each matrix block in the matrix block sequence of the input channel information feedback model.
可选的,为了统一掩码操作,在步骤1210之前,网络设备向多个源侧终端下发相同的掩码策略参数,掩码策略参数是与掩码操作相关的参数。Optionally, in order to unify the masking operation, before step 1210, the network device delivers the same masking policy parameter to multiple source-side terminals, where the masking policy parameter is a parameter related to the masking operation.
可选的,掩码策略参数包括如下中的至少一种:Optionally, the mask policy parameters include at least one of the following:
·掩码操作对应的矩阵尺寸信息。• The matrix size information corresponding to the mask operation.
其中,矩阵尺寸信息用于指示输入信道信息反馈模型的矩阵块序列中的每个矩阵块的尺寸。Wherein, the matrix size information is used to indicate the size of each matrix block in the matrix block sequence of the input channel information feedback model.
·掩码操作对应的采样信息。• The sampling information corresponding to the mask operation.
其中,采样信息用于指示掩码操作中的采样的执行方式。示例性的,采样信息包括如下中的至少一种:采样方式;采样率。Wherein, the sampling information is used to indicate the execution mode of sampling in the mask operation. Exemplarily, the sampling information includes at least one of the following: sampling mode; sampling rate.
步骤1220:网络设备对多个训练好的第一编码器的模型参数进行聚合计算,得到全局编码器。Step 1220: The network device aggregates and calculates the model parameters of multiple trained first encoders to obtain a global encoder.
其中,聚合计算指的是对一组值进行计算,并返回单个值的一种计算方式。在本申请实施例中,对多个第一编码器的模型参数进行聚合计算,并返回最终的一个全局编码器的模型参数,本申请实施例对聚合计算的具体实现形式不加以限制。Among them, aggregation calculation refers to a calculation method that calculates a set of values and returns a single value. In the embodiment of the present application, the model parameters of multiple first encoders are aggregated and calculated, and a final model parameter of a global encoder is returned. The embodiment of the present application does not limit the specific implementation form of the aggregated calculation.
步骤1230:网络设备向目标侧终端发送第二迁移学习信息。Step 1230: The network device sends the second transfer learning information to the target terminal.
相应的,目标侧终端接收网络设备发送的第二迁移学习信息。第二迁移学习信息用于辅助进行迁移学习。Correspondingly, the target-side terminal receives the second migration learning information sent by the network device. The second transfer learning information is used to assist transfer learning.
其中,第二迁移学习信息包括:全局编码器、掩码操作对应的矩阵尺寸信息,全局编码器是基于掩码操作进行训练得到的。Wherein, the second transfer learning information includes: the global encoder and matrix size information corresponding to the mask operation, and the global encoder is obtained by training based on the mask operation.
可以理解的是,第二迁移学习信息中的掩码操作对应的矩阵尺寸信息即为第一迁移学习信息中的掩码操作对应的矩阵尺寸信息,第二迁移学习信息中的全局编码器是基于多个第一迁移学习信息中的多个第一编码器进行聚合计算得到的。It can be understood that the matrix size information corresponding to the mask operation in the second transfer learning information is the matrix size information corresponding to the mask operation in the first transfer learning information, and the global encoder in the second transfer learning information is based on It is obtained by aggregate calculation of multiple first encoders in the multiple first transfer learning information.
步骤1240:目标侧终端生成第二解码器。Step 1240: The target-side terminal generates a second decoder.
也即,目标侧终端生成一个新的解码器。That is, the target side terminal generates a new decoder.
步骤1250:目标侧终端对全局编码器和第二解码器进行联合训练。Step 1250: The target-side terminal performs joint training on the global encoder and the second decoder.
也即,目标终端利用预训练好的全局编码器,将全局编码器和新的第二解码器在新的数据集下一起联合训练,以完成迁移学习。That is, the target terminal uses the pre-trained global encoder to jointly train the global encoder and the new second decoder under the new data set, so as to complete transfer learning.
步骤1260:目标侧终端向网络设备发送训练好的第二解码器。Step 1260: The target-side terminal sends the trained second decoder to the network device.
综上所述,本实施例提供的技术方案,多个源侧终端协作训练得到一个共享的全局编码器,多个终端设备下的数据冗余程度更高,利用掩码操作可以更加显著地减小数据冗余,有利于增强模型提取潜在特征的表征能力,同时加速模型的预训练速度。To sum up, in the technical solution provided by this embodiment, multiple source-side terminals cooperate to train to obtain a shared global encoder, and the data redundancy under multiple terminal devices is higher, and the mask operation can be used to significantly reduce Small data redundancy is conducive to enhancing the representation ability of the model to extract potential features, and at the same time speeding up the pre-training speed of the model.
示例性的,结合图13对上述实现方式进行示例性的说明,如图13所示,执行如下步骤:Exemplarily, the above-mentioned implementation manner is exemplarily described in conjunction with FIG. 13 , as shown in FIG. 13 , the following steps are performed:
步骤1301,统一掩码策略:网络设备统一配置掩码策略参数,然后统一下发到候选的n个源侧终端:源侧终端1,源侧终端2,...,源侧终端n。Step 1301, unify the masking strategy: the network device uniformly configures masking strategy parameters, and then distributes them to n candidate source-side terminals: source-side terminal 1, source-side terminal 2, . . . , source-side terminal n.
步骤1302,预训练编码器:源侧终端各自执行掩码操作,将屏蔽后的掩码信道信息作为输入,训练编码器-译码器。Step 1302, pre-training the encoder: the source-side terminals each perform a masking operation, and use the masked channel information as input to train the encoder-decoder.
对于单个终端设备,基于掩码策略的自编码器网络架构是一致的。组件包括:掩码操作、编码器、码字填充、解码器四部分。每个终端设备都需要以上四个组件,各个终端设备的工作架构和流程可以参见图9所示实施例,这里不再详细赘述。For a single end device, the masking strategy-based autoencoder network architecture is consistent. The components include: mask operation, encoder, codeword filling, and decoder. Each terminal device requires the above four components, and the working structure and flow of each terminal device can refer to the embodiment shown in FIG. 9 , which will not be described in detail here.
在本实施例中,解码器和编码器可以采用非对称设计,相较于编码器的参数规模,解码器可以适当地减小网络层数和参数量,从而减少预训练时间。In this embodiment, the decoder and encoder can adopt an asymmetric design. Compared with the parameter scale of the encoder, the decoder can appropriately reduce the number of network layers and parameters, thereby reducing the pre-training time.
步骤1303,上传编码器:各个源侧终端将解码器部分删除,以节省设备内存资源,只保留编码器部分,并上传编码器同步给网络设备。Step 1303, uploading the encoder: each source-side terminal deletes the decoder part to save device memory resources, retains only the encoder part, and uploads the encoder to the network device for synchronization.
步骤1304,聚合计算:基站服务器或者空中计算节点,对各个协作的源侧终端的编码器模型参数进行聚合计算,得到全局编码器。Step 1304, aggregation calculation: the base station server or the over-the-air computing node performs aggregation calculation on the encoder model parameters of each coordinated source-side terminal to obtain a global encoder.
步骤1305,下发全局编码器和矩阵尺寸信息:网络设备,如基站服务器或者空中计算节点,将全局编码器和掩码操作对应的矩阵尺寸信息下发给目标侧终端。Step 1305, delivering global encoder and matrix size information: network devices, such as base station servers or air computing nodes, deliver the global encoder and matrix size information corresponding to mask operations to the target terminal.
可以理解的是,目标侧终端可以是多个,而且不限于源侧终端,网络设备下的所有终端都可以作为候选的目标侧终端,具体要看系统策略。It can be understood that there may be multiple target-side terminals, and it is not limited to source-side terminals. All terminals under the network device can be used as candidate target-side terminals, depending on system policies.
步骤1306,微调阶段:目标侧终端利用矩阵尺寸信息,将已有信道信息数据处理成矩阵块序列,不需要做掩码,直接将完整矩阵块序列输入给编码器-解码器。Step 1306, fine-tuning stage: the terminal on the target side uses the matrix size information to process the existing channel information data into a matrix block sequence, without masking, and directly inputs the complete matrix block sequence to the encoder-decoder.
需要注意的是此处的编码器是全局编码器,但是目标侧终端需要重新生成一个初始化状态的解码器。这里的编码器模型大小可以适当增大参数规模,以求得到更好的译码性能。It should be noted that the encoder here is a global encoder, but the target-side terminal needs to regenerate an initialized decoder. The size of the encoder model here can appropriately increase the parameter scale in order to obtain better decoding performance.
步骤1307,上传编码器:编码器是一种最终需要部署在接收端的模型,所以目标侧终端还要将训练好的译码器发送给网络设备,保证网络设备可以对目标侧终端的编码器发出的码字做出正确解析,恢复成完整的信道信息。Step 1307, upload the encoder: the encoder is a model that needs to be deployed at the receiving end, so the target terminal must also send the trained decoder to the network device to ensure that the network device can send the encoder to the target terminal The codewords are correctly analyzed and restored to complete channel information.
如上述步骤所示,在实现迁移学习的过程中,各个参与方无需分享本地设备中的数据,充分保证了参与方的数据隐私性和安全性。As shown in the above steps, in the process of implementing transfer learning, each participant does not need to share the data in the local device, which fully guarantees the data privacy and security of the participants.
需要说明的是,上述方法实施例可以分别单独实施,也可以组合实施,本申请对此不进行限制。It should be noted that the foregoing method embodiments may be implemented individually or in combination, which is not limited in the present application.
在上述各个实施例中,由源侧终端执行的步骤可以单独实现成为源侧终端一侧的信道信息反馈模型的训练方法,由目标侧终端执行的步骤可以单独实现成为目标侧终端一侧的信道信息反馈模型的训练方法,由网络设备执行的步骤可以单独实现成为网络设备一侧的信道信息反馈模型的训练方法。In each of the above-mentioned embodiments, the steps performed by the source-side terminal can independently realize the training method of the channel information feedback model on the side of the source-side terminal, and the steps performed by the target-side terminal can independently realize the channel information on the side of the target-side terminal. In the training method of the information feedback model, the steps performed by the network device can be independently implemented as the training method of the channel information feedback model on the network device side.
图14示出了本申请一个示例性实施例提供的信道信息反馈模型的训练装置的结构框图,该装置可以实现成为源侧终端,或者,实现成为源侧终端中的一部分,该装置包括:掩码模块1402、模型处理模块1404和训练模块1406;Fig. 14 shows a structural block diagram of an apparatus for training a channel information feedback model provided by an exemplary embodiment of the present application. The apparatus can be implemented as a source-side terminal, or can be implemented as a part of a source-side terminal. The apparatus includes: code module 1402, model processing module 1404 and training module 1406;
所述掩码模块1402,用于对初始信道信息进行掩码操作,得到掩码信道信息;The masking module 1402 is configured to perform a masking operation on initial channel information to obtain masked channel information;
所述模型处理模块1404,用于将所述掩码信道信息输入所述信道信息反馈模型,输出恢复信道信息;The model processing module 1404 is configured to input the masked channel information into the channel information feedback model, and output restored channel information;
所述训练模块1406,用于基于所述恢复信道信息与所述初始信道信息之间的误差,对所述信道信息反馈模型进行训练。The training module 1406 is configured to train the channel information feedback model based on the error between the recovered channel information and the initial channel information.
在一个可选的实施例中,所述掩码模块1402,用于:In an optional embodiment, the masking module 1402 is configured to:
将用于表示所述初始信道信息的信道矩阵划分为非重叠的多个矩阵块;dividing the channel matrix used to represent the initial channel information into a plurality of non-overlapping matrix blocks;
为所述矩阵块生成位置索引,组成矩阵块序列;generating a position index for the matrix block to form a sequence of matrix blocks;
对所述矩阵块序列进行采样,并屏蔽所述矩阵块序列中未被采样的矩阵块,得到所述掩码信道信息。Sampling the matrix block sequence and masking unsampled matrix blocks in the matrix block sequence to obtain the masked channel information.
在一个可选的实施例中,所述采样对应的采样方式包括:In an optional embodiment, the sampling method corresponding to the sampling includes:
随机采样;random sampling;
或,or,
栅格采样。Raster sampling.
在一个可选的实施例中,所述信道信息反馈模型包括:第一编码器和第一解码器;In an optional embodiment, the channel information feedback model includes: a first encoder and a first decoder;
所述模型处理模块1404,用于:The model processing module 1404 is configured to:
将所述掩码信道信息作为所述第一编码器的模型输入,经由所述第一编码器对所述掩码信道信息进行压缩处理,得到压缩编码信息;inputting the masked channel information as a model of the first encoder, and compressing the masked channel information via the first encoder to obtain compressed encoded information;
对所述压缩编码信息进行码字填充,得到填充后的所述压缩编码信息;performing codeword padding on the compressed coded information to obtain the filled compressed coded information;
将所述填充后的所述压缩编码信息作为所述第一解码器的模型输入,经由所述第一解码器对所述填充后的所述压缩编码信息进行解压缩处理,得到所述恢复信道信息。inputting the filled compressed coded information as a model of the first decoder, decompressing the filled compressed coded information via the first decoder, to obtain the restored channel information.
在一个可选的实施例中,所述装置还包括:信息上报模块;In an optional embodiment, the device further includes: an information reporting module;
所述信息上报模块,用于在所述信道信息反馈模型的训练完成后,向网络设备发送所述信道信息反馈模型的第一迁移学习信息,所述第一迁移学习信息用于对所述信道信息反馈模型进行迁移学习。The information reporting module is configured to send the first migration learning information of the channel information feedback model to the network device after the training of the channel information feedback model is completed, and the first migration learning information is used for the channel Information feedback model for transfer learning.
在一个可选的实施例中,所述信道信息反馈模型包括:第一编码器,所述第一迁移学习信息包括:In an optional embodiment, the channel information feedback model includes: a first encoder, and the first transfer learning information includes:
所述第一编码器;said first encoder;
所述掩码操作对应的矩阵尺寸信息。The matrix size information corresponding to the mask operation.
在一个可选的实施例中,所述装置还包括:参数接收模块;In an optional embodiment, the device further includes: a parameter receiving module;
所述参数接收模块,用于接收网络设备下发的掩码策略参数,所述掩码策略参数是与所述掩码操作相关的参数。The parameter receiving module is configured to receive a masking policy parameter issued by a network device, and the masking policy parameter is a parameter related to the masking operation.
在一个可选的实施例中,所述掩码策略参数包括如下中的至少一种:In an optional embodiment, the mask policy parameters include at least one of the following:
所述掩码操作对应的矩阵尺寸信息;Matrix size information corresponding to the mask operation;
所述掩码操作对应的采样信息。The sampling information corresponding to the mask operation.
图15示出了本申请一个示例性实施例提供的信道信息反馈模型的训练装置的结构框图,该装置可以实现成为目标侧终端,或者,实现成为目标侧终端中的一部分,该装置包括:解码器生成模块1502、信息接收模块1504、训练模块1506和解码器发送模块1508;Fig. 15 shows a structural block diagram of a training device for a channel information feedback model provided by an exemplary embodiment of the present application. The device can be implemented as a target terminal, or can be implemented as a part of the target terminal. The device includes: decoding A decoder generating module 1502, an information receiving module 1504, a training module 1506 and a decoder sending module 1508;
所述解码器生成模块1502,用于生成所述第二解码器;The decoder generation module 1502, configured to generate the second decoder;
所述信息接收模块1504,用于接收网络设备发送的第二迁移学习信息,所述第二迁移学习信息用于辅助进行迁移学习,所述第二迁移学习信息包括:所述第二编码器和掩码操作对应的矩阵尺寸信息,所述第二编码器是基于所述掩码操作进行训练得到的;The information receiving module 1504 is configured to receive second transfer learning information sent by a network device, the second transfer learning information is used to assist transfer learning, and the second transfer learning information includes: the second encoder and Matrix size information corresponding to the mask operation, the second encoder is obtained by training based on the mask operation;
所述训练模块1506,用于基于所述矩阵尺寸信息,对所述第二编码器和所述第二解码器进行联合训练;The training module 1506 is configured to jointly train the second encoder and the second decoder based on the matrix size information;
所述解码器发送模块1508,用于向所述网络设备发送训练好的所述第二解码器。The decoder sending module 1508 is configured to send the trained second decoder to the network device.
在一个可选的实施例中,所述训练模块1506,用于:In an optional embodiment, the training module 1506 is used for:
按照所述矩阵尺寸信息将用于表示初始信道信息的信道矩阵划分为非重叠的多个矩阵块,所述多个矩阵块组成矩阵块序列;Divide the channel matrix used to represent the initial channel information into a plurality of non-overlapping matrix blocks according to the matrix size information, and the plurality of matrix blocks form a matrix block sequence;
将所述矩阵块序列作为所述第二编码器的模型输入,经由所述第二编码器对所述矩阵块序列进行压缩处理,得到压缩编码信息;The matrix block sequence is input as a model of the second encoder, and the matrix block sequence is compressed through the second encoder to obtain compressed encoding information;
将所述压缩编码信息作为所述第二解码器的模型输入,经由所述第二解码器对所述压缩编码信息进行解压缩处理,得到恢复信道信息;inputting the compressed coded information as a model of the second decoder, and decompressing the compressed coded information via the second decoder to obtain restored channel information;
基于所述恢复信道信息与所述初始信道信息之间的误差,对所述第二编码器和所述第二解码器进行联合训练。The second encoder and the second decoder are jointly trained based on an error between the recovered channel information and the initial channel information.
在一个可选的实施例中,所述第二编码器是由一个源侧终端指示给所述网络设备的编码器。In an optional embodiment, the second encoder is an encoder indicated to the network device by a source-side terminal.
在一个可选的实施例中,所述第二编码器是由所述网络设备对多个编码器的模型参数进行聚合计算后, 得到的全局编码器,所述多个编码器分别来自于多个所述源侧终端。In an optional embodiment, the second encoder is a global encoder obtained by aggregate calculation of model parameters of multiple encoders by the network device, and the multiple encoders come from multiple one of the source-side terminals.
图16示出了本申请一个示例性实施例提供的信道信息反馈模型的训练装置的结构框图,该装置可以实现成为网络设备,或者,实现成为网络设备中的一部分,该装置包括:信息发送模块1602和解码器接收模块1604;Fig. 16 shows a structural block diagram of a training device for a channel information feedback model provided by an exemplary embodiment of the present application. The device can be implemented as a network device, or can be implemented as a part of the network device. The device includes: an information sending module 1602 and a decoder receiving module 1604;
所述信息发送模块1602,用于向目标侧终端发送第二迁移学习信息,所述第二迁移学习信息用于辅助进行迁移学习,所述第二迁移学习信息包括:第二编码器和掩码操作对应的矩阵尺寸信息,所述第二编码器是基于所述掩码操作进行训练得到的;The information sending module 1602 is configured to send second transfer learning information to the target terminal, the second transfer learning information is used to assist in transfer learning, and the second transfer learning information includes: a second encoder and a mask Operating the corresponding matrix size information, the second encoder is obtained by training based on the mask operation;
所述解码器接收模块1604,用于接收所述目标侧终端发送的第二解码器,所述第二解码器是所述目标侧终端基于所述第二迁移学习信息进行迁移学习后,训练得到的。The decoder receiving module 1604 is configured to receive the second decoder sent by the target terminal, where the second decoder is obtained by training after the target terminal performs transfer learning based on the second transfer learning information. of.
在一个可选的实施例中,所述第二编码器是由一个源侧终端指示给网络设备的编码器;In an optional embodiment, the second encoder is an encoder indicated to the network device by a source-side terminal;
所述装置还包括:信息接收模块;The device also includes: an information receiving module;
所述信息接收模块,用于接收一个所述源侧终端发送的第一迁移学习信息,所述第一迁移学习信息用于辅助进行迁移学习,所述第一迁移学习信息包括:第一编码器和所述掩码操作对应的矩阵尺寸信息。The information receiving module is configured to receive a piece of first transfer learning information sent by the source-side terminal, the first transfer learning information is used to assist transfer learning, and the first transfer learning information includes: a first encoder Matrix size information corresponding to the mask operation.
在一个可选的实施例中,所述第二编码器是由网络设备对多个编码器的模型参数进行聚合计算后,得到的全局编码器;In an optional embodiment, the second encoder is a global encoder obtained by aggregate calculation of model parameters of multiple encoders by the network device;
所述装置还包括:信息接收模块和聚合计算模块;The device also includes: an information receiving module and an aggregation calculation module;
所述信息接收模块,用于接收多个所述源侧终端分别发送的多个第一迁移学习信息,所述第一迁移学习信息用于辅助进行迁移学习,所述第一迁移学习信息包括:第一编码器和所述掩码操作对应的矩阵尺寸信息;The information receiving module is configured to receive a plurality of first transfer learning information respectively sent by a plurality of the source-side terminals, the first transfer learning information is used to assist transfer learning, and the first transfer learning information includes: Matrix size information corresponding to the first encoder and the mask operation;
所述聚合计算模块,用于对多个所述训练好的第一编码器的模型参数进行聚合计算,得到所述全局编码器。The aggregation calculation module is configured to perform aggregation calculation on model parameters of multiple trained first encoders to obtain the global encoder.
在一个可选的实施例中,所述装置还包括:参数配置模块;In an optional embodiment, the device further includes: a parameter configuration module;
所述参数配置模块,用于向多个所述源侧终端下发相同的掩码策略参数,所述掩码策略参数是与所述掩码操作相关的参数。The parameter configuration module is configured to deliver the same masking policy parameter to multiple terminals at the source side, where the masking policy parameter is a parameter related to the masking operation.
在一个可选的实施例中,所述掩码策略参数包括如下中的至少一种:In an optional embodiment, the mask policy parameters include at least one of the following:
所述掩码操作对应的矩阵尺寸信息;Matrix size information corresponding to the mask operation;
所述掩码操作对应的采样信息。The sampling information corresponding to the mask operation.
图17示出了本申请一个示例性实施例提供的通信设备(终端设备或网络设备)的结构示意图,该通信设备1700包括:处理器1701、收发器1702和存储器1703。FIG. 17 shows a schematic structural diagram of a communication device (terminal device or network device) provided by an exemplary embodiment of the present application. The communication device 1700 includes: a processor 1701 , a transceiver 1702 and a memory 1703 .
处理器1701包括一个或者一个以上处理核心,处理器1701通过运行软件程序以及模块,从而执行各种功能应用。The processor 1701 includes one or more processing cores, and the processor 1701 executes various functional applications by running software programs and modules.
收发器1702可以用于进行信息的接收和发送,收发器1702可以是一块通信芯片。The transceiver 1702 can be used to receive and send information, and the transceiver 1702 can be a communication chip.
存储器1703可用于存储计算机程序,处理器1701用于执行该计算机程序,以实现上述方法实施例中通信设备执行的各个步骤。The memory 1703 may be used to store a computer program, and the processor 1701 is used to execute the computer program, so as to implement various steps performed by the communication device in the foregoing method embodiments.
此外,存储器1703可以由任何类型的易失性或非易失性存储设备或者它们的组合实现,易失性或非易失性存储设备包括但不限于:随机存储器(Random-Access Memory,RAM)和只读存储器(Read-Only Memory,ROM)、可擦写可编程只读存储器(Erasable Programmable Read-Only Memory,EPROM)、电可擦写可编程只读存储器(Electrically Erasable Programmable Read-Only Memory,EEPROM)、闪存或其他固态存储其技术,只读光盘(Compact Disc Read-Only Memory,CD-ROM)、高密度数字视频光盘(Digital Video Disc,DVD)或其他光学存储、磁带盒、磁带、磁盘存储或其他磁性存储设备。In addition, the memory 1703 can be realized by any type of volatile or non-volatile storage device or their combination, and the volatile or non-volatile storage device includes but not limited to: random access memory (Random-Access Memory, RAM) And read-only memory (Read-Only Memory, ROM), erasable programmable read-only memory (Erasable Programmable Read-Only Memory, EPROM), electrically erasable programmable read-only memory (Electrically Erasable Programmable Read-Only Memory, EEPROM), flash memory or other solid-state storage technologies, compact disc read-only memory (CD-ROM), high-density digital video disc (Digital Video Disc, DVD) or other optical storage, tape cartridges, tapes, disks storage or other magnetic storage devices.
其中,当通信设备实现为源侧终端时,本申请实施例涉及的中的处理器1701和收发器1702,可以执行上述实施例任一所示的方法中,由源侧终端执行的步骤,此处不再赘述。Wherein, when the communication device is implemented as a source-side terminal, the processor 1701 and the transceiver 1702 involved in the embodiment of the present application can perform the steps performed by the source-side terminal in any of the methods shown in the above-mentioned embodiments, where I won't repeat them here.
在一种可能的实现方式中,当通信设备实现为源侧终端时,In a possible implementation manner, when the communication device is implemented as a source-side terminal,
所述处理器1701,用于对初始信道信息进行掩码操作,得到掩码信道信息;The processor 1701 is configured to perform a masking operation on initial channel information to obtain masked channel information;
所述处理器1701,用于将所述掩码信道信息输入信道信息反馈模型,输出恢复信道信息;The processor 1701 is configured to input the masked channel information into a channel information feedback model, and output restored channel information;
所述处理器1701,用于基于所述恢复信道信息与所述初始信道信息之间的误差,对所述信道信息反馈模型进行训练。The processor 1701 is configured to train the channel information feedback model based on an error between the recovered channel information and the initial channel information.
其中,当通信设备实现为目标侧终端时,本申请实施例涉及的中的处理器1701和收发器1702,可以执行上述实施例任一所示的方法中,由目标侧终端执行的步骤,此处不再赘述。Wherein, when the communication device is implemented as a target-side terminal, the processor 1701 and the transceiver 1702 involved in the embodiment of the present application may perform the steps performed by the target-side terminal in any of the methods shown in the above-mentioned embodiments, where I won't repeat them here.
在一种可能的实现方式中,当通信设备实现为目标侧终端时,In a possible implementation manner, when the communication device is implemented as a target-side terminal,
所述处理器1701,用于生成第二解码器;The processor 1701 is configured to generate a second decoder;
所述收发器1702,用于接收网络设备发送的第二迁移学习信息,所述第二迁移学习信息用于辅助进行迁移学习,所述第二迁移学习信息包括:所述第二编码器和掩码操作对应的矩阵尺寸信息,所述第二编码器是基于所述掩码操作进行训练得到的;The transceiver 1702 is configured to receive second transfer learning information sent by a network device, the second transfer learning information is used to assist transfer learning, and the second transfer learning information includes: the second encoder and mask Matrix size information corresponding to the code operation, the second encoder is obtained by training based on the mask operation;
所述处理器1701,用于基于所述矩阵尺寸信息,对所述第二编码器和所述第二解码器进行联合训练;The processor 1701 is configured to jointly train the second encoder and the second decoder based on the matrix size information;
所述收发器1702,用于向所述网络设备发送训练好的所述第二解码器。The transceiver 1702 is configured to send the trained second decoder to the network device.
其中,当通信设备实现为网络设备时,本申请实施例涉及的中的处理器1701和收发器1702,可以执行上述实施例任一所示的方法中,由网络设备执行的步骤,此处不再赘述。Wherein, when the communication device is implemented as a network device, the processor 1701 and the transceiver 1702 involved in the embodiment of the present application can execute the steps performed by the network device in any of the methods shown in the above embodiments, which are not described here Let me repeat.
在一种可能的实现方式中,当通信设备实现为网络设备时,In a possible implementation manner, when the communication device is implemented as a network device,
所述收发器1702,用于向目标侧终端发送第二迁移学习信息,所述第二迁移学习信息用于辅助进行迁移学习,所述第二迁移学习信息包括:第二编码器和掩码操作对应的矩阵尺寸信息,所述第二编码器是基于所述掩码操作进行训练得到的;The transceiver 1702 is configured to send second transfer learning information to the target terminal, the second transfer learning information is used to assist transfer learning, and the second transfer learning information includes: a second encoder and a mask operation Corresponding matrix size information, the second encoder is obtained by training based on the mask operation;
所述收发器1702,用于接收所述目标侧终端发送的第二解码器,所述第二解码器是所述目标侧终端基于所述第二迁移学习信息进行迁移学习后,训练得到的。The transceiver 1702 is configured to receive the second decoder sent by the target-side terminal, where the second decoder is trained by the target-side terminal after performing transfer learning based on the second transfer learning information.
在示例性实施例中,还提供了一种计算机可读存储介质,所述计算机可读存储介质中存储有至少一条指令、至少一段程序、代码集或指令集,所述至少一条指令、所述至少一段程序、所述代码集或指令集由处理器加载并执行以实现上述各个方法实施例提供的由通信设备执行的信道信息反馈模型的训练方法。In an exemplary embodiment, a computer-readable storage medium is also provided, the computer-readable storage medium stores at least one instruction, at least one program, a code set or an instruction set, the at least one instruction, the At least one section of program, the code set or instruction set is loaded and executed by the processor to implement the training method of the channel information feedback model executed by the communication device provided in the above method embodiments.
在示例性实施例中,还提供了一种芯片,所述芯片包括可编程逻辑电路和/或程序指令,当所述芯片在计算机设备上运行时,用于实现上述方面所述的信道信息反馈模型的训练方法。In an exemplary embodiment, a chip is also provided, the chip includes a programmable logic circuit and/or program instructions, and when the chip is run on a computer device, it is used to realize the channel information feedback described in the above aspect The training method of the model.
在示例性实施例中,还提供了一种计算机程序产品,该计算机程序产品在计算机设备的处理器上运行时,使得计算机设备执行上述方面所述的信道信息反馈模型的训练方法。In an exemplary embodiment, a computer program product is also provided. When the computer program product runs on a processor of a computer device, the computer device executes the method for training a channel information feedback model described in the above aspects.
本领域普通技术人员可以理解实现上述实施例的全部或部分步骤可以通过硬件来完成,也可以通过程序来指令相关的硬件完成,所述的程序可以存储于一种计算机可读存储介质中,上述提到的存储介质可以是只读存储器,磁盘或光盘等。Those of ordinary skill in the art can understand that all or part of the steps for implementing the above embodiments can be completed by hardware, and can also be completed by instructing related hardware through a program. The program can be stored in a computer-readable storage medium. The above-mentioned The storage medium mentioned may be a read-only memory, a magnetic disk or an optical disk, and the like.
以上所述仅为本申请的可选实施例,并不用以限制本申请,凡在本申请的精神和原则之内,所作的任何修改、等同替换、改进等,均应包含在本申请的保护范围之内。The above are only optional embodiments of the application, and are not intended to limit the application. Any modifications, equivalent replacements, improvements, etc. made within the spirit and principles of the application shall be included in the protection of the application. within range.

Claims (40)

  1. 一种信道信息反馈模型的训练方法,其特征在于,所述方法包括:A method for training a channel information feedback model, characterized in that the method comprises:
    对初始信道信息进行掩码操作,得到掩码信道信息;Perform a masking operation on the initial channel information to obtain masked channel information;
    将所述掩码信道信息输入所述信道信息反馈模型,输出恢复信道信息;Input the masked channel information into the channel information feedback model, and output the restored channel information;
    基于所述恢复信道信息与所述初始信道信息之间的误差,对所述信道信息反馈模型进行训练。The channel information feedback model is trained based on an error between the restored channel information and the initial channel information.
  2. 根据权利要求1所述的方法,其特征在于,所述对初始信道信息进行掩码操作,得到掩码信道信息,包括:The method according to claim 1, wherein the masking operation is performed on the initial channel information to obtain the masked channel information, comprising:
    将用于表示所述初始信道信息的信道矩阵划分为非重叠的多个矩阵块;dividing the channel matrix used to represent the initial channel information into a plurality of non-overlapping matrix blocks;
    为所述矩阵块生成位置索引,组成矩阵块序列;generating a position index for the matrix block to form a sequence of matrix blocks;
    对所述矩阵块序列进行采样,并屏蔽所述矩阵块序列中未被采样的矩阵块,得到所述掩码信道信息。Sampling the matrix block sequence and masking unsampled matrix blocks in the matrix block sequence to obtain the masked channel information.
  3. 根据权利要求2所述的方法,其特征在于,所述采样对应的采样方式包括:The method according to claim 2, wherein the sampling method corresponding to the sampling comprises:
    随机采样;random sampling;
    或,or,
    栅格采样。Raster sampling.
  4. 根据权利要求1至3任一所述的方法,其特征在于,所述信道信息反馈模型包括:第一编码器和第一解码器;The method according to any one of claims 1 to 3, wherein the channel information feedback model comprises: a first encoder and a first decoder;
    所述将所述掩码信道信息输入所述信道信息反馈模型,得到恢复信道信息,包括:The inputting the masked channel information into the channel information feedback model to obtain the restored channel information includes:
    将所述掩码信道信息作为所述第一编码器的模型输入,经由所述第一编码器对所述掩码信道信息进行压缩处理,得到压缩编码信息;inputting the masked channel information as a model of the first encoder, and compressing the masked channel information via the first encoder to obtain compressed encoded information;
    对所述压缩编码信息进行码字填充,得到填充后的所述压缩编码信息;performing codeword padding on the compressed coded information to obtain the filled compressed coded information;
    将所述填充后的所述压缩编码信息作为所述第一解码器的模型输入,经由所述第一解码器对所述填充后的所述压缩编码信息进行解压缩处理,得到所述恢复信道信息。inputting the filled compressed coded information as a model of the first decoder, decompressing the filled compressed coded information via the first decoder, to obtain the restored channel information.
  5. 根据权利要求1至3任一所述的方法,其特征在于,所述方法还包括:The method according to any one of claims 1 to 3, wherein the method further comprises:
    在所述信道信息反馈模型的训练完成后,向网络设备发送所述信道信息反馈模型的第一迁移学习信息,所述第一迁移学习信息用于对所述信道信息反馈模型进行迁移学习。After the training of the channel information feedback model is completed, the first transfer learning information of the channel information feedback model is sent to the network device, where the first transfer learning information is used to perform transfer learning on the channel information feedback model.
  6. 根据权利要求5所述的方法,其特征在于,所述信道信息反馈模型包括:第一编码器,所述第一迁移学习信息包括:The method according to claim 5, wherein the channel information feedback model includes: a first encoder, and the first transfer learning information includes:
    所述第一编码器;said first encoder;
    所述掩码操作对应的矩阵尺寸信息。The matrix size information corresponding to the mask operation.
  7. 根据权利要求1至3任一所述的方法,其特征在于,所述方法还包括:The method according to any one of claims 1 to 3, wherein the method further comprises:
    接收网络设备下发的掩码策略参数,所述掩码策略参数是与所述掩码操作相关的参数。A masking policy parameter delivered by the network device is received, where the masking policy parameter is a parameter related to the masking operation.
  8. 根据权利要求7所述的方法,其特征在于,所述掩码策略参数包括如下中的至少一种:The method according to claim 7, wherein the mask policy parameters include at least one of the following:
    所述掩码操作对应的矩阵尺寸信息;Matrix size information corresponding to the mask operation;
    所述掩码操作对应的采样信息。The sampling information corresponding to the mask operation.
  9. 一种信道信息反馈模型的训练方法,其特征在于,所述信道信息反馈模型包括:第二编码器和第二解码器,所述方法包括:A method for training a channel information feedback model, wherein the channel information feedback model includes: a second encoder and a second decoder, and the method includes:
    生成所述第二解码器;generating said second decoder;
    接收网络设备发送的第二迁移学习信息,所述第二迁移学习信息用于辅助进行迁移学习,所述第二迁移学习信息包括:所述第二编码器和掩码操作对应的矩阵尺寸信息,所述第二编码器是基于所述掩码操作进行训练得到的;receiving second transfer learning information sent by the network device, where the second transfer learning information is used to assist in transfer learning, where the second transfer learning information includes: matrix size information corresponding to the second encoder and mask operation, The second encoder is obtained by training based on the mask operation;
    基于所述矩阵尺寸信息,对所述第二编码器和所述第二解码器进行联合训练;jointly training the second encoder and the second decoder based on the matrix size information;
    向所述网络设备发送训练好的所述第二解码器。sending the trained second decoder to the network device.
  10. 根据权利要求9所述的方法,其特征在于,所述基于矩阵尺寸信息,对所述第二编码器和所述第二 解码器进行联合训练,包括:The method according to claim 9, wherein the joint training of the second encoder and the second decoder based on the matrix size information includes:
    按照所述矩阵尺寸信息将用于表示初始信道信息的信道矩阵划分为非重叠的多个矩阵块,所述多个矩阵块组成矩阵块序列;Divide the channel matrix used to represent the initial channel information into a plurality of non-overlapping matrix blocks according to the matrix size information, and the plurality of matrix blocks form a matrix block sequence;
    将所述矩阵块序列作为所述第二编码器的模型输入,经由所述第二编码器对所述矩阵块序列进行压缩处理,得到压缩编码信息;The matrix block sequence is input as a model of the second encoder, and the matrix block sequence is compressed through the second encoder to obtain compressed encoding information;
    将所述压缩编码信息作为所述第二解码器的模型输入,经由所述第二解码器对所述压缩编码信息进行解压缩处理,得到恢复信道信息;inputting the compressed coded information as a model of the second decoder, and decompressing the compressed coded information via the second decoder to obtain restored channel information;
    基于所述恢复信道信息与所述初始信道信息之间的误差,对所述第二编码器和所述第二解码器进行联合训练。The second encoder and the second decoder are jointly trained based on an error between the recovered channel information and the initial channel information.
  11. 根据权利要求9或10所述的方法,其特征在于,The method according to claim 9 or 10, characterized in that,
    所述第二编码器是由一个源侧终端指示给所述网络设备的编码器。The second encoder is an encoder indicated to the network device by a source-side terminal.
  12. 根据权利要求9或10所述的方法,其特征在于,The method according to claim 9 or 10, characterized in that,
    所述第二编码器是由所述网络设备对多个编码器的模型参数进行聚合计算后,得到的全局编码器,所述多个编码器分别来自于多个所述源侧终端。The second encoder is a global encoder obtained after the network device aggregates and calculates model parameters of multiple encoders, and the multiple encoders come from multiple source-side terminals respectively.
  13. 一种信道信息反馈模型的训练方法,其特征在于,所述方法包括:A method for training a channel information feedback model, characterized in that the method comprises:
    向目标侧终端发送第二迁移学习信息,所述第二迁移学习信息用于辅助进行迁移学习,所述第二迁移学习信息包括:第二编码器和掩码操作对应的矩阵尺寸信息,所述第二编码器是基于所述掩码操作进行训练得到的;Sending second transfer learning information to the target side terminal, where the second transfer learning information is used to assist transfer learning, where the second transfer learning information includes: matrix size information corresponding to the second encoder and mask operation, the The second encoder is obtained by training based on the mask operation;
    接收所述目标侧终端发送的第二解码器,所述第二解码器是所述目标侧终端基于所述第二迁移学习信息进行迁移学习后,训练得到的。receiving the second decoder sent by the target-side terminal, where the second decoder is trained by the target-side terminal after performing transfer learning based on the second transfer learning information.
  14. 根据权利要求13所述的方法,其特征在于,所述第二编码器是由一个源侧终端指示给网络设备的编码器;The method according to claim 13, wherein the second encoder is an encoder indicated to the network device by a source-side terminal;
    所述方法还包括:The method also includes:
    接收一个所述源侧终端发送的第一迁移学习信息,所述第一迁移学习信息用于辅助进行迁移学习,所述第一迁移学习信息包括:第一编码器和所述掩码操作对应的矩阵尺寸信息。Receive a piece of first transfer learning information sent by the source-side terminal, where the first transfer learning information is used to assist transfer learning, where the first transfer learning information includes: a first encoder corresponding to the mask operation Matrix dimension information.
  15. 根据权利要求13所述的方法,其特征在于,所述第二编码器是由网络设备对多个编码器的模型参数进行聚合计算后,得到的全局编码器;The method according to claim 13, wherein the second encoder is a global encoder obtained after the network device aggregates and calculates the model parameters of multiple encoders;
    所述方法还包括:The method also includes:
    接收多个所述源侧终端分别发送的多个第一迁移学习信息,所述第一迁移学习信息用于辅助进行迁移学习,所述第一迁移学习信息包括:第一编码器和所述掩码操作对应的矩阵尺寸信息;receiving a plurality of first transfer learning information respectively sent by a plurality of source-side terminals, the first transfer learning information is used to assist transfer learning, and the first transfer learning information includes: the first encoder and the mask The matrix size information corresponding to the code operation;
    对多个所述训练好的第一编码器的模型参数进行聚合计算,得到所述全局编码器。Aggregate calculation is performed on the model parameters of multiple trained first encoders to obtain the global encoder.
  16. 根据权利要求15所述的方法,其特征在于,所述方法还包括:The method according to claim 15, further comprising:
    向多个所述源侧终端下发相同的掩码策略参数,所述掩码策略参数是与所述掩码操作相关的参数。delivering the same masking policy parameter to multiple source-side terminals, where the masking policy parameter is a parameter related to the masking operation.
  17. 根据权利要求16所述的方法,其特征在于,所述掩码策略参数包括如下中的至少一种:The method according to claim 16, wherein the mask policy parameters include at least one of the following:
    所述掩码操作对应的矩阵尺寸信息;Matrix size information corresponding to the mask operation;
    所述掩码操作对应的采样信息。The sampling information corresponding to the mask operation.
  18. 一种信道信息反馈模型的训练装置,其特征在于,所述装置包括:掩码模块、模型处理模块和训练模块;A training device for a channel information feedback model, characterized in that the device includes: a mask module, a model processing module and a training module;
    所述掩码模块,用于对初始信道信息进行掩码操作,得到掩码信道信息;The masking module is configured to perform a masking operation on initial channel information to obtain masked channel information;
    所述模型处理模块,用于将所述掩码信道信息输入所述信道信息反馈模型,输出恢复信道信息;The model processing module is configured to input the masked channel information into the channel information feedback model, and output restored channel information;
    所述训练模块,用于基于所述恢复信道信息与所述初始信道信息之间的误差,对所述信道信息反馈模型进行训练。The training module is configured to train the channel information feedback model based on the error between the recovered channel information and the initial channel information.
  19. 根据权利要求18所述的装置,其特征在于,所述掩码模块,用于:The device according to claim 18, wherein the mask module is configured to:
    将用于表示所述初始信道信息的信道矩阵划分为非重叠的多个矩阵块;dividing the channel matrix used to represent the initial channel information into a plurality of non-overlapping matrix blocks;
    为所述矩阵块生成位置索引,组成矩阵块序列;generating a position index for the matrix block to form a sequence of matrix blocks;
    对所述矩阵块序列进行采样,并屏蔽所述矩阵块序列中未被采样的矩阵块,得到所述掩码信道信息。Sampling the matrix block sequence and masking unsampled matrix blocks in the matrix block sequence to obtain the masked channel information.
  20. 根据权利要求19所述的装置,其特征在于,所述采样对应的采样方式包括:The device according to claim 19, wherein the sampling mode corresponding to the sampling comprises:
    随机采样;random sampling;
    或,or,
    栅格采样。Raster sampling.
  21. 根据权利要求18至20任一所述的装置,其特征在于,所述信道信息反馈模型包括:第一编码器和第一解码器;The device according to any one of claims 18 to 20, wherein the channel information feedback model comprises: a first encoder and a first decoder;
    所述模型处理模块,用于:The model processing module is used for:
    将所述掩码信道信息作为所述第一编码器的模型输入,经由所述第一编码器对所述掩码信道信息进行压缩处理,得到压缩编码信息;inputting the masked channel information as a model of the first encoder, and compressing the masked channel information via the first encoder to obtain compressed encoded information;
    对所述压缩编码信息进行码字填充,得到填充后的所述压缩编码信息;performing codeword padding on the compressed coded information to obtain the filled compressed coded information;
    将所述填充后的所述压缩编码信息作为所述第一解码器的模型输入,经由所述第一解码器对所述填充后的所述压缩编码信息进行解压缩处理,得到所述恢复信道信息。inputting the filled compressed coded information as a model of the first decoder, decompressing the filled compressed coded information via the first decoder, to obtain the restored channel information.
  22. 根据权利要求18至20任一所述的装置,其特征在于,所述装置还包括:信息上报模块;The device according to any one of claims 18 to 20, wherein the device further comprises: an information reporting module;
    所述信息上报模块,用于在所述信道信息反馈模型的训练完成后,向网络设备发送所述信道信息反馈模型的第一迁移学习信息,所述第一迁移学习信息用于对所述信道信息反馈模型进行迁移学习。The information reporting module is configured to send the first migration learning information of the channel information feedback model to the network device after the training of the channel information feedback model is completed, and the first migration learning information is used for the channel Information feedback model for transfer learning.
  23. 根据权利要求22所述的装置,其特征在于,所述信道信息反馈模型包括:第一编码器,所述第一迁移学习信息包括:The device according to claim 22, wherein the channel information feedback model includes: a first encoder, and the first transfer learning information includes:
    所述第一编码器;said first encoder;
    所述掩码操作对应的矩阵尺寸信息。The matrix size information corresponding to the mask operation.
  24. 根据权利要求18至20任一所述的装置,其特征在于,所述装置还包括:参数接收模块;The device according to any one of claims 18 to 20, wherein the device further comprises: a parameter receiving module;
    所述参数接收模块,用于接收网络设备下发的掩码策略参数,所述掩码策略参数是与所述掩码操作相关的参数。The parameter receiving module is configured to receive a masking policy parameter issued by a network device, and the masking policy parameter is a parameter related to the masking operation.
  25. 根据权利要求24所述的装置,其特征在于,所述掩码策略参数包括如下中的至少一种:The device according to claim 24, wherein the mask policy parameters include at least one of the following:
    所述掩码操作对应的矩阵尺寸信息;Matrix size information corresponding to the mask operation;
    所述掩码操作对应的采样信息。The sampling information corresponding to the mask operation.
  26. 一种信道信息反馈模型的训练装置,其特征在于,所述信道信息反馈模型包括:第二编码器和第二解码器,所述装置包括:解码器生成模块、信息接收模块、训练模块和解码器发送模块;A training device for a channel information feedback model, characterized in that the channel information feedback model includes: a second encoder and a second decoder, and the device includes: a decoder generation module, an information receiving module, a training module, and a decoder Transmitter module;
    所述解码器生成模块,用于生成所述第二解码器;The decoder generating module is configured to generate the second decoder;
    所述信息接收模块,用于接收网络设备发送的第二迁移学习信息,所述第二迁移学习信息用于辅助进行迁移学习,所述第二迁移学习信息包括:所述第二编码器和掩码操作对应的矩阵尺寸信息,所述第二编码器是基于所述掩码操作进行训练得到的;The information receiving module is configured to receive second transfer learning information sent by a network device, the second transfer learning information is used to assist transfer learning, and the second transfer learning information includes: the second encoder and mask Matrix size information corresponding to the code operation, the second encoder is obtained by training based on the mask operation;
    所述训练模块,用于基于所述矩阵尺寸信息,对所述第二编码器和所述第二解码器进行联合训练;The training module is configured to jointly train the second encoder and the second decoder based on the matrix size information;
    所述解码器发送模块,用于向所述网络设备发送训练好的所述第二解码器。The decoder sending module is configured to send the trained second decoder to the network device.
  27. 根据权利要求26所述的装置,其特征在于,所述训练模块,用于:The device according to claim 26, wherein the training module is used for:
    按照所述矩阵尺寸信息将用于表示初始信道信息的信道矩阵划分为非重叠的多个矩阵块,所述多个矩阵块组成矩阵块序列;Divide the channel matrix used to represent the initial channel information into a plurality of non-overlapping matrix blocks according to the matrix size information, and the plurality of matrix blocks form a matrix block sequence;
    将所述矩阵块序列作为所述第二编码器的模型输入,经由所述第二编码器对所述矩阵块序列进行压缩处理,得到压缩编码信息;The matrix block sequence is input as a model of the second encoder, and the matrix block sequence is compressed through the second encoder to obtain compressed encoding information;
    将所述压缩编码信息作为所述第二解码器的模型输入,经由所述第二解码器对所述压缩编码信息进行解压缩处理,得到恢复信道信息;inputting the compressed coded information as a model of the second decoder, and decompressing the compressed coded information via the second decoder to obtain restored channel information;
    基于所述恢复信道信息与所述初始信道信息之间的误差,对所述第二编码器和所述第二解码器进行联合训练。The second encoder and the second decoder are jointly trained based on an error between the recovered channel information and the initial channel information.
  28. 根据权利要求26或27所述的装置,其特征在于,Apparatus according to claim 26 or 27, characterized in that,
    所述第二编码器是由一个源侧终端指示给所述网络设备的编码器。The second encoder is an encoder indicated to the network device by a source-side terminal.
  29. 根据权利要求26或27所述的装置,其特征在于,Apparatus according to claim 26 or 27, characterized in that,
    所述第二编码器是由所述网络设备对多个编码器的模型参数进行聚合计算后,得到的全局编码器,所述多个编码器分别来自于多个所述源侧终端。The second encoder is a global encoder obtained after the network device aggregates and calculates model parameters of multiple encoders, and the multiple encoders come from multiple source-side terminals respectively.
  30. 一种信道信息反馈模型的训练装置,其特征在于,所述装置包括:信息发送模块和解码器接收模块;A training device for a channel information feedback model, characterized in that the device includes: an information sending module and a decoder receiving module;
    所述信息发送模块,用于向目标侧终端发送第二迁移学习信息,所述第二迁移学习信息用于辅助进行迁移学习,所述第二迁移学习信息包括:第二编码器和掩码操作对应的矩阵尺寸信息,所述第二编码器是基于所述掩码操作进行训练得到的;The information sending module is configured to send second transfer learning information to the target terminal, the second transfer learning information is used to assist transfer learning, and the second transfer learning information includes: a second encoder and a mask operation Corresponding matrix size information, the second encoder is obtained by training based on the mask operation;
    所述解码器接收模块,用于接收所述目标侧终端发送的第二解码器,所述第二解码器是所述目标侧终端基于所述第二迁移学习信息进行迁移学习后,训练得到的。The decoder receiving module is configured to receive the second decoder sent by the target terminal, the second decoder is obtained by training after the target terminal performs transfer learning based on the second transfer learning information .
  31. 根据权利要求30所述的装置,其特征在于,所述第二编码器是由一个源侧终端指示给网络设备的编码器;The device according to claim 30, wherein the second encoder is an encoder indicated to the network device by a source-side terminal;
    所述装置还包括:信息接收模块;The device also includes: an information receiving module;
    所述信息接收模块,用于接收一个所述源侧终端发送的第一迁移学习信息,所述第一迁移学习信息用于辅助进行迁移学习,所述第一迁移学习信息包括:第一编码器和所述掩码操作对应的矩阵尺寸信息。The information receiving module is configured to receive a piece of first transfer learning information sent by the source-side terminal, the first transfer learning information is used to assist transfer learning, and the first transfer learning information includes: a first encoder Matrix size information corresponding to the mask operation.
  32. 根据权利要求30所述的装置,其特征在于,所述第二编码器是由网络设备对多个编码器的模型参数进行聚合计算后,得到的全局编码器;The device according to claim 30, wherein the second encoder is a global encoder obtained by aggregate calculation of model parameters of multiple encoders by network equipment;
    所述装置还包括:信息接收模块和聚合计算模块;The device also includes: an information receiving module and an aggregation calculation module;
    所述信息接收模块,用于接收多个所述源侧终端分别发送的多个第一迁移学习信息,所述第一迁移学习信息用于辅助进行迁移学习,所述第一迁移学习信息包括:第一编码器和所述掩码操作对应的矩阵尺寸信息;The information receiving module is configured to receive a plurality of first transfer learning information respectively sent by a plurality of the source-side terminals, the first transfer learning information is used to assist transfer learning, and the first transfer learning information includes: Matrix size information corresponding to the first encoder and the mask operation;
    所述聚合计算模块,用于对多个所述训练好的第一编码器的模型参数进行聚合计算,得到所述全局编码器。The aggregation calculation module is configured to perform aggregation calculation on model parameters of multiple trained first encoders to obtain the global encoder.
  33. 根据权利要求32所述的装置,其特征在于,所述装置还包括:参数配置模块;The device according to claim 32, further comprising: a parameter configuration module;
    所述参数配置模块,用于向多个所述源侧终端下发相同的掩码策略参数,所述掩码策略参数是与所述掩码操作相关的参数。The parameter configuration module is configured to deliver the same masking policy parameter to multiple terminals at the source side, where the masking policy parameter is a parameter related to the masking operation.
  34. 根据权利要求33所述的装置,其特征在于,所述掩码策略参数包括如下中的至少一种:The device according to claim 33, wherein the mask policy parameters include at least one of the following:
    所述掩码操作对应的矩阵尺寸信息;Matrix size information corresponding to the mask operation;
    所述掩码操作对应的采样信息。The sampling information corresponding to the mask operation.
  35. 一种终端设备,其特征在于,所述终端设备包括:处理器;其中,A terminal device, characterized in that the terminal device includes: a processor; wherein,
    所述处理器,用于对初始信道信息进行掩码操作,得到掩码信道信息;The processor is configured to perform a masking operation on initial channel information to obtain masked channel information;
    所述处理器,用于将所述掩码信道信息输入信道信息反馈模型,输出恢复信道信息;The processor is configured to input the masked channel information into a channel information feedback model, and output restored channel information;
    所述处理器,用于基于所述恢复信道信息与所述初始信道信息之间的误差,对所述信道信息反馈模型进行训练。The processor is configured to train the channel information feedback model based on an error between the recovered channel information and the initial channel information.
  36. 一种终端设备,其特征在于,所述终端设备包括:处理器和与所述处理器相连的收发器;其中,A terminal device, characterized in that the terminal device includes: a processor and a transceiver connected to the processor; wherein,
    所述处理器,用于生成第二解码器;the processor, configured to generate a second decoder;
    所述收发器,用于接收网络设备发送的第二迁移学习信息,所述第二迁移学习信息用于辅助进行迁移学习,所述第二迁移学习信息包括:所述第二编码器和掩码操作对应的矩阵尺寸信息,所述第二编码器是基于所述掩码操作进行训练得到的;The transceiver is configured to receive second transfer learning information sent by a network device, the second transfer learning information is used to assist transfer learning, and the second transfer learning information includes: the second encoder and a mask Operating the corresponding matrix size information, the second encoder is obtained by training based on the mask operation;
    所述处理器,用于基于所述矩阵尺寸信息,对所述第二编码器和所述第二解码器进行联合训练;The processor is configured to jointly train the second encoder and the second decoder based on the matrix size information;
    所述收发器,用于向所述网络设备发送训练好的所述第二解码器。The transceiver is configured to send the trained second decoder to the network device.
  37. 一种网络设备,其特征在于,所述网络设备包括:收发器;其中,A network device, characterized in that the network device includes: a transceiver; wherein,
    所述收发器,用于向目标侧终端发送第二迁移学习信息,所述第二迁移学习信息用于辅助进行迁移学 习,所述第二迁移学习信息包括:第二编码器和掩码操作对应的矩阵尺寸信息,所述第二编码器是基于所述掩码操作进行训练得到的;The transceiver is configured to send second transfer learning information to the target terminal, the second transfer learning information is used to assist in transfer learning, and the second transfer learning information includes: a second encoder corresponding to a mask operation The matrix size information of the second encoder is obtained by training based on the mask operation;
    所述收发器,用于接收所述目标侧终端发送的第二解码器,所述第二解码器是所述目标侧终端基于所述第二迁移学习信息进行迁移学习后,训练得到的。The transceiver is configured to receive the second decoder sent by the target-side terminal, where the second decoder is trained by the target-side terminal after performing transfer learning based on the second transfer learning information.
  38. 一种计算机可读存储介质,其特征在于,所述可读存储介质中存储有可执行指令,所述可执行指令由处理器加载并执行以实现如权利要求1至17任一所述的信道信息反馈模型的训练方法。A computer-readable storage medium, characterized in that executable instructions are stored in the readable storage medium, and the executable instructions are loaded and executed by a processor to implement the channel according to any one of claims 1 to 17 Training methods for information feedback models.
  39. 一种芯片,其特征在于,所述芯片包括可编程逻辑电路和/或程序指令,当所述芯片运行时,用于实现如权利要求1至17任一所述的信道信息反馈模型的训练方法。A kind of chip, it is characterized in that, described chip comprises programmable logic circuit and/or program instruction, when described chip runs, is used for realizing the training method of channel information feedback model as described in any one of claims 1 to 17 .
  40. 一种计算机程序产品或计算机程序,其特征在于,所述计算机程序产品或计算机程序包括计算机指令,所述计算机指令存储在计算机可读存储介质中,处理器从所述计算机可读存储介质读取并执行所述计算机指令,以实现如权利要求1至17任一所述的信道信息反馈模型的训练方法。A computer program product or computer program, characterized in that the computer program product or computer program includes computer instructions, the computer instructions are stored in a computer-readable storage medium, and the processor reads the computer-readable storage medium from the computer-readable storage medium And execute the computer instructions to realize the training method of the channel information feedback model according to any one of claims 1 to 17.
PCT/CN2021/143874 2021-12-31 2021-12-31 Method and device for training channel information feedback model, apparatus, and storage medium WO2023123429A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202180103608.5A CN118120156A (en) 2021-12-31 2021-12-31 Training method, device, equipment and storage medium of channel information feedback model
PCT/CN2021/143874 WO2023123429A1 (en) 2021-12-31 2021-12-31 Method and device for training channel information feedback model, apparatus, and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2021/143874 WO2023123429A1 (en) 2021-12-31 2021-12-31 Method and device for training channel information feedback model, apparatus, and storage medium

Publications (1)

Publication Number Publication Date
WO2023123429A1 true WO2023123429A1 (en) 2023-07-06

Family

ID=86997128

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/143874 WO2023123429A1 (en) 2021-12-31 2021-12-31 Method and device for training channel information feedback model, apparatus, and storage medium

Country Status (2)

Country Link
CN (1) CN118120156A (en)
WO (1) WO2023123429A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108847876A (en) * 2018-07-26 2018-11-20 东南大学 A kind of extensive MIMO time varying channel states Information Compression feedback and method for reconstructing
CN111630803A (en) * 2018-01-22 2020-09-04 Oppo广东移动通信有限公司 Information determination method and device and computer storage medium
CN111901024A (en) * 2020-07-29 2020-11-06 燕山大学 MIMO channel state information feedback method based on fitting depth learning resistance
WO2021230649A1 (en) * 2020-05-13 2021-11-18 한양대학교 에리카산학협력단 Multi-channel resource allocation method and system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111630803A (en) * 2018-01-22 2020-09-04 Oppo广东移动通信有限公司 Information determination method and device and computer storage medium
CN108847876A (en) * 2018-07-26 2018-11-20 东南大学 A kind of extensive MIMO time varying channel states Information Compression feedback and method for reconstructing
WO2021230649A1 (en) * 2020-05-13 2021-11-18 한양대학교 에리카산학협력단 Multi-channel resource allocation method and system
CN111901024A (en) * 2020-07-29 2020-11-06 燕山大学 MIMO channel state information feedback method based on fitting depth learning resistance

Also Published As

Publication number Publication date
CN118120156A (en) 2024-05-31

Similar Documents

Publication Publication Date Title
WO2021189225A1 (en) Machine learning model training method, electronic device and storage medium
CN114223158A (en) Channel state information processing method, electronic equipment and storage medium
WO2023123429A1 (en) Method and device for training channel information feedback model, apparatus, and storage medium
US20230136416A1 (en) Neural network obtaining method and apparatus
DE102016121567A1 (en) Convolutional decoding of system information
WO2022236785A1 (en) Channel information feedback method, receiving end device, and transmitting end device
JP2017526221A (en) Method and apparatus for data transmission
CN114867058A (en) CSI feedback method and device, storage medium, terminal and network equipment
WO2023004563A1 (en) Method for obtaining reference signal and communication devices
US20240154670A1 (en) Method and apparatus for feedback channel status information based on machine learning in wireless communication system
WO2024026792A1 (en) Communication method and apparatus, device, storage medium, chip, and program product
WO2023060503A1 (en) Information processing method and apparatus, device, medium, chip, product, and program
TWI830543B (en) Information processing methods, devices, terminals and network equipment
WO2023015499A1 (en) Wireless communication method and device
WO2024032701A1 (en) Channel state information processing method and apparatus
WO2024032775A1 (en) Quantization method and apparatus
EP4398527A1 (en) Model processing method, electronic device, network device, and terminal device
WO2022199133A1 (en) Method for transmitting compressed codebook, and method for obtaining channel state information matrix
WO2023028948A1 (en) Model processing method, electronic device, network device, and terminal device
WO2023115254A1 (en) Data processing method and device
WO2022217502A1 (en) Information processing method and apparatus, communication device, and storage medium
WO2023077274A1 (en) Csi feedback method and apparatus, and device and storage medium
WO2022151063A1 (en) Information sending method, information receiving method, apparatus, device, and medium
WO2022141544A1 (en) Method and apparatus for transmitting model data
US20240048207A1 (en) Method and apparatus for transmitting and receiving feedback information based on artificial neural network

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21969774

Country of ref document: EP

Kind code of ref document: A1