WO2023115254A1 - Procédé et dispositif de traitement de données - Google Patents

Procédé et dispositif de traitement de données Download PDF

Info

Publication number
WO2023115254A1
WO2023115254A1 PCT/CN2021/139596 CN2021139596W WO2023115254A1 WO 2023115254 A1 WO2023115254 A1 WO 2023115254A1 CN 2021139596 W CN2021139596 W CN 2021139596W WO 2023115254 A1 WO2023115254 A1 WO 2023115254A1
Authority
WO
WIPO (PCT)
Prior art keywords
training data
neural network
codebook
present application
precoding matrix
Prior art date
Application number
PCT/CN2021/139596
Other languages
English (en)
Chinese (zh)
Inventor
肖寒
Original Assignee
Oppo广东移动通信有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Oppo广东移动通信有限公司 filed Critical Oppo广东移动通信有限公司
Priority to PCT/CN2021/139596 priority Critical patent/WO2023115254A1/fr
Publication of WO2023115254A1 publication Critical patent/WO2023115254A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04BTRANSMISSION
    • H04B7/00Radio transmission systems, i.e. using radiation field
    • H04B7/02Diversity systems; Multi-antenna system, i.e. transmission or reception using multiple antennas
    • H04B7/04Diversity systems; Multi-antenna system, i.e. transmission or reception using multiple antennas using two or more spaced independent antennas
    • H04B7/0413MIMO systems
    • H04B7/0417Feedback systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04BTRANSMISSION
    • H04B7/00Radio transmission systems, i.e. using radiation field
    • H04B7/02Diversity systems; Multi-antenna system, i.e. transmission or reception using multiple antennas
    • H04B7/04Diversity systems; Multi-antenna system, i.e. transmission or reception using multiple antennas using two or more spaced independent antennas
    • H04B7/0413MIMO systems
    • H04B7/0456Selection of precoding matrices or codebooks, e.g. using matrices antenna weighting

Definitions

  • the present application relates to the field of communication technology, and more specifically, to a method and device for processing data.
  • a scheme based on a codebook can be used to realize channel information feedback. Due to the elegant design of the codebook, the channel information feedback scheme based on the codebook can adapt to the channel information feedback in different scenarios, that is, the scheme has high generalization.
  • the channel information needs to be mapped to the precoding matrix in the codebook, and the feedback of the channel information is realized by feeding back the precoding matrix. Since the precoding matrix is discrete, the mapping process is quantized and lossy, which leads to low accuracy of the codebook-based precoding scheme.
  • the present application provides a data processing method and device to solve the problem of difficulty in obtaining training data for a neural network-based channel feedback model.
  • a method for processing data includes: using a codebook to generate a plurality of training data, the plurality of training data is used to train a neural network model, and the neural network model is used to perform channel information feedback.
  • a device for processing data includes: a first generation unit, configured to generate a plurality of training data using a codebook, the plurality of training data are used to train a neural network model, and the neural network The network model is used for channel information feedback.
  • a device for processing data including a processor, a memory, and a communication interface, the memory is used to store one or more computer programs, and the processor is used to call the computer programs in the memory to make the The terminal device executes the method described in the first aspect.
  • the embodiment of the present application provides a computer-readable storage medium, the computer-readable storage medium stores a computer program, and the computer program enables the terminal device to perform some or all of the steps in the method of the first aspect above .
  • an embodiment of the present application provides a computer program product, wherein the computer program product includes a non-transitory computer-readable storage medium storing a computer program, and the computer program is operable to enable the terminal to execute the above-mentioned first Some or all of the steps in the method of one aspect.
  • the computer program product can be a software installation package.
  • an embodiment of the present application provides a chip, the chip includes a memory and a processor, and the processor can call and run a computer program from the memory to implement some or all of the steps described in the method of the first aspect above .
  • a computer program product including a program, the program causes a computer to execute the method described in the first aspect.
  • a computer program causes a computer to execute the method described in the first aspect.
  • the present application uses a codebook to generate a plurality of training data, and uses the training data to train a neural network-based channel feedback model. Therefore, the present application can reduce the requirement and overhead of the neural network for acquiring actual scene channel data, and enhance the generalization of the neural network by using the generalization capability of the codebook.
  • Fig. 1 is a wireless communication system applied in the embodiment of the present application.
  • FIG. 2 is a structural diagram of a neural network applicable to an embodiment of the present application.
  • FIG. 3 is a structural diagram of a convolutional neural network applicable to an embodiment of the present application.
  • Fig. 4 is an example diagram of an image compression process based on an autoencoder.
  • Fig. 5 is an example diagram of an image generation process based on a variational autoencoder.
  • Fig. 6 is a schematic diagram of a neural network-based channel feedback process.
  • Fig. 7 is a schematic flowchart of a method for processing data provided by an embodiment of the present application.
  • FIG. 8 is a schematic diagram of a method for grouping precoding matrices of a codebook provided by an embodiment of the present application.
  • FIG. 9 is a channel information feedback model based on a variational autoencoder provided in an embodiment of the present application.
  • FIG. 10 is a schematic structural diagram of a variational autoencoder provided by an embodiment of the present application.
  • Fig. 11 is a schematic structural diagram of an apparatus for data processing provided by an embodiment of the present application.
  • Fig. 12 is a schematic structural diagram of a device provided by an embodiment of the present application.
  • FIG. 1 is a wireless communication system 100 applied in an embodiment of the present application.
  • the wireless communication system 100 may include a network device 110 and a terminal device 120 .
  • the network device 110 may be a device that communicates with the terminal device 120 .
  • the network device 110 can provide communication coverage for a specific geographical area, and can communicate with the terminal device 120 located in the coverage area.
  • Figure 1 exemplarily shows one network device and two terminals.
  • the wireless communication system 100 may include multiple network devices and each network device may include other numbers of terminal devices within the coverage area. The embodiment does not limit this.
  • the wireless communication system 100 may further include other network entities such as a network controller and a mobility management entity, which is not limited in this embodiment of the present application.
  • network entities such as a network controller and a mobility management entity, which is not limited in this embodiment of the present application.
  • the technical solutions of the embodiments of the present application can be applied to various communication systems, for example: the fifth generation (5th generation, 5G) system or new radio (new radio, NR), long term evolution (long term evolution, LTE) system , LTE frequency division duplex (frequency division duplex, FDD) system, LTE time division duplex (time division duplex, TDD), etc.
  • the technical solutions provided in this application can also be applied to future communication systems, such as the sixth generation mobile communication system, and satellite communication systems, and so on.
  • the terminal equipment in the embodiment of the present application may also be called user equipment (user equipment, UE), access terminal, subscriber unit, subscriber station, mobile station, mobile station (mobile station, MS), mobile terminal (mobile terminal, MT) ), remote station, remote terminal, mobile device, user terminal, terminal, wireless communication device, user agent, or user device.
  • the terminal device in the embodiment of the present application may be a device that provides voice and/or data connectivity to users, and can be used to connect people, objects and machines, such as handheld devices with wireless connection functions, vehicle-mounted devices, and the like.
  • the terminal device in the embodiment of the present application can be mobile phone (mobile phone), tablet computer (Pad), notebook computer, palmtop computer, mobile internet device (mobile internet device, MID), wearable device, virtual reality (virtual reality, VR) equipment, augmented reality (augmented reality, AR) equipment, wireless terminals in industrial control, wireless terminals in self driving, wireless terminals in remote medical surgery, smart Wireless terminals in smart grid, wireless terminals in transportation safety, wireless terminals in smart city, wireless terminals in smart home, etc.
  • UE can be used to act as a base station.
  • a UE may act as a scheduling entity that provides sidelink signals between UEs in V2X or D2D, etc.
  • a cell phone and an automobile communicate with each other using sidelink signals. Communication between cellular phones and smart home devices without relaying communication signals through base stations.
  • the network device in this embodiment of the present application may be a device for communicating with a terminal device, and the network device may also be called an access network device or a wireless access network device, for example, the network device may be a base station.
  • the network device in this embodiment of the present application may refer to a radio access network (radio access network, RAN) node (or device) that connects a terminal device to a wireless network.
  • radio access network radio access network, RAN node (or device) that connects a terminal device to a wireless network.
  • the base station can broadly cover various names in the following, or replace with the following names, such as: Node B (NodeB), evolved base station (evolved NodeB, eNB), next generation base station (next generation NodeB, gNB), relay station, Access point, transmission point (transmitting and receiving point, TRP), transmission point (transmitting point, TP), primary station MeNB, secondary station SeNB, multi-standard radio (MSR) node, home base station, network controller, access node , wireless node, access point (access point, AP), transmission node, transceiver node, base band unit (base band unit, BBU), remote radio unit (Remote Radio Unit, RRU), active antenna unit (active antenna unit) , AAU), radio head (remote radio head, RRH), central unit (central unit, CU), distributed unit (distributed unit, DU), positioning nodes, etc.
  • NodeB Node B
  • eNB evolved base station
  • next generation NodeB next generation NodeB
  • a base station may be a macro base station, a micro base station, a relay node, a donor node, or the like, or a combination thereof.
  • a base station may also refer to a communication module, modem or chip used to be set in the aforementioned equipment or device.
  • the base station can also be a mobile switching center, a device that undertakes the function of a base station in D2D, vehicle-to-everything (V2X), machine-to-machine (M2M) communication, and a device in a 6G network.
  • V2X vehicle-to-everything
  • M2M machine-to-machine
  • Base stations can support networks of the same or different access technologies. The embodiment of the present application does not limit the specific technology and specific device form adopted by the network device.
  • Base stations can be fixed or mobile.
  • a helicopter or drone can be configured to act as a mobile base station, and one or more cells can move according to the location of the mobile base station.
  • a helicopter or drone may be configured to serve as a device in communication with another base station.
  • the network device in this embodiment of the present application may refer to a CU or a DU, or, the network device includes a CU and a DU.
  • a gNB may also include an AAU.
  • Network equipment and terminal equipment can be deployed on land, including indoors or outdoors, hand-held or vehicle-mounted; they can also be deployed on water; they can also be deployed on aircraft, balloons and satellites in the air.
  • the scenarios where the network device and the terminal device are located are not limited.
  • a codebook-based scheme can be used to realize channel feature extraction and feedback. That is, after the receiver performs channel estimation, according to the result of channel estimation, according to a certain optimization criterion, select the precoding matrix that best matches the current channel from the pre-set precoding codebook, and pass the feedback link of the air interface to the Precoding matrix index (precoding matrix index, PMI) information is fed back to the transmitter for the transmitter to implement precoding.
  • the receiver may also feed back the measured channel quality indication (CQI) to the transmitter for the transmitter to implement adaptive modulation and coding.
  • Channel feedback may also be called channel state information (channel state information-reference signal, CSI) feedback.
  • Codebook-Based Channel Feedback Scheme Due to the sophisticated design of the codebook, multiple precoding matrices in the codebook can represent channel information in different scenarios. Therefore, the codebook-based channel feedback scheme can adapt to channel information feedback tasks in different scenarios.
  • mapping channel information to the precoding matrix in the codebook is a process of discretely quantizing continuous channel information, which makes the mapping process lossy in quantization, resulting in a decrease in the accuracy of the feedback channel information, thereby reducing the accuracy of the precoding matrix. Encoding performance.
  • AI artificial intelligence
  • Neural networks are commonly used architectures in AI. Common neural networks include convolutional neural network (CNN), recurrent neural network (RNN), deep neural network (DNN), etc.
  • CNN convolutional neural network
  • RNN recurrent neural network
  • DNN deep neural network
  • the neural network applicable to the embodiment of the present application is introduced below with reference to FIG. 2 .
  • the neural network shown in FIG. 2 can be divided into three types according to the position of different layers: input layer 210 , hidden layer 220 and output layer 230 .
  • the first layer is the input layer 210
  • the last layer is the output layer 230
  • the middle layer between the first layer and the last layer is the hidden layer 220 .
  • the input layer 210 is used to input data. Taking a communication system as an example, the input data may be, for example, a received signal received by a receiver.
  • the hidden layer 220 is used to process the input data, for example, to decompress the received signal.
  • the output layer 230 is used for outputting processed output data, for example, outputting a decompressed signal.
  • the neural network includes multiple layers, each layer includes multiple neurons, and the neurons between layers can be fully connected or partially connected. For connected neurons, the output of neurons in the previous layer can be used as the input of neurons in the next layer.
  • neural network deep learning algorithms have been proposed in recent years.
  • the neural network deep learning algorithm introduces more hidden layers in the neural network.
  • This neural network model is widely used in pattern recognition, signal processing, optimization combination, anomaly detection and so on.
  • CNN is a deep neural network with a convolutional structure, and its structure can be shown in Figure 3.
  • the neural network shown in FIG. 3 may include an input layer 310 , a convolutional layer 320 , a pooling layer 330 , a fully connected layer 340 , and an output layer 350 .
  • Each convolutional layer 320 can include many convolution kernels.
  • the convolution kernel is also called an operator. Its function can be regarded as a filter for extracting specific information from the input signal.
  • the convolution kernel can be a weight in essence. matrix, this weight matrix is usually predefined.
  • weight values in these weight matrices need to be obtained through a lot of training in practical applications, and each weight matrix formed by the weight values obtained through training can extract information from the input signal, thereby helping CNN to make correct predictions.
  • the initial convolutional layer often extracts more general features, which can also be called low-level features; as the depth of CNN deepens, the later convolution The features extracted by the layers are getting more and more complex.
  • Pooling layer 330 because it is often necessary to reduce the number of training parameters, it is often necessary to periodically introduce a pooling layer after the convolutional layer, for example, it can be a layer of convolutional layer followed by a layer of pooling layer as shown in Figure 3 , can also be a multi-layer convolutional layer followed by one or more pooling layers. In signal processing, the sole purpose of pooling layers is to reduce the spatial size of the extracted information.
  • the introduction of the convolutional layer 320 and the pooling layer 330 effectively controls the sharp increase of network parameters, limits the number of parameters and taps the characteristics of the local structure, improving the robustness of the algorithm.
  • the fully connected layer 340 after being processed by the convolutional layer 320 and the pooling layer 330, CNN is not enough to output the required output information. Because as mentioned above, the convolutional layer 320 and the pooling layer 330 only extract features and reduce the parameters brought by the input data. However, in order to generate the final output information (eg, the bitstream of the original information transmitted by the transmitter), the CNN also needs to utilize the fully connected layer 340 .
  • the fully connected layer 340 may include a plurality of hidden layers, and the parameters contained in the multi-layer hidden layers may be pre-trained according to relevant training data of a specific task type, for example, the task type may include receiving For another example, the task type may also include performing channel estimation based on the pilot signal received by the receiver.
  • the output layer 350 for outputting results.
  • the output layer 350 is provided with a loss function (for example, a loss function similar to classification cross entropy), which is used to calculate the prediction error, or to evaluate the result (also called predicted value) output by the CNN model and the ideal result (also called The degree of difference between the true value).
  • a loss function for example, a loss function similar to classification cross entropy
  • the CNN model In order to minimize the loss function, the CNN model needs to be trained.
  • the CNN model may be trained using a backpropagation algorithm (BP).
  • BP backpropagation algorithm
  • the training process of BP consists of forward propagation process and back propagation process.
  • forward propagation the propagation from 310 to 350 in Fig. 3 is forward propagation
  • the input data is input into the above layers of the CNN model, processed layer by layer and transmitted to the output layer.
  • the above loss function is minimized as the optimization goal, and transferred to backpropagation (as shown in Figure 3, the propagation from 350 to 310 is backpropagation), and the calculation is obtained layer by layer
  • the partial derivative of the optimization target to the weight of each neuron constitutes the gradient of the optimization target to the weight vector, which is used as the basis for modifying the model weight.
  • the training process of CNN is completed in the weight modification process. When the above error reaches the expected value, the training process of CNN ends.
  • the CNN shown in Figure 3 is only an example of a convolutional neural network.
  • the convolutional neural network can also exist in the form of other network models, and this embodiment of the present application does not make any reference to this. limited.
  • Autoencoders are a class of artificial neural networks used in semi-supervised and unsupervised learning.
  • An autoencoder is a neural network that takes an input signal as the training target.
  • An autoencoder can include an encoder (encoder) and a decoder (decoder).
  • the input of the encoder may be an image to be compressed.
  • a code stream (code) is output.
  • the number of bits occupied by the code stream output by the encoder is generally smaller than the number of bits occupied by the image to be compressed. For example, the number of bits occupied by the code stream output by the encoder shown in FIG. 4 may be less than 784 bits. From this, it can be seen that the encoder can achieve a compressed representation of the entity input to the encoder.
  • the input of the decoder can be code stream.
  • the code stream may be a code stream output by an encoder.
  • the output of the decoder is the decompressed image. It can be seen from Fig. 4 that the decompressed image is consistent with the image to be compressed input to the encoder. Therefore, the decoder can realize the reconstruction of the original entity.
  • the data to be compressed (such as the picture to be compressed in Figure 4) can be used as the input of the self-encoder (ie, the input of the encoder) and the label (ie, the output of the decoder), and the encoder and the decoder for end-to-end joint training.
  • Variational autoencoders introduce probability distributions in the encoder.
  • the probability distribution can make the entity input to the encoder be mapped to the code stream conforming to the probability distribution.
  • different values collected in the probability distribution can be combined, so that the input of the decoder of the variational autoencoder can be multiple different code streams.
  • the decoder can decode different code streams output by the encoder, so as to output multiple entities similar to the input entity.
  • FIG. 5 is an example diagram of an image generation process based on a variational autoencoder.
  • the input of the encoder can be the original image, and the output of the encoder can be the mean and variance.
  • the code stream can be calculated by the mean and variance of the encoder output and the collected values collected from the probability distribution.
  • the code stream can be input into the decoder, and the decoder decodes the code stream and outputs an image similar to the original image.
  • the probability distribution may be a normal distribution.
  • Channel feedback can be realized based on AI, such as neural network-based channel feedback.
  • the network device side can restore the channel information fed back by the terminal device side as much as possible through the neural network.
  • This neural network-based channel feedback can restore channel information, and also provides the possibility of reducing channel feedback overhead on the terminal device side.
  • a deep learning autoencoder can be used to implement channel feedback.
  • the input of the AI-based channel feedback model can be channel information, that is, the channel information can be regarded as the compressed image input to the self-encoder.
  • the AI-based channel feedback model can perform compressed feedback on channel information.
  • the AI-based channel feedback model can reconstruct the compressed channel information, thereby retaining the channel information to a greater extent.
  • FIG. 6 is a schematic diagram of an AI-based channel feedback process.
  • the channel feedback model shown in Fig. 6 includes an encoder and a decoder.
  • the encoder and decoder are respectively deployed at the receiving end (receive, Rx) and the sending end (transmit, Tx).
  • the receiving end can obtain the channel information matrix through channel estimation.
  • the channel information matrix can be compressed and encoded by the neural network of the encoder to form a compressed bit stream (codeword).
  • codeword compressed bit stream
  • the compressed bit stream can be fed back to the receiving end through an air interface feedback link.
  • the sending end can decode or restore the channel information according to the feedback bit stream through the decoder, so as to obtain complete feedback channel information.
  • the AI-based channel feedback model may have the structure shown in FIG. 6 .
  • the encoder may include several fully connected layers, and the decoder may include a residual network.
  • FIG. 6 is only an example, and the present application does not limit the structure of the network model inside the encoder and decoder, and the structure of the network model can be flexibly designed.
  • the channel feedback based on the neural network can directly compress the channel information. Therefore, the accuracy of the channel information fed back by the neural network-based channel feedback scheme is relatively high.
  • the channel feedback scheme based on neural network has better performance.
  • the performance of neural network-based channel feedback schemes is poor when the training data does not match the test data.
  • neural network-based channel feedback schemes perform poorly when the scenarios of training data and test data are inconsistent. Therefore, channel feedback schemes based on neural networks have the problem of low generalization.
  • Related technologies use multi-scenario training data to train a neural network model to improve generalization. In the actual operation process, it is more difficult to obtain a large-scale and high-diversity training data set.
  • FIG. 7 is a schematic flowchart of a method for processing data provided by an embodiment of the present application.
  • the method shown in FIG. 7 may include step S710.
  • Step S710 using the codebook to generate a plurality of training data.
  • Multiple training data can be used to train the neural network model.
  • the neural network model can be used for channel information feedback.
  • Training data may include channel information.
  • the channel information may be a tensor composed of channel feature vectors of multiple subbands and multiple layers.
  • the channel information may be a tensor with a dimension of N ⁇ M ⁇ S ⁇ I.
  • N may be the number of transmitting antennas, and the value of N may be greater than or equal to 2.
  • M may be the number of subbands, and the value of M may be greater than or equal to 1.
  • S may be the number of transmission layers, and the value of S may be greater than or equal to 1.
  • the codebook may be the codebook used in codebook-based channel feedback in the related art.
  • a codebook may include multiple precoding matrices.
  • the present application does not limit the form of the precoding matrix.
  • the precoding matrix may be in the form of a vector, that is, the codebook may include multiple precoding vectors.
  • the codebook may be implemented based on a discrete Fourier transform (discrete fourier transform, DFT) vector, that is, the precoding matrix may include a DFT vector.
  • the length of the DFT vector may be N
  • the number of DFT vectors in the codebook may be N 1 N 2 .
  • N 1 and N 2 may be the number of antenna ports.
  • N 1 may be the number of antenna ports in the first dimension
  • N 2 may be the number of antenna ports in the second dimension.
  • N may be the total number of antenna ports.
  • the precoding matrix in the codebook can represent channel information in different scenarios, that is, the codebook naturally has generalization ability.
  • the multiple pieces of training data generated by using the codebook may also include channel information of multiple scenarios. Therefore, using the training data generated by the codebook to train the neural network model can enhance the generalization of the neural network model.
  • the process of using codebooks to generate training data is much easier than obtaining diverse training data in actual scenarios. Therefore, using the technical solution of the present application, a neural network model with high generalization can be trained more simply, so as to realize high-precision and high-performance channel information feedback.
  • the method shown in FIG. 7 may further include step S720.
  • Step S720 perform oversampling processing on multiple precoding matrices in the codebook.
  • Oversampling can be to insert more precoding matrices between adjacent precoding matrices. It can be understood that each precoding matrix may correspond to a spatial angle orientation, and there is a certain angular interval between adjacent precoding matrices. Through oversampling processing, the angle interval between adjacent precoding matrices can be made smaller, so as to achieve the purpose of fine quantization of angle.
  • the oversampling factors of the first dimension and the second dimension of the two-dimensional array antenna may be O 1 and O 2 respectively.
  • the precoding matrix as a two-dimensional DFT vector as an example, the oversampled two-dimensional DFT vector a m,n can be expressed as:
  • x m and u n are oversampled one-dimensional DFT vectors respectively, and x m and u n can be expressed as:
  • u n [1,...,exp(j2 ⁇ (N 2 ⁇ 1)m)/N 2 O 2 ] T .
  • a precoding matrix can represent a beam.
  • the actual channel information can consist of information from one or more beams. Therefore, one or more precoding matrices can be selected in the codebook to form a training data to simulate real channel information.
  • the present application does not limit the method of selecting the precoding matrix in the codebook to generate the training data.
  • the precoding matrix may be selected in a grouping manner to generate training data.
  • Multiple precoding matrices can belong to multiple groups.
  • the plurality of groups may include a first group, and the plurality of training data may include the first training data.
  • the first training data can be generated using the precoding matrix in the first packet. For example, all the precoding matrices contained in the first training data can be selected from the first group. It should be noted that, the present application does not limit the number of precoding matrices in the first group, for example, it may be one or more.
  • the grouping can be determined according to the beam situation corresponding to the precoding matrix. For example, real channel information often includes information about multiple similar beams. Therefore, precoding matrices corresponding to adjacent or similar beams may be grouped into one group. It can be seen that selecting training data from a group can make the training data closer to real channel information.
  • FIG. 8 is a schematic diagram of a grouping method provided by an embodiment of the present application.
  • a grid point (circular point in FIG. 8) on the two-dimensional grid represents a precoding matrix.
  • a grid block (rectangular box in FIG. 8 ) on a two-dimensional grid can represent a group.
  • the first precoding matrix may be any grid point on the two-dimensional grid.
  • grid point 811 may represent the first precoding matrix.
  • the first group can be any rectangular box on the two-dimensional grid.
  • block 821 may represent the first grouping.
  • the first beam of the grid block ie the first beam (the shaded circular point in Fig. 8)
  • the starting position of the packet can be adjusted.
  • the size of the grid block the number of precoding matrices in the group can be adjusted.
  • the precoding matrices in one group may be continuous or spaced, which is not limited in this application.
  • grouping can be controlled by parameters L 1 , L 2 , s 1 , s 2 , p 1 , p 2 .
  • L 1 represents the size of the grid in the first dimension
  • L 2 represents the size of the grid in the second dimension
  • L 1 ⁇ L 2 represents the size of the grid.
  • s 1 represents the distance between the first beams of adjacent grid blocks in the first dimension.
  • s 2 represents the distance between the first beams of adjacent grid blocks in the second dimension.
  • p 1 represents the distance between adjacent beams in a grid block in the first dimension.
  • p 2 represents the distance between adjacent beams in a grid block in the second dimension.
  • the training data can be selected from the precoding matrix in the packet.
  • MS precoding matrices can be selected in the first group, and further processed (for example, perform dimension conversion) to generate a tensor with a dimension of N ⁇ M ⁇ S ⁇ I, and use this tensor as the first training data.
  • the present application does not limit the selection manner of the precoding matrix.
  • the first training data may randomly select a precoding matrix in the first group.
  • each group can generate at least one training data, and multiple groups can generate multiple training data.
  • the first group can generate the first training data and the second training data
  • the second group can generate the third training data.
  • a data set composed of multiple training data can cover all precoding matrices in the codebook. It can be understood that all precoding matrices in the codebook can cover various scenarios. Therefore, the neural network model trained with this dataset can integrate the high generalization performance of the codebook.
  • precoding matrices in multiple groups may overlap, for example, multiple groups may include a first group and a second group, multiple precoding matrices may include a first precoding matrix, and the first precoding matrix may be both Belonging to the first group also belongs to the second group.
  • the second grouping may be a grid block 822 .
  • the first precoding matrix 811 may belong to both the first group 821 and the second group 822 . It can be seen from the example shown in FIG. 8 that the first group 821 and the second group 822 may overlap in the first dimension. It can be understood that the first group and the second group may overlap in the second dimension. Alternatively, the first grouping and the second grouping overlap in both the first dimension and the second dimension.
  • precoding matrices of multiple groups can make the codebook be divided into more groups, thereby obtaining more training data.
  • the application does not limit the type of neural network model.
  • the neural network model may include a variational autoencoder.
  • the variational autoencoder can map the input training data to a probability distribution, so that multiple code streams conforming to the probability distribution can be obtained from one training data. Multiple codestreams provide more training data for the decoder, which can improve the generalization of the decoder and thus the generalization of the variational autoencoder.
  • the probability distribution may not be introduced, and the variational autoencoder can become an autoencoder, so that the decoder can accurately restore the channel information input to the encoder.
  • FIG. 9 is a channel information feedback model based on a variational autoencoder provided in an embodiment of the present application.
  • Both the input data of the encoder 910 and the output data of the decoder 920 may be channel information.
  • the channel information may be training data generated by a codebook, or actually obtained channel information.
  • the input data of the encoder 910 or the output data of the decoder 920 may include training data generated based on a codebook, or may include training data collected in an actual scene.
  • the input data of the encoder 910 or the output data of the decoder 920 may be channel information obtained after actual channel estimation.
  • the channel information may be reshaped to change the dimension of the channel information.
  • training data or channel information obtained from channel estimation may be reshaped.
  • reshaping can make the input data of the encoder 910 adapt to the model structure of the encoder 910, thereby simplifying the operation of the neural network model.
  • reshaping can make the output data suitable for subsequent processing, thereby simplifying subsequent operations.
  • the shaped channel information can be a three-dimensional tensor, and the dimension format of the three-dimensional tensor can be NM ⁇ S ⁇ I, NS ⁇ M ⁇ I, NI ⁇ M ⁇ S, N ⁇ MS ⁇ I, N ⁇ MI ⁇ S Or N ⁇ M ⁇ SI.
  • the shaped channel information may be a two-dimensional matrix, and the dimension format of the two-dimensional matrix may be NMS ⁇ I, NMI ⁇ S, NSI ⁇ M or MSI ⁇ N.
  • the shaped channel information may be a vector, and the length of the vector may be NMSI.
  • m can be related to the mean
  • v can be related to the variance.
  • m and v can be of length p.
  • the input to the decoder 920 may be a vector of length p.
  • the input to the decoder 920 can be different.
  • the input of the decoder 920 may be a quantized and dequantized vector of the vector s ⁇ exp(v)+m.
  • the vectors v and m are obtained from the output of the encoder 910, and the vector s can be collected from a standard normal distribution.
  • the vector s can also be of length p.
  • the input of the decoder 920 may be a quantized and dequantized vector m, and m may be obtained from the output of the encoder 910 .
  • This application does not limit the specific structure of the encoder or decoder in variational self-encoding.
  • it can be constructed by using one or more network structures of fully connected network, convolutional neural network, residual network, and self-attention mechanism network.
  • FIG. 10 is a schematic structural diagram of a variational autoencoder provided by an embodiment of the present application.
  • the variational autoencoder shown in FIG. 10 includes an encoder 1010 and a decoder 1020 .
  • the encoder 1010 may include a fully connected layer 1012 , a fully connected layer 1013 , a fully connected layer 1014 , a fully connected layer 1015 and a fully connected layer 1016 .
  • the decoder 1020 may include a fully connected layer 1022 , a fully connected layer 1023 and a fully connected layer 1024 .
  • the quantization scheme of the variational autoencoder shown in FIG. 10 may be 3-bit uniform quantization, and the feedback resource may be 48 bits.
  • the fully connected layer 1012 can output a vector with a dimension of 1024.
  • the fully connected layer 1013 can output a vector with a dimension of 256.
  • the fully connected layer 1014 can output a vector with a dimension of 128.
  • Both the input of the fully connected layer 1015 and the fully connected layer 1016 may be the output of the fully connected layer 1014 .
  • the fully connected layer 1015 can output a vector m with a dimension of 16.
  • the fully connected layer 1016 can output a vector v with dimension 16.
  • the input of decoder 1020 may be a vector with dimension 16.
  • the fully connected layer 1022 can output a vector with a dimension of 2048.
  • the fully connected layer 1023 can output a vector with a dimension of 1024.
  • the fully connected layer 1024 can output a vector with a dimension of 768.
  • the multiple training data generated by using the codebook can be used to train the neural network-based channel information feedback model.
  • This application does not limit the training method of the model.
  • W can be used as the input of the encoder, and the output of the decoder can be W ⁇ .
  • W may include training data generated using a codebook, or may include training data collected in actual scenarios.
  • the loss function can be:
  • training data is also referred to as training samples or samples. This application is not limited to this.
  • FIG. 11 is a schematic structural diagram of an apparatus 1100 for processing data provided by an embodiment of the present application.
  • the apparatus 1100 may include a generating unit 1110 .
  • the first generating unit 1110 may be configured to generate a plurality of training data by using a codebook, the plurality of training data are used to train a neural network model, and the neural network model is used for channel information feedback.
  • the neural network model includes a variational autoencoder.
  • the codebook includes multiple precoding matrices.
  • the apparatus 1100 may further include an oversampling unit 1120 .
  • the oversampling unit 1120 may be configured to perform oversampling processing on the multiple precoding matrices.
  • the multiple precoding matrices belong to multiple groups, the multiple groups include a first group, the multiple training data include first training data, and the first generating unit 1110 may include: a second A generating unit, configured to generate the first training data by using the precoding matrix in the first group.
  • the multiple precoding matrices include a first precoding matrix
  • the multiple groups include a second group
  • the precoding matrix includes a discrete Fourier transform (DFT) vector.
  • DFT discrete Fourier transform
  • the device further includes: a reshaping unit, configured to reshape the training data, so as to change the dimensions of the training data.
  • a reshaping unit configured to reshape the training data, so as to change the dimensions of the training data.
  • FIG. 12 is a schematic structural diagram of an apparatus for processing data according to an embodiment of the present application.
  • the dashed line in Figure 12 indicates that the unit or module is optional.
  • the apparatus 1200 may be used to implement the methods described in the foregoing method embodiments.
  • Apparatus 1200 may be a chip, a terminal device or a network device.
  • Apparatus 1200 may include one or more processors 1210 .
  • the processor 1210 can support the device 1200 to implement the methods described in the foregoing method embodiments.
  • the processor 1210 may be a general purpose processor or a special purpose processor.
  • the processor may be a central processing unit (central processing unit, CPU).
  • the processor can also be other general-purpose processors, digital signal processors (digital signal processors, DSPs), application specific integrated circuits (application specific integrated circuits, ASICs), off-the-shelf programmable gate arrays (field programmable gate arrays, FPGAs) Or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc.
  • a general-purpose processor may be a microprocessor, or the processor may be any conventional processor, or the like.
  • Apparatus 1200 may also include one or more memories 1220 .
  • a program is stored in the memory 1220, and the program can be executed by the processor 1210, so that the processor 1210 executes the methods described in the foregoing method embodiments.
  • the memory 1220 may be independent from the processor 1210 or may be integrated in the processor 1210 .
  • the apparatus 1200 may also include a transceiver 1230 .
  • the processor 1210 can communicate with other devices or chips through the transceiver 1230 .
  • the processor 1210 may send and receive data with other devices or chips through the transceiver 1230 .
  • the embodiment of the present application also provides a computer-readable storage medium for storing programs.
  • the computer-readable storage medium can be applied to the terminal or the network device provided in the embodiments of the present application, and the program enables the computer to execute the methods performed by the terminal or the network device in the various embodiments of the present application.
  • the embodiment of the present application also provides a computer program product.
  • the computer program product includes programs.
  • the computer program product can be applied to the terminal or the network device provided in the embodiments of the present application, and the program enables the computer to execute the methods performed by the terminal or the network device in the various embodiments of the present application.
  • the embodiment of the present application also provides a computer program.
  • the computer program can be applied to the terminal or the network device provided in the embodiments of the present application, and the computer program enables the computer to execute the methods performed by the terminal or the network device in the various embodiments of the present application.
  • the "indication" mentioned may be a direct indication, may also be an indirect indication, and may also mean that there is an association relationship.
  • a indicates B which can mean that A directly indicates B, for example, B can be obtained through A; it can also indicate that A indirectly indicates B, for example, A indicates C, and B can be obtained through C; it can also indicate that there is an association between A and B relation.
  • B corresponding to A means that B is associated with A, and B can be determined according to A.
  • determining B according to A does not mean determining B only according to A, and B may also be determined according to A and/or other information.
  • the term "corresponding" may indicate that there is a direct or indirect correspondence between the two, or that there is an association between the two, or that it indicates and is instructed, configures and is configured, etc. relation.
  • predefined or “preconfigured” can be realized by pre-saving corresponding codes, tables or other methods that can be used to indicate relevant information in devices (for example, including terminal devices and network devices).
  • the application does not limit its specific implementation.
  • pre-defined may refer to defined in the protocol.
  • the "protocol” may refer to a standard protocol in the communication field, for example, may include the LTE protocol, the NR protocol, and related protocols applied to future communication systems, which is not limited in the present application.
  • sequence numbers of the above-mentioned processes do not mean the order of execution, and the execution order of each process should be determined by its functions and internal logic, rather than the implementation process of the embodiments of the present application. constitute any limitation.
  • the disclosed systems, devices and methods may be implemented in other ways.
  • the device embodiments described above are only illustrative.
  • the division of the units is only a logical function division. In actual implementation, there may be other division methods.
  • multiple units or components can be combined or May be integrated into another system, or some features may be ignored, or not implemented.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be through some interfaces, and the indirect coupling or communication connection of devices or units may be in electrical, mechanical or other forms.
  • the units described as separate components may or may not be physically separated, and the components shown as units may or may not be physical units, that is, they may be located in one place, or may be distributed to multiple network units. Part or all of the units can be selected according to actual needs to achieve the purpose of the solution of this embodiment.
  • each functional unit in each embodiment of the present application may be integrated into one processing unit, each unit may exist separately physically, or two or more units may be integrated into one unit.
  • all or part of them may be implemented by software, hardware, firmware or any combination thereof.
  • software When implemented using software, it may be implemented in whole or in part in the form of a computer program product.
  • the computer program product includes one or more computer instructions. When the computer program instructions are loaded and executed on the computer, the processes or functions according to the embodiments of the present application will be generated in whole or in part.
  • the computer can be a general purpose computer, a special purpose computer, a computer network, or other programmable devices.
  • the computer instructions may be stored in or transmitted from one computer-readable storage medium to another computer-readable storage medium, for example, the computer instructions may be transmitted from a website, computer, server or data center Transmission to another website site, computer, server or data center by wired (such as coaxial cable, optical fiber, digital subscriber line (DSL)) or wireless (such as infrared, wireless, microwave, etc.).
  • the computer-readable storage medium may be any available medium that can be read by a computer, or a data storage device such as a server or a data center integrated with one or more available media.
  • the available medium may be a magnetic medium (for example, a floppy disk, a hard disk, a magnetic tape), an optical medium (for example, a digital versatile disc (digital video disc, DVD)) or a semiconductor medium (for example, a solid state disk (solid state disk, SSD) )wait.
  • a magnetic medium for example, a floppy disk, a hard disk, a magnetic tape
  • an optical medium for example, a digital versatile disc (digital video disc, DVD)
  • a semiconductor medium for example, a solid state disk (solid state disk, SSD)

Abstract

La présente demande concerne un procédé et un dispositif de traitement de données. Le procédé de traitement de données consiste à : générer une pluralité d'éléments de données d'entraînement à l'aide d'un livre de codes, la pluralité d'éléments de données d'entraînement étant utilisée pour entraîner un modèle de réseau neuronal, et le modèle de réseau neuronal étant utilisé pour effectuer une rétroaction d'informations de canal. Selon la présente demande, la pluralité d'éléments de données d'entraînement sont générés à l'aide du livre de codes, et un modèle de rétroaction de canal basé sur un réseau neuronal est entraîné à l'aide des données d'entraînement. Par conséquent, selon la présente demande, l'exigence et le surdébit d'obtention de données de canal de scène réelle par le réseau neuronal peuvent être réduits, et la généralisation du réseau neuronal est améliorée en utilisant la capacité de généralisation du livre de codes.
PCT/CN2021/139596 2021-12-20 2021-12-20 Procédé et dispositif de traitement de données WO2023115254A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2021/139596 WO2023115254A1 (fr) 2021-12-20 2021-12-20 Procédé et dispositif de traitement de données

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2021/139596 WO2023115254A1 (fr) 2021-12-20 2021-12-20 Procédé et dispositif de traitement de données

Publications (1)

Publication Number Publication Date
WO2023115254A1 true WO2023115254A1 (fr) 2023-06-29

Family

ID=86900943

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/139596 WO2023115254A1 (fr) 2021-12-20 2021-12-20 Procédé et dispositif de traitement de données

Country Status (1)

Country Link
WO (1) WO2023115254A1 (fr)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111464220A (zh) * 2020-03-10 2020-07-28 西安交通大学 一种基于深度学习的信道状态信息重建方法
CN112583501A (zh) * 2019-09-30 2021-03-30 华为技术有限公司 信道测量方法和通信装置
US20210250068A1 (en) * 2020-02-10 2021-08-12 Korea University Research And Business Foundation Limited-feedback method and device based on machine learning in wireless communication system
CN113381950A (zh) * 2021-04-25 2021-09-10 清华大学 基于网络聚合策略的高效mimo信道反馈方法及装置
CN113381790A (zh) * 2021-06-09 2021-09-10 东南大学 一种基于ai的环境知识辅助的无线信道反馈方法
WO2021237423A1 (fr) * 2020-05-25 2021-12-02 Oppo广东移动通信有限公司 Procédés de transmission d'informations d'état de canal, dispositif électronique et support de stockage
CN113810086A (zh) * 2020-06-12 2021-12-17 华为技术有限公司 信道信息反馈方法、通信装置及存储介质

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112583501A (zh) * 2019-09-30 2021-03-30 华为技术有限公司 信道测量方法和通信装置
US20210250068A1 (en) * 2020-02-10 2021-08-12 Korea University Research And Business Foundation Limited-feedback method and device based on machine learning in wireless communication system
CN111464220A (zh) * 2020-03-10 2020-07-28 西安交通大学 一种基于深度学习的信道状态信息重建方法
WO2021237423A1 (fr) * 2020-05-25 2021-12-02 Oppo广东移动通信有限公司 Procédés de transmission d'informations d'état de canal, dispositif électronique et support de stockage
CN113810086A (zh) * 2020-06-12 2021-12-17 华为技术有限公司 信道信息反馈方法、通信装置及存储介质
CN113381950A (zh) * 2021-04-25 2021-09-10 清华大学 基于网络聚合策略的高效mimo信道反馈方法及装置
CN113381790A (zh) * 2021-06-09 2021-09-10 东南大学 一种基于ai的环境知识辅助的无线信道反馈方法

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
MOSTAFA HUSSIEN; KIM KHOA NGUYEN; MOHAMED CHERIET: "PRVNet: Variational Autoencoders for Massive MIMO CSI Feedback", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 9 November 2020 (2020-11-09), 201 Olin Library Cornell University Ithaca, NY 14853 , XP081809116 *

Similar Documents

Publication Publication Date Title
WO2021217519A1 (fr) Procédé et appareil de réglage de réseau neuronal
WO2023126007A1 (fr) Procédé et appareil de transmission d'informations de canal
US20230136416A1 (en) Neural network obtaining method and apparatus
WO2022253023A1 (fr) Procédé et appareil de communication
WO2023115254A1 (fr) Procédé et dispositif de traitement de données
CN114492784A (zh) 神经网络的测试方法和装置
WO2023070675A1 (fr) Procédé et appareil de traitement de données
WO2023283785A1 (fr) Procédé de traitement de signal, et récepteur
WO2023016503A1 (fr) Procédé et appareil de communication
WO2024008004A1 (fr) Procédé et appareil de communication
WO2024032775A1 (fr) Procédé et appareil de quantification
WO2023231881A1 (fr) Procédé et appareil d'application de modèle
WO2024046288A1 (fr) Procédé et appareil de communication
WO2023279947A1 (fr) Procédé et appareil de communication
US20240120971A1 (en) Information transmission method and apparatus
WO2023006096A1 (fr) Procédé et appareil de communication
WO2023236803A1 (fr) Procédé de transmission d'informations de canal, appareil de communication et dispositif de communication
WO2023133886A1 (fr) Procédé de rétroaction d'informations de canal, dispositif d'extrémité d'envoi, et dispositif d'extrémité de réception
WO2023125699A1 (fr) Procédé et appareil de communication
WO2024046215A1 (fr) Procédé et appareil de communication
US20240137082A1 (en) Communication method and apparatus
WO2024036631A1 (fr) Procédé et appareil de renvoi d'informations, dispositif, et support de stockage
WO2023231934A1 (fr) Procédé et appareil de communication
WO2023060503A1 (fr) Procédé et appareil de traitement d'informations, dispositif, support, puce, produit et programme
WO2024020793A1 (fr) Procédé de rétroaction d'informations d'état de canal (csi), dispositif terminal et dispositif réseau

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21968411

Country of ref document: EP

Kind code of ref document: A1