WO2023125934A1 - Ai network information transmission method and apparatus, and communication device - Google Patents
Ai network information transmission method and apparatus, and communication device Download PDFInfo
- Publication number
- WO2023125934A1 WO2023125934A1 PCT/CN2022/143952 CN2022143952W WO2023125934A1 WO 2023125934 A1 WO2023125934 A1 WO 2023125934A1 CN 2022143952 W CN2022143952 W CN 2022143952W WO 2023125934 A1 WO2023125934 A1 WO 2023125934A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- network
- parameters
- network information
- information
- compressed
- Prior art date
Links
- 230000005540 biological transmission Effects 0.000 title claims abstract description 142
- 238000000034 method Methods 0.000 title claims abstract description 140
- 230000006854 communication Effects 0.000 title claims abstract description 45
- 238000004891 communication Methods 0.000 title claims abstract description 33
- 238000013473 artificial intelligence Methods 0.000 claims description 341
- 230000014509 gene expression Effects 0.000 claims description 56
- 230000000694 effects Effects 0.000 claims description 22
- 230000006835 compression Effects 0.000 claims description 21
- 238000007906 compression Methods 0.000 claims description 21
- 230000001174 ascending effect Effects 0.000 claims description 3
- 230000006870 function Effects 0.000 description 38
- 238000010586 diagram Methods 0.000 description 11
- 238000007726 management method Methods 0.000 description 11
- 238000012549 training Methods 0.000 description 7
- 238000004590 computer program Methods 0.000 description 6
- 238000005516 engineering process Methods 0.000 description 6
- 238000012545 processing Methods 0.000 description 6
- 238000012546 transfer Methods 0.000 description 6
- 238000013528 artificial neural network Methods 0.000 description 5
- 238000005259 measurement Methods 0.000 description 5
- 230000001360 synchronised effect Effects 0.000 description 5
- 230000003190 augmentative effect Effects 0.000 description 2
- 238000013523 data management Methods 0.000 description 2
- 230000007774 longterm Effects 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- 241000699670 Mus sp. Species 0.000 description 1
- 230000004913 activation Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000007599 discharging Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 239000004984 smart glass Substances 0.000 description 1
- 238000005406 washing Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L69/00—Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
- H04L69/04—Protocols for data compression, e.g. ROHC
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W28/00—Network traffic management; Network resource management
- H04W28/02—Traffic management, e.g. flow control or congestion control
- H04W28/06—Optimizing the usage of the radio link, e.g. header compression, information sizing, discarding information
Definitions
- the present application belongs to the technical field of communication, and in particular relates to an AI network information transmission method, device and communication equipment.
- AI Artificial Intelligence
- communication data can be transmitted between network-side devices and terminals through AI networks.
- AI networks In a communication system, the entire AI network is usually transmitted together, resulting in a large system overhead.
- Embodiments of the present application provide an AI network information transmission method, device, and communication equipment, which can solve the problem of relatively large transmission overhead of communication equipment in AI network transmission in the related art.
- an AI network information transmission method including:
- the first end compresses the AI network information, and the AI network information includes at least one of network structure and network parameters;
- the first end sends the compressed AI network information to the second end.
- an AI network information transmission method including:
- the second end receives the compressed AI network information sent by the first end, where the AI network information includes at least one of network structure and network parameters.
- an AI network information transmission device including:
- a compression module configured to compress AI network information, where the AI network information includes at least one of network structure and network parameters;
- a sending module configured to send the compressed AI network information to the second end.
- an AI network information transmission device including:
- the receiving module is configured to receive the compressed AI network information sent by the first end, where the AI network information includes at least one of network structure and network parameters.
- a communication device including a processor and a memory, the memory stores programs or instructions that can run on the processor, and when the programs or instructions are executed by the processor, the first The steps of the AI network information transmission method described in the aspect, or the steps of implementing the AI network information transmission method described in the second aspect.
- a sixth aspect provides a readable storage medium, on which a program or instruction is stored, and when the program or instruction is executed by a processor, the steps of the AI network information transmission method as described in the first aspect are implemented , or implement the steps of the AI network information transmission method as described in the second aspect.
- a chip in the seventh aspect, there is provided a chip, the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is used to run programs or instructions to implement the AI described in the first aspect A network information transmission method, or realize the AI network information transmission method as described in the second aspect.
- a computer program/program product is provided, the computer program/program product is stored in a storage medium, and the computer program/program product is executed by at least one processor to implement the method described in the first aspect The steps of the AI network information transmission method, or the steps of realizing the AI network information transmission method as described in the second aspect.
- the first end can send compressed AI network information to the second end, the AI network information includes at least one of the network structure and network parameters, and there is no need to include all The network structure and network parameters of the entire AI network are transmitted together, so that the network structure and network parameters of the AI network can be sent separately, which can effectively reduce the transmission overhead in the communication process.
- FIG. 1 is a block diagram of a wireless communication system to which an embodiment of the present application is applicable;
- FIG. 2 is a flow chart of an AI network information transmission method provided by an embodiment of the present application.
- Fig. 3 is a flow chart of another AI network information transmission method provided by the embodiment of the present application.
- FIG. 4 is a structural diagram of an AI network information transmission device provided by an embodiment of the present application.
- Fig. 5 is a structural diagram of another AI network information transmission device provided by the embodiment of the present application.
- FIG. 6 is a structural diagram of a communication device provided by an embodiment of the present application.
- FIG. 7 is a structural diagram of a terminal provided in an embodiment of the present application.
- FIG. 8 is a structural diagram of a network-side device provided by an embodiment of the present application.
- FIG. 9 is a structural diagram of another network-side device provided by an embodiment of the present application.
- first, second and the like in the specification and claims of the present application are used to distinguish similar objects, and are not used to describe a specific sequence or sequence. It is to be understood that the terms so used are interchangeable under appropriate circumstances such that the embodiments of the application are capable of operation in sequences other than those illustrated or described herein and that "first" and “second” distinguish objects. It is usually one category, and the number of objects is not limited. For example, there may be one or more first objects.
- “and/or” in the description and claims means at least one of the connected objects, and the character “/” generally means that the related objects are an "or” relationship.
- LTE Long Term Evolution
- LTE-Advanced LTE-Advanced
- LTE-A Long Term Evolution-Advanced
- CDMA Code Division Multiple Access
- TDMA Time Division Multiple Access
- FDMA Frequency Division Multiple Access
- OFDMA Orthogonal Frequency Division Multiple Access
- SC-FDMA Single-carrier Frequency Division Multiple Access
- system and “network” in the embodiments of the present application are often used interchangeably, and the described technologies can be used for the above-mentioned systems and radio technologies as well as other systems and radio technologies.
- NR New Radio
- the following description describes the New Radio (NR) system for illustrative purposes, and uses NR terminology in most of the following descriptions, but these techniques can also be applied to applications other than NR system applications, such as the 6th generation (6 th Generation, 6G) communication system.
- 6G 6th Generation
- Fig. 1 shows a block diagram of a wireless communication system to which the embodiment of the present application is applicable.
- the wireless communication system includes a terminal 11 and a network side device 12 .
- the terminal 11 can be a mobile phone, a tablet computer (Tablet Personal Computer), a laptop computer (Laptop Computer) or a notebook computer, a personal digital assistant (Personal Digital Assistant, PDA), a palmtop computer, a netbook, a super mobile personal computer (ultra-mobile personal computer, UMPC), mobile Internet device (Mobile Internet Device, MID), augmented reality (augmented reality, AR) / virtual reality (virtual reality, VR) equipment, robot, wearable device (Wearable Device) , Vehicle User Equipment (VUE), Pedestrian User Equipment (PUE), smart home (home equipment with wireless communication functions, such as refrigerators, TVs, washing machines or furniture, etc.), game consoles, personal computers (personal computer, PC), teller machine or self-service machine and other terminal side devices, wearable devices include: smart watches, smart bracelet
- the network side device 12 may include an access network device or a core network device, where the access network device may also be called a radio access network device, a radio access network (Radio Access Network, RAN), a radio access network function, or a wireless network. access network unit.
- the access network equipment may include a base station, a wireless local area network (Wireless Local Area Network, WLAN) access point, or a WiFi node, etc., and the base station may be called a node B, an evolved node B (Evolved Node B, eNB), an access point, or a base station.
- BTS Base Transceiver Station
- BSS Basic Service Set
- ESS Extended Service Set
- Home Node B Home Evolved Node B
- TRP Transmitting Receiving Point
- TRP Transmitting Receiving Point
- the core network equipment may include but not limited to at least one of the following: core network node, core network function, mobility management entity (Mobility Management Entity, MME), access mobility management function (Access and Mobility Management Function, AMF), session management function (Session Management Function, SMF), User Plane Function (UPF), Policy Control Function (Policy Control Function, PCF), Policy and Charging Rules Function (PCRF), edge application service Discovery function (Edge Application Server Discovery Function, EASDF), unified data management (Unified Data Management, UDM), unified data storage (Unified Data Repository, UDR), home subscriber server (Home Subscriber Server, HSS), centralized network configuration ( Centralized network configuration, CNC), network storage function (Network Repository Function, NRF), network exposure function (Network Exposure Function, NEF), local NEF (Local NEF, or L-NEF), binding support function (Binding Support Function, BSF), application function (Application Function, AF), etc.
- MME mobility management entity
- AMF Access and Mobility Management Function
- Figure 2 is a flow chart of an AI network information transmission method provided in the embodiment of the present application, as shown in Figure 2, the method includes the following steps:
- Step 201 the first end compresses the AI network information, and the AI network information includes at least one of network structure and network parameters;
- Step 202 the first end sends the compressed AI network information to the second end.
- the first end and the second end are communication devices with sending and receiving functions.
- the first end is one of the network-side device and the terminal
- the second end is the other of the network-side device and the terminal; or, the first end and the second end are different nodes of a terminal; or, the first end and the second end are different nodes of a network side device.
- the network side equipment may include access network equipment (for example: base station) and core network equipment.
- the first end may be an access network device, and the second end may be a core network device; or, the first end may be a terminal, and the second end may be a core network device or an access network device; or, the second end may be a core network device or an access network device;
- the one end and the second end are different nodes of the access network equipment; or, the first end and the second end are different nodes of the core network equipment, etc., and the embodiments of the present application do not list them one by one .
- the AI network information includes at least one of network structure and network parameters.
- the AI network information may be the network structure and/or network parameters of a certain AI network, or the network structures and/or network parameters of multiple AI networks.
- an AI network may also be called an AI neural network, an AI model, or the like.
- the network parameters include weight parameters, hyperparameters and the like of the AI network.
- the compressing the AI network information may refer to compressing the AI network information into a file corresponding to the preset model expression method according to the preset model expression method.
- the so-called model expression method is a kind of data Structure, which describes the AI network structure, network parameters and other information according to certain rules.
- the AI network information includes network structure and/or network parameters
- the compression of the AI network information by the first end also includes compressing the network structure and/or weight parameters, and compressing the compressed network structure and/or or the weight parameter is sent to the second end.
- the AI network information only includes the network structure, the first end only compresses and sends the network structure; or, the AI network information may only include network parameters, then the first end only compresses and sends the network parameters; or , the AI network information includes part of the network structure and part of the network parameters, and the first end compresses and sends the part of the network structure and part of the network parameters.
- the specific information content included in the AI network information may also be in other situations, which will not be described in detail here.
- the AI network information includes at least one of the network structure and network parameters, and there is no need to transmit the entire AI network including all network structures and network parameters together during the communication process, so that the network of the AI network
- the structure and network parameters can be sent separately, which can effectively reduce the transmission overhead in the communication process.
- the AI network information may include network structure and network parameters, and the step 201 may include any of the following:
- the first end combines and compresses the network structure and the network parameters
- the first end compresses the network structure and the network parameters respectively.
- the first end can combine and compress the network structure and network parameters of the AI network into a transmission file based on a preset model expression, or can also compress the network structure and weights based on a preset model expression.
- the parameters are compressed separately into two separate transfer files.
- the step 202 may include any of the following:
- the first end sends the compressed network structure and compressed network parameters to the second end in combination;
- the first end sends the compressed network structure and the compressed network parameters to the second end respectively.
- the first end compresses the network structure and network parameters respectively, such as compressing into two independent transmission files, and compresses the two compressed
- the transmission files are sent to the second end together, or the two compressed transmission files can be sent separately, or only one of the compressed transmission files can be sent, so that the first end can receive the compressed AI network information
- the transmission method is more flexible.
- the step 201 may also include any of the following:
- the first end converts the AI network information into a corresponding transmission file based on a preset model representation, and compresses the transmission file;
- the first end compresses the AI network information based on a preset data format
- the first end obtains the AI network information to be sent and the existing AI network information of the second end, and calculates the AI network information between the AI network information to be sent and the existing AI network information of the second end. Network information difference is compressed;
- the first end obtains the AI network information to be sent and the AI network information of the preset AI network, and calculates the AI network information difference between the AI network information to be sent and the AI network information of the preset AI network to compress.
- the preset model expression method may be an AI network expression method common to both the first end and the second end, such as open neural network exchange (ONNX), TensorFlow, and the like.
- Open neural network exchange ONNX
- TensorFlow is a machine learning framework.
- the first end can convert the AI network information into a corresponding transmission file based on ONNX, and then compress the transmission file and send it to the second end. After the second end decompresses the compressed transmission file, it can also Based on ONNX, the decompressed transmission file is converted into its own applicable network structure and network parameters.
- AI network information saved under the two different neural network frameworks is different and cannot be read directly.
- the AI network information is converted into the corresponding transmission file through the preset model expression method and then compressed and transmitted. It also enables two communication devices using different neural network frameworks to realize the reading and application of AI network information.
- the first end may also compress the AI network information based on a preset data format, such as the protobuf data format used by ONNX, where protobuf is a data exchange format.
- a preset data format such as the protobuf data format used by ONNX, where protobuf is a data exchange format.
- the first end may also compress the AI network information difference between the AI network information to be sent and the existing AI network information at the second end, and the first end only needs to send the compressed AI network information difference value to the second end, and then there is no need to compress and send the same AI network information between the AI network information to be sent and the existing AI network information of the second end, which can effectively save the transmission of the first end overhead.
- the first end may compress the AI network information difference between the AI network information to be sent and the AI network information of the preset AI network, and the first end only needs to send the compressed AI network information difference to the second end to save the transmission overhead of the first end.
- the preset AI network may be a predetermined protocol or a high-level configuration, such as some fixed AI network templates, or may also be a common AI network template for communication devices.
- the preset AI network includes initial values of network structure and network parameters.
- the AI network information difference includes at least one of the following:
- the reference value being the maximum value in the network parameters
- the preset model expression method includes any one of the following: a model expression method stipulated in an agreement, and a user-defined model expression method.
- the self-defined model expression method may refer to a data structure defined by the communication device, which is used to describe the network structure and network parameters of the AI network.
- the content of the self-defined model expression includes at least one of the following: the network structure of the AI network, the attributes of the network parameters of the AI network, and the values of the network parameters of the AI network.
- the attribute of the network parameter includes information such as the name, identification, and dimension of the network parameter; the value of the parameter of the AI network may be one or more.
- the representation of the network structure in the preset model representation includes at least one of the following:
- the updated numerical position in the network parameters of the AI network is the updated numerical position in the network parameters of the AI network.
- association relationship between the network structures may refer to the connection relationship between the input and output of each network structure (or also called nodes), for example, the output of the first node is connected to the input of the second node, and the output of the first node is connected to the input of the second node.
- the output of the second node is connected to the input of the third node, and so on.
- the attribute of the network parameter includes information such as a name, an identifier, and a dimension of the network parameter.
- the first end converts the AI network information into a corresponding transmission file based on a preset model expression method, and compresses the transmission file, including:
- the first end converts the AI network information into at least one transmission file based on at least one preset model representation, and one preset model representation corresponds to at least one transmission file;
- the first end merges and compresses the at least one transmission file, or the first end compresses the at least one transmission file separately and then merges them.
- one preset model expression may correspond to at least one transmission file, for example, the network structure and network parameters are respectively converted based on a preset model expression into corresponding transfer files.
- the first end converts the AI network information into at least one transmission file based on at least one preset model expression, merge the at least one transmission file together for compression, and send the compressed transmission file, or
- the at least one transmission file mentioned above is compressed independently, and then combined and sent together after compression, or each compressed transmission file can also be sent separately.
- the AI network information includes network structure and network parameters
- the first end converts the AI network information into a corresponding transmission file based on a preset model expression, and compresses the transmission file, including:
- the first end converts the network structure into a first transmission file based on a preset model representation, converts the network parameters into a second transmission file based on the preset model representation, and converts the first The first transmission file and the second transmission file are respectively compressed.
- the first end can convert the network structure and network parameters into corresponding transmission files based on the preset model expression, and then obtain two transmission files, and compress the two transmission files respectively, and can separately
- the two compressed transmission files may be sent, or they may be combined and sent together.
- the first end is a base station
- the second end is a terminal.
- the base station may save the network structure of the trained AI network into a transmission file in a corresponding format based on a preset model expression Compress and send to the terminal; if network parameters are required, the base station saves the network parameters based on the preset model expression method as a transmission file for compression and sends to the terminal.
- the transmission file corresponding to the network structure and the transmission file corresponding to the network parameters may be of different file types.
- the base station may send the compressed transmission file through the data channel.
- the compressed AI network information when the AI network information includes network parameters, the compressed AI network information also includes compressed network parameters, and the first end sends the compressed AI network information to the second end ,include:
- the first end sends the compressed network parameters to the second end according to the priority order of the compressed network parameters based on the priority order of the compressed network parameters.
- the base station when it sends compressed network parameters, it may divide the compressed network parameters into N groups according to a preset priority order, Groups with high priority are delivered first, and groups with low priority are delivered later or not delivered when transmission resources are limited.
- the first end sends the compressed network parameters to the second end according to the priority order based on the priority order of the compressed network parameters, including:
- the first end discards the grouped network parameters and sends the remaining network parameters in a preset order, the preset order being the grouped Network parameters are prioritized in order from low to high.
- the first end is a base station
- the second end is a terminal.
- the base station transmits the compressed network parameters
- it may divide the compressed network parameters into N groups according to a preset priority order, and transmit the compressed network parameters in the base station.
- the resource is less than the preset threshold, that is, when the transmission resources of the base station are limited, or when there is a burst of high-priority traffic
- the base station discards the network parameters in order of priority from low to high, that is, the priority
- the network parameters of the lowest group are discarded first until the transmission resources of the base station are sufficient to send the network parameters of the remaining groups, which ensures that the network parameters sent by the base station are network parameters with higher priority.
- the terminal may use a default value agreed in the protocol or be 0 for the network parameters not received.
- the method before the first end compresses the AI network information, the method further includes:
- the first end receives first request information sent by the second end, and the first request information is used to request acquisition of target AI network information;
- the first end compresses the AI network information, including:
- the first end compresses the target AI network information.
- the second end can obtain the specified target AI network information based on the first request information, and then the first end compresses the target AI network information based on the first request information and sends it to the second end.
- the method before the first end compresses the target AI network information, the method further includes:
- the first end judges whether the target AI network information needs to be updated
- the first end compresses the target AI network information, including:
- the first end compresses the updated target AI network information.
- the first end can determine which target AI network information the second end wants to obtain based on the first update request. Before compressing the target AI network information, the first end may determine whether the target AI network information needs to be updated, if necessary, update the target AI network information, and update the target AI network information The network information is compressed and then sent to the second end; if the target AI network information does not need to be updated, the first end may directly compress the target AI network information and then send it to the second end.
- the second end sends a weight parameter update request (Request of Weight Updating) to the first end, specifying that one or some specific weight parameters need to be obtained, and the first end determines that an update is required based on the weight parameter update request , to update the specified weight parameters, and send the updated weight parameters to the second end after being compressed based on the preset model representation.
- a weight parameter update request Request of Weight Updating
- the first request information includes at least one of the following:
- the network effect measurement value may mean that the second end calculates the effect of the AI network according to a certain agreed method, and reports the calculated value, and the first end judges whether AI is required based on the calculated value carried in the first request information.
- Updates to network information For example, for channel state information (Channel State Information, CSI) prediction, the terminal (second end) calculates the correlation between the predicted result and the actual measurement result and reports it to the base station (first end). Effect measurement value.
- the base station judges whether to update the AI network information according to the correlation reported by the terminal and the correlation calculated by its own AI network. If the correlation is poor, it means that the channel changes too fast and does not need to be updated, because the base station side The AI network also performed poorly. If the correlation calculated by the AI network on the base station side is significantly better than the correlation reported by the terminal, it means that the parameters of the AI network of the terminal are outdated and cannot match the current channel, and the base station needs to resend the AI network parameters.
- CSI Channel State Information
- the target AI network information includes first target network parameters, and the first end compresses the target AI network information, including:
- the first end converts the attributes and parameter values of the first target network parameters into a preset format based on a preset model representation and then compresses them;
- the attribute of the first target network parameter includes at least one of name, dimension, and length.
- the terminal when the terminal initially accesses, it may select an AI network from the preset AI networks, and send the identifier of the AI network to the terminal,
- the attributes and parameter values of the network parameters of the AI network can also be converted into a preset format and then compressed, for example, compressed according to a preset model expression, and then sent to the terminal.
- the method further includes:
- the first end calculates the correlation between the network parameters of the first cell and the network parameters of the second cell, and acquires a second target network parameter, where the second target network parameter includes at least one of the following items: the correlation is less than Network parameters with preset thresholds and the first N network parameters in a preset sequence, where the preset sequence is a sequence in which the correlation of network parameters is arranged in ascending order;
- the first end compresses the AI network information, including:
- the first end compresses the second target network parameters
- the first end sends the compressed AI network information to the second end, including:
- the first end sends the compressed second target network parameters to the second end.
- the network side device may send complete network parameters of an AI network to the terminal, or may also send partial network parameters to the terminal. For example, when a terminal switches from a first cell to a second cell, the second cell may obtain the network parameters corresponding to the first cell from the first cell, calculate the correlation of each network parameter, and the network side device will set the correlation to be less than the preset The threshold network parameters or the first N network parameters with less correlation are compressed and sent to the terminal, and the terminal may only update the received network parameters.
- the first end and the second end can interact in advance with the attributes of the network parameters to be transferred, including name, dimension, length, etc., and then pass the network parameters to be transferred through the preset
- the model expression is converted into the corresponding file and then compressed and delivered.
- the order of all network parameters can be considered known, and the position of the network parameters that need to be updated in the overall list can be exchanged through bitmaps or combination numbers.
- the first end when describing the network parameters based on the preset model expression, can compress and transmit the attributes of the network parameters and the parameter values together, and the second end can judge based on the attributes of the received network parameters The length of the network parameters and the corresponding weights.
- a network parameter can be a single number or a list of multiple numbers.
- the first end can indicate the position of the changed value through the value position. The first end Only the changed value is transmitted, and the second end only updates the received value.
- the first end transmits complete network parameters, it can also indicate the position of the non-zero value in the network parameter through the value position, the first end only transmits the non-zero value, and the second end treats the value that has not been received as 0 or a default value.
- the numerical location indication may be an additional indication independent of network parameters, or a radio resource control (Radio Resource Control, RRC) configuration or a medium access control control element (Medium Access Control Element, MACCE) configuration.
- RRC Radio Resource Control
- MACCE Medium Access Control Element
- the terminal and the base station use the joint AI network for CSI feedback, that is, the terminal converts the channel information into several bits of CSI feedback information through the AI network, and reports it to the base station.
- the network recovers the channel information.
- the base station Since the network of the base station and the terminal needs to be jointly trained, and the channel conditions of different cells are different, new network parameters may also be required. Therefore, when the terminal accesses the network, the base station needs to send the network parameters used by the terminal to the terminal.
- the CSI feedback network can be divided into two parts, the terminal coding part and the base station decoding part.
- the base station only needs to send the AI network of the terminal coding part to the terminal.
- the base station can save the network structure and network parameters of the AI network to be sent into a file corresponding to the preset model expression method, such as an ONNX file or a pth file of pytorch, and then pass the entire file through the Packet Data Convergence Protocol (Packet Data Convergence Protocol, PDCP) layer compressed and sent to the terminal.
- the terminal After receiving the file, the terminal loads it into its own AI framework according to the file format to realize inference or continue training using the AI network.
- pytorch is a deep learning framework
- pth is a file type in pytorch.
- the base station saves the AI network structure as a file corresponding to a preset model expression method, such as a TensorFlow meta file, and then sends the meta file to the terminal after ZIP compression.
- a preset model expression method such as a TensorFlow meta file
- the terminal After the terminal receives the file, it builds the corresponding AI network under its own AI framework.
- the weight parameter that has not been received defaults to 0 or a fixed initial value.
- the terminal can use this initial value for training, or it can be used directly This default value is inferred.
- meta is a file type in TensorFlow.
- the base station sends a certain index value (index) to indicate the corresponding AI network template, and then saves the network parameters as a file corresponding to the preset model expression , such as the data file of TensorFlow, or the index and network parameters of the network template can be saved together as a file corresponding to a custom network expression method and sent to the terminal.
- index is a file type in TensorFlow.
- the terminal After receiving the corresponding file, the terminal loads the corresponding network structure in its own AI framework according to the index of the pre-defined AI network template, initializes its own network parameters according to the network parameters of the AI network template, and then transfers the network parameters sent by the base station. The parameters are updated to the AI network. If some network parameters are not received, the initial values in the AI template are used.
- the terminal can use the AI network to perform channel measurement, including CSI reference signal (CSI Reference Signal, CSI-RS), demodulation reference signal (Demodulation Reference Signal, DMRS) and other channel estimation and radio resource management (Radio resource management, RRM) measurement, etc.
- CSI reference signal CSI Reference Signal
- DMRS demodulation Reference Signal
- RRM Radio resource management
- the network structure of the AI network used by the terminal in the new cell can generally be considered to be consistent with the AI network structure of the old cell.
- the new The cell can send its own trained network parameters with the same network structure as the terminal to the terminal.
- the terminal continues to train and infer based on these parameters, which can improve real-time efficiency, reduce the number of training times, and quickly converge the network.
- the new cell can obtain the network parameters sent to the terminal last time from the old cell, compare its own network parameters with the network parameters sent to the user last time, and calculate the average correlation of weight values, for example:
- the base station judges which network parameters need to be updated according to the correlation between the two weight parameters, and selects M2 network parameters to be updated from the total M1 network parameters to update.
- the base station can first inform the terminal which network parameters need to be updated. Since the terminal knows the network structure, it naturally knows a total of M1 network parameters. The base station and the terminal use the same sorting method for the M1 network parameters. This sorting method corresponds to the preset The base station notifies the terminal of the position of the M2 network parameters that need to be updated in this sort by means of a bitmap (bitmap) or combination number, etc. After receiving the notification from the base station, the terminal determines which M2 network parameters need to be updated , and then receive specific parameter values.
- bitmap bitmap
- the base station Since it has been determined in advance which network parameters need to be updated, the base station directly compresses the values of the network parameters that need to be updated into a data file in the preset model expression mode, without including weight name, dimension and other information, and directly transmits the values, and the terminal receives the values Afterwards, according to the dimensions of each network parameter that you know, analyze the parameter value of this network parameter at the corresponding position, and update the network parameters in your own AI framework.
- the base station can directly compress the name and value of the network parameter to be updated into a file in a preset model expression mode, and the terminal first parses the name of the network parameter, and then obtains the corresponding value according to the name.
- the base station first calculates the corresponding correlation among all K values of the network parameter, and if the correlation is greater than a certain fixed threshold, it is considered that this value does not need to be updated, Then compress the position of all the values that need to be updated into the file corresponding to the preset model expression method through the method of bitmap or combination number and the value that needs to be transmitted. After receiving the compressed file, the terminal will first extract which network parameters It needs to be updated, and then judge which values of this network parameter need to be updated, and only update the values of the corresponding positions.
- the network parameters that need to be updated, and the numerical position that needs to be updated in each network parameter can be indicated in advance or configured to the terminal, and then the corresponding numerical information can be directly compressed and sent to the terminal.
- the so-called preset model expression method is a data structure, according to a certain Rules describe the AI network and describe information such as AI network parameters.
- the preset model expression methods include TensorFlow, PyTorch, ONNX, etc., which describe the AI network description according to fixed rules, such as describing the network as a combination of several independent nodes, or describing it as a combination of several layers , each layer has an independent function, the specific function is expressed by some weight parameters and activation functions, etc., different frameworks have their own definition of the network, in order to uniformly transmit in the communication system, you can define some communication requirements network description.
- the network structure and network parameters of the AI network should be separated to support the transfer of independent network structures and independent network parameters.
- the network structure can be based on the node, define the basic functions of the node, map all the weights to the input and output, and use the same number to represent the connection through the input and output, so as to describe the entire network through the number of nodes and input and output.
- For network parameters include at least one of the following:
- the position information of the effective value in the parameter, the parameter that has not changed or is close to 0 may not be passed, and the corresponding position information is required to indicate the position of the effective value.
- some functions only require the terminal to use the AI network, and do not require joint training with network-side devices, such as channel prediction and positioning on the terminal side.
- These AI networks are affected by the channel environment. As the terminal moves, it needs to be constantly Train and update the network.
- terminal B enters the area of terminal A, terminal A can pass the trained AI network to terminal B, and terminal B performs training based on terminal A's AI network. Since terminal A and terminal B are in similar areas, the channel The information has a certain correlation, so using the AI network of terminal A to continue training can help terminal B converge faster.
- the training complexity can be reduced by passing the trained AI network, or even directly use the AI network trained by other nodes.
- Terminal A and terminal B have the same AI network structure for the same function. The two need to interact mainly with network parameters.
- terminal A can divide the network parameters into N parts, respectively In different time slots (slots), it is transmitted to terminal B.
- information such as network parameter names and dimensions can be uniformly compressed with parameter values or can be directly transmitted after independent configuration.
- terminal B After receiving part of the network parameters of terminal A, terminal B can directly update the network parameters of its corresponding location, and use the updated network to directly infer, or terminal B can wait until all network parameters are received, and update its corresponding network parameters together.
- Network parameters which mainly depends on whether some network parameters can work independently. This capability can be notified by terminal A to terminal B, or configured through a common base station. If it can work independently, terminal B will update each time it receives new network parameters. You can update your own network, otherwise you have to wait until all are received and update together.
- Terminal B can also specify a network parameter that needs to be updated, and send information such as the name and/or dimension of the network parameter that needs to be updated to terminal A, and can also include information such as the location of the resource that is expected to receive the network parameter.
- Terminal A according to terminal B According to the order of the network parameters required by terminal A, the corresponding network parameter values are directly compressed into a file corresponding to a preset model expression method and sent to terminal B. After terminal B receives it, it updates the network it needs parameter.
- FIG. 3 is another AI network information transmission method provided by the embodiment of the present application. As shown in FIG. 3, the AI network information transmission method includes the following steps:
- Step 301 The second end receives the compressed AI network information sent by the first end, where the AI network information includes at least one of network structure and network parameters.
- the first end and the second end are communication devices with sending and receiving functions.
- the first end is one of the network-side device and the terminal
- the second end is the other of the network-side device and the terminal; or, the first end and the second end are different nodes of a terminal; or, the first end and the second end are different nodes of a network side device.
- the network side equipment may include access network equipment (for example: base station) and core network equipment.
- the first end may be an access network device, and the second end may be a core network device; or, the first end may be a terminal, and the second end may be a core network device or an access network device; or, the second end may be a core network device or an access network device;
- the one end and the second end are different nodes of the access network equipment; or, the first end and the second end are different nodes of the core network equipment, etc., and the embodiments of the present application do not list them one by one .
- the AI network information includes at least one of network structure and network parameters.
- the AI network information may be the network structure and/or network parameters of a certain AI network, or the network structures and/or network parameters of multiple AI networks.
- an AI network may also be called an AI neural network, an AI model, or the like.
- the network parameters include weight parameters, hyperparameters and the like of the AI network.
- the first end compresses the AI network information, and sends the compressed AI network information to the second end.
- the compressing the AI network information may refer to compressing the AI network information into a file corresponding to the preset model representation method according to the preset model representation method, the so-called model representation method is a data structure, Describe the AI network structure, network parameters and other information according to certain rules.
- the first terminal compressing the AI network information reference may be made to the specific description in the method embodiment shown in FIG. 2 above, and details are not repeated here.
- the AI network information includes network structure and/or network parameters
- the compression of the AI network information by the first end also includes compressing the network structure and/or weight parameters, and compressing the compressed network structure and/or weight parameters Parameters are sent to the second end.
- the AI network information only includes the network structure, the first end only compresses and sends the network structure; or, the AI network information may only include network parameters, then the first end only compresses and sends the network parameters; or , the AI network information includes part of the network structure and part of the network parameters, and the first end compresses and sends the part of the network structure and part of the network parameters.
- the specific information content included in the AI network information may also be in other situations, which will not be described in detail here.
- the AI network information includes at least one of the network structure and network parameters, and there is no need to transmit the entire AI network including all network structures and network parameters together during the communication process, so that the network of the AI network
- the structure and network parameters can be sent separately, which can effectively reduce the transmission overhead in the communication process.
- the method further includes:
- the second terminal sends first request information to the first terminal, where the first request information is used to request acquisition of target AI network information;
- the step 301 includes:
- the second end receives the compressed target AI network information sent by the first end.
- the first request information includes at least one of the following:
- the AI network information transmission method provided by the embodiment of the present application is applied to the second end, corresponding to the AI network information transmission method applied to the first end provided in the embodiment of FIG. 2 above.
- the specific implementation process of the relevant steps reference may be made to the description in the above-mentioned method embodiment shown in FIG. 2 , and details are not repeated here to avoid repetition.
- the second end receives the compressed AI network information sent by the first end, the AI network information includes at least one of the network structure and network parameters, and there is no need to include all network information during the communication process.
- the structure and network parameters of the entire AI network are compressed and transmitted together, so that the network structure and network parameters of the AI network can be sent separately, which can effectively reduce the transmission overhead in the communication process.
- the AI network information transmission method provided in the embodiment of the present application may be executed by an AI network information transmission device.
- the AI network information transmission device provided in the embodiment of the present application is described by taking the AI network information transmission device executing the AI network information transmission method as an example.
- FIG. 4 is a structural diagram of an AI network information transmission device provided in an embodiment of the present application. As shown in FIG. 4, the AI network information transmission device 400 includes:
- a compression module 401 configured to compress AI network information, where the AI network information includes at least one of network structure and network parameters;
- the sending module 402 is configured to send the compressed AI network information to the second end.
- the AI network information includes network structure and network parameters
- the compression module 401 is configured to perform any of the following:
- the sending module 402 is configured to perform any of the following:
- the compressed network structure and the compressed network parameters are respectively sent to the second end.
- the compression module 401 is configured to perform any of the following:
- the AI network information difference includes at least one of the following:
- the reference value being the maximum value in the network parameters
- the preset model expression manner includes any one of the following: a protocol-agreed model expression manner, and a user-defined model expression manner.
- the content of the self-defined model expression includes at least one of the following: the network structure of the AI network, the attributes of the network parameters of the AI network, and the values of the network parameters of the AI network.
- the representation of the network structure in the preset model representation includes at least one of the following:
- the updated numerical position in the network parameters of the AI network is the updated numerical position in the network parameters of the AI network.
- the compression module 401 is also used for:
- the AI network information includes network structure and network parameters
- the compression module 401 is also used for:
- the compressed AI network information includes compressed network parameters
- the sending module 402 is further configured to:
- the sending module 402 is also configured to:
- the grouped network parameters are discarded and the remaining network parameters are sent according to a preset order, the preset order being the order of priority of the grouped network parameters from low to high.
- the device also includes:
- a receiving module configured to receive first request information sent by the second end, where the first request information is used to request acquisition of target AI network information;
- the compression module 401 is further configured to: compress the target AI network information.
- the first request information includes at least one of the following:
- the device also includes:
- a judging module configured to judge whether the target AI network information needs to be updated
- the compression module 401 is also used for:
- the target AI network information includes first target network parameters
- the compression module 401 is further configured to:
- the attribute of the first target network parameter includes at least one of name, dimension, and length.
- the device is one of the network side device and the terminal, and the second end is the other of the network side device and the terminal; or,
- the device and the second end are different nodes of a terminal; or,
- the device and the second end are different nodes of the network side equipment.
- the device is a network side device
- the second end is a terminal
- the AI network information includes network parameters
- the method further includes:
- the first end calculates the correlation between the network parameters of the first cell and the network parameters of the second cell, and obtains a second target network parameter, where the second target network parameter includes at least one of the following items: the correlation is less than Network parameters with preset thresholds, and the first N network parameters in a preset sequence, where the preset sequence is a sequence in which the correlation of network parameters is arranged in ascending order;
- the compression module 401 is also used for:
- the sending module 402 is also used for:
- the device can send compressed AI network information to the second end, the AI network information includes at least one of the network structure and network parameters, and there is no need to include all network information during the communication process.
- the structure and network parameters of the entire AI network are transmitted together, so that the network structure and network parameters of the AI network can be sent separately, which can effectively reduce the transmission overhead in the communication process.
- the AI network information transmission apparatus 400 in the embodiment of the present application may be an electronic device, such as an electronic device with an operating system, or a component of the electronic device, such as an integrated circuit or a chip.
- the electronic device may be a terminal, or other devices other than the terminal.
- the terminal may include, but not limited to, the types of terminal 11 listed above, and other devices may be servers, Network Attached Storage (NAS), etc., which are not specifically limited in this embodiment of the present application.
- NAS Network Attached Storage
- the AI network information transmission device 400 provided in the embodiment of the present application can realize each process implemented in the method embodiment shown in FIG. 2 and achieve the same technical effect. To avoid repetition, details are not repeated here.
- FIG. 5 is a structural diagram of another AI network information transmission device provided in the embodiment of the present application. As shown in FIG. 5, the AI network information transmission device 500 includes:
- the receiving module 501 is configured to receive compressed AI network information sent by the first end, where the AI network information includes at least one of network structure and network parameters.
- the device also includes:
- a sending module configured to send first request information to the first end, where the first request information is used to request acquisition of target AI network information;
- the receiving module 501 is also used for:
- the first request information includes at least one of the following:
- the first end is one of the network side device and the terminal, and the device is the other of the network side device and the terminal; or,
- the first end and the device are different nodes of terminals; or,
- the first end and the device are different nodes of network side equipment.
- the device receives the compressed AI network information sent by the first end, the AI network information includes at least one of the network structure and network parameters, and there is no need to include all network information during the communication process.
- the structure and network parameters of the entire AI network are compressed and transmitted together, so that the network structure and network parameters of the AI network can be sent separately, which can effectively reduce the transmission overhead in the communication process.
- the AI network information transmission device 500 provided in the embodiment of the present application can realize each process implemented in the method embodiment shown in FIG. 3 and achieve the same technical effect. To avoid repetition, details are not repeated here.
- the embodiment of the present application further provides a communication device 600, including a processor 601 and a memory 602, and the memory 602 stores programs or instructions that can run on the processor 601.
- a communication device 600 including a processor 601 and a memory 602 stores programs or instructions that can run on the processor 601.
- the programs or instructions are executed by the processor 601, the various steps of the embodiment of the AI network information transmission method described above in FIG. 2 or FIG. 3 can be realized, and the same technical effect can be achieved. To avoid repetition, details are not repeated here.
- FIG. 7 is a schematic diagram of a hardware structure of a terminal implementing an embodiment of the present application.
- the terminal 700 includes, but is not limited to: a radio frequency unit 701, a network module 702, an audio output unit 703, an input unit 704, a sensor 705, a display unit 706, a user input unit 707, an interface unit 708, a memory 709, and a processor 710. At least some parts.
- the terminal 700 may also include a power supply (such as a battery) for supplying power to various components, and the power supply may be logically connected to the processor 710 through the power management system, so as to manage charging, discharging, and power consumption through the power management system. Management and other functions.
- a power supply such as a battery
- the terminal structure shown in FIG. 7 does not constitute a limitation on the terminal, and the terminal may include more or fewer components than shown in the figure, or combine some components, or arrange different components, which will not be repeated here.
- the input unit 704 may include a graphics processing unit (Graphics Processing Unit, GPU) 7041 and a microphone 7042, and the graphics processor 7041 is used by the image capture device (such as the image data of the still picture or video obtained by the camera) for processing.
- the display unit 706 may include a display panel 7061, and the display panel 7061 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like.
- the user input unit 707 includes at least one of a touch panel 7071 and other input devices 7072 .
- the touch panel 7071 is also called a touch screen.
- the touch panel 7071 may include two parts, a touch detection device and a touch controller.
- Other input devices 7072 may include, but are not limited to, physical keyboards, function keys (such as volume control buttons, switch buttons, etc.), trackballs, mice, and joysticks, which will not be described in detail here.
- the radio frequency unit 701 may transmit the downlink data from the network side device to the processor 710 for processing after receiving the downlink data; in addition, the radio frequency unit 701 may send uplink data to the network side device.
- the radio frequency unit 701 includes, but is not limited to, an antenna, an amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like.
- the memory 709 can be used to store software programs or instructions as well as various data.
- the memory 709 may mainly include a first storage area for storing programs or instructions and a second storage area for storing data, wherein the first storage area may store an operating system, an application program or instructions required by at least one function (such as a sound playing function, image playback function, etc.), etc.
- memory 709 may include volatile memory or nonvolatile memory, or, memory 709 may include both volatile and nonvolatile memory.
- the non-volatile memory can be read-only memory (Read-Only Memory, ROM), programmable read-only memory (Programmable ROM, PROM), erasable programmable read-only memory (Erasable PROM, EPROM), electronically programmable Erase Programmable Read-Only Memory (Electrically EPROM, EEPROM) or Flash.
- ROM Read-Only Memory
- PROM programmable read-only memory
- Erasable PROM Erasable PROM
- EPROM erasable programmable read-only memory
- Electrical EPROM Electrical EPROM
- EEPROM electronically programmable Erase Programmable Read-Only Memory
- Volatile memory can be random access memory (Random Access Memory, RAM), static random access memory (Static RAM, SRAM), dynamic random access memory (Dynamic RAM, DRAM), synchronous dynamic random access memory (Synchronous DRAM, SDRAM), double data rate synchronous dynamic random access memory (Double Data Rate SDRAM, DDRSDRAM), enhanced synchronous dynamic random access memory (Enhanced SDRAM, ESDRAM), synchronous connection dynamic random access memory (Synch link DRAM , SLDRAM) and Direct Memory Bus Random Access Memory (Direct Rambus RAM, DRRAM).
- RAM Random Access Memory
- SRAM static random access memory
- DRAM dynamic random access memory
- DRAM synchronous dynamic random access memory
- SDRAM double data rate synchronous dynamic random access memory
- Double Data Rate SDRAM Double Data Rate SDRAM
- DDRSDRAM double data rate synchronous dynamic random access memory
- Enhanced SDRAM, ESDRAM enhanced synchronous dynamic random access memory
- Synch link DRAM , SLDRAM
- Direct Memory Bus Random Access Memory Direct Rambus
- the processor 710 may include one or more processing units; optionally, the processor 710 integrates an application processor and a modem processor, wherein the application processor mainly handles operations related to the operating system, user interface, and application programs, etc., Modem processors mainly process wireless communication signals, such as baseband processors. It can be understood that the foregoing modem processor may not be integrated into the processor 710 .
- the terminal 700 is the first end.
- the processor 710 is configured to: compress the AI network information, where the AI network information includes at least one of network structure and network parameters;
- the radio frequency unit 701 is configured to: send the compressed AI network information to the second end.
- the AI network information includes network structure and network parameters
- the processor 710 is configured to perform any of the following:
- the radio frequency unit 701 is further configured to perform any of the following:
- the compressed network structure and the compressed network parameters are respectively sent to the second end.
- processor 710 is configured to perform any of the following:
- the AI network information difference includes at least one of the following:
- the reference value being the maximum value in the network parameters
- the preset model expression manner includes any one of the following: a protocol-agreed model expression manner, and a user-defined model expression manner.
- the content of the self-defined model expression includes at least one of the following: the network structure of the AI network, the attributes of the network parameters of the AI network, and the values of the network parameters of the AI network.
- the representation of the network structure in the preset model representation includes at least one of the following:
- the updated numerical position in the network parameters of the AI network is the updated numerical position in the network parameters of the AI network.
- processor 710 is further configured to:
- the AI network information includes network structure and network parameters
- the processor 710 is further configured to:
- the compressed AI network information includes compressed network parameters
- the radio frequency unit 701 is further configured to:
- the radio frequency unit 701 is also used for:
- the grouped network parameters are discarded and the remaining network parameters are sent according to a preset order, the preset order being the order of priority of the grouped network parameters from low to high.
- the radio frequency unit 701 is also used for:
- the processor 710 is also used for:
- the first request information includes at least one of the following:
- processor 710 is further configured to:
- the target AI network information includes first target network parameters
- the processor 710 is further configured to:
- the attribute of the first target network parameter includes at least one of name, dimension, and length.
- the terminal 700 is a second terminal.
- the radio frequency unit 701 is further configured to: receive compressed AI network information sent by the first end, where the AI network information includes at least one of network structure and network parameters.
- the radio frequency unit 701 is also used for:
- the first request information includes at least one of the following:
- the technical solution provided by this application enables the network structure and network parameters of the AI network to be sent separately, thereby effectively reducing the transmission overhead in the communication process.
- the embodiment of the present application also provides a network-side device.
- the various implementation processes and implementation methods of the above-mentioned method embodiments shown in FIG. 2 and FIG. 3 can be applied to the network-side device embodiment, and can achieve the same technical effect.
- the embodiment of the present application also provides a network side device.
- the network side device 800 includes: an antenna 81 , a radio frequency device 82 , a baseband device 83 , a processor 84 and a memory 85 .
- the antenna 81 is connected to a radio frequency device 82 .
- the radio frequency device 82 receives information through the antenna 81, and sends the received information to the baseband device 83 for processing.
- the baseband device 83 processes the information to be sent and sends it to the radio frequency device 82
- the radio frequency device 82 processes the received information and sends it out through the antenna 81 .
- the method performed by the network side device in the above embodiments may be implemented in the baseband device 83, where the baseband device 83 includes a baseband processor.
- the baseband device 83 can include at least one baseband board, for example, a plurality of chips are arranged on the baseband board, as shown in FIG.
- the program executes the network device operations shown in the above method embodiments.
- the network side device may also include a network interface 86, such as a common public radio interface (common public radio interface, CPRI).
- a network interface 86 such as a common public radio interface (common public radio interface, CPRI).
- the network-side device 800 in this embodiment of the present disclosure further includes: instructions or programs stored in the memory 85 and operable on the processor 84, and the processor 84 calls the instructions or programs in the memory 85 to execute FIG. 4 or FIG. 5
- the methods executed by each module shown in the figure achieve the same technical effect, so in order to avoid repetition, they are not repeated here.
- the embodiment of the present application also provides another network side device.
- the network side device 900 includes: a processor 901 , a network interface 902 and a memory 903 .
- the network interface 902 is, for example, a common public radio interface (common public radio interface, CPRI).
- the network-side device 900 in this embodiment of the present disclosure further includes: instructions or programs stored in the memory 903 and executable on the processor 901, and the processor 901 calls the instructions or programs in the memory 903 to execute FIG. 4 or FIG. 5
- the methods executed by each module shown in the figure achieve the same technical effect, so in order to avoid repetition, they are not repeated here.
- the embodiment of the present application also provides a readable storage medium, on which a program or instruction is stored, and when the program or instruction is executed by the processor, each process of the method embodiment described above in FIG. 2 or FIG. 3 is implemented. , and can achieve the same technical effect, in order to avoid repetition, it will not be repeated here.
- the processor is the processor in the terminal described in the foregoing embodiments.
- the readable storage medium includes a computer-readable storage medium, such as a computer read-only memory ROM, a random access memory RAM, a magnetic disk or an optical disk, and the like.
- the embodiment of the present application further provides a chip, the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is used to run programs or instructions to implement the above-mentioned Figure 2 or Figure 3.
- the chip includes a processor and a communication interface
- the communication interface is coupled to the processor
- the processor is used to run programs or instructions to implement the above-mentioned Figure 2 or Figure 3.
- the chip mentioned in the embodiment of the present application may also be called a system-on-chip, a system-on-chip, a system-on-a-chip, or a system-on-a-chip.
- the embodiment of the present application further provides a computer program/program product, the computer program/program product is stored in a storage medium, and the computer program/program product is executed by at least one processor to implement the above-mentioned Figure 2 or Figure 3
- the various processes of the method embodiments can achieve the same technical effect, and are not repeated here to avoid repetition.
- the embodiment of the present application also provides a communication system, including: a terminal and a network-side device, the terminal can be used to perform the steps of the method described in Figure 2 above, and the network-side device can be used to perform the method described in Figure 3 above or, the terminal may be used to perform the steps of the method described in FIG. 3 above, and the network side device may be used to perform the steps of the method described in FIG. 2 above.
- a communication system including: a terminal and a network-side device
- the terminal can be used to perform the steps of the method described in Figure 2 above
- the network-side device can be used to perform the method described in Figure 3 above or
- the terminal may be used to perform the steps of the method described in FIG. 3 above
- the network side device may be used to perform the steps of the method described in FIG. 2 above.
- the term “comprising”, “comprising” or any other variation thereof is intended to cover a non-exclusive inclusion such that a process, method, article or apparatus comprising a set of elements includes not only those elements, It also includes other elements not expressly listed, or elements inherent in the process, method, article, or device. Without further limitations, an element defined by the phrase “comprising a " does not preclude the presence of additional identical elements in the process, method, article, or apparatus comprising that element.
- the scope of the methods and devices in the embodiments of the present application is not limited to performing functions in the order shown or discussed, and may also include performing functions in a substantially simultaneous manner or in reverse order according to the functions involved. Functions are performed, for example, the described methods may be performed in an order different from that described, and various steps may also be added, omitted, or combined. Additionally, features described with reference to certain examples may be combined in other examples.
Landscapes
- Engineering & Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Computer Security & Cryptography (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Evolutionary Computation (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Information Transfer Between Computers (AREA)
- Compression, Expansion, Code Conversion, And Decoders (AREA)
- Mobile Radio Communication Systems (AREA)
Abstract
The present application belongs to the technical field of communications. Disclosed are an AI network information transmission method and apparatus, and a communication device. The AI network information transmission method in the embodiments of the present application comprises: a first terminal compressing AI network information, wherein the AI network information comprises at least one of a network structure and a network parameter; and the first terminal sending the compressed AI network information to a second terminal.
Description
相关申请的交叉引用Cross References to Related Applications
本申请主张在2021年12月31日在中国提交的中国专利申请No.202111666710.4的优先权,其全部内容通过引用包含于此。This application claims priority to Chinese Patent Application No. 202111666710.4 filed in China on December 31, 2021, the entire contents of which are hereby incorporated by reference.
本申请属于通信技术领域,具体涉及一种AI网络信息传输方法、装置及通信设备。The present application belongs to the technical field of communication, and in particular relates to an AI network information transmission method, device and communication equipment.
人工智能(Artificial Intelligence,AI)是研究和开发用于模拟、延伸、扩展人的智能的理论、方法、技术及应用系统的一门新的技术科学,受到人们的广泛关注,针对AI的应用也越来越广泛。目前,人们已经开始研究将AI网络应用在通信系统中,例如网络侧设备和终端之间可以通过AI网络来传输通信数据。在通信系统中,通常是将整个AI网络一起传递,造成系统开销较大。Artificial Intelligence (AI) is a new technical science that researches and develops theories, methods, technologies and application systems for simulating, extending and expanding human intelligence. more and more widely. At present, people have begun to study the application of AI networks in communication systems. For example, communication data can be transmitted between network-side devices and terminals through AI networks. In a communication system, the entire AI network is usually transmitted together, resulting in a large system overhead.
发明内容Contents of the invention
本申请实施例提供一种AI网络信息传输方法、装置及通信设备,能够解决相关技术中通信设备传输AI网络传输开销较大的问题。Embodiments of the present application provide an AI network information transmission method, device, and communication equipment, which can solve the problem of relatively large transmission overhead of communication equipment in AI network transmission in the related art.
第一方面,提供了一种AI网络信息传输方法,包括:In the first aspect, an AI network information transmission method is provided, including:
第一端对AI网络信息进行压缩,所述AI网络信息包括网络结构和网络参数中的至少一项;The first end compresses the AI network information, and the AI network information includes at least one of network structure and network parameters;
所述第一端向第二端发送压缩后的所述AI网络信息。The first end sends the compressed AI network information to the second end.
第二方面,提供了一种AI网络信息传输方法,包括:In the second aspect, an AI network information transmission method is provided, including:
第二端接收第一端发送的压缩后的AI网络信息,所述AI网络信息包括网络结构和网络参数中的至少一项。The second end receives the compressed AI network information sent by the first end, where the AI network information includes at least one of network structure and network parameters.
第三方面,提供了一种AI网络信息传输装置,包括:In a third aspect, an AI network information transmission device is provided, including:
压缩模块,用于对AI网络信息进行压缩,所述AI网络信息包括网络结构和网络参数中的至少一项;A compression module, configured to compress AI network information, where the AI network information includes at least one of network structure and network parameters;
发送模块,用于向第二端发送压缩后的所述AI网络信息。A sending module, configured to send the compressed AI network information to the second end.
第四方面,提供了一种AI网络信息传输装置,包括:In the fourth aspect, an AI network information transmission device is provided, including:
接收模块,用于接收第一端发送的压缩后的AI网络信息,所述AI网络信息包括网络结构和网络参数中的至少一项。The receiving module is configured to receive the compressed AI network information sent by the first end, where the AI network information includes at least one of network structure and network parameters.
第五方面,提供了一种通信设备,包括处理器和存储器,所述存储器存储可在所述处理器上运行的程序或指令,所述程序或指令被所述处理器执行时实现如第一方面所述的AI网络信息传输方法的步骤,或者实现如第二方面所述的AI网络信息传输方法的步骤。In a fifth aspect, a communication device is provided, including a processor and a memory, the memory stores programs or instructions that can run on the processor, and when the programs or instructions are executed by the processor, the first The steps of the AI network information transmission method described in the aspect, or the steps of implementing the AI network information transmission method described in the second aspect.
第六方面,提供了一种可读存储介质,所述可读存储介质上存储程序或指令,所述程序或指令被处理器执行时实现如第一方面所述的AI网络信息传输方法的步骤,或者实现如第二方面所述的AI网络信息传输方法的步骤。A sixth aspect provides a readable storage medium, on which a program or instruction is stored, and when the program or instruction is executed by a processor, the steps of the AI network information transmission method as described in the first aspect are implemented , or implement the steps of the AI network information transmission method as described in the second aspect.
第七方面,提供了一种芯片,所述芯片包括处理器和通信接口,所述通信接口和所述处理器耦合,所述处理器用于运行程序或指令,实现如第一方面所述的AI网络信息传输方法,或实现如第二方面所述的AI网络信息传输方法。In the seventh aspect, there is provided a chip, the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is used to run programs or instructions to implement the AI described in the first aspect A network information transmission method, or realize the AI network information transmission method as described in the second aspect.
第八方面,提供了一种计算机程序/程序产品,所述计算机程序/程序产品被存储在存储介质中,所述计算机程序/程序产品被至少一个处理器执行以实现如第一方面所述的AI网络信息传输方法的步骤,或者实现如第二方面所述的AI网络信息传输方法的步骤。In an eighth aspect, a computer program/program product is provided, the computer program/program product is stored in a storage medium, and the computer program/program product is executed by at least one processor to implement the method described in the first aspect The steps of the AI network information transmission method, or the steps of realizing the AI network information transmission method as described in the second aspect.
在本申请实施例中,第一端能够向第二端发送压缩的AI网络信息,所述AI网络信息包括网络结构和网络参数中的至少一项,进而在通信过程中也就无需将包括全部网络结构和网络参数的整个AI网络一起进行传输,使得AI网络的网络结构和网络参数可以分开发送,进而能够有效降低通信过程中的传输开销。In the embodiment of the present application, the first end can send compressed AI network information to the second end, the AI network information includes at least one of the network structure and network parameters, and there is no need to include all The network structure and network parameters of the entire AI network are transmitted together, so that the network structure and network parameters of the AI network can be sent separately, which can effectively reduce the transmission overhead in the communication process.
图1是本申请实施例可应用的一种无线通信系统的框图;FIG. 1 is a block diagram of a wireless communication system to which an embodiment of the present application is applicable;
图2是本申请实施例提供的一种AI网络信息传输方法的流程图;FIG. 2 is a flow chart of an AI network information transmission method provided by an embodiment of the present application;
图3是本申请实施例提供的另一种AI网络信息传输方法的流程图;Fig. 3 is a flow chart of another AI network information transmission method provided by the embodiment of the present application;
图4是本申请实施例提供的一种AI网络信息传输装置的结构图;FIG. 4 is a structural diagram of an AI network information transmission device provided by an embodiment of the present application;
图5是本申请实施例提供的另一种AI网络信息传输装置的结构图;Fig. 5 is a structural diagram of another AI network information transmission device provided by the embodiment of the present application;
图6是本申请实施例提供的一种通信设备的结构图;FIG. 6 is a structural diagram of a communication device provided by an embodiment of the present application;
图7是本申请实施例提供的一种终端的结构图;FIG. 7 is a structural diagram of a terminal provided in an embodiment of the present application;
图8是本申请实施例提供的一种网络侧设备的结构图;FIG. 8 is a structural diagram of a network-side device provided by an embodiment of the present application;
图9是本申请实施例提供的另一种网络侧设备的结构图。FIG. 9 is a structural diagram of another network-side device provided by an embodiment of the present application.
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚描述,显然,所描述的实施例是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员所获得的所有其他实施例,都属于本申请保护的范围。The technical solutions in the embodiments of the present application will be clearly described below in conjunction with the drawings in the embodiments of the present application. Obviously, the described embodiments are part of the embodiments of the present application, but not all of them. All other embodiments obtained by persons of ordinary skill in the art based on the embodiments in this application belong to the protection scope of this application.
本申请的说明书和权利要求书中的术语“第一”、“第二”等是用于区别类似的对象,而不用于描述特定的顺序或先后次序。应该理解这样使用的术语在适当情况下可以互换,以便本申请的实施例能够以除了在这里图示或描述的那些以外的顺序实施,且“第一”、“第二”所区别的对象通常为一类,并不限定对象的个数,例如第一对象可以是一个,也可以是多个。此外,说明书以及权利要求中“和/或”表示所连接对象的至少其中之一,字符“/”一般表示前后关联对象是一种“或”的关系。The terms "first", "second" and the like in the specification and claims of the present application are used to distinguish similar objects, and are not used to describe a specific sequence or sequence. It is to be understood that the terms so used are interchangeable under appropriate circumstances such that the embodiments of the application are capable of operation in sequences other than those illustrated or described herein and that "first" and "second" distinguish objects. It is usually one category, and the number of objects is not limited. For example, there may be one or more first objects. In addition, "and/or" in the description and claims means at least one of the connected objects, and the character "/" generally means that the related objects are an "or" relationship.
值得指出的是,本申请实施例所描述的技术不限于长期演进型(Long Term Evolution,LTE)/LTE的演进(LTE-Advanced,LTE-A)系统,还可用于其他无线通信系统,诸如码分多址(Code Division Multiple Access,CDMA)、时分多址(Time Division Multiple Access,TDMA)、频分多址(Frequency Division Multiple Access,FDMA)、正交频分多址(Orthogonal Frequency Division Multiple Access,OFDMA)、单载波频分多址(Single-carrier Frequency Division Multiple Access,SC-FDMA)和其他系统。本申请实施例中的术语“系统”和“网络”常被可互换地使用,所描述的技术既可用于以上提及的系统 和无线电技术,也可用于其他系统和无线电技术。以下描述出于示例目的描述了新空口(New Radio,NR)系统,并且在以下大部分描述中使用NR术语,但是这些技术也可应用于NR系统应用以外的应用,如第6代(6
th Generation,6G)通信系统。
It is worth pointing out that the technology described in the embodiment of this application is not limited to the Long Term Evolution (Long Term Evolution, LTE)/LTE-Advanced (LTE-Advanced, LTE-A) system, and can also be used in other wireless communication systems, such as code Code Division Multiple Access (CDMA), Time Division Multiple Access (TDMA), Frequency Division Multiple Access (FDMA), Orthogonal Frequency Division Multiple Access, OFDMA), Single-carrier Frequency Division Multiple Access (Single-carrier Frequency Division Multiple Access, SC-FDMA) and other systems. The terms "system" and "network" in the embodiments of the present application are often used interchangeably, and the described technologies can be used for the above-mentioned systems and radio technologies as well as other systems and radio technologies. The following description describes the New Radio (NR) system for illustrative purposes, and uses NR terminology in most of the following descriptions, but these techniques can also be applied to applications other than NR system applications, such as the 6th generation (6 th Generation, 6G) communication system.
图1示出本申请实施例可应用的一种无线通信系统的框图。无线通信系统包括终端11和网络侧设备12。其中,终端11可以是手机、平板电脑(Tablet Personal Computer)、膝上型电脑(Laptop Computer)或称为笔记本电脑、个人数字助理(Personal Digital Assistant,PDA)、掌上电脑、上网本、超级移动个人计算机(ultra-mobile personal computer,UMPC)、移动上网装置(Mobile Internet Device,MID)、增强现实(augmented reality,AR)/虚拟现实(virtual reality,VR)设备、机器人、可穿戴式设备(Wearable Device)、车载设备(Vehicle User Equipment,VUE)、行人终端(Pedestrian User Equipment,PUE)、智能家居(具有无线通信功能的家居设备,如冰箱、电视、洗衣机或者家具等)、游戏机、个人计算机(personal computer,PC)、柜员机或者自助机等终端侧设备,可穿戴式设备包括:智能手表、智能手环、智能耳机、智能眼镜、智能首饰(智能手镯、智能手链、智能戒指、智能项链、智能脚镯、智能脚链等)、智能腕带、智能服装等。需要说明的是,在本申请实施例并不限定终端11的具体类型。网络侧设备12可以包括接入网设备或核心网设备,其中,接入网设备也可以称为无线接入网设备、无线接入网(Radio Access Network,RAN)、无线接入网功能或无线接入网单元。接入网设备可以包括基站、无线局域网(Wireless Local Area Network,WLAN)接入点或WiFi节点等,基站可被称为节点B、演进节点B(Evolved Node B,eNB)、接入点、基收发机站(Base Transceiver Station,BTS)、无线电基站、无线电收发机、基本服务集(Basic Service Set,BSS)、扩展服务集(Extended Service Set,ESS)、家用B节点、家用演进型B节点、发送接收点(Transmitting Receiving Point,TRP)或所述领域中其他某个合适的术语,只要达到相同的技术效果,所述基站不限于特定技术词汇,需要说明的是,在本申请实施例中仅以NR系统中的基站为例进行介绍,并不限定基站的具体类型。核心网设备可以包含但不限于如下至少一项:核心网节点、核心网功能、移动管理实体(Mobility Management Entity,MME)、接入移动管理功能(Access and Mobility Management Function,AMF)、会话管理功能(Session Management Function,SMF)、用户平面功能(User Plane Function,UPF)、策略控制功能(Policy Control Function,PCF)、策略与计费规则功能单元(Policy and Charging Rules Function,PCRF)、边缘应用服务发现功能(Edge Application Server Discovery Function,EASDF)、统一数据管理(Unified Data Management,UDM),统一数据仓储(Unified Data Repository,UDR)、归属用户服务器(Home Subscriber Server,HSS)、集中式网络配置(Centralized network configuration,CNC)、网络存储功能(Network Repository Function,NRF),网络开放功能(Network Exposure Function,NEF)、本地NEF(Local NEF,或L-NEF)、绑定支持功能(Binding Support Function,BSF)、应用功能(Application Function,AF)等。需要说明的是,在本申请实施例中仅以NR系统中的核心网设备为例进行介绍,并不限定核心网设备的具体类型。Fig. 1 shows a block diagram of a wireless communication system to which the embodiment of the present application is applicable. The wireless communication system includes a terminal 11 and a network side device 12 . Wherein, the terminal 11 can be a mobile phone, a tablet computer (Tablet Personal Computer), a laptop computer (Laptop Computer) or a notebook computer, a personal digital assistant (Personal Digital Assistant, PDA), a palmtop computer, a netbook, a super mobile personal computer (ultra-mobile personal computer, UMPC), mobile Internet device (Mobile Internet Device, MID), augmented reality (augmented reality, AR) / virtual reality (virtual reality, VR) equipment, robot, wearable device (Wearable Device) , Vehicle User Equipment (VUE), Pedestrian User Equipment (PUE), smart home (home equipment with wireless communication functions, such as refrigerators, TVs, washing machines or furniture, etc.), game consoles, personal computers (personal computer, PC), teller machine or self-service machine and other terminal side devices, wearable devices include: smart watches, smart bracelets, smart headphones, smart glasses, smart jewelry (smart bracelets, smart bracelets, smart rings, smart necklaces, smart feet bracelets, smart anklets, etc.), smart wristbands, smart clothing, etc. It should be noted that, the embodiment of the present application does not limit the specific type of the terminal 11 . The network side device 12 may include an access network device or a core network device, where the access network device may also be called a radio access network device, a radio access network (Radio Access Network, RAN), a radio access network function, or a wireless network. access network unit. The access network equipment may include a base station, a wireless local area network (Wireless Local Area Network, WLAN) access point, or a WiFi node, etc., and the base station may be called a node B, an evolved node B (Evolved Node B, eNB), an access point, or a base station. Base Transceiver Station (BTS), radio base station, radio transceiver, Basic Service Set (BSS), Extended Service Set (ESS), Home Node B, Home Evolved Node B, Transmitting Receiving Point (Transmitting Receiving Point, TRP) or some other suitable term in the field, as long as the same technical effect is achieved, the base station is not limited to specific technical terms. It should be noted that, in the embodiment of this application, only The base station in the NR system is taken as an example for introduction, and the specific type of the base station is not limited. The core network equipment may include but not limited to at least one of the following: core network node, core network function, mobility management entity (Mobility Management Entity, MME), access mobility management function (Access and Mobility Management Function, AMF), session management function (Session Management Function, SMF), User Plane Function (UPF), Policy Control Function (Policy Control Function, PCF), Policy and Charging Rules Function (PCRF), edge application service Discovery function (Edge Application Server Discovery Function, EASDF), unified data management (Unified Data Management, UDM), unified data storage (Unified Data Repository, UDR), home subscriber server (Home Subscriber Server, HSS), centralized network configuration ( Centralized network configuration, CNC), network storage function (Network Repository Function, NRF), network exposure function (Network Exposure Function, NEF), local NEF (Local NEF, or L-NEF), binding support function (Binding Support Function, BSF), application function (Application Function, AF), etc. It should be noted that, in the embodiment of the present application, only the core network equipment in the NR system is used as an example for introduction, and the specific type of the core network equipment is not limited.
下面结合附图,通过一些实施例及其应用场景对本申请实施例提供的AI网络信息传输方法进行详细地说明。The following describes in detail the AI network information transmission method provided by the embodiment of the present application through some embodiments and application scenarios with reference to the accompanying drawings.
请参照图2,图2是本申请实施例提供的一种AI网络信息传输方法的流程图,如图2所示,所述方法包括以下步骤:Please refer to Figure 2, Figure 2 is a flow chart of an AI network information transmission method provided in the embodiment of the present application, as shown in Figure 2, the method includes the following steps:
步骤201、第一端对AI网络信息进行压缩,所述AI网络信息包括网络结构和网络参数中的至少一项; Step 201, the first end compresses the AI network information, and the AI network information includes at least one of network structure and network parameters;
步骤202、所述第一端向第二端发送压缩后的所述AI网络信息。 Step 202, the first end sends the compressed AI network information to the second end.
本申请实施例中,所述第一端和所述第二端为具有发送和接收功能的通信设备。In the embodiment of the present application, the first end and the second end are communication devices with sending and receiving functions.
可选地,所述第一端为网络侧设备和终端中的一者,所述第二端为网络侧设备和终端中的另一者;或者,所述第一端和所述第二端为终端的不同节点;或者,所述第一端和所述第二端为网络侧设备的不同节点。Optionally, the first end is one of the network-side device and the terminal, and the second end is the other of the network-side device and the terminal; or, the first end and the second end are different nodes of a terminal; or, the first end and the second end are different nodes of a network side device.
需要说明地,所述网络侧设备可以包括接入网设备(例如:基站)和核心网设备。可选地,所述第一端可以为接入网设备,第二端为核心网设备;或者,第一端为终端,第二端为核心网设备或接入网设备;或者,所述第一端和所述第二端为接入网设备的不同节点;或者,所述第一端和所述第二端 为核心网设备的不同节点,等等,本申请实施例不做一一列举。It should be noted that the network side equipment may include access network equipment (for example: base station) and core network equipment. Optionally, the first end may be an access network device, and the second end may be a core network device; or, the first end may be a terminal, and the second end may be a core network device or an access network device; or, the second end may be a core network device or an access network device; The one end and the second end are different nodes of the access network equipment; or, the first end and the second end are different nodes of the core network equipment, etc., and the embodiments of the present application do not list them one by one .
本申请实施例中,所述AI网络信息包括网络结构和网络参数中的至少一项。可选地,所述AI网络信息可以是某一个AI网络的网络结构和/或网络参数,或者是多个AI网络的网络结构和/或网络参数。在一些实施例中,AI网络也可以称为AI神经网络、AI模型等。其中,所述网络参数包括AI网络的权重参数、超参数等。In this embodiment of the present application, the AI network information includes at least one of network structure and network parameters. Optionally, the AI network information may be the network structure and/or network parameters of a certain AI network, or the network structures and/or network parameters of multiple AI networks. In some embodiments, an AI network may also be called an AI neural network, an AI model, or the like. Wherein, the network parameters include weight parameters, hyperparameters and the like of the AI network.
需要说明地,所述对AI网络信息进行压缩,可以是指按照预设的模型表述方式将所述AI网络信息压缩到预设的模型表述方式对应的文件中,所谓模型表述方式为一种数据结构,按照一定的规则描述AI网络结构、网络参数等信息。It should be noted that the compressing the AI network information may refer to compressing the AI network information into a file corresponding to the preset model expression method according to the preset model expression method. The so-called model expression method is a kind of data Structure, which describes the AI network structure, network parameters and other information according to certain rules.
可选地,所述AI网络信息包括网络结构和/或网络参数,第一端对AI网络信息的压缩也就包括对网络结构和/或权重参数进行压缩,并将压缩后的网络结构和/或权重参数发送给第二端。例如,AI网络信息仅包括网络结构,则第一端仅对网络结构进行压缩并发送;或者,AI网络信息也可以是仅包括网络参数,则第一端仅对网络参数进行压缩并发送;或者,AI网络信息包括部分网络结构和部分网络参数,则第一端对该部分网络结构和部分网络参数进行压缩并发送。当然,所述AI网络信息包括的具体信息内容还可以是其他的情况,此处不做过多赘述。Optionally, the AI network information includes network structure and/or network parameters, and the compression of the AI network information by the first end also includes compressing the network structure and/or weight parameters, and compressing the compressed network structure and/or or the weight parameter is sent to the second end. For example, if the AI network information only includes the network structure, the first end only compresses and sends the network structure; or, the AI network information may only include network parameters, then the first end only compresses and sends the network parameters; or , the AI network information includes part of the network structure and part of the network parameters, and the first end compresses and sends the part of the network structure and part of the network parameters. Of course, the specific information content included in the AI network information may also be in other situations, which will not be described in detail here.
本申请实施例中,AI网络信息包括网络结构和网络参数中的至少一项,进而在通信过程中也就无需将包括全部网络结构和网络参数的整个AI网络一起进行传输,使得AI网络的网络结构和网络参数可以分开发送,进而能够有效降低通信过程中的传输开销。In the embodiment of the present application, the AI network information includes at least one of the network structure and network parameters, and there is no need to transmit the entire AI network including all network structures and network parameters together during the communication process, so that the network of the AI network The structure and network parameters can be sent separately, which can effectively reduce the transmission overhead in the communication process.
可选地,所述AI网络信息可以包括网络结构和网络参数,所述步骤201可以包括如下任意一项:Optionally, the AI network information may include network structure and network parameters, and the step 201 may include any of the following:
所述第一端对所述网络结构和所述网络参数进行合并压缩;The first end combines and compresses the network structure and the network parameters;
所述第一端对所述网络结构和所述网络参数分别进行压缩。The first end compresses the network structure and the network parameters respectively.
例如,所述第一端可以是基于预设的模型表述方式将AI网络的网络结构和网络参数合并压缩成一个传输文件,或者也可以是基于预设的模型表述方式将所述网络结构和权重参数分别压缩成两个独立的传输文件。For example, the first end can combine and compress the network structure and network parameters of the AI network into a transmission file based on a preset model expression, or can also compress the network structure and weights based on a preset model expression. The parameters are compressed separately into two separate transfer files.
可选地,在所述第一端对所述网络结构和所述网络参数分别进行压缩的情况下,所述步骤202可以包括如下任意一项:Optionally, in the case where the first end compresses the network structure and the network parameters respectively, the step 202 may include any of the following:
所述第一端向第二端合并发送压缩后的网络结构和压缩后的网络参数;The first end sends the compressed network structure and compressed network parameters to the second end in combination;
所述第一端向第二端分别发送压缩后的网络结构和压缩后的网络参数。The first end sends the compressed network structure and the compressed network parameters to the second end respectively.
示例性地,在AI网络信息包括网络结构和网络参数的情况下,第一端对所述网络结构和网络参数分别进行压缩,例如压缩成两个独立的传输文件,并将压缩后的两个传输文件一起发送给第二端,或者也可以是分开发送所述压缩后的两个传输文件,或者也可以是只发送其中一个压缩后的传输文件,使得第一端对于压缩后的AI网络信息的传输方式更为灵活。Exemplarily, in the case that the AI network information includes network structure and network parameters, the first end compresses the network structure and network parameters respectively, such as compressing into two independent transmission files, and compresses the two compressed The transmission files are sent to the second end together, or the two compressed transmission files can be sent separately, or only one of the compressed transmission files can be sent, so that the first end can receive the compressed AI network information The transmission method is more flexible.
可选地,所述步骤201还可以包括如下任意一项:Optionally, the step 201 may also include any of the following:
所述第一端基于预设的模型表述方式将AI网络信息转换成对应的传输文件,对所述传输文件进行压缩;The first end converts the AI network information into a corresponding transmission file based on a preset model representation, and compresses the transmission file;
所述第一端基于预设的数据格式对所述AI网络信息进行压缩;The first end compresses the AI network information based on a preset data format;
所述第一端获取待发送的AI网络信息和所述第二端已有的AI网络信息,对所述待发送的AI网络信息与所述第二端已有的AI网络信息之间的AI网络信息差值进行压缩;The first end obtains the AI network information to be sent and the existing AI network information of the second end, and calculates the AI network information between the AI network information to be sent and the existing AI network information of the second end. Network information difference is compressed;
所述第一端获取待发送的AI网络信息和预设AI网络的AI网络信息,对所述待发送的AI网络信息与所述预设AI网络的AI网络信息之间的AI网络信息差值进行压缩。The first end obtains the AI network information to be sent and the AI network information of the preset AI network, and calculates the AI network information difference between the AI network information to be sent and the AI network information of the preset AI network to compress.
其中,所述预设的模型表述方式可以是第一端和第二端都通用的AI网络表述方式,例如开放式神经网络交互(open neural network exchange,ONNX)、TensorFlow等。其中,TensorFlow是一种机器学习框架。示例性地,第一端可以是基于ONNX将AI网络信息转换成对应的传输文件,然后对该传输文件进行压缩后发送给第二端,第二端对压缩的传输文件进行解压后,同样能够基于ONNX将解压后的传输文件转换成自身适用的网络结构和网络参数。Wherein, the preset model expression method may be an AI network expression method common to both the first end and the second end, such as open neural network exchange (ONNX), TensorFlow, and the like. Among them, TensorFlow is a machine learning framework. Exemplarily, the first end can convert the AI network information into a corresponding transmission file based on ONNX, and then compress the transmission file and send it to the second end. After the second end decompresses the compressed transmission file, it can also Based on ONNX, the decompressed transmission file is converted into its own applicable network structure and network parameters.
需要说明地,两个不同的神经网络框架下的AI网络信息保存的文件结构不同,无法直接读取,通过预设的模型表述方式将AI网络信息转换成对应的传输文件后再进行压缩传输,也就使得使用不同的神经网络框架的两个通信设备能够实现AI网络信息的读取以及应用。It should be noted that the file structure of the AI network information saved under the two different neural network frameworks is different and cannot be read directly. The AI network information is converted into the corresponding transmission file through the preset model expression method and then compressed and transmitted. It also enables two communication devices using different neural network frameworks to realize the reading and application of AI network information.
可选地,所述第一端也可以是基于预设的数据格式对所述AI网络信息进行压缩,例如ONNX使用的protobuf数据格式等,其中,protobuf是一种数据交换格式。Optionally, the first end may also compress the AI network information based on a preset data format, such as the protobuf data format used by ONNX, where protobuf is a data exchange format.
或者,所述第一端也可以是对待发送的AI网络信息与第二端已有的AI网络信息之间的AI网络信息差值进行压缩,第一端只需要发送压缩后的AI网络信息差值给第二端,进而对于待发送的AI网络信息与第二端已有的AI网络信息之间相同的AI网络信息也就无需进行压缩和发送,这样也就能够有效节省第一端的传输开销。Alternatively, the first end may also compress the AI network information difference between the AI network information to be sent and the existing AI network information at the second end, and the first end only needs to send the compressed AI network information difference value to the second end, and then there is no need to compress and send the same AI network information between the AI network information to be sent and the existing AI network information of the second end, which can effectively save the transmission of the first end overhead.
又或者,所述第一端可以是对待发送的AI网络信息与预设AI网络的AI网络信息之间的AI网络信息差值进行压缩,第一端只需要发送压缩后的AI网络信息差值给第二端,以节省第一端的传输开销。其中,所述预设AI网络可以是协议预定或高层配置,例如某些固定的AI网络模板,或者也可以是通信设备通用的AI网络模板。所述预设AI网络包括网络结构的初始值和网络参数的初始值。Alternatively, the first end may compress the AI network information difference between the AI network information to be sent and the AI network information of the preset AI network, and the first end only needs to send the compressed AI network information difference to the second end to save the transmission overhead of the first end. Wherein, the preset AI network may be a predetermined protocol or a high-level configuration, such as some fixed AI network templates, or may also be a common AI network template for communication devices. The preset AI network includes initial values of network structure and network parameters.
可选地,所述AI网络信息差值包括如下至少一项:Optionally, the AI network information difference includes at least one of the following:
指定的网络参数;specified network parameters;
网络参数的索引;index of network parameters;
修改的网络参数;Modified network parameters;
修改的网络参数中修改的参数值;The modified parameter value in the modified network parameter;
修改的网络参数中修改的参数值的位置;The position of the modified parameter value in the modified network parameter;
修改的网络参数中修改的参考值的位置,所述参考值为所述网络参数中的最大值;the location of the modified reference value in the modified network parameters, the reference value being the maximum value in the network parameters;
修改的网络参数中的非零值;A non-zero value in the modified network parameter;
修改的网络参数中的非零值的位置;the location of non-zero values in the modified network parameters;
新增的网络结构;Newly added network structure;
删除的网络结构;Deleted network structures;
修改的网络结构。Modified network structure.
本申请实施例中,所述预设的模型表述方式包括如下任意一项:协议约定的模型表述方式、自定义的模型表述方式。其中,所述自定义的模型表述 方式可以是指通信设备自定义的一种数据结构,用于表述AI网络的网络结构和网络参数。In the embodiment of the present application, the preset model expression method includes any one of the following: a model expression method stipulated in an agreement, and a user-defined model expression method. Wherein, the self-defined model expression method may refer to a data structure defined by the communication device, which is used to describe the network structure and network parameters of the AI network.
可选地,所述自定义的模型表述方式的内容包括如下至少一项:AI网络的网络结构、AI网络的网络参数的属性、AI网络的网络参数的数值。其中,所述网络参数的属性包括网络参数的名称、标识、维度等信息;所述AI网络的参数的数值可以是一个或多个。Optionally, the content of the self-defined model expression includes at least one of the following: the network structure of the AI network, the attributes of the network parameters of the AI network, and the values of the network parameters of the AI network. Wherein, the attribute of the network parameter includes information such as the name, identification, and dimension of the network parameter; the value of the parameter of the AI network may be one or more.
可选地,所述预设的模型表述方式中网络结构的表述方式包括如下至少一项:Optionally, the representation of the network structure in the preset model representation includes at least one of the following:
AI网络的网络结构之间的关联关系;The relationship between the network structures of the AI network;
AI网络的网络参数的属性;Attributes of the network parameters of the AI network;
AI网络的网络参数中非零值的位置;The location of non-zero values in the network parameters of the AI network;
AI网络的网络参数中的更新数值位置。The updated numerical position in the network parameters of the AI network.
需要说明地,所述网络结构之间的关联关系可以是指各个网络结构(或者也称节点)的输入输出之间的连接关系,例如第一个节点的输出连接第二个节点的输入,第二个节点的输出连接第三个节点的输入,等等。所述网络参数的属性包括网络参数的名称、标识、维度等信息。It should be noted that the association relationship between the network structures may refer to the connection relationship between the input and output of each network structure (or also called nodes), for example, the output of the first node is connected to the input of the second node, and the output of the first node is connected to the input of the second node. The output of the second node is connected to the input of the third node, and so on. The attribute of the network parameter includes information such as a name, an identifier, and a dimension of the network parameter.
本申请实施例中,所述第一端基于预设的模型表述方式将AI网络信息转换成对应的传输文件,对所述传输文件进行压缩,包括:In the embodiment of the present application, the first end converts the AI network information into a corresponding transmission file based on a preset model expression method, and compresses the transmission file, including:
所述第一端基于至少一个预设的模型表述方式将AI网络信息转换成至少一个传输文件,一个所述预设的模型表述方式对应至少一个传输文件;The first end converts the AI network information into at least one transmission file based on at least one preset model representation, and one preset model representation corresponds to at least one transmission file;
所述第一端对所述至少一个传输文件进行合并压缩,或者,所述第一端对所述至少一个传输文件分别进行压缩后再合并。The first end merges and compresses the at least one transmission file, or the first end compresses the at least one transmission file separately and then merges them.
需要说明地,所述预设的模型表述方式可以是有多个,一个预设的模型表述方式可以是对应至少一个传输文件,例如基于一个预设的模型表述方式将网络结构和网络参数分别转换成对应的传输文件。当然,也可以是基于预设的模型表述方式将网络结构和网络参数转换成一个传输文件。It should be noted that there may be multiple preset model expressions, and one preset model expression may correspond to at least one transmission file, for example, the network structure and network parameters are respectively converted based on a preset model expression into corresponding transfer files. Of course, it is also possible to convert the network structure and network parameters into a transmission file based on a preset model expression manner.
第一端在将AI网络信息基于至少一个预设的模型表述方式转换成至少一个传输文件后,将所述至少一个传输文件合并在一起进行压缩,并发送压缩后的传输文件,或者是将所述至少一个传输文件分别进行独立压缩,压缩 之后再合并到一起进行发送,或者也可以是分别发送每一个压缩后的传输文件。After the first end converts the AI network information into at least one transmission file based on at least one preset model expression, merge the at least one transmission file together for compression, and send the compressed transmission file, or The at least one transmission file mentioned above is compressed independently, and then combined and sent together after compression, or each compressed transmission file can also be sent separately.
可选地,所述AI网络信息包括网络结构和网络参数,所述第一端基于预设的模型表述方式将AI网络信息转换成对应的传输文件,对所述传输文件进行压缩,包括:Optionally, the AI network information includes network structure and network parameters, and the first end converts the AI network information into a corresponding transmission file based on a preset model expression, and compresses the transmission file, including:
所述第一端基于预设的模型表述方式将所述网络结构转换成第一传输文件,并基于所述预设的模型表述方式将所述网络参数转换成第二传输文件,将所述第一传输文件和所述第二传输文件分别进行压缩。The first end converts the network structure into a first transmission file based on a preset model representation, converts the network parameters into a second transmission file based on the preset model representation, and converts the first The first transmission file and the second transmission file are respectively compressed.
也就是说,第一端能够基于预设的模型表述方式分别将网络结构和网络参数转换成对应的传输文件,进而得到两个传输文件,并对这两个传输文件分别进行压缩,并能够分别发送压缩后的这两个传输文件,或者也可以是合并在一起发送。That is to say, the first end can convert the network structure and network parameters into corresponding transmission files based on the preset model expression, and then obtain two transmission files, and compress the two transmission files respectively, and can separately The two compressed transmission files may be sent, or they may be combined and sent together.
例如,所述第一端为基站,第二端为终端,终端在初始接入的时候,基站可以是将训练好的AI网络的网络结构基于预设的模型表述方式保存成对应格式的传输文件进行压缩,并发送给终端;如果需要网络参数,基站再将网络参数基于所述预设的模型表述方式保存成传输文件进行压缩,并发送给终端。可选地,网络结构对应的传输文件与网络参数对应的传输文件可以是不同的文件类型。另外,基站可以是通过数据信道发送压缩后的传输文件。For example, the first end is a base station, and the second end is a terminal. When the terminal initially accesses, the base station may save the network structure of the trained AI network into a transmission file in a corresponding format based on a preset model expression Compress and send to the terminal; if network parameters are required, the base station saves the network parameters based on the preset model expression method as a transmission file for compression and sends to the terminal. Optionally, the transmission file corresponding to the network structure and the transmission file corresponding to the network parameters may be of different file types. In addition, the base station may send the compressed transmission file through the data channel.
可选地,在所述AI网络信息包括网络参数的情况下,压缩后的AI网络信息也就包括压缩后的网络参数,所述第一端向第二端发送压缩后的所述AI网络信息,包括:Optionally, when the AI network information includes network parameters, the compressed AI network information also includes compressed network parameters, and the first end sends the compressed AI network information to the second end ,include:
所述第一端基于所述压缩后的网络参数的优先级顺序,按照所述优先级顺序向第二端发送所述压缩后的网络参数。The first end sends the compressed network parameters to the second end according to the priority order of the compressed network parameters based on the priority order of the compressed network parameters.
例如,同样以所述第一端为基站,第二端为终端为例,基站在发送压缩后的网络参数的时候,可以是将压缩后的网络参数按照预设的优先级顺序分成N组,优先级高的组先下发,优先级低的组后下发或者在传输资源有限的情况下不下发。For example, also taking the first end as a base station and the second end as a terminal as an example, when the base station sends compressed network parameters, it may divide the compressed network parameters into N groups according to a preset priority order, Groups with high priority are delivered first, and groups with low priority are delivered later or not delivered when transmission resources are limited.
可选地,所述第一端基于所述压缩后的网络参数的优先级顺序,按照所述优先级顺序向第二端发送所述压缩后的网络参数,包括:Optionally, the first end sends the compressed network parameters to the second end according to the priority order based on the priority order of the compressed network parameters, including:
所述第一端基于所述压缩后的网络参数的优先级顺序,对所述压缩后的网络参数进行分组;grouping the compressed network parameters by the first end based on the priority order of the compressed network parameters;
在一些特定场景下,例如传输资源小于预设阈值的情况等,所述第一端按照预设顺序对分组后的网络参数进行丢弃并发送剩余的网络参数,所述预设顺序为分组后的网络参数的优先级从低至高的顺序。In some specific scenarios, such as when the transmission resource is less than the preset threshold, the first end discards the grouped network parameters and sends the remaining network parameters in a preset order, the preset order being the grouped Network parameters are prioritized in order from low to high.
例如,所述第一端为基站,第二端为终端,基站在发送压缩后的网络参数的时候,可以是将压缩后的网络参数按照预设的优先级顺序分成N组,在基站的传输资源小于预设阈值的情况下,也即基站的传输资源有限的时候,或者有突发的高优先级业务的时候,基站按照优先级从低至高的顺序对网络参数进行丢弃,也即优先级最低的组的网络参数最先被丢弃,直至基站的传输资源足够发送剩余组的网络参数,也就能够确保基站发送的网络参数为优先级较高的网络参数。For example, the first end is a base station, and the second end is a terminal. When the base station transmits the compressed network parameters, it may divide the compressed network parameters into N groups according to a preset priority order, and transmit the compressed network parameters in the base station. When the resource is less than the preset threshold, that is, when the transmission resources of the base station are limited, or when there is a burst of high-priority traffic, the base station discards the network parameters in order of priority from low to high, that is, the priority The network parameters of the lowest group are discarded first until the transmission resources of the base station are sufficient to send the network parameters of the remaining groups, which ensures that the network parameters sent by the base station are network parameters with higher priority.
需要说明地,终端在接收到基站发送的网络结构和网络参数后,对于没有接收到的网络参数可以是使用协议约定的默认值或是0。It should be noted that, after receiving the network structure and network parameters sent by the base station, the terminal may use a default value agreed in the protocol or be 0 for the network parameters not received.
本申请实施例中,在所述第一端对AI网络信息进行压缩之前,所述方法还包括:In the embodiment of the present application, before the first end compresses the AI network information, the method further includes:
所述第一端接收所述第二端发送的第一请求信息,所述第一请求信息用于请求获取目标AI网络信息;The first end receives first request information sent by the second end, and the first request information is used to request acquisition of target AI network information;
这种情况下,所述第一端对AI网络信息进行压缩,包括:In this case, the first end compresses the AI network information, including:
所述第一端对所述目标AI网络信息进行压缩。The first end compresses the target AI network information.
也就是说,第二端能够基于第一请求信息来获取指定的目标AI网络信息,进而第一端基于所述第一请求信息对目标AI网络信息进行压缩后发送给第二端。That is to say, the second end can obtain the specified target AI network information based on the first request information, and then the first end compresses the target AI network information based on the first request information and sends it to the second end.
可选地,所述第一端对所述目标AI网络信息进行压缩之前,所述方法还包括:Optionally, before the first end compresses the target AI network information, the method further includes:
所述第一端判断是否需要对所述目标AI网络信息进行更新;The first end judges whether the target AI network information needs to be updated;
在判定需要对所述目标AI网络信息进行更新的情况下,更新所述目标AI网络信息;When it is determined that the target AI network information needs to be updated, update the target AI network information;
这种情况下,所述第一端对所述目标AI网络信息进行压缩,包括:In this case, the first end compresses the target AI network information, including:
所述第一端对更新后的所述目标AI网络信息进行压缩。The first end compresses the updated target AI network information.
需要说明地,第一端在接收到第二端发送的第一更新请求后,基于所述第一更新请求也即能够确定第二端想要获取的目标AI网络信息是哪些,在第一端对所述目标AI网络信息进行压缩之前,第一端可以判断是否需要对所述目标AI网络信息进行更新,如果需要更新,则更新所述目标AI网络信息,并对更新后的所述目标AI网络信息进行压缩后发送给第二端;若不需要对目标AI网络信息进行更新,则第一端可以是直接将所述目标AI网络信息进行压缩,然后发给第二端。It should be noted that after receiving the first update request sent by the second end, the first end can determine which target AI network information the second end wants to obtain based on the first update request. Before compressing the target AI network information, the first end may determine whether the target AI network information needs to be updated, if necessary, update the target AI network information, and update the target AI network information The network information is compressed and then sent to the second end; if the target AI network information does not need to be updated, the first end may directly compress the target AI network information and then send it to the second end.
例如,第二端向第一端发送权重参数更新请求(Request of Weight Updating),指定需要获取某个或某些特定的权重参数,第一端基于该权重参数更新请求,确定需要更新的情况下,对指定的权重参数进行更新,并基于预设的模型表述方式对更新的权重参数进行压缩后发送给第二端。For example, the second end sends a weight parameter update request (Request of Weight Updating) to the first end, specifying that one or some specific weight parameters need to be obtained, and the first end determines that an update is required based on the weight parameter update request , to update the specified weight parameters, and send the updated weight parameters to the second end after being compressed based on the preset model representation.
可选地,所述第一请求信息包括如下至少一项:Optionally, the first request information includes at least one of the following:
请求的网络参数的名称;the name of the requested network parameter;
请求的网络参数的标识;Identification of the requested network parameters;
网络结构更新请求;network structure update request;
网络参数更新请求;Network parameter update request;
AI网络的网络效果度量值。Network Effect Measures for AI Networks.
其中,所述网络效果度量值可以是指第二端按照某种约定的方法计算AI网络的效果,上报这个计算值,第一端基于第一请求信息中携带的这个计算值判读是否需要进行AI网络信息的更新。例如,针对信道状态信息(Channel State Information,CSI)预测,终端(第二端)计算预测的结果和某次实测的结果的相关性并上报给基站(第一端),该相关性也即网络效果度量值,基站根据终端上报的相关性和自己的AI网络计算的相关性比较,判断是否需要更新AI网络信息,如果相关性都较差,说明是信道变化过快,不用更新,因为基站侧的AI网络也性能不好。如果基站侧的AI网络计算的相关性明显优于终端上报的相关性,说明终端的AI网络的参数已经过时,无法匹配当前信道,基站需要重新发送AI网络参数。Wherein, the network effect measurement value may mean that the second end calculates the effect of the AI network according to a certain agreed method, and reports the calculated value, and the first end judges whether AI is required based on the calculated value carried in the first request information. Updates to network information. For example, for channel state information (Channel State Information, CSI) prediction, the terminal (second end) calculates the correlation between the predicted result and the actual measurement result and reports it to the base station (first end). Effect measurement value. The base station judges whether to update the AI network information according to the correlation reported by the terminal and the correlation calculated by its own AI network. If the correlation is poor, it means that the channel changes too fast and does not need to be updated, because the base station side The AI network also performed poorly. If the correlation calculated by the AI network on the base station side is significantly better than the correlation reported by the terminal, it means that the parameters of the AI network of the terminal are outdated and cannot match the current channel, and the base station needs to resend the AI network parameters.
可选地,所述目标AI网络信息包括第一目标网络参数,所述第一端对所 述目标AI网络信息进行压缩,包括:Optionally, the target AI network information includes first target network parameters, and the first end compresses the target AI network information, including:
所述第一端基于预设的模型表述方式将所述第一目标网络参数的属性和参数值转换成预设格式后进行压缩;The first end converts the attributes and parameter values of the first target network parameters into a preset format based on a preset model representation and then compresses them;
其中,所述第一目标网络参数的属性包括名称、维度、长度中的至少一项。Wherein, the attribute of the first target network parameter includes at least one of name, dimension, and length.
示例性地,以所述第一端为基站,第二端为终端,终端在初始接入的时候,可以在预设AI网络中选择一种AI网络,将该AI网络的标识发送给终端,还可以将该AI网络的网络参数的属性和参数值转换成预设格式后进行压缩,例如按照预设的模型表述方式进行压缩,然后发送给终端。Exemplarily, taking the first end as the base station and the second end as the terminal, when the terminal initially accesses, it may select an AI network from the preset AI networks, and send the identifier of the AI network to the terminal, The attributes and parameter values of the network parameters of the AI network can also be converted into a preset format and then compressed, for example, compressed according to a preset model expression, and then sent to the terminal.
可选地,在所述第一端为网络侧设备,所述第二端为终端,所述AI网络信息包括网络参数,所述第二端从第一小区切换到第二小区的情况下,所述第一端对AI网络信息进行压缩之前,所述方法还包括:Optionally, when the first end is a network side device, the second end is a terminal, the AI network information includes network parameters, and the second end is handed over from the first cell to the second cell, Before the first end compresses the AI network information, the method further includes:
所述第一端计算所述第一小区的网络参数与所述第二小区的网络参数的相关性,获取第二目标网络参数,所述第二目标网络参数包括如下至少一项:相关性小于预设阈值的网络参数、预设序列中的前N个网络参数,所述预设序列为网络参数的相关性按照从小到大顺序排列的序列;The first end calculates the correlation between the network parameters of the first cell and the network parameters of the second cell, and acquires a second target network parameter, where the second target network parameter includes at least one of the following items: the correlation is less than Network parameters with preset thresholds and the first N network parameters in a preset sequence, where the preset sequence is a sequence in which the correlation of network parameters is arranged in ascending order;
这种情况下,所述第一端对AI网络信息进行压缩,包括:In this case, the first end compresses the AI network information, including:
所述第一端对所述第二目标网络参数进行压缩;The first end compresses the second target network parameters;
所述第一端向第二端发送压缩后的所述AI网络信息,包括:The first end sends the compressed AI network information to the second end, including:
所述第一端向第二端发送压缩后的所述第二目标网络参数。The first end sends the compressed second target network parameters to the second end.
示例性地,终端在切换小区的情况下,网络侧设备(例如基站)可以是将某个AI网络完整的网络参数发送给终端,或者也可以是将部分网络参数发送给终端。例如,终端从第一小区切换至第二小区,第二小区可以是从第一小区获取第一小区对应使用的网络参数,计算每个网络参数的相关性,网络侧设备将相关性小于预设阈值的网络参数或者是相关性较小的前N个网络参数进行压缩后发送给终端,终端可以只更新接收到的网络参数。Exemplarily, when the terminal switches cells, the network side device (such as a base station) may send complete network parameters of an AI network to the terminal, or may also send partial network parameters to the terminal. For example, when a terminal switches from a first cell to a second cell, the second cell may obtain the network parameters corresponding to the first cell from the first cell, calculate the correlation of each network parameter, and the network side device will set the correlation to be less than the preset The threshold network parameters or the first N network parameters with less correlation are compressed and sent to the terminal, and the terminal may only update the received network parameters.
本申请实施例中,针对网络参数的更新和传递,第一端和第二端可以提前交互需要传递的网络参数的属性,包括名称、维度、长度等,然后将需要传递的网络参数通过预设的模型表述方式转换成对应的文件后进行压缩再传 递。其中,当网络结构已知的时候,所有的网络参数的顺序可以认为是已知的,通过位图(bitmap)或者组合数等方式可以交互需要更新的网络参数在整体列表中的位置。In this embodiment of the application, for the update and transfer of network parameters, the first end and the second end can interact in advance with the attributes of the network parameters to be transferred, including name, dimension, length, etc., and then pass the network parameters to be transferred through the preset The model expression is converted into the corresponding file and then compressed and delivered. Wherein, when the network structure is known, the order of all network parameters can be considered known, and the position of the network parameters that need to be updated in the overall list can be exchanged through bitmaps or combination numbers.
可选地,在基于预设的模型表述方式来进行网络参数描述的时候,第一端可以是将网络参数的属性和参数值一起压缩并传递,第二端基于接收到的网络参数的属性判断网络参数的长度以及对应的权重。Optionally, when describing the network parameters based on the preset model expression, the first end can compress and transmit the attributes of the network parameters and the parameter values together, and the second end can judge based on the attributes of the received network parameters The length of the network parameters and the corresponding weights.
网络参数可以是一个数也可以是多个数组成的列表,在更新网络参数的时候,如果列表中只有部分数值发生变化,第一端可以通过数值位置来指示变更的数值的位置,第一端只传递变更的数值,第二端也只更新接收到的数值。或者,第一端在传递完整的网络参数的时候,也可以通过数值位置指示网络参数中非零数值的位置,第一端只传递非零的数值,第二端将没有收到的数值当作0或者某个预设值。其中,数值位置指示可以是独立于网络参数额外指示,或者是无线资源控制(Radio Resource Control,RRC)配置或者是媒体接入控制控制元素(Medium Access Control Control Element,MACCE)配置。A network parameter can be a single number or a list of multiple numbers. When updating network parameters, if only part of the values in the list change, the first end can indicate the position of the changed value through the value position. The first end Only the changed value is transmitted, and the second end only updates the received value. Alternatively, when the first end transmits complete network parameters, it can also indicate the position of the non-zero value in the network parameter through the value position, the first end only transmits the non-zero value, and the second end treats the value that has not been received as 0 or a default value. Wherein, the numerical location indication may be an additional indication independent of network parameters, or a radio resource control (Radio Resource Control, RRC) configuration or a medium access control control element (Medium Access Control Control Element, MACCE) configuration.
为更好地理解本申请实施例提供的技术方案,以下通过几个具体的实施例进行举例说明。In order to better understand the technical solutions provided by the embodiments of the present application, several specific examples are given below for illustration.
实施例一Embodiment one
终端和基站使用联合的AI网络进行CSI反馈,即终端通过AI网络将信道信息转换成若干比特(bit)的CSI反馈信息,并上报给基站,基站接收终端反馈的bit信息,通过基站侧的AI网络将信道信息恢复出来。The terminal and the base station use the joint AI network for CSI feedback, that is, the terminal converts the channel information into several bits of CSI feedback information through the AI network, and reports it to the base station. The network recovers the channel information.
由于基站和终端的网络需要进行联合训练,不同的小区信道情况不同,也可能需要新的网络参数,因此当终端接入网络的时候,基站需要将终端使用的网络参数发送给终端。Since the network of the base station and the terminal needs to be jointly trained, and the channel conditions of different cells are different, new network parameters may also be required. Therefore, when the terminal accesses the network, the base station needs to send the network parameters used by the terminal to the terminal.
CSI反馈的网络可以分成两个部分,终端编码部分和基站解码部分,通常,基站只需要将终端编码部分的AI网络发送给终端。基站可以将需要发送的AI网络的网络结构和网络参数保存成预设的模型表述方式对应的文件,例如ONNX文件或者pytorch的pth文件,然后将整个文件通过分组数据汇聚协议(Packet Data Convergence Protocol,PDCP)层压缩之后发送给终端。终 端接收到该文件之后,根据文件格式加载到自己的AI框架下,实现利用AI网络的推断或者继续训练。其中,pytorch是一种深度学习框架,pth是pytorch中的一种文件类型。The CSI feedback network can be divided into two parts, the terminal coding part and the base station decoding part. Usually, the base station only needs to send the AI network of the terminal coding part to the terminal. The base station can save the network structure and network parameters of the AI network to be sent into a file corresponding to the preset model expression method, such as an ONNX file or a pth file of pytorch, and then pass the entire file through the Packet Data Convergence Protocol (Packet Data Convergence Protocol, PDCP) layer compressed and sent to the terminal. After receiving the file, the terminal loads it into its own AI framework according to the file format to realize inference or continue training using the AI network. Among them, pytorch is a deep learning framework, and pth is a file type in pytorch.
或者,基站将AI网络结构保存成预设的模型表述方式对应的文件,例如TensorFlow的meta文件,然后将这个meta文件经过ZIP压缩发送给终端。终端接收到该文件之后,在自己的AI框架下搭建对应的AI网络,没有收到的权重参数默认为0或者某个固定的起始值,终端可以使用这个初始值进行训练,也可以直接使用这个默认值进行推断。其中,meta是TensorFlow中的一种文件类型。Alternatively, the base station saves the AI network structure as a file corresponding to a preset model expression method, such as a TensorFlow meta file, and then sends the meta file to the terminal after ZIP compression. After the terminal receives the file, it builds the corresponding AI network under its own AI framework. The weight parameter that has not been received defaults to 0 or a fixed initial value. The terminal can use this initial value for training, or it can be used directly This default value is inferred. Among them, meta is a file type in TensorFlow.
或者,根据预先定义的AI网络模板(也即上述预设AI网络),基站发送某个索引值(index)指示对应的AI网络模板,然后将网络参数保存成预设的模型表述方式对应的文件,例如TensorFlow的data文件,也可以是将这个网络模板的index和网络参数一起保存成某种自定义的网络表述方法对应的文件,发送给终端。其中,data是TensorFlow中的一种文件类型。终端接收到对应的文件之后,根据预先定义的AI网络模板的index,在自己的AI框架内加载对应的网络结构,并根据AI网络模板的网络参数初始化自己的网络参数,然后将基站发送的网络参数更新到AI网络中,如果某些网络参数没有收到,则使用所述AI模板中的初始值。Or, according to the pre-defined AI network template (that is, the above-mentioned preset AI network), the base station sends a certain index value (index) to indicate the corresponding AI network template, and then saves the network parameters as a file corresponding to the preset model expression , such as the data file of TensorFlow, or the index and network parameters of the network template can be saved together as a file corresponding to a custom network expression method and sent to the terminal. Among them, data is a file type in TensorFlow. After receiving the corresponding file, the terminal loads the corresponding network structure in its own AI framework according to the index of the pre-defined AI network template, initializes its own network parameters according to the network parameters of the AI network template, and then transfers the network parameters sent by the base station. The parameters are updated to the AI network. If some network parameters are not received, the initial values in the AI template are used.
实施例二Embodiment two
在终端测量下行信道的过程中,终端可以使用AI网络进行信道测量,包括CSI参考信号(CSI Reference Signal,CSI-RS)、解调参考信号(Demodulation Reference Signal,DMRS)等信道估计以及无线资源管理(Radio resource management,RRM)测量等。同样,由于不同小区的信道差异性,当终端切换小区的时候,往往需要切换使用的网络,因为旧小区的网络已经无法支持新小区下信道的变化情况,此时新的小区需要向终端发送新的网络参数。In the process of the terminal measuring the downlink channel, the terminal can use the AI network to perform channel measurement, including CSI reference signal (CSI Reference Signal, CSI-RS), demodulation reference signal (Demodulation Reference Signal, DMRS) and other channel estimation and radio resource management (Radio resource management, RRM) measurement, etc. Similarly, due to the channel differences of different cells, when the terminal switches cells, it often needs to switch the used network, because the network of the old cell can no longer support the channel changes in the new cell, and the new cell needs to send a new message to the terminal. network parameters.
终端在新的小区使用的AI网络的网络结构通常可以认为和旧的小区的AI网络结构一致,但是由于信道质量的变化,需要重新进行训练,浪费时间也难以满足实时性的需要,因此新的小区可以将自己训练好的与终端网络结构相同的网络参数发送给终端,终端在这些参数基础上继续训练和推断,可 以提高实时效率,减少训练次数,快速收敛网络。The network structure of the AI network used by the terminal in the new cell can generally be considered to be consistent with the AI network structure of the old cell. However, due to changes in channel quality, retraining is required, which wastes time and is difficult to meet real-time requirements. Therefore, the new The cell can send its own trained network parameters with the same network structure as the terminal to the terminal. The terminal continues to train and infer based on these parameters, which can improve real-time efficiency, reduce the number of training times, and quickly converge the network.
新小区可以从旧小区处获得上一次发送给终端的网络参数,将自己的网络参数和上一次发送给用户的网络参数进行比较,计算权重值的平均相关性,例如:The new cell can obtain the network parameters sent to the terminal last time from the old cell, compare its own network parameters with the network parameters sent to the user last time, and calculate the average correlation of weight values, for example:
其中,
表示上一次发送给终端的网络参数的第i个数值,
表示这次发给终端的网络参数的第i个数值,N为这个网络参数的数值个数,计算的C越小代表越相关。具体相关性计算公式可以有很多种,基站根据两次权重参数相关性判断哪些网络参数需要更新,在总共的M1个网络参数中选择需要更新的M2个网络参数进行更新。
in, Indicates the i-th value of the network parameter sent to the terminal last time, Indicates the i-th value of the network parameter sent to the terminal this time, N is the number of values of this network parameter, and the smaller the calculated C, the more relevant it is. There are many specific correlation calculation formulas. The base station judges which network parameters need to be updated according to the correlation between the two weight parameters, and selects M2 network parameters to be updated from the total M1 network parameters to update.
基站可以先通知终端哪些网络参数需要更新,由于终端已知网络结构,自然已知总共的M1个网络参数,基站和终端对这M1个网络参数采用相同的排序方式,这个排序方式对应于预设的模型表述方式,基站将需要更新的M2个网络参数在这个排序中的位置以位图(bitmap)或者组合数等方法通知终端,终端接收到基站的通知之后,确定哪M2个网络参数需要更新,然后接收具体的参数数值。The base station can first inform the terminal which network parameters need to be updated. Since the terminal knows the network structure, it naturally knows a total of M1 network parameters. The base station and the terminal use the same sorting method for the M1 network parameters. This sorting method corresponds to the preset The base station notifies the terminal of the position of the M2 network parameters that need to be updated in this sort by means of a bitmap (bitmap) or combination number, etc. After receiving the notification from the base station, the terminal determines which M2 network parameters need to be updated , and then receive specific parameter values.
由于已经提前确定哪些网络参数需要更新,基站直接将需要更新的网络参数的数值压缩成预设的模型表述方式中的数据文件,无需包含权重名称、维度等信息,直接传递数值,终端接收到数值之后,根据自己已知的每个网络参数的维度在对应的位置解析这个网络参数的参数数值,更新自己的AI框架内的网络参数。Since it has been determined in advance which network parameters need to be updated, the base station directly compresses the values of the network parameters that need to be updated into a data file in the preset model expression mode, without including weight name, dimension and other information, and directly transmits the values, and the terminal receives the values Afterwards, according to the dimensions of each network parameter that you know, analyze the parameter value of this network parameter at the corresponding position, and update the network parameters in your own AI framework.
或者,基站可以直接将需要更新的网络参数名称和数值一起压缩到预设的模型表述方式的文件中,终端首先解析网络参数名称,然后根据这个名称获取对应的数值。Alternatively, the base station can directly compress the name and value of the network parameter to be updated into a file in a preset model expression mode, and the terminal first parses the name of the network parameter, and then obtains the corresponding value according to the name.
可选地,针对每一个需要更新的网络参数,基站首先在这个网络参数的全部的K个数值中分别计算对应的相关性,如果相关性大于某一固定阈值, 则认为这个数值可以不更新,然后将所有需要更新的数值的位置通过bitmap或者组合数的方法与需要传递的数值一起压缩到预设的模型表述方式对应的文件中,终端收到压缩后的文件之后,首先解出来哪些网络参数需要更新,然后再判断这个网络参数的哪些数值需要更新,只更新对应位置的数值。Optionally, for each network parameter that needs to be updated, the base station first calculates the corresponding correlation among all K values of the network parameter, and if the correlation is greater than a certain fixed threshold, it is considered that this value does not need to be updated, Then compress the position of all the values that need to be updated into the file corresponding to the preset model expression method through the method of bitmap or combination number and the value that needs to be transmitted. After receiving the compressed file, the terminal will first extract which network parameters It needs to be updated, and then judge which values of this network parameter need to be updated, and only update the values of the corresponding positions.
具体的,需要更新的网络参数,每个网络参数中需要更新的数值位置可以提前指示或者配置给终端,然后直接将对应的数值信息压缩之后发送给终端。Specifically, the network parameters that need to be updated, and the numerical position that needs to be updated in each network parameter can be indicated in advance or configured to the terminal, and then the corresponding numerical information can be directly compressed and sent to the terminal.
实施例三Embodiment three
两个使用AI网络的节点交互AI网络信息的时候需要将AI网络信息压缩到某种预设的模型表述方式对应的文件中,所谓预设的模型表述方式即为一种数据结构,按照一定的规则描述AI网络,描述AI网络参数等信息。When two nodes using the AI network exchange AI network information, it is necessary to compress the AI network information into a file corresponding to a certain preset model expression method. The so-called preset model expression method is a data structure, according to a certain Rules describe the AI network and describe information such as AI network parameters.
例如,预设的模型表述方式包括TensorFlow、PyTorch、ONNX等,是将AI网络描述按照固定的规则进行描述,例如将网络描述成若干个独立的节点的组合,或者是描述成若干个层的组合,每个层有独立的功能,具体的功能通过一些权重参数以及激活函数等方式表述,不同的框架都有自己对网络的定义方式,为了统一在通信系统中传递,可以定义一些符合通信需求的网络描述方式。For example, the preset model expression methods include TensorFlow, PyTorch, ONNX, etc., which describe the AI network description according to fixed rules, such as describing the network as a combination of several independent nodes, or describing it as a combination of several layers , each layer has an independent function, the specific function is expressed by some weight parameters and activation functions, etc., different frameworks have their own definition of the network, in order to uniformly transmit in the communication system, you can define some communication requirements network description.
AI网络的网络结构和网络参数应当分离,可以支持独立的网络结构的传递和独立的网络参数的传递。网络结构可以是以节点为基础单元,定义节点的基本功能,将所有的权重映射到输入输出上,通过输入输出使用相同编号表示连接,从而通过节点和输入输出的编号描述整个网络。The network structure and network parameters of the AI network should be separated to support the transfer of independent network structures and independent network parameters. The network structure can be based on the node, define the basic functions of the node, map all the weights to the input and output, and use the same number to represent the connection through the input and output, so as to describe the entire network through the number of nodes and input and output.
对于网络参数,包括以下至少一项:For network parameters, include at least one of the following:
1.参数的编号,和网络结构中描述的编号一致;1. The number of parameters is consistent with the number described in the network structure;
2.参数的维度,需要和与之连接的其他参数匹配;2. The dimension of the parameter needs to match the other parameters connected to it;
3.参数的数值,即为具体参数内容,和维度相匹配;3. The value of the parameter, that is, the specific parameter content, matches the dimension;
4.参数中有效数值位置信息,为了减少参数传递时的开销,没有变化或者本身接近0的参数可以不传递,需要对应的位置信息指示有效数值的位置。4. In order to reduce the overhead of parameter transfer, the position information of the effective value in the parameter, the parameter that has not changed or is close to 0 may not be passed, and the corresponding position information is required to indicate the position of the effective value.
这些内容可以任意组合成独立的文件,将完整的网络参数表示成差分之后的结果,即部分参数不传递,或者将完整的网络表示成部分权重的结果, 即对应压缩操作。These contents can be combined into independent files arbitrarily, and the complete network parameters are expressed as the result after the difference, that is, some parameters are not passed, or the complete network is expressed as the result of partial weights, that is, the corresponding compression operation.
实施例四Embodiment four
在一些场景下,有些功能只需要终端使用AI网络,并不需要和网络侧设备联合训练,例如终端侧的信道预测,定位等,这些AI网络受信道环境影响,随着终端的移动,需要不断训练,更新网络。当终端B进入到终端A的区域的时候,终端A可以将已经训练好的AI网络传递给终端B,终端B基于终端A的AI网络进行训练,由于终端A和终端B在相近的区域,信道信息有一定相关性,因此使用终端A的AI网络继续训练可以帮助终端B更快地收敛。In some scenarios, some functions only require the terminal to use the AI network, and do not require joint training with network-side devices, such as channel prediction and positioning on the terminal side. These AI networks are affected by the channel environment. As the terminal moves, it needs to be constantly Train and update the network. When terminal B enters the area of terminal A, terminal A can pass the trained AI network to terminal B, and terminal B performs training based on terminal A's AI network. Since terminal A and terminal B are in similar areas, the channel The information has a certain correlation, so using the AI network of terminal A to continue training can help terminal B converge faster.
在任意两个有关联的节点之间,都可以通过传递已经训练好的AI网络降低训练复杂度,甚至直接使用别的节点训练好的AI网络。Between any two associated nodes, the training complexity can be reduced by passing the trained AI network, or even directly use the AI network trained by other nodes.
终端A和终端B针对同一功能的AI网络的结构是相同的,二者需要交互的主要是网络参数,考虑到副链路(sidelink)的传输限制,终端A可以将网络参数分成N份,分别在不同的时隙(slot)传递给终端B,每次传递的时候,网络参数名称、维度等信息可以是参数数值统一压缩也可以是独立配置之后直接传递数值。终端B接收到终端A的部分网络参数之后,可以直接更新自己对应位置的网络参数,并使用更新后的网络直接进行推断,或者终端B可以等到全部网络参数接收完整之后,一起更新自己的对应的网络参数,这一点主要取决于部分网络参数是否可以独立工作,这个能力可以由终端A通知终端B,或者通过共同的基站配置,如果可以独立工作,终端B每次接收到新的网络参数之后都可以更新自身网络,否则就要等到全部接收完毕之后一起更新。Terminal A and terminal B have the same AI network structure for the same function. The two need to interact mainly with network parameters. Considering the transmission limitation of the sidelink, terminal A can divide the network parameters into N parts, respectively In different time slots (slots), it is transmitted to terminal B. When transmitting each time, information such as network parameter names and dimensions can be uniformly compressed with parameter values or can be directly transmitted after independent configuration. After receiving part of the network parameters of terminal A, terminal B can directly update the network parameters of its corresponding location, and use the updated network to directly infer, or terminal B can wait until all network parameters are received, and update its corresponding network parameters together. Network parameters, which mainly depends on whether some network parameters can work independently. This capability can be notified by terminal A to terminal B, or configured through a common base station. If it can work independently, terminal B will update each time it receives new network parameters. You can update your own network, otherwise you have to wait until all are received and update together.
终端B也可以指定某个需要更新的网络参数,将需要更新的网络参数的名称和/或维度等信息发送给终端A,可以同时包括期望接收网络参数的资源位置等信息,终端A根据终端B的信令,将对应的网络参数的数值按照终端A需要的网络参数的顺序,直接压缩成某种预设的模型表述方式对应的文件发送给终端B,终端B接收之后,更新自己需要的网络参数。Terminal B can also specify a network parameter that needs to be updated, and send information such as the name and/or dimension of the network parameter that needs to be updated to terminal A, and can also include information such as the location of the resource that is expected to receive the network parameter. Terminal A according to terminal B According to the order of the network parameters required by terminal A, the corresponding network parameter values are directly compressed into a file corresponding to a preset model expression method and sent to terminal B. After terminal B receives it, it updates the network it needs parameter.
请参照图3,图3是本申请实施例提供的另一种AI网络信息传输方法,如图3所示,所述AI网络信息传输方法包括以下步骤:Please refer to FIG. 3. FIG. 3 is another AI network information transmission method provided by the embodiment of the present application. As shown in FIG. 3, the AI network information transmission method includes the following steps:
步骤301、第二端接收第一端发送的压缩后的AI网络信息,所述AI网络信息包括网络结构和网络参数中的至少一项。 Step 301. The second end receives the compressed AI network information sent by the first end, where the AI network information includes at least one of network structure and network parameters.
本申请实施例中,所述第一端和所述第二端为具有发送和接收功能的通信设备。In the embodiment of the present application, the first end and the second end are communication devices with sending and receiving functions.
可选地,所述第一端为网络侧设备和终端中的一者,所述第二端为网络侧设备和终端中的另一者;或者,所述第一端和所述第二端为终端的不同节点;或者,所述第一端和所述第二端为网络侧设备的不同节点。Optionally, the first end is one of the network-side device and the terminal, and the second end is the other of the network-side device and the terminal; or, the first end and the second end are different nodes of a terminal; or, the first end and the second end are different nodes of a network side device.
需要说明地,所述网络侧设备可以包括接入网设备(例如:基站)和核心网设备。可选地,所述第一端可以为接入网设备,第二端为核心网设备;或者,第一端为终端,第二端为核心网设备或接入网设备;或者,所述第一端和所述第二端为接入网设备的不同节点;或者,所述第一端和所述第二端为核心网设备的不同节点,等等,本申请实施例不做一一列举。It should be noted that the network side equipment may include access network equipment (for example: base station) and core network equipment. Optionally, the first end may be an access network device, and the second end may be a core network device; or, the first end may be a terminal, and the second end may be a core network device or an access network device; or, the second end may be a core network device or an access network device; The one end and the second end are different nodes of the access network equipment; or, the first end and the second end are different nodes of the core network equipment, etc., and the embodiments of the present application do not list them one by one .
本申请实施例中,所述AI网络信息包括网络结构和网络参数中的至少一项。可选地,所述AI网络信息可以是某一个AI网络的网络结构和/或网络参数,或者是多个AI网络的网络结构和/或网络参数。在一些实施例中,AI网络也可以称为AI神经网络、AI模型等。其中,所述网络参数包括AI网络的权重参数、超参数等。In this embodiment of the present application, the AI network information includes at least one of network structure and network parameters. Optionally, the AI network information may be the network structure and/or network parameters of a certain AI network, or the network structures and/or network parameters of multiple AI networks. In some embodiments, an AI network may also be called an AI neural network, an AI model, or the like. Wherein, the network parameters include weight parameters, hyperparameters and the like of the AI network.
可以理解地,第一端对AI网络信息进行压缩,将压缩后的AI网络信息发送给第二端。其中,所述对AI网络信息进行压缩,可以是指按照预设的模型表述方式将所述AI网络信息压缩到预设的模型表述方式对应的文件中,所谓模型表述方式为一种数据结构,按照一定的规则描述AI网络结构、网络参数等信息。所述第一端对AI网络信息进行压缩的实现过程可以参照上述图2所述方法实施例中的具体描述,此处不再赘述。Understandably, the first end compresses the AI network information, and sends the compressed AI network information to the second end. Wherein, the compressing the AI network information may refer to compressing the AI network information into a file corresponding to the preset model representation method according to the preset model representation method, the so-called model representation method is a data structure, Describe the AI network structure, network parameters and other information according to certain rules. For the implementation process of the first terminal compressing the AI network information, reference may be made to the specific description in the method embodiment shown in FIG. 2 above, and details are not repeated here.
其中,所述AI网络信息包括网络结构和/或网络参数,第一端对AI网络信息的压缩也就包括对网络结构和/或权重参数进行压缩,并将压缩后的网络结构和/或权重参数发送给第二端。例如,AI网络信息仅包括网络结构,则第一端仅对网络结构进行压缩并发送;或者,AI网络信息也可以是仅包括网络参数,则第一端仅对网络参数进行压缩并发送;或者,AI网络信息包括部分网络结构和部分网络参数,则第一端对该部分网络结构和部分网络参数进行 压缩并发送。当然,所述AI网络信息包括的具体信息内容还可以是其他的情况,此处不做过多赘述。Wherein, the AI network information includes network structure and/or network parameters, and the compression of the AI network information by the first end also includes compressing the network structure and/or weight parameters, and compressing the compressed network structure and/or weight parameters Parameters are sent to the second end. For example, if the AI network information only includes the network structure, the first end only compresses and sends the network structure; or, the AI network information may only include network parameters, then the first end only compresses and sends the network parameters; or , the AI network information includes part of the network structure and part of the network parameters, and the first end compresses and sends the part of the network structure and part of the network parameters. Of course, the specific information content included in the AI network information may also be in other situations, which will not be described in detail here.
本申请实施例中,AI网络信息包括网络结构和网络参数中的至少一项,进而在通信过程中也就无需将包括全部网络结构和网络参数的整个AI网络一起进行传输,使得AI网络的网络结构和网络参数可以分开发送,进而能够有效降低通信过程中的传输开销。In the embodiment of the present application, the AI network information includes at least one of the network structure and network parameters, and there is no need to transmit the entire AI network including all network structures and network parameters together during the communication process, so that the network of the AI network The structure and network parameters can be sent separately, which can effectively reduce the transmission overhead in the communication process.
可选地,所述步骤301之前,所述方法还包括:Optionally, before the step 301, the method further includes:
所述第二端向所述第一端发送第一请求信息,所述第一请求信息用于请求获取目标AI网络信息;The second terminal sends first request information to the first terminal, where the first request information is used to request acquisition of target AI network information;
这种情况下,所述步骤301包括:In this case, the step 301 includes:
所述第二端接收所述第一端发送的压缩后的所述目标AI网络信息。The second end receives the compressed target AI network information sent by the first end.
可选地,所述第一请求信息包括如下至少一项:Optionally, the first request information includes at least one of the following:
请求的网络参数的名称;the name of the requested network parameter;
请求的网络参数的标识;Identification of the requested network parameters;
网络结构更新请求;network structure update request;
网络参数更新请求;Network parameter update request;
AI网络的网络效果度量值。Network Effect Measures for AI Networks.
需要说明地,本申请实施例所提供的AI网络信息传输方法应用于第二端,与上述图2实施例中所提供的应用于第一端的AI网络信息传输方法相对应,本申请实施例中相关步骤的具体实现过程可以参照上述图2所述方法实施例中的描述,为避免重复,此处不再赘述。It should be noted that the AI network information transmission method provided by the embodiment of the present application is applied to the second end, corresponding to the AI network information transmission method applied to the first end provided in the embodiment of FIG. 2 above. For the specific implementation process of the relevant steps, reference may be made to the description in the above-mentioned method embodiment shown in FIG. 2 , and details are not repeated here to avoid repetition.
本申请实施例中,第二端接收第一端发送的压缩的AI网络信息,所述AI网络信息包括网络结构和网络参数中的至少一项,进而在通信过程中也就无需将包括全部网络结构和网络参数的整个AI网络一起进行压缩传输,使得AI网络的网络结构和网络参数可以分开发送,进而能够有效降低通信过程中的传输开销。In the embodiment of the present application, the second end receives the compressed AI network information sent by the first end, the AI network information includes at least one of the network structure and network parameters, and there is no need to include all network information during the communication process. The structure and network parameters of the entire AI network are compressed and transmitted together, so that the network structure and network parameters of the AI network can be sent separately, which can effectively reduce the transmission overhead in the communication process.
本申请实施例提供的AI网络信息传输方法,执行主体可以为AI网络信息传输装置。本申请实施例中以AI网络信息传输装置执行AI网络信息传输方法为例,说明本申请实施例提供的AI网络信息传输装置。The AI network information transmission method provided in the embodiment of the present application may be executed by an AI network information transmission device. In the embodiment of the present application, the AI network information transmission device provided in the embodiment of the present application is described by taking the AI network information transmission device executing the AI network information transmission method as an example.
请参照图4,图4是本申请实施例提供的一种AI网络信息传输装置的结构图,如图4所示,AI网络信息传输装置400包括:Please refer to FIG. 4. FIG. 4 is a structural diagram of an AI network information transmission device provided in an embodiment of the present application. As shown in FIG. 4, the AI network information transmission device 400 includes:
压缩模块401,用于对AI网络信息进行压缩,所述AI网络信息包括网络结构和网络参数中的至少一项;A compression module 401, configured to compress AI network information, where the AI network information includes at least one of network structure and network parameters;
发送模块402,用于向第二端发送压缩后的所述AI网络信息。The sending module 402 is configured to send the compressed AI network information to the second end.
可选地,所述AI网络信息包括网络结构和网络参数,所述压缩模块401用于执行如下任意一项:Optionally, the AI network information includes network structure and network parameters, and the compression module 401 is configured to perform any of the following:
对所述网络结构和所述网络参数进行合并压缩;Combining and compressing the network structure and the network parameters;
对所述网络结构和所述网络参数分别进行压缩。Compressing the network structure and the network parameters respectively.
可选地,在所述压缩模块401对所述网络结构和所述网络参数分别进行压缩的情况下,所述发送模块402用于执行如下任意一项:Optionally, when the compression module 401 compresses the network structure and the network parameters respectively, the sending module 402 is configured to perform any of the following:
向第二端合并发送压缩后的网络结构和压缩后的网络参数;Combining and sending the compressed network structure and compressed network parameters to the second end;
向第二端分别发送压缩后的网络结构和压缩后的网络参数。The compressed network structure and the compressed network parameters are respectively sent to the second end.
可选地,所述压缩模块401用于执行如下任意一项:Optionally, the compression module 401 is configured to perform any of the following:
基于预设的模型表述方式将AI网络信息转换成对应的传输文件,对所述传输文件进行压缩;Convert the AI network information into a corresponding transmission file based on the preset model expression method, and compress the transmission file;
基于预设的数据格式对所述AI网络信息进行压缩;Compressing the AI network information based on a preset data format;
获取待发送的AI网络信息和所述第二端已有的AI网络信息,对所述待发送的AI网络信息与所述第二端已有的AI网络信息之间的AI网络信息差值进行压缩;Acquiring the AI network information to be sent and the existing AI network information of the second end, and performing an AI network information difference between the AI network information to be sent and the existing AI network information of the second end compression;
获取待发送的AI网络信息和预设AI网络的AI网络信息,对所述待发送的AI网络信息与所述预设AI网络的AI网络信息之间的AI网络信息差值进行压缩。Acquiring the AI network information to be sent and the AI network information of the preset AI network, and compressing the AI network information difference between the AI network information to be sent and the AI network information of the preset AI network.
可选地,所述AI网络信息差值包括如下至少一项:Optionally, the AI network information difference includes at least one of the following:
指定的网络参数;specified network parameters;
网络参数的索引;index of network parameters;
修改的网络参数;Modified network parameters;
修改的网络参数中修改的参数值;The modified parameter value in the modified network parameter;
修改的网络参数中修改的参数值的位置;The position of the modified parameter value in the modified network parameter;
修改的网络参数中修改的参考值的位置,所述参考值为所述网络参数中的最大值;the location of the modified reference value in the modified network parameters, the reference value being the maximum value in the network parameters;
修改的网络参数中的非零值;A non-zero value in the modified network parameter;
修改的网络参数中的非零值的位置;the location of non-zero values in the modified network parameters;
新增的网络结构;Newly added network structure;
删除的网络结构;Deleted network structure;
修改的网络结构。Modified network structure.
可选地,所述预设的模型表述方式包括如下任意一项:协议约定的模型表述方式、自定义的模型表述方式。Optionally, the preset model expression manner includes any one of the following: a protocol-agreed model expression manner, and a user-defined model expression manner.
可选地,所述自定义的模型表述方式的内容包括如下至少一项:AI网络的网络结构、AI网络的网络参数的属性、AI网络的网络参数的数值。Optionally, the content of the self-defined model expression includes at least one of the following: the network structure of the AI network, the attributes of the network parameters of the AI network, and the values of the network parameters of the AI network.
可选地,所述预设的模型表述方式中网络结构的表述方式包括如下至少一项:Optionally, the representation of the network structure in the preset model representation includes at least one of the following:
AI网络的网络结构之间的关联关系;The relationship between the network structures of the AI network;
AI网络的网络参数的属性;Attributes of the network parameters of the AI network;
AI网络的网络参数中非零值的位置;The location of non-zero values in the network parameters of the AI network;
AI网络的网络参数中的更新数值位置。The updated numerical position in the network parameters of the AI network.
可选地,所述压缩模块401还用于:Optionally, the compression module 401 is also used for:
基于至少一个预设的模型表述方式将AI网络信息转换成至少一个传输文件,一个所述预设的模型表述方式对应至少一个传输文件;Converting AI network information into at least one transmission file based on at least one preset model representation, where one preset model representation corresponds to at least one transmission file;
对所述至少一个传输文件进行合并压缩,或者,对所述至少一个传输文件分别进行压缩后再合并。Combining and compressing the at least one transmission file, or compressing and merging the at least one transmission file respectively.
可选地,所述AI网络信息包括网络结构和网络参数,所述压缩模块401还用于:Optionally, the AI network information includes network structure and network parameters, and the compression module 401 is also used for:
基于预设的模型表述方式将所述网络结构转换成第一传输文件,并基于所述预设的模型表述方式将所述网络参数转换成第二传输文件,将所述第一传输文件和所述第二传输文件分别进行压缩。Converting the network structure into a first transmission file based on a preset model representation, converting the network parameters into a second transmission file based on the preset model representation, and converting the first transmission file and the The above-mentioned second transmission files are respectively compressed.
可选地,所述压缩后的AI网络信息包括压缩后的网络参数,所述发送模块402还用于:Optionally, the compressed AI network information includes compressed network parameters, and the sending module 402 is further configured to:
基于所述压缩后的网络参数的优先级顺序,按照所述优先级顺序向第二端发送所述压缩后的网络参数。Based on the priority order of the compressed network parameters, send the compressed network parameters to the second end according to the priority order.
可选地,所述发送模块402还用于:Optionally, the sending module 402 is also configured to:
基于所述压缩后的网络参数的优先级顺序,对所述压缩后的网络参数进行分组;grouping the compressed network parameters based on the priority order of the compressed network parameters;
在传输资源小于预设阈值的情况下,按照预设顺序对分组后的网络参数进行丢弃并发送剩余的网络参数,所述预设顺序为分组后的网络参数的优先级从低至高的顺序。When the transmission resource is less than the preset threshold, the grouped network parameters are discarded and the remaining network parameters are sent according to a preset order, the preset order being the order of priority of the grouped network parameters from low to high.
可选地,所述装置还包括:Optionally, the device also includes:
接收模块,用于接收所述第二端发送的第一请求信息,所述第一请求信息用于请求获取目标AI网络信息;A receiving module, configured to receive first request information sent by the second end, where the first request information is used to request acquisition of target AI network information;
所述压缩模块401还用于:对所述目标AI网络信息进行压缩。The compression module 401 is further configured to: compress the target AI network information.
可选地,所述第一请求信息包括如下至少一项:Optionally, the first request information includes at least one of the following:
请求的网络参数的名称;the name of the requested network parameter;
请求的网络参数的标识;Identification of the requested network parameters;
网络结构更新请求;network structure update request;
网络参数更新请求;Network parameter update request;
AI网络的网络效果度量值。Network Effect Measures for AI Networks.
可选地,所述装置还包括:Optionally, the device also includes:
判断模块,用于判断是否需要对所述目标AI网络信息进行更新;A judging module, configured to judge whether the target AI network information needs to be updated;
在判定需要对所述目标AI网络信息进行更新的情况下,更新所述目标AI网络信息;When it is determined that the target AI network information needs to be updated, update the target AI network information;
所述压缩模块401还用于:The compression module 401 is also used for:
对更新后的所述目标AI网络信息进行压缩。Compressing the updated target AI network information.
可选地,所述目标AI网络信息包括第一目标网络参数,所述压缩模块401还用于:Optionally, the target AI network information includes first target network parameters, and the compression module 401 is further configured to:
基于预设的模型表述方式将所述第一目标网络参数的属性和参数值转换成预设格式后进行压缩;converting the attributes and parameter values of the first target network parameters into a preset format based on a preset model expression manner, and then compressing;
其中,所述第一目标网络参数的属性包括名称、维度、长度中的至少一 项。Wherein, the attribute of the first target network parameter includes at least one of name, dimension, and length.
可选地,所述装置为网络侧设备和终端中的一者,所述第二端为网络侧设备和终端中的另一者;或者,Optionally, the device is one of the network side device and the terminal, and the second end is the other of the network side device and the terminal; or,
所述装置和所述第二端为终端的不同节点;或者,The device and the second end are different nodes of a terminal; or,
所述装置和所述第二端为网络侧设备的不同节点。The device and the second end are different nodes of the network side equipment.
可选地,所述装置为网络侧设备,所述第二端为终端,所述AI网络信息包括网络参数,在所述第二端从第一小区切换到第二小区的情况下,所述第一端对AI网络信息进行压缩之前,所述方法还包括:Optionally, the device is a network side device, the second end is a terminal, the AI network information includes network parameters, and when the second end is handed over from a first cell to a second cell, the Before the first end compresses the AI network information, the method further includes:
所述第一端计算所述第一小区的网络参数与所述第二小区的网络参数的相关性,获取第二目标网络参数,所述第二目标网络参数包括如下至少一项:相关性小于预设阈值的网络参数、预设序列中的前N个网络参数,所述预设序列为网络参数的相关性按照从小到大顺序排列的序列;The first end calculates the correlation between the network parameters of the first cell and the network parameters of the second cell, and obtains a second target network parameter, where the second target network parameter includes at least one of the following items: the correlation is less than Network parameters with preset thresholds, and the first N network parameters in a preset sequence, where the preset sequence is a sequence in which the correlation of network parameters is arranged in ascending order;
所述压缩模块401还用于:The compression module 401 is also used for:
对所述第二目标网络参数进行压缩;Compressing the second target network parameters;
所述发送模块402还用于:The sending module 402 is also used for:
向第二端发送压缩后的所述第二目标网络参数。Sending the compressed second target network parameters to the second end.
本申请实施例中,所述装置能够向第二端发送压缩的AI网络信息,所述AI网络信息包括网络结构和网络参数中的至少一项,进而在通信过程中也就无需将包括全部网络结构和网络参数的整个AI网络一起进行传输,使得AI网络的网络结构和网络参数可以分开发送,进而能够有效降低通信过程中的传输开销。In the embodiment of the present application, the device can send compressed AI network information to the second end, the AI network information includes at least one of the network structure and network parameters, and there is no need to include all network information during the communication process. The structure and network parameters of the entire AI network are transmitted together, so that the network structure and network parameters of the AI network can be sent separately, which can effectively reduce the transmission overhead in the communication process.
本申请实施例中的AI网络信息传输装置400可以是电子设备,例如具有操作系统的电子设备,也可以是电子设备中的部件,例如集成电路或芯片。该电子设备可以是终端,也可以为除终端之外的其他设备。示例性的,终端可以包括但不限于上述所列举的终端11的类型,其他设备可以为服务器、网络附属存储器(Network Attached Storage,NAS)等,本申请实施例不作具体限定。The AI network information transmission apparatus 400 in the embodiment of the present application may be an electronic device, such as an electronic device with an operating system, or a component of the electronic device, such as an integrated circuit or a chip. The electronic device may be a terminal, or other devices other than the terminal. Exemplarily, the terminal may include, but not limited to, the types of terminal 11 listed above, and other devices may be servers, Network Attached Storage (NAS), etc., which are not specifically limited in this embodiment of the present application.
本申请实施例提供的AI网络信息传输装置400能够实现图2所述方法实施例实现的各个过程,并达到相同的技术效果,为避免重复,这里不再赘述。The AI network information transmission device 400 provided in the embodiment of the present application can realize each process implemented in the method embodiment shown in FIG. 2 and achieve the same technical effect. To avoid repetition, details are not repeated here.
请参照图5,图5是本申请实施例提供的另一种AI网络信息传输装置的结构图,如图5所示,所述AI网络信息传输装置500包括:Please refer to FIG. 5. FIG. 5 is a structural diagram of another AI network information transmission device provided in the embodiment of the present application. As shown in FIG. 5, the AI network information transmission device 500 includes:
接收模块501,用于接收第一端发送的压缩后的AI网络信息,所述AI网络信息包括网络结构和网络参数中的至少一项。The receiving module 501 is configured to receive compressed AI network information sent by the first end, where the AI network information includes at least one of network structure and network parameters.
可选地,所述装置还包括:Optionally, the device also includes:
发送模块,用于向所述第一端发送第一请求信息,所述第一请求信息用于请求获取目标AI网络信息;A sending module, configured to send first request information to the first end, where the first request information is used to request acquisition of target AI network information;
所述接收模块501还用于:The receiving module 501 is also used for:
接收所述第一端发送的压缩后的所述目标AI网络信息。Receive the compressed target AI network information sent by the first end.
可选地,所述第一请求信息包括如下至少一项:Optionally, the first request information includes at least one of the following:
请求的网络参数的名称;the name of the requested network parameter;
请求的网络参数的标识;Identification of the requested network parameters;
网络结构更新请求;network structure update request;
网络参数更新请求;Network parameter update request;
AI网络的网络效果度量值。Network Effect Measures for AI Networks.
可选地,所述第一端为网络侧设备和终端中的一者,所述装置为网络侧设备和终端中的另一者;或者,Optionally, the first end is one of the network side device and the terminal, and the device is the other of the network side device and the terminal; or,
所述第一端和所述装置为终端的不同节点;或者,The first end and the device are different nodes of terminals; or,
所述第一端和所述装置为网络侧设备的不同节点。The first end and the device are different nodes of network side equipment.
本申请实施例中,所述装置接收第一端发送的压缩的AI网络信息,所述AI网络信息包括网络结构和网络参数中的至少一项,进而在通信过程中也就无需将包括全部网络结构和网络参数的整个AI网络一起进行压缩传输,使得AI网络的网络结构和网络参数可以分开发送,进而能够有效降低通信过程中的传输开销。In the embodiment of the present application, the device receives the compressed AI network information sent by the first end, the AI network information includes at least one of the network structure and network parameters, and there is no need to include all network information during the communication process. The structure and network parameters of the entire AI network are compressed and transmitted together, so that the network structure and network parameters of the AI network can be sent separately, which can effectively reduce the transmission overhead in the communication process.
本申请实施例提供的AI网络信息传输装置500能够实现图3所述方法实施例实现的各个过程,并达到相同的技术效果,为避免重复,这里不再赘述。The AI network information transmission device 500 provided in the embodiment of the present application can realize each process implemented in the method embodiment shown in FIG. 3 and achieve the same technical effect. To avoid repetition, details are not repeated here.
可选地,如图6所示,本申请实施例还提供一种通信设备600,包括处理器601和存储器602,存储器602上存储有可在所述处理器601上运行的程序或指令,该程序或指令被处理器601执行时实现上述图2或图3所述的 AI网络信息传输方法实施例的各个步骤,且能达到相同的技术效果。为避免重复,这里不再赘述。Optionally, as shown in FIG. 6 , the embodiment of the present application further provides a communication device 600, including a processor 601 and a memory 602, and the memory 602 stores programs or instructions that can run on the processor 601. The When the programs or instructions are executed by the processor 601, the various steps of the embodiment of the AI network information transmission method described above in FIG. 2 or FIG. 3 can be realized, and the same technical effect can be achieved. To avoid repetition, details are not repeated here.
本申请实施例还提供一种终端,上述图2或图3方法实施例的各个实施过程和实现方式均可适用于该终端实施例中,且能达到相同的技术效果。具体地,图7为实现本申请实施例的一种终端的硬件结构示意图。The embodiment of the present application also provides a terminal, and each implementation process and implementation manner of the above-mentioned method embodiment in FIG. 2 or FIG. 3 can be applied to the terminal embodiment, and can achieve the same technical effect. Specifically, FIG. 7 is a schematic diagram of a hardware structure of a terminal implementing an embodiment of the present application.
该终端700包括但不限于:射频单元701、网络模块702、音频输出单元703、输入单元704、传感器705、显示单元706、用户输入单元707、接口单元708、存储器709以及处理器710等中的至少部分部件。The terminal 700 includes, but is not limited to: a radio frequency unit 701, a network module 702, an audio output unit 703, an input unit 704, a sensor 705, a display unit 706, a user input unit 707, an interface unit 708, a memory 709, and a processor 710. At least some parts.
本领域技术人员可以理解,终端700还可以包括给各个部件供电的电源(比如电池),电源可以通过电源管理系统与处理器710逻辑相连,从而通过电源管理系统实现管理充电、放电、以及功耗管理等功能。图7中示出的终端结构并不构成对终端的限定,终端可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件布置,在此不再赘述。Those skilled in the art can understand that the terminal 700 may also include a power supply (such as a battery) for supplying power to various components, and the power supply may be logically connected to the processor 710 through the power management system, so as to manage charging, discharging, and power consumption through the power management system. Management and other functions. The terminal structure shown in FIG. 7 does not constitute a limitation on the terminal, and the terminal may include more or fewer components than shown in the figure, or combine some components, or arrange different components, which will not be repeated here.
应理解的是,本申请实施例中,输入单元704可以包括图形处理单元(Graphics Processing Unit,GPU)7041和麦克风7042,图形处理器7041对在视频捕获模式或图像捕获模式中由图像捕获装置(如摄像头)获得的静态图片或视频的图像数据进行处理。显示单元706可包括显示面板7061,可以采用液晶显示器、有机发光二极管等形式来配置显示面板7061。用户输入单元707包括触控面板7071以及其他输入设备7072中的至少一种。触控面板7071,也称为触摸屏。触控面板7071可包括触摸检测装置和触摸控制器两个部分。其他输入设备7072可以包括但不限于物理键盘、功能键(比如音量控制按键、开关按键等)、轨迹球、鼠标、操作杆,在此不再赘述。It should be understood that, in this embodiment of the present application, the input unit 704 may include a graphics processing unit (Graphics Processing Unit, GPU) 7041 and a microphone 7042, and the graphics processor 7041 is used by the image capture device ( Such as the image data of the still picture or video obtained by the camera) for processing. The display unit 706 may include a display panel 7061, and the display panel 7061 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like. The user input unit 707 includes at least one of a touch panel 7071 and other input devices 7072 . The touch panel 7071 is also called a touch screen. The touch panel 7071 may include two parts, a touch detection device and a touch controller. Other input devices 7072 may include, but are not limited to, physical keyboards, function keys (such as volume control buttons, switch buttons, etc.), trackballs, mice, and joysticks, which will not be described in detail here.
本申请实施例中,射频单元701接收来自网络侧设备的下行数据后,可以传输给处理器710进行处理;另外,射频单元701可以向网络侧设备发送上行数据。通常,射频单元701包括但不限于天线、放大器、收发信机、耦合器、低噪声放大器、双工器等。In the embodiment of the present application, the radio frequency unit 701 may transmit the downlink data from the network side device to the processor 710 for processing after receiving the downlink data; in addition, the radio frequency unit 701 may send uplink data to the network side device. Generally, the radio frequency unit 701 includes, but is not limited to, an antenna, an amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like.
存储器709可用于存储软件程序或指令以及各种数据。存储器709可主要包括存储程序或指令的第一存储区和存储数据的第二存储区,其中,第一存储区可存储操作系统、至少一个功能所需的应用程序或指令(比如声音播 放功能、图像播放功能等)等。此外,存储器709可以包括易失性存储器或非易失性存储器,或者,存储器709可以包括易失性和非易失性存储器两者。其中,非易失性存储器可以是只读存储器(Read-Only Memory,ROM)、可编程只读存储器(Programmable ROM,PROM)、可擦除可编程只读存储器(Erasable PROM,EPROM)、电可擦除可编程只读存储器(Electrically EPROM,EEPROM)或闪存。易失性存储器可以是随机存取存储器(Random Access Memory,RAM),静态随机存取存储器(Static RAM,SRAM)、动态随机存取存储器(Dynamic RAM,DRAM)、同步动态随机存取存储器(Synchronous DRAM,SDRAM)、双倍数据速率同步动态随机存取存储器(Double Data Rate SDRAM,DDRSDRAM)、增强型同步动态随机存取存储器(Enhanced SDRAM,ESDRAM)、同步连接动态随机存取存储器(Synch link DRAM,SLDRAM)和直接内存总线随机存取存储器(Direct Rambus RAM,DRRAM)。本申请实施例中的存储器709包括但不限于这些和任意其它适合类型的存储器。The memory 709 can be used to store software programs or instructions as well as various data. The memory 709 may mainly include a first storage area for storing programs or instructions and a second storage area for storing data, wherein the first storage area may store an operating system, an application program or instructions required by at least one function (such as a sound playing function, image playback function, etc.), etc. Furthermore, memory 709 may include volatile memory or nonvolatile memory, or, memory 709 may include both volatile and nonvolatile memory. Among them, the non-volatile memory can be read-only memory (Read-Only Memory, ROM), programmable read-only memory (Programmable ROM, PROM), erasable programmable read-only memory (Erasable PROM, EPROM), electronically programmable Erase Programmable Read-Only Memory (Electrically EPROM, EEPROM) or Flash. Volatile memory can be random access memory (Random Access Memory, RAM), static random access memory (Static RAM, SRAM), dynamic random access memory (Dynamic RAM, DRAM), synchronous dynamic random access memory (Synchronous DRAM, SDRAM), double data rate synchronous dynamic random access memory (Double Data Rate SDRAM, DDRSDRAM), enhanced synchronous dynamic random access memory (Enhanced SDRAM, ESDRAM), synchronous connection dynamic random access memory (Synch link DRAM , SLDRAM) and Direct Memory Bus Random Access Memory (Direct Rambus RAM, DRRAM). The memory 709 in the embodiment of the present application includes but is not limited to these and any other suitable types of memory.
处理器710可包括一个或多个处理单元;可选地,处理器710集成应用处理器和调制解调处理器,其中,应用处理器主要处理涉及操作系统、用户界面和应用程序等的操作,调制解调处理器主要处理无线通信信号,如基带处理器。可以理解的是,上述调制解调处理器也可以不集成到处理器710中。The processor 710 may include one or more processing units; optionally, the processor 710 integrates an application processor and a modem processor, wherein the application processor mainly handles operations related to the operating system, user interface, and application programs, etc., Modem processors mainly process wireless communication signals, such as baseband processors. It can be understood that the foregoing modem processor may not be integrated into the processor 710 .
在本申请实施例的一种实施方式中,终端700为第一端。其中,处理器710用于:对AI网络信息进行压缩,所述AI网络信息包括网络结构和网络参数中的至少一项;In an implementation manner of the embodiment of this application, the terminal 700 is the first end. Wherein, the processor 710 is configured to: compress the AI network information, where the AI network information includes at least one of network structure and network parameters;
射频单元701用于:向第二端发送压缩后的所述AI网络信息。The radio frequency unit 701 is configured to: send the compressed AI network information to the second end.
可选地,所述AI网络信息包括网络结构和网络参数,所述处理器710用于执行如下任意一项:Optionally, the AI network information includes network structure and network parameters, and the processor 710 is configured to perform any of the following:
对所述网络结构和所述网络参数进行合并压缩;Combining and compressing the network structure and the network parameters;
对所述网络结构和所述网络参数分别进行压缩。Compressing the network structure and the network parameters respectively.
可选地,在所述处理器710用于对所述网络结构和所述网络参数分别进行压缩的情况下,所述射频单元701还用于执行如下任意一项:Optionally, when the processor 710 is configured to compress the network structure and the network parameters respectively, the radio frequency unit 701 is further configured to perform any of the following:
向第二端合并发送压缩后的网络结构和压缩后的网络参数;Combining and sending the compressed network structure and compressed network parameters to the second end;
向第二端分别发送压缩后的网络结构和压缩后的网络参数。The compressed network structure and the compressed network parameters are respectively sent to the second end.
可选地,所述处理器710用于执行如下任意一项:Optionally, the processor 710 is configured to perform any of the following:
基于预设的模型表述方式将AI网络信息转换成对应的传输文件,对所述传输文件进行压缩;Convert the AI network information into a corresponding transmission file based on the preset model expression method, and compress the transmission file;
基于预设的数据格式对所述AI网络信息进行压缩;Compressing the AI network information based on a preset data format;
获取待发送的AI网络信息和所述第二端已有的AI网络信息,对所述待发送的AI网络信息与所述第二端已有的AI网络信息之间的AI网络信息差值进行压缩;Acquiring the AI network information to be sent and the existing AI network information of the second end, and performing an AI network information difference between the AI network information to be sent and the existing AI network information of the second end compression;
获取待发送的AI网络信息和预设AI网络的AI网络信息,对所述待发送的AI网络信息与所述预设AI网络的AI网络信息之间的AI网络信息差值进行压缩。Acquiring the AI network information to be sent and the AI network information of the preset AI network, and compressing the AI network information difference between the AI network information to be sent and the AI network information of the preset AI network.
可选地,所述AI网络信息差值包括如下至少一项:Optionally, the AI network information difference includes at least one of the following:
指定的网络参数;specified network parameters;
网络参数的索引;index of network parameters;
修改的网络参数;Modified network parameters;
修改的网络参数中修改的参数值;The modified parameter value in the modified network parameter;
修改的网络参数中修改的参数值的位置;The position of the modified parameter value in the modified network parameter;
修改的网络参数中修改的参考值的位置,所述参考值为所述网络参数中的最大值;the location of the modified reference value in the modified network parameters, the reference value being the maximum value in the network parameters;
修改的网络参数中的非零值;A non-zero value in the modified network parameter;
修改的网络参数中的非零值的位置;the location of non-zero values in the modified network parameters;
新增的网络结构;Newly added network structure;
删除的网络结构;Deleted network structure;
修改的网络结构。Modified network structure.
可选地,所述预设的模型表述方式包括如下任意一项:协议约定的模型表述方式、自定义的模型表述方式。Optionally, the preset model expression manner includes any one of the following: a protocol-agreed model expression manner, and a user-defined model expression manner.
可选地,所述自定义的模型表述方式的内容包括如下至少一项:AI网络的网络结构、AI网络的网络参数的属性、AI网络的网络参数的数值。Optionally, the content of the self-defined model expression includes at least one of the following: the network structure of the AI network, the attributes of the network parameters of the AI network, and the values of the network parameters of the AI network.
可选地,所述预设的模型表述方式中网络结构的表述方式包括如下至少一项:Optionally, the representation of the network structure in the preset model representation includes at least one of the following:
AI网络的网络结构之间的关联关系;The relationship between the network structures of the AI network;
AI网络的网络参数的属性;Attributes of the network parameters of the AI network;
AI网络的网络参数中非零值的位置;The location of non-zero values in the network parameters of the AI network;
AI网络的网络参数中的更新数值位置。The updated numerical position in the network parameters of the AI network.
可选地,所述处理器还710用于:Optionally, the processor 710 is further configured to:
基于至少一个预设的模型表述方式将AI网络信息转换成至少一个传输文件,一个所述预设的模型表述方式对应至少一个传输文件;Converting the AI network information into at least one transmission file based on at least one preset model representation, where one preset model representation corresponds to at least one transmission file;
对所述至少一个传输文件进行合并压缩,或者,对所述至少一个传输文件分别进行压缩后再合并。Combining and compressing the at least one transmission file, or compressing and merging the at least one transmission file respectively.
可选地,所述AI网络信息包括网络结构和网络参数,所述处理器710还用于:Optionally, the AI network information includes network structure and network parameters, and the processor 710 is further configured to:
基于预设的模型表述方式将所述网络结构转换成第一传输文件,并基于所述预设的模型表述方式将所述网络参数转换成第二传输文件,将所述第一传输文件和所述第二传输文件分别进行压缩。Converting the network structure into a first transmission file based on a preset model representation, converting the network parameters into a second transmission file based on the preset model representation, and converting the first transmission file and the The above-mentioned second transmission files are respectively compressed.
可选地,所述压缩后的AI网络信息包括压缩后的网络参数,所述射频单元701还用于:Optionally, the compressed AI network information includes compressed network parameters, and the radio frequency unit 701 is further configured to:
基于所述压缩后的网络参数的优先级顺序,按照所述优先级顺序向第二端发送所述压缩后的网络参数。Based on the priority order of the compressed network parameters, send the compressed network parameters to the second end according to the priority order.
可选地,所述射频单元701还用于:Optionally, the radio frequency unit 701 is also used for:
基于所述压缩后的网络参数的优先级顺序,对所述压缩后的网络参数进行分组;grouping the compressed network parameters based on the priority order of the compressed network parameters;
在传输资源小于预设阈值的情况下,按照预设顺序对分组后的网络参数进行丢弃并发送剩余的网络参数,所述预设顺序为分组后的网络参数的优先级从低至高的顺序。When the transmission resource is less than the preset threshold, the grouped network parameters are discarded and the remaining network parameters are sent according to a preset order, the preset order being the order of priority of the grouped network parameters from low to high.
可选地,所述射频单元701还用于:Optionally, the radio frequency unit 701 is also used for:
接收所述第二端发送的第一请求信息,所述第一请求信息用于请求获取目标AI网络信息;receiving first request information sent by the second end, where the first request information is used to request acquisition of target AI network information;
所述处理器710还用于:The processor 710 is also used for:
对所述目标AI网络信息进行压缩。Compressing the target AI network information.
可选地,所述第一请求信息包括如下至少一项:Optionally, the first request information includes at least one of the following:
请求的网络参数的名称;the name of the requested network parameter;
请求的网络参数的标识;Identification of the requested network parameters;
网络结构更新请求;network structure update request;
网络参数更新请求;Network parameter update request;
AI网络的网络效果度量值。Network Effect Measures for AI Networks.
可选地,所述处理器710还用于:Optionally, the processor 710 is further configured to:
判断是否需要对所述目标AI网络信息进行更新;Judging whether it is necessary to update the target AI network information;
在判定需要对所述目标AI网络信息进行更新的情况下,更新所述目标AI网络信息;When it is determined that the target AI network information needs to be updated, update the target AI network information;
对更新后的所述目标AI网络信息进行压缩。Compressing the updated target AI network information.
可选地,所述目标AI网络信息包括第一目标网络参数,所述处理器710还用于:Optionally, the target AI network information includes first target network parameters, and the processor 710 is further configured to:
基于预设的模型表述方式将所述第一目标网络参数的属性和参数值转换成预设格式后进行压缩;converting the attributes and parameter values of the first target network parameters into a preset format based on a preset model expression manner, and then compressing;
其中,所述第一目标网络参数的属性包括名称、维度、长度中的至少一项。Wherein, the attribute of the first target network parameter includes at least one of name, dimension, and length.
在本申请实施例的另一种实现方式中,所述终端700为第二端。其中,所述射频单元701还用于:接收第一端发送的压缩后的AI网络信息,所述AI网络信息包括网络结构和网络参数中的至少一项。In another implementation manner of the embodiment of this application, the terminal 700 is a second terminal. Wherein, the radio frequency unit 701 is further configured to: receive compressed AI network information sent by the first end, where the AI network information includes at least one of network structure and network parameters.
可选地,所述射频单元701还用于:Optionally, the radio frequency unit 701 is also used for:
向所述第一端发送第一请求信息,所述第一请求信息用于请求获取目标AI网络信息;Sending first request information to the first end, where the first request information is used to request acquisition of target AI network information;
接收所述第一端发送的压缩后的所述目标AI网络信息。Receive the compressed target AI network information sent by the first end.
可选地,所述第一请求信息包括如下至少一项:Optionally, the first request information includes at least one of the following:
请求的网络参数的名称;the name of the requested network parameter;
请求的网络参数的标识;Identification of the requested network parameters;
网络结构更新请求;network structure update request;
网络参数更新请求;Network parameter update request;
AI网络的网络效果度量值。Network Effect Measures for AI Networks.
本申请提供的技术方案,使得AI网络的网络结构和网络参数可以分开发送,进而能够有效降低通信过程中的传输开销。The technical solution provided by this application enables the network structure and network parameters of the AI network to be sent separately, thereby effectively reducing the transmission overhead in the communication process.
本申请实施例还提供一种网络侧设备,上述图2和图3所述方法实施例的各个实施过程和实现方式均可适用于该网络侧设备实施例中,且能达到相同的技术效果。The embodiment of the present application also provides a network-side device. The various implementation processes and implementation methods of the above-mentioned method embodiments shown in FIG. 2 and FIG. 3 can be applied to the network-side device embodiment, and can achieve the same technical effect.
具体地,本申请实施例还提供了一种网络侧设备。如图8所示,该网络侧设备800包括:天线81、射频装置82、基带装置83、处理器84和存储器85。天线81与射频装置82连接。在上行方向上,射频装置82通过天线81接收信息,将接收的信息发送给基带装置83进行处理。在下行方向上,基带装置83对要发送的信息进行处理,并发送给射频装置82,射频装置82对收到的信息进行处理后经过天线81发送出去。Specifically, the embodiment of the present application also provides a network side device. As shown in FIG. 8 , the network side device 800 includes: an antenna 81 , a radio frequency device 82 , a baseband device 83 , a processor 84 and a memory 85 . The antenna 81 is connected to a radio frequency device 82 . In the uplink direction, the radio frequency device 82 receives information through the antenna 81, and sends the received information to the baseband device 83 for processing. In the downlink direction, the baseband device 83 processes the information to be sent and sends it to the radio frequency device 82 , and the radio frequency device 82 processes the received information and sends it out through the antenna 81 .
以上实施例中网络侧设备执行的方法可以在基带装置83中实现,该基带装置83包括基带处理器。The method performed by the network side device in the above embodiments may be implemented in the baseband device 83, where the baseband device 83 includes a baseband processor.
基带装置83例如可以包括至少一个基带板,该基带板上设置有多个芯片,如图8所示,其中一个芯片例如为基带处理器,通过总线接口与存储器85连接,以调用存储器85中的程序,执行以上方法实施例中所示的网络设备操作。The baseband device 83 can include at least one baseband board, for example, a plurality of chips are arranged on the baseband board, as shown in FIG. The program executes the network device operations shown in the above method embodiments.
该网络侧设备还可以包括网络接口86,该接口例如为通用公共无线接口(common public radio interface,CPRI)。The network side device may also include a network interface 86, such as a common public radio interface (common public radio interface, CPRI).
具体地,本公开实施例的网络侧设备800还包括:存储在存储器85上并可在处理器84上运行的指令或程序,处理器84调用存储器85中的指令或程序执行图4或图5所示各模块执行的方法,并达到相同的技术效果,为避免重复,故不在此赘述。Specifically, the network-side device 800 in this embodiment of the present disclosure further includes: instructions or programs stored in the memory 85 and operable on the processor 84, and the processor 84 calls the instructions or programs in the memory 85 to execute FIG. 4 or FIG. 5 The methods executed by each module shown in the figure achieve the same technical effect, so in order to avoid repetition, they are not repeated here.
具体地,本申请实施例还提供了另一种网络侧设备。如图9所示,该网络侧设备900包括:处理器901、网络接口902和存储器903。其中,网络接口902例如为通用公共无线接口(common public radio interface,CPRI)。Specifically, the embodiment of the present application also provides another network side device. As shown in FIG. 9 , the network side device 900 includes: a processor 901 , a network interface 902 and a memory 903 . Wherein, the network interface 902 is, for example, a common public radio interface (common public radio interface, CPRI).
具体地,本公开实施例的网络侧设备900还包括:存储在存储器903上并可在处理器901上运行的指令或程序,处理器901调用存储器903中的指令或程序执行图4或图5所示各模块执行的方法,并达到相同的技术效果, 为避免重复,故不在此赘述。Specifically, the network-side device 900 in this embodiment of the present disclosure further includes: instructions or programs stored in the memory 903 and executable on the processor 901, and the processor 901 calls the instructions or programs in the memory 903 to execute FIG. 4 or FIG. 5 The methods executed by each module shown in the figure achieve the same technical effect, so in order to avoid repetition, they are not repeated here.
本申请实施例还提供一种可读存储介质,所述可读存储介质上存储有程序或指令,该程序或指令被处理器执行时实现上述图2或图3所述方法实施例的各个过程,且能达到相同的技术效果,为避免重复,这里不再赘述。The embodiment of the present application also provides a readable storage medium, on which a program or instruction is stored, and when the program or instruction is executed by the processor, each process of the method embodiment described above in FIG. 2 or FIG. 3 is implemented. , and can achieve the same technical effect, in order to avoid repetition, it will not be repeated here.
其中,所述处理器为上述实施例中所述的终端中的处理器。所述可读存储介质,包括计算机可读存储介质,如计算机只读存储器ROM、随机存取存储器RAM、磁碟或者光盘等。Wherein, the processor is the processor in the terminal described in the foregoing embodiments. The readable storage medium includes a computer-readable storage medium, such as a computer read-only memory ROM, a random access memory RAM, a magnetic disk or an optical disk, and the like.
本申请实施例另提供了一种芯片,所述芯片包括处理器和通信接口,所述通信接口和所述处理器耦合,所述处理器用于运行程序或指令,实现上述图2或图3所述方法实施例的各个过程,且能达到相同的技术效果,为避免重复,这里不再赘述。The embodiment of the present application further provides a chip, the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is used to run programs or instructions to implement the above-mentioned Figure 2 or Figure 3. Each process of the method embodiment described above can achieve the same technical effect, so in order to avoid repetition, details are not repeated here.
应理解,本申请实施例提到的芯片还可以称为系统级芯片,系统芯片,芯片系统或片上系统芯片等。It should be understood that the chip mentioned in the embodiment of the present application may also be called a system-on-chip, a system-on-chip, a system-on-a-chip, or a system-on-a-chip.
本申请实施例另提供了一种计算机程序/程序产品,所述计算机程序/程序产品被存储在存储介质中,所述计算机程序/程序产品被至少一个处理器执行以实现上述图2或图3所述方法实施例的各个过程,且能达到相同的技术效果,为避免重复,这里不再赘述。The embodiment of the present application further provides a computer program/program product, the computer program/program product is stored in a storage medium, and the computer program/program product is executed by at least one processor to implement the above-mentioned Figure 2 or Figure 3 The various processes of the method embodiments can achieve the same technical effect, and are not repeated here to avoid repetition.
本申请实施例还提供了一种通信系统,包括:终端及网络侧设备,所述终端可用于执行如上述图2所述方法的步骤,所述网络侧设备可用于执行如上图3所述方法的步骤;或者,所述终端可用于执行如上述图3所述方法的步骤,所述网络侧设备可用于执行如上图2所述方法的步骤。The embodiment of the present application also provides a communication system, including: a terminal and a network-side device, the terminal can be used to perform the steps of the method described in Figure 2 above, and the network-side device can be used to perform the method described in Figure 3 above or, the terminal may be used to perform the steps of the method described in FIG. 3 above, and the network side device may be used to perform the steps of the method described in FIG. 2 above.
需要说明的是,在本文中,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者装置不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者装置所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括该要素的过程、方法、物品或者装置中还存在另外的相同要素。此外,需要指出的是,本申请实施方式中的方法和装置的范围不限按示出或讨论的顺序来执行功能,还可包括根据所涉及的功能按基本同时的方式或按相反的顺序来执行功能,例 如,可以按不同于所描述的次序来执行所描述的方法,并且还可以添加、省去、或组合各种步骤。另外,参照某些示例所描述的特征可在其他示例中被组合。It should be noted that, in this document, the term "comprising", "comprising" or any other variation thereof is intended to cover a non-exclusive inclusion such that a process, method, article or apparatus comprising a set of elements includes not only those elements, It also includes other elements not expressly listed, or elements inherent in the process, method, article, or device. Without further limitations, an element defined by the phrase "comprising a ..." does not preclude the presence of additional identical elements in the process, method, article, or apparatus comprising that element. In addition, it should be pointed out that the scope of the methods and devices in the embodiments of the present application is not limited to performing functions in the order shown or discussed, and may also include performing functions in a substantially simultaneous manner or in reverse order according to the functions involved. Functions are performed, for example, the described methods may be performed in an order different from that described, and various steps may also be added, omitted, or combined. Additionally, features described with reference to certain examples may be combined in other examples.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。基于这样的理解,本申请的技术方案本质上或者说对相关技术做出贡献的部分可以以计算机软件产品的形式体现出来,该计算机软件产品存储在一个存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台终端(可以是手机,计算机,服务器,空调器,或者网络设备等)执行本申请各个实施例所述的方法。Through the description of the above embodiments, those skilled in the art can clearly understand that the methods of the above embodiments can be implemented by means of software plus a necessary general-purpose hardware platform, and of course also by hardware, but in many cases the former is better implementation. Based on this understanding, the essence of the technical solution of this application or the part that contributes to related technologies can be embodied in the form of computer software products, which are stored in a storage medium (such as ROM/RAM, disk, CD) contains several instructions to enable a terminal (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) to execute the methods described in various embodiments of the present application.
上面结合附图对本申请的实施例进行了描述,但是本申请并不局限于上述的具体实施方式,上述的具体实施方式仅仅是示意性的,而不是限制性的,本领域的普通技术人员在本申请的启示下,在不脱离本申请宗旨和权利要求所保护的范围情况下,还可做出很多形式,均属于本申请的保护之内。The embodiments of the present application have been described above in conjunction with the accompanying drawings, but the present application is not limited to the above-mentioned specific implementations. The above-mentioned specific implementations are only illustrative and not restrictive. Those of ordinary skill in the art will Under the inspiration of this application, without departing from the purpose of this application and the scope of protection of the claims, many forms can also be made, all of which belong to the protection of this application.
Claims (26)
- 一种人工智能AI网络信息传输方法,包括:An artificial intelligence AI network information transmission method, comprising:第一端对AI网络信息进行压缩,所述AI网络信息包括网络结构和网络参数中的至少一项;The first end compresses the AI network information, and the AI network information includes at least one of network structure and network parameters;所述第一端向第二端发送压缩后的所述AI网络信息。The first end sends the compressed AI network information to the second end.
- 根据权利要求1所述的方法,其中,所述AI网络信息包括网络结构和网络参数,所述第一端对AI网络信息进行压缩,包括如下任意一项:The method according to claim 1, wherein the AI network information includes network structure and network parameters, and the first end compresses the AI network information, including any of the following:所述第一端对所述网络结构和所述网络参数进行合并压缩;The first end combines and compresses the network structure and the network parameters;所述第一端对所述网络结构和所述网络参数分别进行压缩。The first end compresses the network structure and the network parameters respectively.
- 根据权利要求2所述的方法,其中,在所述第一端对所述网络结构和所述网络参数分别进行压缩的情况下,所述第一端向第二端发送压缩后的所述AI网络信息,包括如下任意一项:The method according to claim 2, wherein, when the first end compresses the network structure and the network parameters respectively, the first end sends the compressed AI to the second end Network information, including any of the following:所述第一端向第二端合并发送压缩后的网络结构和压缩后的网络参数;The first end sends the compressed network structure and compressed network parameters to the second end in combination;所述第一端向第二端分别发送压缩后的网络结构和压缩后的网络参数。The first end sends the compressed network structure and the compressed network parameters to the second end respectively.
- 根据权利要求1所述的方法,其中,所述第一端对AI网络信息进行压缩,包括如下任意一项:The method according to claim 1, wherein the first end compresses the AI network information, including any of the following:所述第一端基于预设的模型表述方式将AI网络信息转换成对应的传输文件,对所述传输文件进行压缩;The first end converts the AI network information into a corresponding transmission file based on a preset model representation, and compresses the transmission file;所述第一端基于预设的数据格式对所述AI网络信息进行压缩;The first end compresses the AI network information based on a preset data format;所述第一端获取待发送的AI网络信息和所述第二端已有的AI网络信息,对所述待发送的AI网络信息与所述第二端已有的AI网络信息之间的AI网络信息差值进行压缩;The first end obtains the AI network information to be sent and the existing AI network information of the second end, and calculates the AI network information between the AI network information to be sent and the existing AI network information of the second end. Network information difference is compressed;所述第一端获取待发送的AI网络信息和预设AI网络的AI网络信息,对所述待发送的AI网络信息与所述预设AI网络的AI网络信息之间的AI网络信息差值进行压缩。The first end obtains the AI network information to be sent and the AI network information of the preset AI network, and calculates the AI network information difference between the AI network information to be sent and the AI network information of the preset AI network to compress.
- 根据权利要求4所述的方法,其中,所述AI网络信息差值包括如下至少一项:The method according to claim 4, wherein the AI network information difference includes at least one of the following:指定的网络参数;specified network parameters;网络参数的索引;index of network parameters;修改的网络参数;Modified network parameters;修改的网络参数中修改的参数值;The modified parameter value in the modified network parameter;修改的网络参数中修改的参数值的位置;The position of the modified parameter value in the modified network parameter;修改的网络参数中修改的参考值的位置,所述参考值为所述网络参数中的最大值;the location of the modified reference value in the modified network parameters, the reference value being the maximum value in the network parameters;修改的网络参数中的非零值;A non-zero value in the modified network parameter;修改的网络参数中的非零值的位置;the location of non-zero values in the modified network parameters;新增的网络结构;Newly added network structure;删除的网络结构;Deleted network structure;修改的网络结构。Modified network structure.
- 根据权利要求4所述的方法,其中,所述预设的模型表述方式包括如下任意一项:协议约定的模型表述方式、自定义的模型表述方式。The method according to claim 4, wherein, the preset model expression method includes any one of the following: a protocol-agreed model expression method, and a self-defined model expression method.
- 根据权利要求6所述的方法,其中,所述自定义的模型表述方式的内容包括如下至少一项:AI网络的网络结构、AI网络的网络参数的属性、AI网络的网络参数的数值。The method according to claim 6, wherein the content of the self-defined model expression includes at least one of the following: the network structure of the AI network, the attributes of the network parameters of the AI network, and the values of the network parameters of the AI network.
- 根据权利要求4所述的方法,其中,所述预设的模型表述方式中网络结构的表述方式包括如下至少一项:The method according to claim 4, wherein the expression of the network structure in the preset model expression includes at least one of the following:AI网络的网络结构之间的关联关系;The relationship between the network structures of the AI network;AI网络的网络参数的属性;Attributes of the network parameters of the AI network;AI网络的网络参数中非零值的位置;The location of non-zero values in the network parameters of the AI network;AI网络的网络参数中的更新数值位置。The updated numerical position in the network parameters of the AI network.
- 根据权利要求4所述的方法,其中,所述第一端基于预设的模型表述方式将AI网络信息转换成对应的传输文件,对所述传输文件进行压缩,包括:The method according to claim 4, wherein the first end converts the AI network information into a corresponding transmission file based on a preset model expression, and compresses the transmission file, including:所述第一端基于至少一个预设的模型表述方式将AI网络信息转换成至少一个传输文件,一个所述预设的模型表述方式对应至少一个传输文件;The first end converts the AI network information into at least one transmission file based on at least one preset model representation, and one preset model representation corresponds to at least one transmission file;所述第一端对所述至少一个传输文件进行合并压缩,或者,所述第一端对所述至少一个传输文件分别进行压缩后再合并。The first end merges and compresses the at least one transmission file, or the first end compresses the at least one transmission file separately and then merges them.
- 根据权利要求4所述的方法,其中,所述AI网络信息包括网络结构 和网络参数,所述第一端基于预设的模型表述方式将AI网络信息转换成对应的传输文件,对所述传输文件进行压缩,包括:The method according to claim 4, wherein the AI network information includes network structure and network parameters, and the first end converts the AI network information into a corresponding transmission file based on a preset model representation, and the transmission Files are compressed, including:所述第一端基于预设的模型表述方式将所述网络结构转换成第一传输文件,并基于所述预设的模型表述方式将所述网络参数转换成第二传输文件,将所述第一传输文件和所述第二传输文件分别进行压缩。The first end converts the network structure into a first transmission file based on a preset model representation, converts the network parameters into a second transmission file based on the preset model representation, and converts the first The first transmission file and the second transmission file are respectively compressed.
- 根据权利要求1-10中任一项所述的方法,其中,所述压缩后的AI网络信息包括压缩后的网络参数,所述第一端向第二端发送压缩后的所述AI网络信息,包括:The method according to any one of claims 1-10, wherein the compressed AI network information includes compressed network parameters, and the first terminal sends the compressed AI network information to the second terminal ,include:所述第一端基于所述压缩后的网络参数的优先级顺序,按照所述优先级顺序向第二端发送所述压缩后的网络参数。The first end sends the compressed network parameters to the second end according to the priority order of the compressed network parameters based on the priority order of the compressed network parameters.
- 根据权利要求11所述的方法,其中,所述第一端基于所述压缩后的网络参数的优先级顺序,按照所述优先级顺序向第二端发送所述压缩后的网络参数,包括:The method according to claim 11, wherein the first end sends the compressed network parameters to the second end according to the priority order based on the priority order of the compressed network parameters, comprising:所述第一端基于所述压缩后的网络参数的优先级顺序,对所述压缩后的网络参数进行分组;grouping the compressed network parameters by the first end based on the priority order of the compressed network parameters;在传输资源小于预设阈值的情况下,所述第一端按照预设顺序对分组后的网络参数进行丢弃并发送剩余的网络参数,所述预设顺序为分组后的网络参数的优先级从低至高的顺序。When the transmission resource is less than the preset threshold, the first end discards the grouped network parameters according to a preset order and sends the remaining network parameters, and the preset order is that the priority of the grouped network parameters starts from Lowest to highest order.
- 根据权利要求1-10中任一项所述的方法,其中,所述第一端对AI网络信息进行压缩之前,所述方法还包括:The method according to any one of claims 1-10, wherein, before the first end compresses the AI network information, the method further includes:所述第一端接收所述第二端发送的第一请求信息,所述第一请求信息用于请求获取目标AI网络信息;The first end receives first request information sent by the second end, and the first request information is used to request acquisition of target AI network information;所述第一端对AI网络信息进行压缩,包括:The first end compresses the AI network information, including:所述第一端对所述目标AI网络信息进行压缩。The first end compresses the target AI network information.
- 根据权利要求13所述的方法,其中,所述第一请求信息包括如下至少一项:The method according to claim 13, wherein the first request information includes at least one of the following:请求的网络参数的名称;the name of the requested network parameter;请求的网络参数的标识;Identification of the requested network parameters;网络结构更新请求;network structure update request;网络参数更新请求;Network parameter update request;AI网络的网络效果度量值。Network Effect Measures for AI Networks.
- 根据权利要求13所述的方法,其中,所述第一端对所述目标AI网络信息进行压缩之前,所述方法还包括:The method according to claim 13, wherein, before the first end compresses the target AI network information, the method further comprises:所述第一端判断是否需要对所述目标AI网络信息进行更新;The first end judges whether the target AI network information needs to be updated;在判定需要对所述目标AI网络信息进行更新的情况下,更新所述目标AI网络信息;When it is determined that the target AI network information needs to be updated, update the target AI network information;所述第一端对所述目标AI网络信息进行压缩,包括:The first end compresses the target AI network information, including:所述第一端对更新后的所述目标AI网络信息进行压缩。The first end compresses the updated target AI network information.
- 根据权利要求13所述的方法,其中,所述目标AI网络信息包括第一目标网络参数,所述第一端对所述目标AI网络信息进行压缩,包括:The method according to claim 13, wherein the target AI network information includes first target network parameters, and the first end compresses the target AI network information, including:所述第一端基于预设的模型表述方式将所述第一目标网络参数的属性和参数值转换成预设格式后进行压缩;The first end converts the attributes and parameter values of the first target network parameters into a preset format based on a preset model representation and then compresses them;其中,所述第一目标网络参数的属性包括名称、维度、长度中的至少一项。Wherein, the attribute of the first target network parameter includes at least one of name, dimension, and length.
- 根据权利要求1-10中任一项所述的方法,其中,所述第一端为网络侧设备和终端中的一者,所述第二端为网络侧设备和终端中的另一者;或者,The method according to any one of claims 1-10, wherein the first end is one of a network-side device and a terminal, and the second end is the other of a network-side device and a terminal; or,所述第一端和所述第二端为终端的不同节点;或者,The first end and the second end are different nodes of a terminal; or,所述第一端和所述第二端为网络侧设备的不同节点。The first end and the second end are different nodes of the network side equipment.
- 根据权利要求1-10中任一项所述的方法,其中,所述第一端为网络侧设备,所述第二端为终端,所述AI网络信息包括网络参数,在所述第二端从第一小区切换到第二小区的情况下,所述第一端对AI网络信息进行压缩之前,所述方法还包括:The method according to any one of claims 1-10, wherein the first end is a network-side device, the second end is a terminal, the AI network information includes network parameters, and at the second end In the case of switching from the first cell to the second cell, before the first end compresses the AI network information, the method further includes:所述第一端计算所述第一小区的网络参数与所述第二小区的网络参数的相关性,获取第二目标网络参数,所述第二目标网络参数包括如下至少一项:相关性小于预设阈值的网络参数、预设序列中的前N个网络参数,所述预设序列为网络参数的相关性按照从小到大顺序排列的序列;The first end calculates the correlation between the network parameters of the first cell and the network parameters of the second cell, and acquires a second target network parameter, where the second target network parameter includes at least one of the following items: the correlation is less than Network parameters with preset thresholds and the first N network parameters in a preset sequence, where the preset sequence is a sequence in which the correlation of network parameters is arranged in ascending order;所述第一端对AI网络信息进行压缩,包括:The first end compresses the AI network information, including:所述第一端对所述第二目标网络参数进行压缩;The first end compresses the second target network parameters;所述第一端向第二端发送压缩后的所述AI网络信息,包括:The first end sends the compressed AI network information to the second end, including:所述第一端向第二端发送压缩后的所述第二目标网络参数。The first end sends the compressed second target network parameters to the second end.
- 一种AI网络信息传输方法,包括:An AI network information transmission method, comprising:第二端接收第一端发送的压缩后的AI网络信息,所述AI网络信息包括网络结构和网络参数中的至少一项。The second end receives the compressed AI network information sent by the first end, where the AI network information includes at least one of network structure and network parameters.
- 根据权利要求19所述的方法,其中,所述第二端接收第一端发送的压缩后的AI网络信息之前,所述方法还包括:The method according to claim 19, wherein, before the second end receives the compressed AI network information sent by the first end, the method further comprises:所述第二端向所述第一端发送第一请求信息,所述第一请求信息用于请求获取目标AI网络信息;The second terminal sends first request information to the first terminal, where the first request information is used to request acquisition of target AI network information;所述第二端接收第一端发送的压缩后的AI网络信息,包括:The second end receives the compressed AI network information sent by the first end, including:所述第二端接收所述第一端发送的压缩后的所述目标AI网络信息。The second end receives the compressed target AI network information sent by the first end.
- 根据权利要求20所述的方法,其中,所述第一请求信息包括如下至少一项:The method according to claim 20, wherein the first request information includes at least one of the following:请求的网络参数的名称;the name of the requested network parameter;请求的网络参数的标识;Identification of the requested network parameters;网络结构更新请求;network structure update request;网络参数更新请求;Network parameter update request;AI网络的网络效果度量值。Network Effect Measures for AI Networks.
- 根据权利要求19所述的方法,其中,所述第一端为网络侧设备和终端中的一者,所述第二端为网络侧设备和终端中的另一者;或者,The method according to claim 19, wherein the first end is one of the network-side device and the terminal, and the second end is the other of the network-side device and the terminal; or,所述第一端和所述第二端为终端的不同节点;或者,The first end and the second end are different nodes of a terminal; or,所述第一端和所述第二端为网络侧设备的不同节点。The first end and the second end are different nodes of the network side equipment.
- 一种AI网络信息传输装置,包括:An AI network information transmission device, comprising:压缩模块,用于对AI网络信息进行压缩,所述AI网络信息包括网络结构和网络参数中的至少一项;A compression module, configured to compress AI network information, where the AI network information includes at least one of network structure and network parameters;发送模块,用于向第二端发送压缩后的所述AI网络信息。A sending module, configured to send the compressed AI network information to the second end.
- 一种AI网络信息传输装置,包括:An AI network information transmission device, comprising:接收模块,用于接收第一端发送的压缩后的AI网络信息,所述AI网络信息包括网络结构和网络参数中的至少一项。The receiving module is configured to receive the compressed AI network information sent by the first end, where the AI network information includes at least one of network structure and network parameters.
- 一种通信设备,包括处理器和存储器,所述存储器存储可在所述处理器上运行的程序或指令,所述程序或指令被所述处理器执行时实现如权利要求1-18中任一项所述的AI网络信息传输方法的步骤,或者实现如权利要求19-22中任一项所述的AI网络信息传输方法的步骤。A communication device, comprising a processor and a memory, the memory stores programs or instructions that can run on the processor, and when the programs or instructions are executed by the processor, any one of claims 1-18 is implemented The steps of the AI network information transmission method described in claim 1, or the steps of the AI network information transmission method described in any one of claims 19-22.
- 一种可读存储介质,所述可读存储介质上存储程序或指令,所述程序或指令被处理器执行时实现如权利要求1-18中任一项所述的AI网络信息传输方法的步骤,或者实现如权利要求19-22中任一项所述的AI网络信息传输方法的步骤。A readable storage medium, on which a program or instruction is stored, and when the program or instruction is executed by a processor, the steps of the AI network information transmission method according to any one of claims 1-18 are realized , or realize the steps of the AI network information transmission method described in any one of claims 19-22.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111666710.4A CN116418880A (en) | 2021-12-31 | 2021-12-31 | AI network information transmission method and device and communication equipment |
CN202111666710.4 | 2021-12-31 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2023125934A1 true WO2023125934A1 (en) | 2023-07-06 |
Family
ID=86998170
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2022/143952 WO2023125934A1 (en) | 2021-12-31 | 2022-12-30 | Ai network information transmission method and apparatus, and communication device |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN116418880A (en) |
WO (1) | WO2023125934A1 (en) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108229644A (en) * | 2016-12-15 | 2018-06-29 | 上海寒武纪信息科技有限公司 | The device of compression/de-compression neural network model, device and method |
CN112445823A (en) * | 2019-09-04 | 2021-03-05 | 华为技术有限公司 | Searching method of neural network structure, image processing method and device |
CN112907309A (en) * | 2019-11-19 | 2021-06-04 | 阿里巴巴集团控股有限公司 | Model updating method, resource recommendation method, device, equipment and system |
CN113011575A (en) * | 2019-12-19 | 2021-06-22 | 华为技术有限公司 | Neural network model updating method, image processing method and device |
-
2021
- 2021-12-31 CN CN202111666710.4A patent/CN116418880A/en active Pending
-
2022
- 2022-12-30 WO PCT/CN2022/143952 patent/WO2023125934A1/en unknown
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108229644A (en) * | 2016-12-15 | 2018-06-29 | 上海寒武纪信息科技有限公司 | The device of compression/de-compression neural network model, device and method |
CN112445823A (en) * | 2019-09-04 | 2021-03-05 | 华为技术有限公司 | Searching method of neural network structure, image processing method and device |
CN112907309A (en) * | 2019-11-19 | 2021-06-04 | 阿里巴巴集团控股有限公司 | Model updating method, resource recommendation method, device, equipment and system |
CN113011575A (en) * | 2019-12-19 | 2021-06-22 | 华为技术有限公司 | Neural network model updating method, image processing method and device |
Also Published As
Publication number | Publication date |
---|---|
CN116418880A (en) | 2023-07-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20240322982A1 (en) | Method for reporting csi report, terminal device, and network device | |
WO2019095898A1 (en) | Signal transmission method, and terminal and network device | |
WO2024012235A1 (en) | Method, apparatus and device for analyzing artificial intelligence request | |
WO2023125934A1 (en) | Ai network information transmission method and apparatus, and communication device | |
WO2023280143A1 (en) | Ai task control method, terminal, base station, and storage medium | |
CN116567691A (en) | Information transmission method, device, terminal and network side equipment | |
WO2023125932A1 (en) | Ai network information transmission method and apparatus, and communication device | |
WO2023125933A1 (en) | Ai network information transmission method and apparatus, and communication device | |
WO2023088268A1 (en) | Ai information transmission method and device | |
WO2024088119A1 (en) | Data processing method and apparatus, and terminal and network-side device | |
WO2024149156A1 (en) | Information transmission method and apparatus, and terminal and network-side device | |
WO2023098535A1 (en) | Information interaction method and apparatus, and communication device | |
WO2023151585A1 (en) | Terminal target surface capability reporting and acquiring methods, terminal, and network device | |
WO2024149157A1 (en) | Csi transmission method and apparatus, terminal and network side device | |
WO2023185756A1 (en) | Information transmission methods and apparatus, terminal and network-side device | |
WO2024027618A1 (en) | Gap configuration method and apparatus, network side device and storage medium | |
WO2024027683A1 (en) | Model matching method and apparatus, communication device, and readable storage medium | |
WO2023143416A1 (en) | Information processing method, terminal, and network function | |
WO2023186091A1 (en) | Sample determination method, apparatus and device | |
WO2024032694A1 (en) | Csi prediction processing method and apparatus, communication device, and readable storage medium | |
WO2024078456A1 (en) | Uplink control information transmission method and apparatus, and terminal | |
WO2023125870A1 (en) | Communication method and apparatus, communication device, and readable storage medium | |
WO2023207947A1 (en) | Cell reconfiguration method, apparatus, terminal, and network side device | |
WO2023217141A1 (en) | Terminal positioning method, terminal, and network side device | |
WO2024027582A1 (en) | Model updating method and apparatus, communication device, and readable storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 22915182 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |