WO2024156232A1 - Communication method and device - Google Patents

Communication method and device Download PDF

Info

Publication number
WO2024156232A1
WO2024156232A1 PCT/CN2023/138554 CN2023138554W WO2024156232A1 WO 2024156232 A1 WO2024156232 A1 WO 2024156232A1 CN 2023138554 W CN2023138554 W CN 2023138554W WO 2024156232 A1 WO2024156232 A1 WO 2024156232A1
Authority
WO
WIPO (PCT)
Prior art keywords
model
information
reasoning
communication device
models
Prior art date
Application number
PCT/CN2023/138554
Other languages
French (fr)
Chinese (zh)
Inventor
黄谢田
曹龙雨
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Publication of WO2024156232A1 publication Critical patent/WO2024156232A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/04Inference or reasoning models
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/14Network analysis or design
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/16Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks using machine learning or artificial intelligence

Definitions

  • the present application relates to the field of communication technology, and in particular to a communication method and device.
  • AI artificial intelligence
  • ML machine learning
  • CN core network
  • RAN radio access network
  • the current solution proposes that the model training functional network element (or entity) provides multiple models with different performances to the model reasoning functional network element (or entity), and the model reasoning functional network element (or entity) selects a suitable model for reasoning.
  • the model reasoning functional network element (or entity) cannot determine the training capability of the model training functional network element (or entity). If multiple models are requested from the model training functional network element (or entity), the request may fail.
  • the present application proposes a communication method and device that can effectively improve the use effect of the model to ensure the performance of intelligent reasoning (or analysis).
  • the present application provides a communication method, which can be executed by a first communication device, or by a component of the first communication device (such as a processor, a chip, or a chip system, etc.), and the present application does not specifically limit this.
  • the method may specifically include the following steps: the first communication device receives model request information, and the model request information includes reasoning requirement information; the first communication device determines a first model according to the model request information, and the first model is a multi-model; the first communication device sends a first message, and the first message includes information about the first model.
  • the first communication device is regarded as the training party of the model, and the first communication device may be, but is not limited to: a model training function network element, or a model training function entity, or a communication device including a model training function.
  • the first communication device may be a NWDAF network element including a model training function module, or a network element management system (EMS) device, or an access network device (such as a base station), etc.
  • EMS network element management system
  • the first communication device receives model request information, which includes reasoning requirement information.
  • the first communication device determines a more suitable multi-model (i.e., the first model) based on the reasoning requirement information in the model request information, and then sends the information of the first model through the first information.
  • the reasoning end i.e., the second communication device including the model reasoning function module
  • receives the information of the first model based on the information of the first model
  • the reasoning and combination of multiple models using the first model can obtain a more accurate reasoning result. Therefore, this method can effectively improve the use effect of the model to ensure the performance of intelligent reasoning.
  • the method before the first communication device receives the model request information, the method further includes: the first communication device sends training capability indication information, where the training capability indication information is used to indicate that the first communication device supports multi-model training.
  • the first communication device can effectively perform multi-model training after receiving the model request information.
  • the first communication device Information determining the first model, can include the following implementation methods:
  • Implementation method one the model request information also includes a multi-model training strategy; the first communication device performs training according to the inference requirement information and the multi-model training strategy to obtain multiple sub-models of the first model.
  • the first communication device determines a multi-model training strategy based on the reasoning requirement information; and performs training based on the reasoning requirement information and the multi-model training strategy to obtain multiple sub-models of the first model.
  • the multi-model training strategy includes one or more of the following: data processing strategy, training algorithm, training mode, number of sub-models, and type of sub-models.
  • the first communication device can effectively train multiple sub-models of the first model based on the model request information.
  • the first communication device determines the first model according to the model request information, including: the first communication device determines the first model from at least one preset multiple model according to the inference requirement information.
  • the inference requirement information includes one or more of the following: the type of inference, the performance requirement of inference, the speed requirement of inference, and the power consumption requirement of inference.
  • the first communication device can directly and quickly select the first model from at least one trained multiple models based on the inference requirement information.
  • the model request information also includes multi-model indication information, and the multi-model indication information is used to indicate that the model requested to be trained or obtained is a multi-model.
  • the multi-model indication information may also be carried in the inference requirement information included in the model request information, or the multi-model indication information may be sent separately to the first communication device, which is not specifically limited in the embodiments of the present application.
  • the first communication device can effectively and accurately train or provide multiple models for the second communication device.
  • the information of the first model includes model information of the first model and information of multiple sub-models of the first model, and the information of each sub-model includes one or more of the following: identification information of the sub-model, the level of the sub-model, the performance of the sub-model, and performance constraints; the multiple sub-models include multiple first-level sub-models and one second-level sub-model, and the second-level sub-model is used to aggregate the reasoning information of the multiple first-level sub-models; or the multiple sub-models are all first-level sub-models, and the information of the first model also includes aggregation method and/or weight information.
  • the information of multiple sub-models of the first model and the performance information of the sub-models, as well as the combination method between the reasoning information of multiple sub-models can be effectively determined.
  • the method further includes: the first communication device sending reasoning performance information of the first model, where the reasoning performance information of the first model includes one or more of the following:
  • the performance of the first model the size information of the first model, the power consumption of the reasoning of the first model, the reasoning speed of the first model, and the computing power of the first model.
  • the receiving end (such as a third communication device including a model management function network element) that receives the reasoning performance information of the first model can also effectively adjust the first model based on the reasoning performance information of the first model. For example, based on the reasoning performance information of the first model and the actual reasoning requirement information, the number of sub-models of the first model can be appropriately reduced.
  • the present application provides a communication method, which can be executed by a second communication device or by a component of the second communication device (such as a processor, a chip, or a chip system, etc.), and the present application does not specifically limit this.
  • the method may specifically include the following steps: the second communication device receives second information, the second information includes information of a first model, the first model is determined according to reasoning requirement information, and the first model is a multi-model; the second communication device obtains reasoning information of the first model based on the information of the first model.
  • the second communication device serves as a model reasoning party, and the second communication device may be the following but is not limited to: a model reasoning function network element, or a model reasoning function entity, or a communication device including a model reasoning function.
  • the second communication device may be a NWDAF network element including a model reasoning function module, or a network element management system (EMS) device, or an access network device (such as a base station), etc.
  • EMS network element management system
  • the second communication device receives the information of the first model. Since the first model is a multi-model determined according to the reasoning requirement information, the second communication device uses the information of the first model and the first model to perform reasoning and combination of the multi-models to obtain a reasoning result with higher accuracy. Therefore, this method can effectively improve the use effect of the model to ensure the performance of intelligent reasoning.
  • the method before the second communication device receives the second information, the method further includes: the second communication device sends reasoning capability information and the reasoning requirement information; the reasoning capability information includes reasoning capability indication information, and one or more of the following: reasoning computing power and storage space; the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning; the reasoning requirement information includes one or more of the following: the type of reasoning, the performance requirement of reasoning, the speed requirement of reasoning, and the power consumption requirement of reasoning.
  • the second communication device sends its own reasoning capability information and reasoning requirement information, which can not only ensure that the second communication device can effectively perform multi-model reasoning in the future, but also ensure the performance of the second communication device in performing reasoning based on multiple models in the future.
  • the information of the first model includes model information of the first model and information of multiple sub-models of the first model, and the information of each sub-model includes one or more of the following: identification information of the sub-model, the level of the sub-model, the performance of the sub-model, and performance constraints.
  • the second communication device can accurately obtain information about multiple sub-models of the first model, so as to effectively use these sub-models for reasoning later.
  • the multiple sub-models include multiple first-level sub-models and one second-level sub-model, and the second-level sub-model is used to aggregate the reasoning information of the multiple first-level sub-models;
  • the second communication device obtains the reasoning information of the first model based on the information of the first model, including: the second communication device uses the multiple first-level sub-models to perform reasoning respectively based on the information of the multiple first-level sub-models to obtain the reasoning information of the multiple first-level sub-models;
  • the second communication device uses the second-level sub-model to aggregate the reasoning information of the multiple first-level sub-models to obtain the reasoning information of the first model.
  • the second communication device can use multiple first-level sub-models of the first sub-model to perform reasoning respectively, and use the second-level sub-model to effectively combine the reasoning information of the multiple first-level sub-models, so as to obtain the reasoning information of the first model.
  • the multiple sub-models are all first-level sub-models, and the information of the first model also includes an aggregation method and/or weight information; the second communication device obtains reasoning information of the first model based on the information of the first model, including: the second communication device uses the multiple sub-models to perform reasoning respectively based on the information of the multiple sub-models to obtain the reasoning information of the multiple sub-models; the second communication device aggregates the reasoning information of the multiple sub-models according to the aggregation method and/or weight information to obtain the reasoning information of the first model.
  • the second communication device can also use multiple sub-models of the first sub-model to perform reasoning separately, and use specified aggregation methods and/or weight information to effectively combine the reasoning information of the multiple sub-models to obtain the reasoning information of the first model.
  • the present application provides a communication method, which can be executed by a third communication device or by a component of the third communication device (such as a processor, a chip, or a chip system, etc.), and the present application does not specifically limit this.
  • the method may specifically include the following steps: the third communication device receives training capability indication information of the first communication device; the training capability indication information is used to indicate that the first communication device supports multi-model training; the third communication device receives reasoning requirement information and reasoning capability information of the second communication device; the reasoning capability information includes reasoning capability indication information, and the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning; the third communication device sends model request information to the first communication device, and the model request information includes the reasoning requirement information; the third communication device receives first information from the first communication device, the first information includes information of the first model, the first model is a multi-model, and the first model is determined according to the reasoning requirement information; the third communication device sends second information to the second communication device, and the second information includes information of the first model.
  • the third communication device serves as a model manager, and the third communication device may be, but is not limited to: a model management function network element, or a model management function entity, or a communication device including a model management function.
  • the third communication device is a network management system (NMS) device including a model management function module.
  • NMS network management system
  • the third communication device receives the training capability indication information of the first communication device and the reasoning requirement information and reasoning capability information of the second communication device, and determines that the first communication device supports multi-model training and the second communication device supports multi-model reasoning; then the third communication device sends a model request information carrying the reasoning requirement information to the first communication device, and the third communication device can effectively receive the first information carrying the information of the first model from the first communication device, the information of the first model is determined according to the reasoning requirement information and the first model is a multi-model; the third communication device then sends the information of the first model to the second communication device through the second information; after the second communication device receives the information of the first model, based on the information of the first model, the first model can be effectively used to perform multi-model reasoning and combination, and obtain a reasoning result with higher accuracy. Therefore, this method can effectively improve the use effect of the model to ensure the performance of intelligent reasoning.
  • the inference requirement information of the second communication device includes one or more of the following: the type of inference, the performance requirement of inference, the speed requirement of inference, and the power consumption requirement of inference.
  • the reasoning capability information of the second communication device further includes one or more of the following: The computing power and storage space for reasoning.
  • the model request information also includes multi-model indication information, and the multi-model indication information is used to indicate that the model requested to be trained or obtained is a multi-model.
  • the multi-model indication information may also be carried in the inference requirement information included in the model request information, or the third communication device may send the multi-model indication information to the first communication device separately. This embodiment of the present application does not specifically limit this.
  • the first communication device can effectively and accurately train or provide multiple models.
  • the method also includes: the third communication device receives the reasoning performance information of the first model from the first communication device; the third communication device adjusts the number of sub-models in the first model according to the reasoning requirement information and the reasoning capability information of the second communication device, as well as the reasoning performance information of the first model and the information of the first model; the reasoning performance information of the first model includes one or more of the following: the performance of the first model, the size information of the first model, the power consumption of the reasoning of the first model, the reasoning speed of the first model, and the computing power of the first model.
  • the third communication device can adjust the number of sub-models of the first model (such as reducing the number of sub-models of the first model) based on the reasoning performance information of the first model and the reasoning requirement information and reasoning capability information of the second communication device to ensure that the reasoning performance of the sub-models of the first model actually used is better.
  • the information of the first model includes model information of the first model and information of multiple sub-models of the first model, and the information of each sub-model includes one or more of the following: identification information of the sub-model, the level of the sub-model, the performance of the sub-model, and performance constraints; the multiple sub-models include multiple first-level sub-models and one second-level sub-model, and the second-level sub-model is used to aggregate the reasoning information of the multiple first-level sub-models; or the multiple sub-models are all first-level sub-models, and the information of the first model also includes aggregation method and/or weight information.
  • the model information of the first model may be the identification, name, type, etc. of the first model.
  • the identification information of the sub-model may be the storage address of the sub-model, or the unique identifier of the sub-model, etc.
  • the manner in which the information of the multiple sub-models of the first model and the reasoning information of the sub-models are combined can be accurately determined, so that the second communication device can subsequently effectively use these sub-models to perform combined reasoning.
  • an embodiment of the present application further provides a communication device, which may be the first communication device of the first aspect, or a component (e.g., a chip, or a chip system, or a circuit) in the first communication device, or a device that can be used in combination with the first communication device.
  • the first communication device may be, but is not limited to, a model training function network element, or a model training function entity, or a communication device including a model training function.
  • the communication device may include a module or unit corresponding to the method/operation/step/action described in the first aspect, and the module or unit may be a hardware circuit, or software, or a hardware circuit combined with software.
  • the communication device may include a communication module (or a transceiver module) and a processing module. The processing module is used to call the communication module to perform the communication (i.e., receiving and/or sending) function.
  • the communication device includes a communication unit (or a transceiver unit) and a processing unit; the processing unit can be used to call the communication unit to perform communication (i.e., receiving and/or sending) functions; wherein the communication unit is used to receive model request information, and the model request information includes reasoning requirement information; the processing unit is used to determine a first model based on the model request information, and the first model is a multi-model; the communication unit is also used to send first information, and the first information includes information of the first model.
  • the processing unit can be used to call the communication unit to perform communication (i.e., receiving and/or sending) functions
  • the communication unit is used to receive model request information, and the model request information includes reasoning requirement information
  • the processing unit is used to determine a first model based on the model request information, and the first model is a multi-model
  • the communication unit is also used to send first information, and the first information includes information of the first model.
  • the communication unit is further used to: before receiving the model request information, send training capability indication information, where the training capability indication information is used to indicate that the first communication device supports multi-model training.
  • the model request information when used to request training of multiple models, the model request information also includes a training strategy for the multiple models; when the processing unit determines the first model according to the model request information, it is specifically used to: train according to the reasoning requirement information and the training strategy for the multiple models to obtain multiple sub-models of the first model; or when the model request information is used to request training of multiple models; when the processing unit determines the first model according to the model request information, it is specifically used to: determine the training strategy for the multiple models according to the reasoning requirement information; and train according to the reasoning requirement information and the training strategy for the multiple models to obtain multiple sub-models of the first model; wherein the training strategy for the multiple models includes one or more of the following: data processing strategy, training algorithm, training mode, number of sub-models, and type of sub-models.
  • the processing unit when determining the first model based on the model request information, is specifically used to: determine the first model from at least one preset multiple models based on the reasoning requirement information.
  • the model request information also includes multi-model indication information, and the multi-model indication information is used to indicate that the model requested to be trained or obtained is a multi-model.
  • the reasoning requirement information includes one or more of the following: the type of reasoning, the performance requirement of reasoning, the speed requirement of reasoning, and the power consumption requirement of reasoning.
  • the information of the first model includes model information of the first model and information of multiple sub-models of the first model, and the information of each sub-model includes one or more of the following: identification information of the sub-model, the level of the sub-model, the performance of the sub-model, and performance constraints; the multiple sub-models include multiple first-level sub-models and one second-level sub-model, and the second-level sub-model is used to aggregate the reasoning information of the multiple first-level sub-models; or the multiple sub-models are all first-level sub-models, and the information of the first model also includes aggregation method and/or weight information.
  • the communication unit is also used to send reasoning performance information of the first model, and the reasoning performance information of the first model includes one or more of the following: performance of the first model, size information of the first model, power consumption of reasoning of the first model, reasoning speed of the first model, and computing power of the first model.
  • an embodiment of the present application further provides a communication device, which can be used for the second communication device of the second aspect, or can be a component (for example, a chip, or a chip system, or a circuit) in the second communication device, or a device that can be used in combination with the second communication device.
  • the second communication device can be, but is not limited to: a model reasoning function network element, or a model reasoning function entity, or a communication device including a model reasoning function.
  • the communication device may include a module or unit corresponding to the method/operation/step/action described in the second aspect, and the module or unit may be a hardware circuit, or software, or a hardware circuit combined with software.
  • the communication device may include a processing module and a communication module (or a transceiver model). The processing module is used to call the communication module to perform the communication (i.e., receiving and/or sending) function.
  • the communication device includes a communication unit (or a transceiver unit) and a processing unit; wherein the communication unit receives second information, the second information includes information of a first model, the first model is determined based on reasoning requirement information, and the first model is a multi-model; the processing unit is used to obtain reasoning information of the first model based on the information of the first model.
  • the communication unit is also used to: send reasoning capability information and the reasoning requirement information before receiving the second information;
  • the reasoning capability information includes reasoning capability indication information, and one or more of the following: computing power and storage space for reasoning;
  • the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning;
  • the reasoning requirement information includes one or more of the following: the type of reasoning, the performance requirement of reasoning, the speed requirement of reasoning, and the power consumption requirement of reasoning.
  • the information of the first model includes model information of the first model and information of multiple sub-models of the first model, and the information of each sub-model includes one or more of the following: identification information of the sub-model, the level of the sub-model, the performance of the sub-model, and performance constraints.
  • the multiple sub-models include multiple first-level sub-models and one second-level sub-model, and the second-level sub-model is used to aggregate the reasoning information of the multiple first-level sub-models;
  • the processing unit When obtaining the inference information of the first model based on the information of the first model, the processing unit is specifically used to: based on the information of the multiple first-level sub-models, use the multiple first-level sub-models to perform inference respectively to obtain the inference information of the multiple first-level sub-models; use the second-level sub-model to aggregate the inference information of the multiple first-level sub-models to obtain the inference information of the first model.
  • the multiple sub-models are all first-level sub-models, and the information of the first model also includes an aggregation method and/or weight information; when the processing unit obtains the reasoning information of the first model based on the information of the first model, it is specifically used to: based on the information of the multiple sub-models, use the multiple sub-models to perform reasoning respectively to obtain the reasoning information of the multiple sub-models; aggregate the reasoning information of the multiple sub-models according to the aggregation method and/or weight information to obtain the reasoning information of the first model.
  • an embodiment of the present application further provides a communication device, which can be used for the third communication device of the third aspect, or can be a component (for example, a chip, or a chip system, or a circuit) in the third communication device, or a device that can be used in combination with the third communication device.
  • the third communication device can be, but is not limited to, a model management function network element, or a model management function entity, or a communication device including a model management function.
  • the communication device may include a module or unit corresponding to the method/operation/step/action described in the third aspect, and the module or unit may be a hardware circuit, or software, or a combination of a hardware circuit and software.
  • the communication device may include a processing module and a transceiver module. The processing module is used to call the communication module (or the transceiver module) to perform the communication (ie, receiving and/or sending) function.
  • the communication device includes a communication unit (transceiver unit) and a processing unit; wherein the processing unit is used to call the communication unit to perform communication (i.e., receiving and/or sending) functions; the communication unit is used to receive training capability indication information of a first communication device; the training capability indication information is used to indicate that the first communication device supports multi-model training; and receive reasoning requirement information and reasoning capability information of a second communication device; the reasoning capability information includes reasoning capability indication information, and the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning; the communication unit is also used to send model request information to the first communication device, and the model request information includes the reasoning requirement information; receive first information from the first communication device, the first information includes information of a first model, the first model is a multi-model, and the first model is determined based on the reasoning requirement information; and send second information to the second communication device, the second information includes information of the first model.
  • the processing unit is used to call the communication unit to perform communication (i.e., receiving and/or sending)
  • the reasoning requirement information includes one or more of the following: the type of reasoning, the performance requirement of reasoning, the speed requirement of reasoning, and the power consumption requirement of reasoning.
  • the reasoning capability information further includes one or more of the following: reasoning computing power and storage space.
  • the model request information also includes multi-model indication information, and the multi-model indication information is used to indicate that the model requested to be trained or obtained is a multi-model.
  • the communication unit is further used to: receive reasoning performance information of the first model from the first communication device; the processing unit is further used to adjust the number of sub-models in the first model according to the reasoning requirement information and the reasoning capability information of the second communication device, as well as the reasoning information of the first model and the information of the first model; the reasoning performance information of the first model includes one or more of the following: the performance of the first model, the size information of the first model, the power consumption of the reasoning of the first model, the reasoning speed of the first model, and the computing power of the first model.
  • the information of the first model includes model information of the first model and information of multiple sub-models of the first model, and the information of each sub-model includes one or more of the following: identification information of the sub-model, the level of the sub-model, the performance of the sub-model, and performance constraints; the multiple sub-models include multiple first-level sub-models and one second-level sub-model, and the second-level sub-model is used to aggregate the reasoning information of the multiple first-level sub-models; or the multiple sub-models are all first-level sub-models, and the information of the first model also includes aggregation method and/or weight information.
  • a communication device in an embodiment of the present application, and the device includes: at least one processor and an interface circuit; the interface circuit is used to provide input and/or output of programs or instructions to the at least one processor; the at least one processor is used to execute the program or instructions so that the communication device can implement the method provided by the above-mentioned first aspect or any possible implementation method thereof, or can implement the method provided by the above-mentioned second aspect or any possible implementation method thereof, or can implement the method provided by the above-mentioned third aspect or any possible implementation method thereof.
  • a computer storage medium in an embodiment of the present application, in which a software program is stored.
  • the software program is read and executed by one or more processors, the method provided by the first aspect or any possible implementation thereof can be implemented, or the method provided by the second aspect or any possible implementation thereof can be implemented, or the method provided by the third aspect or any possible implementation thereof can be implemented.
  • a computer program product comprising instructions is provided in an embodiment of the present application.
  • the computer executes the method provided in the first aspect or any possible implementation manner thereof, or the computer executes the method provided in the second aspect or any possible implementation manner thereof, or the computer executes the method provided in the third aspect or any possible implementation manner thereof.
  • a chip system in an embodiment of the present application, which chip system includes a processor for supporting a device to implement the functions involved in the above-mentioned first aspect, or for supporting a device to implement the functions involved in the above-mentioned second aspect, or for supporting a device to implement the functions involved in the above-mentioned third aspect.
  • the chip system further includes a memory, and the memory is used to store necessary program instructions and data.
  • the chip system can be composed of a chip, or can include a chip and other discrete devices.
  • a chip system is also provided in an embodiment of the present application, which includes a processor and an interface, wherein the interface is used to obtain a program or instruction, and the processor is used to call the program or instruction to implement or support the device to implement the function involved in the first aspect, or the processor is used to call the program or instruction to implement or support the device to implement the function involved in the second aspect, or the processor is used to call the program or instruction to implement or support the device to implement the function involved in the third aspect.
  • the chip system further includes a memory, the memory being used to store program instructions necessary for the terminal device.
  • the chip system can be composed of chips or include chips and other discrete devices.
  • FIG1 is a schematic diagram of a solution flow chart for improving the use effect of the model
  • FIG. 2 is an example diagram of two system logic architectures provided in an embodiment of the present application.
  • FIG3A is a schematic diagram of a first practical deployment architecture to which the method of an embodiment of the present application can be applied;
  • FIG3B is a schematic diagram of a second practical deployment architecture to which the method of the embodiment of the present application can be applied;
  • FIG3C is a schematic diagram of a third practical deployment architecture to which the method of the embodiment of the present application can be applied;
  • FIG3D is a schematic diagram of a fourth actual deployment architecture to which the method of the embodiment of the present application can be applied;
  • FIG3E is a schematic diagram of a fifth practical deployment architecture to which the method of the embodiment of the present application can be applied;
  • FIG4A is a flow chart of a communication method provided in an embodiment of the present application.
  • FIG4B is a schematic diagram of a flow chart of another communication method provided in an embodiment of the present application.
  • FIG4C is an example diagram of a multi-model training and reasoning process provided in an embodiment of the present application.
  • FIG5 is a schematic diagram of a flow chart of a first embodiment provided in the embodiments of the present application.
  • FIG6 is a schematic diagram of a flow chart of a second embodiment provided in the present application.
  • FIG7 is a schematic diagram of a flow chart of a third embodiment provided in the embodiments of the present application.
  • FIG8 is a schematic diagram of a flow chart of a fourth embodiment provided in the embodiments of the present application.
  • FIG9 is a schematic diagram of a flow chart of a fifth embodiment provided in the embodiments of the present application.
  • FIG10 is a schematic diagram of a flow chart of a sixth embodiment provided in an embodiment of the present application.
  • FIG11 is a schematic diagram of a flow chart of a seventh embodiment provided in the embodiments of the present application.
  • FIG12 is a schematic diagram of the structure of a communication device provided in an embodiment of the present application.
  • FIG13 is a schematic diagram of the structure of another communication device provided in an embodiment of the present application.
  • FIG. 14 is a schematic diagram of the device structure of a chip provided in an embodiment of the present application.
  • AI artificial intelligence
  • ML machine learning
  • the current protocol defines a management data analytics service (MDAS).
  • MDAS producers can process and analyze data related to network, service events and status based on AI/ML technology, and provide analysis reports for network and service operations.
  • the network data analysis function (NWDAF) can perform network data analysis based on ML models, obtain data analysis results, and provide them to the network, network management and applications for policy decision-making.
  • the current main research is to support the definition of the functional framework of RAN intelligence, that is, to make necessary enhancements based on the current RAN architecture and interfaces to support network intelligence.
  • model training functional network element provides a plurality of models with different performances to the model reasoning functional network element (or entity), and the model reasoning functional network element (or entity) selects a suitable model for reasoning.
  • the steps for implementation include: S101: the model reasoning functional entity sends a model request to the model training functional entity, and the model request includes the reasoning type (such as coverage problem analysis, cell traffic prediction, etc.) and performance requirements (such as precision, accuracy, etc.); S102: the model training functional entity trains multiple models according to the reasoning type and performance requirements in the model request; S103: the model The model training functional entity sends a model response to the model reasoning functional entity, where the model response includes a reasoning type and a list of multiple models, for example, ⁇ model ID 1, model ID 2, ...>.
  • the model response includes a reasoning type and a list of multiple models, for example, ⁇ model ID 1, model ID 2, ...>.
  • the model reasoning functional entity cannot determine the training capability of the model training functional entity. If multiple models are requested from the model training functional entity, the request may fail.
  • the number of models trained by the model training functional entity is uncertain and the provided models are not necessarily suitable for the reasoning of the model reasoning functional entity, these may lead to the scheme being infeasible or the reasoning effect being unsatisfactory, thereby failing to effectively improve the intelligent reasoning (or analysis) performance of the model.
  • the present application proposes a communication method that can effectively improve the use effect of the model, thereby ensuring the performance of intelligent reasoning.
  • the method proposed in the present application can be applied to the 5G system architecture, and can also be applied to but not limited to the long-term evolution (LTE) communication system, and various wireless communication systems that will evolve in the future.
  • LTE long-term evolution
  • the system logical architecture applicable to the embodiments of the present application mainly includes a model management functional entity, a model training functional entity and a model reasoning functional entity.
  • Figure 2 shows example diagrams of several system logical architectures of the embodiments of the present application.
  • the model management functional entity, the model training functional entity and the model reasoning functional entity can all be independent logical entities, and there is an interface for communication between any two entities; as shown in (2) in Figure 2, the model management functional entity is optional, and the model management functional entity can be set inside the model training functional entity.
  • the model training functional entity and the model reasoning functional entity are independent logical entities, and there is an interface for communication between the two.
  • Figure 3A shows an architecture that supports model reasoning in the management domain
  • the model management function entity can be deployed in the network management system (network management system, NMS) equipment
  • the model training function entity and the model reasoning function entity can be deployed in the element management system (element management system, EMS) equipment
  • Figure 3B shows an architecture that supports model reasoning in the RAN domain
  • the model management function entity can be deployed in the NMS
  • the model training function entity can be deployed in the EMS
  • the model reasoning function entity can be deployed in the RAN domain equipment, such as a base station
  • FIG3C shows an architecture supporting model reasoning in the CN domain.
  • the model training functional entity can be deployed in the NWDAF (MTLF), and the model reasoning functional entity can be deployed in the NWDAF (AnLF).
  • the network repository function (NRF) is mainly used for the management of network functions (NFs), including NF registration/update/deregistration, NF discovery, etc. Since the NWDAF can be regarded as a NF, the NRF can manage the NWDAF.
  • FIG3D and FIG3E respectively show an architecture supporting air interface-related model reasoning in the RAN domain.
  • the model training functional entity can be deployed in the base station, and the model reasoning functional entity can be deployed in the user equipment (UE); optionally, the model reasoning functional entity can also be deployed in the base station when both sides can support reasoning.
  • the model training functional entity and the model reasoning functional entity are deployed on both the base station and the UE side, which can be applied to the scenario of joint training of the base station and the UE in the bilateral model scenario.
  • NMS Network Management System
  • NMS can also be called a cross-domain management system, which is responsible for the operation, management and maintenance of the network.
  • EMS Element Management System
  • EMS can also be called a domain management system or a single domain management system, which is used to manage one or more network elements of a certain category.
  • NMS and EMS can also be collectively referred to as 3GPP management system, or operations administration and maintenance (OAM) module.
  • OAM operations administration and maintenance
  • Radio Access Network provides wireless access to user devices, allowing users to access the network.
  • Core network CN mainly provides user connection, user management and service carrying, and provides an interface to the external network as a bearer network.
  • NWDAF network element responsible for data analysis in the core network domain, such as detecting abnormal user behavior and analyzing slice load.
  • Model management functional entity responsible for model-related lifecycle management, including training strategy configuration, etc.
  • Model training functional entity responsible for model training and generating an ML model after model training is completed.
  • Model reasoning functional entity responsible for the reasoning of the model, using the model to obtain the reasoning output or reasoning result.
  • AnLF is the analysis logic function of the NWDAF network element.
  • MTLF Model training logic function for NWDAF network elements.
  • Base station A device in a mobile communication system that connects the fixed part with the wireless part and is connected to mobile terminals through wireless channels in the air.
  • UE User terminal equipment, a device that allows users to access the network.
  • FIGS 3A-3E is not limited to including only the entities shown in the figures, but may also include other devices not shown in the figures, which will not be listed one by one in this application.
  • model management functional entity, model training functional entity, and model reasoning functional entity included in the NMS, EMS, RAN, UE, and NWDAF in Figures 3A-3E can also be called model management function, model training function, model reasoning function, and can also be called model management function network element/module, model training function network element/module, and model reasoning function network element/module.
  • the network element or function can be a network element in a hardware device, a software function running on dedicated hardware, or a virtualized function instantiated on a platform (e.g., a cloud platform).
  • a platform e.g., a cloud platform.
  • the above network element or function can be implemented by one device, or by multiple devices together, or can be a functional module in one device, which is not specifically limited in the embodiments of the present application.
  • the embodiments of the present application are explained by taking the network element as an example, and the XX network element is directly referred to as XX, for example, the SMF network element is referred to as SMF.
  • SMF SMF network element
  • the names of all messages and information in this application are only examples and may be other names, which are not limited in this application.
  • the message or information from network element 1 to network element 2 may be a message sent directly from network element 1 to network element 2, or may be sent indirectly, for example, network element 1 first sends a message to network element 3, and network element 3 then sends a message to network element 2, and finally the message or information is sent to network element 2 through one or more network elements.
  • indication may include direct indication, indirect indication, explicit indication, and implicit indication.
  • indication information may include direct indication, indirect indication, explicit indication, and implicit indication.
  • the information indicated by the indication information is referred to as the information to be indicated.
  • the information to be indicated can be directly indicated, such as the information to be indicated itself or the index of the information to be indicated.
  • the information to be indicated can also be indirectly indicated by indicating other information, wherein there is an association relationship between the other information and the information to be indicated. It is also possible to indicate only a part of the information to be indicated, while the other parts of the information to be indicated are known or agreed in advance.
  • the indication of specific information can also be achieved with the help of the arrangement order of each information agreed in advance (such as specified by the protocol), thereby reducing the indication overhead to a certain extent.
  • the information to be indicated can be sent as a whole, or divided into multiple sub-information and sent separately, and the sending period and/or sending time of these sub-information can be the same or different.
  • the specific sending method is not limited in this application.
  • the sending period and/or sending time of these sub-information can be pre-defined, for example, pre-defined according to the protocol, or configured by the transmitting device by sending configuration information to the receiving device.
  • the configuration information can include, for example, but not limited to, one or a combination of at least two of radio resource control signaling, media access control (media access control, MAC) layer signaling and physical layer signaling.
  • radio resource control signaling for example, radio resource control (radio resource control, RRC) signaling
  • MAC layer signaling for example, includes MAC control element (control element, CE)
  • physical layer signaling for example, includes downlink control information (downlink control information, DCI).
  • An embodiment of the present application provides a communication method, which can be applied to but not limited to the actual deployment architecture shown in Figures 3A-3E, and the method can be executed by the network element involved in the present application, or by the chip corresponding to the network element involved.
  • the network element in the present application can be a physical entity network element or a virtual network element.
  • the present application does not specifically limit the form of the network element involved.
  • FIG4A is a flow chart of a communication method proposed in an embodiment of the present application.
  • the method can be executed by a transceiver and/or processor of a first communication device (or a second communication device or a third communication device), or by a chip corresponding to the transceiver and/or processor.
  • the embodiment can also be implemented by a controller or control device connected to the first communication device (or a second communication device or a third communication device), and the controller or control device is used to manage at least one device including the first communication device (or a second communication device or a third communication device).
  • the present application does not make specific restrictions on the specific form of the communication device that executes this embodiment.
  • the third communication device receives training capability indication information from the first communication device, where the training capability indication information is used to indicate that the first communication device supports multi-model training.
  • the first communication device can be used as a model training end (or model determination end), and the first communication device can be, but is not limited to: a model training function network element, or a model training function entity, or a communication device including a model training function, such as a NWDAF network element including a model training function module, or a network element management system (EMS) device, or an access network device (such as a base station), etc.
  • the third communication device can be, but is not limited to: a model management function network element, or a model management function entity, or a communication device including a model management function, such as a network management system (NMS) device including a model management function module.
  • NMS network management system
  • the third communication device may first send training capability query information to the first communication device, and after the first communication device receives the training capability query information, it may send the training capability indication information to the third communication device. In other embodiments, the first communication device may also actively report (i.e., send) the training capability indication information to the third communication device.
  • the third communication device receives reasoning requirement information and reasoning capability information of the second communication device, where the reasoning capability information includes reasoning capability indication information, and the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning.
  • the second communication device can serve as a model reasoning end (or model usage end), and the second communication device can be but is not limited to: a model reasoning function network element, or a model reasoning function entity, or a communication device including a model reasoning function, such as a NWDAF network element including a model reasoning function module, or a network element management system (EMS) device, or an access network device (such as a base station), etc.
  • a model reasoning function network element or a model reasoning function entity
  • a communication device including a model reasoning function such as a NWDAF network element including a model reasoning function module, or a network element management system (EMS) device, or an access network device (such as a base station), etc.
  • EMS network element management system
  • first communication device, the second communication device, and the third communication device can all be independent devices, or the first communication device, the second communication device, and the third communication device can be respectively located in independent devices; or the first communication device and the second communication device are located in the same device; or the first communication device and the second communication device are the same device; therefore, the embodiment of the present application does not specifically limit the specific form of the first communication device, the second communication device, and the third communication device, the device where each communication device is located, and the location.
  • the third communication device first sends reasoning requirement query information and reasoning capability query information to the second communication device. After the second communication device receives the reasoning requirement query information and reasoning capability query information, it sends the reasoning requirement information and the reasoning capability query information to the third communication device.
  • the above-mentioned reasoning requirement query information and reasoning capability query information can be sent separately by the third communication device, or they can be sent simultaneously, that is, the time when the third communication device sends the reasoning requirement query information and the reasoning capability query information is not limited.
  • the reasoning requirement query information and the reasoning capability query information can be carried in the same message and sent by the third communication device, or they can be carried in different messages and sent by the third communication device, and the embodiments of the present application do not limit this.
  • the second communication device may also actively report reasoning requirement information and reasoning capability information to the third communication device, where the reasoning capability information includes reasoning capability indication information, and the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning.
  • the above-mentioned reasoning requirement information and reasoning capability information can be sent separately by the second communication device, or sent simultaneously by the second communication device, that is, the time when the second communication device sends the reasoning requirement information and the reasoning capability information is not limited.
  • the reasoning requirement information and the reasoning capability information can be carried in the same message and sent by the second communication device, or they can be carried in different messages and sent by the second communication device, and the embodiments of the present application do not limit this.
  • the reasoning requirement information of the second communication device may include but is not limited to one or more of the type of reasoning (for example, coverage problem analysis, cell traffic prediction, etc.), performance requirements of reasoning (such as precision, accuracy, etc.), speed requirements of reasoning, and power consumption requirements of reasoning.
  • the type of reasoning for example, coverage problem analysis, cell traffic prediction, etc.
  • performance requirements of reasoning such as precision, accuracy, etc.
  • speed requirements of reasoning speed requirements of reasoning
  • power consumption requirements of reasoning may include but is not limited to one or more of the type of reasoning (for example, coverage problem analysis, cell traffic prediction, etc.), performance requirements of reasoning (such as precision, accuracy, etc.), speed requirements of reasoning, and power consumption requirements of reasoning.
  • the third communication device sends model request information to the first communication device, wherein the model request information includes the inference requirement information.
  • the first communication device receives the model request information.
  • the model request information also includes multi-model indication information, where the multi-model indication information is used to indicate that the model requested to be trained or obtained is a multi-model.
  • the multi-model indication information may also be carried in the inference requirement information included in the model request information, or the third communication device may separately send the multi-model indication information to the first communication device.
  • the present application does not make any specific limitations on this.
  • the first communication device determines a first model according to the model request information, where the first model is a multi-model.
  • the first communication device when the model request information is used to request training of multiple models, the first communication device, based on the model request information, Determining the first model may include but is not limited to the following methods:
  • Method 1 If the model request information also includes a multi-model training strategy, the first communication device can perform training according to the inference requirement information and the multi-model training strategy to obtain multiple sub-models of the first model.
  • Method 2 The first communication device first determines a multi-model training strategy based on the inference requirement information; and then performs training based on the inference requirement information and the multi-model training strategy to obtain multiple sub-models of the first model.
  • the multi-model training strategy may include but is not limited to data processing strategy, training algorithm, training mode, number of sub-models, type of sub-models, etc.
  • the first communication device determines the first model according to the model request information, which may include: the first communication device determines the first model from at least one preset multiple model according to the inference requirement information.
  • the at least one preset multiple model may be a multiple model that has been trained in advance by the first communication device.
  • S405A The first communication device sends first information, where the first information includes information of the first model.
  • the third communication device receives the first information from the first communication device, and the information of the first model includes model information of the first model and information of multiple sub-models of the first model.
  • the information of each sub-model includes but is not limited to identification information of the sub-model, the level of the sub-model, the performance of the sub-model, and performance constraints.
  • the identification information of the submodel may be, but is not limited to: the name (identification) of the submodel, the storage address information of the submodel, and the unique identification number of the submodel.
  • the level of the submodel may include a first-level submodel and a second-level submodel; wherein the second-level submodel may be used to aggregate the reasoning information of multiple first-level submodels or multiple first-level submodels.
  • the multiple sub-models of the first model include multiple first-level sub-models and one second-level sub-model, and the second-level sub-model is used to aggregate the reasoning information of the multiple first-level sub-models.
  • the multiple sub-models of the first model are all first-level sub-models, and the information of the first model also includes aggregation method and/or weight information.
  • the first communication device also sends the reasoning performance information of the first model to the third communication device, and correspondingly, the third communication device receives the reasoning performance information of the first model;
  • the reasoning performance information of the first model may include but is not limited to the performance of the first model, the size information of the first model, the power consumption of the reasoning of the first model, the reasoning speed of the first model, the computing power of the first model, etc.
  • the third communication device sends second information to the second communication device, where the second information includes information of the first model.
  • the second communication device receives the second information.
  • the third communication device before the third communication device executes step S406A, the third communication device further executes the following steps:
  • the third communication device adjusts the number of sub-models in the first model according to the reasoning requirement information and the reasoning capability information of the second communication device, as well as the reasoning information of the first model and the information of the first model. By adjusting the number of sub-models of the first model, it can be ensured that the effect of actually using the first model is better.
  • the second communication device obtains inference information of the first model based on the information of the first model.
  • the multiple sub-models of the first model include multiple first-level sub-models and one second-level sub-model; the second communication device obtains the reasoning information of the first model based on the information of the first model, which may include:
  • the second communication device uses the multiple first-level sub-models to perform reasoning respectively based on the information of the multiple first-level sub-models to obtain reasoning information of the multiple first-level sub-models; the second communication device then uses the second-level sub-model to aggregate the reasoning information of the multiple first-level sub-models to obtain the reasoning information of the first model.
  • the second communication device obtains the reasoning information of the first model based on the information of the first model, which may include: the second communication device uses the multiple sub-models to perform reasoning respectively based on the information of the multiple sub-models to obtain the reasoning information of the multiple sub-models; the second communication device then aggregates the reasoning information of the multiple sub-models according to the aggregation method and/or weight information to obtain the reasoning information of the first model.
  • the first communication device after the first communication device receives the model request information, it can determine a more appropriate multi-model (i.e., the first model) according to the reasoning requirement information in the model request information, and then send the information of the first model through the first information; when the reasoning end (i.e., the second communication device including the model reasoning function module) receives the information of the first model, based on the information of the first model, the multi-model reasoning and combination are performed using the first model, and a reasoning result with higher accuracy can be obtained. Therefore, the use effect of the model can be effectively improved, and the performance of intelligent reasoning (or analysis) can be guaranteed.
  • a more appropriate multi-model i.e., the first model
  • the reasoning end i.e., the second communication device including the model reasoning function module
  • Figure 4B is a flow chart of another communication method proposed in an embodiment of the present application.
  • the method can be executed by a transceiver and/or processor of a first communication device (or a second communication device), or by a chip corresponding to the transceiver and/or processor.
  • the embodiment can also be implemented by a controller or control device connected to the first communication device (or a second communication device), and the controller or control device is used to manage at least one device including the first communication device (or a second communication device).
  • the present application does not make specific restrictions on the specific form of the communication device that executes this embodiment. Please refer to Figure 4B, the specific process of the method is as follows:
  • S401B The first communication device receives model request information from the second communication device, where the model request information includes reasoning requirement information.
  • the reasoning requirement information may include, but is not limited to, the type of reasoning, the performance requirement of reasoning, the speed requirement of reasoning, and the power consumption requirement of reasoning.
  • the first communication device may be used as a model training end (or a model determination end), and the first communication device may be, but is not limited to: a model training function network element, or a model training function entity, or a communication device including a model training function, such as a NWDAF network element including a model training function module, or a network element management system (EMS) device, or an access network device (such as a base station), etc.
  • a model training function network element or a model training function entity
  • a communication device including a model training function such as a NWDAF network element including a model training function module, or a network element management system (EMS) device, or an access network device (such as a base station), etc.
  • NWDAF network element including a model training function module
  • EMS network element management system
  • access network device such as a base station
  • the second communication device may be used as a model reasoning end (or a model use end), and the second communication device may be, but is not limited to: a model reasoning function network element, or a model reasoning function entity, or a communication device including a model reasoning function, such as a NWDAF network element including a model reasoning function module, or a network element management system (EMS) device, or an access network device (such as a base station), etc.
  • a model reasoning function network element or a model reasoning function entity
  • a communication device including a model reasoning function such as a NWDAF network element including a model reasoning function module, or a network element management system (EMS) device, or an access network device (such as a base station), etc.
  • first communication device and the second communication device can be independent devices, or the first communication device and the second communication device can be located in independent devices respectively; or the first communication device and the second communication device are located in the same device; or the first communication device and the second communication device are the same device; therefore, the embodiment of the present application does not specifically limit the specific form of the first communication device and the second communication device, the device where each communication device is located, and the location.
  • the first communication device may first send training capability indication information to the second communication device, and the training capability indication information is used to indicate that the first communication device supports multi-model training.
  • the first communication device may also receive reasoning capability information from the second communication device, and the reasoning capability information includes but is not limited to: reasoning capability indication information, and the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning.
  • the reasoning capability information may also include the computing power and storage space for reasoning.
  • the first communication device determines a first model according to the model request information, where the first model is a multi-model.
  • step S402B when the first communication device executes the step S402B, specific reference may be made to the above-mentioned step S404A, which will not be repeated here.
  • the first communication device sends first information, the first information including information of the first model.
  • the second communication device receives the first information (ie, second information).
  • the information of the first model includes model information of the first model and information of multiple sub-models of the first model.
  • the information of each sub-model includes but is not limited to identification information of the sub-model, the level of the sub-model, the performance of the sub-model, and performance constraints.
  • the identification information of the submodel may be, but is not limited to: the name (identification) of the submodel, the storage address information of the submodel, and the unique identification symbol of the submodel.
  • the level of the submodel may include a first-level submodel and a second-level submodel; wherein the second-level submodel may be used to aggregate the reasoning information of multiple first-level submodels or multiple first-level submodels.
  • S404B The second communication device obtains inference information of the first model based on the information of the first model.
  • step S404B when the second communication device executes the step S404B, specific reference may be made to the above-mentioned step S407A, which will not be repeated here.
  • FIG4C shows an example diagram of a multi-model training and reasoning process proposed in an embodiment of the present application.
  • the model training function will start multiple learner training models according to the reasoning requirement information of the model reasoning function.
  • the training method may include: the model training function first splits the original data set into multiple sub-data sets according to a certain strategy, that is, data set 1, data set 2...data set n, where n is a positive integer; then, the model training function uses each sub-data set to train the sub-learner, thereby obtaining multiple sub-learner models, that is, sub-learner 1, sub-learner 2...sub-learner n in the figure.
  • the model reasoning function After the model reasoning function obtains the information of the multiple sub-learner models from the model training function, the reasoning data can be input into the multiple sub-learners (that is, sub-learner 1, sub-learner 2...sub-learner n) according to the information of the multiple sub-learner models to obtain corresponding reasoning outputs, that is, reasoning output 1, reasoning output 2...reasoning output n; then, the model reasoning function combines these reasoning outputs according to a preset aggregation method (such as voting method, simple average method, weighted average method, linear hybrid method) to obtain the final reasoning output.
  • a preset aggregation method such as voting method, simple average method, weighted average method, linear hybrid method
  • the method can obtain Appropriate multiple models are used to implement reasoning, thereby improving the reasoning (or analysis) performance of the model.
  • Embodiment 1 is a diagrammatic representation of Embodiment 1:
  • the solution of the present application is applied to the deployment architecture shown in FIG. 3A above, by enhancing the model training and deployment process of the OAM (NMS/EMS) domain to support reasoning and analysis of the management domain based on the combination of multiple learner models.
  • the first communication device and the second communication device in the solution of the present application are the same network element management system device (EMS for short) in FIG. 3A that includes a model training function module and a model reasoning function module
  • the third communication device in the solution of the present application is a network management system device (NMS for short) in FIG. 3A that includes a model management function module.
  • EMS network element management system device
  • NMS network management system device
  • S501a The model management function module in the NMS sends reasoning requirement query information to the model reasoning function module in the EMS.
  • the model reasoning function module in the EMS receives the reasoning requirement query information, and the reasoning requirement query information is used to query the reasoning requirement of the model reasoning function module.
  • S501b The model reasoning function module in the EMS sends reasoning requirement information to the model management function module in the NMS.
  • the model management function module in the NMS receives the reasoning requirement information.
  • step S501a may also be omitted, that is, the model management function module in the NMS does not send the reasoning requirement query information to the model reasoning function module in the EMS, but the model reasoning function module in the EMS actively reports (i.e., sends) the reasoning requirement information to the model management function module in the NMS.
  • the above-mentioned reasoning requirement information may include: reasoning type requirements (such as coverage problem analysis, cell traffic prediction, etc.), reasoning performance requirements (such as reasoning precision, accuracy, etc.), reasoning speed requirements (optional), and reasoning power consumption requirements (optional).
  • the reasoning speed requirement can also be called the reasoning delay requirement, which indicates the requirement for the time to execute reasoning, for example, the time for a single reasoning execution is less than 1s;
  • the reasoning power consumption requirement indicates the requirement for the power consumption of executing reasoning, for example, the energy consumed by a single reasoning is less than 5J.
  • the model management function module in the NMS sends reasoning capability query information to the model reasoning function module in the EMS.
  • the reasoning capability query information is used to query (or obtain) the reasoning capability of the model reasoning function module in the EMS.
  • S502b The model reasoning function module in the EMS sends reasoning capability information to the model management function module in the NMS.
  • the model management function module in the NMS receives the reasoning capability information.
  • step S502a may also be omitted, that is, the model management function module in the NMS does not send reasoning capability query information to the model reasoning function module in the EMS, but the model reasoning function module in the EMS actively reports (i.e., sends) the reasoning capability information to the model management function module in the NMS.
  • the above-mentioned reasoning capability information may include: reasoning capability indication information, reasoning computing power (optional), and storage space (optional).
  • the storage space may be the size of the storage space occupied by the model reasoning, or the storage address, etc.
  • the reasoning capability indication information is used to indicate whether the model reasoning function in the EMS supports multi-model reasoning.
  • the corresponding value of the reasoning capability indication information is true (yes/correct), it means that multi-model reasoning is supported; if the corresponding value of the reasoning capability indication information is false (no/error), it means that multi-model reasoning is not supported; or the reasoning capability indication information is a specific value, if the value of the reasoning capability indication information is 1, it means that multi-model reasoning is supported, and if the value of the reasoning capability indication information is 0, it means that multi-model reasoning is not supported.
  • the inference computing power may refer to the computing power information available at the model inference function module, such as the available hardware resource information and the utilization rate of the hardware resources, wherein the hardware resources may include general computing power, such as the central processing unit (CPU), and high-performance computing power, such as the graphics processing unit (GPU), the neural network processing unit (NPU), etc.
  • the hardware resource information may be the original hardware information, which may include the hardware type, the number of cores, the processing frequency, etc., or the quantified computing power, which may usually be measured by the number of floating-point operations per second (FLOPS) supported.
  • the model management function module in the NMS sends training capability query information to the model training function module in the EMS.
  • the model training function module in the EMS receives the training capability query information.
  • the training capability query information is used to query (or obtain) the training capability of the model training function model in the EMS.
  • S503b The model training function module in the EMS sends training capability information to the model management function module in the NMS.
  • model management function module in the NMS may not send the training capability query information to the model training function module in the EMS, that is, the above step S503a is not executed, but the model training function module in the EMS actively sends the training capability query information to the model management function module in the NMS.
  • the module reports (i.e. sends) its own training capability information.
  • the above-mentioned training capability information is used to notify (or indicate) whether the model training function model in the EMS supports multi-model training.
  • the model training function module in the EMS sends training capability indication information to the model management function module in the NMS. If the corresponding value of the training capability indication information is true (yes or correct), it means that multi-model training is supported; if the corresponding value of the training capability indication information is false (no/error), it means that multi-model training is not supported; or the training capability indication information is a specific value. If the value of the training capability indication information is 1, it means that multi-model training is supported; if the value of the training capability indication information is 0, it means that multi-model training is not supported.
  • the above steps S501a-S501b, S502a-S502b, S503a-S503b are the query and reporting process of reasoning requirement information and capability information, which are optional steps, or can be completed offline.
  • the embodiment of the present application does not specifically limit the order of executing the above-mentioned steps of querying and reporting reasoning requirement information (i.e., S501a-S501b), querying and reporting reasoning capability information (i.e., S502a-S502b), and querying and reporting training capability information (i.e., S503a-S503b).
  • the model management function module in the NMS can determine the reasoning requirements of the model reasoning function module in the EMS, the training capability of the model training function module in the EMS, and the reasoning capability of the model reasoning function module in the EMS, so as to further perform the following steps:
  • the model management function module in the NMS sends a model training request to the model training function module in the EMS.
  • the model training function module in the EMS receives the model training request.
  • the model training request includes: model identification (or inference type), multi-model training instructions, and multi-model training strategy (optional).
  • the model identifier may also be expressed as an inference type.
  • the multi-model training indication is used to indicate whether to perform multi-model training.
  • the multi-model training strategy is used to indicate the training method, which may include a data processing strategy and a training algorithm indication.
  • the data processing strategy may include input data sampling and feature sampling.
  • input data sampling indicates sampling the original data to form multiple sub-data sets, each of which is used to train a sub-model.
  • Feature sampling indicates sampling the features of the original data, and data with different features can form multiple different sub-data sets, each of which is used to train a sub-model.
  • the training algorithm indication may include: the number of sub-models, the type of model, and the hyper-parameter configuration.
  • the number of sub-models indicates the number of sub-models that constitute the multi-model
  • the type of model indicates the model type of different sub-models, such as random forest model and convolutional neural network model
  • the hyper-parameter configuration indicates the hyper-parameters of model training, such as the number of layers, number of iterations, and learning rate of the neural network model.
  • the model training function module in the EMS performs multi-model training according to the model training request to obtain a first model (multi-model).
  • the model training function module in the EMS can determine the multi-model training strategy by itself based on the obtained reasoning requirement information.
  • the specific content of the multi-model training strategy can refer to the content of the multi-model training strategy in the above step S504, which will not be introduced in detail here.
  • the model training function module in the EMS sends a model training report to the model management function module in the NMS.
  • the model management function module in the NMS receives the model training report.
  • the model training report may include: model information of the first model, indication information of multiple models, a list of sub-models of the first model (i.e., information of multiple sub-models), an aggregation method (also called a combination method), a weight factor (optional), the performance of the first model, the size of the first model, the computational complexity of the first model, the inference speed of the first model, and the inference energy consumption of the first model.
  • the model information of the first model may be information used to identify the first model, such as a name or a unique identifier.
  • the indication information of the multiple models is used to indicate that the first model trained by the model training function module in the EMS is a multiple model.
  • the list of sub-models of the first model contains a series of information about the multiple sub-models constituting the first model, and the information of each sub-model includes: identification information of the sub-model, the level of the sub-model (also referred to as the category of the sub-model), the performance of the sub-model, and the performance constraint of the sub-model.
  • the identification information of the sub-model can be a unique identifier of the sub-model or a storage address of the sub-model.
  • the level of the sub-model can include a first-level sub-model (equivalent to the first-level sub-model in the present application) and a second-level sub-model (equivalent to the second-level sub-model in the present application, also referred to as an aggregate model), and the second-level sub-model is used to aggregate the reasoning information of the first-level sub-model.
  • the model training report if the list of sub-models of the first model contains only multiple first-level sub-models, then the model training report also includes an aggregation method, which can be but is not limited to voting method, simple average method, weighted average method, linear mixing method. When the aggregation method included in the model training report is weighted average method or linear mixing method, then the model training report should also include the weight factor corresponding to each sub-model. If the list of sub-models of the first model contains multiple first-level sub-models and one second-level sub-model, then the model training report may not include the aggregation method (also called the combination method) and the weight factor.
  • an aggregation method can be but is not limited to voting method, simple average method, weighted average method, linear mixing method.
  • the model training report should also include the weight factor corresponding to each sub-model. If the list of sub-models of the first model contains multiple first-level sub-models and one second-level sub-model, then the model training report may
  • the model management function module in the NMS adjusts the first model and determines the sub-model to be actually deployed based on the reasoning requirement information and reasoning capability information from the EMS and the first model information in the model training report.
  • the model management function module in the NMS can adjust the number and aggregation method of sub-models actually deployed of the first model based on the reasoning requirement information and reasoning capability information of the model reasoning function module in the EMS and the model information in the training report.
  • the model management function module in the NMS reduces the number of sub-models actually deployed in the first model according to the performance of each sub-model and the performance constraints of each sub-model.
  • the model management function module in the NMS sends model deployment request information to the model reasoning function module in the EMS.
  • the model reasoning function module in the EMS receives the model deployment request information, which is used to request the deployment of the sub-model of the first model.
  • the model deployment request information includes the model information of the first model, the indication information of multiple models, the list of sub-models of the first model (i.e., the information of multiple sub-models), the aggregation method (also called the combination method), and the weight factor (optional).
  • the model reasoning function module in the EMS actually deploys the sub-model of the first model trained by the model training function module in the EMS based on the information of the sub-model of the first model contained in the model deployment request.
  • the model reasoning function module in the EMS sends a model deployment response message to the model management function module in the NMS. Accordingly, the model management function module in the NMS receives the model deployment response message to determine (or know) that the model reasoning function module in the EMS has completed the deployment of the sub-model of the first model.
  • the model reasoning function module in the EMS performs multi-model reasoning based on the first model to obtain a reasoning result.
  • the model reasoning function module in the EMS inputs the data to be inferred (i.e., input data) into the sub-models actually deployed in the first model, respectively, to obtain the reasoning results of the corresponding sub-models. Furthermore, the model reasoning function model can combine the reasoning results of each sub-model based on an aggregation method to obtain the final reasoning result. Alternatively, the model reasoning function model can input the reasoning results obtained by each first-level sub-model of the first model into the second-level sub-model, and output the combined final reasoning result.
  • the model reasoning function model in the EMS can determine the multiple first-level sub-models according to the information of the multiple first-level sub-models of the first model included in the model reasoning request information, use the multiple first-level sub-models to perform reasoning respectively, and then obtain the reasoning results of the multiple first-level sub-models based on the storage addresses of the multiple first-level sub-models, and finally use the aggregation method to aggregate (or combine) the reasoning results of the multiple first-level sub-models to obtain the final reasoning result.
  • the model reasoning function model in the EMS can determine the multiple first-level sub-models based on the information of the multiple first-level sub-models of the first model included in the model reasoning request information, use the multiple first-level sub-models to perform reasoning respectively, and then obtain the reasoning results of the multiple first-level sub-models based on the storage addresses of the multiple first-level sub-models, and finally use the second-level sub-model to aggregate (or combine) the reasoning results of the multiple first-level sub-models to obtain the final reasoning result.
  • the model training function module and the model reasoning function module in the EMS can correspondingly feedback the training capability (whether to support the training of multiple models), the reasoning capability (whether to support the reasoning of multiple models) and the reasoning requirement information of the multiple models to the model management function module in the NMS; in the EMS, the model management function module adds the training instructions of multiple models and the training strategy of multiple models to the model training request sent by the model training function module, and then the model training function module can generate the first model (i.e., multiple models) based on the model training request, send the information of the first model to the model management function module, add the indication information of multiple models and the information of the sub-models of the first model to the information of the first model, and then the model management module adjusts the number of sub-models of the first model based on the information of the first model, and the reasoning requirements and reasoning capabilities of the model reasoning function module to determine the first model actually deployed.
  • the model management function module adds the training instructions of multiple models and the training strategy of multiple models to the model training request sent by
  • the scheme of the present application supports the reasoning function module of the management domain to perform multi-model combined reasoning, and supports the use of the most appropriate multi-model according to the reasoning requirements, which can effectively improve the reasoning (or analysis) effect of the model.
  • Embodiment 2 is a diagrammatic representation of Embodiment 1:
  • the solution of the present application is applied to the deployment architecture shown in FIG. 3B above, and the model training and deployment process of the OAM (NMS/EMS) domain is enhanced to support reasoning and analysis of the RAN domain based on the combination of multiple learner models.
  • This second embodiment is similar to the steps of the first embodiment above.
  • the model reasoning function module is located in the RAN/gNB, that is, the first communication device in the solution of the present application is a network element management system device (EMS for short) including a model training function module, the second communication device in the solution of the present application is a RAN/gNB including a model reasoning function module, and the third communication device in the solution of the present application is a network management system device (NMS for short) including a model management function module.
  • EMS network element management system device
  • NMS network management system device
  • the model reasoning function module in the RAN/gNB and the model management function module in the NMS device are similar.
  • the modules can interact directly with each other or forward through EMS.
  • the solution of the embodiment of the present application can be implemented when the model reasoning function model in the RAN/gNB supports multi-model reasoning. Referring to Figure 6, the specific process of the second embodiment is as follows:
  • the model management function module in the NMS sends reasoning requirement query information to the model reasoning function module in the RAN/gNB.
  • the model management function module in the NMS can forward the reasoning requirement query information to the model reasoning function module in the RAN/gNB through the EMS.
  • the reasoning capability query information is used to query (or obtain) the reasoning capability of the model reasoning function module in the RAN/gNB.
  • the model reasoning function module in the RAN/gNB sends reasoning requirement information to the model management function module in the NMS.
  • the model reasoning function module in the RAN/gNB can forward the reasoning requirement information to the model management function module in the NMS through the EMS.
  • the reasoning capability information may include: reasoning capability indication information, reasoning computing power (optional), and storage space (optional).
  • step S601b The content of the reasoning capability information in step S601b can refer to the above step S501b, which will not be described in detail here.
  • the model management function module in the NMS sends reasoning capability query information to the model reasoning function module in the RAN/gNB.
  • the model management function module in the NMS may forward the reasoning capability query information to the model reasoning function module in the RAN/gNB through the EMS.
  • the reasoning capability query information is used to query (or obtain) the reasoning capability of the model reasoning function module in the RAN/gNB.
  • the model reasoning function module in the RAN/gNB sends reasoning capability information to the model management function module in the NMS.
  • the model reasoning function module in the RAN/gNB may forward the reasoning capability information to the model management function module in the NMS through the EMS.
  • the reasoning capability information may include: reasoning capability indication information, reasoning computing power (optional), and storage space (optional).
  • the content of the reasoning capability information in step S602b can refer to the specific description in the above step S502b, which will not be repeated here.
  • step S603a The model management function module in the NMS sends training capability query information to the model training function module in the EMS.
  • the specific description of step S603a can be found in the above step S503a, which will not be repeated here.
  • step S603b The model training function module in the EMS sends the training capability information to the model management function module in the NMS.
  • the specific description of step S603b can be found in the above step S503b, which will not be repeated here.
  • steps S601a-S601b, S602a-S602b, S603a-S603b are the query and reporting process of reasoning requirement information and capability information, which are optional steps, or can be completed offline.
  • the embodiment of the present application does not specifically limit the order of executing the above-mentioned steps of querying and reporting reasoning requirement information (i.e., S601a-S601b), querying and reporting reasoning capability information (i.e., S602a-S602b), and querying and reporting training capability information (i.e., S603a-S603b).
  • step S604 The model management function module in the NMS sends a model training request to the model training function module in the EMS.
  • the specific description of step S604 can be found in the above step S504, which will not be repeated here.
  • step S605 The model training function module in the EMS performs multi-model training according to the model training request to obtain a first model (multi-model).
  • the specific description of step S605 can be found in the above step S505, which will not be repeated here.
  • step S606 The model training function module in the EMS sends a model training report to the model management function module in the NMS.
  • the specific description of step S606 can be found in the above step S506, which will not be repeated here.
  • step S607 The model management function module in the NMS determines the model to be actually deployed. That is, the model management function module in the NMS adjusts the first model and determines the sub-model to be actually deployed based on the reasoning requirement information and reasoning capability information, as well as the first model information in the training report.
  • the specific description of step S607 can be found in the above step S507, which will not be repeated here.
  • the model management function module in the NMS sends model deployment request information to the model reasoning function module in the RAN/gNB.
  • the model management function module in the NMS can forward the model deployment request information to the model reasoning function module in the RAN/gNB through the EMS.
  • the content of the model deployment request information in step S608a can refer to the specific description in the above step S508a, which will not be repeated here.
  • the model reasoning function module in the RAN/gNB sends model deployment response information to the model management function module in the NMS.
  • the model reasoning function module in the RAN/gNB can send the model deployment response information to the model management function module in the NMS through the EMS.
  • the content of the model deployment response information in step S608b can refer to the description of the model deployment response information in the above step S508b.
  • the model reasoning function module in the RAN/gNB performs reasoning based on the first model to obtain a reasoning result.
  • the model reasoning function module in the RAN/gNB inputs the data to be reasoned (i.e., input data) into the sub-models actually deployed in the first model, respectively, to obtain the reasoning results of the corresponding sub-models.
  • the reasoning results of each sub-model are combined to obtain the final reasoning result.
  • the model reasoning function model inputs the reasoning results obtained by each first-level sub-model of the first model into the second-level sub-model, and outputs the combined final reasoning result.
  • the model reasoning function module in the RAN/gNB executes step S609.
  • the specific description of the model reasoning function module in the above-mentioned EMS executing step S509 can be referred to, and will not be repeated here.
  • the solution of the present application supports the reasoning function module of the RAN domain to perform reasoning and combination of multiple models, and supports the use of the most appropriate model for reasoning according to the reasoning requirements, which can effectively improve the reasoning (or analysis) effect of the model.
  • the solution of the present application is applied to the deployment architecture shown in FIG. 3C above, and the model discovery and subscription process of NWDAF is enhanced to support NWDAF reasoning (or analysis) based on multi-model combination, thereby improving the reasoning (or analysis) effect of NWDAF.
  • the first communication device in the solution of the present application is the first NWDAF network element including the model training function module in FIG. 3C
  • the second communication device in the solution of the present application is the second NWDAF network element including the model reasoning function module in FIG. 3C.
  • the specific process of the third embodiment is as follows:
  • the first NWDAF network element sends NF registration request information to the NRF network element.
  • the NRF network element receives the NF registration request information.
  • the NF registration request information may include an inference identifier (also called an analysis identifier), an indication of the ability to support multi-model training, inference performance, inference speed (optional), and inference power consumption (optional).
  • an inference identifier also called an analysis identifier
  • an indication of the ability to support multi-model training inference performance, inference speed (optional), and inference power consumption (optional).
  • the NRF network element sends the response information of the NF registration request to the first NWDAF network element.
  • the first NWDAF network element receives the response information of the NF registration request.
  • the model training function module of the first NWDAF network element reports its own model training capability information to the NRF network element.
  • the first NWDAF network element is taken as an example to introduce its reporting of its own model training capability information to the NRF network element.
  • Each NWDAF network element including model training functions can report its own model training capability information to the NRF network element by referring to the above steps S701-S702.
  • S703 The NWDAF consumer sends analysis subscription request information to the second NWDAF network element.
  • the second NWDAF network element receives the analysis subscription request information.
  • the second NWDAF network element sends the response information of the analysis subscription request to the NWDAF consumer.
  • the NWDAF consumer receives the response information of the analysis subscription request.
  • the second NWDAF network element sends NF discovery request information to the NRF network element.
  • the NRF network element receives the NF discovery request information
  • the NF discovery request information (equivalent to the reasoning requirement information in the above-mentioned solution of the present application) may include: NF type, reasoning identifier, reasoning performance requirement, multi-model training capability indication information, reasoning speed requirement, and reasoning power consumption requirement.
  • the NRF network element sends a response message of the NF discovery request to the first NWDAF network element.
  • the first NWDAF network element receives the response message of the NF discovery request.
  • the response message of the NF discovery request includes the address of the NWDAF network element that has the model training function and supports multi-model training.
  • the NWDAF network element that has the model training function and supports multi-model training takes the first NWDAF network element as an example
  • the response information of the NF discovery request includes the address of the first NWDAF network element.
  • the second NWDAF network element sends model subscription request information to the first NWDAF network element.
  • the first NWDAF network element receives the model subscription request information.
  • the model subscription request information (equivalent to the model request information in the above-mentioned solution of the present application) includes an inference identifier, inference performance requirements, indication information for requesting multi-model inference (the indication information is used to request multiple models), inference speed requirements, and inference power consumption requirements (these are equivalent to the inference requirement information in the above-mentioned solution of the present application).
  • the first NWDAF network element sends response information of the model subscription request to the second NWDAF network element.
  • the second NWDAF network element receives the response information of the model subscription request.
  • the first NWDAF network element performs multi-model training according to the model reading request information to obtain a first model (ie, multi-model).
  • step S709 is an optional step, that is, step S709 may be executed or not executed.
  • the first NWDAF network element may, based on the inference requirement information included in the model subscription request information, Directly select a suitable multi-model (ie, the first model) from at least one trained multi-model.
  • the first NWDAF network element sends model notification information to the second NWDAF network element.
  • the second NWDAF network element receives the model notification information, which includes an inference identifier, indication information of the first model, a sub-model list of the first model (i.e., information of multiple sub-models), an aggregation method (also referred to as a combination method), a weight factor (optional), and performance information of the first model.
  • the indication information of the first model is used to indicate that the first model is a multi-model.
  • the information of each sub-model may include but is not limited to: identification information of the sub-model (such as a unique identifier or storage address information of the sub-model), the level of the sub-model (also referred to as the category of the sub-model), the performance of the sub-model, and the performance constraint of the sub-model.
  • the second NWDAF network element performs multi-model reasoning (or analysis) based on the model notification information to obtain a reasoning (or analysis) result.
  • the step S711 may be specifically executed with reference to the manner in which the model reasoning function module in the above step S509 or S609 performs reasoning based on the first model to obtain a reasoning result, which will not be described in detail here.
  • the second NWDAF network element sends notification information of the reasoning result to the NWDAF consumer, where the notification information of the reasoning result includes the reasoning identifier, the reasoning result or the analysis result.
  • the NWDAF consumer receives the notification information of the reasoning result.
  • the first NWDAF network element including the model training function module and the second NWDAF network element including the model reasoning function module can report the corresponding training capabilities and model capabilities, and reasoning requirement information to the NRF network element, and the first NWDAF network element including the model training function module and the second NWDAF network element including the model reasoning function module can exchange reasoning requirements and multi-model indication information. Therefore, the solution of this third embodiment supports NWDAF to select (or train) appropriate multi-models according to reasoning requirements, and supports NWDAF to perform multi-model combined reasoning, thereby effectively improving the reasoning (or analysis) effect of NWDAF.
  • Embodiment 4 is a diagrammatic representation of Embodiment 4:
  • the solution of this application is applied to the deployment architecture shown in FIG. 3D above, by enhancing the RAN model deployment and switching process to support intelligent reasoning (or analysis) based on multi-model reasoning.
  • the first communication device in this application solution can be a base station gNB (i.e., source gNB ⁇ target gNB) including a model training function module in FIG. 3D
  • the second communication device in this application solution is UE1 including a model reasoning function model in FIG. 3D.
  • the specific process of this fourth embodiment is as follows:
  • the target gNB sends the AI capability information of the target gNB to the source gNB.
  • the source gNB receives the AI capability information of the target gNB.
  • the source gNB sends the AI capability information of the source gNB to the target gNB.
  • the target gNB receives the AI capability information of the source gNB.
  • the above steps S801a and S801b take a source gNB and a target gNB as an example to introduce the process of exchanging AI capability information between the source gNB and the target gNB.
  • the gNB that exchanges AI capability information with the source gNB is not limited to the target gNB.
  • the target gNB and the source gNB go online, they can exchange AI capability information through the Xn interface.
  • the AI capability information may include: AI switch, training capability indication information (i.e., supporting multi-model training).
  • S802a The source gNB sends reasoning capability query information to UE1.
  • the source gNB sends reasoning capability query information to UE1, and the reasoning capability query information is used to query (or obtain) the reasoning capability information of UE1.
  • S802b UE1 sends reasoning capability information to the source gNB.
  • the UE1 may send reasoning capability information to the source gNB through the Uu interface.
  • the source gNB does not send reasoning capability query information to the UE1, but the UE1 actively reports its own reasoning capability information to the source gNB through the Uu interface.
  • the reasoning capability information of UE1 may include: AI switch, storage space size, reasoning computing power, multi-model reasoning indication information (i.e., indicating support for multi-model reasoning), and remaining power of UE1 (optional).
  • the embodiment of the present application does not specifically limit the time sequence of executing the process of exchanging AI capability information between the source gNB and the target gNB (i.e., the above steps S801a-S801b) and the process of UE1 reporting reasoning capability information to the source gNB (i.e., the above steps S802a-S802b).
  • the source gNB performs multi-model training based on the reasoning capability information of UE1 to obtain a first model (i.e., a multi-model).
  • This step S803 is an optional step, which may be executed or not.
  • the source gNB may also directly select a suitable multi-model (i.e., the first model) from at least one trained multi-model based on the reasoning capability information of the UE1.
  • a suitable multi-model i.e., the first model
  • S804 The source gNB sends notification information of the first model to UE1.
  • the source gNB may send notification information of the first model to the UE1 through the Uu interface, and correspondingly, the UE1 receives the notification information of the first model.
  • the notification information of the first model may include: the identifier of the first model, indication information of the first model (i.e., used to indicate that the first model is a multi-model), a list of sub-models of the first model (i.e., information of multiple sub-models), an aggregation method (also called a combination method), a weight factor (optional), and the performance of the first model.
  • the source gNB may perform multi-model inference based on the information of the first model to obtain an inference result.
  • This step S805a is an optional step. If the first model is a bilateral model, the source gNB executes this step S805a; if the first model is a unilateral UE model, the source gNB does not execute this step S805a.
  • the UE1 may perform multi-model reasoning based on the information of the first model to obtain a reasoning result.
  • the source gNB also determines whether it is necessary to switch the base station (gNB) accessed by UE1 based on the received signal strength reported by UE1. For example, when the signal strength received by the source gNB from UE1 is lower than the set threshold, it is determined to trigger the switch.
  • gNB base station
  • the source gNB sends an RRC connection reconfiguration message to UE1, which includes measurement configuration information.
  • the source gNB may send the RRC connection reconfiguration message to the UE1 via the Uu interface.
  • the UE1 receives the RRC connection reconfiguration message via the Uu interface.
  • S807 UE1 performs measurement based on the measurement configuration information and obtains a measurement report of UE1.
  • S808 UE1 sends a measurement report of UE1 to the source gNB.
  • the UE1 may send a measurement report of the UE1 to the source gNB via the Uu interface.
  • the source gNB receives the measurement report of the UE1 via the Uu interface.
  • the source gNB determines the target gNB based on the measurement report of UE1 and the AI capability information of the neighboring station.
  • the target gNB generally has the ability to train multiple models.
  • gNB1, gNB2, and gNB3 refer to the above steps S801a and S801b to exchange AI capability information with the source gNB, so that the source gNB obtains the AI capability information of gNB1, the AI capability information of gNB2, and the AI capability information of gNB3; and then in this step, the source gNB can select a suitable gNB1 as the target gNB from the three base stations based on the AI capability information of the three base stations and the measurement report of UE1.
  • the source gNB sends the handover request information of UE1 to the target gNB.
  • the target gNB receives the handover request information of UE1.
  • the handover request information of UE1 includes the identifier of UE1, the indication information of the first multi-model (or the indication information of using the first model), the identifier of the first model, and the information of the sub-model of the first model.
  • the target gNB performs multi-model training based on the handover request information of UE1 to obtain a second model (i.e., multi-model).
  • step S811 is an optional step, that is, step S811 may be executed or not. If the target gNB does not execute step S811, the target gNB may directly select a multi-model (i.e., the second model) from at least one trained multi-model based on the handover request information of the UE1; or the target gNB may directly use the multi-model (i.e., the first model) of the source gNB.
  • a multi-model i.e., the second model
  • the target gNB may directly use the multi-model (i.e., the first model) of the source gNB.
  • S812 The target gNB and UE1 complete random access.
  • step S812 the target gNB performs a random access process with the UE1 so that the UE1 successfully accesses the target gNB for communication.
  • the random access process can be specifically implemented by referring to the existing random access method and will not be described in detail here.
  • S813 The target gNB sends notification information of the second model to UE1.
  • the target gNB may send notification information of the second model to UE1 through the Uu interface, and correspondingly, the UE1 receives notification information of the second model through the Uu interface.
  • the notification information of the second model may include model information of the second model (such as the name, identifier, type, etc. of the second model), indication information of the second model (used to indicate that the second model is a multi-model), a list of multiple sub-models of the second model (i.e., information of multiple sub-models), aggregation method, weight factor (optional), and performance of the second model.
  • model information of the second model such as the name, identifier, type, etc. of the second model
  • indication information of the second model used to indicate that the second model is a multi-model
  • a list of multiple sub-models of the second model i.e., information of multiple sub-models
  • aggregation method i.e., information of multiple sub-models
  • the list of each sub-model may include: the level of the sub-model (also referred to as the category of the sub-model), the identification information of the sub-model (such as the storage address of the sub-model, the unique identifier).
  • the level of sub-models may include a first-level sub-model and a second-level sub-model, wherein the first-level sub-model is used for reasoning or analysis, and the second-level sub-model is used to aggregate (or combine) reasoning information of multiple first-level sub-models.
  • S814a The target gNB performs multi-model inference based on the notification information of the second model to obtain an inference result.
  • the target gNB uses the sub-models of the second model for reasoning based on the information of the second model in the notification information of the second model to obtain reasoning results of multiple sub-models, and combines them in an aggregation manner to obtain the final reasoning result; or uses the first-level sub-models of the second model for reasoning to obtain reasoning results of multiple first-level sub-models, and then uses the second-level sub-model to combine the reasoning results of these first-level sub-models to obtain the final reasoning result.
  • the step S814a is an optional step, and the step S814a may be performed or may not be performed.
  • S814b UE1 performs multi-model inference based on the notification information of the second model to obtain an inference result.
  • the model reasoning function module in UE1 can perform reasoning using the multiple sub-models based on the information of the multiple sub-models of the second model in the notification information of the second model to obtain the reasoning information of the multiple sub-models, and then use an aggregation method to aggregate or combine the reasoning information of the multiple sub-models to obtain the reasoning information of the second model; or the model reasoning function module in UE1 can use the second sub-model to aggregate or combine the reasoning information of the multiple first-level sub-models to obtain the reasoning information of the second model.
  • step S814b may also refer to the above-mentioned step S509 or S609 or S711, which will not be described in detail here.
  • the multi-model training capability information interaction between new base stations and the multi-model training/inference capability interaction between the base station and the terminal are added, and the indication information of the multi-model is added in the model notification information.
  • the present application scheme supports the UE to perform combined reasoning based on multiple models to improve the reasoning (or analysis) effect of the model.
  • the source gNB determines to trigger the switching of a new gNB
  • the source gNB will, based on the multi-model capability information of each gNB, preferably use a base station with multi-model training capability as the target gNB for switching. After the UE switches to access the target gNB, it can still perform combined reasoning based on multiple models to improve the reasoning (or analysis) effect of the model.
  • the solution of the present application is applied to the deployment architecture shown in FIG. 3E above, that is, a scenario in which joint training of a base station and a UE is performed in a bilateral model scenario.
  • the first communication device and the second communication device in the solution of the present application may be a base station (gNB) or a UE (e.g., UE1) including a model training function model and a model reasoning function module.
  • gNB base station
  • UE1 e.g., UE1
  • FIG. 9 the specific process of the fifth embodiment is as follows:
  • the target gNB sends the AI capability information of the target gNB to the source gNB.
  • the AI capability information may include: AI switch, support for multi-model training capability indication (yes/no).
  • the source gNB sends the AI capability information of the source gNB to the target gNB.
  • the source gNB and multiple gNBs can exchange their respective AI capability information through the corresponding Xn interface.
  • This embodiment 5 takes the source gNB and the target gNB as an example.
  • S902a The source gNB sends capability query information to UE1.
  • the source gNB may send capability query information to the source UE1 through the Uu interface, and the capability query information is used to query (or request) the capability information of the UE1.
  • the UE1 receives the capability query information through the Uu interface.
  • S902b UE1 sends UE1’s capability information to the source gNB.
  • UE1 may send the capability information of UE1 to the source gNB through the Uu interface, and correspondingly, the source gNB receives the capability information of UE1 through the Uu interface.
  • the source gNB does not send the capability query information to UE1, but UE1 actively sends the capability information of UE1 to the source gNB.
  • the capability information of UE1 may include: AI switch, storage space, computing power, support for multi-model reasoning indication information, and remaining power (optional).
  • the embodiment of the present application does not specifically limit the order of executing the steps of exchanging AI capability information between the above-mentioned base stations (ie, S901a-S901b), and the steps of querying and reporting UE1 capability information (ie, S902a-S902b).
  • the source gNB and UE1 negotiate a joint training strategy.
  • the joint training strategy may include: multi-model training indication information, multi-model joint training mode, data processing strategy, number of sub-models, model type, and hyper-parameter configuration.
  • the multi-model training indication information is used to indicate the training of multiple models.
  • the multi-model joint training mode can be one-to-one, many-to-one, one-to-many, or many-to-many; one-to-one means that the multiple models on the source gNB and the UE1 side are both regarded as an overall model, and the output of the gNB's overall model is used as the input of the UE1's overall model; many-to-one means that the multiple models on the UE1 side are regarded as an overall model, and the source gNB
  • the multiple outputs of the multi-model on the source gNB side are used as the input of the overall model of UE1; one-to-many means that the multiple models on the source gNB side are used as an overall model, and the output of the overall model of the source gNB is used as the input of the multi-model of UE1; many-to-many means that the multiple outputs of the multi-model on the source gNB side are used
  • the data processing strategy can be input data sampling, feature sampling, etc.
  • the input data sampling and feature sampling can be specifically described in the above embodiments and will not be described in detail here.
  • the source gNB performs multi-model training to obtain a multi-model of the source gNB.
  • S904b UE1 performs multi-model training to obtain a multi-model of UE1.
  • steps S904a and S904b can be executed synchronously, and when the source gNB and the UE1 perform multi-model training respectively, the intermediate parameters of their respective multi-model training, such as gradients or intermediate inference results, are exchanged according to the joint training strategy in the above step S903.
  • the source gNB performs inference based on the multiple models and aggregation method of the source gNB to obtain an inference result.
  • the source gNB uses its own trained sub-models to perform reasoning separately, and then uses the aggregation method to combine the reasoning results of the sub-models to obtain the final reasoning result.
  • S905b UE1 performs reasoning based on the multiple models and aggregation method of UE1 to obtain a reasoning result.
  • the UE1 uses each sub-model trained by itself to perform reasoning respectively, and then uses an aggregation method to combine the reasoning results of each sub-model to obtain a final reasoning result.
  • step S905b may be executed first, and then step S905a may be executed, that is, after UE1 obtains the final inference result using the multi-model trained by itself, the final inference result on the UE1 side is reported to the source gNB.
  • the source gNB may use the final inference result on the UE1 side as the input of the multi-model of the source gNB to obtain the final inference result on the source gNB side.
  • S906 The source gNB sends measurement configuration information to UE1.
  • UE1 receives the measurement configuration information.
  • S907 UE1 performs measurement based on the measurement configuration information and obtains a measurement report of UE1.
  • S908 UE1 sends a measurement report of UE1 to the source gNB.
  • the source gNB receives the measurement report of UE1.
  • the source gNB selects the target gNB based on the measurement report of UE1 and the AI capability of the neighboring station.
  • the source gNB preferably selects a base station with multi-model training capability as the target gNB based on the measurement report of UE1 and the AI capability of the neighboring base station.
  • the source gNB sends the handover request information of UE1 to the target gNB.
  • the target gNB receives the handover request information of UE1, and the UE handover request information includes the identification information of UE1 and the multi-model indication information, and the multi-model indication information is used to request the use of the multi-model.
  • the source gNB may also send its own multi-model and the usage information of the multi-model of the source gNB to the target gNB, then the target gNB and UE1 directly reuse the multi-model of the source gNB and the multi-model previously trained by UE1, and there is no need to perform multi-model training separately, that is, the following steps S913a and S913b are not executed.
  • the handover request information of the UE1 sent by the source gNB to the target gNB also includes the identifier of the multi-model trained by the source gNB, the list of the multi-model, the aggregation method, and the weight factor (optional).
  • the source gNB may also send the joint training strategy previously negotiated with the UE1 to the target gNB. In this case, there is no need to repeatedly negotiate the joint training strategy between the target gNB and the UE1, that is, the following step S912 is not executed.
  • S911 The target gNB completes random access with UE1.
  • the target gNB performs a random access process with the UE1 so that the UE1 successfully accesses the target gNB for communication.
  • the specific random access process is implemented with reference to the existing random access process and will not be described in detail here.
  • S912 The target gNB and UE1 negotiate a joint training strategy.
  • the joint training strategy may include: multi-model training instructions, multi-model joint training mode, data processing strategy, number of sub-models, model type, and hyper-parameter configuration.
  • Step 912 is an optional step. If in the above step S910, the switching request information of the UE1 includes the joint training strategy, then the target gNB and the UE1 do not need to renegotiate the joint training strategy.
  • S913a UE1 performs multi-model training to obtain a multi-model of UE1.
  • the target gNB performs multi-model training to obtain a multi-model of the target gNB.
  • the target gNB and UE1 when the target gNB and UE1 perform multi-model training respectively, they exchange intermediate parameters of the multi-model training, such as gradients or intermediate inference results, according to the negotiated joint training strategy.
  • S914a UE1 performs reasoning and combination based on each sub-model of its own multi-model to obtain a reasoning result.
  • the UE1 uses each sub-model of the multi-model trained by itself to perform reasoning respectively, and then uses an aggregation method to combine the reasoning results of each sub-model to obtain a final reasoning result.
  • S914b The target gNB performs inference and combination based on each sub-model of its own multi-model to obtain an inference result.
  • the target gNB uses each sub-model of its own trained multi-model (or each sub-model of the source gNB's multi-model) to perform inference respectively, and then uses the aggregation method to combine the inference results of each sub-model to obtain the final inference result.
  • step S914a after UE1 obtains the final inference result using its own trained multi-model, it also reports the final inference result on the UE1 side to the target gNB.
  • the target gNB can use the final inference result on the UE1 side as the input of the multi-model of the target gNB (or the multi-model of the source gNB) to obtain the final inference result on the target gNB side.
  • This fifth embodiment supports multi-model combined reasoning in the scenario where the base station and UE jointly train a bilateral model, which can improve the reasoning (or analysis) effect of the model.
  • a base station with multi-model training capability is preferred, and the UE can still perform combined reasoning based on multiple models after switching.
  • this embodiment six another multi-model usage scenario is mainly aimed at, that is, the model training function can encapsulate multiple models into a large model, the model reasoning function can be unaware of the internal structure of the large model (i.e., multiple models), and the model selection and deployment process is added.
  • This embodiment six is introduced with a general logical architecture, as shown in Figure 10, the specific process of this embodiment six is as follows:
  • S1001a The model management function sends reasoning requirement query information to the model reasoning function.
  • the model management function forwards the reasoning requirement query information to the model reasoning function through the model training function.
  • This step S1001a can be cross-referenced with the above-mentioned step S501a or S601a.
  • S1001b The model reasoning function sends reasoning requirement information to the model management function.
  • the model reasoning function forwards the reasoning requirement information to the model management function via the model training function.
  • the model reasoning function proactively reports (ie, sends) the reasoning requirement information to the model management function.
  • the reasoning requirement information includes: reasoning type requirement, reasoning accuracy requirement, reasoning speed requirement, and reasoning energy consumption requirement.
  • the reasoning accuracy requirement, reasoning speed requirement, and reasoning energy consumption requirement can also be collectively referred to as reasoning performance requirements.
  • the reasoning speed requirement can also be called the reasoning latency requirement, which indicates the requirement for reasoning execution time, for example: a single reasoning execution time is less than 1s; the reasoning energy consumption requirement indicates the requirement for reasoning energy consumption, for example: a single reasoning consumes less than 5J of energy.
  • step S1001b may refer to the above step S501b or S601b, which will not be repeated here.
  • S1002a The model management function sends reasoning capability query information to the model reasoning function.
  • the model management function forwards the reasoning capability query information to the model reasoning function via the model training function.
  • the step S1002a can be cross-referenced with the above-mentioned step S502a or S602a.
  • S1002b The model reasoning function sends reasoning capability information to the model management function.
  • the model reasoning function forwards the reasoning capability information to the model management function via the model training function.
  • the model reasoning function proactively reports (ie, sends) the reasoning capability information to the model management function.
  • the reasoning capability information includes: reasoning computing power (optional), storage space (optional), power, etc.
  • the reasoning computing power indicates the computing power information available at the reasoning function, including available hardware resource information and hardware resource utilization.
  • the hardware resource information can be the original hardware information, including hardware type, number of cores, processing frequency, etc., or it can be the quantified computing power.
  • step S1002b may refer to the above step S502b or S602b, which will not be repeated here.
  • step S1003a The model management function sends training capability query information to the model training function.
  • This step S1003a can be cross-referenced with the above-mentioned step S503a or S603a.
  • the model training function sends training capability information to the model management function.
  • the model training function can actively report the training capability information to the model management function.
  • the training capability information includes training computing power and the upper limit of model accuracy that can be achieved.
  • the step S1003b may be cross-referenced with the above-mentioned step S503b or S603b.
  • the model management function sends model training request information to the model training function.
  • the model training request information includes: model identification/inference type, and model training strategy information.
  • the training strategy information of the model is determined based on the reasoning requirement information, the reasoning capability information and the training capability information; the training strategy information of the model is used to indicate the training method, which may include: multiple model training instruction information, data processing strategy, training algorithm instruction, etc.
  • the model management function may also send the original reasoning requirement information and reasoning capability information to the model training function, and the model training function itself determines the model training strategy based on the reasoning requirement information, reasoning capability information, and training capability information.
  • step S1004 may refer to the above-mentioned S504 or step S604, which will not be repeated here.
  • the model training function performs model training according to the model training request information to obtain a first model (ie, a multi-model).
  • step S1005 may be described in detail with reference to step S505 or S605, and will not be repeated here.
  • the model training function sends a model training report to the model management function.
  • the model training report includes: identification information of the first model, accuracy of the model, accuracy constraints, size of the first model, inference computing power of the first model, inference speed of the first model, and inference energy consumption of the first model.
  • step S1006 may be described in detail with reference to step S506 or S606, and will not be repeated here.
  • the model management function can determine the actually deployed model based on the original reasoning requirement information, reasoning capability information and the model training report.
  • the model management function in step S1004 also instructs the model training function to train multiple specified models (which may be multiple models similar to the first model)
  • the model training report fed back by the model training function may include multiple models with different performances, that is, multiple multiple models similar to the first model, but the performance of each model is different; the model management function can determine the actually deployed model (for example, the first model) from the multiple models with different performances based on the reasoning requirement information, reasoning capability information and the model training report.
  • S1008a The model management function sends model deployment request information to the model reasoning function.
  • the model reasoning function receives the model deployment request information, and the model deployment request information includes the actually deployed model information (such as the first model information).
  • the model deployment request information includes: identification information of the first model actually deployed.
  • the model deployment request information may also include other information of the first model, such as the accuracy of the first model, the accuracy constraint of the first model, the size of the first model, the reasoning computing power of the first model, the reasoning speed of the first model, and the reasoning energy consumption of the first model.
  • S1008b The model reasoning function sends model deployment response information to the model management function.
  • the model management function receives the model deployment response information to determine (or know) that the model reasoning function has completed the model deployment.
  • the model reasoning function performs reasoning based on the actually deployed model to obtain a reasoning result.
  • the model reasoning function uses the first model to perform reasoning based on the first model information (i.e., the model information actually deployed) to obtain a reasoning result.
  • the model reasoning function inputs the information to be reasoned into the first model and outputs a reasoning result.
  • a process of querying/reporting reasoning requirements is added between the model management function and the model reasoning function.
  • the reasoning requirements include reasoning speed requirements, reasoning energy consumption requirements, etc.
  • a process of querying/reporting reasoning capabilities is also added between the model management function and the model reasoning function.
  • the reasoning capabilities include reasoning computing power, storage space, power, etc.
  • a process of querying/reporting training capabilities is added between the model management function and the model training function.
  • the model training report sent by the model training function to the model management function includes relevant information of the model, such as model size, model reasoning computing power, model reasoning speed, model reasoning energy consumption, etc.
  • the model management function can determine the model training strategy based on the reasoning requirements, reasoning capabilities, and training capabilities.
  • the model management function can also determine the actual deployed model based on the reasoning requirements, reasoning capabilities, and the information of the model in the model training report. Therefore, in the sixth embodiment, querying/reporting reasoning requirements, reasoning capabilities, and training capabilities are added, and the trained model can be determined based on the reasoning requirements, reasoning capabilities, and training capabilities, and the best model for actual deployment can be determined based on the information of the reasoning requirements, reasoning capabilities, and actual models, thereby improving the reasoning (or analysis) effect of the model.
  • the application scenario of this seventh embodiment is similar to that of the sixth embodiment, except that there is no model management function in the seventh embodiment, and only the interaction between the model training function and the model reasoning function is involved.
  • the seventh embodiment is described in a general logical architecture, and specifically, it can be applied to the deployment architecture 3C-3E. Referring to FIG11 , the specific process of the seventh embodiment is as follows:
  • the model inference function sends model training request information to the model training function.
  • the model training function receives the model training request information of the model inference function, and the model training request information includes: Model identification/inference type, inference accuracy requirement, inference speed requirement, inference energy consumption requirement, and request for multiple model instructions.
  • the value of requesting multiple models indication information is yes/no, to indicate whether it is necessary to provide multiple models that meet the reasoning requirements.
  • the requesting multiple models indication information indicates yes (that is, indicating that multiple models that meet the reasoning requirements need to be provided), it can further indicate the number of models required, for example, indicating that 5 models are required.
  • the model training function performs model training according to the request of the model inference function to obtain a first model (ie, a multi-model).
  • step S1102 is an optional step. If the first model is a multi-model that has been trained in advance by the model training function, then step S1102 may not be performed.
  • S1103 The model training function sends the first model information to the model reasoning function.
  • the first model information includes identification information of the first model (such as name, type), accuracy of the first model, accuracy constraint, size of the first model, inference computing power of the first model, inference speed of the first model, and inference energy consumption of the first model.
  • step S1101 indicates that multiple models are required (which may be multiple models similar to the first model)
  • the first model information in step S1103 is a list containing multiple models with different performances, that is, it contains multiple model information similar to the first model, but the performance of each model is different.
  • the model reasoning function selects a suitable model based on the reasoning requirement information, the reasoning capability information and the first model information.
  • step S1104 is executed.
  • the model reasoning function selects a suitable model from the multiple models similar to the first model based on the reasoning requirement information, the reasoning capability information and the list.
  • the model reasoning function performs model reasoning to obtain reasoning results.
  • the model reasoning function may use the first model to perform reasoning to obtain a reasoning result, or the model reasoning function may use the model selected in the above step S1104 to perform reasoning to obtain a reasoning result.
  • the inference speed requirement, inference energy consumption requirement, and request for multiple model indication information are added to the model training request information (or model request information) sent by the model inference function;
  • the model training function can determine the returned model information according to the inference requirements, and add other information of the model to the model information, such as the size of the model, the model inference computing power, the model inference speed, and the model inference energy consumption;
  • the model inference function can determine the model actually used based on the inference requirements, the inference ability, and the actual model information.
  • This embodiment seven supports the determination of model information according to the inference requirements, and supports the selection of the most appropriate model for inference (or analysis) based on the inference requirements, the inference ability, and the actual model information, thereby improving the inference (or analysis) effect of the model.
  • the communication device provided in the embodiment of the present application is described below.
  • an embodiment of the present application provides a communication device, which can be used to perform the operation performed by the first communication device in the above method embodiment.
  • the communication device can also be a first communication device, a processor of the first communication device, or a chip.
  • the device includes a module or unit corresponding to the method/operation/step/action described by the first communication device in the above embodiment, and the module or unit can be a hardware circuit, or software, or a hardware circuit combined with software.
  • the communication device can have a structure as shown in Figure 12.
  • the communication device 1200 may include a communication unit 1201 (also referred to as a transceiver unit) and a processing unit 1202.
  • the communication unit 1201 is equivalent to a communication module (or a transceiver module), and the processing unit 1202 is equivalent to a processing module.
  • the processing unit 1202 may be used to call the communication unit 1201 to perform a receiving and/or sending function, and the communication unit 1201 may implement a corresponding communication function.
  • the communication unit 1201 may include a receiving unit and/or a sending unit.
  • the receiving unit may be used to receive information and/or data
  • the sending unit may be used to send information and/or data.
  • the communication unit 1201 may also be referred to as a communication interface or a transceiver module.
  • the communication device 1200 may further include a storage unit 1203, which is equivalent to a storage module and can be used to store instructions and/or data.
  • the processing unit 1202 can read the instructions and/or data in the storage module so that the communication device implements the aforementioned method embodiment.
  • the communication device 1200 can be used to perform the actions performed by the first communication device in the above method embodiment.
  • the communication device 1200 can be the first communication device or a component that can be configured in the first communication device.
  • the communication unit 1201 is used to perform the sending-related operations on the first communication device side in the above method embodiment
  • the processing unit 1202 is used to perform the processing-related operations on the first communication device side in the above method embodiment.
  • the communication unit 1201 may include a sending unit and a receiving unit.
  • the sending unit is used to perform the sending operation in the above method embodiment.
  • the receiving unit is used to perform the receiving operation in the above method embodiment.
  • the communication device 1200 may include a sending unit but not a receiving unit.
  • the communication device 1200 may include a receiving unit but not a sending unit. Specifically, it may depend on whether the above solution executed by the communication device 1200 includes a sending action and a receiving action.
  • the communication device 1200 is used to execute the actions executed by the first communication device in the embodiment shown in FIG. 4A or FIG. 4B above.
  • the communication unit 1201 is used to receive model request information, and the model request information includes reasoning requirement information; the processing unit 1202 is used to determine the first model according to the model request information, and the first model is a multi-model; the communication unit 1201 is also used to send first information, and the first information includes information of the first model.
  • the processing unit 1202 in the above embodiment may be implemented by at least one processor or processor-related circuits.
  • the communication unit 1201 may be implemented by a transceiver or transceiver-related circuits.
  • the storage unit may be implemented by at least one memory.
  • an embodiment of the present application provides a communication device, which can be used to perform the operation performed by the second communication device in the above method embodiment.
  • the communication device can also be a second communication device, a processor of the second communication device, or a chip.
  • the device includes a module or unit corresponding to the method/operation/step/action described by the second communication device in the above embodiment, and the module or unit can be a hardware circuit, or software, or a hardware circuit combined with software.
  • the communication device can also have a structure as shown in Figure 12.
  • the communication device 1200 may include a communication unit 1201 (also referred to as a transceiver unit) and a processing unit 1202.
  • the communication unit 1201 is equivalent to a communication module (or a transceiver module), and the processing unit 1202 is equivalent to a processing module.
  • the processing unit 1202 may be used to call the communication unit 1201 to perform a receiving and/or sending function, and the communication unit 1201 may implement a corresponding communication function.
  • the communication unit 1201 may include a receiving unit and/or a sending unit.
  • the receiving unit may be used to receive information and/or data
  • the sending unit may be used to send information and/or data.
  • the communication unit 1201 may also be referred to as a communication interface or a transceiver module.
  • the communication device 1200 may further include a storage unit 1203, which is equivalent to a storage module and can be used to store instructions and/or data.
  • the processing unit 1202 can read the instructions and/or data in the storage module so that the communication device implements the aforementioned method embodiment.
  • the communication device 1200 can be used to perform the actions performed by the second communication device in the above method embodiment.
  • the communication device 1200 can be a first communication device or a component that can be configured in the second communication device.
  • the communication unit 1201 is used to perform the sending-related operations on the second communication device side in the above method embodiment, and the processing unit 1202 is used to perform the processing-related operations on the second communication device side in the above method embodiment.
  • the communication unit 1201 may include a sending unit and a receiving unit.
  • the sending unit is used to perform the sending operation in the above method embodiment.
  • the receiving unit is used to perform the receiving operation in the above method embodiment.
  • the communication device 1200 may include a sending unit but not a receiving unit.
  • the communication device 1200 may include a receiving unit but not a sending unit. Specifically, it may depend on whether the above solution executed by the communication device 1200 includes a sending action and a receiving action.
  • the communication device 1200 is used to execute the actions executed by the second communication device in the embodiment shown in FIG. 4A or FIG. 4B above.
  • the communication unit 1201 is used to receive second information, which includes information of a first model, where the first model is determined based on reasoning requirement information, and the first model is a multi-model; the processing unit 1202 is used to obtain reasoning information of the first model based on the information of the first model.
  • the processing unit 1202 in the above embodiment may be implemented by at least one processor or processor-related circuits.
  • the communication unit 1201 may be implemented by a transceiver or transceiver-related circuits.
  • the storage unit may be implemented by at least one memory.
  • an embodiment of the present application provides a communication device, which can be used to perform the operations performed by the third communication device in the above method embodiment.
  • the communication device can also be a third communication device, a processor of the third communication device, or a chip.
  • the device includes a module or unit corresponding to the method/operation/step/action described by the third communication device in the above embodiment.
  • the module or unit can be a hardware circuit, or software, or a combination of a hardware circuit and software.
  • the communication device can also have the following The structure shown in Figure 12.
  • the communication device 1200 may include a processing unit 1202, and optionally, a communication unit 1201.
  • the communication unit 1201 is equivalent to a transceiver module
  • the processing unit 1202 is equivalent to a processing module.
  • the processing unit 1202 may be used to call the communication unit 1201 to perform a receiving and/or sending function, and the communication unit 1201 may implement a corresponding communication function.
  • the communication unit 1201 may include a receiving unit and/or a sending unit.
  • the receiving unit may be used to receive information and/or data
  • the sending unit may be used to send information and/or data.
  • the communication unit 1201 may also be called a communication interface or a transceiver module.
  • the communication device 1200 may further include a storage unit 1203, which is equivalent to a storage module and can be used to store instructions and/or data.
  • the processing unit 1202 can read the instructions and/or data in the storage module so that the communication device implements the aforementioned method embodiment.
  • the communication device 1200 may be used to perform the actions performed by the third communication device in the above method embodiment.
  • the communication device 1200 may be a third communication device or a component that may be configured in a third communication device.
  • the communication unit 1201 is used to perform the sending-related operations on the third communication device side in the above method embodiment
  • the processing unit 1202 is used to perform the processing-related operations on the third communication device side in the above method embodiment.
  • the communication unit 1201 may include a sending unit and a receiving unit.
  • the sending unit is used to perform the sending operation in the above method embodiment.
  • the receiving unit is used to perform the receiving operation in the above method embodiment.
  • the communication device 1200 may include a sending unit but not a receiving unit.
  • the communication device 1200 may include a receiving unit but not a sending unit. Specifically, it may depend on whether the above solution executed by the communication device 1200 includes a sending action and a receiving action.
  • the communication device 1200 is used to execute the actions executed by the third communication device in the embodiment shown in FIG. 4A above.
  • the communication unit 1201 is used to receive training capability indication information of a first communication device; the training capability indication information is used to indicate that the first communication device supports multi-model training; and receive reasoning requirement information and reasoning capability information of a second communication device; the reasoning capability information includes reasoning capability indication information, and the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning;
  • the communication unit 1201 is further configured to send model request information to the first communication device, wherein the model request information includes the reasoning requirement information; and receive first information from the first communication device, wherein the first information includes information of a first model, wherein the first model is a multi-model, and the first model is determined according to the reasoning requirement information;
  • the communication unit 1201 is further configured to send second information to the second communication device, where the second information includes information of the first model.
  • the processing unit 1202 in the above embodiment may be implemented by at least one processor or processor-related circuits.
  • the communication unit 1201 may be implemented by a transceiver or transceiver-related circuits.
  • the storage unit may be implemented by at least one memory.
  • the embodiment of the present application also provides a communication device, as shown in FIG13, which is a schematic diagram of a communication device provided by the present application.
  • the communication device 1300 can be the first communication device, the processor of the first communication device, or the chip in the above embodiment.
  • the communication device 1300 can be used to perform the operation performed by the first communication device in the above method embodiment.
  • the communication device 1300 includes: a processor 1302.
  • the communication device 1300 can also include a communication interface 1301, a memory 1303, and a communication bus 1304.
  • the communication interface 1301, the processor 1302, and the memory 1303 can be connected to each other through the communication bus 1304;
  • the communication bus 1304 can be a peripheral component interconnect standard (PCI) bus or an extended industry standard architecture (EISA) bus, etc.
  • the communication bus 1304 can be divided into an address bus, a data bus, a control bus, etc.
  • FIG13 shows only one thick line, but this does not mean that there is only one bus or one type of bus.
  • Processor 1302 may be a CPU, a microprocessor, an ASIC, or one or more integrated circuits for controlling the execution of the program of the present application.
  • the communication interface 1301 uses any transceiver-like device to communicate with other devices or communication networks, such as Ethernet, radio access network (RAN), wireless local area networks (WLAN), wired access networks, etc.
  • RAN radio access network
  • WLAN wireless local area networks
  • wired access networks etc.
  • the memory 1303 may be a ROM or other type of static storage device that can store static information and instructions, a RAM or other type of dynamic storage device that can store information and instructions, or an electrically erasable programmable read-only memory (EPROM).
  • PROM Electrically erasable programmable read-only memory
  • EEPROM Electrically erasable programmable read-only memory
  • CD-ROM compact disc read-only memory
  • optical disk storage including compressed optical disk, laser disk, optical disk, digital versatile disk, Blu-ray disk, etc.
  • magnetic disk storage medium or other magnetic storage device or any other medium that can be used to carry or store the desired program code in the form of instructions or data structures and can be accessed by the computer, but is not limited to this.
  • the memory can be independent and connected to the processor through the communication bus 1304.
  • the memory can also be integrated with the processor.
  • the memory 1303 is used to store computer-executable instructions for executing the solution of the present application, and the execution is controlled by the processor 1302.
  • the processor 1302 is used to execute the computer-executable instructions stored in the memory 1303, thereby realizing the communication method provided in the above embodiment of the present application.
  • the computer-executable instructions in the embodiments of the present application may also be referred to as application code, which is not specifically limited in the embodiments of the present application.
  • FIG14 is a schematic diagram of the device structure of a chip provided in an embodiment of the present application.
  • the chip 1400 includes an interface circuit 1401 and one or more processors 1402.
  • the chip 1400 may also include a bus.
  • the processor 1402 may be an integrated circuit chip with signal processing capabilities.
  • each step of the above-mentioned eye tracking method can be completed by an integrated logic circuit of hardware in the processor 1402 or instructions in the form of software.
  • the above-mentioned processor 1402 may be a general-purpose processor, a digital communicator (DSP), an application-specific integrated circuit (ASIC), a field programmable gate array (FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components.
  • DSP digital communicator
  • ASIC application-specific integrated circuit
  • FPGA field programmable gate array
  • the various methods and steps disclosed in the embodiments of the present application can be implemented or executed.
  • the general-purpose processor may be a microprocessor or the processor may also
  • the interface circuit 1401 can be used to send or receive data, instructions or information.
  • the processor 1402 can use the data, instructions or other information received by the interface circuit 1401 to process, and can send the processing completion information through the interface circuit 1401.
  • the chip further includes a memory 1403, which may include a read-only memory and a random access memory, and provides operation instructions and data to the processor.
  • a portion of the memory 1403 may also include a non-volatile random access memory (NVRAM).
  • NVRAM non-volatile random access memory
  • the memory stores executable software modules or data structures
  • the processor can perform corresponding operations by calling operation instructions stored in the memory (the operation instructions can be stored in the operating system).
  • the chip can be used in the first communication device (second communication device, third communication device) involved in the embodiment of the present application.
  • the interface circuit 1401 can be used to output the execution result of the processor 1402.
  • the communication method provided by one or more embodiments of the present application can refer to the aforementioned embodiments, which will not be repeated here.
  • interface circuit 1401 and the processor 1402 can be implemented through hardware design, software design, or a combination of hardware and software, and there is no limitation here.
  • An embodiment of the present application also provides a computer-readable storage medium, on which computer instructions for implementing the method executed by the first communication device in the above method embodiment are stored, and/or computer instructions for implementing the method executed by the second communication device in the above method embodiment are stored, and/or computer instructions for implementing the method executed by the third communication device in the above method embodiment are stored.
  • the computer when the computer program is executed by a computer, the computer can implement the method performed by the first communication device in the above method embodiment.
  • An embodiment of the present application also provides a computer program product comprising instructions, which, when executed by a computer, enables the computer to implement the method performed by the first communication device in the above method embodiment, and/or when executed by a computer, enables the computer to implement the method performed by the second communication device in the above method embodiment, and/or when executed by a computer, enables the computer to implement the method performed by the third communication device in the above method embodiment.
  • An embodiment of the present application also provides a chip device, including a processor, for calling a computer program or computer instruction stored in the memory so that the processor executes a communication method of the embodiment shown in FIG. 4A or FIG. 4B above.
  • the input of the chip device corresponds to the receiving operation in the embodiment shown in FIG. 4A or FIG. 4B
  • the output of the chip device corresponds to the sending operation in the embodiment shown in FIG. 4A or FIG. 4B .
  • the processor is coupled to the memory via an interface.
  • the chip device further comprises a memory, in which computer programs or computer instructions are stored.
  • the processor mentioned in any of the above places may be a general-purpose central processing unit, a microprocessor, an application-specific integrated circuit (ASIC), or one or more integrated circuits for controlling the execution of a program of a communication method of the embodiment shown in FIG. 4A or FIG. 4B.
  • the memory mentioned in any of the above places may be a read-only memory (ROM) or other types of static storage devices that can store static information and instructions, a random access memory (RAM), etc.
  • the communication devices may also include a hardware layer, an operating system layer running on the hardware layer, and an application layer running on the operating system layer.
  • the hardware layer may include hardware such as a central processing unit (CPU), a memory management unit (MMU), and a memory (also called main memory).
  • the operating system of the operating system layer may be any one or more computer operating systems that implement business processing through processes, such as Linux operating system, Unix operating system, Android operating system, iOS operating system, or Windows operating system.
  • the application layer may include applications such as browsers, address books, word processing software, and instant messaging software.
  • each functional module in each embodiment of the present application may be integrated into a processor, or may exist physically separately, or two or more modules may be integrated into one module.
  • the above-mentioned integrated modules may be implemented in the form of hardware or in the form of software functional modules.
  • Computer-readable media include computer storage media and communication media, wherein the communication media include any medium that facilitates the transmission of a computer program from one place to another.
  • the storage medium can be any available medium that a computer can access.
  • a computer-readable medium may include RAM, ROM, electrically erasable programmable read-only memory (EEPROM), compact disc read-only memory (CD-ROM) or other optical disc storage, disk storage medium or other magnetic storage device, or any other medium that can be used to carry or store the desired program code in the form of an instruction or data structure and can be accessed by a computer.
  • EEPROM electrically erasable programmable read-only memory
  • CD-ROM compact disc read-only memory
  • Any connection can be appropriately a computer-readable medium.
  • disk and disc include compact disc (CD), laser disc, optical disc, digital video disc (DVD), floppy disk, and Blu-ray disc, where disks usually copy data magnetically and discs use lasers to copy data optically.

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Theoretical Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • Medical Informatics (AREA)
  • Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • Mobile Radio Communication Systems (AREA)

Abstract

The present application discloses a communication method and device. The method comprises: a first communication device receives model request information, the model request information comprising inference demand information; the first communication device determines a first model according to the model request information, the first model being a multi-model; the first communication device sends first information, the first information comprising information of the first model. Therefore, after receiving the model request information, the first communication device determines a suitable multi-model (i.e., the first model) according to the inference demand information in the model request information, and then sends the information of the first model by means of the first information; after receiving the information of the first model, on the basis of the information of the first model, an inference end carries out multi-model inference by using the first model, so as to obtain an inference result having relatively high accuracy. Therefore, the method can effectively improve the use effect of a model, thereby ensuring the performance of intelligent inference (or analysis).

Description

一种通信方法和装置A communication method and device
相关申请的交叉引用CROSS-REFERENCE TO RELATED APPLICATIONS
本申请要求在2023年01月29日提交中国专利局、申请号为202310115449.1、申请名称为“一种通信方法和装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application filed with the China Patent Office on January 29, 2023, with application number 202310115449.1 and application name “A Communication Method and Device”, the entire contents of which are incorporated by reference in this application.
技术领域Technical Field
本申请涉及本发明涉及通信技术领域,尤其涉及一种通信方法和装置。The present application relates to the field of communication technology, and in particular to a communication method and device.
背景技术Background technique
随着网络的智能化和自动化水平的不断提升,人工智能(artificial intelligence,AI)和机器学习(machine learning,ML)技术应用的领域也越来越广泛,例如在管理、核心网(core network,CN)和无线接入网(radio access network,RAN)等领域;由于当前模型训练和推理的基本架构已经确定,且支持各域的模型使用,因而如何提高模型的使用效果,以保证智能分析性能是目前进一步需要探讨的问题。With the continuous improvement of network intelligence and automation, the application fields of artificial intelligence (AI) and machine learning (ML) technologies are becoming more and more extensive, such as in management, core network (CN) and radio access network (RAN). Since the basic architecture of model training and reasoning has been determined and supports the use of models in various domains, how to improve the use effect of the model to ensure the performance of intelligent analysis is an issue that needs to be further explored.
为了提高模型的使用效果,目前方案中提出由模型训练功能网元(或实体)向模型推理功能网元(或实体)提供多个不同性能的模型,模型推理功能网元(或实体)从中选择合适的模型进行推理。然而,这种方案中存在一些明显缺陷,例如,模型推理功能网元(或实体)无法确定模型训练功能网元(或实体)的训练能力,若向模型训练功能网元(或实体)请求多个模型时可能会请求失败,另外,由于模型训练功能网元(或实体)训练的模型数量不确定,且所提供的模型不一定适用于模型推理功能网元(或实体)的推理,这些均可能导致方案不可行或者推理的效果不理想,从而不能有效地提高模型的智能分析性能。In order to improve the use effect of the model, the current solution proposes that the model training functional network element (or entity) provides multiple models with different performances to the model reasoning functional network element (or entity), and the model reasoning functional network element (or entity) selects a suitable model for reasoning. However, there are some obvious defects in this solution. For example, the model reasoning functional network element (or entity) cannot determine the training capability of the model training functional network element (or entity). If multiple models are requested from the model training functional network element (or entity), the request may fail. In addition, since the number of models trained by the model training functional network element (or entity) is uncertain, and the provided model may not be suitable for the reasoning of the model reasoning functional network element (or entity), these may make the solution infeasible or the reasoning effect unsatisfactory, thereby failing to effectively improve the intelligent analysis performance of the model.
鉴于上述可知,在多模型的使用中,如何获取合适的模型进行推理,以提高模型分析性能是目前亟待解决的技术问题之一。In view of the above, when using multiple models, how to obtain a suitable model for reasoning to improve model analysis performance is one of the technical problems that needs to be solved urgently.
发明内容Summary of the invention
本申请提出一种通信方法和装置,可以有效地提高模型的使用效果,以保证智能推理(或分析)的性能。The present application proposes a communication method and device that can effectively improve the use effect of the model to ensure the performance of intelligent reasoning (or analysis).
第一方面,本申请提供一种通信方法,该方法可以由第一通信装置执行,也可以由第一通信装置的部件(例如处理器、芯片、或芯片系统等)执行,本申请对此不做具体限定。该方法具体可包括以下步骤:第一通信装置接收模型请求信息,该模型请求信息中包括推理需求信息;该第一通信装置根据该模型请求信息,确定第一模型,该第一模型为多模型;该第一通信装置发送第一信息,该第一信息中包括该第一模型的信息。In a first aspect, the present application provides a communication method, which can be executed by a first communication device, or by a component of the first communication device (such as a processor, a chip, or a chip system, etc.), and the present application does not specifically limit this. The method may specifically include the following steps: the first communication device receives model request information, and the model request information includes reasoning requirement information; the first communication device determines a first model according to the model request information, and the first model is a multi-model; the first communication device sends a first message, and the first message includes information about the first model.
在本申请实施例中,该第一通信装置视为模型的训练方,该第一通信装置可以为但不限于为:模型训练功能网元、或模型训练功能实体、或包含模型训练功能的通信装置。示例性地,该第一通信装置可以为包含模型训练功能模块的NWDAF网元、或者网元管理系统(EMS)设备、或者接入网设备(例如基站)等。In the embodiment of the present application, the first communication device is regarded as the training party of the model, and the first communication device may be, but is not limited to: a model training function network element, or a model training function entity, or a communication device including a model training function. Exemplarily, the first communication device may be a NWDAF network element including a model training function module, or a network element management system (EMS) device, or an access network device (such as a base station), etc.
在本申请方案中,第一通信装置接收模型请求信息,该模型请求信息中包括推理需求信息,该第一通信装置将根据该模型请求信息中的推理需求信息,确定较合适的多模型(即第一模型),再通过第一信息发送该第一模型的信息。从而可知,当推理端(即包含模型推理功能模块的第二通信装置)接收到该第一模型的信息后,基于该第一模型的信息,并利用该第一模型进行多模型的推理和结合,可以得到准确性较高的推理结果。因此,通过该方法可以有效地提高模型的使用效果,以保证智能推理的性能。In the present application scheme, the first communication device receives model request information, which includes reasoning requirement information. The first communication device determines a more suitable multi-model (i.e., the first model) based on the reasoning requirement information in the model request information, and then sends the information of the first model through the first information. It can be seen that when the reasoning end (i.e., the second communication device including the model reasoning function module) receives the information of the first model, based on the information of the first model, the reasoning and combination of multiple models using the first model can obtain a more accurate reasoning result. Therefore, this method can effectively improve the use effect of the model to ensure the performance of intelligent reasoning.
一种可能的实施方式中,该第一通信装置接收模型请求信息之前,该方法还包括:该第一通信装置发送训练能力指示信息,该训练能力指示信息用于指示该第一通信装置支持多模型的训练。In a possible implementation, before the first communication device receives the model request information, the method further includes: the first communication device sends training capability indication information, where the training capability indication information is used to indicate that the first communication device supports multi-model training.
通过该实施方式,可以保证该第一通信装置接收到模型请求信息后,能有效地执行多模型的训练。Through this implementation, it can be ensured that the first communication device can effectively perform multi-model training after receiving the model request information.
一种可能的实施方式中,该模型请求信息用于请求训练多模型时,该第一通信装置根据该模型请求 信息,确定第一模型,可以包括以下几种实现方式:In a possible implementation manner, when the model request information is used to request training of multiple models, the first communication device Information, determining the first model, can include the following implementation methods:
实现方式一:该模型请求信息中还包括多模型的训练策略;该第一通信装置根据该推理需求信息和该多模型的训练策略进行训练,得到该第一模型的多个子模型。Implementation method one: the model request information also includes a multi-model training strategy; the first communication device performs training according to the inference requirement information and the multi-model training strategy to obtain multiple sub-models of the first model.
实现方式二:该第一通信装置根据该推理需求信息,确定多模型的训练策略;以及根据该推理需求信息和该多模型的训练策略进行训练,得到该第一模型的多个子模型。Implementation method two: The first communication device determines a multi-model training strategy based on the reasoning requirement information; and performs training based on the reasoning requirement information and the multi-model training strategy to obtain multiple sub-models of the first model.
在本申请实施例中,该多模型的训练策略包括以下一项或多项:数据处理策略、训练的算法、训练的模式、子模型的数量、子模型的类型。In an embodiment of the present application, the multi-model training strategy includes one or more of the following: data processing strategy, training algorithm, training mode, number of sub-models, and type of sub-models.
通过该实施方式,该第一通信装置可以基于该模型请求信息,能有效的训练出该第一模型的多个子模型。Through this implementation, the first communication device can effectively train multiple sub-models of the first model based on the model request information.
一种可能的实施方式中,该模型请求信息用于请求获取多模型时,该第一通信装置根据该模型请求信息,确定第一模型,包括:该第一通信装置根据该推理需求信息,从至少一个预设的多模型中确定该第一模型。示例性的,该推理需求信息包括以下一项或多项:推理的类型、推理的性能需求、推理的速度需求、推理的功耗需求。In a possible implementation, when the model request information is used to request to obtain multiple models, the first communication device determines the first model according to the model request information, including: the first communication device determines the first model from at least one preset multiple model according to the inference requirement information. Exemplarily, the inference requirement information includes one or more of the following: the type of inference, the performance requirement of inference, the speed requirement of inference, and the power consumption requirement of inference.
通过该实施方式,该第一通信装置根据该推理需求信息,可以直接且快速地从已训练好的至少一个多模型中选出该第一模型。Through this implementation, the first communication device can directly and quickly select the first model from at least one trained multiple models based on the inference requirement information.
一种可能的实施方式,该模型请求信息中还包括多模型指示信息,该多模型指示信息用于指示请求训练或获取的模型为多模型。In a possible implementation manner, the model request information also includes multi-model indication information, and the multi-model indication information is used to indicate that the model requested to be trained or obtained is a multi-model.
该多模型指示信息也可以携带在该模型请求信息包含的推理需求信息中,或者该多模型指示信息被单独的发送给该第一通信装置,本申请实施例对此不做具体限定。The multi-model indication information may also be carried in the inference requirement information included in the model request information, or the multi-model indication information may be sent separately to the first communication device, which is not specifically limited in the embodiments of the present application.
通过该实施方式,该第一通信装置可以有效地且准确地为该第二通信装置训练或提供多模型。Through this implementation, the first communication device can effectively and accurately train or provide multiple models for the second communication device.
一种可能的实施方式中,该第一模型的信息中包括该第一模型的模型信息和该第一模型的多个子模型的信息,每个子模型的信息包括以下一项或多项:子模型的标识信息、子模型的级别、子模型的性能、性能约束;该多个子模型中包括多个第一级子模型和一个第二级子模型,该第二级子模型用于聚合该多个第一级子模型的推理信息;或者该多个子模型均为第一级子模型,该第一模型的信息中还包括聚合方法和/或权重信息。In one possible implementation, the information of the first model includes model information of the first model and information of multiple sub-models of the first model, and the information of each sub-model includes one or more of the following: identification information of the sub-model, the level of the sub-model, the performance of the sub-model, and performance constraints; the multiple sub-models include multiple first-level sub-models and one second-level sub-model, and the second-level sub-model is used to aggregate the reasoning information of the multiple first-level sub-models; or the multiple sub-models are all first-level sub-models, and the information of the first model also includes aggregation method and/or weight information.
通过该实施方式,可以有效地的确定该第一模型的多个子模型的信息和子模型的性能信息,以及多个子模型的推理信息之间的结合方式(即可以利用聚合方式结合,也可以通过第二级子模型结合)。Through this implementation, the information of multiple sub-models of the first model and the performance information of the sub-models, as well as the combination method between the reasoning information of multiple sub-models (that is, they can be combined in an aggregation manner or through second-level sub-models) can be effectively determined.
一种可能的实施方式中,该方法还包括:该第一通信装置发送该第一模型的推理性能信息,该第一模型的推理性能信息包括以下一项或多项:In a possible implementation manner, the method further includes: the first communication device sending reasoning performance information of the first model, where the reasoning performance information of the first model includes one or more of the following:
该第一模型的性能、该第一模型的大小信息、该第一模型的推理的功耗、该第一模型的推理速度、该第一模型的算力。The performance of the first model, the size information of the first model, the power consumption of the reasoning of the first model, the reasoning speed of the first model, and the computing power of the first model.
通过该实施方式,接收该第一模型的推理性能信息的接收端(如包含模型管理功能网元的第三通信装置)还可以基于该第一模型的推理性能信息对该第一模型进行有效的调整。例如基于该第一模型的推理性能信息和实际推理需求信息,适当减少该第一模型的子模型的数量。Through this implementation, the receiving end (such as a third communication device including a model management function network element) that receives the reasoning performance information of the first model can also effectively adjust the first model based on the reasoning performance information of the first model. For example, based on the reasoning performance information of the first model and the actual reasoning requirement information, the number of sub-models of the first model can be appropriately reduced.
第二方面,本申请提供一种通信方法,该方法可以由第二通信装置执行,也可以由第二通信装置的部件(例如处理器、芯片、或芯片系统等)执行,本申请对此不做具体限定。该方法具体可包括以下步骤:第二通信装置接收第二信息,该第二信息中包括第一模型的信息,该第一模型是根据推理需求信息确定的,该第一模型为多模型;该第二通信装置基于该第一模型的信息,得到该第一模型的推理信息。In a second aspect, the present application provides a communication method, which can be executed by a second communication device or by a component of the second communication device (such as a processor, a chip, or a chip system, etc.), and the present application does not specifically limit this. The method may specifically include the following steps: the second communication device receives second information, the second information includes information of a first model, the first model is determined according to reasoning requirement information, and the first model is a multi-model; the second communication device obtains reasoning information of the first model based on the information of the first model.
在本申请实施例中,该第二通信装置作为模型推理方,该第二通信装置可以为以下但不限于为:模型推理功能网元、或模型推理功能实体、或包含模型推理功能的通信装置。示例性地,该第二通信装置可以为包含模型推理功能模块的NWDAF网元、或者网元管理系统(EMS)设备、或者接入网设备(例如基站)等。In the embodiment of the present application, the second communication device serves as a model reasoning party, and the second communication device may be the following but is not limited to: a model reasoning function network element, or a model reasoning function entity, or a communication device including a model reasoning function. Exemplarily, the second communication device may be a NWDAF network element including a model reasoning function module, or a network element management system (EMS) device, or an access network device (such as a base station), etc.
在本申请方案中,第二通信装置接收到第一模型的信息,由于该第一模型是根据推理需求信息确定的多模型,该第二通信装置基于该第一模型的信息,并利用该第一模型进行多模型的推理和结合,可以得到准确性较高的推理结果。因此,通过该方法可以有效地提高模型的使用效果,以保证智能推理的性能。In the present application, the second communication device receives the information of the first model. Since the first model is a multi-model determined according to the reasoning requirement information, the second communication device uses the information of the first model and the first model to perform reasoning and combination of the multi-models to obtain a reasoning result with higher accuracy. Therefore, this method can effectively improve the use effect of the model to ensure the performance of intelligent reasoning.
一种可能的实施方式中,该第二通信装置接收第二信息之前,该方法还包括:该第二通信装置发送推理能力信息和该推理需求信息;该推理能力信息包括推理能力指示信息,以及下述一项或多项:推理 的算力、存储空间;该推理能力指示信息用于指示该第二通信装置支持多模型的推理;该推理需求信息包括下述一项或多项:推理的类型、推理的性能需求、推理的速度需求、推理的功耗需求。In a possible implementation manner, before the second communication device receives the second information, the method further includes: the second communication device sends reasoning capability information and the reasoning requirement information; the reasoning capability information includes reasoning capability indication information, and one or more of the following: reasoning computing power and storage space; the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning; the reasoning requirement information includes one or more of the following: the type of reasoning, the performance requirement of reasoning, the speed requirement of reasoning, and the power consumption requirement of reasoning.
通过该实施方式,该第二通信装置发送自身的推理能力信息和推理需求信息,不仅可以保证该第二通信装置后续能有效地进行多模型的推理,也可以保证该第二通信装置后续基于多模型执行推理的性能。Through this implementation, the second communication device sends its own reasoning capability information and reasoning requirement information, which can not only ensure that the second communication device can effectively perform multi-model reasoning in the future, but also ensure the performance of the second communication device in performing reasoning based on multiple models in the future.
在一种可能的实施方式中,该第一模型的信息中包括该第一模型的模型信息和该第一模型的多个子模型的信息,每个子模型的信息包括以下一项或多项:子模型的标识信息、子模型的级别、子模型的性能、性能约束。In a possible implementation, the information of the first model includes model information of the first model and information of multiple sub-models of the first model, and the information of each sub-model includes one or more of the following: identification information of the sub-model, the level of the sub-model, the performance of the sub-model, and performance constraints.
通过该实施方式,该第二通信装置可以准确地获得该第一模型的多个子模型的信息,以便于后续有效地利用这些子模型进行推理。Through this implementation, the second communication device can accurately obtain information about multiple sub-models of the first model, so as to effectively use these sub-models for reasoning later.
在一种可能的实施方式中,该多个子模型中包括多个第一级子模型和一个第二级子模型,该第二级的子模型用于聚合该多个第一级子模型的推理信息;该第二通信装置基于该第一模型的信息,得到该第一模型的推理信息,包括:该第二通信装置基于该多个第一级子模型的信息,利用该多个第一级子模型分别进行推理,得到该多个第一级子模型的推理信息;该第二通信装置使用该第二级子模型对该多个第一级子模型的推理信息进行聚合,得到该第一模型的推理信息。In a possible implementation, the multiple sub-models include multiple first-level sub-models and one second-level sub-model, and the second-level sub-model is used to aggregate the reasoning information of the multiple first-level sub-models; the second communication device obtains the reasoning information of the first model based on the information of the first model, including: the second communication device uses the multiple first-level sub-models to perform reasoning respectively based on the information of the multiple first-level sub-models to obtain the reasoning information of the multiple first-level sub-models; the second communication device uses the second-level sub-model to aggregate the reasoning information of the multiple first-level sub-models to obtain the reasoning information of the first model.
通过该实施方式,该第二通信装置可以利用该第一子模型的多个第一级子模型分别进行推理,并采用第二级子模型对该多个第一级子模型的推理信息进行有效的结合,从而得到该第一模型的推理信息。Through this implementation, the second communication device can use multiple first-level sub-models of the first sub-model to perform reasoning respectively, and use the second-level sub-model to effectively combine the reasoning information of the multiple first-level sub-models, so as to obtain the reasoning information of the first model.
在一种可能的实施方式中,该多个子模型均为第一级子模型,该第一模型的信息中还包括聚合方法和/或权重信息;该第二通信装置基于该第一模型的信息,得到该第一模型的推理信息,包括:该第二通信装置基于该多个子模型的信息,利用该多个子模型分别进行推理,得到该多个子模型的推理信息;该第二通信装置根据该聚合方法和/或权重信息对该多个子模型的推理信息进行聚合,得到该第一模型的推理信息。In a possible implementation, the multiple sub-models are all first-level sub-models, and the information of the first model also includes an aggregation method and/or weight information; the second communication device obtains reasoning information of the first model based on the information of the first model, including: the second communication device uses the multiple sub-models to perform reasoning respectively based on the information of the multiple sub-models to obtain the reasoning information of the multiple sub-models; the second communication device aggregates the reasoning information of the multiple sub-models according to the aggregation method and/or weight information to obtain the reasoning information of the first model.
通过该实施方式,该第二通信装置还可以利用该第一子模型的多个子模型分别进行推理,并采用指定的聚合方法和/或权重信息将该多个子模型的推理信息进行有效的结合,从而得到该第一模型的推理信息。Through this implementation, the second communication device can also use multiple sub-models of the first sub-model to perform reasoning separately, and use specified aggregation methods and/or weight information to effectively combine the reasoning information of the multiple sub-models to obtain the reasoning information of the first model.
第三方面,本申请提供一种通信方法,该方法可以由第三通信装置执行,也可以由第三通信装置的部件(例如处理器、芯片、或芯片系统等)执行,本申请对此不做具体限定。该方法具体可包括以下步骤:第三通信装置接收第一通信装置的训练能力指示信息;该训练能力指示信息用于指示该第一通信装置支持多模型的训练;该第三通信装置接收第二通信装置的推理需求信息和推理能力信息;该推理能力信息中包括推理能力指示信息,该推理能力指示信息用于指示该第二通信装置支持多模型的推理;该第三通信装置向该第一通信装置发送模型请求信息,该模型请求信息中包括该推理需求信息;该第三通信装置从该第一通信装置接收第一信息,该第一信息中包括第一模型的信息,该第一模型为多模型,该第一模型是根据该推理需求信息确定的;该第三通信装置向该第二通信装置发送第二信息,该第二信息中包括该第一模型的信息。In a third aspect, the present application provides a communication method, which can be executed by a third communication device or by a component of the third communication device (such as a processor, a chip, or a chip system, etc.), and the present application does not specifically limit this. The method may specifically include the following steps: the third communication device receives training capability indication information of the first communication device; the training capability indication information is used to indicate that the first communication device supports multi-model training; the third communication device receives reasoning requirement information and reasoning capability information of the second communication device; the reasoning capability information includes reasoning capability indication information, and the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning; the third communication device sends model request information to the first communication device, and the model request information includes the reasoning requirement information; the third communication device receives first information from the first communication device, the first information includes information of the first model, the first model is a multi-model, and the first model is determined according to the reasoning requirement information; the third communication device sends second information to the second communication device, and the second information includes information of the first model.
在本申请实施例中,该第三通信装置作为模型管理方,该第三通信装置可以为但不限于为:模型管理功能网元、或模型管理功能实体、或包含模型管理功能的通信装置。示例性地,该第三通信装置为包含模型管理功能模块的网络管理系统(NMS)设备。In the embodiment of the present application, the third communication device serves as a model manager, and the third communication device may be, but is not limited to: a model management function network element, or a model management function entity, or a communication device including a model management function. Exemplarily, the third communication device is a network management system (NMS) device including a model management function module.
在本申请方案中,第三通信装置接收第一通信装置的训练能力指示信息和第二通信装置的推理需求信息和推理能力信息,确定该第一通信装置支持多模型的训练,以及该第二通信装置支持多模型的推理;进而该第三通信装置向该第一通信装置发送携带该推理需求信息的模型请求信息,该第三通信装置可以有效地从该第一通信装置收到携带第一模型的信息的第一信息,该第一模型的信息是根据该推理需求信息确定的且该第一模型为多模型;该第三通信装置再通过第二信息将该第一模型的信息发送给第二通信装置;该第二通信装置收到该第一模型的信息后,基于该第一模型的信息,可以有效地利用该第一模型进行多模型的推理和结合,得到准确性较高的推理结果。因此,通过该方法可以有效地提高模型的使用效果,以保证智能推理的性能。In the present application scheme, the third communication device receives the training capability indication information of the first communication device and the reasoning requirement information and reasoning capability information of the second communication device, and determines that the first communication device supports multi-model training and the second communication device supports multi-model reasoning; then the third communication device sends a model request information carrying the reasoning requirement information to the first communication device, and the third communication device can effectively receive the first information carrying the information of the first model from the first communication device, the information of the first model is determined according to the reasoning requirement information and the first model is a multi-model; the third communication device then sends the information of the first model to the second communication device through the second information; after the second communication device receives the information of the first model, based on the information of the first model, the first model can be effectively used to perform multi-model reasoning and combination, and obtain a reasoning result with higher accuracy. Therefore, this method can effectively improve the use effect of the model to ensure the performance of intelligent reasoning.
一种可能的实施方式中,该第二通信装置的推理需求信息包括以下一项或多项:推理的类型、推理的性能需求、推理的速度需求、推理的功耗需求。通过该实施方式,可以保证该第二通信装置后续基于多模型执行推理的性能。In a possible implementation, the inference requirement information of the second communication device includes one or more of the following: the type of inference, the performance requirement of inference, the speed requirement of inference, and the power consumption requirement of inference. Through this implementation, the performance of the second communication device in subsequent inference based on multiple models can be guaranteed.
一种可能的实施方式中,该第二通信装置的推理能力信息还包括以下一项或多项: 推理的算力、存储空间。通过该实施方式,可以保证该第二通信装置有效地进行多模型推理。In a possible implementation manner, the reasoning capability information of the second communication device further includes one or more of the following: The computing power and storage space for reasoning. Through this implementation, it can be ensured that the second communication device effectively performs multi-model reasoning.
一种可能的实施方式,该模型请求信息中还包括多模型指示信息,该多模型指示信息用于指示请求训练或获取的模型为多模型。In a possible implementation manner, the model request information also includes multi-model indication information, and the multi-model indication information is used to indicate that the model requested to be trained or obtained is a multi-model.
该多模型指示信息也可以携带在该模型请求信息包含的推理需求信息中,或者该第三通信装置单独的将该多模型指示信息发送给第一通信装置,本申请实施例对此不做具体限定。The multi-model indication information may also be carried in the inference requirement information included in the model request information, or the third communication device may send the multi-model indication information to the first communication device separately. This embodiment of the present application does not specifically limit this.
通过该实施方式,可以保证该第一通信装置有效地且准确的训练或提供多模型。Through this implementation, it can be ensured that the first communication device can effectively and accurately train or provide multiple models.
一种可能的实施方式中,该方法还包括:该第三通信装置从该第一通信装置接收该第一模型的推理性能信息;该第三通信装置根据该第二通信装置的推理需求信息和该推理能力信息,以及该第一模型的推理性能信息和该第一模型的信息,调整该第一模型中的子模型的数量;该第一模型的推理性能信息包括以下一项或多项:该第一模型的性能、该第一模型的大小信息、该第一模型的推理的功耗、该第一模型的推理速度、该第一模型的算力。In a possible implementation, the method also includes: the third communication device receives the reasoning performance information of the first model from the first communication device; the third communication device adjusts the number of sub-models in the first model according to the reasoning requirement information and the reasoning capability information of the second communication device, as well as the reasoning performance information of the first model and the information of the first model; the reasoning performance information of the first model includes one or more of the following: the performance of the first model, the size information of the first model, the power consumption of the reasoning of the first model, the reasoning speed of the first model, and the computing power of the first model.
通过该实施方式,该第三通信装置在收到该第一模型的推理性能信息后,可以基于该第一模型的推理性能信息,以及该第二通信装置的推理需求信息和推理能力信息,对该第一模型的子模型的数量进行调整(如减少该第一模型的子模型的数量),以保证实际使用的该第一模型的子模型的推理性能更佳。Through this implementation, after receiving the reasoning performance information of the first model, the third communication device can adjust the number of sub-models of the first model (such as reducing the number of sub-models of the first model) based on the reasoning performance information of the first model and the reasoning requirement information and reasoning capability information of the second communication device to ensure that the reasoning performance of the sub-models of the first model actually used is better.
一种可能的实施方式中,该第一模型的信息中包括该第一模型的模型信息和该第一模型的多个子模型的信息,每个子模型的信息包括以下一项或多项:子模型的标识信息、子模型的级别、子模型的性能、性能约束;该多个子模型中包括多个第一级子模型和一个第二级子模型,该第二级子模型用于聚合该多个第一级子模型的推理信息;或者该多个子模型均为第一级子模型,该第一模型的信息中还包括聚合方法和/或权重信息。In one possible implementation, the information of the first model includes model information of the first model and information of multiple sub-models of the first model, and the information of each sub-model includes one or more of the following: identification information of the sub-model, the level of the sub-model, the performance of the sub-model, and performance constraints; the multiple sub-models include multiple first-level sub-models and one second-level sub-model, and the second-level sub-model is used to aggregate the reasoning information of the multiple first-level sub-models; or the multiple sub-models are all first-level sub-models, and the information of the first model also includes aggregation method and/or weight information.
在本申请实施例中,该第一模型的模型信息可以为该第一模型的标识、名称、类型等。该子模型的标识信息可以为子模型的存储地址,或者子模型的唯一标识符等。In the embodiment of the present application, the model information of the first model may be the identification, name, type, etc. of the first model. The identification information of the sub-model may be the storage address of the sub-model, or the unique identifier of the sub-model, etc.
通过该实施方式,可以准确地确定该第一模型的多个子模型的信息和子模型的推理信息之间结合的方式,以便于后续该第二通信装置有效地利用这些子模型进行结合推理。Through this implementation, the manner in which the information of the multiple sub-models of the first model and the reasoning information of the sub-models are combined can be accurately determined, so that the second communication device can subsequently effectively use these sub-models to perform combined reasoning.
第四方面,本申请实施例还提供一种通信装置,该通信装置可以是第一方面的第一通信装置,该通信装置也可以是第一通信装置中的部件(例如,芯片,或者芯片系统,或者电路),或者是能够和该第一通信装置匹配使用的装置。在本申请实施例中,该第一通信装置可以为但不限于为:模型训练功能网元、或者模型训练功能实体、或者包含模型训练功能的通信装置。In a fourth aspect, an embodiment of the present application further provides a communication device, which may be the first communication device of the first aspect, or a component (e.g., a chip, or a chip system, or a circuit) in the first communication device, or a device that can be used in combination with the first communication device. In an embodiment of the present application, the first communication device may be, but is not limited to, a model training function network element, or a model training function entity, or a communication device including a model training function.
一种可能的实现方式中,该通信装置可以包括执行第一方面中所描述的方法/操作/步骤/动作所一一对应的模块或单元,该模块或单元可以是硬件电路,也可是软件,也可以是硬件电路结合软件实现。一种可能的实现方式中,该通信装置可以包括通信模块(或收发模块)和处理模块。处理模块用于调用通信模块执行通信(即接收和/或发送)的功能。In one possible implementation, the communication device may include a module or unit corresponding to the method/operation/step/action described in the first aspect, and the module or unit may be a hardware circuit, or software, or a hardware circuit combined with software. In one possible implementation, the communication device may include a communication module (or a transceiver module) and a processing module. The processing module is used to call the communication module to perform the communication (i.e., receiving and/or sending) function.
一种可能的实现方式中,该通信装置包括通信单元(或收发单元)、处理单元;所述处理单元可以用于调用通信单元执行通信(即接收和/或发送)的功能;其中,所述通信单元,用于接收模型请求信息,所述模型请求信息中包括推理需求信息;所述处理单元,用于根据所述模型请求信息,确定第一模型,所述第一模型为多模型;所述通信单元,还用于发送第一信息,所述第一信息中包括所述第一模型的信息。In one possible implementation, the communication device includes a communication unit (or a transceiver unit) and a processing unit; the processing unit can be used to call the communication unit to perform communication (i.e., receiving and/or sending) functions; wherein the communication unit is used to receive model request information, and the model request information includes reasoning requirement information; the processing unit is used to determine a first model based on the model request information, and the first model is a multi-model; the communication unit is also used to send first information, and the first information includes information of the first model.
一种可能的实现方式中,所述通信单元,还用于:接收模型请求信息之前,发送训练能力指示信息,所述训练能力指示信息用于指示所述第一通信装置支持多模型的训练。In a possible implementation, the communication unit is further used to: before receiving the model request information, send training capability indication information, where the training capability indication information is used to indicate that the first communication device supports multi-model training.
一种可能的实现方式中,所述模型请求信息用于请求训练多模型时,所述模型请求信息中还包括多模型的训练策略;所述处理单元在根据所述模型请求信息,确定第一模型时,具体用于:根据所述推理需求信息和所述多模型的训练策略进行训练,得到所述第一模型的多个子模型;或者所述模型请求信息用于请求训练多模型时;所述处理单元在根据所述模型请求信息,确定第一模型时,具体用于:根据所述推理需求信息,确定多模型的训练策略;以及根据所述推理需求信息和所述多模型的训练策略进行训练,得到所述第一模型的多个子模型;其中,所述多模型的训练策略包括以下一项或多项:数据处理策略、训练的算法、训练的模式、子模型的数量、子模型的类型。In one possible implementation, when the model request information is used to request training of multiple models, the model request information also includes a training strategy for the multiple models; when the processing unit determines the first model according to the model request information, it is specifically used to: train according to the reasoning requirement information and the training strategy for the multiple models to obtain multiple sub-models of the first model; or when the model request information is used to request training of multiple models; when the processing unit determines the first model according to the model request information, it is specifically used to: determine the training strategy for the multiple models according to the reasoning requirement information; and train according to the reasoning requirement information and the training strategy for the multiple models to obtain multiple sub-models of the first model; wherein the training strategy for the multiple models includes one or more of the following: data processing strategy, training algorithm, training mode, number of sub-models, and type of sub-models.
一种可能的实现方式中,所述模型请求信息用于请求获取多模型时,所述处理单元在根据所述模型请求信息,确定第一模型时,具体用于:根据所述推理需求信息,从至少一个预设的多模型中确定所述第一模型。 In a possible implementation, when the model request information is used to request acquisition of multiple models, the processing unit, when determining the first model based on the model request information, is specifically used to: determine the first model from at least one preset multiple models based on the reasoning requirement information.
一种可能的实现方式,所述模型请求信息中还包括多模型指示信息,所述多模型指示信息用于指示请求训练或获取的模型为多模型。In a possible implementation, the model request information also includes multi-model indication information, and the multi-model indication information is used to indicate that the model requested to be trained or obtained is a multi-model.
一种可能的实现方式中,所述推理需求信息包括以下一项或多项:推理的类型、推理的性能需求、推理的速度需求、推理的功耗需求。In a possible implementation, the reasoning requirement information includes one or more of the following: the type of reasoning, the performance requirement of reasoning, the speed requirement of reasoning, and the power consumption requirement of reasoning.
一种可能的实现方式中,所述第一模型的信息中包括所述第一模型的模型信息和所述第一模型的多个子模型的信息,每个子模型的信息包括以下一项或多项:子模型的标识信息、子模型的级别、子模型的性能、性能约束;所述多个子模型中包括多个第一级子模型和一个第二级子模型,所述第二级子模型用于聚合所述多个第一级子模型的推理信息;或者所述多个子模型均为第一级子模型,所述第一模型的信息中还包括聚合方法和/或权重信息。In one possible implementation, the information of the first model includes model information of the first model and information of multiple sub-models of the first model, and the information of each sub-model includes one or more of the following: identification information of the sub-model, the level of the sub-model, the performance of the sub-model, and performance constraints; the multiple sub-models include multiple first-level sub-models and one second-level sub-model, and the second-level sub-model is used to aggregate the reasoning information of the multiple first-level sub-models; or the multiple sub-models are all first-level sub-models, and the information of the first model also includes aggregation method and/or weight information.
一种可能的实现方式中,所述通信单元,还用于发送所述第一模型的推理性能信息,所述第一模型的推理性能信息包括以下一项或多项:所述第一模型的性能、所述第一模型的大小信息、所述第一模型的推理的功耗、所述第一模型的推理速度、所述第一模型的算力。In one possible implementation, the communication unit is also used to send reasoning performance information of the first model, and the reasoning performance information of the first model includes one or more of the following: performance of the first model, size information of the first model, power consumption of reasoning of the first model, reasoning speed of the first model, and computing power of the first model.
第五方面,本申请实施例还提供一种通信装置,该通信装置可以用于第二方面的第二通信装置,也可以是该第二通信装置中的部件(例如,芯片,或者芯片系统,或者电路),或者是能够和该第二通信装置匹配使用的装置。在本申请实施例中,该第二通信装置可以为但不限于为:模型推理功能网元、或者模型推理功能实体、或者包含模型推理功能的通信装置。In a fifth aspect, an embodiment of the present application further provides a communication device, which can be used for the second communication device of the second aspect, or can be a component (for example, a chip, or a chip system, or a circuit) in the second communication device, or a device that can be used in combination with the second communication device. In an embodiment of the present application, the second communication device can be, but is not limited to: a model reasoning function network element, or a model reasoning function entity, or a communication device including a model reasoning function.
一种可能的实现方式中,该通信装置可以包括执行第二方面中所描述的方法/操作/步骤/动作所一一对应的模块或单元,该模块或单元可以是硬件电路,也可是软件,也可以是硬件电路结合软件实现。一种可能的实现方式中,该通信装置可以包括处理模块和通信模块(或收发模型)。处理模块用于调用通信模块执行通信(即接收和/或发送)的功能。In one possible implementation, the communication device may include a module or unit corresponding to the method/operation/step/action described in the second aspect, and the module or unit may be a hardware circuit, or software, or a hardware circuit combined with software. In one possible implementation, the communication device may include a processing module and a communication module (or a transceiver model). The processing module is used to call the communication module to perform the communication (i.e., receiving and/or sending) function.
一种可能的实现方式中,该通信装置包括通信单元(或收发单元)和处理单元;其中,所述通信单元,接收第二信息,所述第二信息中包括第一模型的信息,所述第一模型是根据推理需求信息确定的,且所述第一模型为多模型;所述处理单元,用于基于所述第一模型的信息,得到所述第一模型的推理信息。In one possible implementation, the communication device includes a communication unit (or a transceiver unit) and a processing unit; wherein the communication unit receives second information, the second information includes information of a first model, the first model is determined based on reasoning requirement information, and the first model is a multi-model; the processing unit is used to obtain reasoning information of the first model based on the information of the first model.
一种可能的实现方式中,所述通信单元,还用于:在接收第二信息之前,发送推理能力信息和所述推理需求信息;所述推理能力信息包括推理能力指示信息,以及下述一项或多项:推理的算力、存储空间;所述推理能力指示信息用于指示所述第二通信装置支持多模型的推理;所述推理需求信息包括下述一项或多项:推理的类型、推理的性能需求、推理的速度需求、推理的功耗需求。In one possible implementation, the communication unit is also used to: send reasoning capability information and the reasoning requirement information before receiving the second information; the reasoning capability information includes reasoning capability indication information, and one or more of the following: computing power and storage space for reasoning; the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning; the reasoning requirement information includes one or more of the following: the type of reasoning, the performance requirement of reasoning, the speed requirement of reasoning, and the power consumption requirement of reasoning.
一种可能的实现方式中,所述第一模型的信息中包括所述第一模型的模型信息和所述第一模型的多个子模型的信息,每个子模型的信息包括以下一项或多项:子模型的标识信息、子模型的级别、子模型的性能、性能约束。In one possible implementation, the information of the first model includes model information of the first model and information of multiple sub-models of the first model, and the information of each sub-model includes one or more of the following: identification information of the sub-model, the level of the sub-model, the performance of the sub-model, and performance constraints.
一种可能的实现方式中,所述多个子模型中包括多个第一级子模型和一个第二级子模型,所述第二级的子模型用于聚合所述多个第一级子模型的推理信息;In a possible implementation, the multiple sub-models include multiple first-level sub-models and one second-level sub-model, and the second-level sub-model is used to aggregate the reasoning information of the multiple first-level sub-models;
所述处理单元,在基于所述第一模型的信息,得到所述第一模型的推理信息时,具体用于:基于所述多个第一级子模型的信息,利用所述多个第一级子模型分别进行推理,得到所述多个第一级子模型的推理信息;使用所述第二级子模型对所述多个第一级子模型的推理信息进行聚合,得到所述第一模型的推理信息。When obtaining the inference information of the first model based on the information of the first model, the processing unit is specifically used to: based on the information of the multiple first-level sub-models, use the multiple first-level sub-models to perform inference respectively to obtain the inference information of the multiple first-level sub-models; use the second-level sub-model to aggregate the inference information of the multiple first-level sub-models to obtain the inference information of the first model.
一种可能的实现方式中,所述多个子模型均为第一级子模型,所述第一模型的信息中还包括聚合方法和/或权重信息;所述处理单元,在基于所述第一模型的信息,得到所述第一模型的推理信息时,具体用于:基于所述多个子模型的信息,利用所述多个子模型分别进行推理,得到所述多个子模型的推理信息;根据所述聚合方法和/或权重信息对所述多个子模型的推理信息进行聚合,得到所述第一模型的推理信息。In one possible implementation, the multiple sub-models are all first-level sub-models, and the information of the first model also includes an aggregation method and/or weight information; when the processing unit obtains the reasoning information of the first model based on the information of the first model, it is specifically used to: based on the information of the multiple sub-models, use the multiple sub-models to perform reasoning respectively to obtain the reasoning information of the multiple sub-models; aggregate the reasoning information of the multiple sub-models according to the aggregation method and/or weight information to obtain the reasoning information of the first model.
第六方面,本申请实施例还提供一种通信装置,该通信装置可以用于第三方面的第三通信装置,也可以是该第三通信装置中的部件(例如,芯片,或者芯片系统,或者电路),或者是能够和该第三通信装置匹配使用的装置。在本申请实施例中,该第三通信装置可以为但不限于为:模型管理功能网元、或者模型管理功能实体、或者包含模型管理功能的通信装置。In a sixth aspect, an embodiment of the present application further provides a communication device, which can be used for the third communication device of the third aspect, or can be a component (for example, a chip, or a chip system, or a circuit) in the third communication device, or a device that can be used in combination with the third communication device. In an embodiment of the present application, the third communication device can be, but is not limited to, a model management function network element, or a model management function entity, or a communication device including a model management function.
一种可能的实现方式中,该通信装置可以包括执行第三方面中所描述的方法/操作/步骤/动作所一一对应的模块或单元,该模块或单元可以是硬件电路,也可是软件,也可以是硬件电路结合软件实现。一 种可能的实现方式中,该通信装置可以包括处理模块和收发模块。处理模块用于调用通信模块(或收发模块)执行通信(即接收和/或发送)的功能。In a possible implementation, the communication device may include a module or unit corresponding to the method/operation/step/action described in the third aspect, and the module or unit may be a hardware circuit, or software, or a combination of a hardware circuit and software. In a possible implementation, the communication device may include a processing module and a transceiver module. The processing module is used to call the communication module (or the transceiver module) to perform the communication (ie, receiving and/or sending) function.
一种可能的实现方式中,该通信装置包括通信单元(收发单元)和处理单元;其中,所述处理单元用于调用通信单元执行通信(即接收和/或发送)的功能;所述通信单元,用于接收第一通信装置的训练能力指示信息;所述训练能力指示信息用于指示所述第一通信装置支持多模型的训练;以及接收第二通信装置的推理需求信息和推理能力信息;所述推理能力信息中包括推理能力指示信息,所述推理能力指示信息用于指示所述第二通信装置支持多模型的推理;所述通信单元,还用于向所述第一通信装置发送模型请求信息,所述模型请求信息中包括所述推理需求信息;从所述第一通信装置接收第一信息,所述第一信息中包括第一模型的信息,所述第一模型为多模型,所述第一模型是根据所述推理需求信息确定的;以及向所述第二通信装置发送第二信息,所述第二信息中包括所述第一模型的信息。In one possible implementation, the communication device includes a communication unit (transceiver unit) and a processing unit; wherein the processing unit is used to call the communication unit to perform communication (i.e., receiving and/or sending) functions; the communication unit is used to receive training capability indication information of a first communication device; the training capability indication information is used to indicate that the first communication device supports multi-model training; and receive reasoning requirement information and reasoning capability information of a second communication device; the reasoning capability information includes reasoning capability indication information, and the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning; the communication unit is also used to send model request information to the first communication device, and the model request information includes the reasoning requirement information; receive first information from the first communication device, the first information includes information of a first model, the first model is a multi-model, and the first model is determined based on the reasoning requirement information; and send second information to the second communication device, the second information includes information of the first model.
一种可能的实现方式中,所述推理需求信息包括以下一项或多项:推理的类型、推理的性能需求、推理的速度需求、推理的功耗需求。In a possible implementation, the reasoning requirement information includes one or more of the following: the type of reasoning, the performance requirement of reasoning, the speed requirement of reasoning, and the power consumption requirement of reasoning.
一种可能的实现方式中,所述推理能力信息还包括以下一项或多项:推理的算力、存储空间。In a possible implementation, the reasoning capability information further includes one or more of the following: reasoning computing power and storage space.
一种可能的实现方式,所述模型请求信息中还包括多模型指示信息,所述多模型指示信息用于指示请求训练或获取的模型为多模型。In a possible implementation, the model request information also includes multi-model indication information, and the multi-model indication information is used to indicate that the model requested to be trained or obtained is a multi-model.
一种可能的实现方式中,所述通信单元,还用于:从所述第一通信装置接收所述第一模型的推理性能信息;所述处理单元,还用于根据所述第二通信装置的推理需求信息和所述推理能力信息,以及所述第一模型的推理信息和所述第一模型的信息,调整所述第一模型中的子模型的数量;所述第一模型的推理性能信息包括以下一项或多项:所述第一模型的性能、所述第一模型的大小信息、所述第一模型的推理的功耗、所述第一模型的推理速度、所述第一模型的算力。In one possible implementation, the communication unit is further used to: receive reasoning performance information of the first model from the first communication device; the processing unit is further used to adjust the number of sub-models in the first model according to the reasoning requirement information and the reasoning capability information of the second communication device, as well as the reasoning information of the first model and the information of the first model; the reasoning performance information of the first model includes one or more of the following: the performance of the first model, the size information of the first model, the power consumption of the reasoning of the first model, the reasoning speed of the first model, and the computing power of the first model.
一种可能的实现方式中,所述第一模型的信息中包括所述第一模型的模型信息和所述第一模型的多个子模型的信息,每个子模型的信息包括以下一项或多项:子模型的标识信息、子模型的级别、子模型的性能、性能约束;所述多个子模型中包括多个第一级子模型和一个第二级子模型,所述第二级子模型用于聚合所述多个第一级子模型的推理信息;或者所述多个子模型均为第一级子模型,所述第一模型的信息中还包括聚合方式和/或权重信息。In one possible implementation, the information of the first model includes model information of the first model and information of multiple sub-models of the first model, and the information of each sub-model includes one or more of the following: identification information of the sub-model, the level of the sub-model, the performance of the sub-model, and performance constraints; the multiple sub-models include multiple first-level sub-models and one second-level sub-model, and the second-level sub-model is used to aggregate the reasoning information of the multiple first-level sub-models; or the multiple sub-models are all first-level sub-models, and the information of the first model also includes aggregation method and/or weight information.
第七方面,本申请实施例中提供一种通信装置,该装置包括:至少一个处理器和接口电路;所述接口电路用于为所述至少一个处理器提供程序或指令的输入和/或输出;所述至少一个处理器用于执行所述程序或者指令以使得所述通信装置可实现上述第一方面或其中任意一种可能的实施方式提供的方法,或者可实现上述第二方面或其中任意一种可能的实施方式提供的方法,或者可实现上述第三方面或其中任意一种可能的实施方式提供的方法。In the seventh aspect, a communication device is provided in an embodiment of the present application, and the device includes: at least one processor and an interface circuit; the interface circuit is used to provide input and/or output of programs or instructions to the at least one processor; the at least one processor is used to execute the program or instructions so that the communication device can implement the method provided by the above-mentioned first aspect or any possible implementation method thereof, or can implement the method provided by the above-mentioned second aspect or any possible implementation method thereof, or can implement the method provided by the above-mentioned third aspect or any possible implementation method thereof.
第八方面,本申请实施例中提供一种计算机存储介质,该存储介质中存储软件程序,该软件程序在被一个或多个处理器读取并执行时,可实现上述第一方面或其中任意一种可能的实施方式提供的方法,或者可实现上述第二方面或其中任意一种可能的实施方式提供的方法,或者可实现上述第三方面或其中任意一种可能的实施方式提供的方法。In an eighth aspect, a computer storage medium is provided in an embodiment of the present application, in which a software program is stored. When the software program is read and executed by one or more processors, the method provided by the first aspect or any possible implementation thereof can be implemented, or the method provided by the second aspect or any possible implementation thereof can be implemented, or the method provided by the third aspect or any possible implementation thereof can be implemented.
第九方面,本申请实施例中提供一种包含指令的计算机程序产品,当指令在计算机上运行时,使得计算机执行上述第一方面或其中任一种可能的实施方式提供的方法,或者使得计算机执行上述第二方面或其中任一种可能的实施方式提供的方法,或者使得计算机执行上述第三方面或其中任一种可能的实施方式提供的方法。In the ninth aspect, a computer program product comprising instructions is provided in an embodiment of the present application. When the instructions are executed on a computer, the computer executes the method provided in the first aspect or any possible implementation manner thereof, or the computer executes the method provided in the second aspect or any possible implementation manner thereof, or the computer executes the method provided in the third aspect or any possible implementation manner thereof.
第十方面,本申请实施例中提供一种芯片系统,该芯片系统包括处理器,用于支持设备实现上述第一方面中所涉及的功能,或者用于支持设备实现上述第二方面中所涉及的功能,或者用于支持设备实现上述第三方面中所涉及的功能。In the tenth aspect, a chip system is provided in an embodiment of the present application, which chip system includes a processor for supporting a device to implement the functions involved in the above-mentioned first aspect, or for supporting a device to implement the functions involved in the above-mentioned second aspect, or for supporting a device to implement the functions involved in the above-mentioned third aspect.
在一种可能的设计中,所述芯片系统还包括存储器,所述存储器,用于保存必要的程序指令和数据。该芯片系统,可以由芯片构成,也可以包含芯片和其他分立器件。In a possible design, the chip system further includes a memory, and the memory is used to store necessary program instructions and data. The chip system can be composed of a chip, or can include a chip and other discrete devices.
第十一方面,本申请实施例中还提供一种芯片系统,该芯片系统包括处理器和接口,所述接口用于获取程序或指令,所述处理器用于调用所述程序或指令以实现或者支持设备实现第一方面所涉及的功能,或者所述处理器用于调用所述程序或指令以实现或者支持设备实现第二方面所涉及的功能,或者所述处理器用于调用所述程序或指令以实现或者支持设备实现第三方面所涉及的功能。In the eleventh aspect, a chip system is also provided in an embodiment of the present application, which includes a processor and an interface, wherein the interface is used to obtain a program or instruction, and the processor is used to call the program or instruction to implement or support the device to implement the function involved in the first aspect, or the processor is used to call the program or instruction to implement or support the device to implement the function involved in the second aspect, or the processor is used to call the program or instruction to implement or support the device to implement the function involved in the third aspect.
在一种可能的设计中,所述芯片系统还包括存储器,所述存储器,用于保存终端设备必要的程序指 令和数据。该芯片系统,可以由芯片构成,也可以包括芯片和其他分立器件。In a possible design, the chip system further includes a memory, the memory being used to store program instructions necessary for the terminal device. The chip system can be composed of chips or include chips and other discrete devices.
上述第四方面至第六方面以及第四方面至第六方面中的任意一种可能的实现可以达到的技术效果,可以参照上述第一方面至第三方面以及第一方面至第三方面中任意一种可能的实施方式所能达到的技术效果,这里不再重复赘述。The technical effects that can be achieved by any possible implementation of the above-mentioned fourth to sixth aspects and the fourth to sixth aspects can refer to the technical effects that can be achieved by any possible implementation of the above-mentioned first to third aspects and the first to third aspects, and will not be repeated here.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
图1为一种提高模型的使用效果的方案流程示意图;FIG1 is a schematic diagram of a solution flow chart for improving the use effect of the model;
图2为本申请实施例提供的两种系统逻辑架构的示例图;FIG. 2 is an example diagram of two system logic architectures provided in an embodiment of the present application;
图3A为本申请实施例的方法所能应用的第一种实际部署架构示意图;FIG3A is a schematic diagram of a first practical deployment architecture to which the method of an embodiment of the present application can be applied;
图3B为本申请实施例的方法所能应用的第二种实际部署架构示意图;FIG3B is a schematic diagram of a second practical deployment architecture to which the method of the embodiment of the present application can be applied;
图3C为本申请实施例的方法所能应用的第三种实际部署架构示意图;FIG3C is a schematic diagram of a third practical deployment architecture to which the method of the embodiment of the present application can be applied;
图3D为本申请实施例的方法所能应用的第四种实际部署架构示意图;FIG3D is a schematic diagram of a fourth actual deployment architecture to which the method of the embodiment of the present application can be applied;
图3E为本申请实施例的方法所能应用的第五种实际部署架构示意图;FIG3E is a schematic diagram of a fifth practical deployment architecture to which the method of the embodiment of the present application can be applied;
图4A为本申请实施例提供的一种通信方法的流程示意图;FIG4A is a flow chart of a communication method provided in an embodiment of the present application;
图4B为本申请实施例提供的另一种通信方法的流程示意图;FIG4B is a schematic diagram of a flow chart of another communication method provided in an embodiment of the present application;
图4C为本申请实施例提供的一种多模型训练和推理过程的示例图;FIG4C is an example diagram of a multi-model training and reasoning process provided in an embodiment of the present application;
图5为本申请实施例提供的第一个实施例的流程示意图;FIG5 is a schematic diagram of a flow chart of a first embodiment provided in the embodiments of the present application;
图6为本申请实施例提供的第二个实施例的流程示意图;FIG6 is a schematic diagram of a flow chart of a second embodiment provided in the present application;
图7为本申请实施例提供的第三个实施例的流程示意图;FIG7 is a schematic diagram of a flow chart of a third embodiment provided in the embodiments of the present application;
图8为本申请实施例提供的第四个实施例的流程示意图;FIG8 is a schematic diagram of a flow chart of a fourth embodiment provided in the embodiments of the present application;
图9为本申请实施例提供的第五个实施例的流程示意图;FIG9 is a schematic diagram of a flow chart of a fifth embodiment provided in the embodiments of the present application;
图10为本申请实施例提供的第六个实施例的流程示意图;FIG10 is a schematic diagram of a flow chart of a sixth embodiment provided in an embodiment of the present application;
图11为本申请实施例提供的第七个实施例的流程示意图;FIG11 is a schematic diagram of a flow chart of a seventh embodiment provided in the embodiments of the present application;
图12为本申请实施例提供的一种通信装置的结构示意图;FIG12 is a schematic diagram of the structure of a communication device provided in an embodiment of the present application;
图13为本申请实施例提供的另一种通信装置的结构示意图;FIG13 is a schematic diagram of the structure of another communication device provided in an embodiment of the present application;
图14为本申请实施例提供的一种芯片的装置结构示意图。FIG. 14 is a schematic diagram of the device structure of a chip provided in an embodiment of the present application.
具体实施方式Detailed ways
为了使本申请的目的、技术方案和优点更加清楚,下面将结合附图对本申请作进一步地详细描述。方法实施例中的具体操作方法也可以应用于装置实施例或系统实施例中。其中,在本申请的描述中,除非另有说明,“多个”的含义是两个或两个以上。In order to make the purpose, technical solution and advantages of the present application clearer, the present application will be further described in detail below with reference to the accompanying drawings. The specific operation method in the method embodiment can also be applied to the device embodiment or the system embodiment. In the description of the present application, unless otherwise specified, the meaning of "multiple" is two or more.
随着网络的智能化和自动化水平的不断提升,人工智能(artificial intelligence,AI)和机器学习(machine learning,ML)技术应用的领域也越来越广泛,主要包括管理域、核心网(core network,CN)域和无线接入网(radio access network,RAN)域等。With the continuous improvement of network intelligence and automation, the application fields of artificial intelligence (AI) and machine learning (ML) technologies are becoming more and more extensive, mainly including management domain, core network (CN) domain and radio access network (RAN) domain.
在管理域的智能化中,目前协议定义了管理数据分析服务(management data analytics service,MDAS),MDAS生产者可以基于AI/ML技术对网络、服务事件及状态相关的数据进行处理和分析,并提供分析报告用于网络和服务运营。在核心网域的智能化中,可以由网络数据分析功能(network data analytics function,NWDAF)基于ML模型进行网络数据分析,得到数据分析结果,并提供给网络、网管及应用以执行策略决策使用。在RAN域的智能化中,目前的主要研究是支持RAN智能化的功能框架定义,即旨在基于当前RAN的架构和接口进行必要的增强,以支持网络智能化。In the intelligentization of the management domain, the current protocol defines a management data analytics service (MDAS). MDAS producers can process and analyze data related to network, service events and status based on AI/ML technology, and provide analysis reports for network and service operations. In the intelligentization of the core network domain, the network data analysis function (NWDAF) can perform network data analysis based on ML models, obtain data analysis results, and provide them to the network, network management and applications for policy decision-making. In the intelligentization of the RAN domain, the current main research is to support the definition of the functional framework of RAN intelligence, that is, to make necessary enhancements based on the current RAN architecture and interfaces to support network intelligence.
由于当前标准中模型训练和推理的基本架构已经相应的确定,且支持各领域的模型使用。因而如何提高模型的使用效果,以保证智能分析性能是目前需要进一步探讨的问题。Since the basic architecture of model training and reasoning in the current standard has been determined accordingly and supports the use of models in various fields, how to improve the use effect of the model to ensure the performance of intelligent analysis is an issue that needs further discussion.
在一种提高模型的使用效果的方案中,提出由模型训练功能网元(或实体)向模型推理功能网元(或实体)提供多个不同性能的模型,模型推理功能网元(或实体)从中选择合适的模型进行推理。示例性地,如图1所示,其实现的步骤包括:S101:模型推理功能实体向模型训练功能实体发送模型请求,该模型请求中包括推理类型(例如覆盖问题分析、小区流量预测等)和性能需求(例如精度、准确度等);S102:模型训练功能实体根据该模型请求中的推理的类型和性能需求,训练得到多个模型;S103:模 型训练功能实体向模型推理功能实体发送模型响应,该模型响应中包括推理类型和多个模型的列表,例如,<模型标识1,模型标识2,…>。In a scheme for improving the use effect of the model, it is proposed that the model training functional network element (or entity) provides a plurality of models with different performances to the model reasoning functional network element (or entity), and the model reasoning functional network element (or entity) selects a suitable model for reasoning. Exemplarily, as shown in FIG1 , the steps for implementation include: S101: the model reasoning functional entity sends a model request to the model training functional entity, and the model request includes the reasoning type (such as coverage problem analysis, cell traffic prediction, etc.) and performance requirements (such as precision, accuracy, etc.); S102: the model training functional entity trains multiple models according to the reasoning type and performance requirements in the model request; S103: the model The model training functional entity sends a model response to the model reasoning functional entity, where the model response includes a reasoning type and a list of multiple models, for example, <model ID 1, model ID 2, …>.
然而,上述这种方案存在一些明显缺陷,如模型推理功能实体无法确定模型训练功能实体的训练能力,若向模型训练功能实体请求多个模型时可能会请求失败,另外,由于模型训练功能实体训练的模型数量不确定,且所提供的模型不一定适用于模型推理功能实体的推理,这些均可能导致方案不可行或者推理的效果不理想,从而不能有效地提高模型的智能推理(或分析)性能。However, the above scheme has some obvious defects. For example, the model reasoning functional entity cannot determine the training capability of the model training functional entity. If multiple models are requested from the model training functional entity, the request may fail. In addition, since the number of models trained by the model training functional entity is uncertain and the provided models are not necessarily suitable for the reasoning of the model reasoning functional entity, these may lead to the scheme being infeasible or the reasoning effect being unsatisfactory, thereby failing to effectively improve the intelligent reasoning (or analysis) performance of the model.
鉴于上述可知,在多模型的使用中,如何获取合适的模型进行推理,以提高模型分析性能是目前亟待解决的技术问题之一。In view of the above, when using multiple models, how to obtain a suitable model for reasoning to improve model analysis performance is one of the technical problems that needs to be solved urgently.
因此,本申请提出一种通信方法,可以有效地提高模型的使用效果,从而可保证智能推理的性能。本申请提出的该方法可以适用于5G系统架构中,还可以适用于但不限于长期演进(long term evolution,LTE)通信系统,以及未来演进的各种无线通信系统中。Therefore, the present application proposes a communication method that can effectively improve the use effect of the model, thereby ensuring the performance of intelligent reasoning. The method proposed in the present application can be applied to the 5G system architecture, and can also be applied to but not limited to the long-term evolution (LTE) communication system, and various wireless communication systems that will evolve in the future.
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行描述。The technical solutions in the embodiments of the present application will be described below in conjunction with the drawings in the embodiments of the present application.
本申请实施例所适用的系统逻辑架构中主要包括模型管理功能实体、模型训练功能实体以及模型推理功能实体,图2示出了本申请实施例的几种系统逻辑架构的示例图,如图2中(1)所示,模型管理功能实体、模型训练功能实体以及模型推理功能实体可以都是相互独立的逻辑实体,且任意两个实体之间有接口进行通信;如图2中(2)所示,模型管理功能实体为可选的,模型管理功能实体可以设在模型训练功能实体的内部,模型训练功能实体和模型推理功能实体是相互独立的逻辑实体,且两者之间有接口进行通信。The system logical architecture applicable to the embodiments of the present application mainly includes a model management functional entity, a model training functional entity and a model reasoning functional entity. Figure 2 shows example diagrams of several system logical architectures of the embodiments of the present application. As shown in (1) in Figure 2, the model management functional entity, the model training functional entity and the model reasoning functional entity can all be independent logical entities, and there is an interface for communication between any two entities; as shown in (2) in Figure 2, the model management functional entity is optional, and the model management functional entity can be set inside the model training functional entity. The model training functional entity and the model reasoning functional entity are independent logical entities, and there is an interface for communication between the two.
此外,本申请还提供了本申请实施例的方法所能应用的几种实际部署架构,如图3A-3E所示。其中,图3A示出了一种支持管理域的模型推理的架构,模型管理功能实体可以部署在网络管理系统(network management system,NMS)设备,模型训练功能实体和模型推理功能实体可以部署在网元管理系统(element management system,EMS)设备。图3B示出了一种支持RAN域的模型推理的架构,模型管理功能实体可以部署在NMS,模型训练功能实体可以部署在EMS,模型推理功能实体可以部署在RAN域设备,例如基站。图3C出示了一种支持CN域的模型推理的架构,模型训练功能实体可以部署在NWDAF(MTLF),模型推理功能实体可以部署在NWDAF(AnLF),其中,网络存储功能(network repository function,NRF)主要用于网络功能(network runction,NF)的管理,包括NF注册/更新/去注册、NF发现等,由于NWDAF可以视为一种NF,NRF可以对NWDAF进行管理。图3D和图3E分别示出了一种支持RAN域的空口相关的模型推理的架构,在图3D中,模型训练功能实体可以部署在基站,模型推理功能实体可以部署在用户设备(user equipment,UE);可选的,在双边均能支持推理下,基站中也可以部署模型推理功能实体。图3E中,基站和UE侧都部署有模型训练功能实体和模型推理功能实体,可适用于双边模型场景下基站和UE联合训练的场景。In addition, the present application also provides several practical deployment architectures to which the method of the embodiment of the present application can be applied, as shown in Figures 3A-3E. Among them, Figure 3A shows an architecture that supports model reasoning in the management domain, the model management function entity can be deployed in the network management system (network management system, NMS) equipment, the model training function entity and the model reasoning function entity can be deployed in the element management system (element management system, EMS) equipment. Figure 3B shows an architecture that supports model reasoning in the RAN domain, the model management function entity can be deployed in the NMS, the model training function entity can be deployed in the EMS, and the model reasoning function entity can be deployed in the RAN domain equipment, such as a base station. FIG3C shows an architecture supporting model reasoning in the CN domain. The model training functional entity can be deployed in the NWDAF (MTLF), and the model reasoning functional entity can be deployed in the NWDAF (AnLF). The network repository function (NRF) is mainly used for the management of network functions (NFs), including NF registration/update/deregistration, NF discovery, etc. Since the NWDAF can be regarded as a NF, the NRF can manage the NWDAF. FIG3D and FIG3E respectively show an architecture supporting air interface-related model reasoning in the RAN domain. In FIG3D, the model training functional entity can be deployed in the base station, and the model reasoning functional entity can be deployed in the user equipment (UE); optionally, the model reasoning functional entity can also be deployed in the base station when both sides can support reasoning. In FIG3E, the model training functional entity and the model reasoning functional entity are deployed on both the base station and the UE side, which can be applied to the scenario of joint training of the base station and the UE in the bilateral model scenario.
下面对上述实际部署架构中的网元/模块/设备的功能进行相应介绍。The functions of the network elements/modules/devices in the above actual deployment architecture are introduced below.
网络管理系统NMS:NMS也可以称为跨域管理系统,用于负责网络的运行、管理和维护功能。Network Management System NMS: NMS can also be called a cross-domain management system, which is responsible for the operation, management and maintenance of the network.
网元管理系统EMS:EMS也可以称为域管理系统或者单域管理系统,用于管理一个或多个某个类别的网元。Element Management System (EMS): EMS can also be called a domain management system or a single domain management system, which is used to manage one or more network elements of a certain category.
上述的NMS和EMS还可以统称为3GPP管理系统,或者操作维护管理(operations administration and maintenance,OAM)模块。The above-mentioned NMS and EMS can also be collectively referred to as 3GPP management system, or operations administration and maintenance (OAM) module.
无线接入网RAN:为用户设备提供无线接入,使用户可以接入到网络中。Radio Access Network (RAN): provides wireless access to user devices, allowing users to access the network.
核心网CN:主要是提供用户连接、对用户的管理以及对业务完成承载,作为承载网络提供到外部网络的接口。Core network CN: mainly provides user connection, user management and service carrying, and provides an interface to the external network as a bearer network.
网络数据分析功能NWDAF网元:负责核心网域的数据分析,如检测用户异常行为、分析切片的负载等。Network data analysis function NWDAF network element: responsible for data analysis in the core network domain, such as detecting abnormal user behavior and analyzing slice load.
模型管理功能实体:负责模型相关的生命周期管理,包括训练策略配置等。Model management functional entity: responsible for model-related lifecycle management, including training strategy configuration, etc.
模型训练功能实体:负责模型的训练,并在模型训练结束之后生成ML模型。Model training functional entity: responsible for model training and generating an ML model after model training is completed.
模型推理功能实体:负责模型的推理,利用模型得到推理输出或推理结果。Model reasoning functional entity: responsible for the reasoning of the model, using the model to obtain the reasoning output or reasoning result.
AnLF:为NWDAF网元的分析逻辑功能。AnLF: is the analysis logic function of the NWDAF network element.
MTLF:为NWDAF网元的模型训练逻辑功能。MTLF: Model training logic function for NWDAF network elements.
基站:移动通信系统中,连接固定部分与无线部分,并通过空中的无线通道与移动终端相连的设备。 Base station: A device in a mobile communication system that connects the fixed part with the wireless part and is connected to mobile terminals through wireless channels in the air.
UE:用户终端设备,允许用户接入网络的设备。UE: User terminal equipment, a device that allows users to access the network.
图3A-3E所示的部署架构中不限于仅包含图中所示的实体,还可以包含其它未在图中表示的设备,具体本申请在此处不再一一列举。The deployment architecture shown in Figures 3A-3E is not limited to including only the entities shown in the figures, but may also include other devices not shown in the figures, which will not be listed one by one in this application.
图3A-3E中的NMS、EMS、RAN、UE、NWDAF所包含的模型管理功能实体、模型训练功能实体、以及模型推理功能实体,也可以称为模型管理功能、模型训练功能、模型推理功能,还可以称为模型管理功能网元/模块、模型训练功能网元/模块、模型推理功能网元/模块。The model management functional entity, model training functional entity, and model reasoning functional entity included in the NMS, EMS, RAN, UE, and NWDAF in Figures 3A-3E can also be called model management function, model training function, model reasoning function, and can also be called model management function network element/module, model training function network element/module, and model reasoning function network element/module.
在本申请实施例中,网元或者功能既可以是硬件设备中的网络元件,也可以是在专用硬件上运行软件功能,或者是平台(例如,云平台)上实例化的虚拟化功能。作为一种可能的实现方法,上述网元或者功能可以由一个设备实现,也可以由多个设备共同实现,还可以是一个设备内的一个功能模块,本申请实施例对此不作具体限定。In the embodiments of the present application, the network element or function can be a network element in a hardware device, a software function running on dedicated hardware, or a virtualized function instantiated on a platform (e.g., a cloud platform). As a possible implementation method, the above network element or function can be implemented by one device, or by multiple devices together, or can be a functional module in one device, which is not specifically limited in the embodiments of the present application.
为方便说明,本申请实施例以网元为例进行说明,并将XX网元直接简称为XX,例如,SMF网元简称为SMF。应理解,本申请中所有网元的名称仅仅作为示例,在未来通信中还可以称为其它名称,或者在未来通信中本申请涉及的网元还可以通过其它具有相同功能的实体或者设备等来替代,本申请对此均不作限定。这里做统一说明,后续不再赘述。For the convenience of explanation, the embodiments of the present application are explained by taking the network element as an example, and the XX network element is directly referred to as XX, for example, the SMF network element is referred to as SMF. It should be understood that the names of all network elements in the present application are only examples, and they may be called other names in future communications, or the network elements involved in the present application may be replaced by other entities or devices with the same functions in future communications, and the present application does not limit this. A unified explanation is given here, and no further description will be given later.
需要说明的是,本申请中所有消息和信息的名称仅仅作为示例,可以是其它名称,本申请对此均不作限定。应理解,网元1到网元2的消息或信息,可以是网元1直接向网元2发送的消息,也可以是间接发送,例如网元1先向网元3发送消息,网元3再向网元2发送消息,最终消息或信息通过一个或多个网元发送到了网元2。It should be noted that the names of all messages and information in this application are only examples and may be other names, which are not limited in this application. It should be understood that the message or information from network element 1 to network element 2 may be a message sent directly from network element 1 to network element 2, or may be sent indirectly, for example, network element 1 first sends a message to network element 3, and network element 3 then sends a message to network element 2, and finally the message or information is sent to network element 2 through one or more network elements.
在本申请实施例中,“推理”和“分析”的意思可视为相同,均基于模型实现,例如“基于多模型进行推理,得到推理结果”相当于“基于多模型进行分析,得到分析结果”,“推理类型”相当于“分析类型”,“推理标识”相当于“分析标识”。In the embodiments of the present application, the meanings of "reasoning" and "analysis" can be regarded as the same, and both are implemented based on models. For example, "performing reasoning based on multiple models to obtain reasoning results" is equivalent to "performing analysis based on multiple models to obtain analysis results", "reasoning type" is equivalent to "analysis type", and "reasoning identifier" is equivalent to "analysis identifier".
此外,在本申请中,“指示”可以包括直接指示、间接指示、显示指示、隐式指示。当描述某一指示信息用于指示A时,可以理解为该指示信息携带A、直接指示A,或间接指示A。In addition, in this application, "indication" may include direct indication, indirect indication, explicit indication, and implicit indication. When describing that a certain indication information is used to indicate A, it can be understood that the indication information carries A, directly indicates A, or indirectly indicates A.
本申请中,指示信息所指示的信息,称为待指示信息。在具体实现过程中,对待指示信息进行指示的方式有很多种,例如但不限于,可以直接指示待指示信息,如待指示信息本身或者该待指示信息的索引等。也可以通过指示其他信息来间接指示待指示信息,其中该其他信息与待指示信息之间存在关联关系。还可以仅仅指示待指示信息的一部分,而待指示信息的其他部分则是已知的或者提前约定的。例如,还可以借助预先约定(例如协议规定)的各个信息的排列顺序来实现对特定信息的指示,从而在一定程度上降低指示开销。In this application, the information indicated by the indication information is referred to as the information to be indicated. In the specific implementation process, there are many ways to indicate the information to be indicated, such as but not limited to, the information to be indicated can be directly indicated, such as the information to be indicated itself or the index of the information to be indicated. The information to be indicated can also be indirectly indicated by indicating other information, wherein there is an association relationship between the other information and the information to be indicated. It is also possible to indicate only a part of the information to be indicated, while the other parts of the information to be indicated are known or agreed in advance. For example, the indication of specific information can also be achieved with the help of the arrangement order of each information agreed in advance (such as specified by the protocol), thereby reducing the indication overhead to a certain extent.
待指示信息可以作为一个整体一起发送,也可以分成多个子信息分开发送,而且这些子信息的发送周期和/或发送时机可以相同,也可以不同。具体发送方法本申请不进行限定。其中,这些子信息的发送周期和/或发送时机可以是预先定义的,例如根据协议预先定义的,也可以是发射端设备通过向接收端设备发送配置信息来配置的。其中,该配置信息可以例如但不限于包括无线资源控制信令、媒体接入控制(media access control,MAC)层信令和物理层信令中的一种或者至少两种的组合。其中,无线资源控制信令例如包无线资源控制(radio resource control,RRC)信令;MAC层信令例如包括MAC控制元素(control element,CE);物理层信令例如包括下行控制信息(downlink control information,DCI)。The information to be indicated can be sent as a whole, or divided into multiple sub-information and sent separately, and the sending period and/or sending time of these sub-information can be the same or different. The specific sending method is not limited in this application. Among them, the sending period and/or sending time of these sub-information can be pre-defined, for example, pre-defined according to the protocol, or configured by the transmitting device by sending configuration information to the receiving device. Among them, the configuration information can include, for example, but not limited to, one or a combination of at least two of radio resource control signaling, media access control (media access control, MAC) layer signaling and physical layer signaling. Among them, radio resource control signaling, for example, radio resource control (radio resource control, RRC) signaling; MAC layer signaling, for example, includes MAC control element (control element, CE); physical layer signaling, for example, includes downlink control information (downlink control information, DCI).
下面结合具体实施例介绍本申请的技术方案。The technical solution of the present application is introduced below in conjunction with specific embodiments.
本申请实施例提供了一种通信方法,该方法可适用于但不限于图3A-3E所示的实际部署架构中,并且该方法可以由本申请涉及到的网元执行,或者由涉及到的网元对应的芯片执行,本申请中的网元可以为物理上的实体网元,也可以是虚拟的网元,本申请对涉及的网元的形态不做具体限定。An embodiment of the present application provides a communication method, which can be applied to but not limited to the actual deployment architecture shown in Figures 3A-3E, and the method can be executed by the network element involved in the present application, or by the chip corresponding to the network element involved. The network element in the present application can be a physical entity network element or a virtual network element. The present application does not specifically limit the form of the network element involved.
图4A为本申请实施例提出的一种通信方法的流程示意图。该方法可以由第一通信装置(也可以是第二通信装置、第三通信装置)的收发器和/或处理器执行,也可以由该收发器和/或处理器对应的芯片执行。或者该实施例还可由该第一通信装置(也可以是第二通信装置、第三通信装置)所连接的控制器或控制设备实现,该控制器或控制设备用于管理包括该第一通信装置(也可以是第二通信装置、第三通信装置)在内的至少一个装置。并且针对执行该实施例的通信装置的具体形态,本申请不做具体限定。并且,需要说明的是,下文中提及的“第一”、“第二”等序数词是用于对多个对象进行区分,以便于描述,并不用于限定多个对象的顺序、时序、优先级或者重要程度。请参阅图4A,该方法的具体流程 如下:FIG4A is a flow chart of a communication method proposed in an embodiment of the present application. The method can be executed by a transceiver and/or processor of a first communication device (or a second communication device or a third communication device), or by a chip corresponding to the transceiver and/or processor. Alternatively, the embodiment can also be implemented by a controller or control device connected to the first communication device (or a second communication device or a third communication device), and the controller or control device is used to manage at least one device including the first communication device (or a second communication device or a third communication device). And the present application does not make specific restrictions on the specific form of the communication device that executes this embodiment. In addition, it should be noted that the ordinal numbers such as "first" and "second" mentioned below are used to distinguish multiple objects for the purpose of description, and are not used to limit the order, timing, priority or importance of multiple objects. Please refer to FIG4A for the specific flow of the method. as follows:
S401A:第三通信装置接收第一通信装置的训练能力指示信息,该训练能力指示信息用于指示该第一通信装置支持多模型的训练。S401A: The third communication device receives training capability indication information from the first communication device, where the training capability indication information is used to indicate that the first communication device supports multi-model training.
在本申请实施例中,该第一通信装置可以作为模型训练端(或模型确定端),该第一通信装置可以为但不限于为:模型训练功能网元、或者模型训练功能实体、或者包含模型训练功能的通信装置,例如包含模型训练功能模块的NWDAF网元、或者网元管理系统(EMS)设备、或者接入网设备(如基站)等。该第三通信装置可以为但不限于为:模型管理功能网元、或者模型管理功能实体、或者包含模型管理功能的通信装置,例如包含模型管理功能模块的网络管理系统(NMS)设备。In the embodiment of the present application, the first communication device can be used as a model training end (or model determination end), and the first communication device can be, but is not limited to: a model training function network element, or a model training function entity, or a communication device including a model training function, such as a NWDAF network element including a model training function module, or a network element management system (EMS) device, or an access network device (such as a base station), etc. The third communication device can be, but is not limited to: a model management function network element, or a model management function entity, or a communication device including a model management function, such as a network management system (NMS) device including a model management function module.
在一些实施例中,该第三通信装置可以先向该第一通信装置发送训练能力查询信息,该第一通信装置接收该训练能力查询信息后,再向该第三通信装置发送该训练能力指示信息。在另一些实施例中,该第一通信装置也可以主动向该第三通信装置上报(即发送)该训练能力指示信息。In some embodiments, the third communication device may first send training capability query information to the first communication device, and after the first communication device receives the training capability query information, it may send the training capability indication information to the third communication device. In other embodiments, the first communication device may also actively report (i.e., send) the training capability indication information to the third communication device.
S402A:第三通信装置接收第二通信装置的推理需求信息和推理能力信息,该推理能力信息中包括推理能力指示信息,该推理能力指示信息用于指示该第二通信装置支持多模型的推理。S402A: The third communication device receives reasoning requirement information and reasoning capability information of the second communication device, where the reasoning capability information includes reasoning capability indication information, and the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning.
在本申请实施例中,该第二通信装置可以作为模型推理端(或模型使用端),该第二通信装置可以为但不限于为:模型推理功能网元、或者模型推理功能实体、或者包含模型推理功能的通信装置,例如包含模型推理功能模块的NWDAF网元、或者网元管理系统(EMS)设备、或者接入网设备(如基站)等。In an embodiment of the present application, the second communication device can serve as a model reasoning end (or model usage end), and the second communication device can be but is not limited to: a model reasoning function network element, or a model reasoning function entity, or a communication device including a model reasoning function, such as a NWDAF network element including a model reasoning function module, or a network element management system (EMS) device, or an access network device (such as a base station), etc.
此外,该第一通信装置和该第二通信装置,以及该第三通信装置均可以为相互独立的设备,或者该第一通信装置和该第二通信装置以及该第三通信装置可以分别位于相互独立的设备中;又或者该第一通信装置和该第二通信装置位于同一个装置中;或者该第一通信装置和该第二通信装置为同一设备;因此,本申请实施例对该第一通信装置、该第二通信装置以及该第三通信装置的具体形式、各通信装置所在设备以及所在位置不做具体限定。In addition, the first communication device, the second communication device, and the third communication device can all be independent devices, or the first communication device, the second communication device, and the third communication device can be respectively located in independent devices; or the first communication device and the second communication device are located in the same device; or the first communication device and the second communication device are the same device; therefore, the embodiment of the present application does not specifically limit the specific form of the first communication device, the second communication device, and the third communication device, the device where each communication device is located, and the location.
在一些实施例中,该第三通信装置先向该第二通信装置发送推理需求查询信息和推理能力查询信息,该第二通信装置接收该推理需求查询信息和推理能力查询信息后,再向该第三通信装置发送该推理需求信息和该推理能力信息。In some embodiments, the third communication device first sends reasoning requirement query information and reasoning capability query information to the second communication device. After the second communication device receives the reasoning requirement query information and reasoning capability query information, it sends the reasoning requirement information and the reasoning capability query information to the third communication device.
上述的推理需求查询信息和推理能力查询信息可以由第三通信装置分别单独发送,也可以同时发送,即该第三通信装置发送推理需求查询信息和推理能力查询信息的时间先后不限。此外,该推理需求查询信息和推理能力查询信息可以携带在同一消息中由该第三通信装置发送,也可以分别携带在不同消息中由该第三通信装置发送,本申请实施例对此也不做限定。The above-mentioned reasoning requirement query information and reasoning capability query information can be sent separately by the third communication device, or they can be sent simultaneously, that is, the time when the third communication device sends the reasoning requirement query information and the reasoning capability query information is not limited. In addition, the reasoning requirement query information and the reasoning capability query information can be carried in the same message and sent by the third communication device, or they can be carried in different messages and sent by the third communication device, and the embodiments of the present application do not limit this.
在另一些实施例中,该第二通信装置也可以主动向该第三通信装置上报推理需求信息和推理能力信息,该推理能力信息中包括推理能力指示信息,该推理能力指示信息用于指示该第二通信装置支持多模型的推理。In other embodiments, the second communication device may also actively report reasoning requirement information and reasoning capability information to the third communication device, where the reasoning capability information includes reasoning capability indication information, and the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning.
同理,上述的推理需求信息和推理能力信息可以由第二通信装置分别单独发送,或者由该第二通信装置同时发送,即该第二通信装置发送该推理需求信息和推理能力信息的时间先后不限。此外,该推理需求信息和推理能力信息可以携带在同一消息中并由该第二通信装置发送,也可以分别携带在不同消息中并由该第二通信装置发送,本申请实施例对此也不做限定。Similarly, the above-mentioned reasoning requirement information and reasoning capability information can be sent separately by the second communication device, or sent simultaneously by the second communication device, that is, the time when the second communication device sends the reasoning requirement information and the reasoning capability information is not limited. In addition, the reasoning requirement information and the reasoning capability information can be carried in the same message and sent by the second communication device, or they can be carried in different messages and sent by the second communication device, and the embodiments of the present application do not limit this.
在本申请实施例中,该第二通信装置的推理需求信息可以包括但不限于包括推理的类型(例如覆盖问题分析、小区流量预测等)、推理的性能需求(如精度、准确度等)、推理的速度需求、推理的功耗需求中的一项或多项。In an embodiment of the present application, the reasoning requirement information of the second communication device may include but is not limited to one or more of the type of reasoning (for example, coverage problem analysis, cell traffic prediction, etc.), performance requirements of reasoning (such as precision, accuracy, etc.), speed requirements of reasoning, and power consumption requirements of reasoning.
在本申请实施例对执行上述步骤S401A和S402A的先后顺序不做具体限定。In the embodiment of the present application, there is no specific limitation on the order of executing the above steps S401A and S402A.
S403A:第三通信装置向该第一通信装置发送模型请求信息,该模型请求信息中包括该推理需求信息。相应的,该第一通信装置接收该模型请求信息。S403A: The third communication device sends model request information to the first communication device, wherein the model request information includes the inference requirement information. Correspondingly, the first communication device receives the model request information.
在一种可能的实施方式中,该模型请求信息中还包括多模型指示信息,该多模型指示信息用于指示请求训练或获取的模型为多模型。In a possible implementation, the model request information also includes multi-model indication information, where the multi-model indication information is used to indicate that the model requested to be trained or obtained is a multi-model.
在本申请实施例中,该多模型指示信息也可以携带在该模型请求信息包含的推理需求信息中,或者该第三通信装置单独的将该多模型指示信息发送给该第一通信装置,本申请对此不做具体限定。In an embodiment of the present application, the multi-model indication information may also be carried in the inference requirement information included in the model request information, or the third communication device may separately send the multi-model indication information to the first communication device. The present application does not make any specific limitations on this.
S404A:该第一通信装置根据该模型请求信息,确定第一模型,该第一模型为多模型。S404A: The first communication device determines a first model according to the model request information, where the first model is a multi-model.
在一种实施方式中,该模型请求信息用于请求训练多模型时,该第一通信装置根据该模型请求信息, 确定第一模型,可以包括但不限于以下方式:In one embodiment, when the model request information is used to request training of multiple models, the first communication device, based on the model request information, Determining the first model may include but is not limited to the following methods:
方式1:该模型请求信息中还包括多模型的训练策略,那么该第一通信装置可以根据该推理需求信息和该多模型的训练策略进行训练,得到该第一模型的多个子模型。Method 1: If the model request information also includes a multi-model training strategy, the first communication device can perform training according to the inference requirement information and the multi-model training strategy to obtain multiple sub-models of the first model.
方式2:该第一通信装置先根据该推理需求信息,确定多模型的训练策略;再根据该推理需求信息和该多模型的训练策略进行训练,得到该第一模型的多个子模型。Method 2: The first communication device first determines a multi-model training strategy based on the inference requirement information; and then performs training based on the inference requirement information and the multi-model training strategy to obtain multiple sub-models of the first model.
在本申请实施例中,该多模型的训练策略可以包括但不限于数据处理策略、训练的算法、训练的模式、子模型的数量、子模型的类型等。In an embodiment of the present application, the multi-model training strategy may include but is not limited to data processing strategy, training algorithm, training mode, number of sub-models, type of sub-models, etc.
在另一种实施方式中,该模型请求信息用于请求获取多模型时,该第一通信装置根据该模型请求信息,确定第一模型,可以包括:该第一通信装置根据该推理需求信息,从至少一个预设的多模型中确定该第一模型。在本申请实施例中,该至少一个预设的多模型可以为该第一通信装置提前已经训练好的多模型。In another embodiment, when the model request information is used to request to obtain multiple models, the first communication device determines the first model according to the model request information, which may include: the first communication device determines the first model from at least one preset multiple model according to the inference requirement information. In an embodiment of the present application, the at least one preset multiple model may be a multiple model that has been trained in advance by the first communication device.
S405A:第一通信装置发送第一信息,该第一信息中包括第一模型的信息。S405A: The first communication device sends first information, where the first information includes information of the first model.
相应的,该第三通信装置从该第一通信装置接收该第一信息,该第一模型的信息中包括该第一模型的模型信息和该第一模型的多个子模型的信息,每个子模型的信息包括但不限于子模型的标识信息、子模型的级别、子模型的性能、性能约束。Correspondingly, the third communication device receives the first information from the first communication device, and the information of the first model includes model information of the first model and information of multiple sub-models of the first model. The information of each sub-model includes but is not limited to identification information of the sub-model, the level of the sub-model, the performance of the sub-model, and performance constraints.
示例性地,子模型的标识信息可以为但不限于为:子模型的名称(标识)、子模型的存储地址信息、子模型的唯一标识符号。子模型的级别可以包括第一级子模型和第二级子模型;其中,第二级子模型可以用于聚合多个第一级子模型或者多个第一级子模型的推理信息。Exemplarily, the identification information of the submodel may be, but is not limited to: the name (identification) of the submodel, the storage address information of the submodel, and the unique identification number of the submodel. The level of the submodel may include a first-level submodel and a second-level submodel; wherein the second-level submodel may be used to aggregate the reasoning information of multiple first-level submodels or multiple first-level submodels.
在一些实施例中,该第一模型的多个子模型中包括多个第一级子模型和一个第二级子模型,该第二级子模型用于聚合该多个第一级子模型的推理信息。In some embodiments, the multiple sub-models of the first model include multiple first-level sub-models and one second-level sub-model, and the second-level sub-model is used to aggregate the reasoning information of the multiple first-level sub-models.
在另一些实施例中,该第一模型的多个子模型均为第一级子模型,该第一模型的信息中还包括聚合方法和/或权重信息。In other embodiments, the multiple sub-models of the first model are all first-level sub-models, and the information of the first model also includes aggregation method and/or weight information.
在一种实施方式中,该第一通信装置还向第三通信装置发送该第一模型的推理性能信息,相应的,该第三通信装置接收该第一模型的推理性能信息;该第一模型的推理性能信息可以包括但不限于该第一模型的性能、该第一模型的大小信息、该第一模型的推理的功耗、该第一模型的推理速度、该第一模型的算力等。In one embodiment, the first communication device also sends the reasoning performance information of the first model to the third communication device, and correspondingly, the third communication device receives the reasoning performance information of the first model; the reasoning performance information of the first model may include but is not limited to the performance of the first model, the size information of the first model, the power consumption of the reasoning of the first model, the reasoning speed of the first model, the computing power of the first model, etc.
S406A:第三通信装置向第二通信装置发送第二信息,该第二信息中包括第一模型的信息。S406A: The third communication device sends second information to the second communication device, where the second information includes information of the first model.
相应的,该第二通信装置接收该第二信息。Correspondingly, the second communication device receives the second information.
在一种实施方式中,该第三通信装置在执行该步骤S406A之前,该第三通信装置还执行以下步骤:In one implementation, before the third communication device executes step S406A, the third communication device further executes the following steps:
该第三通信装置根据该第二通信装置的推理需求信息和该推理能力信息,以及该第一模型的推理信息和该第一模型的信息,调整该第一模型中的子模型的数量。通过调整该第一模型的子模型的数量,可以保证实际使用该第一模型的效果较优。The third communication device adjusts the number of sub-models in the first model according to the reasoning requirement information and the reasoning capability information of the second communication device, as well as the reasoning information of the first model and the information of the first model. By adjusting the number of sub-models of the first model, it can be ensured that the effect of actually using the first model is better.
S407A:第二通信装置基于该第一模型的信息,得到该第一模型的推理信息。S407A: The second communication device obtains inference information of the first model based on the information of the first model.
在一种实施方式中,该第一模型的多个子模型中包括多个第一级子模型和一个第二级子模型;该第二通信装置基于该第一模型的信息,得到该第一模型的推理信息,可以包括:In one implementation, the multiple sub-models of the first model include multiple first-level sub-models and one second-level sub-model; the second communication device obtains the reasoning information of the first model based on the information of the first model, which may include:
该第二通信装置基于该多个第一级子模型的信息,利用该多个第一级子模型分别进行推理,得到该多个第一级子模型的推理信息;该第二通信装置再使用该第二级子模型对该多个第一级子模型的推理信息进行聚合,得到该第一模型的推理信息。The second communication device uses the multiple first-level sub-models to perform reasoning respectively based on the information of the multiple first-level sub-models to obtain reasoning information of the multiple first-level sub-models; the second communication device then uses the second-level sub-model to aggregate the reasoning information of the multiple first-level sub-models to obtain the reasoning information of the first model.
在另一种实施方式中,该第一模型的多个子模型均为第一级子模型,该第一模型的信息中还包括聚合方法和/或权重信息时;该第二通信装置基于该第一模型的信息,得到该第一模型的推理信息,可以包括:该第二通信装置基于该多个子模型的信息,利用该多个子模型分别进行推理,得到该多个子模型的推理信息;该第二通信装置再根据该聚合方法和/或权重信息对该多个子模型的推理信息进行聚合,得到该第一模型的推理信息。In another embodiment, when the multiple sub-models of the first model are all first-level sub-models, and the information of the first model also includes aggregation method and/or weight information; the second communication device obtains the reasoning information of the first model based on the information of the first model, which may include: the second communication device uses the multiple sub-models to perform reasoning respectively based on the information of the multiple sub-models to obtain the reasoning information of the multiple sub-models; the second communication device then aggregates the reasoning information of the multiple sub-models according to the aggregation method and/or weight information to obtain the reasoning information of the first model.
综上所述,图4A所示实施例中,第一通信装置接收模型请求信息后,可以根据该模型请求信息中的推理需求信息,确定较合适的多模型(即第一模型),再通过第一信息发送该第一模型的信息;当推理端(即包含模型推理功能模块的第二通信装置)接收到该第一模型的信息后,基于该第一模型的信息,并利用该第一模型进行多模型的推理和结合,可以得到准确性较高的推理结果。因此,可以有效地提高模型的使用效果,保证智能推理(或分析)的性能。 In summary, in the embodiment shown in FIG4A, after the first communication device receives the model request information, it can determine a more appropriate multi-model (i.e., the first model) according to the reasoning requirement information in the model request information, and then send the information of the first model through the first information; when the reasoning end (i.e., the second communication device including the model reasoning function module) receives the information of the first model, based on the information of the first model, the multi-model reasoning and combination are performed using the first model, and a reasoning result with higher accuracy can be obtained. Therefore, the use effect of the model can be effectively improved, and the performance of intelligent reasoning (or analysis) can be guaranteed.
图4B为本申请实施例提出的另一种通信方法的流程示意图。该方法可以由第一通信装置(也可以是第二通信装置)的收发器和/或处理器执行,也可以由该收发器和/或处理器对应的芯片执行。或者该实施例还可由该第一通信装置(也可以是第二通信装置)所连接的控制器或控制设备实现,该控制器或控制设备用于管理包括该第一通信装置(也可以是第二通信装置)在内的至少一个装置。并且针对执行该实施例的通信装置的具体形态,本申请不做具体限定。请参阅图4B,该方法的具体流程如下:Figure 4B is a flow chart of another communication method proposed in an embodiment of the present application. The method can be executed by a transceiver and/or processor of a first communication device (or a second communication device), or by a chip corresponding to the transceiver and/or processor. Alternatively, the embodiment can also be implemented by a controller or control device connected to the first communication device (or a second communication device), and the controller or control device is used to manage at least one device including the first communication device (or a second communication device). And the present application does not make specific restrictions on the specific form of the communication device that executes this embodiment. Please refer to Figure 4B, the specific process of the method is as follows:
S401B:第一通信装置从第二通信装置接收模型请求信息,该模型请求信息中包括推理需求信息。S401B: The first communication device receives model request information from the second communication device, where the model request information includes reasoning requirement information.
在本申请实施例中,该推理需求信息可以包括但不限于推理的类型、推理的性能需求、推理的速度需求、推理的功耗需求。In an embodiment of the present application, the reasoning requirement information may include, but is not limited to, the type of reasoning, the performance requirement of reasoning, the speed requirement of reasoning, and the power consumption requirement of reasoning.
在本申请实施例中,该第一通信装置可以作为模型训练端(或模型确定端),该第一通信装置可以为但不限于为:模型训练功能网元、或者模型训练功能实体、或者包含模型训练功能的通信装置,例如包含模型训练功能模块的NWDAF网元、或者网元管理系统(EMS)设备、或者接入网设备(如基站)等。该第二通信装置可以作为模型推理端(或模型使用端),该第二通信装置可以为但不限于为:模型推理功能网元、或者模型推理功能实体、或者包含模型推理功能的通信装置,例如包含模型推理功能模块的NWDAF网元、或者网元管理系统(EMS)设备、或者接入网设备(如基站)等。In an embodiment of the present application, the first communication device may be used as a model training end (or a model determination end), and the first communication device may be, but is not limited to: a model training function network element, or a model training function entity, or a communication device including a model training function, such as a NWDAF network element including a model training function module, or a network element management system (EMS) device, or an access network device (such as a base station), etc. The second communication device may be used as a model reasoning end (or a model use end), and the second communication device may be, but is not limited to: a model reasoning function network element, or a model reasoning function entity, or a communication device including a model reasoning function, such as a NWDAF network element including a model reasoning function module, or a network element management system (EMS) device, or an access network device (such as a base station), etc.
此外,该第一通信装置和该第二通信装置均可以为相互独立的设备,或者该第一通信装置和该第二通信装置可以分别位于相互独立的设备中;又或者该第一通信装置和该第二通信装置位于同一设备中;或者该第一通信装置和该第二通信装置为同一个装置;因此,本申请实施例对该第一通信装置和该第二通信装置的具体形式、各通信装置所在设备以及所在位置不做具体限定。In addition, the first communication device and the second communication device can be independent devices, or the first communication device and the second communication device can be located in independent devices respectively; or the first communication device and the second communication device are located in the same device; or the first communication device and the second communication device are the same device; therefore, the embodiment of the present application does not specifically limit the specific form of the first communication device and the second communication device, the device where each communication device is located, and the location.
在一些实施例中,该第一通信装置从第二通信装置接收该模型请求信息之前,该第一通信装置可以先向该第二通信装置发送训练能力指示信息,该训练能力指示信息用于指示该第一通信装置支持多模型的训练。在另一些实施例中,该第一通信装置还可以从该第二通信装置接收推理能力信息,该推理能力信息包括但不限于:推理能力指示信息,该推理能力指示信息用于指示该第二通信装置支持多模型的推理。可选的,该推理能力信息中还可以包括推理的算力、存储空间。In some embodiments, before the first communication device receives the model request information from the second communication device, the first communication device may first send training capability indication information to the second communication device, and the training capability indication information is used to indicate that the first communication device supports multi-model training. In other embodiments, the first communication device may also receive reasoning capability information from the second communication device, and the reasoning capability information includes but is not limited to: reasoning capability indication information, and the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning. Optionally, the reasoning capability information may also include the computing power and storage space for reasoning.
S402B:该第一通信装置根据该模型请求信息,确定第一模型,该第一模型为多模型。S402B: The first communication device determines a first model according to the model request information, where the first model is a multi-model.
在本申请实施例中,该第一通信装置在执行该步骤S402B时,具体可以参考上述步骤S404A,此处不再赘述。In the embodiment of the present application, when the first communication device executes the step S402B, specific reference may be made to the above-mentioned step S404A, which will not be repeated here.
S403B:该第一通信装置发送第一信息,该第一信息中包括该第一模型的信息。相应的,该第二通信装置接收该第一信息(即第二信息)。S403B: The first communication device sends first information, the first information including information of the first model. Correspondingly, the second communication device receives the first information (ie, second information).
在本申请实施例中,该第一模型的信息中包括该第一模型的模型信息和该第一模型的多个子模型的信息,每个子模型的信息包括但不限于子模型的标识信息、子模型的级别、子模型的性能、性能约束。In an embodiment of the present application, the information of the first model includes model information of the first model and information of multiple sub-models of the first model. The information of each sub-model includes but is not limited to identification information of the sub-model, the level of the sub-model, the performance of the sub-model, and performance constraints.
示例性地,子模型的标识信息可以为但不限于为:子模型的名称(标识)、子模型的存储地址信息、子模型的唯一标识符号。子模型的级别可以包括第一级子模型、第二级子模型;其中,第二级子模型可以用于聚合多个第一级子模型或者多个第一级子模型的推理信息。Exemplarily, the identification information of the submodel may be, but is not limited to: the name (identification) of the submodel, the storage address information of the submodel, and the unique identification symbol of the submodel. The level of the submodel may include a first-level submodel and a second-level submodel; wherein the second-level submodel may be used to aggregate the reasoning information of multiple first-level submodels or multiple first-level submodels.
S404B:该第二通信装置基于该第一模型的信息,得到该第一模型的推理信息。S404B: The second communication device obtains inference information of the first model based on the information of the first model.
在本申请实施例中,该第二通信装置在执行该步骤S404B时,具体可以参考上述步骤S407A,此处不再赘述。In the embodiment of the present application, when the second communication device executes the step S404B, specific reference may be made to the above-mentioned step S407A, which will not be repeated here.
示例性地,图4C示出了一种本申请实施例提出的多模型训练和推理过程的示例图,在图4C中,模型训练功能会根据模型推理功能的推理需求信息,启动多个学习器训练模型,训练的方式可以包括:模型训练功能先按照某种策略将原始数据集拆分为多个子数据集,即数据集1、数据集2…数据集n,n为正整数;然后,模型训练功能利用每个子数据集训练子学习器,从而获得多个子学习器模型,即图中的子学习器1、子学习器2…子学习器n。模型推理功能从模型训练功能中获得该多个子学习器模型的信息之后,可以根据该多个子学习器模型的信息,将推理数据分别输入到该多个子学习器(即子学习器1、子学习器2…子学习器n)中,得到相应的推理输出,即推理输出1、推理输出2…推理输出n;然后,模型推理功能按照预设的聚合方式(如投票法、简单平均法、加权平均法、线性混合法)对这些推理输出进行结合,得到最终推理输出。For example, FIG4C shows an example diagram of a multi-model training and reasoning process proposed in an embodiment of the present application. In FIG4C, the model training function will start multiple learner training models according to the reasoning requirement information of the model reasoning function. The training method may include: the model training function first splits the original data set into multiple sub-data sets according to a certain strategy, that is, data set 1, data set 2...data set n, where n is a positive integer; then, the model training function uses each sub-data set to train the sub-learner, thereby obtaining multiple sub-learner models, that is, sub-learner 1, sub-learner 2...sub-learner n in the figure. After the model reasoning function obtains the information of the multiple sub-learner models from the model training function, the reasoning data can be input into the multiple sub-learners (that is, sub-learner 1, sub-learner 2...sub-learner n) according to the information of the multiple sub-learner models to obtain corresponding reasoning outputs, that is, reasoning output 1, reasoning output 2...reasoning output n; then, the model reasoning function combines these reasoning outputs according to a preset aggregation method (such as voting method, simple average method, weighted average method, linear hybrid method) to obtain the final reasoning output.
综上所述,图4B和图4C所示的实施例中,在第一通信装置支持多模型训练的情况下,向该第一通信装置请求多模型,可以避免多模型请求失败,而且该第一通信装置可以基于该第二通信装置的推理需求,确定(或训练)合适的且满足推理需求的多模型。因此,在多模型的使用中,通过该方法可以获得 合适的多模型来实现推理,从而可以提高模型的推理(或分析)性能。In summary, in the embodiments shown in FIG. 4B and FIG. 4C , when the first communication device supports multi-model training, requesting the multi-model from the first communication device can avoid the failure of the multi-model request, and the first communication device can determine (or train) a suitable multi-model that meets the reasoning requirements based on the reasoning requirements of the second communication device. Therefore, in the use of the multi-model, the method can obtain Appropriate multiple models are used to implement reasoning, thereby improving the reasoning (or analysis) performance of the model.
下面的几个具体的实施例中,针对不同的应用场景,进一步的详细阐述上述本申请方案提出的一种通信方法。In the following specific embodiments, a communication method proposed by the above-mentioned solution of the present application is further elaborated in detail for different application scenarios.
实施例一:Embodiment 1:
在该实施例一中,将本申请方案应用在上述图3A所示的部署架构中,通过增强OAM(NMS/EMS)域的模型训练与部署流程,以支持基于多学习器模型结合的管理域的推理和分析。本申请方案中的第一通信装置和第二通信装置为图3A中包含模型训练功能模块和模型推理功能模块的同一网元管理系统设备(简称EMS),本申请方案中的第三通信装置为图3A中包含模型管理功能模块的网络管理系统设备(简称NMS)。参考图5所示,该实施例一的具体流程如下:In this first embodiment, the solution of the present application is applied to the deployment architecture shown in FIG. 3A above, by enhancing the model training and deployment process of the OAM (NMS/EMS) domain to support reasoning and analysis of the management domain based on the combination of multiple learner models. The first communication device and the second communication device in the solution of the present application are the same network element management system device (EMS for short) in FIG. 3A that includes a model training function module and a model reasoning function module, and the third communication device in the solution of the present application is a network management system device (NMS for short) in FIG. 3A that includes a model management function module. Referring to FIG. 5, the specific process of the first embodiment is as follows:
S501a:NMS中的模型管理功能模块向EMS中的模型推理功能模块发送推理需求查询信息。S501a: The model management function module in the NMS sends reasoning requirement query information to the model reasoning function module in the EMS.
相应的,该EMS中的模型推理功能模块接收该推理需求查询信息,该推理需求查询信息用于查询该模型推理功能模块的推理需求。Correspondingly, the model reasoning function module in the EMS receives the reasoning requirement query information, and the reasoning requirement query information is used to query the reasoning requirement of the model reasoning function module.
S501b:EMS中的模型推理功能模块向NMS中的模型管理功能模块发送推理需求信息。S501b: The model reasoning function module in the EMS sends reasoning requirement information to the model management function module in the NMS.
相应的,该NMS中的模型管理功能模块接收该推理需求信息。Correspondingly, the model management function module in the NMS receives the reasoning requirement information.
一种实施方式中,上述步骤S501a也可以省略,即NMS中的模型管理功能模块不向EMS中的模型推理功能模块发送推理需求查询信息,而是由该EMS中的模型推理功能模块主动向NMS中的模型管理功能模块上报(即发送)该推理需求信息。In one implementation, the above step S501a may also be omitted, that is, the model management function module in the NMS does not send the reasoning requirement query information to the model reasoning function module in the EMS, but the model reasoning function module in the EMS actively reports (i.e., sends) the reasoning requirement information to the model management function module in the NMS.
上述推理需求信息中可以包括:推理类型需求(例如覆盖问题分析、小区流量预测等)、推理性能需求(例如推理的精度、准确度等)、推理速度需求(可选的)、推理功耗需求(可选的)。其中,该推理速度需求也可以称为推理时延需求,表示对执行推理的时间的需求,例如,单次执行推理的时间小于1s;该推理功耗需求表示对执行推理的功耗的需求,例如,单次推理消耗的能量小于5J。The above-mentioned reasoning requirement information may include: reasoning type requirements (such as coverage problem analysis, cell traffic prediction, etc.), reasoning performance requirements (such as reasoning precision, accuracy, etc.), reasoning speed requirements (optional), and reasoning power consumption requirements (optional). Among them, the reasoning speed requirement can also be called the reasoning delay requirement, which indicates the requirement for the time to execute reasoning, for example, the time for a single reasoning execution is less than 1s; the reasoning power consumption requirement indicates the requirement for the power consumption of executing reasoning, for example, the energy consumed by a single reasoning is less than 5J.
S502a:NMS中的模型管理功能模块向EMS中的模型推理功能模块发送推理能力查询信息。该推理能力查询信息用于查询(或获知)该EMS中的模型推理功能模块的推理能力。S502a: The model management function module in the NMS sends reasoning capability query information to the model reasoning function module in the EMS. The reasoning capability query information is used to query (or obtain) the reasoning capability of the model reasoning function module in the EMS.
S502b:EMS中的模型推理功能模块向NMS中的模型管理功能模块发送推理能力信息。S502b: The model reasoning function module in the EMS sends reasoning capability information to the model management function module in the NMS.
相应的,该NMS中的模型管理功能模块接收该推理能力信息。Correspondingly, the model management function module in the NMS receives the reasoning capability information.
一种实施方式中,上述步骤S502a也可以省略,即NMS中的模型管理功能模块不向该EMS中的模型推理功能模块发送推理能力查询信息,而是由该EMS中的模型推理功能模块主动向NMS中的模型管理功能模块上报(即发送)该推理能力信息。In one implementation, the above step S502a may also be omitted, that is, the model management function module in the NMS does not send reasoning capability query information to the model reasoning function module in the EMS, but the model reasoning function module in the EMS actively reports (i.e., sends) the reasoning capability information to the model management function module in the NMS.
上述推理能力信息中可以包括:推理能力指示信息、推理算力(可选的)、存储空间(可选的)。该存储空间可以为模型推理所占用的存储空间的大小,或者存储的地址等。其中,该推理能力指示信息用于指示EMS中的模型推理功能是否支持多模型的推理。The above-mentioned reasoning capability information may include: reasoning capability indication information, reasoning computing power (optional), and storage space (optional). The storage space may be the size of the storage space occupied by the model reasoning, or the storage address, etc. Among them, the reasoning capability indication information is used to indicate whether the model reasoning function in the EMS supports multi-model reasoning.
示例性地,若该推理能力指示信息对应取值为true(是/正确),则表示支持多模型的推理;若该推理能力指示信息对应取值为false(否/错误),则表示不支持多模型的推理;或者该推理能力指示信息为具体的值,若该推理能力指示信息的值为1,则表示支持多模型的推理,若该推理能力指示信息的值为0,则表示不支持多模型的推理。Exemplarily, if the corresponding value of the reasoning capability indication information is true (yes/correct), it means that multi-model reasoning is supported; if the corresponding value of the reasoning capability indication information is false (no/error), it means that multi-model reasoning is not supported; or the reasoning capability indication information is a specific value, if the value of the reasoning capability indication information is 1, it means that multi-model reasoning is supported, and if the value of the reasoning capability indication information is 0, it means that multi-model reasoning is not supported.
该推理算力可以指模型推理功能模块处可用的算力信息,比如可用的硬件资源信息和硬件资源的利用率,其中,硬件资源可以包括通用算力,例如中央处理单元(central processing unit,CPU),以及高性能算力,例如图形处理单元(graphics processing unit,GPU)、神经网络处理单元(Neural Network Processing Unit,NPU)等。硬件资源信息可以是原始硬件信息,可以包括硬件类型、核数、处理频率等,也可以是量化后的运算能力,通常可以用支持的每秒浮点运算次数(floating-point operations per second,FLOPS)来衡量。The inference computing power may refer to the computing power information available at the model inference function module, such as the available hardware resource information and the utilization rate of the hardware resources, wherein the hardware resources may include general computing power, such as the central processing unit (CPU), and high-performance computing power, such as the graphics processing unit (GPU), the neural network processing unit (NPU), etc. The hardware resource information may be the original hardware information, which may include the hardware type, the number of cores, the processing frequency, etc., or the quantified computing power, which may usually be measured by the number of floating-point operations per second (FLOPS) supported.
S503a:NMS中的模型管理功能模块向EMS中的模型训练功能模块发送训练能力查询信息。相应的,该EMS中的模型训练功能模块接收该训练能力查询信息。该训练能力查询信息用于查询(或获知)该EMS中模型训练功能模型的训练能力。S503a: The model management function module in the NMS sends training capability query information to the model training function module in the EMS. Correspondingly, the model training function module in the EMS receives the training capability query information. The training capability query information is used to query (or obtain) the training capability of the model training function model in the EMS.
S503b:EMS中的模型训练功能模块向NMS中的模型管理功能模块发送训练能力信息。S503b: The model training function module in the EMS sends training capability information to the model management function module in the NMS.
当然,该NMS中的模型管理功能模块也可以不向该EMS中的模型训练功能模块发送训练能力查询信息,即不执行上述步骤S503a,而是由该EMS中的模型训练功能模块主动向该NMS中的模型管理功能 模块上报(即发送)自身的训练能力信息。Of course, the model management function module in the NMS may not send the training capability query information to the model training function module in the EMS, that is, the above step S503a is not executed, but the model training function module in the EMS actively sends the training capability query information to the model management function module in the NMS. The module reports (i.e. sends) its own training capability information.
其中,上述训练能力信息用于通知(或指示)该EMS中的模型训练功能模型是否支持多模型的训练。Among them, the above-mentioned training capability information is used to notify (or indicate) whether the model training function model in the EMS supports multi-model training.
示例性地,该EMS中的模型训练功能模块向NMS中的模型管理功能模块发送训练能力指示信息,若该训练能力指示信息对应取值为true(是或正确),则表示支持多模型的训练,若该训练能力指示信息对应取值为false(否/错误),则表示不支持多模型的训练;或者该训练能力指示信息为具体的值,若该训练能力指示信息的值为1,则表示支持多模型的训练,若该训练能力指示信息的值为0,则表示不支持多模型的训练。Exemplarily, the model training function module in the EMS sends training capability indication information to the model management function module in the NMS. If the corresponding value of the training capability indication information is true (yes or correct), it means that multi-model training is supported; if the corresponding value of the training capability indication information is false (no/error), it means that multi-model training is not supported; or the training capability indication information is a specific value. If the value of the training capability indication information is 1, it means that multi-model training is supported; if the value of the training capability indication information is 0, it means that multi-model training is not supported.
上述步骤S501a-S501b,S502a-S502b,S503a-S503b为推理需求信息和能力信息的查询和上报过程,属于可选的步骤,或者可以离线(线下)完成。另外,本申请实施例对执行上述推理需求信息查询和上报的步骤(即S501a-S501b),推理能力信息查询和上报的步骤(即S502a-S502b),以及训练能力信息查询和上报的步骤(即S503a-S503b)的先后顺序不做具体限定。The above steps S501a-S501b, S502a-S502b, S503a-S503b are the query and reporting process of reasoning requirement information and capability information, which are optional steps, or can be completed offline. In addition, the embodiment of the present application does not specifically limit the order of executing the above-mentioned steps of querying and reporting reasoning requirement information (i.e., S501a-S501b), querying and reporting reasoning capability information (i.e., S502a-S502b), and querying and reporting training capability information (i.e., S503a-S503b).
通过上述步骤S501a-S501b,S502a-S502b,S503a-S503b,NMS中的模型管理功能模块可以确定EMS中的模型推理功能模块的推理需求,以及EMS中的模型训练功能模块的训练能力和EMS中的模型推理功能模块的推理能力,从而可以进一步执行下述的步骤:Through the above steps S501a-S501b, S502a-S502b, S503a-S503b, the model management function module in the NMS can determine the reasoning requirements of the model reasoning function module in the EMS, the training capability of the model training function module in the EMS, and the reasoning capability of the model reasoning function module in the EMS, so as to further perform the following steps:
S504:NMS中的模型管理功能模块向EMS中的模型训练功能模块发送模型训练请求。相应的,该EMS中的模型训练功能模块接收该模型训练请求。S504: The model management function module in the NMS sends a model training request to the model training function module in the EMS. Correspondingly, the model training function module in the EMS receives the model training request.
该模型训练请求中包括:模型标识(或推理类型)、多模型训练指示、多模型的训练策略(可选的)。The model training request includes: model identification (or inference type), multi-model training instructions, and multi-model training strategy (optional).
其中,模型标识也可以表示为推理类型。该多模型训练指示用于指示是否进行多模型的训练。该多模型的训练策略用于指示训练的方法,可以包括处理数据的策略、训练算法指示。The model identifier may also be expressed as an inference type. The multi-model training indication is used to indicate whether to perform multi-model training. The multi-model training strategy is used to indicate the training method, which may include a data processing strategy and a training algorithm indication.
在上述中,处理数据的策略可以包括输入数据采样、特征抽样。其中,输入数据采样是指示对原始数据进行抽样,以构成多个子数据集,每个子数据集用于训练得到一个子模型。特征抽样是指示对原始数据的特征进行抽样,不同特征的数据可以构成多个不同的子数据集,每个子数据集用于训练得到一个子模型。In the above, the data processing strategy may include input data sampling and feature sampling. Among them, input data sampling indicates sampling the original data to form multiple sub-data sets, each of which is used to train a sub-model. Feature sampling indicates sampling the features of the original data, and data with different features can form multiple different sub-data sets, each of which is used to train a sub-model.
训练算法指示可以包括:子模型的个数、模型的类型、超参配置。其中,子模型的个数是指示构成多模型的子模型的个数;模型的类型是指示不同子模型的模型类型,例如随机森林模型、卷积神经网络模型;超参配置是指示模型训练的超参,例如神经网络模型的层数、迭代次数、学习率等。The training algorithm indication may include: the number of sub-models, the type of model, and the hyper-parameter configuration. The number of sub-models indicates the number of sub-models that constitute the multi-model; the type of model indicates the model type of different sub-models, such as random forest model and convolutional neural network model; the hyper-parameter configuration indicates the hyper-parameters of model training, such as the number of layers, number of iterations, and learning rate of the neural network model.
S505:EMS中的模型训练功能模块根据该模型训练请求进行多模型训练,得到第一模型(多模型)。S505: The model training function module in the EMS performs multi-model training according to the model training request to obtain a first model (multi-model).
若该模型训练请求中不包括多模型的训练策略,那么该EMS中的模型训练功能模块可以由自身根据获得的推理需求信息,确定多模型的训练策略。该多模型的训练策略的具体内容可以参照上述步骤S504中的多模型的训练策略的内容,此处不再详细介绍。If the model training request does not include a multi-model training strategy, the model training function module in the EMS can determine the multi-model training strategy by itself based on the obtained reasoning requirement information. The specific content of the multi-model training strategy can refer to the content of the multi-model training strategy in the above step S504, which will not be introduced in detail here.
S506:EMS中的模型训练功能模块向NMS中的模型管理功能模块发送模型训练报告。相应的,该NMS中的模型管理功能模块接收该模型训练报告。S506: The model training function module in the EMS sends a model training report to the model management function module in the NMS. Correspondingly, the model management function module in the NMS receives the model training report.
该模型训练报告中可以包括:该第一模型的模型信息、多模型的指示信息、该第一模型的子模型的列表(即多个子模型的信息)、聚合方法(也可称为结合方法)、权重因子(可选的)、该第一模型的性能、该第一模型的大小、该第一模型的计算量、该第一模型的推理速度、该第一模型的推理能耗。The model training report may include: model information of the first model, indication information of multiple models, a list of sub-models of the first model (i.e., information of multiple sub-models), an aggregation method (also called a combination method), a weight factor (optional), the performance of the first model, the size of the first model, the computational complexity of the first model, the inference speed of the first model, and the inference energy consumption of the first model.
示例性地,该第一模型的模型信息可以为用于标识该第一模型的信息,例如名称或唯一标识符。Exemplarily, the model information of the first model may be information used to identify the first model, such as a name or a unique identifier.
示例性地,该多模型的指示信息用于指示该EMS中的模型训练功能模块训练得到的第一模型为多模型。该第一模型的子模型的列表中包含一系列构成该第一模型的多个子模型的信息,每个子模型的信息中包括:子模型的标识信息、子模型的级别(也可称为子模型的类别)、子模型的性能、子模型的性能约束。该子模型的标识信息可以为子模型的唯一标识符或该子模型的存储地址。该子模型的级别可以包括一级子模型(相当于本申请方案中的第一级子模型)和二级子模型(相当于本申请方案中的第二级子模型,也可称为聚合模型),该二级子模型用于聚合一级子模型的推理信息。Exemplarily, the indication information of the multiple models is used to indicate that the first model trained by the model training function module in the EMS is a multiple model. The list of sub-models of the first model contains a series of information about the multiple sub-models constituting the first model, and the information of each sub-model includes: identification information of the sub-model, the level of the sub-model (also referred to as the category of the sub-model), the performance of the sub-model, and the performance constraint of the sub-model. The identification information of the sub-model can be a unique identifier of the sub-model or a storage address of the sub-model. The level of the sub-model can include a first-level sub-model (equivalent to the first-level sub-model in the present application) and a second-level sub-model (equivalent to the second-level sub-model in the present application, also referred to as an aggregate model), and the second-level sub-model is used to aggregate the reasoning information of the first-level sub-model.
在该模型训练报告中,若该第一模型的子模型的列表中仅包含多个一级子模型,那么该模型训练报告中还包括聚合方法,该聚合方法可以为但不限于为投票法、简单平均法、加权平均法、线性混合法。当该模型训练报告中包含的聚合方法为加权平均法或线性混合法,那么该模型训练报告中还应包括各子模型对应的权重因子。若该第一模型的子模型的列表中包含多个一级子模型和一个二级子模型,那么该模型训练报告中可以不包括聚合方法(也可称为结合方法)和权重因子。 In the model training report, if the list of sub-models of the first model contains only multiple first-level sub-models, then the model training report also includes an aggregation method, which can be but is not limited to voting method, simple average method, weighted average method, linear mixing method. When the aggregation method included in the model training report is weighted average method or linear mixing method, then the model training report should also include the weight factor corresponding to each sub-model. If the list of sub-models of the first model contains multiple first-level sub-models and one second-level sub-model, then the model training report may not include the aggregation method (also called the combination method) and the weight factor.
S507:NMS中的模型管理功能模块确定实际部署的模型。S507: The model management function module in the NMS determines the model actually deployed.
即该NMS中的模型管理功能模块基于来自EMS的推理需求信息和推理能力信息,以及该模型训练报告中的第一模型信息,调整该第一模型并确定实际部署的子模型。That is, the model management function module in the NMS adjusts the first model and determines the sub-model to be actually deployed based on the reasoning requirement information and reasoning capability information from the EMS and the first model information in the model training report.
在该实施例一中,该NMS中的模型管理功能模块可以基于EMS中的模型推理功能模块的推理需求信息和推理能力信息,以及该训练报告中的模型信息,对实际部署该第一模型的子模型的个数和聚合方法进行调整。In this first embodiment, the model management function module in the NMS can adjust the number and aggregation method of sub-models actually deployed of the first model based on the reasoning requirement information and reasoning capability information of the model reasoning function module in the EMS and the model information in the training report.
例如,该NMS中的模型管理功能模块根据各子模型的性能和各子模型的性能约束,减少实际部署该第一模型的子模型的个数。For example, the model management function module in the NMS reduces the number of sub-models actually deployed in the first model according to the performance of each sub-model and the performance constraints of each sub-model.
S508a:NMS中的模型管理功能模块向EMS中的模型推理功能模块发送模型部署请求信息。S508a: The model management function module in the NMS sends model deployment request information to the model reasoning function module in the EMS.
相应的,该EMS中的模型推理功能模块接收该模型部署请求信息,该模型部署请求信息用于请求对该第一模型的子模型进行部署。该模型部署请求信息中包括该第一模型的模型信息、多模型的指示信息、该第一模型的子模型的列表(即多个子模型的信息)、聚合方法(也可称为结合方法)、权重因子(可选的)。Correspondingly, the model reasoning function module in the EMS receives the model deployment request information, which is used to request the deployment of the sub-model of the first model. The model deployment request information includes the model information of the first model, the indication information of multiple models, the list of sub-models of the first model (i.e., the information of multiple sub-models), the aggregation method (also called the combination method), and the weight factor (optional).
进而该EMS中的模型推理功能模块基于该模型部署请求中包含的该第一模型的子模型的信息,对该EMS中的模型训练功能模块训练出的该第一模型的子模型进行实际部署。Then, the model reasoning function module in the EMS actually deploys the sub-model of the first model trained by the model training function module in the EMS based on the information of the sub-model of the first model contained in the model deployment request.
S508b:EMS中的模型推理功能模块向NMS中的模型管理功能模块发送模型部署响应信息。相应的,该NMS中的模型管理功能模块接收到该模型部署响应信息,以确定(或知晓)该EMS中的模型推理功能模块部署该第一模型的子模型完成。S508b: The model reasoning function module in the EMS sends a model deployment response message to the model management function module in the NMS. Accordingly, the model management function module in the NMS receives the model deployment response message to determine (or know) that the model reasoning function module in the EMS has completed the deployment of the sub-model of the first model.
S509:EMS中的模型推理功能模块基于该第一模型进行多模型的推理,得到推理结果。S509: The model reasoning function module in the EMS performs multi-model reasoning based on the first model to obtain a reasoning result.
即该EMS中的模型推理功能模块将待推理的数据(即输入数据),分别输入到该第一模型的实际部署的子模型中,得到对应的子模型的推理结果。进一步的,该模型推理功能模型可以基于聚合方式对各子模型的推理结果进行结合,得到最终推理结果。或者该模型推理功能模型可以将该第一模型的各一级子模型得到的推理结果输入到二级子模型中,输出结合后的最终推理结果。That is, the model reasoning function module in the EMS inputs the data to be inferred (i.e., input data) into the sub-models actually deployed in the first model, respectively, to obtain the reasoning results of the corresponding sub-models. Furthermore, the model reasoning function model can combine the reasoning results of each sub-model based on an aggregation method to obtain the final reasoning result. Alternatively, the model reasoning function model can input the reasoning results obtained by each first-level sub-model of the first model into the second-level sub-model, and output the combined final reasoning result.
例如,EMS中的模型推理功能模型根据该模型推理请求信息中包括的该第一模型的多个一级子模型的信息,可以确定该多个一级子模型,利用该多个一级子模型分别进行推理,然后基于该多个一级子模型的存储地址,获得该多个一级子模型的推理结果,最后采用聚合方法对该多个一级子模型的推理结果进行聚合(或结合),得到最终的推理结果。For example, the model reasoning function model in the EMS can determine the multiple first-level sub-models according to the information of the multiple first-level sub-models of the first model included in the model reasoning request information, use the multiple first-level sub-models to perform reasoning respectively, and then obtain the reasoning results of the multiple first-level sub-models based on the storage addresses of the multiple first-level sub-models, and finally use the aggregation method to aggregate (or combine) the reasoning results of the multiple first-level sub-models to obtain the final reasoning result.
或者该EMS中的模型推理功能模型根据该模型推理请求信息中包括的该第一模型的多个一级子模型的信息,可以确定该多个一级子模型,利用该多个一级子模型分别进行推理,然后基于该多个一级子模型的存储地址,获得该多个一级子模型的推理结果,最后使用二级子模型对该多个一级子模型的推理结果进行聚合(或结合),得到最终的推理结果。Alternatively, the model reasoning function model in the EMS can determine the multiple first-level sub-models based on the information of the multiple first-level sub-models of the first model included in the model reasoning request information, use the multiple first-level sub-models to perform reasoning respectively, and then obtain the reasoning results of the multiple first-level sub-models based on the storage addresses of the multiple first-level sub-models, and finally use the second-level sub-model to aggregate (or combine) the reasoning results of the multiple first-level sub-models to obtain the final reasoning result.
在该实施例一中,EMS中的模型训练功能模块和模型推理功能模块可以对应地向NMS中的模型管理功能模块反馈多模型的训练能力(是否支持多模型的训练)、推理能力(是否支持多模型的推理)以及推理需求信息;在EMS中,该模型管理功能模块向模型训练功能模块发送的模型训练请求中增加多模型的训练指示和多模型的训练策略,进而该模型训练功能模块可以基于该模型训练请求生成第一模型(即多模型),向该模型管理功能模块发送该第一模型的信息,该第一模型的信息中增加多模型的指示信息和该第一模型的子模型的信息,再由该模型管理模块基于该第一模型的信息、以及模型推理功能模块的推理需求和推理能力,对该第一模型的子模型数量进行调整,以确定实际部署的第一模型。因此,通过该实施例一,可知本申请的方案支持管理域的推理功能模块进行多模型结合推理,并支持根据推理需求使用最合适的多模型,可以有效地提高模型的推理(或分析)效果。In this embodiment 1, the model training function module and the model reasoning function module in the EMS can correspondingly feedback the training capability (whether to support the training of multiple models), the reasoning capability (whether to support the reasoning of multiple models) and the reasoning requirement information of the multiple models to the model management function module in the NMS; in the EMS, the model management function module adds the training instructions of multiple models and the training strategy of multiple models to the model training request sent by the model training function module, and then the model training function module can generate the first model (i.e., multiple models) based on the model training request, send the information of the first model to the model management function module, add the indication information of multiple models and the information of the sub-models of the first model to the information of the first model, and then the model management module adjusts the number of sub-models of the first model based on the information of the first model, and the reasoning requirements and reasoning capabilities of the model reasoning function module to determine the first model actually deployed. Therefore, through this embodiment 1, it can be known that the scheme of the present application supports the reasoning function module of the management domain to perform multi-model combined reasoning, and supports the use of the most appropriate multi-model according to the reasoning requirements, which can effectively improve the reasoning (or analysis) effect of the model.
实施例二:Embodiment 2:
在该实施例二中,将本申请方案应用在上述图3B所示的部署架构中,通过增强OAM(NMS/EMS)域的模型训练与部署流程,以支持基于多学习器模型结合的RAN域的推理和分析。该实施例二与上述的实施例一步骤类似,该实施例二的区别是模型推理功能模块位于RAN/gNB中,即本申请方案中的第一通信装置为包含模型训练功能模块的网元管理系统设备(简称EMS),本申请方案中的第二通信装置为包含模型推理功能模块的RAN/gNB,本申请方案中的第三通信装置为包含模型管理功能模块的网络管理系统设备(简称NMS)。此外,RAN/gNB中的模型推理功能模块与NMS设备中的模型管理功能 模块之间可以直接进行交互,也可以通过EMS进行转发,本申请实施例的方案可以在RAN/gNB中的模型推理功能模型支持多模型的推理的情况下可实现。参考图6所示,该实施例二的具体流程如下:In this second embodiment, the solution of the present application is applied to the deployment architecture shown in FIG. 3B above, and the model training and deployment process of the OAM (NMS/EMS) domain is enhanced to support reasoning and analysis of the RAN domain based on the combination of multiple learner models. This second embodiment is similar to the steps of the first embodiment above. The difference of this second embodiment is that the model reasoning function module is located in the RAN/gNB, that is, the first communication device in the solution of the present application is a network element management system device (EMS for short) including a model training function module, the second communication device in the solution of the present application is a RAN/gNB including a model reasoning function module, and the third communication device in the solution of the present application is a network management system device (NMS for short) including a model management function module. In addition, the model reasoning function module in the RAN/gNB and the model management function module in the NMS device are similar. The modules can interact directly with each other or forward through EMS. The solution of the embodiment of the present application can be implemented when the model reasoning function model in the RAN/gNB supports multi-model reasoning. Referring to Figure 6, the specific process of the second embodiment is as follows:
S601a:NMS中的模型管理功能模块向RAN/gNB中的模型推理功能模块发送推理需求查询信息。S601a: The model management function module in the NMS sends reasoning requirement query information to the model reasoning function module in the RAN/gNB.
一种可选的实现方式中,该NMS中的模型管理功能模块可以通过EMS向该RAN/gNB中的模型推理功能模块转发该推理需求查询信息。该推理能力查询信息用于查询(或获知)该RAN/gNB中的模型推理功能模块的推理能力。In an optional implementation, the model management function module in the NMS can forward the reasoning requirement query information to the model reasoning function module in the RAN/gNB through the EMS. The reasoning capability query information is used to query (or obtain) the reasoning capability of the model reasoning function module in the RAN/gNB.
S601b:RAN/gNB中的模型推理功能模块向NMS中的模型管理功能模块发送推理需求信息。S601b: The model reasoning function module in the RAN/gNB sends reasoning requirement information to the model management function module in the NMS.
一种可选的实现方式中,该RAN/gNB中的模型推理功能模块可以通过EMS向NMS中的模型管理功能模块转发该推理需求信息。该推理能力信息中可以包括:推理能力指示信息、推理算力(可选的)、存储空间(可选的)。In an optional implementation, the model reasoning function module in the RAN/gNB can forward the reasoning requirement information to the model management function module in the NMS through the EMS. The reasoning capability information may include: reasoning capability indication information, reasoning computing power (optional), and storage space (optional).
该步骤S601b中的该推理能力信息中的内容可以参考上述步骤S501b,此处不再具体赘述。The content of the reasoning capability information in step S601b can refer to the above step S501b, which will not be described in detail here.
S602a:NMS中的模型管理功能模块向RAN/gNB中的模型推理功能模块发送推理能力查询信息。S602a: The model management function module in the NMS sends reasoning capability query information to the model reasoning function module in the RAN/gNB.
一种可选的实现方式中,该NMS中的模型管理功能模块可以通过EMS向该RAN/gNB中的模型推理功能模块转发该推理能力查询信息。该推理能力查询信息用于查询(或获知)该RAN/gNB中的模型推理功能模块的推理能力。In an optional implementation, the model management function module in the NMS may forward the reasoning capability query information to the model reasoning function module in the RAN/gNB through the EMS. The reasoning capability query information is used to query (or obtain) the reasoning capability of the model reasoning function module in the RAN/gNB.
S602b:RAN/gNB中的模型推理功能模块向NMS中的模型管理功能模块发送推理能力信息。S602b: The model reasoning function module in the RAN/gNB sends reasoning capability information to the model management function module in the NMS.
一种可选的实现方式中,该RAN/gNB中的模型推理功能模块可以通过EMS向该NMS中的模型管理功能模块转发该推理能力信息。该推理能力信息中可以包括:推理能力指示信息、推理算力(可选的)、存储空间(可选的)。In an optional implementation, the model reasoning function module in the RAN/gNB may forward the reasoning capability information to the model management function module in the NMS through the EMS. The reasoning capability information may include: reasoning capability indication information, reasoning computing power (optional), and storage space (optional).
该步骤S602b中的推理能力信息中的内容可以参考上述步骤S502b中的具体描述,此处不再赘述。The content of the reasoning capability information in step S602b can refer to the specific description in the above step S502b, which will not be repeated here.
S603a:NMS中的模型管理功能模块向EMS中的模型训练功能模块发送训练能力查询信息。该步骤S603a可以参见上述步骤S503a中的具体描述,此处不再赘述。S603a: The model management function module in the NMS sends training capability query information to the model training function module in the EMS. The specific description of step S603a can be found in the above step S503a, which will not be repeated here.
S603b:EMS中的模型训练功能模块向NMS中的模型管理功能模块发送训练能力信息。该步骤S603b可以参见上述步骤S503b中的具体描述,此处不再赘述。S603b: The model training function module in the EMS sends the training capability information to the model management function module in the NMS. The specific description of step S603b can be found in the above step S503b, which will not be repeated here.
上述步骤S601a-S601b,S602a-S602b,S603a-S603b为推理需求信息和能力信息的查询和上报过程,属于可选的步骤,或者可以离线(线下)完成。另外,本申请实施例对执行上述推理需求信息查询和上报的步骤(即S601a-S601b),推理能力信息查询和上报的步骤(即S602a-S602b),以及训练能力信息查询和上报的步骤(即S603a-S603b)的先后顺序不做具体限定。The above steps S601a-S601b, S602a-S602b, S603a-S603b are the query and reporting process of reasoning requirement information and capability information, which are optional steps, or can be completed offline. In addition, the embodiment of the present application does not specifically limit the order of executing the above-mentioned steps of querying and reporting reasoning requirement information (i.e., S601a-S601b), querying and reporting reasoning capability information (i.e., S602a-S602b), and querying and reporting training capability information (i.e., S603a-S603b).
S604:NMS中的模型管理功能模块向EMS中的模型训练功能模块发送模型训练请求。该步骤S604可以参见上述步骤S504中的具体描述,此处不再赘述。S604: The model management function module in the NMS sends a model training request to the model training function module in the EMS. The specific description of step S604 can be found in the above step S504, which will not be repeated here.
S605:EMS中的模型训练功能模块根据该模型训练请求进行多模型训练,得到第一模型(多模型)。该步骤S605可以参见上述步骤S505中的具体描述,此处不再赘述。S605: The model training function module in the EMS performs multi-model training according to the model training request to obtain a first model (multi-model). The specific description of step S605 can be found in the above step S505, which will not be repeated here.
S606:EMS中的模型训练功能模块向NMS中的模型管理功能模块发送模型训练报告。该步骤S606可以参见上述步骤S506中的具体描述,此处不再赘述。S606: The model training function module in the EMS sends a model training report to the model management function module in the NMS. The specific description of step S606 can be found in the above step S506, which will not be repeated here.
S607:NMS中的模型管理功能模块确定实际部署的模型。即NMS中的模型管理功能模块基于推理需求信息和推理能力信息,以及该训练报告中的第一模型信息,调整该第一模型并确定实际部署的子模型。该步骤S607可以参见上述步骤S507中的具体描述,此处不再赘述。S607: The model management function module in the NMS determines the model to be actually deployed. That is, the model management function module in the NMS adjusts the first model and determines the sub-model to be actually deployed based on the reasoning requirement information and reasoning capability information, as well as the first model information in the training report. The specific description of step S607 can be found in the above step S507, which will not be repeated here.
S608a:NMS中的模型管理功能模块向RAN/gNB中的模型推理功能模块发送模型部署请求信息。S608a: The model management function module in the NMS sends model deployment request information to the model reasoning function module in the RAN/gNB.
一种可选的实现方式中,该NMS中的模型管理功能模块可以通过EMS向RAN/gNB中的模型推理功能模块转发该模型部署请求信息。该步骤S608a中该模型部署请求信息的内容可以参见上述步骤S508a中的具体描述,此处不再赘述。In an optional implementation, the model management function module in the NMS can forward the model deployment request information to the model reasoning function module in the RAN/gNB through the EMS. The content of the model deployment request information in step S608a can refer to the specific description in the above step S508a, which will not be repeated here.
S608b:RAN/gNB中的模型推理功能模块向NMS中的模型管理功能模块发送模型部署响应信息。S608b: The model reasoning function module in the RAN/gNB sends model deployment response information to the model management function module in the NMS.
一种可选的实现方式中,该RAN/gNB中的模型推理功能模块可以通过EMS向NMS中的模型管理功能模块发送该模型部署响应信息。该步骤S608b中该模型部署响应信息的内容可以参考上述步骤S508b中的模型部署响应信息的描述。In an optional implementation, the model reasoning function module in the RAN/gNB can send the model deployment response information to the model management function module in the NMS through the EMS. The content of the model deployment response information in step S608b can refer to the description of the model deployment response information in the above step S508b.
S609:RAN/gNB中的模型推理功能模块基于该第一模型进行推理,得到推理结果。S609: The model reasoning function module in the RAN/gNB performs reasoning based on the first model to obtain a reasoning result.
示例性地,RAN/gNB中的模型推理功能模块将待推理的数据(即输入数据),分别输入到该第一模型的实际部署的子模型中,得到对应的子模型的推理结果。进一步的,该模型推理功能模型基于聚合 方式对各子模型的推理结果进行结合,得到最终推理结果。或者该模型推理功能模型将该第一模型的各一级子模型得到的推理结果输入到二级子模型中,输出结合后的最终推理结果。Exemplarily, the model reasoning function module in the RAN/gNB inputs the data to be reasoned (i.e., input data) into the sub-models actually deployed in the first model, respectively, to obtain the reasoning results of the corresponding sub-models. The reasoning results of each sub-model are combined to obtain the final reasoning result. Alternatively, the model reasoning function model inputs the reasoning results obtained by each first-level sub-model of the first model into the second-level sub-model, and outputs the combined final reasoning result.
RAN/gNB中的模型推理功能模块执行该步骤S609,可以参考上述EMS中的模型推理功能模块执行步骤S509的具体描述,此处不再赘述。The model reasoning function module in the RAN/gNB executes step S609. The specific description of the model reasoning function module in the above-mentioned EMS executing step S509 can be referred to, and will not be repeated here.
与上述实施例一相比,通过该实施例二,可知本申请的方案支持RAN域的推理功能模块进行多模型的推理和结合,并支持根据推理需求使用最合适的模型进行推理,可以有效的提高模型的推理(或分析)效果。Compared with the above-mentioned embodiment 1, through this embodiment 2, it can be seen that the solution of the present application supports the reasoning function module of the RAN domain to perform reasoning and combination of multiple models, and supports the use of the most appropriate model for reasoning according to the reasoning requirements, which can effectively improve the reasoning (or analysis) effect of the model.
实施例三:Embodiment three:
在该实施例三中,将本申请方案应用在上述图3C所示的部署架构中,通过增强NWDAF的模型发现与订阅流程,以支持基于多模型结合的NWDAF推理(或分析),提高NWDAF的推理(或分析)效果。本申请方案中的第一通信装置为图3C中包含模型训练功能模块的第一NWDAF网元,本申请方案中的第二通信装置为图3C中包含模型推理功能模块的第二NWDAF网元。参考图7所示,该实施例三的具体流程如下:In this third embodiment, the solution of the present application is applied to the deployment architecture shown in FIG. 3C above, and the model discovery and subscription process of NWDAF is enhanced to support NWDAF reasoning (or analysis) based on multi-model combination, thereby improving the reasoning (or analysis) effect of NWDAF. The first communication device in the solution of the present application is the first NWDAF network element including the model training function module in FIG. 3C, and the second communication device in the solution of the present application is the second NWDAF network element including the model reasoning function module in FIG. 3C. Referring to FIG. 7, the specific process of the third embodiment is as follows:
S701:第一NWDAF网元向NRF网元发送NF注册请求信息。相应的,该NRF网元接收该NF注册请求信息。S701: The first NWDAF network element sends NF registration request information to the NRF network element. Correspondingly, the NRF network element receives the NF registration request information.
该NF注册请求信息中可以包括推理标识(也可称为分析标识)、支持多模型训练的能力指示、推理性能、推理速度(可选的)、推理功耗(可选的)。The NF registration request information may include an inference identifier (also called an analysis identifier), an indication of the ability to support multi-model training, inference performance, inference speed (optional), and inference power consumption (optional).
S702:NRF网元向第一NWDAF网元发送NF注册请求的响应信息。相应的,该第一NWDAF网元接收该NF注册请求的响应信息。S702: The NRF network element sends the response information of the NF registration request to the first NWDAF network element. Correspondingly, the first NWDAF network element receives the response information of the NF registration request.
通过上述步骤S701-S702,该第一NWDAF网元的模型训练功能模块向NRF网元上报自身的模型训练能力信息。Through the above steps S701-S702, the model training function module of the first NWDAF network element reports its own model training capability information to the NRF network element.
在本申请实施例中,以第一NWDAF网元为例介绍其向NRF网元上报自身的模型训练能力信息,而实际中,可能有多个包含模型训练功能的NWDAF网元向NRF网元上报自身的模型训练能力信息,每个包含模型训练功能的NWDAF网元均可以参考上述步骤S701-S702的方式向NRF网元上报自身的模型训练能力信息。In the embodiment of the present application, the first NWDAF network element is taken as an example to introduce its reporting of its own model training capability information to the NRF network element. In practice, there may be multiple NWDAF network elements including model training functions that report their own model training capability information to the NRF network element. Each NWDAF network element including model training functions can report its own model training capability information to the NRF network element by referring to the above steps S701-S702.
S703:NWDAF消费者向第二NWDAF网元发送分析订阅请求信息。相应的,该第二NWDAF网元接收该分析订阅请求信息。S703: The NWDAF consumer sends analysis subscription request information to the second NWDAF network element. Correspondingly, the second NWDAF network element receives the analysis subscription request information.
S704:第二NWDAF网元向该NWDAF消费者发送分析订阅请求的响应信息。相应的,该NWDAF消费者接收该分析订阅请求的响应信息。S704: The second NWDAF network element sends the response information of the analysis subscription request to the NWDAF consumer. Correspondingly, the NWDAF consumer receives the response information of the analysis subscription request.
S705:第二NWDAF网元向NRF网元发送NF发现请求信息。相应的,该NRF网元接收该NF发现请求信息,该NF发现请求信息(相当于上述本申请方案中的推理需求信息)中可以包括:NF类型、推理标识、推理的性能需求、多模型的训练能力指示信息、推理的速度需求、推理的功耗需求。S705: The second NWDAF network element sends NF discovery request information to the NRF network element. Correspondingly, the NRF network element receives the NF discovery request information, and the NF discovery request information (equivalent to the reasoning requirement information in the above-mentioned solution of the present application) may include: NF type, reasoning identifier, reasoning performance requirement, multi-model training capability indication information, reasoning speed requirement, and reasoning power consumption requirement.
S706:NRF网元向第一NWDAF网元发送NF发现请求的响应信息。相应的,该第一NWDAF网元接收该NF发现请求的响应信息。该NF发现请求的响应信息中包括具备模型训练功能且支持多模型训练的NWDAF网元地址。S706: The NRF network element sends a response message of the NF discovery request to the first NWDAF network element. Correspondingly, the first NWDAF network element receives the response message of the NF discovery request. The response message of the NF discovery request includes the address of the NWDAF network element that has the model training function and supports multi-model training.
由于在该实施例三中,该具备模型训练功能且支持多模型训练的NWDAF网元以第一NWDAF网元为例,因此,该NF发现请求的响应信息中包括该第一NWDAF网元的地址。Since in the third embodiment, the NWDAF network element that has the model training function and supports multi-model training takes the first NWDAF network element as an example, the response information of the NF discovery request includes the address of the first NWDAF network element.
S707:第二NWDAF网元向第一NWDAF网元发送模型订阅请求信息。相应的,该第一NWDAF网元接收该模型订阅请求信息。S707: The second NWDAF network element sends model subscription request information to the first NWDAF network element. Correspondingly, the first NWDAF network element receives the model subscription request information.
该模型订阅请求信息(相当于上述本申请方案中的模型请求信息)中包括推理标识、推理的性能需求、请求多模型推理的指示信息(该指示信息用于请求多模型)、推理的速度需求、推理的功耗需求(这些相当于上述本申请方案中的推理需求信息)。The model subscription request information (equivalent to the model request information in the above-mentioned solution of the present application) includes an inference identifier, inference performance requirements, indication information for requesting multi-model inference (the indication information is used to request multiple models), inference speed requirements, and inference power consumption requirements (these are equivalent to the inference requirement information in the above-mentioned solution of the present application).
S708:第一NWDAF网元向第二NWDAF网元发送模型订阅请求的响应信息。相应的,该第二NWDAF网元接收该模型订阅请求的响应信息。S708: The first NWDAF network element sends response information of the model subscription request to the second NWDAF network element. Correspondingly, the second NWDAF network element receives the response information of the model subscription request.
S709:第一NWDAF网元根据该模型请阅请求信息进行多模型训练,得到第一模型(即多模型)。S709: The first NWDAF network element performs multi-model training according to the model reading request information to obtain a first model (ie, multi-model).
在该实施例三中,该步骤S709为可选的步骤,即该步骤S709可以执行,也可以不执行。当不执行该步骤S709时,那么该第一NWDAF网元可以根据该模型订阅请求信息中包含的推理需求信息, 直接从至少一个已训练好的多模型中,选择合适的多模型(即该第一模型)。In the third embodiment, step S709 is an optional step, that is, step S709 may be executed or not executed. When step S709 is not executed, the first NWDAF network element may, based on the inference requirement information included in the model subscription request information, Directly select a suitable multi-model (ie, the first model) from at least one trained multi-model.
S710:第一NWDAF网元向第二NWDAF网元发送模型通知信息。相应的,该第二NWDAF网元接收该模型通知信息,该模型通知信息中包括推理标识、该第一模型的指示信息、该第一模型的子模型列表(即多个子模型的信息)、聚合方法(也可称为结合方法)、权重因子(可选的)、该第一模型的性能信息。S710: The first NWDAF network element sends model notification information to the second NWDAF network element. Correspondingly, the second NWDAF network element receives the model notification information, which includes an inference identifier, indication information of the first model, a sub-model list of the first model (i.e., information of multiple sub-models), an aggregation method (also referred to as a combination method), a weight factor (optional), and performance information of the first model.
该第一模型的指示信息用于指示该第一模型为多模型。每个子模型的信息中可以包括但不限于:子模型的标识信息(例如子模型的唯一标识符或存储地址信息)、子模型的级别(也可称为子模型的类别)、子模型的性能、子模型的性能约束。The indication information of the first model is used to indicate that the first model is a multi-model. The information of each sub-model may include but is not limited to: identification information of the sub-model (such as a unique identifier or storage address information of the sub-model), the level of the sub-model (also referred to as the category of the sub-model), the performance of the sub-model, and the performance constraint of the sub-model.
上述聚合方式、第一模型的性能、以及子模型的相关信息的具体解释,均可以参考上述实施例一中的具体介绍,此处不再赘述。For the specific explanation of the above-mentioned aggregation method, the performance of the first model, and the relevant information of the sub-model, please refer to the specific introduction in the above-mentioned embodiment 1, which will not be repeated here.
S711:第二NWDAF网元基于该模型通知信息进行多模型的推理(或分析),得到推理(或分析)结果。S711: The second NWDAF network element performs multi-model reasoning (or analysis) based on the model notification information to obtain a reasoning (or analysis) result.
示例性地,该步骤S711具体可以参考上述步骤S509或S609中的模型推理功能模块基于该第一模型进行推理得到推理结果的方式来执行,此处不再具体赘述。Exemplarily, the step S711 may be specifically executed with reference to the manner in which the model reasoning function module in the above step S509 or S609 performs reasoning based on the first model to obtain a reasoning result, which will not be described in detail here.
S712:第二NWDAF网元向NWDAF消费者发送推理结果的通知信息,该推理结果通知信息中包括推理标识、推理结果或分析结果。相应的,该NWDAF消费者接收该推理结果的通知信息。S712: The second NWDAF network element sends notification information of the reasoning result to the NWDAF consumer, where the notification information of the reasoning result includes the reasoning identifier, the reasoning result or the analysis result. Correspondingly, the NWDAF consumer receives the notification information of the reasoning result.
在该实施例三中,包含模型训练功能模块的第一NWDAF网元和包含模型推理功能模块的第二NWDAF网元可以向NRF网元上报对应的训练能力和模型能力、推理需求信息,并且包含模型训练功能模块的第一NWDAF网元和包含模型推理功能模块的第二NWDAF网元之间可以交互推理的需求,还可以交互多模型的指示信息。因此,该实施例三的方案支持NWDAF根据推理需求选择(或训练)合适的多模型,并且支持NWDAF进行多模型结合推理,从而可以有效地提高NWDAF的推理(或分析)效果。In this third embodiment, the first NWDAF network element including the model training function module and the second NWDAF network element including the model reasoning function module can report the corresponding training capabilities and model capabilities, and reasoning requirement information to the NRF network element, and the first NWDAF network element including the model training function module and the second NWDAF network element including the model reasoning function module can exchange reasoning requirements and multi-model indication information. Therefore, the solution of this third embodiment supports NWDAF to select (or train) appropriate multi-models according to reasoning requirements, and supports NWDAF to perform multi-model combined reasoning, thereby effectively improving the reasoning (or analysis) effect of NWDAF.
实施例四:Embodiment 4:
在该实施例四中,将本申请方案应用在上述图3D所示的部署架构中,通过增强RAN模型部署和切换流程,以支持基于多模型推理的智能化推理(或分析)。本申请方案中的第一通信装置可以为图3D中包含模型训练功能模块的基站gNB(即源gNB\目标gNB),本申请方案中的第二通信装置为图3D中包含模型推理功能模型的UE1。参考图8所示,该实施例四的具体流程如下:In this fourth embodiment, the solution of this application is applied to the deployment architecture shown in FIG. 3D above, by enhancing the RAN model deployment and switching process to support intelligent reasoning (or analysis) based on multi-model reasoning. The first communication device in this application solution can be a base station gNB (i.e., source gNB\target gNB) including a model training function module in FIG. 3D, and the second communication device in this application solution is UE1 including a model reasoning function model in FIG. 3D. Referring to FIG. 8, the specific process of this fourth embodiment is as follows:
S801a:目标gNB向源gNB发送目标gNB的AI能力信息。相应的,该源gNB接收该目标gNB的AI能力信息。S801a: The target gNB sends the AI capability information of the target gNB to the source gNB. Correspondingly, the source gNB receives the AI capability information of the target gNB.
S801b:源gNB向目标gNB发送源gNB的AI能力信息。相应的,该目标gNB接收该源gNB的AI能力信息。S801b: The source gNB sends the AI capability information of the source gNB to the target gNB. Correspondingly, the target gNB receives the AI capability information of the source gNB.
上述步骤S801a和S801b是以一个源gNB和一个目标gNB为例来介绍源gNB和目标gNB之间交互AI能力信息的过程,在实际应用中,与源gNB交互AI能力信息的gNB不限于该目标gNB。示例性地,目标gNB和源gNB在上线后,可以通过Xn接口交互AI能力信息,该AI能力信息可以包括:AI开关,训练能力指示信息(即支持多模型的训练)。The above steps S801a and S801b take a source gNB and a target gNB as an example to introduce the process of exchanging AI capability information between the source gNB and the target gNB. In practical applications, the gNB that exchanges AI capability information with the source gNB is not limited to the target gNB. Exemplarily, after the target gNB and the source gNB go online, they can exchange AI capability information through the Xn interface. The AI capability information may include: AI switch, training capability indication information (i.e., supporting multi-model training).
S802a:源gNB向UE1发送推理能力查询信息。S802a: The source gNB sends reasoning capability query information to UE1.
示例性的,以UE1为例,该UE1接入源gNB之后,该源gNB向UE1发送推理能力查询信息,该推理能力查询信息用于查询(或获取)该UE1的推理能力信息。Exemplarily, taking UE1 as an example, after UE1 accesses the source gNB, the source gNB sends reasoning capability query information to UE1, and the reasoning capability query information is used to query (or obtain) the reasoning capability information of UE1.
S802b:UE1向源gNB发送推理能力信息。S802b: UE1 sends reasoning capability information to the source gNB.
示例性地,该UE1可以通过Uu接口向该源gNB发送推理能力信息。或者,UE1接入源gNB之后,该源gNB不向该UE1发送推理能力查询信息,而是由该UE1通过Uu接口主动向该源gNB上报自身的推理能力信息。Exemplarily, the UE1 may send reasoning capability information to the source gNB through the Uu interface. Alternatively, after the UE1 accesses the source gNB, the source gNB does not send reasoning capability query information to the UE1, but the UE1 actively reports its own reasoning capability information to the source gNB through the Uu interface.
其中,该UE1的推理能力信息中可以包括:AI开关,存储空间大小,推理的算力,多模型的推理指示信息(即指示支持多模型的推理)、UE1的剩余电量(可选的)。Among them, the reasoning capability information of UE1 may include: AI switch, storage space size, reasoning computing power, multi-model reasoning indication information (i.e., indicating support for multi-model reasoning), and remaining power of UE1 (optional).
本申请实施例对执行源gNB与目标gNB之间交互AI能力信息的过程(即上述步骤S801a-S801b)和UE1向源gNB上报推理能力信息的过程(即上述步骤S802a-S802b)的时间先后顺序不做具体限定。The embodiment of the present application does not specifically limit the time sequence of executing the process of exchanging AI capability information between the source gNB and the target gNB (i.e., the above steps S801a-S801b) and the process of UE1 reporting reasoning capability information to the source gNB (i.e., the above steps S802a-S802b).
S803:源gNB基于UE1的推理能力信息进行多模型训练,得到第一模型(即多模型)。该步骤S803为可选的步骤,即可以执行,也可以不执行。 S803: The source gNB performs multi-model training based on the reasoning capability information of UE1 to obtain a first model (i.e., a multi-model). This step S803 is an optional step, which may be executed or not.
在一种可能的实施方式中,该源gNB也可以基于该UE1的推理能力信息,直接从至少一个已训练好的多模型中选择合适的多模型(即第一模型)。In a possible implementation, the source gNB may also directly select a suitable multi-model (i.e., the first model) from at least one trained multi-model based on the reasoning capability information of the UE1.
S804:源gNB向该UE1发送该第一模型的通知信息。S804: The source gNB sends notification information of the first model to UE1.
示例性地,该源gNB可以通过Uu接口向该UE1下发该第一模型的通知信息,相应的,该UE1接收该第一模型的通知信息。Exemplarily, the source gNB may send notification information of the first model to the UE1 through the Uu interface, and correspondingly, the UE1 receives the notification information of the first model.
该第一模型的通知信息中可以包括:该第一模型的标识、该第一模型的指示信息(即用于指示该第一模型为多模型)、该第一模型的子模型的列表(即多个子模型的信息)、聚合方法(也可称为结合方法)、权重因子(可选的)、该第一模型的性能。The notification information of the first model may include: the identifier of the first model, indication information of the first model (i.e., used to indicate that the first model is a multi-model), a list of sub-models of the first model (i.e., information of multiple sub-models), an aggregation method (also called a combination method), a weight factor (optional), and the performance of the first model.
S805a:源gNB可以基于该第一模型的信息进行多模型的推理,得到推理结果。S805a: The source gNB may perform multi-model inference based on the information of the first model to obtain an inference result.
该步骤S805a为可选的步骤。若该第一模型为双边模型,那么该源gNB执行该步骤S805a;若该第一模型为单边的UE模型,那么该源gNB不执行该步骤S805a。This step S805a is an optional step. If the first model is a bilateral model, the source gNB executes this step S805a; if the first model is a unilateral UE model, the source gNB does not execute this step S805a.
S805b:该UE1可以基于该第一模型的信息进行多模型的推理,得到推理结果。S805b: The UE1 may perform multi-model reasoning based on the information of the first model to obtain a reasoning result.
此时,源gNB还根据UE1的上报的接收信号强度,判断是否需要切换UE1接入的基站(gNB)。例如,当源gNB接收来自UE1的信号强度低于设定的阈值时,则确定触发切换。At this time, the source gNB also determines whether it is necessary to switch the base station (gNB) accessed by UE1 based on the received signal strength reported by UE1. For example, when the signal strength received by the source gNB from UE1 is lower than the set threshold, it is determined to trigger the switch.
当源gNB确定切换UE1接入的gNB时,执行以下步骤:When the source gNB determines to switch the gNB to which UE1 is to be connected, the following steps are performed:
S806:源gNB向UE1发送RRC连接重配消息,该RRC连接重配消息中包括测量配置信息。S806: The source gNB sends an RRC connection reconfiguration message to UE1, which includes measurement configuration information.
示例性地,该源gNB可以通过Uu接口向该UE1发送该RRC连接重配消息。相应的,该UE1通过该Uu接口接收该RRC连接重配消息。Exemplarily, the source gNB may send the RRC connection reconfiguration message to the UE1 via the Uu interface. Correspondingly, the UE1 receives the RRC connection reconfiguration message via the Uu interface.
S807:UE1基于该测量配置信息执行测量,得到该UE1的测量报告。S807: UE1 performs measurement based on the measurement configuration information and obtains a measurement report of UE1.
S808:UE1向源gNB发送该UE1的测量报告。S808: UE1 sends a measurement report of UE1 to the source gNB.
示例性地,该UE1可以通过Uu接口向源gNB发送UE1的测量报告。相应的,源gNB通过该Uu接口接收该UE1的测量报告。Exemplarily, the UE1 may send a measurement report of the UE1 to the source gNB via the Uu interface. Correspondingly, the source gNB receives the measurement report of the UE1 via the Uu interface.
S809:源gNB根据UE1的测量报告和邻站的AI能力信息,确定目标gNB。S809: The source gNB determines the target gNB based on the measurement report of UE1 and the AI capability information of the neighboring station.
可以理解,目标gNB一般具备多模型训练的能力。例如,gNB1、gNB2、gNB3参考上述步骤S801a和S801b与该源gNB交互AI能力信息,使得该源gNB获得gNB1的AI能力信息、gNB2的AI能力信息、以及gNB3的AI能力信息;进而在该步骤中该源gNB可以根据这三个基站的AI能力信息和UE1的测量报告,从这三个基站中选择合适的gNB1作为目标gNB。It can be understood that the target gNB generally has the ability to train multiple models. For example, gNB1, gNB2, and gNB3 refer to the above steps S801a and S801b to exchange AI capability information with the source gNB, so that the source gNB obtains the AI capability information of gNB1, the AI capability information of gNB2, and the AI capability information of gNB3; and then in this step, the source gNB can select a suitable gNB1 as the target gNB from the three base stations based on the AI capability information of the three base stations and the measurement report of UE1.
S810:源gNB向目标gNB发送UE1的切换请求信息。相应的,该目标gNB接收该UE1的切换请求信息。该UE1的切换请求信息中包括UE1的标识、该第一多模型的指示信息(或使用该第一模型的指示信息)、该第一模型的标识、该第一模型的子模型的信息。S810: The source gNB sends the handover request information of UE1 to the target gNB. Correspondingly, the target gNB receives the handover request information of UE1. The handover request information of UE1 includes the identifier of UE1, the indication information of the first multi-model (or the indication information of using the first model), the identifier of the first model, and the information of the sub-model of the first model.
S811:目标gNB基于该UE1的切换请求信息进行多模型训练,得到第二模型(即多模型)。S811: The target gNB performs multi-model training based on the handover request information of UE1 to obtain a second model (i.e., multi-model).
该步骤S811为可选的步骤,即可以执行该步骤S811,也可以不执行该步骤S811。若该目标gNB不执行该步骤S811时,该目标gNB可以基于该UE1的切换请求信息,直接从已训练好的至少一个多模型中选择一个多模型(即第二模型);或者该目标gNB可以直接使用上述源gNB的多模型(即第一模型)。This step S811 is an optional step, that is, step S811 may be executed or not. If the target gNB does not execute step S811, the target gNB may directly select a multi-model (i.e., the second model) from at least one trained multi-model based on the handover request information of the UE1; or the target gNB may directly use the multi-model (i.e., the first model) of the source gNB.
S812:目标gNB与UE1完成随机接入。S812: The target gNB and UE1 complete random access.
在该步骤S812中,该目标gNB与该UE1进行随机接入流程,以使得该UE1成功接入该目标gNB进行通信,该随机接入的流程具体可以参考现有的随机接入方法实现,此处不再具体描述。In step S812, the target gNB performs a random access process with the UE1 so that the UE1 successfully accesses the target gNB for communication. The random access process can be specifically implemented by referring to the existing random access method and will not be described in detail here.
S813:目标gNB向UE1发送第二模型的通知信息。S813: The target gNB sends notification information of the second model to UE1.
示例性地,该目标gNB可以通过Uu接口向UE1发送该第二模型的通知信息,相应的,该UE1通过Uu接口接收该第二模型的通知信息。Exemplarily, the target gNB may send notification information of the second model to UE1 through the Uu interface, and correspondingly, the UE1 receives notification information of the second model through the Uu interface.
该第二模型的通知信息中可以包括第二模型的模型信息(如第二模型的名称、标识、类型等)、第二模型的指示信息(用于指示该第二模型为多模型)、该第二模型的多个子模型列表(即多个子模型的信息)、聚合方法,权重因子(可选的)、该第二模型的性能。The notification information of the second model may include model information of the second model (such as the name, identifier, type, etc. of the second model), indication information of the second model (used to indicate that the second model is a multi-model), a list of multiple sub-models of the second model (i.e., information of multiple sub-models), aggregation method, weight factor (optional), and performance of the second model.
其中,每个子模型的列表(或每个子模型的信息)可以包括:子模型的级别(也可称为子模型的类别),子模型的标识信息(如子模型的存储地址、唯一标识符)。The list of each sub-model (or the information of each sub-model) may include: the level of the sub-model (also referred to as the category of the sub-model), the identification information of the sub-model (such as the storage address of the sub-model, the unique identifier).
子模型的级别可包括第一级子模型和第二级子模型,该第一级子模型用于推理或分析,该第二级子模型用于聚合(或结合)多个第一级子模型的推理信息。The level of sub-models may include a first-level sub-model and a second-level sub-model, wherein the first-level sub-model is used for reasoning or analysis, and the second-level sub-model is used to aggregate (or combine) reasoning information of multiple first-level sub-models.
上述聚合方式、该第二模型的性能、以及该第二模型的子模型的相关信息的具体解释,均可以参考 上述实施例一中聚合方式、第一模型的性能以及该第一模型的子模型的相关信息的具体介绍,此处不再赘述。For detailed explanations of the above-mentioned aggregation method, the performance of the second model, and the related information of the sub-models of the second model, please refer to The detailed introduction of the aggregation method, the performance of the first model and the related information of the sub-models of the first model in the above-mentioned embodiment 1 will not be repeated here.
S814a:目标gNB基于该第二模型的通知信息进行多模型的推理,得到推理结果。S814a: The target gNB performs multi-model inference based on the notification information of the second model to obtain an inference result.
即该目标gNB基于该第二模型的通知信息中第二模型的信息,分别利用该第二模型的子模型进行推理,得到多个子模型的推理结果,并采用聚合方式进行结合,得到最终的推理结果;或者分别利用该第二模型的一级子模型进行推理,得到多个一级子模型的推理结果,再利用二级子模型对这些一级子模型的推理结果进行结合,得到最终的推理结果。That is, the target gNB uses the sub-models of the second model for reasoning based on the information of the second model in the notification information of the second model to obtain reasoning results of multiple sub-models, and combines them in an aggregation manner to obtain the final reasoning result; or uses the first-level sub-models of the second model for reasoning to obtain reasoning results of multiple first-level sub-models, and then uses the second-level sub-model to combine the reasoning results of these first-level sub-models to obtain the final reasoning result.
该步骤S814a为可选的步骤,可以执行该步骤S814a,也可以不执行该步骤S814a。The step S814a is an optional step, and the step S814a may be performed or may not be performed.
S814b:UE1基于该第二模型的通知信息进行多模型的推理,得到推理结果。S814b: UE1 performs multi-model inference based on the notification information of the second model to obtain an inference result.
UE1中的模型推理功能模块可以基于该第二模型的通知信息中该第二模型的多个子模型的信息,利用该多个子模型进行推理,得到该多个子模型的推理信息,再采用聚合方法对该多个子模型的推理信息进行聚合或结合,得到该第二模型的推理信息;或者该UE1中的模型推理功能模块可以使用第二子模型对该多个第一级子模型的推理信息进行聚合或结合,得到该第二模型的推理信息。The model reasoning function module in UE1 can perform reasoning using the multiple sub-models based on the information of the multiple sub-models of the second model in the notification information of the second model to obtain the reasoning information of the multiple sub-models, and then use an aggregation method to aggregate or combine the reasoning information of the multiple sub-models to obtain the reasoning information of the second model; or the model reasoning function module in UE1 can use the second sub-model to aggregate or combine the reasoning information of the multiple first-level sub-models to obtain the reasoning information of the second model.
该步骤S814b具体也可以参考上述步骤S509或S609或S711,此处不再具体赘述。The details of step S814b may also refer to the above-mentioned step S509 or S609 or S711, which will not be described in detail here.
通过上述实施例四中,新增基站之间的多模型训练能力信息交互,以及基站和终端之间的多模型训练/推理能力交互,在模型通知信息中增加多模型的指示信息。通过该实施例四,可知本申请方案支持UE基于多模型进行结合推理,以提高模型的推理(或分析)效果。此外,当源gNB确定触发切换新的gNB时,该源gNB会根据各gNB的多模型能力信息,优选具备多模型训练能力的基站作为切换的目标gNB,UE切换接入到目标gNB后,依然可以基于多模型进行结合推理,以提高模型的推理(或分析)效果。Through the above-mentioned fourth embodiment, the multi-model training capability information interaction between new base stations and the multi-model training/inference capability interaction between the base station and the terminal are added, and the indication information of the multi-model is added in the model notification information. Through this fourth embodiment, it can be known that the present application scheme supports the UE to perform combined reasoning based on multiple models to improve the reasoning (or analysis) effect of the model. In addition, when the source gNB determines to trigger the switching of a new gNB, the source gNB will, based on the multi-model capability information of each gNB, preferably use a base station with multi-model training capability as the target gNB for switching. After the UE switches to access the target gNB, it can still perform combined reasoning based on multiple models to improve the reasoning (or analysis) effect of the model.
实施例五:Embodiment five:
在该实施例五中,将本申请方案应用在上述图3E所示的部署架构中,即在使用双边模型场景下执行基站和UE的联合训练的场景。在该实施例五中,本申请方案中的第一通信装置和第二通信装置可以为包含模型训练功能模型和模型推理功能模块的基站(gNB)或UE(例如UE1)。参考图9所示,该实施例五的具体流程如下:In this fifth embodiment, the solution of the present application is applied to the deployment architecture shown in FIG. 3E above, that is, a scenario in which joint training of a base station and a UE is performed in a bilateral model scenario. In this fifth embodiment, the first communication device and the second communication device in the solution of the present application may be a base station (gNB) or a UE (e.g., UE1) including a model training function model and a model reasoning function module. Referring to FIG. 9 , the specific process of the fifth embodiment is as follows:
S901a:目标gNB向源gNB发送目标gNB的AI能力信息。S901a: The target gNB sends the AI capability information of the target gNB to the source gNB.
示例性地,目标gNB和源gNB上线后,可以通过Xn接口交互AI能力信息,该AI能力信息可以包括:AI开关,支持多模型训练能力指示(是/否)。Exemplarily, after the target gNB and the source gNB come online, they can exchange AI capability information through the Xn interface. The AI capability information may include: AI switch, support for multi-model training capability indication (yes/no).
S901b:源gNB向目标gNB发送源gNB的AI能力信息。S901b: The source gNB sends the AI capability information of the source gNB to the target gNB.
参考上述步骤S901a和S901b,源gNB与多个gNB(包括目标gNB)之间可以通过对应的Xn接口交互各自的AI能力信息,该实施例五以源gNB和目标gNB为例来介绍。Referring to the above steps S901a and S901b, the source gNB and multiple gNBs (including the target gNB) can exchange their respective AI capability information through the corresponding Xn interface. This embodiment 5 takes the source gNB and the target gNB as an example.
S902a:源gNB向UE1发送能力查询信息。S902a: The source gNB sends capability query information to UE1.
示例性地,UE1接入源gNB后,源gNB可以通过Uu接口向该源UE1发送能力查询信息,该能力查询信息用于查询(或请求)该UE1的能力信息。相应的,该UE1通过该Uu接口接收该能力查询信息。Exemplarily, after UE1 accesses the source gNB, the source gNB may send capability query information to the source UE1 through the Uu interface, and the capability query information is used to query (or request) the capability information of the UE1. Correspondingly, the UE1 receives the capability query information through the Uu interface.
S902b:UE1向源gNB发送UE1的能力信息。S902b: UE1 sends UE1’s capability information to the source gNB.
示例性地,UE1可以通过Uu接口向该源gNB发送该UE1的能力信息,相应的,该源gNB通过该Uu接口接收该UE1的能力信息。或者源gNB不向UE1发送能力查询信息,而是由该UE1主动向源gNB发送该UE1的能力信息。Exemplarily, UE1 may send the capability information of UE1 to the source gNB through the Uu interface, and correspondingly, the source gNB receives the capability information of UE1 through the Uu interface. Alternatively, the source gNB does not send the capability query information to UE1, but UE1 actively sends the capability information of UE1 to the source gNB.
其中,该UE1的能力信息可以包括:AI开关、存储空间、算力、支持多模型推理指示信息、剩余电量(可选的)。Among them, the capability information of UE1 may include: AI switch, storage space, computing power, support for multi-model reasoning indication information, and remaining power (optional).
本申请实施例对执行上述基站之间交互AI能力信息的步骤(即S901a-S901b),以及UE1能力信息的查询与上报的步骤(即S902a-S902b)的先后顺序不做具体限定。The embodiment of the present application does not specifically limit the order of executing the steps of exchanging AI capability information between the above-mentioned base stations (ie, S901a-S901b), and the steps of querying and reporting UE1 capability information (ie, S902a-S902b).
S903:源gNB和UE1之间协商联合训练的策略。该联合训练的策略可以包括:多模型训练指示信息、多模型联合训练模式、数据处理策略、子模型的个数、模型的类型、超参配置。S903: The source gNB and UE1 negotiate a joint training strategy. The joint training strategy may include: multi-model training indication information, multi-model joint training mode, data processing strategy, number of sub-models, model type, and hyper-parameter configuration.
其中,该多模型训练指示信息用于指示训练多模型。该多模型联合训练模式可以是一对一、或多对一、或一对多、或多对多;其中,一对一表示该源gNB和该UE1侧的多模型都作为一个整体模型,gNB的整体模型的输出作为UE1的整体模型的输入;多对一表示UE1侧的多模型作为一个整体模型,源gNB 的多模型的多个输出作为UE1的整体模型的输入;一对多表示源gNB侧的多模型作为一个整体模型,源gNB的整体模型的输出作为UE1的多模型的输入;多对多表示源gNB侧的多模型的多个输出作为UE1的多模型的输入。The multi-model training indication information is used to indicate the training of multiple models. The multi-model joint training mode can be one-to-one, many-to-one, one-to-many, or many-to-many; one-to-one means that the multiple models on the source gNB and the UE1 side are both regarded as an overall model, and the output of the gNB's overall model is used as the input of the UE1's overall model; many-to-one means that the multiple models on the UE1 side are regarded as an overall model, and the source gNB The multiple outputs of the multi-model on the source gNB side are used as the input of the overall model of UE1; one-to-many means that the multiple models on the source gNB side are used as an overall model, and the output of the overall model of the source gNB is used as the input of the multi-model of UE1; many-to-many means that the multiple outputs of the multi-model on the source gNB side are used as the input of the multi-model of UE1.
其中,数据处理策略可以是输入数据采样、特征抽样等,该输入数据采样和特征抽样具体可以参见上述实施例中的介绍,此处不再具体描述。Among them, the data processing strategy can be input data sampling, feature sampling, etc. The input data sampling and feature sampling can be specifically described in the above embodiments and will not be described in detail here.
S904a:源gNB进行多模型训练,得到该源gNB的多模型。S904a: The source gNB performs multi-model training to obtain a multi-model of the source gNB.
S904b:UE1进行多模型训练,得到该UE1的多模型。S904b: UE1 performs multi-model training to obtain a multi-model of UE1.
上述步骤S904a和S904b可以同步执行,并且该源gNB和该UE1分别进行多模型训练时,按照上述步骤S903中的联合训练的策略交互各自多模型训练的中间参数,例如,梯度或者中间推理结果。The above steps S904a and S904b can be executed synchronously, and when the source gNB and the UE1 perform multi-model training respectively, the intermediate parameters of their respective multi-model training, such as gradients or intermediate inference results, are exchanged according to the joint training strategy in the above step S903.
S905a:源gNB基于该源gNB的多模型和聚合方法进行推理,得到推理结果。S905a: The source gNB performs inference based on the multiple models and aggregation method of the source gNB to obtain an inference result.
具体的,该源gNB利用自身训练好的各子模型分别进行推理,再使用聚合方法对该各子模型的推理结果进行结合,得到最终的推理结果。Specifically, the source gNB uses its own trained sub-models to perform reasoning separately, and then uses the aggregation method to combine the reasoning results of the sub-models to obtain the final reasoning result.
S905b:UE1基于该UE1的多模型和聚合方法进行推理,得到推理结果。S905b: UE1 performs reasoning based on the multiple models and aggregation method of UE1 to obtain a reasoning result.
具体的,该UE1利用自身训练好的各子模型分别进行推理,再使用聚合方法对该各子模型的推理结果进行结合,得到最终的推理结果。Specifically, the UE1 uses each sub-model trained by itself to perform reasoning respectively, and then uses an aggregation method to combine the reasoning results of each sub-model to obtain a final reasoning result.
在一种实施方式中,可以先执行上述步骤S905b,再执行步骤S905a,即UE1利用自身训练好的多模型得到最终的推理结果之后,还将UE1侧的最终的推理结果上报给源gNB,该源gNB可以将该UE1侧的最终的推理结果作该源gNB的多模型的输入,得到该源gNB侧的最终的推理结果。In one implementation, the above step S905b may be executed first, and then step S905a may be executed, that is, after UE1 obtains the final inference result using the multi-model trained by itself, the final inference result on the UE1 side is reported to the source gNB. The source gNB may use the final inference result on the UE1 side as the input of the multi-model of the source gNB to obtain the final inference result on the source gNB side.
当源gNB确定需要切换UE1接入的基站时,则继续执行下述步骤:When the source gNB determines that the base station to which UE1 accesses needs to be switched, the following steps are continued:
S906:源gNB向UE1发送测量配置信息。相应的,该UE1接收该测量配置信息。S906: The source gNB sends measurement configuration information to UE1. Correspondingly, UE1 receives the measurement configuration information.
S907:UE1基于该测量配置信息执行测量,得到该UE1的测量报告。S907: UE1 performs measurement based on the measurement configuration information and obtains a measurement report of UE1.
S908:UE1向源gNB发送该UE1的测量报告。相应的,源gNB接收该UE1的测量报告。S908: UE1 sends a measurement report of UE1 to the source gNB. Correspondingly, the source gNB receives the measurement report of UE1.
S909:源gNB根据该UE1的测量报告和邻站的AI能力,选择目标gNB。S909: The source gNB selects the target gNB based on the measurement report of UE1 and the AI capability of the neighboring station.
该源gNB根据该UE1的测量报告和邻站的AI能力,优选具备多模型训练能力的基站作为目标gNB。The source gNB preferably selects a base station with multi-model training capability as the target gNB based on the measurement report of UE1 and the AI capability of the neighboring base station.
S910:源gNB向目标gNB发送UE1的切换请求信息。相应的,该目标gNB接收该UE1的切换请求信息,该UE切换请求信息中包含UE1的标识信息和多模型指示信息,该多模型指示信息用于请求使用多模型。S910: The source gNB sends the handover request information of UE1 to the target gNB. Correspondingly, the target gNB receives the handover request information of UE1, and the UE handover request information includes the identification information of UE1 and the multi-model indication information, and the multi-model indication information is used to request the use of the multi-model.
可选的,在该步骤S910中,源gNB还可以将自己的多模型以及源gNB的多模型的使用信息发送给目标gNB,那么该目标gNB和UE1直接复用源gNB的多模型和UE1之前训练的多模型,无需再分别进行多模型训练,即不执行下述步骤S913a和S913b。该源gNB向该目标gNB发送的该UE1的切换请求信息还包含该源gNB训练的多模型的标识、该多模型的列表、聚合方法、权重因子(可选的)。Optionally, in step S910, the source gNB may also send its own multi-model and the usage information of the multi-model of the source gNB to the target gNB, then the target gNB and UE1 directly reuse the multi-model of the source gNB and the multi-model previously trained by UE1, and there is no need to perform multi-model training separately, that is, the following steps S913a and S913b are not executed. The handover request information of the UE1 sent by the source gNB to the target gNB also includes the identifier of the multi-model trained by the source gNB, the list of the multi-model, the aggregation method, and the weight factor (optional).
可选的,该步骤S910中源gNB还可以将之前与该UE1协商好的联合训练的策略发送给目标gNB,那么该目标gNB和该UE1之间无需重复协商联合训练的策略,即不执行下述的步骤S912。Optionally, in step S910, the source gNB may also send the joint training strategy previously negotiated with the UE1 to the target gNB. In this case, there is no need to repeatedly negotiate the joint training strategy between the target gNB and the UE1, that is, the following step S912 is not executed.
S911:目标gNB与UE1完成随机接入。S911: The target gNB completes random access with UE1.
该目标gNB与该UE1进行随机接入流程,以使得该UE1成功接入该目标gNB进行通信,具体的随机接入流程参考现有的随机接入流程实现,此处不再具体描述。The target gNB performs a random access process with the UE1 so that the UE1 successfully accesses the target gNB for communication. The specific random access process is implemented with reference to the existing random access process and will not be described in detail here.
S912:目标gNB和UE1之间协商联合训练的策略。S912: The target gNB and UE1 negotiate a joint training strategy.
该联合训练的策略可以包括:多模型训练指示、多模型联合训练模式、数据处理策略、子模型个数、模型类型、超参配置。The joint training strategy may include: multi-model training instructions, multi-model joint training mode, data processing strategy, number of sub-models, model type, and hyper-parameter configuration.
该步骤912为可选的步骤,如果在上述步骤S910中,该UE1的切换请求信息中包含联合训练的策略,那么该目标gNB与该UE1之间可以不用再重新协商联合训练的策略。Step 912 is an optional step. If in the above step S910, the switching request information of the UE1 includes the joint training strategy, then the target gNB and the UE1 do not need to renegotiate the joint training strategy.
S913a:UE1进行多模型训练,得到该UE1的多模型。S913a: UE1 performs multi-model training to obtain a multi-model of UE1.
S913b:目标gNB进行多模型的训练,得到该目标gNB的多模型。S913b: The target gNB performs multi-model training to obtain a multi-model of the target gNB.
在上述中,目标gNB和UE1分别进行多模型训练时,并按照协商好的联合训练的策略交互多模型训练的中间参数,例如梯度或者中间推理结果。In the above, when the target gNB and UE1 perform multi-model training respectively, they exchange intermediate parameters of the multi-model training, such as gradients or intermediate inference results, according to the negotiated joint training strategy.
S914a:UE1基于自身的多模型的各子模型进行推理和结合,得到推理结果。S914a: UE1 performs reasoning and combination based on each sub-model of its own multi-model to obtain a reasoning result.
该UE1利用自身训练好的多模型的各子模型分别进行推理,再使用聚合方法对该各子模型的推理结果进行结合,得到最终的推理结果。 The UE1 uses each sub-model of the multi-model trained by itself to perform reasoning respectively, and then uses an aggregation method to combine the reasoning results of each sub-model to obtain a final reasoning result.
S914b:目标gNB基于自身的多模型的各子模型进行推理和结合,得到推理结果。S914b: The target gNB performs inference and combination based on each sub-model of its own multi-model to obtain an inference result.
该目标gNB利用自身训练好的多模型的各子模型(或者源gNB的多模型的各子模型)分别进行推理,再使用聚合方法对该各子模型的推理结果进行结合,得到最终的推理结果。The target gNB uses each sub-model of its own trained multi-model (or each sub-model of the source gNB's multi-model) to perform inference respectively, and then uses the aggregation method to combine the inference results of each sub-model to obtain the final inference result.
在一种实施方式中,在上述步骤S914a中,UE1利用自身训练好的多模型得到最终的推理结果之后,还将UE1侧的最终的推理结果上报给目标gNB,该目标gNB可以将该UE1侧的最终的推理结果作该目标gNB的多模型(或者源gNB的多模型)的输入,得到该目标gNB侧的最终的推理结果。In one embodiment, in the above step S914a, after UE1 obtains the final inference result using its own trained multi-model, it also reports the final inference result on the UE1 side to the target gNB. The target gNB can use the final inference result on the UE1 side as the input of the multi-model of the target gNB (or the multi-model of the source gNB) to obtain the final inference result on the target gNB side.
该实施例五支持在基站和UE联合训练双边模型场景下的多模型结合推理,可以提高模型的推理(或分析)效果。另外,当切换UE接入的基站时,优选具备多模型训练能力的基站,UE切换后仍然可以基于多模型进行结合推理。This fifth embodiment supports multi-model combined reasoning in the scenario where the base station and UE jointly train a bilateral model, which can improve the reasoning (or analysis) effect of the model. In addition, when switching the base station accessed by the UE, a base station with multi-model training capability is preferred, and the UE can still perform combined reasoning based on multiple models after switching.
实施例六:Embodiment six:
在该实施例六主要针对另一种多模型的使用场景,即模型训练功能可以将多模型封装为一个大模型,模型推理功能可以不感知该大模型(即多模型)的内部结构,并且增加模型选择和部署流程。该实施例六以通用的逻辑架构进行介绍,参考图10所示,该实施例六的具体流程如下:In this embodiment six, another multi-model usage scenario is mainly aimed at, that is, the model training function can encapsulate multiple models into a large model, the model reasoning function can be unaware of the internal structure of the large model (i.e., multiple models), and the model selection and deployment process is added. This embodiment six is introduced with a general logical architecture, as shown in Figure 10, the specific process of this embodiment six is as follows:
S1001a:模型管理功能向模型推理功能发送推理需求查询信息。S1001a: The model management function sends reasoning requirement query information to the model reasoning function.
在一种可选的实施方式中,该模型管理功能通过模型训练功能向该模型推理功能转发该推理需求查询信息。In an optional implementation, the model management function forwards the reasoning requirement query information to the model reasoning function through the model training function.
该步骤S1001a可以与上述步骤S501a或S601a相互参考。This step S1001a can be cross-referenced with the above-mentioned step S501a or S601a.
S1001b:模型推理功能向模型管理功能发送推理需求信息。S1001b: The model reasoning function sends reasoning requirement information to the model management function.
在一种可选的实施方式中,该模型推理功能通过模型训练功能向该模型管理功能转发该推理需求信息。In an optional implementation, the model reasoning function forwards the reasoning requirement information to the model management function via the model training function.
在另一种可选的实施方式中,该模型推理功能主动向该模型管理功能上报(即发送)该推理需求信息。In another optional implementation, the model reasoning function proactively reports (ie, sends) the reasoning requirement information to the model management function.
该推理需求信息中包括:推理的类型需求、推理的精度需求、推理的速度需求、推理的能耗需求。其中,推理的精度需求、推理的速度需求和推理的能耗需求也可以统一称为推理的性能需求。推理速度需求也可称为推理时延需求,表示对推理执行时间的需求,例如:单次推理执行时间小于1s;推理的能耗需求表示对推理耗能的需求,例如:单次推理消耗的能量小于5J。The reasoning requirement information includes: reasoning type requirement, reasoning accuracy requirement, reasoning speed requirement, and reasoning energy consumption requirement. Among them, the reasoning accuracy requirement, reasoning speed requirement, and reasoning energy consumption requirement can also be collectively referred to as reasoning performance requirements. The reasoning speed requirement can also be called the reasoning latency requirement, which indicates the requirement for reasoning execution time, for example: a single reasoning execution time is less than 1s; the reasoning energy consumption requirement indicates the requirement for reasoning energy consumption, for example: a single reasoning consumes less than 5J of energy.
该步骤S1001b可以参考上述步骤S501b或S601b的具体描述,此处不再赘述。The specific description of step S1001b may refer to the above step S501b or S601b, which will not be repeated here.
S1002a:模型管理功能向模型推理功能发送推理能力查询信息。S1002a: The model management function sends reasoning capability query information to the model reasoning function.
在一种可选的实施方式中,该模型管理功能通过模型训练功能向该模型推理功能转发该推理能力查询信息。In an optional implementation, the model management function forwards the reasoning capability query information to the model reasoning function via the model training function.
该步骤S1002a可以与上述步骤S502a或S602a相互参考。The step S1002a can be cross-referenced with the above-mentioned step S502a or S602a.
S1002b:模型推理功能向模型管理功能发送推理能力信息。S1002b: The model reasoning function sends reasoning capability information to the model management function.
在一种可选的实施方式中,该模型推理功能通过模型训练功能向该模型管理功能转发该推理能力信息。In an optional implementation, the model reasoning function forwards the reasoning capability information to the model management function via the model training function.
在另一种可选的实施方式中,该模型推理功能主动向该模型管理功能上报(即发送)该推理能力信息。In another optional implementation, the model reasoning function proactively reports (ie, sends) the reasoning capability information to the model management function.
该推理能力信息中包括:推理算力(可选的)、存储空间(可选的)、电量等。该推理算力指示推理功能处可用的算力信息,包括可用的硬件资源信息和硬件资源利用率。硬件资源信息可以是原始硬件信息,包括硬件类型、核数、处理频率等,也可以是量化后的运算能力。The reasoning capability information includes: reasoning computing power (optional), storage space (optional), power, etc. The reasoning computing power indicates the computing power information available at the reasoning function, including available hardware resource information and hardware resource utilization. The hardware resource information can be the original hardware information, including hardware type, number of cores, processing frequency, etc., or it can be the quantified computing power.
该步骤S1002b可以参考上述步骤S502b或S602b的具体描述,此处不再赘述。The specific description of step S1002b may refer to the above step S502b or S602b, which will not be repeated here.
S1003a:模型管理功能向模型训练功能发送训练能力查询信息。该步骤S1003a可以与上述步骤S503a或S603a相互参考。S1003a: The model management function sends training capability query information to the model training function. This step S1003a can be cross-referenced with the above-mentioned step S503a or S603a.
S1003b:模型训练功能向模型管理功能发送训练能力信息。示例性的,该模型训练功能可以主动向该模型管理功能上报该训练能力信息。该训练能力信息包括训练算力、可以达到的模型精度上限。S1003b: The model training function sends training capability information to the model management function. Exemplarily, the model training function can actively report the training capability information to the model management function. The training capability information includes training computing power and the upper limit of model accuracy that can be achieved.
该步骤S1003b可以与参考上述步骤S503b或S603b相互参考。The step S1003b may be cross-referenced with the above-mentioned step S503b or S603b.
S1004:模型管理功能向模型训练功能发送模型训练请求信息。S1004: The model management function sends model training request information to the model training function.
该模型训练请求信息中包括:模型标识/推理类型,模型训练的策略信息。 The model training request information includes: model identification/inference type, and model training strategy information.
其中,该模型的训练策略信息是根据推理需求信息、推理能力信息和训练能力信息确定;该模型的训练策略信息用于指示训练方法,其中可以包括:多个模型训练的指示信息、数据处理策略、训练算法指示等。或者该模型管理功能也可以向该模型训练功能发送原始的推理需求信息和推理能力信息,由该模型训练功能自身根据该推理需求信息和推理的能力信息,以及训练能力信息,确定模型的训练策略。The training strategy information of the model is determined based on the reasoning requirement information, the reasoning capability information and the training capability information; the training strategy information of the model is used to indicate the training method, which may include: multiple model training instruction information, data processing strategy, training algorithm instruction, etc. Alternatively, the model management function may also send the original reasoning requirement information and reasoning capability information to the model training function, and the model training function itself determines the model training strategy based on the reasoning requirement information, reasoning capability information, and training capability information.
该步骤S1004可以参考上述S504或步骤S604中的具体描述,此处不再赘述。The specific description of step S1004 may refer to the above-mentioned S504 or step S604, which will not be repeated here.
S1005:模型训练功能根据该模型训练请求信息进行模型训练,得到第一模型(即多模型)。S1005: The model training function performs model training according to the model training request information to obtain a first model (ie, a multi-model).
该步骤S1005可以与参考步骤S505或S605中的具体描述,此处不再赘述。The step S1005 may be described in detail with reference to step S505 or S605, and will not be repeated here.
S1006:模型训练功能向模型管理功能发送模型训练报告。S1006: The model training function sends a model training report to the model management function.
该模型训练报告包括:该第一模型的标识信息,模型的精度,精度约束,该第一模型的大小,该第一模型的推理算力,该第一模型的推理速度,该第一模型的推理能耗。The model training report includes: identification information of the first model, accuracy of the model, accuracy constraints, size of the first model, inference computing power of the first model, inference speed of the first model, and inference energy consumption of the first model.
该步骤S1006可以与参考步骤S506或S606中的具体描述,此处不再赘述。The step S1006 may be described in detail with reference to step S506 or S606, and will not be repeated here.
S1007:模型管理功能确定实际部署的模型。S1007: The model management function determines the model actually deployed.
即该模型管理功能可以基于原始的推理需求信息、推理能力信息以及该模型训练报告,确定实际部署的模型。That is, the model management function can determine the actually deployed model based on the original reasoning requirement information, reasoning capability information and the model training report.
例如,如果步骤S1004中模型管理功能还向模型训练功能指示训练指定的多个模型(可以为多个类似第一模型的多模型),那么在步骤S1006中,该模型训练功能反馈的模型训练报告中可以包括多个不同性能的模型,即多个类似第一模型的多模型,但各模型的性能不同;该模型管理功能可以根据推理需求信息、推理能力信息以及该模型训练报告,从该多个不同性能的模型中,确定实际部署的模型(例如第一模型)。For example, if the model management function in step S1004 also instructs the model training function to train multiple specified models (which may be multiple models similar to the first model), then in step S1006, the model training report fed back by the model training function may include multiple models with different performances, that is, multiple multiple models similar to the first model, but the performance of each model is different; the model management function can determine the actually deployed model (for example, the first model) from the multiple models with different performances based on the reasoning requirement information, reasoning capability information and the model training report.
S1008a:模型管理功能向模型推理功能发送模型部署请求信息。S1008a: The model management function sends model deployment request information to the model reasoning function.
相应的,该模型推理功能接收该模型部署请求信息,该模型部署请求信息中包括实际部署的模型信息(例如第一模型信息)。Correspondingly, the model reasoning function receives the model deployment request information, and the model deployment request information includes the actually deployed model information (such as the first model information).
示例性地,该模型部署请求信息中包括:实际部署的第一模型的标识信息。可选的,该模型部署请求信息中还可以包括该第一模型的其它信息,例如该第一模型的精度,该第一模型的精度约束,该第一模型的大小,该第一模型的推理算力,该第一模型的推理速度,该第一模型的推理能耗。Exemplarily, the model deployment request information includes: identification information of the first model actually deployed. Optionally, the model deployment request information may also include other information of the first model, such as the accuracy of the first model, the accuracy constraint of the first model, the size of the first model, the reasoning computing power of the first model, the reasoning speed of the first model, and the reasoning energy consumption of the first model.
S1008b:模型推理功能向模型管理功能发送模型部署响应信息。S1008b: The model reasoning function sends model deployment response information to the model management function.
相应的,该模型管理功能接收该模型部署响应信息,以确定(或知晓)该模型推理功能部署模型完成。Correspondingly, the model management function receives the model deployment response information to determine (or know) that the model reasoning function has completed the model deployment.
S1009:模型推理功能基于实际部署的模型进行推理,得到推理结果。S1009: The model reasoning function performs reasoning based on the actually deployed model to obtain a reasoning result.
示例性地,该模型推理功能根据第一模型信息(即实际部署的模型信息),利用该第一模型进行推理,得到推理结果。例如该模型推理功能将待推理的信息输入到该第一模型中,输出推理结果。Exemplarily, the model reasoning function uses the first model to perform reasoning based on the first model information (i.e., the model information actually deployed) to obtain a reasoning result. For example, the model reasoning function inputs the information to be reasoned into the first model and outputs a reasoning result.
在该实施例六中,模型管理功能与模型推理功能之间增加推理需求查询/上报的流程,推理需求包括推理速度需求、推理能耗需求等,该模型管理功能与模型推理功能之间还增加推理能力查询/上报的流程,推理能力包括推理算力、存储空间、电量等,并且该模型管理功能与模型训练功能之间增加训练能力查询/上报的流程。该模型训练功能向模型管理功能发送的模型训练报告中包括模型的相关信息,比如模型大小、模型推理算力、模型推理速度、模型推理能耗等。该模型管理功能可以基于推理需求、推理能力、训练能力确定模型训练策略,该模型管理功能还可以基于推理需求、推理能力,以及模型训练报告中模型的信息确定实际部署的模型。因此,该实施例六中增加推理需求查询/上报、推理能力查询/上报、训练能力查询/上报,并且可以基于推理需求、推理能力、训练能力确定训练的模型,以及基于推理需求、推理能力和实际模型的信息,确定实际部署的最佳的模型,从而可以提高模型的推理(或分析)效果。In the sixth embodiment, a process of querying/reporting reasoning requirements is added between the model management function and the model reasoning function. The reasoning requirements include reasoning speed requirements, reasoning energy consumption requirements, etc. A process of querying/reporting reasoning capabilities is also added between the model management function and the model reasoning function. The reasoning capabilities include reasoning computing power, storage space, power, etc., and a process of querying/reporting training capabilities is added between the model management function and the model training function. The model training report sent by the model training function to the model management function includes relevant information of the model, such as model size, model reasoning computing power, model reasoning speed, model reasoning energy consumption, etc. The model management function can determine the model training strategy based on the reasoning requirements, reasoning capabilities, and training capabilities. The model management function can also determine the actual deployed model based on the reasoning requirements, reasoning capabilities, and the information of the model in the model training report. Therefore, in the sixth embodiment, querying/reporting reasoning requirements, reasoning capabilities, and training capabilities are added, and the trained model can be determined based on the reasoning requirements, reasoning capabilities, and training capabilities, and the best model for actual deployment can be determined based on the information of the reasoning requirements, reasoning capabilities, and actual models, thereby improving the reasoning (or analysis) effect of the model.
实施例七:Embodiment seven:
在该实施例七的应用场景与上述实施例六的应用场景类似,区别在于实施例七中不存在模型管理功能,只涉及模型训练功能和模型推理功能之间的交互。实施例七以通用的逻辑架构来描述,具体的,可适用于部署架构3C-3E。参考图11所示,该实施例七的具体流程如下:The application scenario of this seventh embodiment is similar to that of the sixth embodiment, except that there is no model management function in the seventh embodiment, and only the interaction between the model training function and the model reasoning function is involved. The seventh embodiment is described in a general logical architecture, and specifically, it can be applied to the deployment architecture 3C-3E. Referring to FIG11 , the specific process of the seventh embodiment is as follows:
S1101:模型推理功能向模型训练功能发送模型训练请求信息。S1101: The model inference function sends model training request information to the model training function.
相应的,该模型训练功能接收该模型推理功能的模型训练请求信息,该模型训练请求信息中包括: 模型标识/推理类型,推理精度需求、推理速度需求、推理能耗需求、请求多个模型指示。Correspondingly, the model training function receives the model training request information of the model inference function, and the model training request information includes: Model identification/inference type, inference accuracy requirement, inference speed requirement, inference energy consumption requirement, and request for multiple model instructions.
其中,请求多个模型指示信息取值为是/否,以指示是否需要提供满足推理需求的多个模型,当该请求多个模型指示信息指示为是时(即指示需要提供满足推理需求的多个模型时),还可以进一步指示需要的模型的数量,例如指示需要5个模型。Among them, the value of requesting multiple models indication information is yes/no, to indicate whether it is necessary to provide multiple models that meet the reasoning requirements. When the requesting multiple models indication information indicates yes (that is, indicating that multiple models that meet the reasoning requirements need to be provided), it can further indicate the number of models required, for example, indicating that 5 models are required.
S1102:模型训练功能根据该模型推理功能的请求进行模型训练,得到第一模型(即多模型)。S1102: The model training function performs model training according to the request of the model inference function to obtain a first model (ie, a multi-model).
该步骤S1102为可选的步骤。如果该第一模型为该模型训练功能提前已经训练好的多模型,那么可以不执行该步骤S1102。This step S1102 is an optional step. If the first model is a multi-model that has been trained in advance by the model training function, then step S1102 may not be performed.
S1103:模型训练功能向模型推理功能发送该第一模型信息。S1103: The model training function sends the first model information to the model reasoning function.
示例性地,该第一模型信息中包括该第一模型的标识信息(如名称、类型),该第一模型的精度,精度约束,该第一模型的大小,该第一模型的推理算力,该第一模型的推理速度,该第一模型的推理能耗。Exemplarily, the first model information includes identification information of the first model (such as name, type), accuracy of the first model, accuracy constraint, size of the first model, inference computing power of the first model, inference speed of the first model, and inference energy consumption of the first model.
如果上述步骤S1101中指示需要多个模型(可以为多个类似第一模型的多模型)时,那么该步骤S1103中的该第一模型信息为一个列表,该列表包含多个不同性能的模型,即包含多个类似第一模型的模型信息,但各模型的性能不同。If the above step S1101 indicates that multiple models are required (which may be multiple models similar to the first model), then the first model information in step S1103 is a list containing multiple models with different performances, that is, it contains multiple model information similar to the first model, but the performance of each model is different.
S1104:模型推理功能基于推理需求信息、推理能力信息以及该第一模型信息,选择合适的模型。S1104: The model reasoning function selects a suitable model based on the reasoning requirement information, the reasoning capability information and the first model information.
如果上述步骤S1103中返回的第一模型信息为包含多个类似第一模型信息的列表,那么执行该步骤S1104。示例性地,该模型推理功能基于推理需求信息、推理能力信息以及该列表,从该多个类似第一模型的模型中,选择出一个合适的模型。If the first model information returned in step S1103 is a list containing multiple similar first model information, then step S1104 is executed. Exemplarily, the model reasoning function selects a suitable model from the multiple models similar to the first model based on the reasoning requirement information, the reasoning capability information and the list.
S1105:模型推理功能进行模型的推理,得到推理结果。S1105: The model reasoning function performs model reasoning to obtain reasoning results.
该模型推理功能可以使用第一模型进行推理,得到推理结果。或者该模型推理功能可以使用上述步骤S1104中选择的模型进行推理,得到推理结果。The model reasoning function may use the first model to perform reasoning to obtain a reasoning result, or the model reasoning function may use the model selected in the above step S1104 to perform reasoning to obtain a reasoning result.
在该实施例七中,在模型推理功能发送的模型训练请求信息(或模型请求信息)中增加推理速度需求、推理能耗需求、请求多个模型指示信息;模型训练功能可以根据推理需求确定返回的模型信息,并且在模型信息中增加模型的其他信息,例如模型的大小、模型推理算力、模型推理速度、模型推理能耗;模型推理功能可以基于推理需求、推理能力和实际模型信息,确定实际使用的模型。该实施例七中支持根据推理需求确定模型信息,并且支持基于推理需求、推理能力和实际模型信息选择最合适的模型进行推理(或分析),从而可以提高模型的推理(或分析)效果。In this embodiment seven, the inference speed requirement, inference energy consumption requirement, and request for multiple model indication information are added to the model training request information (or model request information) sent by the model inference function; the model training function can determine the returned model information according to the inference requirements, and add other information of the model to the model information, such as the size of the model, the model inference computing power, the model inference speed, and the model inference energy consumption; the model inference function can determine the model actually used based on the inference requirements, the inference ability, and the actual model information. This embodiment seven supports the determination of model information according to the inference requirements, and supports the selection of the most appropriate model for inference (or analysis) based on the inference requirements, the inference ability, and the actual model information, thereby improving the inference (or analysis) effect of the model.
下面对本申请实施例提供的通信装置进行描述。The communication device provided in the embodiment of the present application is described below.
基于同一技术构思,本申请实施例提供一种通信装置,该通信装置可以用于执行上述方法实施例中由第一通信装置所执行的操作。该通信装置还可以为第一通信装置、第一通信装置的处理器、或芯片。该装置包括执行上述实施例中第一通信装置所描述的方法/操作/步骤/动作所一一对应的模块或单元,该模块或单元可以是硬件电路,也可是软件,也可以是硬件电路结合软件实现。该通信装置可以具有如图12所示的结构。Based on the same technical concept, an embodiment of the present application provides a communication device, which can be used to perform the operation performed by the first communication device in the above method embodiment. The communication device can also be a first communication device, a processor of the first communication device, or a chip. The device includes a module or unit corresponding to the method/operation/step/action described by the first communication device in the above embodiment, and the module or unit can be a hardware circuit, or software, or a hardware circuit combined with software. The communication device can have a structure as shown in Figure 12.
如图12所示,该通信装置1200可以包括通信单元1201(也可以称为收发单元)和处理单元1202,该通信单元1201相当于通信模块(或收发模块),该处理单元1202相当于处理模块,所述处理单元1202可以用于调用所述通信单元1201执行接收和/或发送的功能,所述通信单元1201可以实现相应的通信功能,具体的,所述通信单元1201可以包括接收单元和/或发送单元,接收单元可以用于接收信息和/或数据等,发送单元可以用于发送信息和/或数据。通信单元1201还可以称为通信接口或收发模块。As shown in FIG. 12 , the communication device 1200 may include a communication unit 1201 (also referred to as a transceiver unit) and a processing unit 1202. The communication unit 1201 is equivalent to a communication module (or a transceiver module), and the processing unit 1202 is equivalent to a processing module. The processing unit 1202 may be used to call the communication unit 1201 to perform a receiving and/or sending function, and the communication unit 1201 may implement a corresponding communication function. Specifically, the communication unit 1201 may include a receiving unit and/or a sending unit. The receiving unit may be used to receive information and/or data, and the sending unit may be used to send information and/or data. The communication unit 1201 may also be referred to as a communication interface or a transceiver module.
可选地,该通信装置1200还可以包括存储单元1203,存储单元1203相当于存储模块,可以用于存储指令和/或数据,处理单元1202可以读取存储模块中的指令和/或数据,以使得通信装置实现前述方法实施例。Optionally, the communication device 1200 may further include a storage unit 1203, which is equivalent to a storage module and can be used to store instructions and/or data. The processing unit 1202 can read the instructions and/or data in the storage module so that the communication device implements the aforementioned method embodiment.
该通信装置1200可以用于执行上文方法实施例中第一通信装置所执行的动作。该通信装置1200可以为第一通信装置或者可配置于第一通信装置的部件。通信单元1201用于执行上文方法实施例中第一通信装置侧的发送相关的操作,处理单元1202用于执行上文方法实施例中第一通信装置侧的处理相关的操作。The communication device 1200 can be used to perform the actions performed by the first communication device in the above method embodiment. The communication device 1200 can be the first communication device or a component that can be configured in the first communication device. The communication unit 1201 is used to perform the sending-related operations on the first communication device side in the above method embodiment, and the processing unit 1202 is used to perform the processing-related operations on the first communication device side in the above method embodiment.
可选地,通信单元1201可以包括发送单元和接收单元。发送单元用于执行上述方法实施例中的发送操作。接收单元用于执行上述方法实施例中的接收操作。 Optionally, the communication unit 1201 may include a sending unit and a receiving unit. The sending unit is used to perform the sending operation in the above method embodiment. The receiving unit is used to perform the receiving operation in the above method embodiment.
需要说明的是,通信装置1200可以包括发送单元,而不包括接收单元。或者,通信装置1200可以包括接收单元,而不包括发送单元。具体可以视通信装置1200执行的上述方案中是否包括发送动作和接收动作。It should be noted that the communication device 1200 may include a sending unit but not a receiving unit. Alternatively, the communication device 1200 may include a receiving unit but not a sending unit. Specifically, it may depend on whether the above solution executed by the communication device 1200 includes a sending action and a receiving action.
作为一种示例,该通信装置1200用于执行上文图4A或图4B所示的实施例中第一通信装置所执行的动作。As an example, the communication device 1200 is used to execute the actions executed by the first communication device in the embodiment shown in FIG. 4A or FIG. 4B above.
例如,所述通信单元1201,用于接收模型请求信息,所述模型请求信息中包括推理需求信息;所述处理单元1202,用于根据所述模型请求信息,确定第一模型,所述第一模型为多模型;所述通信单元1201,还用于发送第一信息,所述第一信息中包括所述第一模型的信息。For example, the communication unit 1201 is used to receive model request information, and the model request information includes reasoning requirement information; the processing unit 1202 is used to determine the first model according to the model request information, and the first model is a multi-model; the communication unit 1201 is also used to send first information, and the first information includes information of the first model.
应理解,各单元执行上述相应过程的具体过程在上述方法实施例中已经详细说明,为了简洁,在此不再赘述。It should be understood that the specific process of each unit executing the above corresponding process has been described in detail in the above method embodiment, and for the sake of brevity, it will not be repeated here.
上文实施例中的处理单元1202可以由至少一个处理器或处理器相关电路实现。通信单元1201可以由收发器或收发器相关电路实现。存储单元可以通过至少一个存储器实现。The processing unit 1202 in the above embodiment may be implemented by at least one processor or processor-related circuits. The communication unit 1201 may be implemented by a transceiver or transceiver-related circuits. The storage unit may be implemented by at least one memory.
基于同一技术构思,本申请实施例提供一种通信装置,该通信装置可以用于执行上述方法实施例中由第二通信装置所执行的操作。该通信装置还可以为第二通信装置、第二通信装置的处理器、或芯片。该装置包括执行上述实施例中第二通信装置所描述的方法/操作/步骤/动作所一一对应的模块或单元,该模块或单元可以是硬件电路,也可是软件,也可以是硬件电路结合软件实现。该通信装置也可以具有如图12所示的结构。Based on the same technical concept, an embodiment of the present application provides a communication device, which can be used to perform the operation performed by the second communication device in the above method embodiment. The communication device can also be a second communication device, a processor of the second communication device, or a chip. The device includes a module or unit corresponding to the method/operation/step/action described by the second communication device in the above embodiment, and the module or unit can be a hardware circuit, or software, or a hardware circuit combined with software. The communication device can also have a structure as shown in Figure 12.
如图12所示,该通信装置1200可以包括通信单元1201(也可以称为收发单元)和处理单元1202,该通信单元1201相当于通信模块(或收发模块),该处理单元1202相当于处理模块,所述处理单元1202可以用于调用所述通信单元1201执行接收和/或发送的功能,所述通信单元1201可以实现相应的通信功能,具体的,所述通信单元1201具体可以包括接收单元和/或发送单元,接收单元可以用于接收信息和/或数据等,发送单元可以用于发送信息和/或数据。通信单元1201还可以称为通信接口或收发模块。As shown in FIG. 12 , the communication device 1200 may include a communication unit 1201 (also referred to as a transceiver unit) and a processing unit 1202. The communication unit 1201 is equivalent to a communication module (or a transceiver module), and the processing unit 1202 is equivalent to a processing module. The processing unit 1202 may be used to call the communication unit 1201 to perform a receiving and/or sending function, and the communication unit 1201 may implement a corresponding communication function. Specifically, the communication unit 1201 may include a receiving unit and/or a sending unit. The receiving unit may be used to receive information and/or data, and the sending unit may be used to send information and/or data. The communication unit 1201 may also be referred to as a communication interface or a transceiver module.
可选地,该通信装置1200还可以包括存储单元1203,存储单元1203相当于存储模块,可以用于存储指令和/或数据,处理单元1202可以读取存储模块中的指令和/或数据,以使得通信装置实现前述方法实施例。Optionally, the communication device 1200 may further include a storage unit 1203, which is equivalent to a storage module and can be used to store instructions and/or data. The processing unit 1202 can read the instructions and/or data in the storage module so that the communication device implements the aforementioned method embodiment.
该通信装置1200可以用于执行上文方法实施例中第二通信装置所执行的动作。该通信装置1200可以为第一通信装置或者可配置于第二通信装置的部件。通信单元1201用于执行上文方法实施例中第二通信装置侧的发送相关的操作,处理单元1202用于执行上文方法实施例中第二通信装置侧的处理相关的操作。The communication device 1200 can be used to perform the actions performed by the second communication device in the above method embodiment. The communication device 1200 can be a first communication device or a component that can be configured in the second communication device. The communication unit 1201 is used to perform the sending-related operations on the second communication device side in the above method embodiment, and the processing unit 1202 is used to perform the processing-related operations on the second communication device side in the above method embodiment.
可选地,通信单元1201可以包括发送单元和接收单元。发送单元用于执行上述方法实施例中的发送操作。接收单元用于执行上述方法实施例中的接收操作。Optionally, the communication unit 1201 may include a sending unit and a receiving unit. The sending unit is used to perform the sending operation in the above method embodiment. The receiving unit is used to perform the receiving operation in the above method embodiment.
需要说明的是,通信装置1200可以包括发送单元,而不包括接收单元。或者,通信装置1200可以包括接收单元,而不包括发送单元。具体可以视通信装置1200执行的上述方案中是否包括发送动作和接收动作。It should be noted that the communication device 1200 may include a sending unit but not a receiving unit. Alternatively, the communication device 1200 may include a receiving unit but not a sending unit. Specifically, it may depend on whether the above solution executed by the communication device 1200 includes a sending action and a receiving action.
作为一种示例,该通信装置1200用于执行上文图4A或图4B所示的实施例中第二通信装置所执行的动作。As an example, the communication device 1200 is used to execute the actions executed by the second communication device in the embodiment shown in FIG. 4A or FIG. 4B above.
例如,所述通信单元1201,用于接收第二信息,所述第二信息中包括第一模型的信息,所述第一模型是根据推理需求信息确定的,且所述第一模型为多模型;所述处理单元1202,用于基于所述第一模型的信息,得到所述第一模型的推理信息。For example, the communication unit 1201 is used to receive second information, which includes information of a first model, where the first model is determined based on reasoning requirement information, and the first model is a multi-model; the processing unit 1202 is used to obtain reasoning information of the first model based on the information of the first model.
应理解,各单元执行上述相应过程的具体过程在上述方法实施例中已经详细说明,为了简洁,在此不再赘述。It should be understood that the specific process of each unit executing the above corresponding process has been described in detail in the above method embodiment, and for the sake of brevity, it will not be repeated here.
上文实施例中的处理单元1202可以由至少一个处理器或处理器相关电路实现。通信单元1201可以由收发器或收发器相关电路实现。存储单元可以通过至少一个存储器实现。The processing unit 1202 in the above embodiment may be implemented by at least one processor or processor-related circuits. The communication unit 1201 may be implemented by a transceiver or transceiver-related circuits. The storage unit may be implemented by at least one memory.
基于同一技术构思,本申请实施例提供一种通信装置,该通信装置可以用于执行上述方法实施例中由第三通信装置所执行的操作。该通信装置还可以为第三通信装置、第三通信装置的处理器、或芯片。该装置包括执行上述实施例中第三通信装置所描述的方法/操作/步骤/动作所一一对应的模块或单元,该模块或单元可以是硬件电路,也可是软件,也可以是硬件电路结合软件实现。该通信装置也可以具有如 图12所示的结构。Based on the same technical concept, an embodiment of the present application provides a communication device, which can be used to perform the operations performed by the third communication device in the above method embodiment. The communication device can also be a third communication device, a processor of the third communication device, or a chip. The device includes a module or unit corresponding to the method/operation/step/action described by the third communication device in the above embodiment. The module or unit can be a hardware circuit, or software, or a combination of a hardware circuit and software. The communication device can also have the following The structure shown in Figure 12.
如图12所示,该通信装置1200可包括处理单元1202,可选的,还包括通信单元1201,该通信单元1201相当于收发模块,该处理单元1202相当于处理模块,所述处理单元1202可以用于调用所述通信单元1201执行接收和/或发送的功能,所述通信单元1201可以实现相应的通信功能,具体的,所述通信单元1201具体可以包括接收单元和/或发送单元,接收单元可以用于接收信息和/或数据等,发送单元可以用于发送信息和/或数据。通信单元1201还可以称为通信接口或收发模块。As shown in FIG. 12 , the communication device 1200 may include a processing unit 1202, and optionally, a communication unit 1201. The communication unit 1201 is equivalent to a transceiver module, and the processing unit 1202 is equivalent to a processing module. The processing unit 1202 may be used to call the communication unit 1201 to perform a receiving and/or sending function, and the communication unit 1201 may implement a corresponding communication function. Specifically, the communication unit 1201 may include a receiving unit and/or a sending unit. The receiving unit may be used to receive information and/or data, and the sending unit may be used to send information and/or data. The communication unit 1201 may also be called a communication interface or a transceiver module.
可选地,该通信装置1200还可以包括存储单元1203,存储单元1203相当于存储模块,可以用于存储指令和/或数据,处理单元1202可以读取存储模块中的指令和/或数据,以使得通信装置实现前述方法实施例。Optionally, the communication device 1200 may further include a storage unit 1203, which is equivalent to a storage module and can be used to store instructions and/or data. The processing unit 1202 can read the instructions and/or data in the storage module so that the communication device implements the aforementioned method embodiment.
该通信装置1200可以用于执行上文方法实施例中第三通信装置所执行的动作。该通信装置1200可以为第三通信装置或者可配置于第三通信装置的部件。通信单元1201用于执行上文方法实施例中第三通信装置侧的发送相关的操作,处理单元1202用于执行上文方法实施例中第三通信装置侧的处理相关的操作。The communication device 1200 may be used to perform the actions performed by the third communication device in the above method embodiment. The communication device 1200 may be a third communication device or a component that may be configured in a third communication device. The communication unit 1201 is used to perform the sending-related operations on the third communication device side in the above method embodiment, and the processing unit 1202 is used to perform the processing-related operations on the third communication device side in the above method embodiment.
可选地,通信单元1201可以包括发送单元和接收单元。发送单元用于执行上述方法实施例中的发送操作。接收单元用于执行上述方法实施例中的接收操作。Optionally, the communication unit 1201 may include a sending unit and a receiving unit. The sending unit is used to perform the sending operation in the above method embodiment. The receiving unit is used to perform the receiving operation in the above method embodiment.
需要说明的是,通信装置1200可以包括发送单元,而不包括接收单元。或者,通信装置1200可以包括接收单元,而不包括发送单元。具体可以视通信装置1200执行的上述方案中是否包括发送动作和接收动作。It should be noted that the communication device 1200 may include a sending unit but not a receiving unit. Alternatively, the communication device 1200 may include a receiving unit but not a sending unit. Specifically, it may depend on whether the above solution executed by the communication device 1200 includes a sending action and a receiving action.
作为一种示例,该通信装置1200用于执行上文图4A所示的实施例中第三通信装置所执行的动作。As an example, the communication device 1200 is used to execute the actions executed by the third communication device in the embodiment shown in FIG. 4A above.
例如,所述通信单元1201,用于接收第一通信装置的训练能力指示信息;所述训练能力指示信息用于指示所述第一通信装置支持多模型的训练;以及接收第二通信装置的推理需求信息和推理能力信息;所述推理能力信息中包括推理能力指示信息,所述推理能力指示信息用于指示所述第二通信装置支持多模型的推理;For example, the communication unit 1201 is used to receive training capability indication information of a first communication device; the training capability indication information is used to indicate that the first communication device supports multi-model training; and receive reasoning requirement information and reasoning capability information of a second communication device; the reasoning capability information includes reasoning capability indication information, and the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning;
所述通信单元1201,还用于向所述第一通信装置发送模型请求信息,所述模型请求信息中包括所述推理需求信息;以及从所述第一通信装置接收第一信息,所述第一信息中包括第一模型的信息,所述第一模型为多模型,所述第一模型是根据所述推理需求信息确定的;The communication unit 1201 is further configured to send model request information to the first communication device, wherein the model request information includes the reasoning requirement information; and receive first information from the first communication device, wherein the first information includes information of a first model, wherein the first model is a multi-model, and the first model is determined according to the reasoning requirement information;
所述通信单元1201,还用于向所述第二通信装置发送第二信息,所述第二信息中包括所述第一模型的信息。The communication unit 1201 is further configured to send second information to the second communication device, where the second information includes information of the first model.
应理解,各单元执行上述相应过程的具体过程在上述方法实施例中已经详细说明,为了简洁,在此不再赘述。It should be understood that the specific process of each unit executing the above corresponding process has been described in detail in the above method embodiment, and for the sake of brevity, it will not be repeated here.
上文实施例中的处理单元1202可以由至少一个处理器或处理器相关电路实现。通信单元1201可以由收发器或收发器相关电路实现。存储单元可以通过至少一个存储器实现。The processing unit 1202 in the above embodiment may be implemented by at least one processor or processor-related circuits. The communication unit 1201 may be implemented by a transceiver or transceiver-related circuits. The storage unit may be implemented by at least one memory.
基于同一技术构思,本申请实施例还提供了一种通信装置,如图13所示,为本申请提供的一种通信装置示意图,该通信装置1300可以是上述实施例中的第一通信装置、第一通信装置的处理器、或芯片,该通信装置1300可以用于执行上述方法实施例中由第一通信装置所执行的操作。该通信装置1300包括:处理器1302。可选的,该通信装置1300还可以包括通信接口1301、存储器1303、通信总线1304。其中,通信接口1301、处理器1302,以及存储器1303可以通过通信总线1304相互连接;通信总线1304可以是外设部件互连标准(peripheral component interconnect,简称PCI)总线或扩展工业标准结构(extended industry standard architecture,简称EISA)总线等。所述通信总线1304可以分为地址总线、数据总线、控制总线等。为便于表示,图13中仅用一条粗线表示,但并不表示仅有一根总线或一种类型的总线。Based on the same technical concept, the embodiment of the present application also provides a communication device, as shown in FIG13, which is a schematic diagram of a communication device provided by the present application. The communication device 1300 can be the first communication device, the processor of the first communication device, or the chip in the above embodiment. The communication device 1300 can be used to perform the operation performed by the first communication device in the above method embodiment. The communication device 1300 includes: a processor 1302. Optionally, the communication device 1300 can also include a communication interface 1301, a memory 1303, and a communication bus 1304. Among them, the communication interface 1301, the processor 1302, and the memory 1303 can be connected to each other through the communication bus 1304; the communication bus 1304 can be a peripheral component interconnect standard (PCI) bus or an extended industry standard architecture (EISA) bus, etc. The communication bus 1304 can be divided into an address bus, a data bus, a control bus, etc. For ease of representation, FIG13 shows only one thick line, but this does not mean that there is only one bus or one type of bus.
处理器1302可以是一个CPU,微处理器,ASIC,或一个或多个用于控制本申请方案程序执行的集成电路。Processor 1302 may be a CPU, a microprocessor, an ASIC, or one or more integrated circuits for controlling the execution of the program of the present application.
通信接口1301,使用任何收发器一类的装置,用于与其他设备或通信网络通信,如以太网,无线接入网(radio access network,RAN),无线局域网(wireless local area networks,WLAN),有线接入网等。The communication interface 1301 uses any transceiver-like device to communicate with other devices or communication networks, such as Ethernet, radio access network (RAN), wireless local area networks (WLAN), wired access networks, etc.
存储器1303可以是ROM或可存储静态信息和指令的其他类型的静态存储设备,RAM或者可存储信息和指令的其他类型的动态存储设备,也可以是电可擦可编程只读存储器(electrically erasable  programmable read-only memory,EEPROM)、只读光盘(compact disc read-only memory,CD-ROM)或其他光盘存储、光碟存储(包括压缩光碟、激光碟、光碟、数字通用光碟、蓝光光碟等)、磁盘存储介质或者其他磁存储设备、或者能够用于携带或存储具有指令或数据结构形式的期望的程序代码并能够由计算机存取的任何其他介质,但不限于此。存储器可以是独立存在,通过通信总线1304与处理器相连接。存储器也可以和处理器集成在一起。The memory 1303 may be a ROM or other type of static storage device that can store static information and instructions, a RAM or other type of dynamic storage device that can store information and instructions, or an electrically erasable programmable read-only memory (EPROM). Programmable read-only memory (EEPROM), compact disc read-only memory (CD-ROM) or other optical disk storage, optical disk storage (including compressed optical disk, laser disk, optical disk, digital versatile disk, Blu-ray disk, etc.), magnetic disk storage medium or other magnetic storage device, or any other medium that can be used to carry or store the desired program code in the form of instructions or data structures and can be accessed by the computer, but is not limited to this. The memory can be independent and connected to the processor through the communication bus 1304. The memory can also be integrated with the processor.
其中,存储器1303用于存储执行本申请方案的计算机执行指令,并由处理器1302来控制执行。处理器1302用于执行存储器1303中存储的计算机执行指令,从而实现本申请上述实施例提供的通信方法。The memory 1303 is used to store computer-executable instructions for executing the solution of the present application, and the execution is controlled by the processor 1302. The processor 1302 is used to execute the computer-executable instructions stored in the memory 1303, thereby realizing the communication method provided in the above embodiment of the present application.
可选的,本申请实施例中的计算机执行指令也可以称之为应用程序代码,本申请实施例对此不作具体限定。Optionally, the computer-executable instructions in the embodiments of the present application may also be referred to as application code, which is not specifically limited in the embodiments of the present application.
图14为本申请实施例提供的一种芯片的装置结构示意图。该芯片1400包括接口电路1401和一个或多个处理器1402。可选的,所述芯片1400还可以包含总线。其中:处理器1402可能是一种集成电路芯片,具有信号的处理能力。在实现过程中,上述眼球跟踪方法的各步骤可以通过处理器1402中的硬件的集成逻辑电路或者软件形式的指令完成。上述的处理器1402可以是通用处理器、数字通信器(DSP)、专用集成电路(ASIC)、现场可编程门阵列(FPGA)或者其它可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件。可以实现或者执行本申请实施例中的公开的各方法、步骤。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。FIG14 is a schematic diagram of the device structure of a chip provided in an embodiment of the present application. The chip 1400 includes an interface circuit 1401 and one or more processors 1402. Optionally, the chip 1400 may also include a bus. Wherein: the processor 1402 may be an integrated circuit chip with signal processing capabilities. In the implementation process, each step of the above-mentioned eye tracking method can be completed by an integrated logic circuit of hardware in the processor 1402 or instructions in the form of software. The above-mentioned processor 1402 may be a general-purpose processor, a digital communicator (DSP), an application-specific integrated circuit (ASIC), a field programmable gate array (FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components. The various methods and steps disclosed in the embodiments of the present application can be implemented or executed. The general-purpose processor may be a microprocessor or the processor may also be any conventional processor, etc.
接口电路1401可以用于数据、指令或者信息的发送或者接收,处理器1402可以利用接口电路1401接收的数据、指令或者其它信息,进行加工,可以将加工完成信息通过接口电路1401发送出去。The interface circuit 1401 can be used to send or receive data, instructions or information. The processor 1402 can use the data, instructions or other information received by the interface circuit 1401 to process, and can send the processing completion information through the interface circuit 1401.
可选的,芯片还包括存储器1403,存储器1403可以包括只读存储器和随机存取存储器,并向处理器提供操作指令和数据。存储器1403的一部分还可以包括非易失性随机存取存储器(NVRAM)。Optionally, the chip further includes a memory 1403, which may include a read-only memory and a random access memory, and provides operation instructions and data to the processor. A portion of the memory 1403 may also include a non-volatile random access memory (NVRAM).
可选的,存储器存储了可执行软件模块或者数据结构,处理器可以通过调用存储器存储的操作指令(该操作指令可存储在操作系统中),执行相应的操作。Optionally, the memory stores executable software modules or data structures, and the processor can perform corresponding operations by calling operation instructions stored in the memory (the operation instructions can be stored in the operating system).
可选的,芯片可以使用在本申请实施例涉及的第一通信装置(第二通信装置、第三通信装置)中。可选的,接口电路1401可用于输出处理器1402的执行结果。关于本申请的一个或多个实施例提供的通信方法可参考前述各个实施例,这里不再赘述。Optionally, the chip can be used in the first communication device (second communication device, third communication device) involved in the embodiment of the present application. Optionally, the interface circuit 1401 can be used to output the execution result of the processor 1402. The communication method provided by one or more embodiments of the present application can refer to the aforementioned embodiments, which will not be repeated here.
需要说明的,接口电路1401、处理器1402各自对应的功能既可以通过硬件设计实现,也可以通过软件设计来实现,还可以通过软硬件结合的方式来实现,这里不作限制。It should be noted that the corresponding functions of the interface circuit 1401 and the processor 1402 can be implemented through hardware design, software design, or a combination of hardware and software, and there is no limitation here.
本申请实施例还提供一种计算机可读存储介质,其上存储有用于实现上述方法实施例中由第一通信装置执行的方法的计算机指令,和/或其上存储有用于实现上述方法实施例中由第二通信装置执行的方法的计算机指令,和/或其上存储有用于实现上述方法实施例中由第三通信装置执行的方法的计算机指令。An embodiment of the present application also provides a computer-readable storage medium, on which computer instructions for implementing the method executed by the first communication device in the above method embodiment are stored, and/or computer instructions for implementing the method executed by the second communication device in the above method embodiment are stored, and/or computer instructions for implementing the method executed by the third communication device in the above method embodiment are stored.
例如,该计算机程序被计算机执行时,使得该计算机可以实现上述方法实施例中由第一通信装置执行的方法。For example, when the computer program is executed by a computer, the computer can implement the method performed by the first communication device in the above method embodiment.
本申请实施例还提供一种包含指令的计算机程序产品,该指令被计算机执行时使得该计算机实现上述方法实施例中由第一通信装置执行的方法,和/或该指令被计算机执行时使得该计算机实现上述方法实施例中由第二通信装置执行的方法,和/或该指令被计算机执行时使得该计算机实现上述方法实施例中由第三通信装置执行的方法。An embodiment of the present application also provides a computer program product comprising instructions, which, when executed by a computer, enables the computer to implement the method performed by the first communication device in the above method embodiment, and/or when executed by a computer, enables the computer to implement the method performed by the second communication device in the above method embodiment, and/or when executed by a computer, enables the computer to implement the method performed by the third communication device in the above method embodiment.
本申请实施例还提供一种芯片装置,包括处理器,用于调用该存储器中存储的计算机程度或计算机指令,以使得该处理器执行上述图4A或图4B所示的实施例的一种通信方法。An embodiment of the present application also provides a chip device, including a processor, for calling a computer program or computer instruction stored in the memory so that the processor executes a communication method of the embodiment shown in FIG. 4A or FIG. 4B above.
一种可能的实现方式中,该芯片装置的输入对应上述图4A或图4B所示的实施例中的接收操作,该芯片装置的输出对应上述图4A或图4B所示的实施例中的发送操作。In a possible implementation, the input of the chip device corresponds to the receiving operation in the embodiment shown in FIG. 4A or FIG. 4B , and the output of the chip device corresponds to the sending operation in the embodiment shown in FIG. 4A or FIG. 4B .
可选地,该处理器通过接口与存储器耦合。Optionally, the processor is coupled to the memory via an interface.
可选地,该芯片装置还包括存储器,该存储器中存储有计算机程度或计算机指令。Optionally, the chip device further comprises a memory, in which computer programs or computer instructions are stored.
其中,上述任一处提到的处理器,可以是一个通用中央处理器,微处理器,特定应用集成电路(application-specific integrated circuit,ASIC),或一个或多个用于控制上述图4A或图4B所示的实施例的一种通信方法的程序执行的集成电路。上述任一处提到的存储器可以为只读存储器(read-only memory,ROM)或可存储静态信息和指令的其他类型的静态存储设备,随机存取存储器(random access memory,RAM)等。 The processor mentioned in any of the above places may be a general-purpose central processing unit, a microprocessor, an application-specific integrated circuit (ASIC), or one or more integrated circuits for controlling the execution of a program of a communication method of the embodiment shown in FIG. 4A or FIG. 4B. The memory mentioned in any of the above places may be a read-only memory (ROM) or other types of static storage devices that can store static information and instructions, a random access memory (RAM), etc.
需要注意的是,为描述方便和简洁,上述提供的任一种通信装置中相关内容的解释及有益效果均可参考上文提供的对应的眼球跟踪方法实施例,此处不再赘述。It should be noted that, for the sake of convenience and brevity of description, the explanation of the relevant contents and beneficial effects in any of the communication devices provided above can refer to the corresponding eye tracking method embodiments provided above, and will not be repeated here.
本申请中,通信装置之间还可以包括硬件层、运行在硬件层之上的操作系统层,以及运行在操作系统层上的应用层。其中,硬件层可以包括中央处理器(central processing unit,CPU)、内存管理模块(memory management unit,MMU)和内存(也称为主存)等硬件。操作系统层的操作系统可以是任意一种或多种通过进程(process)实现业务处理的计算机操作系统,例如,Linux操作系统、Unix操作系统、Android操作系统、iOS操作系统或windows操作系统等。应用层可以包含浏览器、通讯录、文字处理软件、即时通信软件等应用。In the present application, the communication devices may also include a hardware layer, an operating system layer running on the hardware layer, and an application layer running on the operating system layer. Among them, the hardware layer may include hardware such as a central processing unit (CPU), a memory management unit (MMU), and a memory (also called main memory). The operating system of the operating system layer may be any one or more computer operating systems that implement business processing through processes, such as Linux operating system, Unix operating system, Android operating system, iOS operating system, or Windows operating system. The application layer may include applications such as browsers, address books, word processing software, and instant messaging software.
本申请实施例中对模块的划分是示意性的,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,另外,在本申请各个实施例中的各功能模块可以集成在一个处理器中,也可以是单独物理存在,也可以两个或两个以上模块集成在一个模块中。上述集成的模块既可以采用硬件的形式实现,也可以采用软件功能模块的形式实现。The division of modules in the embodiments of the present application is schematic and is only a logical function division. There may be other division methods in actual implementation. In addition, each functional module in each embodiment of the present application may be integrated into a processor, or may exist physically separately, or two or more modules may be integrated into one module. The above-mentioned integrated modules may be implemented in the form of hardware or in the form of software functional modules.
通过以上的实施方式的描述,所属领域的技术人员可以清楚地了解到本申请实施例可以用硬件实现,或固件实现,或它们的组合方式来实现。当使用软件实现时,可以将上述功能存储在计算机可读介质中或作为计算机可读介质上的一个或多个指令或代码进行传输。计算机可读介质包括计算机存储介质和通信介质,其中通信介质包括便于从一个地方向另一个地方传送计算机程序的任何介质。存储介质可以是计算机能够存取的任何可用介质。以此为例但不限于:计算机可读介质可以包括RAM、ROM、电可擦可编程只读存储器(electrically erasable programmable read only memory,EEPROM)、只读光盘(compact disc read-Only memory,CD-ROM)或其他光盘存储、磁盘存储介质或者其他磁存储设备、或者能够用于携带或存储具有指令或数据结构形式的期望的程序代码并能够由计算机存取的任何其他介质。此外。任何连接可以适当的成为计算机可读介质。例如,如果软件是使用同轴电缆、光纤光缆、双绞线、数字用户线(digital subscriber line,DSL)或者诸如红外线、无线电和微波之类的无线技术从网站、服务器或者其他远程源传输的,那么同轴电缆、光纤光缆、双绞线、DSL或者诸如红外线、无线和微波之类的无线技术包括在所属介质的定影中。如本申请实施例所使用的,盘(disk)和碟(disc)包括压缩光碟(compact disc,CD)、激光碟、光碟、数字通用光碟(digital video disc,DVD)、软盘和蓝光光碟,其中盘通常磁性的复制数据,而碟则用激光来光学的复制数据。上面的组合也应当包括在计算机可读介质的保护范围之内。Through the description of the above implementation mode, it can be clearly understood by those skilled in the art that the embodiments of the present application can be implemented by hardware, firmware, or a combination thereof. When software is used for implementation, the above functions can be stored in a computer-readable medium or transmitted as one or more instructions or codes on a computer-readable medium. Computer-readable media include computer storage media and communication media, wherein the communication media include any medium that facilitates the transmission of a computer program from one place to another. The storage medium can be any available medium that a computer can access. Taking this as an example but not limited to: a computer-readable medium may include RAM, ROM, electrically erasable programmable read-only memory (EEPROM), compact disc read-only memory (CD-ROM) or other optical disc storage, disk storage medium or other magnetic storage device, or any other medium that can be used to carry or store the desired program code in the form of an instruction or data structure and can be accessed by a computer. In addition. Any connection can be appropriately a computer-readable medium. For example, if the software is transmitted from a website, server, or other remote source using coaxial cable, fiber optic cable, twisted pair, digital subscriber line (DSL), or wireless technologies such as infrared, radio, and microwave, then the coaxial cable, fiber optic cable, twisted pair, DSL, or wireless technologies such as infrared, wireless, and microwave are included in the fixation of the medium. As used in the embodiments of the present application, disk and disc include compact disc (CD), laser disc, optical disc, digital video disc (DVD), floppy disk, and Blu-ray disc, where disks usually copy data magnetically and discs use lasers to copy data optically. The above combinations should also be included in the scope of protection of computer-readable media.
总之,以上所述仅为本申请的实施例而已,并非用于限定本申请的保护范围。凡根据本申请的揭露,所作的任何修改、等同替换、改进等,均应包含在本申请的保护范围之内。 In short, the above description is only an embodiment of the present application and is not intended to limit the protection scope of the present application. Any modification, equivalent replacement, improvement, etc. made according to the disclosure of the present application shall be included in the protection scope of the present application.

Claims (26)

  1. 一种通信方法,其特征在于,包括:A communication method, comprising:
    第一通信装置接收模型请求信息,所述模型请求信息中包括推理需求信息;The first communication device receives model request information, wherein the model request information includes reasoning requirement information;
    所述第一通信装置根据所述模型请求信息,确定第一模型,所述第一模型为多模型;The first communication device determines a first model according to the model request information, wherein the first model is a multi-model;
    所述第一通信装置发送第一信息,所述第一信息中包括所述第一模型的信息。The first communication device sends first information, where the first information includes information of the first model.
  2. 根据权利要求1所述的方法,其特征在于,所述第一通信装置接收模型请求信息之前,所述方法还包括:The method according to claim 1, characterized in that before the first communication device receives the model request information, the method further comprises:
    所述第一通信装置发送训练能力指示信息,所述训练能力指示信息用于指示所述第一通信装置支持多模型的训练。The first communication device sends training capability indication information, where the training capability indication information is used to indicate that the first communication device supports multi-model training.
  3. 根据权利要求1所述的方法,其特征在于,所述模型请求信息用于请求训练多模型时,所述模型请求信息中还包括多模型的训练策略;所述第一通信装置根据所述模型请求信息,确定第一模型,包括:所述第一通信装置根据所述推理需求信息和所述多模型的训练策略进行训练,得到所述第一模型的多个子模型;或者The method according to claim 1 is characterized in that when the model request information is used to request training of multiple models, the model request information also includes a training strategy for the multiple models; the first communication device determines the first model according to the model request information, comprising: the first communication device performs training according to the inference requirement information and the training strategy for the multiple models to obtain multiple sub-models of the first model; or
    所述模型请求信息用于请求训练多模型时;所述第一通信装置根据所述模型请求信息,确定第一模型,包括:所述第一通信装置根据所述推理需求信息,确定多模型的训练策略;以及根据所述推理需求信息和所述多模型的训练策略进行训练,得到所述第一模型的多个子模型;When the model request information is used to request training of multiple models; the first communication device determines the first model according to the model request information, including: the first communication device determines the training strategy of the multiple models according to the reasoning requirement information; and performs training according to the reasoning requirement information and the training strategy of the multiple models to obtain multiple sub-models of the first model;
    其中,所述多模型的训练策略包括以下一项或多项:The multi-model training strategy includes one or more of the following:
    数据处理策略、训练的算法、训练的模式、子模型的数量、子模型的类型。Data processing strategy, training algorithm, training mode, number of sub-models, and type of sub-models.
  4. 根据权利要求1所述的方法,其特征在于,所述模型请求信息用于请求获取多模型时,所述第一通信装置根据所述模型请求信息,确定第一模型,包括:The method according to claim 1, wherein when the model request information is used to request to obtain multiple models, the first communication device determines the first model according to the model request information, comprising:
    所述第一通信装置根据所述推理需求信息,从至少一个预设的多模型中确定所述第一模型。The first communication device determines the first model from at least one preset multiple models according to the inference requirement information.
  5. 根据权利要求1至4中任一项所述的方法,其特征在于,所述模型请求信息中还包括多模型指示信息,所述多模型指示信息用于指示请求训练或获取的模型为多模型。The method according to any one of claims 1 to 4 is characterized in that the model request information also includes multi-model indication information, and the multi-model indication information is used to indicate that the model requested to be trained or obtained is a multi-model.
  6. 根据权利要求1至5中任一项所述的方法,其特征在于,所述推理需求信息包括以下一项或多项:The method according to any one of claims 1 to 5, characterized in that the reasoning requirement information includes one or more of the following:
    推理的类型、推理的性能需求、推理的速度需求、推理的功耗需求。The type of inference, the performance requirements of inference, the speed requirements of inference, and the power consumption requirements of inference.
  7. 根据权利要求1至6中任一项所述的方法,其特征在于,所述第一模型的信息中包括所述第一模型的模型信息和所述第一模型的多个子模型的信息,每个子模型的信息包括以下一项或多项:The method according to any one of claims 1 to 6, characterized in that the information of the first model includes model information of the first model and information of multiple sub-models of the first model, and the information of each sub-model includes one or more of the following:
    子模型的标识信息、子模型的级别、子模型的性能、性能约束;Sub-model identification information, sub-model level, sub-model performance, and performance constraints;
    所述多个子模型中包括多个第一级子模型和一个第二级子模型,所述第二级子模型用于聚合所述多个第一级子模型的推理信息;或者所述多个子模型均为第一级子模型,所述第一模型的信息中还包括聚合方法和/或权重信息。The multiple sub-models include multiple first-level sub-models and one second-level sub-model, and the second-level sub-model is used to aggregate the reasoning information of the multiple first-level sub-models; or the multiple sub-models are all first-level sub-models, and the information of the first model also includes aggregation method and/or weight information.
  8. 根据权利要求1至7中任一项所述的方法,其特征在于,所述方法还包括:The method according to any one of claims 1 to 7, characterized in that the method further comprises:
    所述第一通信装置发送所述第一模型的推理性能信息,所述第一模型的推理性能信息包括以下一项或多项:The first communication device sends the reasoning performance information of the first model, where the reasoning performance information of the first model includes one or more of the following:
    所述第一模型的性能、所述第一模型的大小信息、所述第一模型的推理的功耗、所述第一模型的推理速度、所述第一模型的算力。The performance of the first model, the size information of the first model, the power consumption of the reasoning of the first model, the reasoning speed of the first model, and the computing power of the first model.
  9. 根据权利要求1至8中任一项所述的方法,其特征在于,所述第一通信装置为以下任一项:The method according to any one of claims 1 to 8, characterized in that the first communication device is any one of the following:
    模型训练功能网元、模型训练功能实体、包含模型训练功能的通信装置。Model training function network element, model training function entity, and communication device including model training function.
  10. 一种通信方法,其特征在于,包括:A communication method, characterized by comprising:
    第二通信装置接收第二信息,所述第二信息中包括第一模型的信息,所述第一模型是根据推理需求信息确定的,且所述第一模型为多模型;The second communication device receives second information, wherein the second information includes information of a first model, the first model is determined according to reasoning requirement information, and the first model is a multi-model;
    所述第二通信装置基于所述第一模型的信息,得到所述第一模型的推理信息。The second communication device obtains inference information of the first model based on the information of the first model.
  11. 根据权利要求10所述的方法,其特征在于,所述第二通信装置接收第二信息之前,所述方法还包括:The method according to claim 10, characterized in that before the second communication device receives the second information, the method further comprises:
    所述第二通信装置发送推理能力信息和所述推理需求信息;The second communication device sends the reasoning capability information and the reasoning requirement information;
    所述推理能力信息包括推理能力指示信息,以及下述一项或多项:The reasoning capability information includes reasoning capability indication information and one or more of the following:
    推理的算力、存储空间;所述推理能力指示信息用于指示所述第二通信装置支持多模型的推理; The computing power and storage space for reasoning; the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning;
    所述推理需求信息包括下述一项或多项:The reasoning requirement information includes one or more of the following:
    推理的类型、推理的性能需求、推理的速度需求、推理的功耗需求。The type of inference, the performance requirements of inference, the speed requirements of inference, and the power consumption requirements of inference.
  12. 根据权利要求10所述的方法,其特征在于,所述第一模型的信息中包括所述第一模型的模型信息和所述第一模型的多个子模型的信息,每个子模型的信息包括以下一项或多项:The method according to claim 10, characterized in that the information of the first model includes model information of the first model and information of multiple sub-models of the first model, and the information of each sub-model includes one or more of the following:
    子模型的标识信息、子模型的级别、子模型的性能、性能约束。Sub-model identification information, sub-model level, sub-model performance, and performance constraints.
  13. 根据权利要求12所述的方法,其特征在于,所述多个子模型中包括多个第一级子模型和一个第二级子模型,所述第二级的子模型用于聚合所述多个第一级子模型的推理信息;The method according to claim 12, characterized in that the multiple sub-models include multiple first-level sub-models and one second-level sub-model, and the second-level sub-model is used to aggregate the reasoning information of the multiple first-level sub-models;
    所述第二通信装置基于所述第一模型的信息,得到所述第一模型的推理信息,包括:The second communication device obtains the inference information of the first model based on the information of the first model, including:
    所述第二通信装置基于所述多个第一级子模型的信息,利用所述多个第一级子模型分别进行推理,得到所述多个第一级子模型的推理信息;The second communication device performs reasoning using the multiple first-level sub-models respectively based on the information of the multiple first-level sub-models to obtain reasoning information of the multiple first-level sub-models;
    所述第二通信装置使用所述第二级子模型对所述多个第一级子模型的推理信息进行聚合,得到所述第一模型的推理信息。The second communication device aggregates the reasoning information of the plurality of first-level sub-models using the second-level sub-model to obtain the reasoning information of the first model.
  14. 根据权利要求12所述的方法,其特征在于,所述多个子模型均为第一级子模型,所述第一模型的信息中还包括聚合方法和/或权重信息;The method according to claim 12, characterized in that the multiple sub-models are all first-level sub-models, and the information of the first model also includes aggregation method and/or weight information;
    所述第二通信装置基于所述第一模型的信息,得到所述第一模型的推理信息,包括:The second communication device obtains the inference information of the first model based on the information of the first model, including:
    所述第二通信装置基于所述多个子模型的信息,利用所述多个子模型分别进行推理,得到所述多个子模型的推理信息;The second communication device performs reasoning using the multiple sub-models respectively based on the information of the multiple sub-models to obtain reasoning information of the multiple sub-models;
    所述第二通信装置根据所述聚合方法和/或权重信息对所述多个子模型的推理信息进行聚合,得到所述第一模型的推理信息。The second communication device aggregates the reasoning information of the multiple sub-models according to the aggregation method and/or weight information to obtain the reasoning information of the first model.
  15. 根据权利要求10至14任一项所述的方法,其特征在于,所述第二通信装置为以下任一项:The method according to any one of claims 10 to 14, characterized in that the second communication device is any one of the following:
    模型推理功能网元、模型推理功能实体、包含模型推理功能的通信装置。Model reasoning function network element, model reasoning function entity, and communication device including model reasoning function.
  16. 一种通信方法,其特征在于,包括:A communication method, comprising:
    第三通信装置接收第一通信装置的训练能力指示信息;所述训练能力指示信息用于指示所述第一通信装置支持多模型的训练;The third communication device receives the training capability indication information of the first communication device; the training capability indication information is used to indicate that the first communication device supports multi-model training;
    所述第三通信装置接收第二通信装置的推理需求信息和推理能力信息;所述推理能力信息中包括推理能力指示信息,所述推理能力指示信息用于指示所述第二通信装置支持多模型的推理;The third communication device receives the reasoning requirement information and reasoning capability information of the second communication device; the reasoning capability information includes reasoning capability indication information, and the reasoning capability indication information is used to indicate that the second communication device supports multi-model reasoning;
    所述第三通信装置向所述第一通信装置发送模型请求信息,所述模型请求信息中包括所述推理需求信息;The third communication device sends model request information to the first communication device, where the model request information includes the reasoning requirement information;
    所述第三通信装置从所述第一通信装置接收第一信息,所述第一信息中包括第一模型的信息,所述第一模型为多模型,所述第一模型是根据所述推理需求信息确定的;The third communication device receives first information from the first communication device, where the first information includes information of a first model, the first model is a multi-model, and the first model is determined according to the reasoning requirement information;
    所述第三通信装置向所述第二通信装置发送第二信息,所述第二信息中包括所述第一模型的信息。The third communication device sends second information to the second communication device, where the second information includes information of the first model.
  17. 根据权利要求16所述的方法,其特征在于,所述第二通信装置的推理需求信息包括以下一项或多项:The method according to claim 16, characterized in that the inference requirement information of the second communication device includes one or more of the following:
    推理的类型、推理的性能需求、推理的速度需求、推理的功耗需求。The type of inference, the performance requirements of inference, the speed requirements of inference, and the power consumption requirements of inference.
  18. 根据权利要求16所述的方法,其特征在于,所述第二通信装置的推理能力信息还包括以下一项或多项:The method according to claim 16, characterized in that the reasoning capability information of the second communication device further includes one or more of the following:
    推理的算力、存储空间。The computing power and storage space for reasoning.
  19. 根据权利要求16所述的方法,其特征在于,所述模型请求信息中还包括多模型指示信息,所述多模型指示信息用于指示请求训练或获取的模型为多模型。The method according to claim 16 is characterized in that the model request information also includes multi-model indication information, and the multi-model indication information is used to indicate that the model requested to be trained or obtained is a multi-model.
  20. 根据权利要求16所述的方法,其特征在于,所述方法还包括:The method according to claim 16, characterized in that the method further comprises:
    所述第三通信装置从所述第一通信装置接收所述第一模型的推理性能信息;The third communication device receives the inference performance information of the first model from the first communication device;
    所述第三通信装置根据所述第二通信装置的推理需求信息和所述推理能力信息,以及所述第一模型的推理信息和所述第一模型的信息,调整所述第一模型中的子模型的数量;The third communication device adjusts the number of sub-models in the first model according to the reasoning requirement information and the reasoning capability information of the second communication device, the reasoning information of the first model and the information of the first model;
    所述第一模型的推理性能信息包括以下一项或多项:The inference performance information of the first model includes one or more of the following:
    所述第一模型的性能、所述第一模型的大小信息、所述第一模型的推理的功耗、所述第一模型的推理速度、所述第一模型的算力。The performance of the first model, the size information of the first model, the power consumption of the reasoning of the first model, the reasoning speed of the first model, and the computing power of the first model.
  21. 根据权利要求16至20中任一项所述的方法,其特征在于,所述第一模型的信息中包括所述第一模型的模型信息和所述第一模型的多个子模型的信息,每个子模型的信息包括以下一项或多项: The method according to any one of claims 16 to 20, characterized in that the information of the first model includes model information of the first model and information of multiple sub-models of the first model, and the information of each sub-model includes one or more of the following:
    子模型的标识信息、子模型的级别、子模型的性能、性能约束;Sub-model identification information, sub-model level, sub-model performance, and performance constraints;
    所述多个子模型中包括多个第一级子模型和一个第二级子模型,所述第二级子模型用于聚合所述多个第一级子模型的推理信息;或者所述多个子模型均为第一级子模型,所述第一模型的信息中还包括聚合方式和/或权重信息。The multiple sub-models include multiple first-level sub-models and one second-level sub-model, and the second-level sub-model is used to aggregate the reasoning information of the multiple first-level sub-models; or the multiple sub-models are all first-level sub-models, and the information of the first model also includes aggregation method and/or weight information.
  22. 根据权利要求16至21中任一项所述的方法,其特征在于,所述第三通信装置为以下任一项:The method according to any one of claims 16 to 21, characterized in that the third communication device is any one of the following:
    模型管理功能网元、模型管理功能实体、包含模型管理功能的通信装置。Model management function network element, model management function entity, and communication device including model management function.
  23. 一种通信装置,其特征在于,包括用于执行如权利要求1至9中任一项所述方法的单元或模块,或用于执行如权利要求10至15中任一项所述方法的单元或模块,或用于执行如权利要求16至22中任一项所述方法的单元或模块。A communication device, characterized in that it includes a unit or module for executing the method as described in any one of claims 1 to 9, or a unit or module for executing the method as described in any one of claims 10 to 15, or a unit or module for executing the method as described in any one of claims 16 to 22.
  24. 一种通信装置,其特征在于,包括处理器和接口电路,所述接口电路用于接收来自所述通信装置之外的其它通信装置的信号并传输至所述处理器或将来自所述处理器的信号发送给所述通信装置之外的其它通信装置,所述处理器通过逻辑电路或执行代码指令用于实现如权利要求1至9中任一项所述的方法,或用于实现如权利要求10至15中任一项所述的方法,或用于实现如权利要求16至22中任一项所述方法的单元或模块。A communication device, characterized in that it includes a processor and an interface circuit, wherein the interface circuit is used to receive signals from other communication devices outside the communication device and transmit them to the processor or send signals from the processor to other communication devices outside the communication device, and the processor is used to implement the method as described in any one of claims 1 to 9, or to implement the method as described in any one of claims 10 to 15, or to implement a unit or module of the method as described in any one of claims 16 to 22 through a logic circuit or execution code instructions.
  25. 一种计算机程序产品,其特征在于,包括计算机程序,当所述计算机程序被通信装置执行时,实现如权利要求1至9中任一项所述的方法,或实现如权利要求10至15中任一项所述的方法,或实现如权利要求16至22中任一项所述方法。A computer program product, characterized in that it includes a computer program, and when the computer program is executed by a communication device, it implements the method as described in any one of claims 1 to 9, or implements the method as described in any one of claims 10 to 15, or implements the method as described in any one of claims 16 to 22.
  26. 一种计算机可读存储介质,其特征在于,所述存储介质中存储有计算机可读程序或指令,当所述计算机程序或指令被通信装置执行时,实现如权利要求1至9中任一项所述的方法,或实现如权利要求10至15中任一项所述的方法,或实现如权利要求16至22中任一项所述方法。 A computer-readable storage medium, characterized in that a computer-readable program or instruction is stored in the storage medium, and when the computer program or instruction is executed by a communication device, the method as described in any one of claims 1 to 9 is implemented, or the method as described in any one of claims 10 to 15 is implemented, or the method as described in any one of claims 16 to 22 is implemented.
PCT/CN2023/138554 2023-01-29 2023-12-13 Communication method and device WO2024156232A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202310115449.1A CN118413455A (en) 2023-01-29 2023-01-29 Communication method and device
CN202310115449.1 2023-01-29

Publications (1)

Publication Number Publication Date
WO2024156232A1 true WO2024156232A1 (en) 2024-08-02

Family

ID=91969914

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/138554 WO2024156232A1 (en) 2023-01-29 2023-12-13 Communication method and device

Country Status (2)

Country Link
CN (1) CN118413455A (en)
WO (1) WO2024156232A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110324170A (en) * 2018-03-30 2019-10-11 华为技术有限公司 Data analysis equipment, multi-model are total to decision system and method
WO2022028665A1 (en) * 2020-08-03 2022-02-10 Nokia Technologies Oy Distributed training in communication networks
CN114254751A (en) * 2020-09-21 2022-03-29 华为技术有限公司 Collaborative inference method and communication device
US20220156658A1 (en) * 2019-03-05 2022-05-19 Telefonaktiebolaget Lm Ericsson (Publ) System and method for managing resources

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110324170A (en) * 2018-03-30 2019-10-11 华为技术有限公司 Data analysis equipment, multi-model are total to decision system and method
US20220156658A1 (en) * 2019-03-05 2022-05-19 Telefonaktiebolaget Lm Ericsson (Publ) System and method for managing resources
WO2022028665A1 (en) * 2020-08-03 2022-02-10 Nokia Technologies Oy Distributed training in communication networks
CN114254751A (en) * 2020-09-21 2022-03-29 华为技术有限公司 Collaborative inference method and communication device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
YUANYUAN ZHANG, MEDIATEK INC.: "Discussion on RAN2 aspects for LCM", 3GPP DRAFT; R2-2211610; TYPE DISCUSSION, 3RD GENERATION PARTNERSHIP PROJECT (3GPP), MOBILE COMPETENCE CENTRE ; 650, ROUTE DES LUCIOLES ; F-06921 SOPHIA-ANTIPOLIS CEDEX ; FRANCE, vol. 3GPP RAN 2, no. Toulouse, FR; 20221114 - 20221118, 4 November 2022 (2022-11-04), Mobile Competence Centre ; 650, route des Lucioles ; F-06921 Sophia-Antipolis Cedex ; France, XP052215715 *

Also Published As

Publication number Publication date
CN118413455A (en) 2024-07-30

Similar Documents

Publication Publication Date Title
US20210235277A1 (en) Method and apparatus for dynamically allocating radio resources in a wireless communication system
CN114143799A (en) Communication method and device
US11516881B2 (en) Slice management system and slice management method
KR20200014925A (en) Method and apparatus for notifying and sending system messages
EP3961999A1 (en) Method for determining service transmission requirement, apparatus, and system
WO2022073207A1 (en) Model evaluation method and apparatus
WO2016201913A1 (en) Data transmission method, device and system, and computer storage medium
WO2022087930A1 (en) Model configuration method and apparatus
WO2022226713A1 (en) Method and apparatus for determining policy
WO2022028580A1 (en) Reselection determination method, network data analytic function and storage medium
TW202044901A (en) Direct link transmission method and terminal
US20230403223A1 (en) Data analysis apparatus management and control method and communication apparatus
EP4175352A1 (en) Information transmission method and device, storage medium, and electronic device
WO2024156232A1 (en) Communication method and device
WO2024012331A1 (en) Method and apparatus for determining artificial intelligence (ai) model
WO2023207534A1 (en) Network management method, apparatus and system, and storage medium
WO2021152629A1 (en) Method and apparatus for dynamically allocating radio resources in a wireless communication system
CN115460617A (en) Network load prediction method and device based on federal learning, electronic equipment and medium
CN115529310B (en) Communication method, device, related equipment and storage medium
CN111698784B (en) Data transmission method, device, equipment and storage medium
WO2024183608A1 (en) Communication method and related apparatus
WO2024169407A9 (en) Communication method and apparatus
WO2024169407A1 (en) Communication method and apparatus
WO2023185711A1 (en) Communication method and apparatus used for training machine learning model
WO2024061125A1 (en) Communication method and apparatus

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23918232

Country of ref document: EP

Kind code of ref document: A1