WO2022108527A1 - Model processing method, system and apparatus, medium, and electronic device - Google Patents

Model processing method, system and apparatus, medium, and electronic device Download PDF

Info

Publication number
WO2022108527A1
WO2022108527A1 PCT/SG2021/050707 SG2021050707W WO2022108527A1 WO 2022108527 A1 WO2022108527 A1 WO 2022108527A1 SG 2021050707 W SG2021050707 W SG 2021050707W WO 2022108527 A1 WO2022108527 A1 WO 2022108527A1
Authority
WO
WIPO (PCT)
Prior art keywords
model
inference
models
target
service executor
Prior art date
Application number
PCT/SG2021/050707
Other languages
French (fr)
Chinese (zh)
Inventor
陈程
周子凯
余乐乐
解浚源
吴良超
常龙
张力哲
刘小兵
吴迪
Original Assignee
脸萌有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 脸萌有限公司 filed Critical 脸萌有限公司
Publication of WO2022108527A1 publication Critical patent/WO2022108527A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • G06N20/20Ensemble learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/602Providing cryptographic facilities or services
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/04Inference or reasoning models

Definitions

  • the federated learning system can effectively help multiple institutions to complete the joint training of models.
  • the federated learning model is usually composed of multiple sub-models. How to ensure the reliability of the inference service is an important issue when the inference service is performed through the federated learning model.
  • the present disclosure provides a model processing method, the method includes: acquiring multiple sub-models; splicing the multiple sub-models to obtain a target model; after receiving a message sent by an inference service executor for the target model In the case of the model obtaining request, the target model is sent to the inference service executor, so that the inference service executor obtains an inference result through the target model.
  • the present disclosure provides a model processing method. The method includes: an inference service executor sends a model acquisition request for a target model, where the target model is obtained by splicing multiple sub-models; the inference service executes The party receives the target model, and obtains an inference result through the target model.
  • the present disclosure provides a model processing system, the system includes a model optimization platform and a model storage platform; the model optimization platform is used for acquiring multiple sub-models, and splicing the multiple sub-models to obtain target model, and send the target model to the model storage platform; the model storage platform is configured to store the target model in the case of receiving a model acquisition request for the target model sent by the inference service executor The model is sent to the inference service executor, so that the inference service executor obtains an inference result through the target model.
  • the present disclosure provides a model processing device, the device comprising: an acquisition module configured to acquire multiple sub-models; and a splicing module configured to splicing the multiple sub-models to obtain a target model; a target model sending module, configured to send the target model to the inference service executor in the case of receiving a model acquisition request for the target model sent by the inference service executor, so that The inference service executor obtains an inference result through the target model.
  • the present disclosure provides a model processing apparatus, the apparatus comprising: an acquisition request sending module, configured to send a model acquisition request for a target model, wherein the target model is to convert the multiple sub-models obtained by splicing; an inference module, configured to receive the target model, and obtain an inference result through the target model.
  • the model processing apparatus may be set at the execution side of the inference service.
  • the present disclosure provides a computer-readable medium on which a computer program is stored, and when the program is executed by a processing apparatus, implements the steps of the method provided in the first aspect of the present disclosure.
  • the present disclosure provides a computer-readable medium on which a computer program is stored, and when the program is executed by a processing apparatus, implements the steps of the method provided in the second aspect of the present disclosure.
  • the present disclosure provides an electronic device, comprising: a storage device on which a computer program is stored; and a processing device for executing the computer program in the storage device, so as to implement the computer program provided in the first aspect of the present disclosure. the steps of the method.
  • the present disclosure provides an electronic device, comprising: a storage device on which a computer program is stored; and a processing device for executing the computer program in the storage device, so as to implement the computer program provided in the second aspect of the present disclosure. the steps of the method.
  • the present disclosure provides a computer program, comprising: instructions that, when executed by a processor, cause the processor to execute the model processing method according to any one of the foregoing embodiments.
  • the present disclosure provides a computer program product comprising instructions that, when executed by a processor, cause the processor to execute the model processing method according to any one of the foregoing embodiments.
  • FIG. 1 is a schematic diagram of a federated learning model in the related art.
  • Fig. 2 is a flow chart of a model processing method according to some exemplary embodiments.
  • Fig. 3 is a schematic diagram of a target model according to some exemplary embodiments.
  • FIG. 4 is a schematic diagram of a model processing system according to some exemplary embodiments.
  • Fig. 5 is a schematic diagram illustrating an inference service executor obtaining an inference result through a target model according to its own model input data, according to some exemplary embodiments.
  • Fig. 6 is a flow chart of a model processing method according to some exemplary embodiments.
  • Fig. 1 is a schematic diagram of a federated learning model in the related art.
  • Fig. 2 is a flow chart of a model processing method according to some exemplary embodiments.
  • Fig. 3 is a schematic diagram of a target model according to some exemplary embodiments.
  • FIG. 4 is a schematic diagram of a model processing system
  • FIG. 7 is a block diagram of a model processing apparatus according to some exemplary embodiments.
  • Fig. 8 is a block diagram of a model processing apparatus according to some exemplary embodiments.
  • FIG. 9 is a schematic structural diagram of an electronic device according to some exemplary embodiments. DETAILED DESCRIPTION Embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While certain embodiments of the present disclosure are shown in the drawings, it should be understood that the present disclosure may be embodied in various forms and should not be construed as limited to the embodiments set forth herein, but rather are provided for the purpose of A more thorough and complete understanding of the present disclosure.
  • the federated learning system can combine the data of multiple data owners to train a common federated learning model.
  • the federated learning model is trained by combining the data of multiple data owners, and the training data is more comprehensive. Therefore, the accuracy of the federated learning model is higher.
  • FIG. 1 is a schematic diagram of a federated learning model in the related art.
  • the federated learning model includes sub-model A and sub-model B.
  • the sub-model A corresponds to the model training participant 1
  • the model input data X, Y, and Z of the sub-model A are data owned by the model training participant 1.
  • the sub-model B corresponds to the model training participant 2, and the model input data M and N of the sub-model B are data owned by the model training participant 2.
  • each model training participant loads its own sub-models, that is, model training participant 1 loads sub-model A, and model training participant 2 loads sub-model B.
  • the model training participant 1 performs computation through the sub-model A according to the model input data X, Y, and Z. Then model training participant 1 needs to remotely send the data to model training participant 2 through the sending node of sub-model A, so as to transmit the data to the receiving node of sub-model B. The model training participant 2 then obtains an inference result through the sub-model B according to the data received by the receiving node and the model input data M and N.
  • remote communication is required between multiple model training participants to complete the entire inference service, that is, the sending node and the receiving node use remote communication to transmit data, and the communication overhead is large.
  • Fig. 2 is a flow chart of a model processing method according to some exemplary embodiments. As shown in FIG. 2, the method may include S201 ⁇ S203o In S201, multiple sub-models are acquired.
  • FIG. 3 is a schematic diagram of a target model according to some exemplary embodiments.
  • the target model shown in FIG. 3 may be obtained according to the federated learning model shown in FIG. 1 , and there is a connection relationship between the sending node of the sub-model A and the receiving node of the sub-model B.
  • the computing node of the sub-model A connected to the sending node and the computing node of the sub-model B connected to the receiving node can be connected to obtain the target model.
  • the target model is the overall full model obtained by splicing sub-model A and sub-model B together.
  • the present disclosure takes two sub-models as examples for illustration, which does not constitute a limitation on the implementation of the present disclosure. In practical applications, the number of sub-models may be multiple, which is not specifically limited in the present disclosure. .
  • the target model is sent to the inference service executor, so that the inference service executor obtains the inference result through the target model.
  • Inference service can refer to the process by which the server performs computations through the model based on the input data and obtains the result.
  • the user's current shopping intention can be inferred through a model, and then the user can be provided with an inference result that meets his shopping intention and needs.
  • the user's current search intent can be inferred through the model, and then the user can be provided with an inference result that conforms to the user's search intent.
  • one of the model training participants can be used as the inference service executor, load the target model, and obtain the inference result through the target model.
  • the target model is obtained by splicing multiple sub-models, and the inference service executor can directly obtain the inference results through the overall target model, without each model training participant loading its own sub-models, and without the need for multiple model training participants. It can effectively avoid the problem of unstable long-distance communication. It is worth noting that, when it is mentioned in the present disclosure that the inference service executor performs operations of sending, receiving, and processing data, it may be understood that the inference service executor performs these operations through a server device.
  • a target model is obtained by splicing multiple sub-models, and the inference service executor can obtain an inference result through the target model.
  • the target model is obtained by splicing multiple sub-models, and the inference service executor can directly obtain the inference result through the overall target model, without each model training participant loading its own sub-models separately.
  • the entire inference service process can be completed locally on the inference service executor, without the need for remote communication between multiple model training participants to transmit data. In this way, not only the communication overhead can be reduced, but also the problem of unstable remote transmission caused by factors such as network routing can be effectively avoided, the normal operation of the inference service can be ensured, and the reliability of the inference service can be improved.
  • the model processing method shown in FIG. 2 may be applied to a model processing apparatus including a splicing module.
  • the model processing device may be, for example, a cloud server, the acquisition module in the model processing device acquires multiple sub-models, and the splicing module splices the multiple sub-models to obtain the target model.
  • the model processing method shown in FIG. 2 can also be applied to a model processing system.
  • FIG. 4 is a schematic diagram of a model processing system according to some exemplary embodiments. As shown in FIG. 4 , the model processing system may include a model optimization platform 401, a model storage platform 402, a model training platform 403, a model metadata storage platform 404, a model training participant 1, and a model training participant 2.
  • the model training platform 403 is used to train each sub-model, such as sub-model A and sub-model B.
  • the model meta information storage platform 404 may be used to store model related meta information.
  • the model training platform 403 can send multiple sub-models to the model optimization platform 401.
  • the model optimization platform 401 can be used to obtain the multiple sub-models sent by the model training platform 403, splicing the multiple sub-models to obtain the target model, and sending the target model to Model storage platform 402 o
  • the inference service executor may send a model acquisition request for the target model to the model storage platform 402; the model storage platform 402 may send the target model to the inference service executor when receiving the request.
  • the inference service executor may be one of model training participant 1 and model training participant 2, for example. FIG.
  • the step of splicing multiple sub-models in S202 may include: acquiring model meta information, where the model meta information may include a sending node having a sub-model of the sending node and other sub-models that have a connection relationship with the sending node The connection relationship information between the receiving nodes; according to the model meta information, connect the computing node of the sub-model connected to the sending node and the computing nodes of other sub-models connected to the receiving node to connect the multiple sub-models. stitching.
  • Model meta-information may refer to information used to describe a model, and may include connection relationship information between nodes.
  • the model optimization platform 401 can obtain model meta information from the model meta information storage platform 404, and according to the model meta information, connect the The computing nodes of the sub-model are connected with the computing nodes of other sub-models connected to the receiving node. In this way, data can be directly transmitted between two computing nodes that originally need to forward data through the sending node and the receiving node.
  • the inference service executor performs the inference service through the target model, and the inference service executor locally
  • the entire inference service process can be completed without the need for remote communication of data, effectively avoiding the problem of unstable remote transmission caused by factors such as network routing, ensuring the normal operation of the inference service, and improving the reliability of the inference service.
  • multiple sub-models correspond to multiple model training participants one-to-one, and each model training participant has its own model input data.
  • the inference service executor can be one of multiple model training participants, and the inference service executor can obtain inference results through the target model according to its own model input data.
  • the inference service performer may be any of a plurality of model training participants.
  • Fig. 5 is a schematic diagram illustrating an inference service executor obtaining an inference result through a target model according to its own model input data, according to some exemplary embodiments.
  • Figure 5 takes the inference service executor as the model training participant 1 as an example.
  • the model training participant 1 obtains the target model, it can input data X, Y, and Z according to its own model, and obtain an inference result through the target model.
  • the inference service executor may be the party that needs the inference result among the multiple model training participants, that is, the inference result demander.
  • the model training participant 1 needs the final inference result, and the model training participant 1 can be used as the inference service executor to obtain the target model and perform the inference service.
  • the inference service executor may send the inference result to the inference result requester.
  • model training participant 2 needs an inference result, and model training participant 1 can send the inference result to model training participant 2 .
  • other model training participants may not transmit model input data to the inference service executor, and the inference service executor can perform inference services according to its own data, with low communication overhead.
  • the reasoning service is completed by the inference service executor, and there is no need to carry out the process of remote communication and data transmission between multiple model training participants, which reduces the communication overhead and improves the stability of the inference service.
  • multiple sub-models correspond to multiple model training participants one-to-one, each model training participant has its own model input data, and the inference service executor is one of the multiple model training participants .
  • the inference service executor can obtain the inference results in the following ways: Receive encrypted model input data sent by other model training participants except the inference service executor; According to the inference service executor's own model input data and other model training participants The encrypted model input data is obtained, and the inference result is obtained through the target model.
  • model training participants can add model input data. encrypted processing. Send the encrypted model input data to the inference service executor, and the present disclosure does not specifically limit the encryption method.
  • the inference service executor can use its own model input data and the encrypted model input data of other model training participants to obtain inference results through the target model, that is, to perform inference services based on the model input data of each sub-model.
  • the inference service executor may be the model training participant that needs to receive the encrypted model input data of other model training participants with the smallest amount of data.
  • the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor. For example, if the model training participant 1 performs the inference service, the model training participant 2 needs to send the model input data M and N of the sub-model B to the model training participant 1; if the model training participant 2 performs the inference service, Model training participant 1 needs to send the model input data X, Y, and Z of sub-model A to model training participant 2.
  • the model training participant 1 needs to receive the minimum data volume of the model input data, which can be executed by the model training participant 1 as an inference service. square.
  • the inference service executor performs the inference service according to the model input data of each sub-model, the model training participant with the smallest amount of model input data to be received can be used as the inference service executor, which can, to a certain extent, be used as the inference service executor. Reduce communication overhead.
  • the inference service executor performs the calculation through the target model, and there is no need to carry out the process of remote communication and data transmission among multiple model training participants, thereby improving the stability of the inference service process.
  • multiple sub-models correspond to multiple model training participants one-to-one, each model training participant has its own model input data, and the inference service executor is not a model training participant; inference service The executor can obtain the inference result in the following ways: Receive the encrypted model input data sent by each model training participant respectively; Obtain the inference result through the target model according to the encrypted model input data of each model training participant.
  • the inference service executor may not be a model training participant, for example, it may be a preset cloud server.
  • Fig. 6 is a flow chart of a model processing method according to some exemplary embodiments. The method can be applied to the inference service executor, that is, the server device of the inference service executor. As shown in FIG.
  • the method may include S501 and S502.
  • the inference service executor sends a model acquisition request for the target model.
  • the target model is obtained by splicing the multiple sub-models.
  • the inference service executor may send a model acquisition request to the model storage platform, and may also send a model acquisition request to a model processing apparatus including a splicing module, which is not limited in the present disclosure.
  • the inference service executor receives the target model, and obtains an inference result through the target model.
  • the inference service executor can send a model acquisition request for the target model.
  • the target model is obtained by splicing multiple sub-models, and the inference service executor can obtain the inference result through the target model.
  • the inference service executor obtains the inference result through the target model.
  • the inference service executor can complete the entire inference service process locally. In this way, there is no need to perform remote communication of data, which not only reduces communication overhead, but also effectively avoids the problem of unstable remote transmission caused by factors such as network routing, ensures the normal operation of inference services, and improves the reliability of inference services.
  • the multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is one of the multiple model training participants.
  • the step of obtaining the inference result through the target model in S502 may include: inputting data according to the model of the inference service executor itself, and obtaining the inference result through the target model.
  • the multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is one of the multiple model training participants.
  • the step of obtaining the inference result from the target model in S502 may include: receiving encrypted model input data sent by other model training participants except the inference service executor; and according to the model input data of the inference service executor itself , and the encrypted model input data of the other model training participants, and obtain the inference result through the target model.
  • the inference service executor is a model training participant that needs to receive the encrypted model input data of other model training participants with the smallest amount of data.
  • the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor.
  • the multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is not the model training participant.
  • the step of obtaining the inference result through the target model in S502 may include: respectively receiving encrypted model input data sent by each of the model training participants; according to the encrypted model input data of each of the model training participants, The target model obtains the inference result.
  • the method applied to the inference service executor in the foregoing embodiment the specific manner in which each step performs operations has been described in detail in the embodiment of the method applied to the model processing system or the model processing apparatus including the splicing module, here A detailed explanation will not be given.
  • the present disclosure also provides a model processing system, such as the model processing system shown in FIG.
  • the system may include a model optimization platform and a model storage platform; the model optimization platform is used for acquiring multiple sub-models, and storing the multiple sub-models Splicing is performed to obtain a target model, and the target model is sent to the model storage platform; the model storage platform is configured to, upon receiving a model acquisition request for the target model sent by the inference service executor, Send the target model to the inference service executor, so that the inference service executor obtains an inference result through the target model.
  • the model optimization platform is used to obtain model meta information, and the model The meta-information includes connection relationship information between a sending node having a sub-model of the sending node and receiving nodes of other sub-models that have a connection relationship with the sending node; the model optimization platform is configured to, according to the model meta-information,
  • the computing node of the sub-model connected with the sending node is connected with the computing nodes of the other sub-models connected with the receiving node, so as to splicing the multiple sub-models.
  • the model processing apparatus 600 may include: an acquisition module 601, configured to acquire multiple sub-models; a splicing module 602, configured to splicing the multiple sub-models to obtain a target model;
  • the target model sending module 603 is configured to send the target model to the inference service executor in the case of receiving a model acquisition request for the target model sent by the inference service executor, so that all The inference service executor obtains the inference result through the target model.
  • the splicing module 602 may include: an obtaining submodule configured to obtain model meta information, where the model meta information includes a sending node having a sub-model of a sending node and a sending node having a connection with the sending node the connection relationship information between the receiving nodes of other sub-models of the connection relationship; the splicing sub-module is configured to connect the computing node of the sub-model connected to the sending node and the computing node of the sub-model connected to the sending node according to the model meta information.
  • the computing nodes of the other sub-models connected to the receiving node are connected to splicing the multiple sub-models.
  • multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is one of the multiple model training participants.
  • the inference service executor obtains the inference result through the target model according to its own model input data.
  • multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is one of the multiple model training participants. 1.
  • the inference service executor obtains the inference result in the following ways: Receive encrypted model input data sent by other model training participants except the inference service executor; According to the inference service executor's own model The input data and the encrypted model input data of the other model training participants are used to obtain the inference result through the target model.
  • the inference service executor is a model training participant that needs to receive the encrypted model input data of other model training participants with the smallest amount of data.
  • the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor.
  • Fig. 8 is a block diagram of a model processing apparatus according to some exemplary embodiments.
  • the model processing apparatus 700 can be applied to an inference service executor. As shown in FIG.
  • the model processing apparatus 700 may include: an acquisition request sending module 701, configured to send a model acquisition request for a target model, wherein the target model is obtained by splicing the multiple sub-models
  • the inference module 702 is configured to receive the target model and obtain an inference result through the target model.
  • multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is one of the multiple model training participants;
  • the inference module 702 may include: a first inference sub-module, configured to input data according to the model of the inference service executor itself, and obtain the inference result through the target model.
  • the inference module 702 may include: a first receiving sub-module, configured to receive encrypted model input data sent by other model training participants except the inference service executor; a second inference sub-module, configured by It is configured to obtain the inference result through the target model according to the model input data of the inference service executor itself and the encrypted model input data of the other model training participants.
  • the inference service executor is a model training participant that needs to receive the encrypted model input data of other model training participants with the smallest amount of data.
  • the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor.
  • multiple sub-models are in one-to-one correspondence with multiple model training participants, each model training participant has its own model input data, and the inference service executor is not the model training participant;
  • the reasoning The module 702 may include: a second receiving sub-module, configured to respectively receive encrypted model input data sent by each of the model training participants; The encrypted model input data of the training participant is used to obtain the inference result through the target model.
  • Terminal devices in the embodiments of the present disclosure may include, but are not limited to, mobile phones, notebook computers, digital broadcast receivers, PDAs (personal digital assistants), PADs (tablets), PMPs (portable multimedia players), vehicle-mounted terminals (eg, Mobile terminals such as car navigation terminals), etc., and stationary terminals such as digital TVs, desktop computers, and the like.
  • PDAs personal digital assistants
  • PADs tablets
  • PMPs portable multimedia players
  • vehicle-mounted terminals eg, Mobile terminals such as car navigation terminals
  • stationary terminals such as digital TVs, desktop computers, and the like.
  • the electronic device shown in FIG. 9 is only an example, and should not impose any limitation on the function and scope of use of the embodiments of the present disclosure. As shown in FIG.
  • the electronic device 800 may include a processing device (eg, a central processing unit, a graphics processor, etc.) 801, which may be loaded into random access according to a program stored in a read only memory (ROM) 802 or from a storage device 808
  • the program in the memory (RAM) 803 executes various appropriate actions and processes.
  • various programs and data necessary for the operation of the electronic device 800 are also stored.
  • the processing device 801 , the ROM 802 , and the RAM 803 are connected to each other through a bus 804 .
  • An input/output (I/O) interface 805 is also connected to bus 804 .
  • Input devices 806 including, for example, a touch screen, touchpad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, etc.; including, for example, a liquid crystal display (LCD), speakers, vibration an output device 807 for a computer, etc.; a storage device 808 including, for example, a magnetic tape, a hard disk, etc.; exchange data.
  • FIG. 9 shows electronic device 800 having various means, it should be understood that not all of the illustrated means are required to be implemented or available. More or fewer devices may alternatively be implemented or provided.
  • the processes described above with reference to the flowcharts may be implemented as computer software programs.
  • embodiments of the present disclosure include a computer program product comprising a computer program carried on a non-transitory computer readable medium, the computer program containing program code for performing the method illustrated in the flowchart.
  • the computer program may be downloaded and installed from the network via the communication device 809 , or from the storage device 808 , or from the ROM 802 .
  • the processing device 801 When the computer program is executed by the processing device 801, the above-mentioned functions defined in the methods of the embodiments of the present disclosure are executed.
  • the computer-readable medium mentioned above in the present disclosure may be a computer-readable signal medium or a computer-readable storage medium, or any combination of the above two.
  • the computer-readable storage medium can be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus or device, or any combination of the above. More specific examples of computer readable storage media may include, but are not limited to: electrical connections with one or more wires, portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable Programmable read only memory (EPROM or flash memory), optical fiber, portable compact disk read only memory (CD-ROM), optical storage devices, magnetic storage devices, or any suitable combination of the above.
  • a computer-readable storage medium may be any tangible medium that contains or stores a program that can be used by or in conjunction with an instruction execution system, apparatus, or device.
  • a computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave, carrying computer-readable program code therein. Such propagated data signals may take a variety of forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • a computer-readable signal medium can also be any computer-readable medium other than a computer-readable storage medium that can transmit, propagate, or transmit a program for use by or in conjunction with an instruction execution system, apparatus, or device .
  • the program code embodied on the computer-readable medium may be transmitted by any suitable medium, including but not limited to: wire, optical fiber cable, RF (radio frequency), etc., or any suitable combination of the above.
  • the client and the server can use any currently known or future developed network protocol such as HTTP (HyperText Transfer Protocol) to communicate, and can communicate with digital data in any form or medium Communication (eg, communication network) interconnection.
  • HTTP HyperText Transfer Protocol
  • Examples of communication networks include local area networks (“LAN”), wide area networks (“WAN”), the Internet (eg, the Internet), and peer-to-peer networks (eg, ad hoc peer-to-peer networks), as well as any currently known or future development network of.
  • LAN local area networks
  • WAN wide area networks
  • the Internet eg, the Internet
  • peer-to-peer networks eg, ad hoc peer-to-peer networks
  • the above-mentioned computer-readable medium may be included in the above-mentioned electronic device; it may also exist independently without assembled into the electronic device.
  • the computer-readable medium carries one or more programs, and when the one or more programs are executed by the electronic device, the electronic device causes the electronic device to: acquire multiple sub-models; splicing the multiple sub-models to obtain a target model; In the case of receiving the model acquisition request for the target model sent by the inference service executor, send the target model to the inference service executor, so that the inference service executor obtains the target model through the inference service executor inference result.
  • the computer-readable medium carries one or more programs, and when the one or more programs are executed by the electronic device, the electronic device causes the electronic device to: send a model acquisition request for a target model, where the target model is obtained by splicing the multiple sub-models; receiving the target model, and obtaining an inference result through the target model.
  • Computer program code for carrying out operations of the present disclosure may be written in one or more programming languages, including but not limited to object-oriented programming languages such as Java, Smalltalk. C++, and This includes conventional procedural programming languages such as the "C" language or similar programming languages.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer, or entirely on the remote computer or server.
  • the remote computer may be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or may be connected to an external computer (eg, using an Internet service provider to via an Internet connection).
  • LAN local area network
  • WAN wide area network
  • Internet service provider e.g.
  • each block in the flowchart or block diagram may represent a module, program segment, or part of code that contains one or more logic functions for implementing the specified executable instructions.
  • the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
  • each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations can be implemented by dedicated hardware-based systems that perform the specified functions or operations , or can be implemented using a combination of dedicated hardware and computer instructions.
  • the modules involved in the embodiments of the present disclosure may be implemented in a software manner, and may also be implemented in a hardware manner.
  • the name of the module does not constitute a limitation of the module itself under certain circumstances, for example, the splicing module may also be described as a "sub-model splicing module".
  • the functions described herein above may be performed, at least in part, by one or more hardware logic components.
  • exemplary types of hardware logic components include: Field Programmable Gate Arrays (FPGAs), Application Specific Integrated Circuits (ASICs), Application Specific Standard Products (ASSPs), Systems on Chips (SOCs), Complex Programmable Logical Devices (CPLDs) and more.
  • a machine-readable medium may be a tangible medium that may contain or store a program for use by or in connection with an instruction execution system, apparatus, or device.
  • the machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium.
  • Machine-readable media may include, but are not limited to, electronic, magnetic, optical, electromagnetic, infrared, or semiconductor systems, devices, or devices, or any suitable combination of the foregoing.
  • machine-readable storage media would include one or more wire-based electrical connections, portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM or flash memory), fiber optics, portable compact disk read only memory (CD-ROM), optical storage devices, magnetic storage devices, or any suitable combination of the foregoing.
  • RAM random access memory
  • ROM read only memory
  • EPROM or flash memory erasable programmable read only memory
  • CD-ROM compact disk read only memory
  • optical storage devices magnetic storage devices, or any suitable combination of the foregoing.
  • Example 1 provides a model processing method, the method includes: acquiring multiple sub-models; splicing the multiple sub-models to obtain a target model; In the case of a model acquisition request for the target model sent by the inference party, the target model is sent to the inference service executor, so that the inference service executor obtains an inference result through the target model.
  • Example 2 provides the method of Example 1.
  • the splicing of the multiple sub-models includes: acquiring model meta information, where the model meta information includes a sub-model having a sending node The connection relationship information between the sending node of the sending node and the receiving nodes of other sub-models that have a connection relationship with the sending node; The computing nodes of the other sub-models connected to the receiving node are connected to splicing the multiple sub-models.
  • Example 3 provides the method of Example 1, wherein multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data,
  • the inference service executor is one of the multiple model training participants, and the inference service executor obtains the inference result through the target model according to its own model input data.
  • Example 4 provides the method of Example 1, where multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data,
  • the inference service executor is one of the multiple model training participants; the inference service executor obtains the inference result in the following manner: Receiving the data sent by other model training participants except the inference service executor Encrypted model input data; obtaining the inference result through the target model according to the model input data of the inference service executor itself and the encrypted model input data of the other model training participants.
  • Example 5 provides the method of Example 4, wherein the inference service executor is the model training participant that needs to receive the encrypted model input data of other model training participants with the smallest amount of data.
  • the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor.
  • Example 6 provides the method of Example 1, wherein multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data,
  • the inference service executor is not the model training participant; the inference service executor obtains the inference result in the following ways: Respectively receive encrypted model input data sent by each of the model training participants; The encrypted model input data of the model training participant is obtained, and the inference result is obtained through the target model.
  • Example 7 provides a model processing method, the method comprising: an inference service executor sending a model acquisition request for a target model, where the target model is to convert the multiple obtained by splicing the sub-models; the inference service executor receives the target model, and obtains an inference result through the target model.
  • Example 8 provides the method of Example 7, wherein multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data,
  • the inference service executor is one of the multiple model training participants;
  • the obtaining an inference result through the target model includes: inputting data according to the model of the inference service executor itself, and obtaining the inference result through the target model inference result.
  • Example 9 provides the method of Example 7, where multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data,
  • the inference service executor is one of the multiple model training participants; the obtaining an inference result through the target model includes: receiving encrypted model inputs sent by other model training participants except the inference service executor data; obtaining the inference result through the target model according to the model input data of the inference service executor itself and the encrypted model input data of the other model training participants.
  • Example 10 provides the method of Example 9, where the inference service executor is a model that needs to receive encrypted model input data from other model training participants with the smallest amount of data. Train participants.
  • the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor.
  • Example 11 provides the method of Example 7, wherein multiple sub-models correspond to multiple model training participants one-to-one, and each model training participant has its own model input data, The inference service executor is not the model training participant; the obtaining the inference result through the target model includes: respectively receiving encrypted model input data sent by each of the model training participants; participating in the model training according to each of the model training participants. The encrypted model input data of the party is used to obtain the inference result through the target model.
  • Example 12 provides a model processing system, where the system includes a model optimization platform and a model storage platform; the model optimization platform is configured to acquire multiple sub-models, and store the multiple sub-models The models are spliced to obtain a target model, and the target model is sent to the model storage platform; the model storage platform is used to receive a model acquisition request for the target model sent by the inference service executor , sending the target model to the inference service executor, so that the inference service executor obtains an inference result through the target model.
  • Example 13 provides the system of Example 12, and the model optimization platform is configured to obtain model meta information, the model meta information including a sending node having a sub-model of the sending node and a sending node associated with the sending node.
  • the model optimization platform is configured to, according to the model meta-information, connect the computing nodes of the sub-models connected to the sending node and the The computing nodes of the other sub-models connected to the receiving node are connected to splicing the multiple sub-models.
  • Example 14 provides a model processing apparatus, the apparatus comprising: an acquisition module configured to acquire a plurality of sub-models; a stitching module configured to A target model is obtained by splicing multiple sub-models; a target model sending module is configured to send the target model to the target model in the case of receiving a model acquisition request for the target model sent by the inference service executor and the inference service executor, so that the inference service executor obtains an inference result through the target model.
  • Example 15 provides a model processing apparatus, the apparatus comprising: an acquisition request sending module configured to send a model acquisition request for a target model, wherein the target The model is obtained by splicing the multiple sub-models; the inference module is configured to receive the target model and obtain an inference result through the target model.
  • the model processing means may be provided at the execution side of the inference service.
  • Example 16 provides a computer-readable medium on which a computer program is stored, and when the program is executed by a processing apparatus, implements the steps of the method in any one of Examples 1-6 .
  • Example 17 provides a computer-readable medium on which a computer program is stored, and when the program is executed by a processing apparatus, implements the steps of the method in any one of Examples 7-11 .
  • Example 18 provides an electronic device, including: a storage device, on which a computer program is stored; a processing device, for executing the computer program in the storage device, to The steps of implementing the method of any of Examples 1-6.
  • Example 19 provides an electronic device, including: a storage device on which a computer program is stored; and a processing device for executing the computer program in the storage device, to The steps of implementing the method of any of Examples 7-11.

Abstract

The present disclosure relates to a model processing method, system and apparatus, a medium, and an electronic device. The method comprises: acquiring a plurality of sub-models; stitching the plurality of sub-models to obtain a target model; and when receiving a model acquisition request for the target model from an inference service execution party, sending the target model to the inference service execution party, so that the inference service execution party obtains an inference result by means of the target model.

Description

模 型处 理方 法、 系统、 装置 、 介质及 电子 设备 相关 申请的交叉引用 本 申请是以 CN申请号为 202011298789.5,申请日为 2020年 11月 18日的申请为 基础, 并主张其优先权, 该 CN申请的公开内容在此作为整体引入本申请中。 技术领域 本公开涉及计算机技术领域, 具体地, 涉及一种模型处理方法、 系统、 装置、 介 质及电子设备。 背景技术 联邦机器学习, 又称联邦学习、 联合学习, 在机器学习领域取得越来越广泛的应 用。 联邦学习能够解决数据孤岛和数据隐私问题, 在满足用户隐私保护、 数据安全的 要求下, 通过联邦学习系统有效帮助多个机构完成模型的共同训练。 联邦学习模型通 常由多个子模型构成, 在通过联邦学习模型进行推理服务时, 如何保证推理服务的可 靠性是重要问题。 发 明内容 提供该发明 内容部分以便以简要的形式介绍构思, 这些构思将在后面的具体实施 方式部分被详细描述。 该发明内容部分并不旨在标识要求保护的技术方案的关键特征 或必要特征, 也不旨在用于限制所要求的保护的技术方案的范围。 第一方面 , 本公开提供一种模型处理方法, 所述方法包括: 获取多个子模型; 将 所述多个子模型进行拼接, 得到目标模型; 在接收到推理服务执行方发送的针对所述 目标模型的模型获取请求的情况下, 将所述目标模型发送至所述推理服务执行方, 以 使所述推理服务执行方通过所述目标模型得到推理结果。 第二方面 , 本公开提供一种模型处理方法, 所述方法包括: 推理服务执行方发送 针对目标模型的模型获取请求, 所述目标模型是将多个子模型进行拼接得到的; 所述 推理服务执行方接收所述目标模型, 并通过所述目标模型得到推理结果。 第三方面 , 本公开提供一种模型处理系统, 所述系统包括模型优化平台、 模型存 储平台; 所述模型优化平台用于获取多个子模型, 将所述多个子模型进行拼接, 得到 目标模型, 并将所述目标模型发送至所述模型存储平台; 所述模型存储平台用于在接 收到推理服务执行方发送的针对所述目标模型的模型获取请求的情况下, 将所述目标 模型发送至所述推理服务执行方, 以使所述推理服务执行方通过所述目标模型得到推 理结果。 第 四方面, 本公开提供一种模型处理装置, 所述装置包括: 获取模块, 被配置成 用于获取多个子模型; 拼接模块, 被配置成用于将所述多个子模型进行拼接, 得到目 标模型; 目标模型发送模块, 被配置成用于在接收到推理服务执行方发送的针对所述 目标模型的模型获取请求的情况下, 将所述目标模型发送至所述推理服务执行方, 以 使所述推理服务执行方通过所述目标模型得到推理结果。 第五方面 , 本公开提供一种模型处理装置, 所述装置包括: 获取请求发送模块, 被配置成用于发送针对目标模型的模型获取请求, 其中, 所述目标模型是将所述多个 子模型进行拼接得到的; 推理模块, 被配置成用于接收所述目标模型, 并通过所述目 标模型得到推理结果。 例如, 所述模型处理装置可以设置在推理服务执行方。 第六方面 , 本公开提供一种计算机可读介质, 其上存储有计算机程序, 该程序被 处理装置执行时实现本公开第一方面提供的所述方法的步骤。 第七方面 , 本公开提供一种计算机可读介质, 其上存储有计算机程序, 该程序被 处理装置执行时实现本公开第二方面提供的所述方法的步骤。 第八方面 ,本公开提供一种电子设备,包括:存储装置,其上存储有计算机程序; 处理装置, 用于执行所述存储装置中的所述计算机程序, 以实现本公开第一方面提供 的所述方法的步骤。 第九方面 ,本公开提供一种电子设备,包括;存储装置,其上存储有计算机程序; 处理装置, 用于执行所述存储装置中的所述计算机程序, 以实现本公开第二方面提供 的所述方法的步骤。 第十方面 , 本公开提供一种计算机程序, 包括: 指令, 所述指令当由处理器执行 时使所述处理器执行根据上述任一个实施例中的模型处理方法。 第十一方面 , 本公开提供一种计算机程序产品, 包括指令, 所述指令当由处理器 执行时使所述处理器执行根据上述任一个实施例中的模型处理方法。 本公开 的其他特征和优点将在随后的具体实施方式部分予以详细说明。 附图说明 结合 附图并参考以下具体实施方式, 本公开各实施例的上述和其他特征、 优点及 方面将变得更加明显。 贯穿附图中, 相同或相似的附图标记表示相同或相似的元素。 应当理解附图是示意性的, 原件和元素不一定按照比例绘制。 在附图中: 图 1为相关技术中联邦学习模型的示意图。 图 2是根据一些示例性实施例示出的一种模型处理方法的流程图。 图 3是根据一些示例性实施例示出的一种目标模型的示意图。 图 4是根据一些示例性实施例示出的一种模型处理系统的示意图。 图 5是根据一些示例性实施例示出的一种推理服务执行方根据自身的模型输入数 据通过目标模型得到推理结果的示意图。 图 6是根据一些示例性实施例示出的一种模型处理方法的流程图。 图 7是根据一些示例性实施例示出的一种模型处理装置的框图。 图 8是根据一些示例性实施例示出的一种模型处理装置的框图。 图 9是根据一些示例性实施例示出的一种电子设备的结构示意图。 具体实施方式 下面将参照 附图更详细地描述本公开的实施例。 虽然附图中显示了本公开的某些 实施例, 然而应当理解的是, 本公开可以通过各种形式来实现, 而且不应该被解释为 限于这里阐述的实施例, 相反提供这些实施例是为了更加透彻和完整地理解本公开。 应当理解的是, 本公开的附图及实施例仅用于示例性作用, 并非用于限制本公开的保 护范围。 应当理解 , 本公开的方法实施方式中记载的各个步骤可以按照不同的顺序执行, 和 /或并行执行。此外,方法实施方式可以包括附加的步骤和 /或省略执行示出的步骤。 本公开的范围在此方面不受限制。 本文使用 的术语 “包括”及其变形是开放性包括,即 “包括但不限于”。术语 “基 于,, 是 “至少部分地基于” 。 术语 “一个实施例”表示 “至少一个实施例” ; 术语 “另一实施例”表示 “至少一个另外的实施例” : 术语 “一些实施例”表示 “至少一 些实施例” 。 其他术语的相关定义将在下文描述中给出。 需要注意 , 本公开中提及的 “第一” 、 “第二”等概念仅用于对不同的装置、 模 块或单元进行区分, 并非用于限定这些装置、 模块或单元所执行的功能的顺序或者相 互依存关系。 需要注 意, 本公开中提及的 “一个” 、 “多个” 的修饰是示意性而非限制性的, 本领域技术人员应当理解, 除非在上下文另有明确指出, 否则应该理解为 “一个或多 个” O 本 公开实施方式中的多个装置之间所交互的消息或者信 息的名称仅用于说明性 的目的, 而并不是用于对这些消息或信息的范围进行限制。 联邦学习系统可以联合多个数据拥有方的数据, 训练出共有的联邦学习模型。 联 邦学习模型是综合多个数据拥有方的数据训练出的, 训练数据更为全面。 因此联邦学 习模型的准确性更高。 联邦学习模型通常由多个子模型组成, 图 1为相关技术中联邦 学习模型的示意图。 如 图 1所示, 联邦学习模型包括子模型 A和子模型 B。 例如, 子模型 A对应模型 训练参与方 1, 子模型 A的模型输入数据 X、 Y、 Z为模型训练参与方 1拥有的数据。 子模型 B对应模型训练参与方 2, 子模型 B的模型输入数据 M、 N为模型训练参与方 2拥有的数据。 在通过联邦学习模型进行推理服务时, 各个模型训练参与方分别加载各自的子模 型, 即模型训练参与方 1加载子模型 A, 模型训练参与方 2加载子模型 B。 如图 1所 示, 模型训练参与方 1根据模型输入数据 X、 Y、 Z通过子模型 A进行计算。 然后模 型训练参与方 1需要通过子模型 A的发送节点将数据远程发送到模型训练参与方 2, 从而将数据传输给子模型 B的接收节点。 模型训练参与方 2再根据接收节点接收到的 数据以及模型输入数据 M、 N, 通过子模型 B得出推理结果。 这样, 在通过联邦学习模型进行推理服务时, 多个模型训练参与方之间需要进行 远程通信才能完成整个推理服务, 即由发送节点和接收节点采用远程通信的方式传输 数据, 通信开销较大。 而且 , 远程通信容易受到网络路由等因素的影响, 通常不够稳定, 可靠性较低, 使得推理服务的计算过程不够稳定。 例如, 如果发送节点在将数据远程传输给接收节 点时, 由于网络拥塞现象而无法及时将数据传输到接收节点, 则影响整个推理服务的 进行。本公开为了解决相关技术中存在的问题,提供一种模型处理方法、系统、装置、 介质及电子设备。 图 2是根据一些示例性实施例示出的一种模型处理方法的流程图。 如 图 2所示, 该方法可包括 S201~S203o 在 S201中, 获取多个子模型。 在 S202中, 将多个子模型进行拼接, 得到目标模型。 图 3是根据一些示例性实施例示出的一种目标模型的示意图。 图 3所示的目标模型可以是根据图 1所示的联邦学习模型得到的,子模型 A的发 送节点与子模型 B的接收节点之间具有连接关系。 如图 3所示, 可将与该发送节点相 连的子模型 A的计算节点和与该接收节点相连的子模型 B的计算节点进行连接,得到 目标模型。该目标模型即将子模型 A和子模型 B拼接到一起之后得到的整体的全模型。 值得说明的是, 本公开以两个子模型为例作为示意进行解释说明, 但并不构成对 本公开实施方式的限制, 在实际应用中, 子模型的数量可以有多个, 本公开不做具体 限制。 在 S203 中, 在接收到推理服务执行方发送的针对目标模型的模型获取请求的情 况下, 将目标模型发送至推理服务执行方, 以使推理服务执行方通过目标模型得到推 理结果。 推理服 务可以指的是服务器根据输入数据通过模型进行计算并得到结果的过程。 示例地, 以预测用户的购物意图为例, 根据用户的历史购物行为信息, 可通过模型推 理出用户当前的购物意图, 进而可以向用户提供符合其购物意图和需求的推理结果。 再示例地 , 以预测用户的搜索意图为例, 根据用户的历史点击行为信息, 可通过 模型推理出用户当前的搜索意图, 进而可以向用户提供符合其搜索意图的推理结果。 在一可选实施方式中, 可由其中一个模型训练参与方作为推理服务执行方, 加载 目标模型并通过目标模型得到推理结果。 目标模型是由多个子模型进行拼接得到的, 推理服务执行方可直接通过该整体的目标模型得到推理结果, 无需各个模型训练参与 方分别加载自身的子模型, 也无需多个模型训练参与方之间远程通信传输数据, 从而 可以有效避免远程通信不稳定的问题。 值得说明的是, 本公开中当提及推理服务执行方进行发送、 接收、 处理数据的操 作时, 可理解为是推理服务执行方通过服务器设备进行这些操作。 通过上述技术方案, 将多个子模型进行拼接, 得到目标模型, 推理服务执行方可 通过目标模型得到推理结果。 目标模型是由多个子模型进行拼接得到的, 推理服务执 行方可直接通过该整体的目标模型得到推理结果, 无需各个模型训练参与方分别加载 自身的子模型。 在推理服务执行方本地即可完成整个推理服务的过程, 不需要多个模 型训练参与方之间远程通信传输数据。 这样, 不但能降低通信开销, 还能有效避免由于网络路由等因素的影响导致的远 程传输不稳定的问题, 保证推理服务的正常进行, 从而提升推理服务的可靠性。 在一实施例 中, 图 2所示的模型处理方法可以应用于包括拼接模块的模型处理装 置。 该模型处理装置例如可以是云端服务器, 由该模型处理装置中的获取模块获取多 个子模型, 并由拼接模块将多个子模型进行拼接得到目标模型。 在另一实施例 中, 图 2所示的模型处理方法也可应用于模型处理系统。 图 4是根据一些示例性实施例示出的一种模型处理系统的示意图。 如 图 4所示, 该模型处理系统可包括模型优化平台 401、 模型存储平台 402, 还 可包括模型训练平台 403、模型元信息存储平台 404、模型训练参与方 1、模型训练参 与方 2。 模型训练平台 403用于训练出各个子模型, 例如子模型 A、 子模型 B。 模型元信 息存储平台 404可用于存储模型相关的元信息。 模型训练平台 403可将多个子模型发 送至模型优化平台 401 o 模型优化平台 401可用于获取模型训练平台 403发送的多个 子模型, 将多个子模型进行拼接, 得到目标模型, 并将目标模型发送至模型存储平台 402o 推理服务执行方可 向模型存储平台 402发送针对目标模型的模型获取请求; 模型 存储平台 402在接收到该请求的情况下可将目标模型发送至推理服务执行方。 该推理 服务执行方例如可以为模型训练参与方 1和模型训练参与方 2之一。 图 4以包括两个 模型训练参与方为例进行示意, 不构成对本公开实施方式的限制。 本公开 中, S202中将多个子模型进行拼接这一步骤可包括: 获取模型元信息, 该 模型元信息可包括具 有发送节点的子模型的发送节点和与该发送节点具有连接关系 的其他子模型的接收节点之间的连接关系信息; 根据模型元信息, 将与该发送节点相 连的该子模型的计算节点和与该接收节点相连的其他子模型的计算节点进行连接, 以 将多个子模型进行拼接。 模型元信 息可以指的是用于描述模型的信息, 可包括节点之间的连接关系信息。 当本公开提供的模型处理方法应用于图 4所示的模型处理系统时, 模型优化平台 401 可从模型元信息存储平台 404获取模型元信息, 并根据该模型元信息, 将与发送节点 相连的子模型的计算节点和与该接收节点相连的其他子模型的计算节点进行连接。 这样, 原本需要通过发送节点和接收节点转发数据的两个计算节点之间可以直接 传输数据。 由推理服务执行方通过该目标模型进行推理服务, 在推理服务执行方本地 即可完成整个推理服务的过程, 无需再进行数据的远程通信, 有效避免由于网络路由 等因素的影响导致的远程传输不稳定的问题, 保证推理服务的正常进行, 从而提升推 理服务的可靠性。 下面 介绍本公开中确定推理服务执行方以及通过 目标模型得到推理结果的几种 示例性实施方式。 在一可选实施方式中, 多个子模型与多个模型训练参与方一一对应, 每一模型训 练参与方均具有自身的模型输入数据。 推理服务执行方可以为多个模型训练参与方之 一, 推理服务执行方可根据自身的模型输入数据通过目标模型得到推理结果。 在该实施方式中, 推理服务执行方可以是多个模型训练参与方中的任一者。 图 5是根据一些示例性实施例示出的一种推理服务执行方根据自身的模型输入数 据通过目标模型得到推理结果的示意图。 图 5以推理服务执行方为模型训练参与方 1为例。 如图 5所示, 模型训练参与方 1获取到 目标模型之后, 可根据自身的模型输入数据 X、 Y、 Z, 通过目标模型得到推 理结果。 示例地 , 推理服务执行方可以是多个模型训练参与方中需要推理结果的一方, 即 推理结果需求方。 例如模型训练参与方 1需要最终的推理结果, 可由模型训练参与方 1作为推理服务执行方, 获取目标模型并进行推理服务。 或者 , 如果推理服务执行方不为推理结果需求方, 那么推理服务执行方可将推理 结果发送给推理结果需求方。 例如模型训练参与方 2需要推理结果, 模型训练参与方 1可将推理结果发送给模型训练参与方 2。 通过上述方案 , 其他模型训练参与方可以不将模型输入数据传输给推理服务执行 方, 推理服务执行方可根据自身的数据进行推理服务, 通信开销较小。 而且, 由推理 服务执行方完成推理服务, 无需再进行多个模型训练参与方之间远程通信传输数据的 过程, 在降低通信开销的同时, 提升推理服务的稳定性。 在另一可选实施方式中, 多个子模型与多个模型训练参与方一一对应, 每一模型 训练参与方均具有自身的模型输入数据, 推理服务执行方为多个模型训练参与方之一。 推理服务执行方可通过如下方式得到推理结果: 接收除推理服务执行方外的其他 模型训练参与方发送的加密的模型输入数据; 根据推理服务执行方自身的模型输入数 据、 以及其他模型训练参与方的加密的模型输入数据, 通过目标模型得到推理结果。 为保护数据隐私, 保证数据安全, 其他模型训练参与方可对模型输入数据进行加 密处理。 将加密的模型输入数据发送给推理服务执行方, 对于加密方式本公开不做具 体限制。 推理服务执行方可根据自身的模型输入数据、 以及其他模型训练参与方的加 密的模型输入数据, 通过目标模型得到推理结果, 即根据每一子模型的模型输入数据 进行推理服务。 可选地 , 推理服务执行方可以为需要接收其他模型训练参与方的加密的模型输入 数据的数据量最小的模型训练参与方。 例如, 推理服务执行方通过如下方式确定: 确 定各模型训练参与方进行推理 服务所需的其他模型训练参与方发送的加密的模型输 入数据的数据量; 将所需数据量最小的模型训练参与方确定为所述推理服务执行方。 示例地 , 如果由模型训练参与方 1进行推理服务, 模型训练参与方 2需要将子模 型 B的模型输入数据 M、 N发送至模型训练参与方 1; 如果由模型训练参与方 2进行 推理服务, 模型训练参与方 1需要将子模型 A的模型输入数据 X、 Y、 Z发送至模型 训练参与方 2。如果模型输入数据 M、 N的数据量小于模型输入数据 X、 Y、 Z的数据 量, 则模型训练参与方 1需要接收的模型输入数据的数据量最小, 可由模型训练参与 方 1作为推理服务执行方。 通过上述方案 , 如果推理服务执行方根据每一子模型的模型输入数据进行推理服 务, 可将需要接收的模型输入数据的数据量最小的模型训练参与方作为推理服务执行 方, 可以在一定程度上减少通信开销。 并且, 由推理服务执行方通过目标模型进行计 算, 无需再进行多个模型训练参与方之间远程通信传输数据的过程, 提高推理服务过 程的稳定性。 在又一可选实施方式中, 多个子模型与多个模型训练参与方一一对应, 每一模型 训练参与方均具有自身的模型输入数据, 推理服务执行方不为模型训练参与方; 推理 服务执行方可通过如下方式得到推理结果: 分别接收各个模型训练参与方发送的加密的模型输入数据; 根据各个模型训练参 与方的加密的模型输入数据, 通过目标模型得到推理结果。 在该实施方式中, 推理服务执行方可以不为模型训练参与方, 例如可以是预先设 置的云服务器。 各个模型训练参与方可将自身的模型输入数据以加密的形式发送给推 理服务执行方。 推理服务执行方获取到目标模型之后, 可根据各个模型训练参与方的 加密的模型输入数据, 通过目标模型得到推理结果。 通过上述方案 , 由推理服务执行方通过目标模型得到推理结果, 不需要再进行数 据的远程通信, 有效避免由于网络路由等因素的影响导致的远程传输不稳定的问题, 保证推理服务的正常进行, 从而提升推理服务的稳定性和可靠性。 图 6是根据一些示例性实施例示出的一种模型处理方法的流程图。 该方法可应用于推理服务执行方,即推理服务执行方的服务器设备。如图 6所示, 该方法可包括 S501和 S502o 在 S501 中, 推理服务执行方发送针对目标模型的模型获取请求。 所述目标模型 是将所述多个子模型进行拼接得到的。 推理服务执行方可 向模型存储平台发送模型获取请求, 也可向包括拼接模块的模 型处理装置发送模型获取请求, 本公开对此不进行限定。 在 S502中, 推理服务执行方接收目标模型, 并通过该目标模型得到推理结果。 通过上述技术方案, 推理服务执行方可发送针对目标模型的模型获取请求。 目标 模型是将多个子模型进行拼接得到的, 推理服务执行方可通过目标模型得到推理结果 o 由于将多个子模型拼接到一起得到目标模型, 推理服务执行方在通过目标模型得 到推理结果时, 在推理服务执行方本地即可完成整个推理服务的过程。 这样, 无需再 进行数据的远程通信, 不但能降低通信开销, 还能有效避免由于网络路由等因素的影 响导致的远程传输不稳定的问题, 保证推理服务的正常进行, 从而提升推理服务的可 靠性。 可选地 , 多个子模型与多个模型训练参与方一一对应, 每一所述模型训练参与方 均具有自身的模型输入数据, 推理服务执行方为多个所述模型训练参与方之一。 S502 中通过所述目标模型得到推理结果的步骤可包括: 根据所述推理服务执行方自身的模 型输入数据, 通过所述目标模型得到所述推理结果。 可选地 , 多个子模型与多个模型训练参与方一一对应, 每一所述模型训练参与方 均具有自身的模型输入数据, 推理服务执行方为多个所述模型训练参与方之一。 S502 中通过所述目标模型得到推理结果的步骤可包括: 接收除所述推理服务执行方外的其 他模型训练参与方发送的加密的模型输入数据; 根据所述推理服务执行方自身的模型 输入数据、 以及所述其他模型训练参与方的加密的模型输入数据, 通过所述目标模型 得到所述推理结果。 可选地 , 所述推理服务执行方为需要接收其他模型训练参与方的加密的模型输入 数据的数据量最小的模型训练参与方。 例如, 推理服务执行方通过如下方式确定: 确 定各模型训练参与方进行推理 服务所需的其他模型训练参与方发送的加密的模型输 入数据的数据量; 将所需数据量最小的模型训练参与方确定为所述推理服务执行方。 可选地 , 多个子模型与多个模型训练参与方一一对应, 每一所述模型训练参与方 均具有自身的模型输入数据,推理服务执行方不为所述模型训练参与方。 S502中通过 所述目标模型得到推理结果的步骤可包括: 分别接收各个所述模型训练参与方发送的 加密的模型输入数据; 根据各个所述模型训练参与方的加密的模型输入数据, 通过所 述目标模型得到所述推理结果。 关于上述实施例中应用于推理服务执行方的方法, 各个步骤执行操作的具体方式 已经在有关应用于模型处理系统 或包括拼接模块的模型处理装置的方法的实施例中 进行了详细描述, 此处将不做详细阐述说明。 本公开还提供一种模型处理系统, 如图 4所示的模型处理系统, 该系统可包括模 型优化平台、 模型存储平台; 所述模型优化平 台用于获取多个子模型, 将所述多个子模型进行拼接, 得到目标 模型, 并将所述目标模型发送至所述模型存储平台; 所述模型存储平台用于在接收到 推理服务执行方发送的针对所述目标模型的模型获取请求的情况下, 将所述目标模型 发送至所述推理服务执行方, 以使所述推理服务执行方通过所述目标模型得到推理结 果 o 可选地 , 所述模型优化平台用于获取模型元信息, 所述模型元信息包括具有发送 节点的子模型的发送节点和 与所述发送节点具有连接关系的其他子模型的接收节点 之间的连接关系信息; 所述模型优化平台用于根据所述模型元信息, 将与所述发送节 点相连的所述子模型 的计算节点和与所述接收节点相连的所述其他子模型的计算节 点进行连接, 以将所述多个子模型进行拼接。 关于上述实施例中的系统, 其中各个模块执行操作的具体方式已经在有关该方法 的实施例中进行了详细描述, 此处将不做详细阐述说明。 图 7是根据一些示例性实施例示出的一种模型处理装置的框图。 如 图 7所示, 该模型处理装置 600可包括: 获取模块 601, 被配置成用于获取多 个子模型;拼接模块 602,被配置成用于将所述多个子模型进行拼接,得到目标模型; 目标模型发送模块 603, 被配置成用于在接收到推理服务执行方发送的针对所述目标 模型的模型获取请求的情况下, 将所述目标模型发送至所述推理服务执行方, 以使所 述推理服务执行方通过所述目标模型得到推理结果。 可选地 , 所述拼接模块 602, 可包括: 获取子模块, 被配置成用于获取模型元信 息, 所述模型元信息包括具有发送节点的子模型的发送节点和与所述发送节点具有连 接关系的其他子模型的接收节点之间的连接关系信息; 拼接子模块, 被配置成用于根 据所述模型元信息, 将与所述发送节点相连的所述子模型的计算节点和与所述接收节 点相连的所述其他子模型的计算节点进行连接, 以将所述多个子模型进行拼接。 可选地 , 多个子模型与多个模型训练参与方一一对应, 每一所述模型训练参与方 均具有自身的模型输入数据, 所述推理服务执行方为多个所述模型训练参与方之一, 所述推理服务执行方根据自身的模型输入数据通过所述 目标模型得到所述推理结果。 可选地 , 多个子模型与多个模型训练参与方一一对应, 每一所述模型训练参与方 均具有自身的模型输入数据, 所述推理服务执行方为多个所述模型训练参与方之一; 所述推理服务执行方通过如下方式得到所述推理结果: 接收除所述推理服务执行方外 的其他模型训练参与方发送的加密的模型输入数据; 根据所述推理服务执行方自身的 模型输入数据、 以及所述其他模型训练参与方的加密的模型输入数据, 通过所述目标 模型得到所述推理结果。 可选地 , 所述推理服务执行方为需要接收其他模型训练参与方的加密的模型输入 数据的数据量最小的模型训练参与方。 例如, 推理服务执行方通过如下方式确定: 确 定各模型训练参与方进行推理 服务所需的其他模型训练参与方发送的加密的模型输 入数据的数据量; 将所需数据量最小的模型训练参与方确定为所述推理服务执行方。 可选地 , 多个子模型与多个模型训练参与方一一对应, 每一所述模型训练参与方 均具有自身的模型输入数据, 所述推理服务执行方不为所述模型训练参与方; 所述推 理服务执行方通过如下方式得到所述推理结果: 分别接收各个所述模型训练参与方发 送的加密的模型输入数据; 根据各个所述模型训练参与方的加密的模型输入数据, 通 过所述目标模型得到所述推理结果。 图 8是根据一些示例性实施例示出的一种模型处理装置的框图。 该模型处理装置 700可应用于推理服务执行方。如图 8所示,该模型处理装置 700 可包括: 获取请求发送模块 701, 被配置成用于发送针对目标模型的模型获取请求, 其中, 所述目标模型是将所述多个子模型进行拼接得到的; 推理模块 702, 被配置成 用于接收所述目标模型, 并通过所述目标模型得到推理结果。 可选地 , 多个子模型与多个模型训练参与方一一对应, 每一所述模型训练参与方 均具有自身的模型输入数据, 推理服务执行方为多个所述模型训练参与方之一; 所述 推理模块 702, 可包括: 第一推理子模块, 被配置成用于根据所述推理服务执行方自 身的模型输入数据, 通过所述目标模型得到所述推理结果。 可选地 , 多个子模型与多个模型训练参与方一一对应, 每一所述模型训练参与方 均具有自身的模型输入数据, 推理服务执行方为多个所述模型训练参与方之一; 所述 推理模块 702, 可包括: 第一接收子模块, 被配置成用于接收除所述推理服务执行方 外的其他模型训练参与方发送的加密的模型输入数据; 第二推理子模块, 被配置成用 于根据所述推理服务执行方自身的模型输入数据、 以及所述其他模型训练参与方的加 密的模型输入数据, 通过所述目标模型得到所述推理结果。 可选地 , 所述推理服务执行方为需要接收其他模型训练参与方的加密的模型输入 数据的数据量最小的模型训练参与方。 例如, 推理服务执行方通过如下方式确定: 确 定各模型训练参与方进行推理 服务所需的其他模型训练参与方发送的加密的模型输 入数据的数据量; 将所需数据量最小的模型训练参与方确定为所述推理服务执行方。 可选地 , 多个子模型与多个模型训练参与方一一对应, 每一所述模型训练参与方 均具有自身的模型输入数据, 推理服务执行方不为所述模型训练参与方; 所述推理模 块 702, 可包括: 第二接收子模块, 被配置成用于分别接收各个所述模型训练参与方 发送的加密的模型输入数据; 第三推理子模块, 被配置成用于根据各个所述模型训练 参与方的加密的模型输入数据, 通过所述目标模型得到所述推理结果。 下面参考 图 9, 其示出了适于用来实现本公开实施例的电子设备 800的结构示意 图。 本公开实施例中的终端设备可以包括但不限于诸如移动电话、 笔记本电脑、 数字 广播接收器、 PDA(个人数字助理) 、 PAD(平板电脑) 、 PMP(便携式多媒体播放 器) 、 车载终端 (例如车载导航终端) 等等的移动终端以及诸如数字 TV、 台式计算 机等等的固定终端。 图 9示出的电子设备仅仅是一个示例, 不应对本公开实施例的功 能和使用范围带来任何限制。 如 图 9所示,电子设备 800可以包括处理装置(例如中央处理器、图形处理器等) 801, 其可以根据存储在只读存储器 (ROM) 802中的程序或者从存储装置 808加载 到随机访问存储器 (RAM) 803中的程序而执行各种适当的动作和处理。在 RAM 803 中, 还存储有电子设备 800操作所需的各种程序和数据。 处理装置 801、 ROM 802以 及 RAM 803通过总线 804彼此相连。 输入 /输出 (I/O) 接口 805也连接至总线 804。 通常 , 以下装置可以连接至 I/O接口 805: 包括例如触摸屏、 触摸板、 键盘、 鼠 标、摄像头、麦克风、加速度计、陀螺仪等的输入装置 806;包括例如液晶显示器(LCD)、 扬声器、 振动器等的输出装置 807; 包括例如磁带、 硬盘等的存储装置 808; 以及通 信装置 809 o 通信装置 809可以允许电子设备 800与其他设备进行无线或有线通信以 交换数据。 虽然图 9示出了具有各种装置的电子设备 800, 但是应理解的是, 并不要 求实施或具备所有示出的装置。 可以替代地实施或具备更多或更少的装置。 特别地 , 根据本公开的实施例, 上文参考流程图描述的过程可以被实现为计算机 软件程序。 例如, 本公开的实施例包括一种计算机程序产品, 其包括承载在非暂态计 算机可读介质上的计算机程序, 该计算机程序包含用于执行流程图所示的方法的程序 代码。 在这样的实施例中, 该计算机程序可以通过通信装置 809从网络上被下载和安 装, 或者从存储装置 808被安装, 或者从 ROM 802被安装。 在该计算机程序被处理 装置 801执行时, 执行本公开实施例的方法中限定的上述功能。 需要说 明的是, 本公开上述的计算机可读介质可以是计算机可读信号介质或者计 算机可读存储介质或者是上述两者的任意组合。 计算机可读存储介质例如可以是一一 但不限于一一电、 磁、 光、 电磁、 红外线、 或半导体的系统、 装置或器件, 或者任意 以上的组合。 计算机可读存储介质的更具体的例子可以包括但不限于: 具有一个或多 个导线的电连接、 便携式计算机磁盘、 硬盘、 随机访问存储器(RAM) 、 只读存储器 (ROM) 、 可擦式可编程只读存储器(EPROM或闪存) 、 光纤、 便携式紧凑磁盘只 读存储器 (CD-ROM) 、 光存储器件、 磁存储器件、 或者上述的任意合适的组合。 在 本公开中, 计算机可读存储介质可以是任何包含或存储程序的有形介质, 该程序可以 被指令执行系统、 装置或者器件使用或者与其结合使用。 而在本公开中, 计算机可读 信号介质可以包括在基带中或者作为载波一部分传播的数据信号, 其中承载了计算机 可读的程序代码。 这种传播的数据信号可以采用多种形式, 包括但不限于电磁信号、 光信号或上述的任意合适的组合。 计算机可读信号介质还可以是计算机可读存储介质 以外的任何计算机可读介质, 该计算机可读信号介质可以发送、 传播或者传输用于由 指令执行系统、 装置或者器件使用或者与其结合使用的程序。 计算机可读介质上包含 的程序代码可以用任何适当的介质传输, 包括但不限于: 电线、 光缆、 RF(射频)等 等, 或者上述的任意合适的组合。 在一些实施方式 中, 客户端、 服务器可以利用诸如 HTTP (HyperText Transfer Protocol, 超文本传输协议) 之类的任何当前已知或未来研发的网络协议进行通信, 并且可以与任意形式或介质的数字数据通信 (例如, 通信网络) 互连。 通信网络的示 例包括局域网 ( “LAN” ) , 广域网 ( “WAN” ) , 网际网 (例如, 互联网) 以及端 对端网络 (例如, ad hoc端对端网络) , 以及任何当前已知或未来研发的网络。 上述计算机可读介质可以是上述电子设备中所包含的; 也可以是单独存在, 而未 装配入该电子设备中。 上述计算机可读介质承载有一个或者多个程序, 当上述一个或者多个程序被该电 子设备执行时, 使得该电子设备: 获取多个子模型; 将所述多个子模型进行拼接, 得 到目标模型; 在接收到推理服务执行方发送的针对所述目标模型的模型获取请求的情 况下, 将所述目标模型发送至所述推理服务执行方, 以使所述推理服务执行方通过所 述目标模型得到推理结果。 或者 , 上述计算机可读介质承载有一个或者多个程序, 当上述一个或者多个程序 被该电子设备执行时, 使得该电子设备: 发送针对目标模型的模型获取请求, 其中, 所述目标模型是将所述多个子模型进行拼接得到的; 接收所述目标模型, 并通过所述 目标模型得到推理结果。 可 以以一种或多种程序设计语言或其组合来编写用于执行本公开 的操作的计算 机程序代码, 上述程序设计语言包括但不限于面向对象的程序设计语言一诸如 Java、 Smalltalk. C++, 还包括常规的过程式程序设计语言一一诸如 “C”语言或类似的程 序设计语言。程序代码可以完全地在用户计算机上执行、部分地在用户计算机上执行、 作为一个独立的软件包执行、 部分在用户计算机上部分在远程计算机上执行、 或者完 全在远程计算机或服务器上执行。 在涉及远程计算机的情形中, 远程计算机可以通过 任意种类的网络一一包括局域网 (LAN) 或广域网 (WAN) — —连接到用户计算机, 或者, 可以连接到外部计算机 (例如利用因特网服务提供商来通过因特网连接) 。 附图中的流程图和框图, 图示了按照本公开各种实施例的系统、 方法和计算机程 序产品的可能实现的体系架构、 功能和操作。 在这点上, 流程图或框图中的每个方框 可以代表一个模块、 程序段、 或代码的一部分, 该模块、 程序段、 或代码的一部分包 含一个或多个用于实现规定的逻辑功能的可执行指令。 也应当注意, 在有些作为替换 的实现中, 方框中所标注的功能也可以以不同于附图中所标注的顺序发生。 例如, 两 个接连地表示的方框实际上可以基本并行地执行, 它们有时也可以按相反的顺序执行, 这依所涉及的功能而定。 也要注意的是, 框图和 /或流程图中的每个方框、 以及框图和 /或流程图中的方框的组合,可以用执行规定的功能或操作的专用的基于硬件的系统来 实现, 或者可以用专用硬件与计算机指令的组合来实现。 描述于本 公开实施例中所涉及到的模块可以通过软件的方式实现, 也可以通过硬 件的方式来实现。 其中, 模块的名称在某种情况下并不构成对该模块本身的限定, 例 如, 拼接模块还可以被描述为 “子模型拼接模块” 。 本文 中以上描述的功能可以至少部分地由一个或多个硬件逻辑部件来执行。 例如, 非限制性地,可以使用的示范类型的硬件逻辑部件包括:现场可编程门阵列(FPGA)、 专用集成电路 (ASIC) 、 专用标准产品 (ASSP) 、 片上系统 (SOC) 、 复杂可编程 逻辑设备 (CPLD) 等等。 在本公开 的上下文中, 机器可读介质可以是有形的介质, 其可以包含或存储以供 指令执行系统、 装置或设备使用或与指令执行系统、 装置或设备结合地使用的程序。 机器可读介质可以是机器可读信号介质或机器可读储存介质。 机器可读介质可以包括 但不限于电子的、 磁性的、 光学的、 电磁的、 红外的、 或半导体系统、 装置或设备, 或者上述内容的任何合适组合。 机器可读存储介质的更具体示例会包括基于一个或多 个线的电气连接、便携式计算机盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、 可擦除可编程只读存储器 (EPROM或快闪存储器) 、 光纤、 便捷式紧凑盘只读存储 器 (CD-ROM) 、 光学储存设备、 磁储存设备、 或上述内容的任何合适组合。 根据本公开 的一个或多个实施例, 示例 1提供了一种模型处理方法, 所述方法包 括: 获取多个子模型; 将所述多个子模型进行拼接, 得到目标模型; 在接收到推理服 务执行方发送的针对所述目标模型的模型获取请求的情况下, 将所述目标模型发送至 所述推理服务执行方, 以使所述推理服务执行方通过所述目标模型得到推理结果。 根据本公开 的一个或多个实施例, 示例 2提供了示例 1的方法, 所述将所述多个 子模型进行拼接, 包括: 获取模型元信息, 所述模型元信息包括具有发送节点的子模 型的发送节点和与所述发送节 点具有连接关系的其他子模型的接收节点之间的连接 关系信息; 根据所述模型元信息, 将与所述发送节点相连的所述子模型的计算节点和 与所述接收节点相连的所述其他子模型的计算节点进行连接, 以将所述多个子模型进 行拼接。 根据本公开 的一个或多个实施例, 示例 3提供了示例 1的方法, 多个子模型与多 个模型训练参与方一一对应, 每一所述模型训练参与方均具有自身的模型输入数据, 所述推理服务执行方为多个所述模型训练参与方之一, 所述推理服务执行方根据自身 的模型输入数据通过所述目标模型得到所述推理结果。 根据本公开 的一个或多个实施例, 示例 4提供了示例 1的方法, 多个子模型与多 个模型训练参与方一一对应, 每一所述模型训练参与方均具有自身的模型输入数据, 所述推理服务执行方为多个所述模型训练参与方之一; 所述推理服务执行方通过如下 方式得到所述推理结果: 接收除所述推理服务执行方外的其他模型训练参与方发送的 加密的模型输入数据; 根据所述推理服务执行方自身的模型输入数据、 以及所述其他 模型训练参与方的加密的模型输入数据, 通过所述目标模型得到所述推理结果。 根据本公开 的一个或多个实施例, 示例 5提供了示例 4的方法, 所述推理服务执 行方为需要接收其他模 型训练参与方的加密的模型输入数据的数据量最小 的模型训 练参与方。 例如, 推理服务执行方通过如下方式确定: 确定各模型训练参与方进行推 理服务所需的其他模型训练参与方发送的加密的模型输入数据的数据量; 将所需数据 量最小的模型训练参与方确定为所述推理服务执行方。 根据本公开 的一个或多个实施例, 示例 6提供了示例 1的方法, 多个子模型与多 个模型训练参与方一一对应, 每一所述模型训练参与方均具有自身的模型输入数据, 所述推理服务执行方不为所述模型训练参与方; 所述推理服务执行方通过如下方式得 到所述推理结果: 分别接收各个所述模型训练参与方发送的加密的模型输入数据; 根 据各个所述模型训练参与方的加密的模型输入数据, 通过所述目标模型得到所述推理 结果。 根据本公开 的一个或多个实施例, 示例 7提供了一种模型处理方法, 所述方法包 括: 推理服务执行方发送针对目标模型的模型获取请求, 其中, 所述目标模型是将所 述多个子模型进行拼接得到的; 推理服务执行方接收所述目标模型, 并通过所述目标 模型得到推理结果。 根据本公开 的一个或多个实施例, 示例 8提供了示例 7的方法, 多个子模型与多 个模型训练参与方一一对应, 每一所述模型训练参与方均具有自身的模型输入数据, 推理服务执行方为多个所述模型训练参与方之一; 所述通过所述目标模型得到推理结 果, 包括: 根据所述推理服务执行方自身的模型输入数据, 通过所述目标模型得到所 述推理结果。 根据本公开 的一个或多个实施例, 示例 9提供了示例 7的方法, 多个子模型与多 个模型训练参与方一一对应, 每一所述模型训练参与方均具有自身的模型输入数据, 推理服务执行方为多个所述模型训练参与方之一; 所述通过所述目标模型得到推理结 果, 包括: 接收除所述推理服务执行方外的其他模型训练参与方发送的加密的模型输 入数据; 根据所述推理服务执行方自身的模型输入数据、 以及所述其他模型训练参与 方的加密的模型输入数据, 通过所述目标模型得到所述推理结果。 根据本公开 的一个或多个实施例, 示例 10提供了示例 9的方法, 所述推理服务 执行方为需要接收其他模 型训练参与方的加密的模型输入数据的数据量最小 的模型 训练参与方。 例如, 推理服务执行方通过如下方式确定: 确定各模型训练参与方进行 推理服务所需的其他模型训练参与方发送的加密的模型输入数据的数据量; 将所需数 据量最小的模型训练参与方确定为所述推理服务执行方。 根据本公开 的一个或多个实施例, 示例 11提供了示例 7的方法, 多个子模型与 多个模型训练参与方一一对应, 每一所述模型训练参与方均具有自身的模型输入数据, 推理服务执行方不为所述模型训练参与方; 所述通过所述目标模型得到推理结果, 包 括: 分别接收各个所述模型训练参与方发送的加密的模型输入数据; 根据各个所述模 型训练参与方的加密的模型输入数据, 通过所述目标模型得到所述推理结果。 根据本公开 的一个或多个实施例, 示例 12提供了一种模型处理系统, 所述系统 包括模型优化平台、 模型存储平台; 所述模型优化平台用于获取多个子模型, 将所述 多个子模型进行拼接, 得到目标模型, 并将所述目标模型发送至所述模型存储平台; 所述模型存储平台用于在 接收到推理服务执行方发送的针对所述目标模型的模型获 取请求的情况下, 将所述目标模型发送至所述推理服务执行方, 以使所述推理服务执 行方通过所述目标模型得到推理结果。 根据本公开 的一个或多个实施例, 示例 13提供了示例 12的系统, 所述模型优化 平台用于获取模型元信息, 所述模型元信息包括具有发送节点的子模型的发送节点和 与所述发送节点具有连接关系的其他子模型的接收节点之间的连接关系信息; 所述模 型优化平台用于根据所述模型元信息, 将与所述发送节点相连的所述子模型的计算节 点和与所述接收节点相连的所述其他子模型的计算节点进行连接, 以将所述多个子模 型进行拼接。 根据本公开 的一个或多个实施例, 示例 14提供了一种模型处理装置, 所述装置 包括: 获取模块, 被配置成用于获取多个子模型; 拼接模块, 被配置成用于将所述多 个子模型进行拼接, 得到目标模型; 目标模型发送模块, 被配置成用于在接收到推理 服务执行方发送的针对所述目标模型的模型获取请求的情况下, 将所述目标模型发送 至所述推理服务执行方, 以使所述推理服务执行方通过所述目标模型得到推理结果。 根据本公开 的一个或多个实施例, 示例 15提供了一种模型处理装置, 所述装置 包括: 获取请求发送模块, 被配置成用于发送针对目标模型的模型获取请求, 其中, 所述目标模型是将所述多个子模型进行拼接得到的; 推理模块, 被配置成用于接收所 述目标模型, 并通过所述目标模型得到推理结果。 例如, 模型处理装置可以设置在推 理服务执行方。 根据本公开 的一个或多个实施例, 示例 16 提供了一种计算机可读介质, 其上存 储有计算机程序, 该程序被处理装置执行时实现示例 1-6中任一项所述方法的步骤。 根据本公开 的一个或多个实施例, 示例 17 提供了一种计算机可读介质, 其上存 储有计算机程序, 该程序被处理装置执行时实现示例 7-11中任一项所述方法的步骤。 根据本公开 的一个或多个实施例, 示例 18 提供了一种电子设备, 包括: 存储装 置,其上存储有计算机程序;处理装置,用于执行所述存储装置中的所述计算机程序, 以实现示例 1-6中任一项所述方法的步骤。 根据本公开 的一个或多个实施例, 示例 19 提供了一种电子设备, 包括: 存储装 置,其上存储有计算机程序;处理装置,用于执行所述存储装置中的所述计算机程序, 以实现示例 7-11中任一项所述方法的步骤。 以上描述仅为本公开的较佳实施例以及对所运用技术原理的说明。 本领域技术人 员应当理解, 本公开中所涉及的公开范围, 并不限于上述技术特征的特定组合而成的 技术方案, 同时也应涵盖在不脱离上述公开构思的情况下, 由上述技术特征或其等同 特征进行任意组合而形成的其它技术方案。 例如上述特征与本公开中公开的 (但不限 于) 具有类似功能的技术特征进行互相替换而形成的技术方案。 此外 , 虽然采用特定次序描绘了各操作, 但是这不应当理解为要求这些操作以所 示出的特定次序或以顺序次序执行来执行。 在一定环境下, 多任务和并行处理可能是 有利的。 同样地, 虽然在上面论述中包含了若干具体实现细节, 但是这些不应当被解 释为对本公开的范围的限制。 在单独的实施例的上下文中描述的某些特征还可以组合 地实现在单个实施例中。 相反地, 在单个实施例的上下文中描述的各种特征也可以单 独地或以任何合适的子组合的方式实现在多个实施例中。 尽管 已经采用特定于结构特征和 /或方法逻辑动作的语言描述了本主题,但是应当 理解所附权利要求书中所限定的主题未必局 限于上面描述的特定特征或动作。 相反, 上面所描述的特定特征和动作仅仅是实现权利要求书的示例形式。 关于上述实施例中 的装置, 其中各个模块执行操作的具体方式已经在有关该方法的实施例中进行了详细 描述, 此处将不做详细阐述说明。 MODEL PROCESSING METHODS, SYSTEMS, APPARATUS, MEDIA AND ELECTRONIC DEVICES CROSS-REFERENCE TO RELATED APPLICATIONS This application is based on an application with CN application number 202011298789.5 and an application date of November 18, 2020, and claims its priority, the CN application The disclosure of is hereby incorporated into this application in its entirety. FIELD OF THE DISCLOSURE The present disclosure relates to the field of computer technology, and in particular, to a model processing method, system, apparatus, medium, and electronic device. 2. Description of the Related Art Federated machine learning, also known as federated learning and federated learning, has been widely used in the field of machine learning. Federated learning can solve data silos and data privacy issues. Under the requirements of user privacy protection and data security, the federated learning system can effectively help multiple institutions to complete the joint training of models. The federated learning model is usually composed of multiple sub-models. How to ensure the reliability of the inference service is an important issue when the inference service is performed through the federated learning model. SUMMARY This Summary section is provided to introduce in a simplified form concepts that are described in detail in the Detailed Description section that follows. This summary section is not intended to identify key features or essential features of the claimed technical solution, nor is it intended to limit the scope of the claimed technical solution. In a first aspect, the present disclosure provides a model processing method, the method includes: acquiring multiple sub-models; splicing the multiple sub-models to obtain a target model; after receiving a message sent by an inference service executor for the target model In the case of the model obtaining request, the target model is sent to the inference service executor, so that the inference service executor obtains an inference result through the target model. In a second aspect, the present disclosure provides a model processing method. The method includes: an inference service executor sends a model acquisition request for a target model, where the target model is obtained by splicing multiple sub-models; the inference service executes The party receives the target model, and obtains an inference result through the target model. In a third aspect, the present disclosure provides a model processing system, the system includes a model optimization platform and a model storage platform; the model optimization platform is used for acquiring multiple sub-models, and splicing the multiple sub-models to obtain target model, and send the target model to the model storage platform; the model storage platform is configured to store the target model in the case of receiving a model acquisition request for the target model sent by the inference service executor The model is sent to the inference service executor, so that the inference service executor obtains an inference result through the target model. In a fourth aspect, the present disclosure provides a model processing device, the device comprising: an acquisition module configured to acquire multiple sub-models; and a splicing module configured to splicing the multiple sub-models to obtain a target model; a target model sending module, configured to send the target model to the inference service executor in the case of receiving a model acquisition request for the target model sent by the inference service executor, so that The inference service executor obtains an inference result through the target model. In a fifth aspect, the present disclosure provides a model processing apparatus, the apparatus comprising: an acquisition request sending module, configured to send a model acquisition request for a target model, wherein the target model is to convert the multiple sub-models obtained by splicing; an inference module, configured to receive the target model, and obtain an inference result through the target model. For example, the model processing apparatus may be set at the execution side of the inference service. In a sixth aspect, the present disclosure provides a computer-readable medium on which a computer program is stored, and when the program is executed by a processing apparatus, implements the steps of the method provided in the first aspect of the present disclosure. In a seventh aspect, the present disclosure provides a computer-readable medium on which a computer program is stored, and when the program is executed by a processing apparatus, implements the steps of the method provided in the second aspect of the present disclosure. In an eighth aspect, the present disclosure provides an electronic device, comprising: a storage device on which a computer program is stored; and a processing device for executing the computer program in the storage device, so as to implement the computer program provided in the first aspect of the present disclosure. the steps of the method. In a ninth aspect, the present disclosure provides an electronic device, comprising: a storage device on which a computer program is stored; and a processing device for executing the computer program in the storage device, so as to implement the computer program provided in the second aspect of the present disclosure. the steps of the method. In a tenth aspect, the present disclosure provides a computer program, comprising: instructions that, when executed by a processor, cause the processor to execute the model processing method according to any one of the foregoing embodiments. In an eleventh aspect, the present disclosure provides a computer program product comprising instructions that, when executed by a processor, cause the processor to execute the model processing method according to any one of the foregoing embodiments. Other features and advantages of the present disclosure will be described in detail in the detailed description that follows. BRIEF DESCRIPTION OF THE DRAWINGS The above and other features, advantages and aspects of various embodiments of the present disclosure will become more apparent with reference to the following detailed description in conjunction with the accompanying drawings. Throughout the drawings, the same or similar reference numbers refer to the same or similar elements. It should be understood that the drawings are schematic and that the originals and elements are not necessarily drawn to scale. In the drawings: FIG. 1 is a schematic diagram of a federated learning model in the related art. Fig. 2 is a flow chart of a model processing method according to some exemplary embodiments. Fig. 3 is a schematic diagram of a target model according to some exemplary embodiments. FIG. 4 is a schematic diagram of a model processing system according to some exemplary embodiments. Fig. 5 is a schematic diagram illustrating an inference service executor obtaining an inference result through a target model according to its own model input data, according to some exemplary embodiments. Fig. 6 is a flow chart of a model processing method according to some exemplary embodiments. Fig. 7 is a block diagram of a model processing apparatus according to some exemplary embodiments. Fig. 8 is a block diagram of a model processing apparatus according to some exemplary embodiments. FIG. 9 is a schematic structural diagram of an electronic device according to some exemplary embodiments. DETAILED DESCRIPTION Embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While certain embodiments of the present disclosure are shown in the drawings, it should be understood that the present disclosure may be embodied in various forms and should not be construed as limited to the embodiments set forth herein, but rather are provided for the purpose of A more thorough and complete understanding of the present disclosure. It should be understood that the drawings and embodiments of the present disclosure are only used for exemplary purposes, and are not intended to limit the protection scope of the present disclosure. It should be understood that the various steps described in the method embodiments of the present disclosure may be performed in different orders, and/or performed in parallel. Furthermore, method embodiments may include additional steps and/or omit performing the illustrated steps. The scope of the present disclosure is not limited in this regard. As used herein, the term "including" and variations thereof are open-ended inclusions, ie, "including but not limited to". The term "based on," is "based at least in part on". The term "one embodiment" means "at least one embodiment"; the term "another embodiment" means "at least one additional embodiment": the term "some embodiments" Represents "at least some embodiments". Relevant definitions of other terms will be given in the following description. It should be noted that concepts such as "first" and "second" mentioned in this disclosure are only used for different devices and modules. or units are not used to limit the order or phase of the functions performed by these devices, modules or units. interdependence. It should be noted that the modifications of "a" and "plurality" mentioned in the present disclosure are illustrative rather than restrictive, and those skilled in the art should understand that, unless the context clearly indicates otherwise, they should be understood as "one or a plurality of" The names of the messages or information exchanged between the multiple devices in the embodiments of the present disclosure are only used for illustrative purposes, and are not used to limit the scope of these messages or information. The federated learning system can combine the data of multiple data owners to train a common federated learning model. The federated learning model is trained by combining the data of multiple data owners, and the training data is more comprehensive. Therefore, the accuracy of the federated learning model is higher. A federated learning model is usually composed of multiple sub-models, and FIG. 1 is a schematic diagram of a federated learning model in the related art. As shown in Figure 1, the federated learning model includes sub-model A and sub-model B. For example, the sub-model A corresponds to the model training participant 1, and the model input data X, Y, and Z of the sub-model A are data owned by the model training participant 1. The sub-model B corresponds to the model training participant 2, and the model input data M and N of the sub-model B are data owned by the model training participant 2. When performing inference services through the federated learning model, each model training participant loads its own sub-models, that is, model training participant 1 loads sub-model A, and model training participant 2 loads sub-model B. As shown in FIG. 1 , the model training participant 1 performs computation through the sub-model A according to the model input data X, Y, and Z. Then model training participant 1 needs to remotely send the data to model training participant 2 through the sending node of sub-model A, so as to transmit the data to the receiving node of sub-model B. The model training participant 2 then obtains an inference result through the sub-model B according to the data received by the receiving node and the model input data M and N. In this way, when the inference service is performed through the federated learning model, remote communication is required between multiple model training participants to complete the entire inference service, that is, the sending node and the receiving node use remote communication to transmit data, and the communication overhead is large. Moreover, long-distance communication is easily affected by factors such as network routing, and is usually not stable enough and has low reliability, which makes the calculation process of the inference service not stable enough. For example, if the sending node cannot transmit the data to the receiving node in time due to network congestion when the data is remotely transmitted to the receiving node, the entire inference service will be affected. In order to solve the problems existing in the related art, the present disclosure provides a model processing method, system, apparatus, medium and electronic device. Fig. 2 is a flow chart of a model processing method according to some exemplary embodiments. As shown in FIG. 2, the method may include S201~S203o In S201, multiple sub-models are acquired. In S202, multiple sub-models are spliced to obtain a target model. Fig. 3 is a schematic diagram of a target model according to some exemplary embodiments. The target model shown in FIG. 3 may be obtained according to the federated learning model shown in FIG. 1 , and there is a connection relationship between the sending node of the sub-model A and the receiving node of the sub-model B. As shown in FIG. 3 , the computing node of the sub-model A connected to the sending node and the computing node of the sub-model B connected to the receiving node can be connected to obtain the target model. The target model is the overall full model obtained by splicing sub-model A and sub-model B together. It is worth noting that the present disclosure takes two sub-models as examples for illustration, which does not constitute a limitation on the implementation of the present disclosure. In practical applications, the number of sub-models may be multiple, which is not specifically limited in the present disclosure. . In S203, in the case of receiving the model acquisition request for the target model sent by the inference service executor, the target model is sent to the inference service executor, so that the inference service executor obtains the inference result through the target model. Inference service can refer to the process by which the server performs computations through the model based on the input data and obtains the result. For example, taking the prediction of the user's shopping intention as an example, according to the user's historical shopping behavior information, the user's current shopping intention can be inferred through a model, and then the user can be provided with an inference result that meets his shopping intention and needs. As another example, taking the prediction of the user's search intent as an example, according to the user's historical click behavior information, the user's current search intent can be inferred through the model, and then the user can be provided with an inference result that conforms to the user's search intent. In an optional implementation manner, one of the model training participants can be used as the inference service executor, load the target model, and obtain the inference result through the target model. The target model is obtained by splicing multiple sub-models, and the inference service executor can directly obtain the inference results through the overall target model, without each model training participant loading its own sub-models, and without the need for multiple model training participants. It can effectively avoid the problem of unstable long-distance communication. It is worth noting that, when it is mentioned in the present disclosure that the inference service executor performs operations of sending, receiving, and processing data, it may be understood that the inference service executor performs these operations through a server device. Through the above technical solution, a target model is obtained by splicing multiple sub-models, and the inference service executor can obtain an inference result through the target model. The target model is obtained by splicing multiple sub-models, and the inference service executor can directly obtain the inference result through the overall target model, without each model training participant loading its own sub-models separately. The entire inference service process can be completed locally on the inference service executor, without the need for remote communication between multiple model training participants to transmit data. In this way, not only the communication overhead can be reduced, but also the problem of unstable remote transmission caused by factors such as network routing can be effectively avoided, the normal operation of the inference service can be ensured, and the reliability of the inference service can be improved. In one embodiment, the model processing method shown in FIG. 2 may be applied to a model processing apparatus including a splicing module. The model processing device may be, for example, a cloud server, the acquisition module in the model processing device acquires multiple sub-models, and the splicing module splices the multiple sub-models to obtain the target model. In another embodiment, the model processing method shown in FIG. 2 can also be applied to a model processing system. FIG. 4 is a schematic diagram of a model processing system according to some exemplary embodiments. As shown in FIG. 4 , the model processing system may include a model optimization platform 401, a model storage platform 402, a model training platform 403, a model metadata storage platform 404, a model training participant 1, and a model training participant 2. The model training platform 403 is used to train each sub-model, such as sub-model A and sub-model B. The model meta information storage platform 404 may be used to store model related meta information. The model training platform 403 can send multiple sub-models to the model optimization platform 401. The model optimization platform 401 can be used to obtain the multiple sub-models sent by the model training platform 403, splicing the multiple sub-models to obtain the target model, and sending the target model to Model storage platform 402 o The inference service executor may send a model acquisition request for the target model to the model storage platform 402; the model storage platform 402 may send the target model to the inference service executor when receiving the request. The inference service executor may be one of model training participant 1 and model training participant 2, for example. FIG. 4 takes an example of including two model training participants for illustration, which does not constitute a limitation on the embodiments of the present disclosure. In the present disclosure, the step of splicing multiple sub-models in S202 may include: acquiring model meta information, where the model meta information may include a sending node having a sub-model of the sending node and other sub-models that have a connection relationship with the sending node The connection relationship information between the receiving nodes; according to the model meta information, connect the computing node of the sub-model connected to the sending node and the computing nodes of other sub-models connected to the receiving node to connect the multiple sub-models. stitching. Model meta-information may refer to information used to describe a model, and may include connection relationship information between nodes. When the model processing method provided by the present disclosure is applied to the model processing system shown in FIG. 4 , the model optimization platform 401 can obtain model meta information from the model meta information storage platform 404, and according to the model meta information, connect the The computing nodes of the sub-model are connected with the computing nodes of other sub-models connected to the receiving node. In this way, data can be directly transmitted between two computing nodes that originally need to forward data through the sending node and the receiving node. The inference service executor performs the inference service through the target model, and the inference service executor locally The entire inference service process can be completed without the need for remote communication of data, effectively avoiding the problem of unstable remote transmission caused by factors such as network routing, ensuring the normal operation of the inference service, and improving the reliability of the inference service. Several exemplary implementations of determining the inference service executor and obtaining inference results through the target model in the present disclosure are described below. In an optional embodiment, multiple sub-models correspond to multiple model training participants one-to-one, and each model training participant has its own model input data. The inference service executor can be one of multiple model training participants, and the inference service executor can obtain inference results through the target model according to its own model input data. In this embodiment, the inference service performer may be any of a plurality of model training participants. Fig. 5 is a schematic diagram illustrating an inference service executor obtaining an inference result through a target model according to its own model input data, according to some exemplary embodiments. Figure 5 takes the inference service executor as the model training participant 1 as an example. As shown in FIG. 5 , after the model training participant 1 obtains the target model, it can input data X, Y, and Z according to its own model, and obtain an inference result through the target model. For example, the inference service executor may be the party that needs the inference result among the multiple model training participants, that is, the inference result demander. For example, the model training participant 1 needs the final inference result, and the model training participant 1 can be used as the inference service executor to obtain the target model and perform the inference service. Or, if the inference service executor is not the inference result requester, the inference service executor may send the inference result to the inference result requester. For example, model training participant 2 needs an inference result, and model training participant 1 can send the inference result to model training participant 2 . Through the above solution, other model training participants may not transmit model input data to the inference service executor, and the inference service executor can perform inference services according to its own data, with low communication overhead. Moreover, the reasoning service is completed by the inference service executor, and there is no need to carry out the process of remote communication and data transmission between multiple model training participants, which reduces the communication overhead and improves the stability of the inference service. In another optional embodiment, multiple sub-models correspond to multiple model training participants one-to-one, each model training participant has its own model input data, and the inference service executor is one of the multiple model training participants . The inference service executor can obtain the inference results in the following ways: Receive encrypted model input data sent by other model training participants except the inference service executor; According to the inference service executor's own model input data and other model training participants The encrypted model input data is obtained, and the inference result is obtained through the target model. In order to protect data privacy and ensure data security, other model training participants can add model input data. encrypted processing. Send the encrypted model input data to the inference service executor, and the present disclosure does not specifically limit the encryption method. The inference service executor can use its own model input data and the encrypted model input data of other model training participants to obtain inference results through the target model, that is, to perform inference services based on the model input data of each sub-model. Optionally, the inference service executor may be the model training participant that needs to receive the encrypted model input data of other model training participants with the smallest amount of data. For example, the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor. For example, if the model training participant 1 performs the inference service, the model training participant 2 needs to send the model input data M and N of the sub-model B to the model training participant 1; if the model training participant 2 performs the inference service, Model training participant 1 needs to send the model input data X, Y, and Z of sub-model A to model training participant 2. If the data volume of the model input data M and N is smaller than the data volume of the model input data X, Y, and Z, the model training participant 1 needs to receive the minimum data volume of the model input data, which can be executed by the model training participant 1 as an inference service. square. Through the above solution, if the inference service executor performs the inference service according to the model input data of each sub-model, the model training participant with the smallest amount of model input data to be received can be used as the inference service executor, which can, to a certain extent, be used as the inference service executor. Reduce communication overhead. In addition, the inference service executor performs the calculation through the target model, and there is no need to carry out the process of remote communication and data transmission among multiple model training participants, thereby improving the stability of the inference service process. In yet another optional embodiment, multiple sub-models correspond to multiple model training participants one-to-one, each model training participant has its own model input data, and the inference service executor is not a model training participant; inference service The executor can obtain the inference result in the following ways: Receive the encrypted model input data sent by each model training participant respectively; Obtain the inference result through the target model according to the encrypted model input data of each model training participant. In this embodiment, the inference service executor may not be a model training participant, for example, it may be a preset cloud server. Each model training participant can send its own model input data to the inference service executor in encrypted form. After the inference service executor obtains the target model, it can train the participants' encrypted model input data according to each model, and obtain the inference result through the target model. Through the above scheme, the inference service executor obtains the inference result through the target model, and no longer needs to perform remote communication of data, which effectively avoids the problem of unstable remote transmission caused by factors such as network routing. Ensure the normal operation of the inference service, thereby improving the stability and reliability of the inference service. Fig. 6 is a flow chart of a model processing method according to some exemplary embodiments. The method can be applied to the inference service executor, that is, the server device of the inference service executor. As shown in FIG. 6, the method may include S501 and S502. In S501, the inference service executor sends a model acquisition request for the target model. The target model is obtained by splicing the multiple sub-models. The inference service executor may send a model acquisition request to the model storage platform, and may also send a model acquisition request to a model processing apparatus including a splicing module, which is not limited in the present disclosure. In S502, the inference service executor receives the target model, and obtains an inference result through the target model. Through the above technical solution, the inference service executor can send a model acquisition request for the target model. The target model is obtained by splicing multiple sub-models, and the inference service executor can obtain the inference result through the target model. Since the target model is obtained by splicing multiple sub-models together, the inference service executor obtains the inference result through the target model. The inference service executor can complete the entire inference service process locally. In this way, there is no need to perform remote communication of data, which not only reduces communication overhead, but also effectively avoids the problem of unstable remote transmission caused by factors such as network routing, ensures the normal operation of inference services, and improves the reliability of inference services. . Optionally, the multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is one of the multiple model training participants. The step of obtaining the inference result through the target model in S502 may include: inputting data according to the model of the inference service executor itself, and obtaining the inference result through the target model. Optionally, the multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is one of the multiple model training participants. The step of obtaining the inference result from the target model in S502 may include: receiving encrypted model input data sent by other model training participants except the inference service executor; and according to the model input data of the inference service executor itself , and the encrypted model input data of the other model training participants, and obtain the inference result through the target model. Optionally, the inference service executor is a model training participant that needs to receive the encrypted model input data of other model training participants with the smallest amount of data. For example, the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor. Optionally, the multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is not the model training participant. The step of obtaining the inference result through the target model in S502 may include: respectively receiving encrypted model input data sent by each of the model training participants; according to the encrypted model input data of each of the model training participants, The target model obtains the inference result. Regarding the method applied to the inference service executor in the foregoing embodiment, the specific manner in which each step performs operations has been described in detail in the embodiment of the method applied to the model processing system or the model processing apparatus including the splicing module, here A detailed explanation will not be given. The present disclosure also provides a model processing system, such as the model processing system shown in FIG. 4 , the system may include a model optimization platform and a model storage platform; the model optimization platform is used for acquiring multiple sub-models, and storing the multiple sub-models Splicing is performed to obtain a target model, and the target model is sent to the model storage platform; the model storage platform is configured to, upon receiving a model acquisition request for the target model sent by the inference service executor, Send the target model to the inference service executor, so that the inference service executor obtains an inference result through the target model. Optionally, the model optimization platform is used to obtain model meta information, and the model The meta-information includes connection relationship information between a sending node having a sub-model of the sending node and receiving nodes of other sub-models that have a connection relationship with the sending node; the model optimization platform is configured to, according to the model meta-information, The computing node of the sub-model connected with the sending node is connected with the computing nodes of the other sub-models connected with the receiving node, so as to splicing the multiple sub-models. Regarding the system in the above-mentioned embodiment, the specific manner in which each module performs operations has been described in detail in the embodiment of the method, and will not be described in detail here. Fig. 7 is a block diagram of a model processing apparatus according to some exemplary embodiments. As shown in FIG. 7 , the model processing apparatus 600 may include: an acquisition module 601, configured to acquire multiple sub-models; a splicing module 602, configured to splicing the multiple sub-models to obtain a target model; The target model sending module 603 is configured to send the target model to the inference service executor in the case of receiving a model acquisition request for the target model sent by the inference service executor, so that all The inference service executor obtains the inference result through the target model. Optionally, the splicing module 602 may include: an obtaining submodule configured to obtain model meta information, where the model meta information includes a sending node having a sub-model of a sending node and a sending node having a connection with the sending node the connection relationship information between the receiving nodes of other sub-models of the connection relationship; the splicing sub-module is configured to connect the computing node of the sub-model connected to the sending node and the computing node of the sub-model connected to the sending node according to the model meta information. The computing nodes of the other sub-models connected to the receiving node are connected to splicing the multiple sub-models. Optionally, multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is one of the multiple model training participants. First, the inference service executor obtains the inference result through the target model according to its own model input data. Optionally, multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is one of the multiple model training participants. 1. The inference service executor obtains the inference result in the following ways: Receive encrypted model input data sent by other model training participants except the inference service executor; According to the inference service executor's own model The input data and the encrypted model input data of the other model training participants are used to obtain the inference result through the target model. Optionally, the inference service executor is a model training participant that needs to receive the encrypted model input data of other model training participants with the smallest amount of data. For example, the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor. Optionally, multiple sub-models are in one-to-one correspondence with multiple model training participants, each model training participant has its own model input data, and the inference service executor is not the model training participant; The inference service executor obtains the inference result in the following ways: respectively receiving the encrypted model input data sent by each of the model training participants; according to the encrypted model input data of each of the model training participants, through the target The model obtains the inference result. Fig. 8 is a block diagram of a model processing apparatus according to some exemplary embodiments. The model processing apparatus 700 can be applied to an inference service executor. As shown in FIG. 8 , the model processing apparatus 700 may include: an acquisition request sending module 701, configured to send a model acquisition request for a target model, wherein the target model is obtained by splicing the multiple sub-models The inference module 702 is configured to receive the target model and obtain an inference result through the target model. Optionally, multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is one of the multiple model training participants; The inference module 702 may include: a first inference sub-module, configured to input data according to the model of the inference service executor itself, and obtain the inference result through the target model. Optionally, multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is one of the multiple model training participants; The inference module 702 may include: a first receiving sub-module, configured to receive encrypted model input data sent by other model training participants except the inference service executor; a second inference sub-module, configured by It is configured to obtain the inference result through the target model according to the model input data of the inference service executor itself and the encrypted model input data of the other model training participants. Optionally, the inference service executor is a model training participant that needs to receive the encrypted model input data of other model training participants with the smallest amount of data. For example, the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor. Optionally, multiple sub-models are in one-to-one correspondence with multiple model training participants, each model training participant has its own model input data, and the inference service executor is not the model training participant; the reasoning The module 702 may include: a second receiving sub-module, configured to respectively receive encrypted model input data sent by each of the model training participants; The encrypted model input data of the training participant is used to obtain the inference result through the target model. 9, which shows a schematic structural diagram of an electronic device 800 suitable for implementing an embodiment of the present disclosure. Terminal devices in the embodiments of the present disclosure may include, but are not limited to, mobile phones, notebook computers, digital broadcast receivers, PDAs (personal digital assistants), PADs (tablets), PMPs (portable multimedia players), vehicle-mounted terminals (eg, Mobile terminals such as car navigation terminals), etc., and stationary terminals such as digital TVs, desktop computers, and the like. The electronic device shown in FIG. 9 is only an example, and should not impose any limitation on the function and scope of use of the embodiments of the present disclosure. As shown in FIG. 9, the electronic device 800 may include a processing device (eg, a central processing unit, a graphics processor, etc.) 801, which may be loaded into random access according to a program stored in a read only memory (ROM) 802 or from a storage device 808 The program in the memory (RAM) 803 executes various appropriate actions and processes. In the RAM 803, various programs and data necessary for the operation of the electronic device 800 are also stored. The processing device 801 , the ROM 802 , and the RAM 803 are connected to each other through a bus 804 . An input/output (I/O) interface 805 is also connected to bus 804 . Typically, the following devices can be connected to the I/O interface 805: Input devices 806 including, for example, a touch screen, touchpad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, etc.; including, for example, a liquid crystal display (LCD), speakers, vibration an output device 807 for a computer, etc.; a storage device 808 including, for example, a magnetic tape, a hard disk, etc.; exchange data. While FIG. 9 shows electronic device 800 having various means, it should be understood that not all of the illustrated means are required to be implemented or available. More or fewer devices may alternatively be implemented or provided. In particular, according to embodiments of the present disclosure, the processes described above with reference to the flowcharts may be implemented as computer software programs. For example, embodiments of the present disclosure include a computer program product comprising a computer program carried on a non-transitory computer readable medium, the computer program containing program code for performing the method illustrated in the flowchart. In such an embodiment, the computer program may be downloaded and installed from the network via the communication device 809 , or from the storage device 808 , or from the ROM 802 . When the computer program is executed by the processing device 801, the above-mentioned functions defined in the methods of the embodiments of the present disclosure are executed. It should be noted that, the computer-readable medium mentioned above in the present disclosure may be a computer-readable signal medium or a computer-readable storage medium, or any combination of the above two. The computer-readable storage medium can be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus or device, or any combination of the above. More specific examples of computer readable storage media may include, but are not limited to: electrical connections with one or more wires, portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable Programmable read only memory (EPROM or flash memory), optical fiber, portable compact disk read only memory (CD-ROM), optical storage devices, magnetic storage devices, or any suitable combination of the above. In the present disclosure, a computer-readable storage medium may be any tangible medium that contains or stores a program that can be used by or in conjunction with an instruction execution system, apparatus, or device. In the present disclosure, however, a computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave, carrying computer-readable program code therein. Such propagated data signals may take a variety of forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing. A computer-readable signal medium can also be any computer-readable medium other than a computer-readable storage medium that can transmit, propagate, or transmit a program for use by or in conjunction with an instruction execution system, apparatus, or device . The program code embodied on the computer-readable medium may be transmitted by any suitable medium, including but not limited to: wire, optical fiber cable, RF (radio frequency), etc., or any suitable combination of the above. In some embodiments, the client and the server can use any currently known or future developed network protocol such as HTTP (HyperText Transfer Protocol) to communicate, and can communicate with digital data in any form or medium Communication (eg, communication network) interconnection. Examples of communication networks include local area networks ("LAN"), wide area networks ("WAN"), the Internet (eg, the Internet), and peer-to-peer networks (eg, ad hoc peer-to-peer networks), as well as any currently known or future development network of. The above-mentioned computer-readable medium may be included in the above-mentioned electronic device; it may also exist independently without assembled into the electronic device. The computer-readable medium carries one or more programs, and when the one or more programs are executed by the electronic device, the electronic device causes the electronic device to: acquire multiple sub-models; splicing the multiple sub-models to obtain a target model; In the case of receiving the model acquisition request for the target model sent by the inference service executor, send the target model to the inference service executor, so that the inference service executor obtains the target model through the inference service executor inference result. Alternatively, the computer-readable medium carries one or more programs, and when the one or more programs are executed by the electronic device, the electronic device causes the electronic device to: send a model acquisition request for a target model, where the target model is obtained by splicing the multiple sub-models; receiving the target model, and obtaining an inference result through the target model. Computer program code for carrying out operations of the present disclosure may be written in one or more programming languages, including but not limited to object-oriented programming languages such as Java, Smalltalk. C++, and This includes conventional procedural programming languages such as the "C" language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer, or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or may be connected to an external computer (eg, using an Internet service provider to via an Internet connection). The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present disclosure. In this regard, each block in the flowchart or block diagram may represent a module, program segment, or part of code that contains one or more logic functions for implementing the specified executable instructions. It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It is also noted that each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations, can be implemented by dedicated hardware-based systems that perform the specified functions or operations , or can be implemented using a combination of dedicated hardware and computer instructions. The modules involved in the embodiments of the present disclosure may be implemented in a software manner, and may also be implemented in a hardware manner. Wherein, the name of the module does not constitute a limitation of the module itself under certain circumstances, for example, the splicing module may also be described as a "sub-model splicing module". The functions described herein above may be performed, at least in part, by one or more hardware logic components. For example, without limitation, exemplary types of hardware logic components that may be used include: Field Programmable Gate Arrays (FPGAs), Application Specific Integrated Circuits (ASICs), Application Specific Standard Products (ASSPs), Systems on Chips (SOCs), Complex Programmable Logical Devices (CPLDs) and more. In the context of the present disclosure, a machine-readable medium may be a tangible medium that may contain or store a program for use by or in connection with an instruction execution system, apparatus, or device. The machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. Machine-readable media may include, but are not limited to, electronic, magnetic, optical, electromagnetic, infrared, or semiconductor systems, devices, or devices, or any suitable combination of the foregoing. More specific examples of machine-readable storage media would include one or more wire-based electrical connections, portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM or flash memory), fiber optics, portable compact disk read only memory (CD-ROM), optical storage devices, magnetic storage devices, or any suitable combination of the foregoing. According to one or more embodiments of the present disclosure, Example 1 provides a model processing method, the method includes: acquiring multiple sub-models; splicing the multiple sub-models to obtain a target model; In the case of a model acquisition request for the target model sent by the inference party, the target model is sent to the inference service executor, so that the inference service executor obtains an inference result through the target model. According to one or more embodiments of the present disclosure, Example 2 provides the method of Example 1. The splicing of the multiple sub-models includes: acquiring model meta information, where the model meta information includes a sub-model having a sending node The connection relationship information between the sending node of the sending node and the receiving nodes of other sub-models that have a connection relationship with the sending node; The computing nodes of the other sub-models connected to the receiving node are connected to splicing the multiple sub-models. According to one or more embodiments of the present disclosure, Example 3 provides the method of Example 1, wherein multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data, The inference service executor is one of the multiple model training participants, and the inference service executor obtains the inference result through the target model according to its own model input data. According to one or more embodiments of the present disclosure, Example 4 provides the method of Example 1, where multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data, The inference service executor is one of the multiple model training participants; the inference service executor obtains the inference result in the following manner: Receiving the data sent by other model training participants except the inference service executor Encrypted model input data; obtaining the inference result through the target model according to the model input data of the inference service executor itself and the encrypted model input data of the other model training participants. According to one or more embodiments of the present disclosure, Example 5 provides the method of Example 4, wherein the inference service executor is the model training participant that needs to receive the encrypted model input data of other model training participants with the smallest amount of data. For example, the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor. According to one or more embodiments of the present disclosure, Example 6 provides the method of Example 1, wherein multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data, The inference service executor is not the model training participant; the inference service executor obtains the inference result in the following ways: Respectively receive encrypted model input data sent by each of the model training participants; The encrypted model input data of the model training participant is obtained, and the inference result is obtained through the target model. According to one or more embodiments of the present disclosure, Example 7 provides a model processing method, the method comprising: an inference service executor sending a model acquisition request for a target model, where the target model is to convert the multiple obtained by splicing the sub-models; the inference service executor receives the target model, and obtains an inference result through the target model. According to one or more embodiments of the present disclosure, Example 8 provides the method of Example 7, wherein multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data, The inference service executor is one of the multiple model training participants; the obtaining an inference result through the target model includes: inputting data according to the model of the inference service executor itself, and obtaining the inference result through the target model inference result. According to one or more embodiments of the present disclosure, Example 9 provides the method of Example 7, where multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data, The inference service executor is one of the multiple model training participants; the obtaining an inference result through the target model includes: receiving encrypted model inputs sent by other model training participants except the inference service executor data; obtaining the inference result through the target model according to the model input data of the inference service executor itself and the encrypted model input data of the other model training participants. According to one or more embodiments of the present disclosure, Example 10 provides the method of Example 9, where the inference service executor is a model that needs to receive encrypted model input data from other model training participants with the smallest amount of data. Train participants. For example, the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor. According to one or more embodiments of the present disclosure, Example 11 provides the method of Example 7, wherein multiple sub-models correspond to multiple model training participants one-to-one, and each model training participant has its own model input data, The inference service executor is not the model training participant; the obtaining the inference result through the target model includes: respectively receiving encrypted model input data sent by each of the model training participants; participating in the model training according to each of the model training participants. The encrypted model input data of the party is used to obtain the inference result through the target model. According to one or more embodiments of the present disclosure, Example 12 provides a model processing system, where the system includes a model optimization platform and a model storage platform; the model optimization platform is configured to acquire multiple sub-models, and store the multiple sub-models The models are spliced to obtain a target model, and the target model is sent to the model storage platform; the model storage platform is used to receive a model acquisition request for the target model sent by the inference service executor , sending the target model to the inference service executor, so that the inference service executor obtains an inference result through the target model. According to one or more embodiments of the present disclosure, Example 13 provides the system of Example 12, and the model optimization platform is configured to obtain model meta information, the model meta information including a sending node having a sub-model of the sending node and a sending node associated with the sending node. the connection relationship information between the receiving nodes of the other sub-models of which the sending node has a connection relationship; the model optimization platform is configured to, according to the model meta-information, connect the computing nodes of the sub-models connected to the sending node and the The computing nodes of the other sub-models connected to the receiving node are connected to splicing the multiple sub-models. According to one or more embodiments of the present disclosure, Example 14 provides a model processing apparatus, the apparatus comprising: an acquisition module configured to acquire a plurality of sub-models; a stitching module configured to A target model is obtained by splicing multiple sub-models; a target model sending module is configured to send the target model to the target model in the case of receiving a model acquisition request for the target model sent by the inference service executor and the inference service executor, so that the inference service executor obtains an inference result through the target model. According to one or more embodiments of the present disclosure, Example 15 provides a model processing apparatus, the apparatus comprising: an acquisition request sending module configured to send a model acquisition request for a target model, wherein the target The model is obtained by splicing the multiple sub-models; the inference module is configured to receive the target model and obtain an inference result through the target model. For example, the model processing means may be provided at the execution side of the inference service. According to one or more embodiments of the present disclosure, Example 16 provides a computer-readable medium on which a computer program is stored, and when the program is executed by a processing apparatus, implements the steps of the method in any one of Examples 1-6 . According to one or more embodiments of the present disclosure, Example 17 provides a computer-readable medium on which a computer program is stored, and when the program is executed by a processing apparatus, implements the steps of the method in any one of Examples 7-11 . According to one or more embodiments of the present disclosure, Example 18 provides an electronic device, including: a storage device, on which a computer program is stored; a processing device, for executing the computer program in the storage device, to The steps of implementing the method of any of Examples 1-6. According to one or more embodiments of the present disclosure, Example 19 provides an electronic device, including: a storage device on which a computer program is stored; and a processing device for executing the computer program in the storage device, to The steps of implementing the method of any of Examples 7-11. The above description is merely a preferred embodiment of the present disclosure and an illustration of the technical principles employed. Those skilled in the art should understand that the scope of disclosure involved in the present disclosure is not limited to the technical solutions formed by the specific combination of the above-mentioned technical features, and should also cover the technical solutions made of the above-mentioned technical features or Other technical solutions formed by any combination of its equivalent features. For example, a technical solution is formed by replacing the above features with the technical features disclosed in the present disclosure (but not limited to) with similar functions. Additionally, although operations are depicted in a particular order, this should not be construed as requiring that the operations be performed in the particular order shown or in a sequential order. Under certain circumstances, multitasking and parallel processing may be advantageous. Likewise, although the above discussion contains several implementation-specific details, these should not be construed as limitations on the scope of the present disclosure. Certain features that are described in the context of separate embodiments can also be implemented in combination in a single embodiment. Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable subcombination. Although the subject matter has been described in language specific to structural features and/or logical acts of method, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are merely example forms of implementing the claims. Regarding the apparatus in the above-mentioned embodiment, the specific manner in which each module performs operations has been described in detail in the embodiment of the method, and will not be described in detail here.

Claims

权 利 要 求 Rights request
1、 一种模型处理方法, 包括: 获取多个子模型 ; 将所述多个子模型进行拼接 , 得到目标模型; 在接收到推理服务执行方发送的针对所述 目标模型的模型获取请求的情况下, 将 所述目标模型发送至所述推理服务执行方, 以使所述推理服务执行方通过所述目标模 型得到推理结果。 1. A model processing method, comprising: acquiring multiple sub-models; splicing the multiple sub-models to obtain a target model; in the case of receiving a model acquisition request for the target model sent by an inference service executor, The target model is sent to the inference service executor, so that the inference service executor obtains an inference result through the target model.
2、 根据权利要求 1 所述的模型处理方法, 其中, 所述将所述多个子模型进行拼 接包括: 获取模型元信息 , 所述模型元信息包括具有发送节点的子模型的发送节点和与所 述发送节点具有连接关系的其他子模型的接收节点之间的连接关系信息; 根据所述模型元信息 , 将与所述发送节点相连的所述子模型的计算节点和与所述 接收节点相连的所述其他子模型的计算节点进行连接, 以将所述多个子模型进行拼接。 2. The model processing method according to claim 1, wherein the splicing the multiple sub-models comprises: acquiring model meta-information, the model meta-information including the sending node of the sub-model having the sending node and the the connection relationship information between the receiving nodes of other sub-models that the sending node has a connection relationship; The computing nodes of the other sub-models are connected to splicing the multiple sub-models.
3、 根据权利要求 1 所述的模型处理方法, 其中, 所述多个子模型与多个模型训 练参与方一一对应, 每一所述模型训练参与方均具有自身的模型输入数据, 所述推理 服务执行方为所述多个模型训练参与方之一, 所述推理服务执行方根据自身的模型输 入数据通过所述目标模型得到所述推理结果。 3. The model processing method according to claim 1, wherein the multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data, and the reasoning The service executor is one of the multiple model training participants, and the inference service executor obtains the inference result through the target model according to its own model input data.
4、 根据权利要求 1 所述的模型处理方法, 其中, 所述多个子模型与多个模型训 练参与方一一对应, 每一所述模型训练参与方均具有自身的模型输入数据, 所述推理 服务执行方为所述多个模型训练参与方之一, 所述推理服务执行方通过如下方式得到所述推理结果: 接 收除所述推理服务执行方外的其他模型训练参与方发送 的加密的模型输入数 据; 根据所述推理服务执行方 自身的模型输入数据、 以及所述其他模型训练参与方的 加密的模型输入数据, 通过所述目标模型得到所述推理结果。 4. The model processing method according to claim 1, wherein the multiple sub-models correspond to multiple model training participants one-to-one, and each model training participant has its own model input data, and the reasoning The service executor is one of the multiple model training participants, and the inference service executor obtains the inference result in the following manner: Receive the encrypted model sent by other model training participants except the inference service executor input data; obtain the inference result through the target model according to the model input data of the inference service executor itself and the encrypted model input data of the other model training participants.
5、 根据权利要求 4 所述的模型处理方法, 其中, 所述推理服务执行方通过如下 方式确定: 确定 各模型训练参与方进行推理服务所需的其他模型训练参 与方发送的加密的 模型输入数据的数据量; 将所需数据量最小 的模型训练参与方确定为所述推理服务执行方。 5. The model processing method according to claim 4, wherein the inference service executor is determined by: determining the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service The model training participant with the smallest required data amount is determined as the inference service executor.
6、 根据权利要求 1 所述的模型处理方法, 其中, 所述多个子模型与多个模型训 练参与方一一对应, 每一所述模型训练参与方均具有自身的模型输入数据, 所述推理 服务执行方不为所述模型训练参与方, 所述推理服务执行方通过如下方式得到所述推理结果: 分别接收各个所述模型训练参与方发送的加密的模型输入数据; 根据各个所述模型训练参与方的加密的模型输入数据, 通过所述目标模型得到所 述推理结果。 6. The model processing method according to claim 1, wherein the multiple sub-models correspond to multiple model training participants one-to-one, and each model training participant has its own model input data, and the reasoning The service executor is not the model training participant, and the inference service executor obtains the inference result in the following ways: respectively receiving encrypted model input data sent by each of the model training participants; training according to each model The encrypted model input data of the participant, and the inference result is obtained through the target model.
7、 一种模型处理方法, 包括: 推理服务执行方发送针对 目标模型的模型获取请求, 所述目标模型是将多个子模 型进行拼接得到的; 所述推理服务执行方接收所述 目标模型, 并通过所述目标模型得到推理结果。 7. A model processing method, comprising: an inference service executor sending a model acquisition request for a target model, where the target model is obtained by splicing multiple sub-models; the inference service executor receiving the target model, and The inference result is obtained through the target model.
8、 根据权利要求 7 所述的模型处理方法, 其中, 所述多个子模型与多个模型训 练参与方一一对应, 每一所述模型训练参与方均具有自身的模型输入数据, 所述推理 服务执行方为所述多个模型训练参与方之一; 所述通过所述 目标模型得到推理结果包括: 根据所述推理服务执行方 自身的模型输入数据, 通过所述目标模型得到所述推理 结果。 8. The model processing method according to claim 7, wherein the multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data, and the reasoning The service executor is one of the multiple model training participants; the obtaining the inference result through the target model includes: inputting data according to the model of the inference service executor itself, and obtaining the inference result through the target model .
9、 根据权利要求 7 所述的模型处理方法, 其中, 所述多个子模型与多个模型训 练参与方一一对应, 每一所述模型训练参与方均具有自身的模型输入数据, 推理服务 执行方为所述多个模型训练参与方之一; 所述通过所述 目标模型得到推理结果包括: 接 收除所述推理服务执行方外的其他模型训练参与方发送 的加密的模型输入数 据; 根据所述推理服务执行方 自身的模型输入数据、 以及所述其他模型训练参与方的 加密的模型输入数据, 通过所述目标模型得到所述推理结果。 9. The model processing method according to claim 7, wherein the multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executes The party is one of the multiple model training participants; the obtaining the inference result through the target model includes: Receive encrypted model input data sent by other model training participants except the inference service executor; according to the inference service executor's own model input data and the encrypted model input data of the other model training participants , and obtain the inference result through the target model.
10、 根据权利要求 9所述的模型处理方法, 其中, 所述推理服务执行方通过如下 方式确定: 确定 各模型训练参与方进行推理服务所需的其他模型训练参 与方发送的加密的 模型输入数据的数据量; 将所需数据量最小 的模型训练参与方确定为所述推理服务执行方。 10. The model processing method according to claim 9, wherein the inference service executor is determined by: determining the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service The model training participant with the smallest required data amount is determined as the inference service executor.
11、 根据权利要求 7所述的模型处理方法, 其中, 所述多个子模型与多个模型训 练参与方一一对应, 每一所述模型训练参与方均具有自身的模型输入数据, 所述推理 服务执行方不为所述模型训练参与方; 所述通过所述 目标模型得到推理结果包括: 分别接收各个所述模型训练参与方发送的加密的模型输入数据; 根据各个所述模型训练参与方的加密的模型输入数据, 通过所述目标模型得到所 述推理结果。 11. The model processing method according to claim 7, wherein the multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data, and the reasoning The service executor is not the model training participant; the obtaining the inference result through the target model includes: respectively receiving the encrypted model input data sent by each of the model training participants; The encrypted model input data is used to obtain the inference result through the target model.
12、 一种模型处理系统, 包括模型优化平台、 模型存储平台, 所述模型优化平 台用于获取多个子模型, 将所述多个子模型进行拼接, 得到目标 模型, 并将所述目标模型发送至所述模型存储平台, 所述 模型存储平台用于在接收到推理服务执行方发送 的针对所述目标模型的模 型获取请求的情况下, 将所述目标模型发送至所述推理服务执行方, 以使所述推理服 务执行方通过所述目标模型得到推理结果。 12. A model processing system, comprising a model optimization platform and a model storage platform, wherein the model optimization platform is used to obtain multiple sub-models, splicing the multiple sub-models to obtain a target model, and sending the target model to The model storage platform, where the model storage platform is configured to send the target model to the inference service executor in the case of receiving a model acquisition request for the target model sent by the inference service executor, to The inference service executor obtains an inference result through the target model.
13、 根据权利要求 12 所述的模型处理系统, 其中, 所述模型优化平台用于获取 模型元信息, 所述模型元信息包括具有发送节点的子模型的发送节点和与所述发送节 点具有连接关系的其他子模型的接收节点之间的连接关系信息; 所述模型优化平 台用于根据所述模型元信息, 将与所述发送节点相连的所述子模 型的计算节点和与所述接收节点相连的所述其他子模型的计算节点进行连接, 以将所 述多个子模型进行拼接。 13. The model processing system according to claim 12, wherein the model optimization platform is used to obtain model meta information, and the model meta information includes a sending node having a sub-model of a sending node and a sending node having a connection with the sending node The connection relationship information between the receiving nodes of other sub-models of the relationship; the model optimization platform is configured to, according to the model meta-information, The computing node of the type is connected with the computing nodes of the other sub-models connected to the receiving node, so as to splicing the multiple sub-models.
14、 一种模型处理装置, 包括: 获取模块, 被配置成用于获取多个子模型; 拼接模块 , 被配置成用于将所述多个子模型进行拼接, 得到目标模型; 目标模型发送模块, 被配置成用于在接收到推理服务执行方发送的针对所述目标 模型的模型获取请求的情况下, 将所述目标模型发送至所述推理服务执行方, 以使所 述推理服务执行方通过所述目标模型得到推理结果。 14. A model processing device, comprising: an acquisition module, configured to acquire a plurality of sub-models; a splicing module, configured to splicing the plurality of sub-models to obtain a target model; a target model sending module, sent by is configured to send the target model to the inference service executor in the case of receiving a model acquisition request for the target model sent by the inference service executor, so that the inference service executor passes the Describe the target model to get the inference results.
15、 一种模型处理装置, 包括: 获取请求发送模块, 被配置成用于发送针对目标模型的模型获取请求, 其中, 所 述目标模型是将多个子模型进行拼接得到的; 推理模块 ,被配置成用于接收所述目标模型,并通过所述目标模型得到推理结果。 15. A model processing apparatus, comprising: an acquisition request sending module, configured to send a model acquisition request for a target model, wherein the target model is obtained by splicing multiple sub-models; an inference module, configured is used to receive the target model and obtain inference results through the target model.
16、 一种计算机可读介质, 其上存储有计算机程序, 其特征在于, 该程序被处理 装置执行时实现权利要求 1-6中任一项所述方法的步骤。 16. A computer-readable medium on which a computer program is stored, characterized in that, when the program is executed by a processing device, the steps of the method according to any one of claims 1-6 are implemented.
17、 一种计算机可读介质, 其上存储有计算机程序, 其特征在于, 该程序被处理 装置执行时实现权利要求 7-11中任一项所述方法的步骤。 17. A computer-readable medium on which a computer program is stored, characterized in that, when the program is executed by a processing device, the steps of the method according to any one of claims 7-11 are implemented.
18、 一种电子设备, 包括: 存储装置 , 其上存储有计算机程序; 处理装置 , 用于执行所述存储装置中的所述计算机程序, 以实现权利要求 1-6中 任一项所述方法的步骤。 18. An electronic device, comprising: a storage device on which a computer program is stored; a processing device for executing the computer program in the storage device to implement the method according to any one of claims 1-6 A step of.
19^ 一种电子设备, 包括: 存储装置 , 其上存储有计算机程序; 处理装置 , 用于执行所述存储装置中的所述计算机程序, 以实现权利要求 7-11中 任一项所述方法的步骤。 19^ An electronic device, comprising: a storage device on which a computer program is stored; a processing device for executing the computer program in the storage device to implement the method in any one of claims 7-11 A step of.
22 twenty two
20, 一种计算机程序, 包括: 指令 , 所述指令当由处理器执行时使所述处理器执行根据权利要求 1~11 中任一 项所述的模型处理方法。 20. A computer program, comprising: instructions that, when executed by a processor, cause the processor to execute the model processing method according to any one of claims 1 to 11.
21.一种计算机程序产品, 包括指令,所述指令当由处理器执行时使所述处理器执 行根据权利要求 1~11中任一项所述的模型处理方法。 21. A computer program product comprising instructions that, when executed by a processor, cause the processor to perform the model processing method according to any one of claims 1 to 11.
23 twenty three
PCT/SG2021/050707 2020-11-18 2021-11-16 Model processing method, system and apparatus, medium, and electronic device WO2022108527A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011298789.5 2020-11-18
CN202011298789.5A CN112418446B (en) 2020-11-18 2020-11-18 Model processing method, system, device, medium and electronic equipment

Publications (1)

Publication Number Publication Date
WO2022108527A1 true WO2022108527A1 (en) 2022-05-27

Family

ID=74773394

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/SG2021/050707 WO2022108527A1 (en) 2020-11-18 2021-11-16 Model processing method, system and apparatus, medium, and electronic device

Country Status (2)

Country Link
CN (1) CN112418446B (en)
WO (1) WO2022108527A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115374944A (en) * 2022-10-26 2022-11-22 小米汽车科技有限公司 Model reasoning method and device, electronic equipment and storage medium

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112346870B (en) * 2020-11-18 2024-04-16 脸萌有限公司 Model processing method and system
CN112966825B (en) * 2021-04-13 2023-05-23 杭州欣禾圣世科技有限公司 Multi-model fusion parallel reasoning method, device and system based on python

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111460511A (en) * 2020-04-17 2020-07-28 支付宝(杭州)信息技术有限公司 Federal learning and virtual object distribution method and device based on privacy protection
US20200242514A1 (en) * 2016-09-26 2020-07-30 Google Llc Communication Efficient Federated Learning
CN111899076A (en) * 2020-08-12 2020-11-06 科技谷(厦门)信息技术有限公司 Aviation service customization system and method based on federal learning technology platform
CN111898769A (en) * 2020-08-17 2020-11-06 中国银行股份有限公司 Method and system for establishing user behavior period model based on horizontal federal learning

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110633805B (en) * 2019-09-26 2024-04-26 深圳前海微众银行股份有限公司 Longitudinal federal learning system optimization method, device, equipment and readable storage medium
CN111461874A (en) * 2020-04-13 2020-07-28 浙江大学 Credit risk control system and method based on federal mode
CN111753996A (en) * 2020-06-24 2020-10-09 中国建设银行股份有限公司 Optimization method, device, equipment and storage medium of scheme determination model
CN111797999A (en) * 2020-07-10 2020-10-20 深圳前海微众银行股份有限公司 Longitudinal federal modeling optimization method, device, equipment and readable storage medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200242514A1 (en) * 2016-09-26 2020-07-30 Google Llc Communication Efficient Federated Learning
CN111460511A (en) * 2020-04-17 2020-07-28 支付宝(杭州)信息技术有限公司 Federal learning and virtual object distribution method and device based on privacy protection
CN111899076A (en) * 2020-08-12 2020-11-06 科技谷(厦门)信息技术有限公司 Aviation service customization system and method based on federal learning technology platform
CN111898769A (en) * 2020-08-17 2020-11-06 中国银行股份有限公司 Method and system for establishing user behavior period model based on horizontal federal learning

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115374944A (en) * 2022-10-26 2022-11-22 小米汽车科技有限公司 Model reasoning method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN112418446A (en) 2021-02-26
CN112418446B (en) 2024-04-09

Similar Documents

Publication Publication Date Title
WO2022108527A1 (en) Model processing method, system and apparatus, medium, and electronic device
WO2023273725A1 (en) File downloading method and apparatus, storage medium and electronic device
US20220271924A1 (en) Method and apparatus for processing data request
WO2022151966A1 (en) Processing method and apparatus for language model, text generation method and apparatus, and medium
WO2021190129A1 (en) Method and device for page processing, electronic device, and computer-readable storage medium
WO2023173954A1 (en) Data acquisition method and apparatus, storage medium, and electronic device
WO2023078072A1 (en) Byzantine fault tolerance-based asynchronous consensus method and apparatus, server and medium
WO2021233297A1 (en) Resource sharing method and apparatus, electronic device and computer-readable medium
CN111309304B (en) Method, device, medium and electronic equipment for generating IDL file
CN111355784B (en) Method, device, medium and electronic equipment for processing request information
WO2022108521A1 (en) Model processing method and system
US20220269622A1 (en) Data processing methods, apparatuses, electronic devices and computer-readable storage media
WO2022134968A1 (en) Model training method, speech recognition method, apparatuses, medium and device
WO2023033717A2 (en) Data protection method and apparatus, medium, and electronic device
CN112100211B (en) Data storage method, apparatus, electronic device, and computer readable medium
WO2022103338A1 (en) Data processing method and apparatus, and electronic device
WO2022108520A1 (en) Data processing method and apparatus, and medium and electronic device
CN111143355B (en) Data processing method and device
CN116437093A (en) Video frame repair method, apparatus, device, storage medium, and program product
CN111367590A (en) Interrupt event processing method and device
CN112036822B (en) Interaction method and device based on color ropes, medium and electronic equipment
CN111399902B (en) Client source file processing method and device, readable medium and electronic equipment
CN112036821B (en) Quantization method, quantization device, quantization medium and quantization electronic equipment based on grid map planning private line
CN116755889B (en) Data acceleration method, device and equipment applied to server cluster data interaction
CN111626787B (en) Resource issuing method, device, medium and equipment

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21895246

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21895246

Country of ref document: EP

Kind code of ref document: A1