WO2022108527A1 - Procédé, système et appareil de traitement de modèle, support et dispositif électronique - Google Patents

Procédé, système et appareil de traitement de modèle, support et dispositif électronique Download PDF

Info

Publication number
WO2022108527A1
WO2022108527A1 PCT/SG2021/050707 SG2021050707W WO2022108527A1 WO 2022108527 A1 WO2022108527 A1 WO 2022108527A1 SG 2021050707 W SG2021050707 W SG 2021050707W WO 2022108527 A1 WO2022108527 A1 WO 2022108527A1
Authority
WO
WIPO (PCT)
Prior art keywords
model
inference
models
target
service executor
Prior art date
Application number
PCT/SG2021/050707
Other languages
English (en)
Chinese (zh)
Inventor
陈程
周子凯
余乐乐
解浚源
吴良超
常龙
张力哲
刘小兵
吴迪
Original Assignee
脸萌有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 脸萌有限公司 filed Critical 脸萌有限公司
Publication of WO2022108527A1 publication Critical patent/WO2022108527A1/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • G06N20/20Ensemble learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/602Providing cryptographic facilities or services
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/04Inference or reasoning models

Definitions

  • the federated learning system can effectively help multiple institutions to complete the joint training of models.
  • the federated learning model is usually composed of multiple sub-models. How to ensure the reliability of the inference service is an important issue when the inference service is performed through the federated learning model.
  • the present disclosure provides a model processing method, the method includes: acquiring multiple sub-models; splicing the multiple sub-models to obtain a target model; after receiving a message sent by an inference service executor for the target model In the case of the model obtaining request, the target model is sent to the inference service executor, so that the inference service executor obtains an inference result through the target model.
  • the present disclosure provides a model processing method. The method includes: an inference service executor sends a model acquisition request for a target model, where the target model is obtained by splicing multiple sub-models; the inference service executes The party receives the target model, and obtains an inference result through the target model.
  • the present disclosure provides a model processing system, the system includes a model optimization platform and a model storage platform; the model optimization platform is used for acquiring multiple sub-models, and splicing the multiple sub-models to obtain target model, and send the target model to the model storage platform; the model storage platform is configured to store the target model in the case of receiving a model acquisition request for the target model sent by the inference service executor The model is sent to the inference service executor, so that the inference service executor obtains an inference result through the target model.
  • the present disclosure provides a model processing device, the device comprising: an acquisition module configured to acquire multiple sub-models; and a splicing module configured to splicing the multiple sub-models to obtain a target model; a target model sending module, configured to send the target model to the inference service executor in the case of receiving a model acquisition request for the target model sent by the inference service executor, so that The inference service executor obtains an inference result through the target model.
  • the present disclosure provides a model processing apparatus, the apparatus comprising: an acquisition request sending module, configured to send a model acquisition request for a target model, wherein the target model is to convert the multiple sub-models obtained by splicing; an inference module, configured to receive the target model, and obtain an inference result through the target model.
  • the model processing apparatus may be set at the execution side of the inference service.
  • the present disclosure provides a computer-readable medium on which a computer program is stored, and when the program is executed by a processing apparatus, implements the steps of the method provided in the first aspect of the present disclosure.
  • the present disclosure provides a computer-readable medium on which a computer program is stored, and when the program is executed by a processing apparatus, implements the steps of the method provided in the second aspect of the present disclosure.
  • the present disclosure provides an electronic device, comprising: a storage device on which a computer program is stored; and a processing device for executing the computer program in the storage device, so as to implement the computer program provided in the first aspect of the present disclosure. the steps of the method.
  • the present disclosure provides an electronic device, comprising: a storage device on which a computer program is stored; and a processing device for executing the computer program in the storage device, so as to implement the computer program provided in the second aspect of the present disclosure. the steps of the method.
  • the present disclosure provides a computer program, comprising: instructions that, when executed by a processor, cause the processor to execute the model processing method according to any one of the foregoing embodiments.
  • the present disclosure provides a computer program product comprising instructions that, when executed by a processor, cause the processor to execute the model processing method according to any one of the foregoing embodiments.
  • FIG. 1 is a schematic diagram of a federated learning model in the related art.
  • Fig. 2 is a flow chart of a model processing method according to some exemplary embodiments.
  • Fig. 3 is a schematic diagram of a target model according to some exemplary embodiments.
  • FIG. 4 is a schematic diagram of a model processing system according to some exemplary embodiments.
  • Fig. 5 is a schematic diagram illustrating an inference service executor obtaining an inference result through a target model according to its own model input data, according to some exemplary embodiments.
  • Fig. 6 is a flow chart of a model processing method according to some exemplary embodiments.
  • Fig. 1 is a schematic diagram of a federated learning model in the related art.
  • Fig. 2 is a flow chart of a model processing method according to some exemplary embodiments.
  • Fig. 3 is a schematic diagram of a target model according to some exemplary embodiments.
  • FIG. 4 is a schematic diagram of a model processing system
  • FIG. 7 is a block diagram of a model processing apparatus according to some exemplary embodiments.
  • Fig. 8 is a block diagram of a model processing apparatus according to some exemplary embodiments.
  • FIG. 9 is a schematic structural diagram of an electronic device according to some exemplary embodiments. DETAILED DESCRIPTION Embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While certain embodiments of the present disclosure are shown in the drawings, it should be understood that the present disclosure may be embodied in various forms and should not be construed as limited to the embodiments set forth herein, but rather are provided for the purpose of A more thorough and complete understanding of the present disclosure.
  • the federated learning system can combine the data of multiple data owners to train a common federated learning model.
  • the federated learning model is trained by combining the data of multiple data owners, and the training data is more comprehensive. Therefore, the accuracy of the federated learning model is higher.
  • FIG. 1 is a schematic diagram of a federated learning model in the related art.
  • the federated learning model includes sub-model A and sub-model B.
  • the sub-model A corresponds to the model training participant 1
  • the model input data X, Y, and Z of the sub-model A are data owned by the model training participant 1.
  • the sub-model B corresponds to the model training participant 2, and the model input data M and N of the sub-model B are data owned by the model training participant 2.
  • each model training participant loads its own sub-models, that is, model training participant 1 loads sub-model A, and model training participant 2 loads sub-model B.
  • the model training participant 1 performs computation through the sub-model A according to the model input data X, Y, and Z. Then model training participant 1 needs to remotely send the data to model training participant 2 through the sending node of sub-model A, so as to transmit the data to the receiving node of sub-model B. The model training participant 2 then obtains an inference result through the sub-model B according to the data received by the receiving node and the model input data M and N.
  • remote communication is required between multiple model training participants to complete the entire inference service, that is, the sending node and the receiving node use remote communication to transmit data, and the communication overhead is large.
  • Fig. 2 is a flow chart of a model processing method according to some exemplary embodiments. As shown in FIG. 2, the method may include S201 ⁇ S203o In S201, multiple sub-models are acquired.
  • FIG. 3 is a schematic diagram of a target model according to some exemplary embodiments.
  • the target model shown in FIG. 3 may be obtained according to the federated learning model shown in FIG. 1 , and there is a connection relationship between the sending node of the sub-model A and the receiving node of the sub-model B.
  • the computing node of the sub-model A connected to the sending node and the computing node of the sub-model B connected to the receiving node can be connected to obtain the target model.
  • the target model is the overall full model obtained by splicing sub-model A and sub-model B together.
  • the present disclosure takes two sub-models as examples for illustration, which does not constitute a limitation on the implementation of the present disclosure. In practical applications, the number of sub-models may be multiple, which is not specifically limited in the present disclosure. .
  • the target model is sent to the inference service executor, so that the inference service executor obtains the inference result through the target model.
  • Inference service can refer to the process by which the server performs computations through the model based on the input data and obtains the result.
  • the user's current shopping intention can be inferred through a model, and then the user can be provided with an inference result that meets his shopping intention and needs.
  • the user's current search intent can be inferred through the model, and then the user can be provided with an inference result that conforms to the user's search intent.
  • one of the model training participants can be used as the inference service executor, load the target model, and obtain the inference result through the target model.
  • the target model is obtained by splicing multiple sub-models, and the inference service executor can directly obtain the inference results through the overall target model, without each model training participant loading its own sub-models, and without the need for multiple model training participants. It can effectively avoid the problem of unstable long-distance communication. It is worth noting that, when it is mentioned in the present disclosure that the inference service executor performs operations of sending, receiving, and processing data, it may be understood that the inference service executor performs these operations through a server device.
  • a target model is obtained by splicing multiple sub-models, and the inference service executor can obtain an inference result through the target model.
  • the target model is obtained by splicing multiple sub-models, and the inference service executor can directly obtain the inference result through the overall target model, without each model training participant loading its own sub-models separately.
  • the entire inference service process can be completed locally on the inference service executor, without the need for remote communication between multiple model training participants to transmit data. In this way, not only the communication overhead can be reduced, but also the problem of unstable remote transmission caused by factors such as network routing can be effectively avoided, the normal operation of the inference service can be ensured, and the reliability of the inference service can be improved.
  • the model processing method shown in FIG. 2 may be applied to a model processing apparatus including a splicing module.
  • the model processing device may be, for example, a cloud server, the acquisition module in the model processing device acquires multiple sub-models, and the splicing module splices the multiple sub-models to obtain the target model.
  • the model processing method shown in FIG. 2 can also be applied to a model processing system.
  • FIG. 4 is a schematic diagram of a model processing system according to some exemplary embodiments. As shown in FIG. 4 , the model processing system may include a model optimization platform 401, a model storage platform 402, a model training platform 403, a model metadata storage platform 404, a model training participant 1, and a model training participant 2.
  • the model training platform 403 is used to train each sub-model, such as sub-model A and sub-model B.
  • the model meta information storage platform 404 may be used to store model related meta information.
  • the model training platform 403 can send multiple sub-models to the model optimization platform 401.
  • the model optimization platform 401 can be used to obtain the multiple sub-models sent by the model training platform 403, splicing the multiple sub-models to obtain the target model, and sending the target model to Model storage platform 402 o
  • the inference service executor may send a model acquisition request for the target model to the model storage platform 402; the model storage platform 402 may send the target model to the inference service executor when receiving the request.
  • the inference service executor may be one of model training participant 1 and model training participant 2, for example. FIG.
  • the step of splicing multiple sub-models in S202 may include: acquiring model meta information, where the model meta information may include a sending node having a sub-model of the sending node and other sub-models that have a connection relationship with the sending node The connection relationship information between the receiving nodes; according to the model meta information, connect the computing node of the sub-model connected to the sending node and the computing nodes of other sub-models connected to the receiving node to connect the multiple sub-models. stitching.
  • Model meta-information may refer to information used to describe a model, and may include connection relationship information between nodes.
  • the model optimization platform 401 can obtain model meta information from the model meta information storage platform 404, and according to the model meta information, connect the The computing nodes of the sub-model are connected with the computing nodes of other sub-models connected to the receiving node. In this way, data can be directly transmitted between two computing nodes that originally need to forward data through the sending node and the receiving node.
  • the inference service executor performs the inference service through the target model, and the inference service executor locally
  • the entire inference service process can be completed without the need for remote communication of data, effectively avoiding the problem of unstable remote transmission caused by factors such as network routing, ensuring the normal operation of the inference service, and improving the reliability of the inference service.
  • multiple sub-models correspond to multiple model training participants one-to-one, and each model training participant has its own model input data.
  • the inference service executor can be one of multiple model training participants, and the inference service executor can obtain inference results through the target model according to its own model input data.
  • the inference service performer may be any of a plurality of model training participants.
  • Fig. 5 is a schematic diagram illustrating an inference service executor obtaining an inference result through a target model according to its own model input data, according to some exemplary embodiments.
  • Figure 5 takes the inference service executor as the model training participant 1 as an example.
  • the model training participant 1 obtains the target model, it can input data X, Y, and Z according to its own model, and obtain an inference result through the target model.
  • the inference service executor may be the party that needs the inference result among the multiple model training participants, that is, the inference result demander.
  • the model training participant 1 needs the final inference result, and the model training participant 1 can be used as the inference service executor to obtain the target model and perform the inference service.
  • the inference service executor may send the inference result to the inference result requester.
  • model training participant 2 needs an inference result, and model training participant 1 can send the inference result to model training participant 2 .
  • other model training participants may not transmit model input data to the inference service executor, and the inference service executor can perform inference services according to its own data, with low communication overhead.
  • the reasoning service is completed by the inference service executor, and there is no need to carry out the process of remote communication and data transmission between multiple model training participants, which reduces the communication overhead and improves the stability of the inference service.
  • multiple sub-models correspond to multiple model training participants one-to-one, each model training participant has its own model input data, and the inference service executor is one of the multiple model training participants .
  • the inference service executor can obtain the inference results in the following ways: Receive encrypted model input data sent by other model training participants except the inference service executor; According to the inference service executor's own model input data and other model training participants The encrypted model input data is obtained, and the inference result is obtained through the target model.
  • model training participants can add model input data. encrypted processing. Send the encrypted model input data to the inference service executor, and the present disclosure does not specifically limit the encryption method.
  • the inference service executor can use its own model input data and the encrypted model input data of other model training participants to obtain inference results through the target model, that is, to perform inference services based on the model input data of each sub-model.
  • the inference service executor may be the model training participant that needs to receive the encrypted model input data of other model training participants with the smallest amount of data.
  • the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor. For example, if the model training participant 1 performs the inference service, the model training participant 2 needs to send the model input data M and N of the sub-model B to the model training participant 1; if the model training participant 2 performs the inference service, Model training participant 1 needs to send the model input data X, Y, and Z of sub-model A to model training participant 2.
  • the model training participant 1 needs to receive the minimum data volume of the model input data, which can be executed by the model training participant 1 as an inference service. square.
  • the inference service executor performs the inference service according to the model input data of each sub-model, the model training participant with the smallest amount of model input data to be received can be used as the inference service executor, which can, to a certain extent, be used as the inference service executor. Reduce communication overhead.
  • the inference service executor performs the calculation through the target model, and there is no need to carry out the process of remote communication and data transmission among multiple model training participants, thereby improving the stability of the inference service process.
  • multiple sub-models correspond to multiple model training participants one-to-one, each model training participant has its own model input data, and the inference service executor is not a model training participant; inference service The executor can obtain the inference result in the following ways: Receive the encrypted model input data sent by each model training participant respectively; Obtain the inference result through the target model according to the encrypted model input data of each model training participant.
  • the inference service executor may not be a model training participant, for example, it may be a preset cloud server.
  • Fig. 6 is a flow chart of a model processing method according to some exemplary embodiments. The method can be applied to the inference service executor, that is, the server device of the inference service executor. As shown in FIG.
  • the method may include S501 and S502.
  • the inference service executor sends a model acquisition request for the target model.
  • the target model is obtained by splicing the multiple sub-models.
  • the inference service executor may send a model acquisition request to the model storage platform, and may also send a model acquisition request to a model processing apparatus including a splicing module, which is not limited in the present disclosure.
  • the inference service executor receives the target model, and obtains an inference result through the target model.
  • the inference service executor can send a model acquisition request for the target model.
  • the target model is obtained by splicing multiple sub-models, and the inference service executor can obtain the inference result through the target model.
  • the inference service executor obtains the inference result through the target model.
  • the inference service executor can complete the entire inference service process locally. In this way, there is no need to perform remote communication of data, which not only reduces communication overhead, but also effectively avoids the problem of unstable remote transmission caused by factors such as network routing, ensures the normal operation of inference services, and improves the reliability of inference services.
  • the multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is one of the multiple model training participants.
  • the step of obtaining the inference result through the target model in S502 may include: inputting data according to the model of the inference service executor itself, and obtaining the inference result through the target model.
  • the multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is one of the multiple model training participants.
  • the step of obtaining the inference result from the target model in S502 may include: receiving encrypted model input data sent by other model training participants except the inference service executor; and according to the model input data of the inference service executor itself , and the encrypted model input data of the other model training participants, and obtain the inference result through the target model.
  • the inference service executor is a model training participant that needs to receive the encrypted model input data of other model training participants with the smallest amount of data.
  • the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor.
  • the multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is not the model training participant.
  • the step of obtaining the inference result through the target model in S502 may include: respectively receiving encrypted model input data sent by each of the model training participants; according to the encrypted model input data of each of the model training participants, The target model obtains the inference result.
  • the method applied to the inference service executor in the foregoing embodiment the specific manner in which each step performs operations has been described in detail in the embodiment of the method applied to the model processing system or the model processing apparatus including the splicing module, here A detailed explanation will not be given.
  • the present disclosure also provides a model processing system, such as the model processing system shown in FIG.
  • the system may include a model optimization platform and a model storage platform; the model optimization platform is used for acquiring multiple sub-models, and storing the multiple sub-models Splicing is performed to obtain a target model, and the target model is sent to the model storage platform; the model storage platform is configured to, upon receiving a model acquisition request for the target model sent by the inference service executor, Send the target model to the inference service executor, so that the inference service executor obtains an inference result through the target model.
  • the model optimization platform is used to obtain model meta information, and the model The meta-information includes connection relationship information between a sending node having a sub-model of the sending node and receiving nodes of other sub-models that have a connection relationship with the sending node; the model optimization platform is configured to, according to the model meta-information,
  • the computing node of the sub-model connected with the sending node is connected with the computing nodes of the other sub-models connected with the receiving node, so as to splicing the multiple sub-models.
  • the model processing apparatus 600 may include: an acquisition module 601, configured to acquire multiple sub-models; a splicing module 602, configured to splicing the multiple sub-models to obtain a target model;
  • the target model sending module 603 is configured to send the target model to the inference service executor in the case of receiving a model acquisition request for the target model sent by the inference service executor, so that all The inference service executor obtains the inference result through the target model.
  • the splicing module 602 may include: an obtaining submodule configured to obtain model meta information, where the model meta information includes a sending node having a sub-model of a sending node and a sending node having a connection with the sending node the connection relationship information between the receiving nodes of other sub-models of the connection relationship; the splicing sub-module is configured to connect the computing node of the sub-model connected to the sending node and the computing node of the sub-model connected to the sending node according to the model meta information.
  • the computing nodes of the other sub-models connected to the receiving node are connected to splicing the multiple sub-models.
  • multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is one of the multiple model training participants.
  • the inference service executor obtains the inference result through the target model according to its own model input data.
  • multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is one of the multiple model training participants. 1.
  • the inference service executor obtains the inference result in the following ways: Receive encrypted model input data sent by other model training participants except the inference service executor; According to the inference service executor's own model The input data and the encrypted model input data of the other model training participants are used to obtain the inference result through the target model.
  • the inference service executor is a model training participant that needs to receive the encrypted model input data of other model training participants with the smallest amount of data.
  • the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor.
  • Fig. 8 is a block diagram of a model processing apparatus according to some exemplary embodiments.
  • the model processing apparatus 700 can be applied to an inference service executor. As shown in FIG.
  • the model processing apparatus 700 may include: an acquisition request sending module 701, configured to send a model acquisition request for a target model, wherein the target model is obtained by splicing the multiple sub-models
  • the inference module 702 is configured to receive the target model and obtain an inference result through the target model.
  • multiple sub-models are in one-to-one correspondence with multiple model training participants, each of the model training participants has its own model input data, and the inference service executor is one of the multiple model training participants;
  • the inference module 702 may include: a first inference sub-module, configured to input data according to the model of the inference service executor itself, and obtain the inference result through the target model.
  • the inference module 702 may include: a first receiving sub-module, configured to receive encrypted model input data sent by other model training participants except the inference service executor; a second inference sub-module, configured by It is configured to obtain the inference result through the target model according to the model input data of the inference service executor itself and the encrypted model input data of the other model training participants.
  • the inference service executor is a model training participant that needs to receive the encrypted model input data of other model training participants with the smallest amount of data.
  • the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor.
  • multiple sub-models are in one-to-one correspondence with multiple model training participants, each model training participant has its own model input data, and the inference service executor is not the model training participant;
  • the reasoning The module 702 may include: a second receiving sub-module, configured to respectively receive encrypted model input data sent by each of the model training participants; The encrypted model input data of the training participant is used to obtain the inference result through the target model.
  • Terminal devices in the embodiments of the present disclosure may include, but are not limited to, mobile phones, notebook computers, digital broadcast receivers, PDAs (personal digital assistants), PADs (tablets), PMPs (portable multimedia players), vehicle-mounted terminals (eg, Mobile terminals such as car navigation terminals), etc., and stationary terminals such as digital TVs, desktop computers, and the like.
  • PDAs personal digital assistants
  • PADs tablets
  • PMPs portable multimedia players
  • vehicle-mounted terminals eg, Mobile terminals such as car navigation terminals
  • stationary terminals such as digital TVs, desktop computers, and the like.
  • the electronic device shown in FIG. 9 is only an example, and should not impose any limitation on the function and scope of use of the embodiments of the present disclosure. As shown in FIG.
  • the electronic device 800 may include a processing device (eg, a central processing unit, a graphics processor, etc.) 801, which may be loaded into random access according to a program stored in a read only memory (ROM) 802 or from a storage device 808
  • the program in the memory (RAM) 803 executes various appropriate actions and processes.
  • various programs and data necessary for the operation of the electronic device 800 are also stored.
  • the processing device 801 , the ROM 802 , and the RAM 803 are connected to each other through a bus 804 .
  • An input/output (I/O) interface 805 is also connected to bus 804 .
  • Input devices 806 including, for example, a touch screen, touchpad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, etc.; including, for example, a liquid crystal display (LCD), speakers, vibration an output device 807 for a computer, etc.; a storage device 808 including, for example, a magnetic tape, a hard disk, etc.; exchange data.
  • FIG. 9 shows electronic device 800 having various means, it should be understood that not all of the illustrated means are required to be implemented or available. More or fewer devices may alternatively be implemented or provided.
  • the processes described above with reference to the flowcharts may be implemented as computer software programs.
  • embodiments of the present disclosure include a computer program product comprising a computer program carried on a non-transitory computer readable medium, the computer program containing program code for performing the method illustrated in the flowchart.
  • the computer program may be downloaded and installed from the network via the communication device 809 , or from the storage device 808 , or from the ROM 802 .
  • the processing device 801 When the computer program is executed by the processing device 801, the above-mentioned functions defined in the methods of the embodiments of the present disclosure are executed.
  • the computer-readable medium mentioned above in the present disclosure may be a computer-readable signal medium or a computer-readable storage medium, or any combination of the above two.
  • the computer-readable storage medium can be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus or device, or any combination of the above. More specific examples of computer readable storage media may include, but are not limited to: electrical connections with one or more wires, portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable Programmable read only memory (EPROM or flash memory), optical fiber, portable compact disk read only memory (CD-ROM), optical storage devices, magnetic storage devices, or any suitable combination of the above.
  • a computer-readable storage medium may be any tangible medium that contains or stores a program that can be used by or in conjunction with an instruction execution system, apparatus, or device.
  • a computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave, carrying computer-readable program code therein. Such propagated data signals may take a variety of forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • a computer-readable signal medium can also be any computer-readable medium other than a computer-readable storage medium that can transmit, propagate, or transmit a program for use by or in conjunction with an instruction execution system, apparatus, or device .
  • the program code embodied on the computer-readable medium may be transmitted by any suitable medium, including but not limited to: wire, optical fiber cable, RF (radio frequency), etc., or any suitable combination of the above.
  • the client and the server can use any currently known or future developed network protocol such as HTTP (HyperText Transfer Protocol) to communicate, and can communicate with digital data in any form or medium Communication (eg, communication network) interconnection.
  • HTTP HyperText Transfer Protocol
  • Examples of communication networks include local area networks (“LAN”), wide area networks (“WAN”), the Internet (eg, the Internet), and peer-to-peer networks (eg, ad hoc peer-to-peer networks), as well as any currently known or future development network of.
  • LAN local area networks
  • WAN wide area networks
  • the Internet eg, the Internet
  • peer-to-peer networks eg, ad hoc peer-to-peer networks
  • the above-mentioned computer-readable medium may be included in the above-mentioned electronic device; it may also exist independently without assembled into the electronic device.
  • the computer-readable medium carries one or more programs, and when the one or more programs are executed by the electronic device, the electronic device causes the electronic device to: acquire multiple sub-models; splicing the multiple sub-models to obtain a target model; In the case of receiving the model acquisition request for the target model sent by the inference service executor, send the target model to the inference service executor, so that the inference service executor obtains the target model through the inference service executor inference result.
  • the computer-readable medium carries one or more programs, and when the one or more programs are executed by the electronic device, the electronic device causes the electronic device to: send a model acquisition request for a target model, where the target model is obtained by splicing the multiple sub-models; receiving the target model, and obtaining an inference result through the target model.
  • Computer program code for carrying out operations of the present disclosure may be written in one or more programming languages, including but not limited to object-oriented programming languages such as Java, Smalltalk. C++, and This includes conventional procedural programming languages such as the "C" language or similar programming languages.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer, or entirely on the remote computer or server.
  • the remote computer may be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or may be connected to an external computer (eg, using an Internet service provider to via an Internet connection).
  • LAN local area network
  • WAN wide area network
  • Internet service provider e.g.
  • each block in the flowchart or block diagram may represent a module, program segment, or part of code that contains one or more logic functions for implementing the specified executable instructions.
  • the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
  • each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations can be implemented by dedicated hardware-based systems that perform the specified functions or operations , or can be implemented using a combination of dedicated hardware and computer instructions.
  • the modules involved in the embodiments of the present disclosure may be implemented in a software manner, and may also be implemented in a hardware manner.
  • the name of the module does not constitute a limitation of the module itself under certain circumstances, for example, the splicing module may also be described as a "sub-model splicing module".
  • the functions described herein above may be performed, at least in part, by one or more hardware logic components.
  • exemplary types of hardware logic components include: Field Programmable Gate Arrays (FPGAs), Application Specific Integrated Circuits (ASICs), Application Specific Standard Products (ASSPs), Systems on Chips (SOCs), Complex Programmable Logical Devices (CPLDs) and more.
  • a machine-readable medium may be a tangible medium that may contain or store a program for use by or in connection with an instruction execution system, apparatus, or device.
  • the machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium.
  • Machine-readable media may include, but are not limited to, electronic, magnetic, optical, electromagnetic, infrared, or semiconductor systems, devices, or devices, or any suitable combination of the foregoing.
  • machine-readable storage media would include one or more wire-based electrical connections, portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM or flash memory), fiber optics, portable compact disk read only memory (CD-ROM), optical storage devices, magnetic storage devices, or any suitable combination of the foregoing.
  • RAM random access memory
  • ROM read only memory
  • EPROM or flash memory erasable programmable read only memory
  • CD-ROM compact disk read only memory
  • optical storage devices magnetic storage devices, or any suitable combination of the foregoing.
  • Example 1 provides a model processing method, the method includes: acquiring multiple sub-models; splicing the multiple sub-models to obtain a target model; In the case of a model acquisition request for the target model sent by the inference party, the target model is sent to the inference service executor, so that the inference service executor obtains an inference result through the target model.
  • Example 2 provides the method of Example 1.
  • the splicing of the multiple sub-models includes: acquiring model meta information, where the model meta information includes a sub-model having a sending node The connection relationship information between the sending node of the sending node and the receiving nodes of other sub-models that have a connection relationship with the sending node; The computing nodes of the other sub-models connected to the receiving node are connected to splicing the multiple sub-models.
  • Example 3 provides the method of Example 1, wherein multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data,
  • the inference service executor is one of the multiple model training participants, and the inference service executor obtains the inference result through the target model according to its own model input data.
  • Example 4 provides the method of Example 1, where multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data,
  • the inference service executor is one of the multiple model training participants; the inference service executor obtains the inference result in the following manner: Receiving the data sent by other model training participants except the inference service executor Encrypted model input data; obtaining the inference result through the target model according to the model input data of the inference service executor itself and the encrypted model input data of the other model training participants.
  • Example 5 provides the method of Example 4, wherein the inference service executor is the model training participant that needs to receive the encrypted model input data of other model training participants with the smallest amount of data.
  • the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor.
  • Example 6 provides the method of Example 1, wherein multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data,
  • the inference service executor is not the model training participant; the inference service executor obtains the inference result in the following ways: Respectively receive encrypted model input data sent by each of the model training participants; The encrypted model input data of the model training participant is obtained, and the inference result is obtained through the target model.
  • Example 7 provides a model processing method, the method comprising: an inference service executor sending a model acquisition request for a target model, where the target model is to convert the multiple obtained by splicing the sub-models; the inference service executor receives the target model, and obtains an inference result through the target model.
  • Example 8 provides the method of Example 7, wherein multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data,
  • the inference service executor is one of the multiple model training participants;
  • the obtaining an inference result through the target model includes: inputting data according to the model of the inference service executor itself, and obtaining the inference result through the target model inference result.
  • Example 9 provides the method of Example 7, where multiple sub-models are in one-to-one correspondence with multiple model training participants, and each model training participant has its own model input data,
  • the inference service executor is one of the multiple model training participants; the obtaining an inference result through the target model includes: receiving encrypted model inputs sent by other model training participants except the inference service executor data; obtaining the inference result through the target model according to the model input data of the inference service executor itself and the encrypted model input data of the other model training participants.
  • Example 10 provides the method of Example 9, where the inference service executor is a model that needs to receive encrypted model input data from other model training participants with the smallest amount of data. Train participants.
  • the inference service executor determines by: determining the data volume of the encrypted model input data sent by other model training participants required by each model training participant to perform the inference service; Determined to be the inference service executor.
  • Example 11 provides the method of Example 7, wherein multiple sub-models correspond to multiple model training participants one-to-one, and each model training participant has its own model input data, The inference service executor is not the model training participant; the obtaining the inference result through the target model includes: respectively receiving encrypted model input data sent by each of the model training participants; participating in the model training according to each of the model training participants. The encrypted model input data of the party is used to obtain the inference result through the target model.
  • Example 12 provides a model processing system, where the system includes a model optimization platform and a model storage platform; the model optimization platform is configured to acquire multiple sub-models, and store the multiple sub-models The models are spliced to obtain a target model, and the target model is sent to the model storage platform; the model storage platform is used to receive a model acquisition request for the target model sent by the inference service executor , sending the target model to the inference service executor, so that the inference service executor obtains an inference result through the target model.
  • Example 13 provides the system of Example 12, and the model optimization platform is configured to obtain model meta information, the model meta information including a sending node having a sub-model of the sending node and a sending node associated with the sending node.
  • the model optimization platform is configured to, according to the model meta-information, connect the computing nodes of the sub-models connected to the sending node and the The computing nodes of the other sub-models connected to the receiving node are connected to splicing the multiple sub-models.
  • Example 14 provides a model processing apparatus, the apparatus comprising: an acquisition module configured to acquire a plurality of sub-models; a stitching module configured to A target model is obtained by splicing multiple sub-models; a target model sending module is configured to send the target model to the target model in the case of receiving a model acquisition request for the target model sent by the inference service executor and the inference service executor, so that the inference service executor obtains an inference result through the target model.
  • Example 15 provides a model processing apparatus, the apparatus comprising: an acquisition request sending module configured to send a model acquisition request for a target model, wherein the target The model is obtained by splicing the multiple sub-models; the inference module is configured to receive the target model and obtain an inference result through the target model.
  • the model processing means may be provided at the execution side of the inference service.
  • Example 16 provides a computer-readable medium on which a computer program is stored, and when the program is executed by a processing apparatus, implements the steps of the method in any one of Examples 1-6 .
  • Example 17 provides a computer-readable medium on which a computer program is stored, and when the program is executed by a processing apparatus, implements the steps of the method in any one of Examples 7-11 .
  • Example 18 provides an electronic device, including: a storage device, on which a computer program is stored; a processing device, for executing the computer program in the storage device, to The steps of implementing the method of any of Examples 1-6.
  • Example 19 provides an electronic device, including: a storage device on which a computer program is stored; and a processing device for executing the computer program in the storage device, to The steps of implementing the method of any of Examples 7-11.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Medical Informatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Bioethics (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Hardware Design (AREA)
  • Computer Security & Cryptography (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

La présente divulgation concerne un procédé, un système et un appareil de traitement de modèle, un support et un dispositif électronique. Le procédé consiste à : acquérir une pluralité de sous-modèles ; assembler la pluralité de sous-modèles pour obtenir un modèle cible ; et lors de la réception d'une demande d'acquisition de modèle pour le modèle cible en provenance d'une partie d'exécution de service d'inférence, envoyer le modèle cible à la partie d'exécution de service d'inférence, de telle sorte que la partie d'exécution de service d'inférence obtient un résultat d'inférence au moyen du modèle cible.
PCT/SG2021/050707 2020-11-18 2021-11-16 Procédé, système et appareil de traitement de modèle, support et dispositif électronique WO2022108527A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011298789.5A CN112418446B (zh) 2020-11-18 2020-11-18 模型处理方法、系统、装置、介质及电子设备
CN202011298789.5 2020-11-18

Publications (1)

Publication Number Publication Date
WO2022108527A1 true WO2022108527A1 (fr) 2022-05-27

Family

ID=74773394

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/SG2021/050707 WO2022108527A1 (fr) 2020-11-18 2021-11-16 Procédé, système et appareil de traitement de modèle, support et dispositif électronique

Country Status (2)

Country Link
CN (1) CN112418446B (fr)
WO (1) WO2022108527A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115374944A (zh) * 2022-10-26 2022-11-22 小米汽车科技有限公司 模型推理方法、装置、电子设备和存储介质

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112346870B (zh) * 2020-11-18 2024-04-16 脸萌有限公司 模型处理方法及系统
CN112966825B (zh) * 2021-04-13 2023-05-23 杭州欣禾圣世科技有限公司 基于python实现的多模型融合并行推理方法、装置及系统
CN113570061A (zh) * 2021-08-27 2021-10-29 知见科技(江苏)有限公司 一种多模型融合推理方法
CN115618966A (zh) * 2022-10-30 2023-01-17 抖音视界有限公司 用于训练机器学习模型的方法、装置、设备和介质

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111460511A (zh) * 2020-04-17 2020-07-28 支付宝(杭州)信息技术有限公司 基于隐私保护的联邦学习、虚拟对象分配方法和装置
US20200242514A1 (en) * 2016-09-26 2020-07-30 Google Llc Communication Efficient Federated Learning
CN111899076A (zh) * 2020-08-12 2020-11-06 科技谷(厦门)信息技术有限公司 一种基于联邦学习技术平台的航空服务定制化系统及方法
CN111898769A (zh) * 2020-08-17 2020-11-06 中国银行股份有限公司 基于横向联邦学习的建立用户行为周期模型的方法及系统

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110633805B (zh) * 2019-09-26 2024-04-26 深圳前海微众银行股份有限公司 纵向联邦学习系统优化方法、装置、设备及可读存储介质
CN111461874A (zh) * 2020-04-13 2020-07-28 浙江大学 一种基于联邦模式的信贷风险控制系统及方法
CN111753996A (zh) * 2020-06-24 2020-10-09 中国建设银行股份有限公司 一种方案确定模型的优化方法、装置、设备及存储介质
CN111797999A (zh) * 2020-07-10 2020-10-20 深圳前海微众银行股份有限公司 纵向联邦建模优化方法、装置、设备及可读存储介质

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200242514A1 (en) * 2016-09-26 2020-07-30 Google Llc Communication Efficient Federated Learning
CN111460511A (zh) * 2020-04-17 2020-07-28 支付宝(杭州)信息技术有限公司 基于隐私保护的联邦学习、虚拟对象分配方法和装置
CN111899076A (zh) * 2020-08-12 2020-11-06 科技谷(厦门)信息技术有限公司 一种基于联邦学习技术平台的航空服务定制化系统及方法
CN111898769A (zh) * 2020-08-17 2020-11-06 中国银行股份有限公司 基于横向联邦学习的建立用户行为周期模型的方法及系统

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115374944A (zh) * 2022-10-26 2022-11-22 小米汽车科技有限公司 模型推理方法、装置、电子设备和存储介质

Also Published As

Publication number Publication date
CN112418446A (zh) 2021-02-26
CN112418446B (zh) 2024-04-09

Similar Documents

Publication Publication Date Title
WO2022108527A1 (fr) Procédé, système et appareil de traitement de modèle, support et dispositif électronique
WO2022151966A1 (fr) Procédé et appareil de traitement pour modèle de langage, procédé et appareil de génération de texte, et support
WO2023273725A1 (fr) Procédé et appareil de téléchargement de fichier, support de stockage et dispositif électronique
US20220271924A1 (en) Method and apparatus for processing data request
WO2021190129A1 (fr) Procédé et dispositif de traitement de page, dispositif électronique et support de stockage lisible par ordinateur
CN111309304B (zh) 一种生成idl文件的方法、装置、介质和电子设备
WO2023173954A1 (fr) Procédé et appareil d'acquisition de données, support de stockage et dispositif électronique
WO2023078072A1 (fr) Procédé et appareil de consensus asynchrone basé sur la tolérance aux pannes byzantine, serveur et support
WO2023033717A2 (fr) Procédé et appareil de protection de données, support et dispositif électronique
CN111355784B (zh) 一种处理请求信息的方法、装置、介质和电子设备
WO2022108521A1 (fr) Procédé et système de traitement de modèle
US20220269622A1 (en) Data processing methods, apparatuses, electronic devices and computer-readable storage media
CN111399902B (zh) 客户端源文件处理方法、装置、可读介质与电子设备
CN112100211B (zh) 数据存储方法、装置、电子设备和计算机可读介质
WO2022103338A1 (fr) Procédé et appareil de traitement de données et dispositif électronique
WO2022108520A1 (fr) Procédé et appareil de traitement de données, support et dispositif électronique
CN111143355B (zh) 数据处理方法及装置
CN116437093A (zh) 视频帧修复方法、装置、设备、存储介质和程序产品
CN111367590A (zh) 中断事件处理方法及其装置
CN112036822B (zh) 一种基于彩绳的交互方法、装置、介质和电子设备
WO2024188345A1 (fr) Procédé et dispositif de traitement de données
CN112036821B (zh) 基于网格图规划专线的量化方法、装置、介质和电子设备
CN111324512B (zh) 用于生成文本的方法、装置、电子设备和计算机可读介质
CN114785794B (zh) 资源分配方法、装置、设备、介质和系统
CN111626787B (zh) 资源发放方法、装置、介质和设备

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21895246

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21895246

Country of ref document: EP

Kind code of ref document: A1