CN112214285A - Docker-based model service deployment system - Google Patents

Docker-based model service deployment system Download PDF

Info

Publication number
CN112214285A
CN112214285A CN202011135150.5A CN202011135150A CN112214285A CN 112214285 A CN112214285 A CN 112214285A CN 202011135150 A CN202011135150 A CN 202011135150A CN 112214285 A CN112214285 A CN 112214285A
Authority
CN
China
Prior art keywords
model
service
management module
mirror image
docker
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011135150.5A
Other languages
Chinese (zh)
Inventor
洪万福
钱智毅
张宏伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xiamen Yuanting Information Technology Co ltd
Original Assignee
Xiamen Yuanting Information Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xiamen Yuanting Information Technology Co ltd filed Critical Xiamen Yuanting Information Technology Co ltd
Priority to CN202011135150.5A priority Critical patent/CN112214285A/en
Publication of CN112214285A publication Critical patent/CN112214285A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • G06F9/45533Hypervisors; Virtual machine monitors
    • G06F9/45558Hypervisor-specific management and integration aspects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5061Partitioning or combining of resources
    • G06F9/5077Logical partitioning of resources; Management or configuration of virtualized resources
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • G06F9/45533Hypervisors; Virtual machine monitors
    • G06F9/45558Hypervisor-specific management and integration aspects
    • G06F2009/45562Creating, deleting, cloning virtual machine instances
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • G06F9/45533Hypervisors; Virtual machine monitors
    • G06F9/45558Hypervisor-specific management and integration aspects
    • G06F2009/4557Distribution of virtual machine instances; Migration and load balancing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • G06F9/45533Hypervisors; Virtual machine monitors
    • G06F9/45558Hypervisor-specific management and integration aspects
    • G06F2009/45595Network integration; Enabling network access in virtual machine instances

Landscapes

  • Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention relates to a model service deployment system based on Docker, which comprises a model warehouse management module, a mirror image library management module, a resource quota management module, a service management module and a service monitoring module, wherein the model warehouse management module is used for managing the resource quota; the model warehouse management module is used for uniformly managing the models; the mirror image library management module is used for providing basic environment mirror image support for model training or service release; the resource quota management module is used for managing the running resources of each model service and can uniformly schedule various types of cluster resources to support heterogeneous hardware resource scheduling; the service management module is used for deploying the model and managing the life cycle of the model according to the model management mode and the resource quota management mode output by the model warehouse management module and the resource quota management module; the service monitoring module is used for monitoring the operation condition and service record of the model service in real time. The invention can more conveniently deploy the model service and realize the high availability and multi-type model support of the model service.

Description

Docker-based model service deployment system
Technical Field
The invention relates to the field of artificial intelligence, in particular to a model service deployment system based on Docker.
Background
With the development of artificial intelligence, machine learning and deep learning related technologies are applied to many industries and fields. Model deployment artificial intelligence is a very important link in application, modeling is difficult, and deployment is more difficult. For model deployment, single-service manual deployment is common at present, and shutdown updating is needed for model service updating in the later period; the model service operation resources cannot be managed and controlled, and the server operation resources cannot be reasonably distributed; the access rights of various model services cannot be uniformly controlled, and the running state and the access condition of the services cannot be monitored. In order to solve the defects, some model service management systems appear in the industry, model service authentication, service state monitoring and the like can be uniformly managed, and the problem of model service deployment management is solved to a certain extent. However, these systems have some disadvantages, such as that the deployment resources cannot be controlled in fine granularity, flexible expansion is not supported, and gray scale distribution is not supported.
Disclosure of Invention
In order to solve the problems, the invention provides a model service deployment system based on Docker.
The specific scheme is as follows:
a model service deployment system based on Docker comprises a model warehouse management module, a mirror image library management module, a resource quota management module, a service management module and a service monitoring module;
the model warehouse management module is used for uniformly managing the models, and the models are acquired in one or more modes of uploading through local model files and transmitting the models through a butted training platform;
the mirror image library management module is used for providing basic environment mirror image support for model training or service release;
the resource quota management module is used for managing the running resources of each model service and can uniformly schedule various types of cluster resources to support heterogeneous hardware resource scheduling;
the service management module is used for deploying the model and managing the life cycle of the model according to the model management mode and the resource quota management mode output by the model warehouse management module and the resource quota management module;
the service monitoring module is used for monitoring the operation condition and service record of the model service in real time.
Further, the functions of the model warehouse management module further include: online model compression conversion, model version management and one-key service release.
Further, the basic environment image comprises a built-in image and a user-defined image, the built-in image comprises a dependent environment of the fixed frame, and the user-defined image comprises a user-defined environment image except the dependent environment of the fixed frame.
Further, the user-defined mirror image supports the construction in a mode of uploading Dockerfile files or the construction through the selection of online environment mirror image options.
Further, the resource quota management module has functions of resource application, resource approval, resource monitoring and flexible expansion capacity.
Further, the deployment of the model in the service management module supports the deployment of a built-in fixed framework and a user-defined framework.
Furthermore, the service monitoring module provides a visual interface and monitors the performance of the service management module after the service management module deploys each model service and the operation condition of each model service in real time.
Furthermore, the real-time monitored operation status of each model service includes deployment resource usage, request information data, service response data, and service authentication information statistical data.
By adopting the technical scheme, the model service can be more conveniently deployed, and high-availability and multi-type model support of the model service is realized.
Drawings
FIG. 1 is a block diagram of a system according to an embodiment of the present invention.
Detailed Description
The invention is further illustrated by the accompanying drawings. The accompanying drawings, which are incorporated in and constitute a part of this disclosure, illustrate embodiments of the invention and, together with the description, serve to explain the principles of the embodiments. Those skilled in the art will appreciate still other possible embodiments and advantages of the present invention with reference to these figures.
The invention will now be further described with reference to the accompanying drawings and detailed description.
The first embodiment is as follows:
an embodiment of the present invention provides a model service deployment system based on a Docker, as shown in fig. 1, the system includes a model warehouse management module, a mirror library management module, a resource quota management module, a service management module, and a service monitoring module, where:
(1) the model warehouse management module is used for uniformly managing the models, and the models are acquired in one or more modes of uploading through local model files and transmitting the models through a butted training platform.
The model is generated by machine learning or deep learning training.
The model warehouse management module comprises an external model operation interface and is in butt joint with other training platforms through the external model operation interface.
The functions of the model warehouse management module further comprise: online model compression conversion, model version management, one-key service release, and the like.
(2) The mirror image library management module is used for providing basic environment mirror image support for model training or service release so as to realize flexible framework environment support.
The basic environment mirror image comprises a built-in mirror image and a user-defined mirror image, wherein the built-in mirror image comprises a dependent environment of a fixed frame, such as a common frame: sklearn, XGboost, LigthGBM, SparkML, TensorFlow, caffe, Theano, Keras, Pythroch, MXNet, and the like. And if the built-in mirror image does not meet the requirements of the user, the user can configure the user-defined mirror image on line. The user-customized image comprises a user-customized environment image except for the dependent environment of the fixed framework. The user-defined mirror image supports the construction in a mode of uploading Dockerfile files or the construction through the selection of online environment mirror image options.
(3) The resource quota management module is used for managing the running resources of each model service, and can uniformly schedule various types of cluster resources to support heterogeneous hardware resource scheduling. The functions of the resource quota management module in this embodiment include resource application, resource approval, resource monitoring, elastic expansion and contraction, and the like, and the operating resources of each model service can be flexibly controlled in a fine-grained manner.
(4) The service management module is used for deploying the model and managing the life cycle of the model according to the model management mode and the resource quota management mode output by the model warehouse management module and the resource quota management module.
The deployment of the model has high reliability and expandability, and supports built-in fixed frameworks such as PMML, Sklearn, XGboost, LigthGBM and SparkML; deployment of ONNX, TensorFlow, cafe, Theano, Keras, Pythroch, MXNet and user-defined frameworks.
The service management module can provide functions of one-stop model service release management, service authentication management, service version control, gray release, online real-time prediction, batch prediction, model evaluation and the like, wherein the service authentication management supports control of service interface access limitation according to days and access quantity.
(5) The service monitoring module is used for monitoring the operation condition and service record of the model service in real time.
The service monitoring module provides a visual interface and monitors the performance of each model service after the service management module deploys each model service and the running condition of each model service in real time. The real-time monitored operation condition of each model service comprises deployment resource use condition, request information data, service response data and service authentication information statistical data.
The embodiment of the invention provides a visual one-stop model version control system, which can more conveniently deploy model services by the purposes of model service version management, monitoring deployment performance, operation condition and web service indexes, and realizes high availability of the model services and multi-type model support.
While the invention has been particularly shown and described with reference to a preferred embodiment, it will be understood by those skilled in the art that various changes in form and detail may be made therein without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (8)

1. A model service deployment system based on Docker is characterized by comprising a model warehouse management module, a mirror image library management module, a resource quota management module, a service management module and a service monitoring module;
the model warehouse management module is used for uniformly managing the models, and the models are acquired in one or more modes of uploading through local model files and transmitting the models through a butted training platform;
the mirror image library management module is used for providing basic environment mirror image support for model training or service release;
the resource quota management module is used for managing the running resources of each model service and can uniformly schedule various types of cluster resources to support heterogeneous hardware resource scheduling;
the service management module is used for deploying the model and managing the life cycle of the model according to the model management mode and the resource quota management mode output by the model warehouse management module and the resource quota management module;
the service monitoring module is used for monitoring the operation condition and service record of the model service in real time.
2. The Docker-based model service deployment system of claim 1, wherein: the functions of the model warehouse management module further comprise: online model compression conversion, model version management and one-key service release.
3. The Docker-based model service deployment system of claim 1, wherein: the basic environment mirror image comprises a built-in mirror image and a user-defined mirror image, the built-in mirror image comprises a dependent environment of the fixed frame, and the user-defined mirror image comprises a user-defined environment mirror image except the dependent environment of the fixed frame.
4. The Docker-based model service deployment system of claim 3, wherein: the user-defined mirror image supports the construction in a mode of uploading Dockerfile files or the construction through the selection of online environment mirror image options.
5. The Docker-based model service deployment system of claim 1, wherein: the resource quota management module has the functions of resource application, resource approval, resource monitoring and elastic expansion and contraction capacity.
6. The Docker-based model service deployment system of claim 1, wherein: the deployment of the model in the service management module supports the deployment of a built-in fixed framework and a user-defined framework.
7. The Docker-based model service deployment system of claim 1, wherein: the service monitoring module provides a visual interface and monitors the performance of each model service after the service management module deploys each model service and the running condition of each model service in real time.
8. The Docker-based model service deployment system of claim 7, wherein: the real-time monitored operation condition of each model service comprises deployment resource use condition, request information data, service response data and service authentication information statistical data.
CN202011135150.5A 2020-10-22 2020-10-22 Docker-based model service deployment system Pending CN112214285A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011135150.5A CN112214285A (en) 2020-10-22 2020-10-22 Docker-based model service deployment system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011135150.5A CN112214285A (en) 2020-10-22 2020-10-22 Docker-based model service deployment system

Publications (1)

Publication Number Publication Date
CN112214285A true CN112214285A (en) 2021-01-12

Family

ID=74056306

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011135150.5A Pending CN112214285A (en) 2020-10-22 2020-10-22 Docker-based model service deployment system

Country Status (1)

Country Link
CN (1) CN112214285A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113283855A (en) * 2021-04-29 2021-08-20 成都商高智能科技有限公司 Recruitment system based on containerized resource programming

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2014058411A1 (en) * 2012-10-08 2014-04-17 Hewlett-Packard Development Company, L.P. Hybrid cloud environment
WO2017045424A1 (en) * 2015-09-18 2017-03-23 乐视控股(北京)有限公司 Application program deployment system and deployment method
CN109508238A (en) * 2019-01-05 2019-03-22 咪付(广西)网络技术有限公司 A kind of resource management system and method for deep learning
CN109600269A (en) * 2019-01-21 2019-04-09 云南电网有限责任公司信息中心 A kind of cloud management platform based on DCOS
US20190171438A1 (en) * 2017-12-05 2019-06-06 Archemy, Inc. Active adaptation of networked compute devices using vetted reusable software components
CN110413294A (en) * 2019-08-06 2019-11-05 中国工商银行股份有限公司 Service delivery system, method, apparatus and equipment
CN111414233A (en) * 2020-03-20 2020-07-14 京东数字科技控股有限公司 Online model reasoning system
CN111629061A (en) * 2020-05-28 2020-09-04 苏州浪潮智能科技有限公司 Inference service system based on Kubernetes

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2014058411A1 (en) * 2012-10-08 2014-04-17 Hewlett-Packard Development Company, L.P. Hybrid cloud environment
WO2017045424A1 (en) * 2015-09-18 2017-03-23 乐视控股(北京)有限公司 Application program deployment system and deployment method
US20190171438A1 (en) * 2017-12-05 2019-06-06 Archemy, Inc. Active adaptation of networked compute devices using vetted reusable software components
CN109508238A (en) * 2019-01-05 2019-03-22 咪付(广西)网络技术有限公司 A kind of resource management system and method for deep learning
CN109600269A (en) * 2019-01-21 2019-04-09 云南电网有限责任公司信息中心 A kind of cloud management platform based on DCOS
CN110413294A (en) * 2019-08-06 2019-11-05 中国工商银行股份有限公司 Service delivery system, method, apparatus and equipment
CN111414233A (en) * 2020-03-20 2020-07-14 京东数字科技控股有限公司 Online model reasoning system
CN111629061A (en) * 2020-05-28 2020-09-04 苏州浪潮智能科技有限公司 Inference service system based on Kubernetes

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113283855A (en) * 2021-04-29 2021-08-20 成都商高智能科技有限公司 Recruitment system based on containerized resource programming
CN113283855B (en) * 2021-04-29 2024-01-30 成都商高智能科技有限公司 Recruitment system based on containerized resource programming

Similar Documents

Publication Publication Date Title
RU2737566C2 (en) Methods, systems and device for dynamic provision of unlimited system administration with operating configuration m:n of high availability
JP4809772B2 (en) Management based on computer system and distributed application model
CN110109686B (en) Application operation and maintenance method and system based on container management engine
CN1407441B (en) System and method for automatic management computer service and programmable device
CN116170316A (en) Network system, instance management and control method, device and storage medium
CN104508627A (en) Hybrid cloud environment
CN102819478B (en) A kind of data handling system monitoring and management method without agency
CN106657167B (en) Management server, server cluster, and management method
CN111143044B (en) Task scheduling management system, method, device and storage medium thereof
CN111930706B (en) Remote call-based distributed network file storage system and method
CN111984270A (en) Application deployment method and system
CN112114939A (en) Distributed system deployment equipment and method
US20220308559A1 (en) Controlling an Industrial Process Using Virtualized Instances of Control Software
CN112214285A (en) Docker-based model service deployment system
Mendiratta Reliability analysis of clustered computing systems
CN114090192A (en) Docker Swarm cluster resource cleaning optimization method and system
CN114489989A (en) Method and system for parallel scheduling based on proxy client
Achtaich et al. Designing a framework for smart IoT adaptations
CN112306640A (en) Container dispensing method, apparatus, device and medium therefor
CN114489761B (en) Service integration and application integration method based on container cluster
CN115048158A (en) Process arranging and calling method, system and computer equipment thereof
CN113986662A (en) Edge cluster monitoring method and system
CN113010531A (en) Block chain BAAS system task scheduling framework based on directed acyclic graph
CN101364224A (en) Information management system and method
CN111291101A (en) Cluster management method and system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination