CN109617738B - Method, system and non-volatile storage medium for user service scaling - Google Patents

Method, system and non-volatile storage medium for user service scaling Download PDF

Info

Publication number
CN109617738B
CN109617738B CN201811619760.5A CN201811619760A CN109617738B CN 109617738 B CN109617738 B CN 109617738B CN 201811619760 A CN201811619760 A CN 201811619760A CN 109617738 B CN109617738 B CN 109617738B
Authority
CN
China
Prior art keywords
service
classification model
data
training
unit
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201811619760.5A
Other languages
Chinese (zh)
Other versions
CN109617738A (en
Inventor
王远大
宋翔
郑健
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ucloud Technology Co ltd
Original Assignee
Ucloud Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ucloud Technology Co ltd filed Critical Ucloud Technology Co ltd
Priority to CN201811619760.5A priority Critical patent/CN109617738B/en
Publication of CN109617738A publication Critical patent/CN109617738A/en
Application granted granted Critical
Publication of CN109617738B publication Critical patent/CN109617738B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/08Configuration management of networks or network elements
    • H04L41/0896Bandwidth or capacity management, i.e. automatically increasing or decreasing capacities
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/08Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters
    • H04L43/0805Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters by checking availability
    • H04L43/0817Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters by checking availability by checking functioning
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Environmental & Geological Engineering (AREA)
  • Telephonic Communication Services (AREA)

Abstract

The invention provides a method, a system and a non-volatile storage medium for service scaling, wherein the method comprises the following steps: a request forwarding step of forwarding a service request from a user; a data acquisition step, namely receiving the service request, calculating service performance data according to the service request, and collecting hardware data; a conversion step, performing format conversion on the service performance data and the hardware data to obtain training data; a decision step, namely acquiring a trained current classification model, and inputting the training data into the current classification model to obtain a scaling decision; a capacity expansion and reduction step, namely utilizing the capacity expansion and reduction decision to perform service capacity expansion and reduction; and a training step, training the current classification model by using the training data to obtain an updated classification model, and taking the updated classification model as the current classification model.

Description

Method, system and non-volatile storage medium for user service scaling
Technical Field
The invention relates to a method, a system and a non-volatile storage medium for service scaling.
Background
In order to attract more users, each big cloud facilitator builds its own API (application program interface) service, such as the arrhizus API market, the hundredth API store, etc. The automatic capacity expansion service provided by the current cloud provider for the user is mainly to dynamically increase or decrease the allocation amount of the machine according to the use condition of the hardware resources, such as the utilization rate of cpu, memory and the like, so as to reduce the waste of the hardware resources. The amount of the resource demand of the API service of the user needs to be estimated by the user and is distributed in advance. In addition, the user needs to monitor the resource utilization rate by means of an alarm system, and when the resources are insufficient, more resources are added manually.
The current capacity expansion and reduction scheme of the cloud provider mainly increases and decreases hardware resources so as to improve the resource utilization rate. The availability and performance of user API services requires the user himself to ensure. The API service performance on the user side is different, and a phenomenon that the service resource utilization rate is low but the service response time becomes slow may occur, and at this time, parallel service expansion rather than capacity reduction is often required. Therefore, simply through the use condition of hardware resources such as cpu occupancy rate, memory usage amount and the like, high availability and high performance of user services cannot be ensured. The user side needs to increase monitoring and manual capacity expansion, so that the access cost of the user side is increased, and the automation degree is low.
In addition, the existing capacity expansion and contraction strategy is too simple to make an optimal decision. The measurement indexes of the types of hardware resources and the quality of service are more, and if the capacity is expanded and contracted by using simple logic judgment only according to partial indexes, the optimization of a strategy is difficult to ensure, and the waste of resources is inevitable.
Meanwhile, the existing capacity expansion and reduction strategy is unchanged, once the system design is completed, the capacity expansion and reduction strategy is kept unchanged all the time, and self-adaptation can not be carried out according to the current situation.
Disclosure of Invention
The invention provides a method for service scaling, which comprises the following steps:
a request forwarding step of forwarding a service request from a user;
a data acquisition step, namely receiving the service request, calculating service performance data according to the service request, and collecting hardware data;
a conversion step, performing format conversion on the service performance data and the hardware data to obtain training data;
a decision step, namely acquiring a trained current classification model, and inputting the training data into the current classification model to obtain a scaling decision;
a capacity expansion and reduction step, namely utilizing the capacity expansion and reduction decision to perform service capacity expansion and reduction;
and a training step, training the current classification model by using the training data to obtain an updated classification model, and taking the updated classification model as the current classification model.
In the data acquisition step, the user service unit receives the service request and calculates service performance data according to the service request.
Wherein the training data conforms to a format of training performed in the training step.
And when the decision step is carried out next time, the updated classification model is obtained.
Wherein, in the request forwarding step, the routing unit forwards the service request, and the service scaling is scaling the routing unit and/or the user service unit.
The present invention also provides a system for service scalability, the system comprising:
a routing unit for forwarding a service request from a user;
the data acquisition unit is used for receiving the service request, calculating service performance data according to the service request and collecting hardware data;
the conversion unit is used for carrying out format conversion on the service performance data and the hardware data to obtain training data;
the decision unit is used for acquiring a trained current classification model and inputting the training data into the current classification model to obtain a scaling decision;
the capacity expansion and reduction unit is used for carrying out service capacity expansion and reduction by utilizing the capacity expansion and reduction decision;
and the training unit is used for training the current classification model by using the training data to obtain an updated classification model, and taking the updated classification model as the current classification model.
Wherein the data acquisition unit further comprises:
the user service unit receives the service request and calculates the service performance data according to the service request;
a collection unit to collect the hardware data.
Wherein the training data conforms to a format in which training is performed in the training unit.
And when the decision unit makes a next decision, acquiring the updated classification model.
Wherein the service scaling is scaling the at least one routing unit and/or the user service unit.
The present invention further provides a nonvolatile storage medium on which a program for performing service expansion is stored, the program being executed by a computer to implement a method of service expansion, the program including:
a request forwarding instruction for forwarding a service request from a user;
a data acquisition instruction, receiving the service request, calculating service performance data according to the service request, and collecting hardware data;
a conversion instruction, which is used for carrying out format conversion on the service performance data and the hardware data to obtain training data;
a decision instruction is used for acquiring a trained current classification model and inputting the training data into the current classification model to obtain a scaling decision;
a capacity expansion and reduction instruction, which utilizes the capacity expansion and reduction decision to carry out service capacity expansion and reduction;
and training the current classification model by using the training data to obtain an updated classification model, and taking the updated classification model as the current classification model.
The invention can ensure high availability and high performance of user service, reduce access cost and access difficulty of the user side, improve automation degree, ensure optimization of capacity expansion and contraction strategies and prevent waste of hardware resources.
Drawings
FIG. 1 is a block diagram of a system for service scalability according to an embodiment of the present invention;
fig. 2 is a flow diagram of a method for service scaling according to an embodiment of the present invention.
Detailed Description
Embodiments of the present invention will be described in detail below with reference to the accompanying drawings.
Fig. 1 is a block diagram of a system 1 for service scalability according to an embodiment of the present invention. As shown in fig. 1, the system 1 includes a routing unit 11, a data obtaining unit 12, a converting unit 13, a training unit 14, a decision unit 15, and a scaling unit 16. The system 1 may include one or more routing units 11 and one or more data acquisition units 12, among others.
Only one routing unit 11 and one data acquisition unit 12 are shown in fig. 1.
Fig. 2 is a flow diagram of a method for service scaling according to an embodiment of the present invention. An embodiment of the present invention will be described in detail below with reference to fig. 1 and 2.
As shown in fig. 2, in the request forwarding step S21, the routing unit 11 forwards the service request from the user. Wherein, the routing unit 11 forwards the service request to the corresponding data obtaining unit 12 according to the content of the service request.
In the data acquisition step S22, the corresponding data acquisition unit 12 receives the service request, calculates service performance data from the service request, and collects hardware data.
The data acquiring unit 12 includes a user service unit 121 and a collecting unit 122. The user service unit 121 receives the service request and calculates service performance data according to the service request. The service performance data is, for example, the query volume per second of the service (service qps), the service response time, the service failure rate, and the like calculated from the service request. The subscriber service unit 121 transmits the calculated service performance data to the conversion unit 13.
Here, each data acquisition unit 12 includes a user service unit 121 and a collection unit 122 corresponding to the user service unit 121, and the user service unit 121 and the collection unit 122 are located on the same physical machine. The collection unit 122 collects hardware data of the physical machine. The hardware data refers to the usage of the hardware resources of the physical machine where the collection unit 122 is located, such as the usage of cpu, gpu (graphics processor), memory, etc., network bandwidth, etc. The collection unit 122 will transmit the currently collected hardware data to the conversion unit 13.
In the conversion step S23, the conversion unit 13 performs format conversion on the service performance data received from the user service unit 121 and the hardware data received from the collection unit 122 to obtain training data. Wherein the training data format conforms to the format of training performed in training unit 14. The conversion unit 13 will pass the training data to the decision unit 15 and the training unit 14.
In decision step S24, the decision unit 15 obtains the trained current classification model (e.g., classification model a) from the training unit 14, and inputs the training data into the current classification model, resulting in a scaling decision.
The current classification model is a model trained by a machine learning algorithm according to historical data, which is received by the training unit 14 before the current training data is acquired. These historical training data may be training data previously acquired and converted from different data acquisition units 12 in the system 1.
Next, in the scaling step S25, the scaling unit 16 performs service scaling using the scaling decision. Wherein, the service expansion is to expand the routing unit 11 and/or the user service unit 121. Specifically, the capacity expansion and reduction unit 16 determines whether to increase or decrease the routing unit 11 and/or the user service unit 121 in the system 1 according to the capacity expansion and reduction decision.
In training step 26, training section 14 trains classification model a using the training data transmitted from forwarding section 13 to obtain an updated classification model (for example, classification model B), and sets classification model B as the current classification model.
Before the training data is received, the classification model a trained using the history data using the machine learning algorithm is stored in the training unit 14. The decision unit 15 now obtains the classification model a, and obtains the above-mentioned scaling decision.
After receiving the training data, the training unit 14 uses the machine learning algorithm to train (adapt) the classification model a again with the training data, e.g. resulting in a more optimized classification model B. Since training takes time, the decision unit 15 uses the classification model a when making the decision this time, and obtains and uses the optimized classification model B when making the next decision. In this way, the current classification model can be retrained again according to the training data received in real time to obtain a continuously optimized classification model, thereby ensuring the decision optimization made by the decision unit 15.
In the invention, the service performance data can be acquired in real time from the current service request, the use condition of the hardware resource can be acquired in real time, and the service performance data and the use condition of the hardware resource are input into the machine learning classification model which is trained in advance, so that whether the capacity expansion is needed or not and the quantity of the capacity expansion are determined. The introduction of service performance data ensures high availability and high performance of user services, thereby reducing the access cost and the access difficulty of a user side and improving the automation degree.
In addition, after receiving the new training data, the training unit 15 trains the current classification model again by using the new training data to obtain an optimized updated classification model, and the optimized updated classification model is used for determining a scaling capacity decision next time. Therefore, the classification model of the invention is not invariable, but is optimized (self-adaptive) in real time according to new training data, thereby ensuring the optimization of the capacity expansion and contraction strategy and preventing the waste of hardware resources.
Preferably, the system 1 of the present invention can be implemented by a container cluster managed by the conventional kubernets (K8S, an existing container management system), and the expansion and contraction unit can be implemented by an api server in K8S.
The present invention also provides a nonvolatile storage medium on which a program for performing service expansion is stored, the program being executed by a computer to implement a method of service expansion, the program including:
a request forwarding instruction for forwarding a service request from a user;
a data acquisition instruction, receiving the service request, calculating service performance data according to the service request, and collecting hardware data;
a conversion instruction, which is used for carrying out format conversion on the service performance data and the hardware data to obtain training data;
a decision instruction, namely acquiring a trained current classification model, and inputting the training data into the current classification model to obtain a scaling decision;
a capacity expansion and reduction instruction, which utilizes the capacity expansion and reduction decision to carry out service capacity expansion and reduction;
and training the current classification model by using the training data to obtain an updated classification model, and taking the updated classification model as the current classification model.
While the present invention has been described in conjunction with specific embodiments, it is evident that many alternatives, modifications, and variations will be apparent to those skilled in the art in light of the foregoing description. Accordingly, it is intended that such alternatives, modifications, and variations be included within the spirit and scope of the appended claims.

Claims (11)

1. A method for API service scalability, the method comprising:
a request forwarding step of forwarding a service request from a user;
a data acquisition step, namely receiving the service request, calculating service performance data according to the service request, and collecting hardware data;
a conversion step, performing format conversion on the service performance data and the hardware data to obtain training data;
a decision step, namely acquiring a trained current classification model, and inputting the training data into the current classification model to obtain a scaling decision;
a capacity expansion and reduction step, namely utilizing the capacity expansion and reduction decision to perform service capacity expansion and reduction;
and a training step, namely training the current classification model by using the training data to obtain an updated classification model, and taking the updated classification model as the current classification model.
2. The method of claim 1, wherein in the data obtaining step, the user service unit receives the service request and calculates service performance data according to the service request.
3. The method of claim 2, wherein the training data conforms to a format in which training is performed in the training step.
4. The method of claim 3, wherein the updated classification model is obtained the next time the decision step is performed.
5. The method according to any of claims 2-4, wherein in the request forwarding step, a routing unit forwards the service request,
wherein the service scaling is scaling the routing unit and/or the user service unit.
6. A system for API service scale-up, the system comprising:
a routing unit for forwarding a service request from a user;
the data acquisition unit is used for receiving the service request, calculating service performance data according to the service request and collecting hardware data;
the conversion unit is used for carrying out format conversion on the service performance data and the hardware data to obtain training data;
the decision unit is used for acquiring a trained current classification model and inputting the training data into the current classification model to obtain a scaling decision;
the capacity expansion and reduction unit is used for carrying out service capacity expansion and reduction by utilizing the capacity expansion and reduction decision;
and the training unit is used for training the current classification model by using the training data to obtain an updated classification model, and taking the updated classification model as the current classification model.
7. The system of claim 6, wherein the data acquisition unit further comprises:
the user service unit receives the service request and calculates the service performance data according to the service request;
and the collecting unit is used for collecting the hardware data.
8. The system of claim 7, wherein the training data conforms to a format for training in the training unit.
9. The system of claim 8, wherein the updated classification model is obtained when the next decision is made by the decision unit.
10. A system according to any of claims 7-9, characterized in that said service scaling is scaling at least one routing element and/or subscriber service element.
11. A nonvolatile storage medium on which a program for API service expansion is stored, the program being executed by a computer to implement a method of service expansion, characterized by comprising:
a request forwarding instruction for forwarding a service request from a user;
a data acquisition instruction, receiving the service request, calculating service performance data according to the service request, and collecting hardware data;
a conversion instruction, which is used for carrying out format conversion on the service performance data and the hardware data to obtain training data;
a decision instruction, namely acquiring a trained current classification model, and inputting the training data into the current classification model to obtain a scaling decision;
a capacity expansion and reduction instruction, which utilizes the capacity expansion and reduction decision to carry out service capacity expansion and reduction;
and training the current classification model by using the training data to obtain an updated classification model, and taking the updated classification model as the current classification model.
CN201811619760.5A 2018-12-28 2018-12-28 Method, system and non-volatile storage medium for user service scaling Active CN109617738B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811619760.5A CN109617738B (en) 2018-12-28 2018-12-28 Method, system and non-volatile storage medium for user service scaling

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811619760.5A CN109617738B (en) 2018-12-28 2018-12-28 Method, system and non-volatile storage medium for user service scaling

Publications (2)

Publication Number Publication Date
CN109617738A CN109617738A (en) 2019-04-12
CN109617738B true CN109617738B (en) 2022-05-31

Family

ID=66011995

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811619760.5A Active CN109617738B (en) 2018-12-28 2018-12-28 Method, system and non-volatile storage medium for user service scaling

Country Status (1)

Country Link
CN (1) CN109617738B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112311578B (en) * 2019-07-31 2023-04-07 中国移动通信集团浙江有限公司 VNF scheduling method and device based on deep reinforcement learning
CN113032157B (en) * 2021-05-31 2021-08-24 睿至科技集团有限公司 Automatic intelligent server capacity expansion and reduction method and system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106959894A (en) * 2016-01-11 2017-07-18 北京京东尚科信息技术有限公司 Resource allocation methods and device
CN107038377A (en) * 2016-02-03 2017-08-11 阿里巴巴集团控股有限公司 A kind of site certificate method and device, website credit method and device
CN107404523A (en) * 2017-07-21 2017-11-28 中国石油大学(华东) Cloud platform adaptive resource dispatches system and method
CN108259225A (en) * 2017-12-20 2018-07-06 中国联合网络通信集团有限公司 A kind of network capacity extension appraisal procedure, device and server
CN108847956A (en) * 2018-05-08 2018-11-20 国家计算机网络与信息安全管理中心 A kind of scalable appearance method and system of multi-dimensional intelligent of telecommunications network safety service VNF

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106470219A (en) * 2015-08-17 2017-03-01 阿里巴巴集团控股有限公司 The dilatation of computer cluster and capacity reduction method and equipment
CN105631196B (en) * 2015-12-22 2018-04-17 中国科学院软件研究所 A kind of container levels flexible resource feed system and method towards micro services framework
US10338968B2 (en) * 2016-02-05 2019-07-02 Sas Institute Inc. Distributed neuromorphic processing performance accountability
CN108173698B (en) * 2018-01-17 2021-03-30 创新先进技术有限公司 Network service management method, device, server and storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106959894A (en) * 2016-01-11 2017-07-18 北京京东尚科信息技术有限公司 Resource allocation methods and device
CN107038377A (en) * 2016-02-03 2017-08-11 阿里巴巴集团控股有限公司 A kind of site certificate method and device, website credit method and device
CN107404523A (en) * 2017-07-21 2017-11-28 中国石油大学(华东) Cloud platform adaptive resource dispatches system and method
CN108259225A (en) * 2017-12-20 2018-07-06 中国联合网络通信集团有限公司 A kind of network capacity extension appraisal procedure, device and server
CN108847956A (en) * 2018-05-08 2018-11-20 国家计算机网络与信息安全管理中心 A kind of scalable appearance method and system of multi-dimensional intelligent of telecommunications network safety service VNF

Also Published As

Publication number Publication date
CN109617738A (en) 2019-04-12

Similar Documents

Publication Publication Date Title
CN109218355B (en) Load balancing engine, client, distributed computing system and load balancing method
CN111966289B (en) Partition optimization method and system based on Kafka cluster
CN111741073B (en) Electric power data transmission system based on 5G communication network
CN110545258B (en) Streaming media server resource allocation method and device and server
CN109617738B (en) Method, system and non-volatile storage medium for user service scaling
CN105703927A (en) Resource allocation method, network device and network system
CN112365366B (en) Micro-grid management method and system based on intelligent 5G slice
CN103986766A (en) Self-adaptation load balancing job task scheduling method and device
CN110881199A (en) Dynamic allocation method, device and system for network slice resources
CN102158346A (en) Information acquisition system and method based on cloud computing
CN104901989A (en) Field service providing system and method
EP3981111B1 (en) Allocating cloud resources in accordance with predicted deployment growth
CN103747077A (en) Transmission mechanism adjusting method, server side and client
CN112313997B (en) Dynamic allocation processing among nodes in a wireless mesh network
CN102480502B (en) I/O load equilibrium method and I/O server
CN115714814B (en) Edge cache replacement method based on multi-agent reinforcement learning
CN114780244A (en) Container cloud resource elastic allocation method and device, computer equipment and medium
CN110557679B (en) Video content identification method, device, medium and system
CN113840330A (en) Method for establishing connection, gateway equipment, network system and scheduling center
CN109067840A (en) Method, system and the storage medium of artificial intelligence online service
CN116235529A (en) Method for implementing an ad hoc network of a plurality of access network devices and electronic device for implementing the method
CN115460124B (en) Method, device, equipment and storage medium for optimizing transmission link across machine room
CN109474696B (en) Network service method, device, electronic equipment and readable storage medium
CN111913660A (en) Dotting data processing method and system
CN116860823A (en) Dynamic equipment fault diagnosis difficult sample bidirectional mining method based on joint knowledge distillation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant