WO2020098030A1 - 一种请求任务的调度方法及调度中心服务器 - Google Patents

一种请求任务的调度方法及调度中心服务器 Download PDF

Info

Publication number
WO2020098030A1
WO2020098030A1 PCT/CN2018/120101 CN2018120101W WO2020098030A1 WO 2020098030 A1 WO2020098030 A1 WO 2020098030A1 CN 2018120101 W CN2018120101 W CN 2018120101W WO 2020098030 A1 WO2020098030 A1 WO 2020098030A1
Authority
WO
WIPO (PCT)
Prior art keywords
training
weight value
node
scheduled
support vector
Prior art date
Application number
PCT/CN2018/120101
Other languages
English (en)
French (fr)
Inventor
林鹏程
Original Assignee
网宿科技股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 网宿科技股份有限公司 filed Critical 网宿科技股份有限公司
Priority to EP18940424.7A priority Critical patent/EP3879786A1/en
Priority to US16/966,900 priority patent/US20210049424A1/en
Publication of WO2020098030A1 publication Critical patent/WO2020098030A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/50Testing arrangements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/60Scheduling or organising the servicing of application requests, e.g. requests for application data transmissions using the analysis and optimisation of the required network resources
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • G06N20/10Machine learning using kernel methods, e.g. support vector machines [SVM]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • G06N20/20Ensemble learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/94Hardware or software architectures specially adapted for image or video understanding
    • G06V10/95Hardware or software architectures specially adapted for image or video understanding structured as a network, e.g. client-server architectures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/14Network analysis or design
    • H04L41/145Network analysis or design involving simulating, designing, planning or modelling of a network
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/16Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks using machine learning or artificial intelligence
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/04Processing captured monitoring data, e.g. for logfile generation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/06Generation of reports
    • H04L43/065Generation of reports related to network devices
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/08Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters
    • H04L43/0852Delays
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/08Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters
    • H04L43/0876Network utilisation, e.g. volume of load or congestion level
    • H04L43/0894Packet rate

Definitions

  • the present invention relates to the field of Internet technology, and in particular, to a scheduling method for requesting tasks and a scheduling center server.
  • CDN Content Delivery Network, content distribution network
  • the CDN control center when the CDN control center receives a request task sent by the user to load the live stream, it can determine the CDN suitable for processing the request task according to the parameters such as the load number and the jam rate of each CDN node in the current network Node and schedule the requested task to the determined CDN node.
  • the purpose of the present application is to provide a scheduling method of a request task and a dispatch center server, which can improve the scheduling accuracy of the request task.
  • the present application provides a method for scheduling a request task.
  • the method includes: receiving node information reported by a CDN node to be scheduled, and constructing multiple training samples based on the node information; creating a support vector Machine model, the support vector machine model includes a specified number of two classifiers, and the specified number is determined based on the total number of CDN nodes to be scheduled; the support vectors are constructed using the multiple training samples constructed
  • the machine model performs multiple rounds of training, and after each round of training, a corresponding weak classifier is generated, and the weak classifier has a weight value; based on the weight value of each weak classifier, each weak classifier is combined into a final Classifier, and schedule the received new request task in the CDN node to be scheduled through the final classifier.
  • another aspect of the present application also provides a dispatch center server, the dispatch center server includes: a training sample construction unit, configured to receive node information reported by a CDN node to be scheduled, and based on the node information, Construct multiple training samples; a support vector machine model creation unit, used to create a support vector machine model, the support vector machine model includes a specified number of two classifiers, and the specified number is based on the CDN node to be scheduled The total number is determined; the iterative training unit is used to perform multiple rounds of training on the support vector machine model using the constructed multiple training samples, each round of training generates a corresponding weak classifier, and the weak classifier has Weight value; a task scheduling unit for combining each of the weak classifiers into a final classifier based on the weight value of each of the weak classifiers, and using the final classifier to place the received new request task in all Perform scheduling in the CDN node to be scheduled.
  • a training sample construction unit configured to receive node information reported by a CDN node to be
  • another aspect of the present application also provides a dispatch center server, the dispatch center server includes a memory and a processor, the memory is used to store a computer program, and when the computer program is executed by the processor, A method for scheduling the above-mentioned request task is realized.
  • the technical solution provided by the present application can train various node information of the CDN node through machine learning, thereby obtaining a classifier capable of scheduling the requested task.
  • the scheduling center in the CDN can receive various node information reported by multiple CDN nodes to be scheduled, and can construct multiple training samples based on the node information.
  • a support vector machine (Support Vector Machine, SVM) model containing multiple binary classifiers can be created, and then the SVM model can be trained in multiple rounds using the above training samples to generate multiple weak classifiers.
  • the purpose of including multiple binary classifiers in the SVM model is to convert a multiple classification problem into multiple binary classification problems, so that it can smoothly schedule more than two CDN nodes.
  • the weight values of each weak classifier can be determined separately, and according to the determined weight values, the multiple weak classifiers can be combined into a final classifier.
  • the weight value of each weak classifier may indicate the role of the weak classifier in the final classifier.
  • the node information of each CDN node to be scheduled can be analyzed through the final classifier to determine a target CDN node suitable for processing the new request task.
  • the technical solution provided by this application can comprehensively consider various aspects of node information through machine learning, and through continuous correction of training samples, a final classifier with high enough accuracy can be generated, thereby improving the scheduling of requested tasks Precision.
  • FIG. 1 is a schematic structural diagram of a dispatch center server and a CDN node in an embodiment of the present invention
  • FIG. 2 is a flowchart of a method for scheduling a requested task in an embodiment of the present invention
  • FIG. 3 is a schematic diagram of a DAG model in an embodiment of the present invention.
  • FIG. 4 is a schematic diagram of a multi-round training process in an embodiment of the present invention.
  • FIG. 5 is a schematic diagram of functional modules of a dispatch center server in an embodiment of the present invention.
  • FIG. 6 is a schematic structural diagram of a dispatch center server in an embodiment of the present invention.
  • This application provides a scheduling method for requesting tasks, which can be applied to a CDN scheduling center server.
  • the dispatch center server can receive a request task for pulling a DC video stream from a user client, and then can analyze the node information of multiple CDN nodes currently to be scheduled to determine that the To which CDN node the task is scheduled.
  • the method for scheduling a request task may include the following steps.
  • S1 Receive the node information reported by the CDN node to be scheduled, and construct multiple training samples based on the node information.
  • the technical solution provided by this application may include a training phase and a prediction phase.
  • multiple training samples can be constructed in advance.
  • multiple CDN nodes to be scheduled can collect their respective node information, and upload the collected node information to the dispatch center server.
  • the node information may include various information such as performance parameters, load parameters, remaining bandwidth, network delay and bit rate of live streaming.
  • the performance parameter can be represented by A i
  • the load parameter can be represented by L i
  • the amount of remaining bandwidth can be represented by B i
  • the network delay can be represented by P i
  • the bit rate of the live stream can be represented by BR i .
  • the subscript i may represent the node information of the i-th CDN node to be scheduled.
  • the CDN node to be scheduled may periodically report the node information at different times to the control center server at specified time intervals.
  • the scheduling center server can collect the node information reported by the CDN node to be scheduled at different times, and construct the node information reported by each of the CDN nodes to be scheduled at the same time as an information vector.
  • the information vector can be expressed in the following manner:
  • I [A 1 , L 1 , B 1 , P 1 , BR 1 , A 2 , L 2 , B 2 , P 2 , BR 2 ... A n , L n , B n , P n , BR n ] ;
  • I may represent an information vector at the current time, and the information vector may include information about each node sent by the n CDN nodes to be scheduled at the current time.
  • an information vector as shown above can be generated at each different time, and each information vector constructed at different times can be used as the multiple training samples constructed.
  • m training samples may be constructed for m node information reported at different times, and each training sample may be represented in the form of the above-mentioned information vector.
  • S3 Create a support vector machine model, the support vector machine model includes a specified number of two classifiers, and the specified number is determined based on the total number of CDN nodes to be scheduled.
  • an SVM model in order to complete the machine learning process, can be created, which can analyze the node information of multiple CDN nodes to be scheduled to determine to which CDN the request task should be scheduled at present node. Therefore, the created SVM model is actually a multi-classifier, and the number of classifications supported by the multi-classifier is consistent with the number of CDN nodes to be scheduled. For example, there are currently 5 CDN nodes participating in scheduling, then the SVM model needs to select a suitable CDN node from the 5 CDN nodes. Therefore, the SVM model is equivalent to a 5 classifier.
  • the SVM model with multi-classification function can be constructed by multiple two classifiers.
  • a DAG Directed Acyclic Graph, Directed Acyclic Graph
  • FIG. 3 assuming that one CDN node is to be selected from the 5 CDN nodes, and the labels of these 5 CDN nodes may be from 1 to 5, then in FIG. 3, one of the five problems can be disassembled.
  • 1-5 means to select a CDN node from the CDN nodes labeled 1 and 5.
  • the paired numbers in other circles can be deduced by analogy.
  • the SVM model may include a specified number of two classifiers, and the specified number has a certain correlation with the total number of CDN nodes to be scheduled.
  • the first item and the last item in the arithmetic sequence can be determined according to the total number of CDN nodes to be scheduled.
  • the first term of the arithmetic sequence may be 1, and the last term may be q-1, where q may represent the total number of CDN nodes to be scheduled.
  • the sum of the arithmetic sequence may be calculated, and the calculated sum of the arithmetic sequence may be used as the specified number.
  • the specified quantity can be determined according to the following formula:
  • the total number of CDN nodes participating in scheduling can be pre-stated, and then the number of two classifiers that should be included in the SVM model is calculated according to the above formula, and finally these two classifiers can be constructed according to the DAG model Layer by layer.
  • the SVM model when the SVM model is constructed, the SVM model can be used for multiple rounds of training using the multiple training samples.
  • each training sample may be assigned the same initial weight value.
  • the initial weight value may be, for example, 1 / m, where m represents the total number of training samples. In this way, the sum of the weight values of these training samples can be guaranteed to be 1.
  • the dispatch center server can query the node information corresponding to the historical time from the node information reported by the CDN node, and construct the above information vector according to the queried node information to obtain the training sample at the historical time.
  • the training sample already has a standard result: the requested task should be scheduled to the CDN node labeled 5. Therefore, after using the training sample to train the SVM model, if the training result output by the SVM model is also a CDN node labeled 5, it means that the training result for the training sample is consistent with the standard result. If the training result output by the SVM model is not a CDN node labeled 5, it means that the training result for the training sample is inconsistent with the standard result. In this case, the training process needs to be corrected so that when the training is performed again after the correction , So that the training results can be consistent with the standard results.
  • the error of the current round of training can be determined according to the difference between the training results and the standard results of the training samples function. Specifically, for the current training sample among the plurality of training samples, a determination value of the current training sample may be determined. Wherein, if the training result of the current training sample is the same as the standard result, the judgment value may be set to 0, and if the training result of the current training sample is different from the standard result, the judgment value may be set to 1. Then, the product of the initial weight value of the current training sample and the decision value may be calculated, so as to obtain the error contribution value corresponding to the current training sample. In the above manner, the error contribution value of each training sample can be calculated. Finally, the sum of error contribution values corresponding to each of the training samples can be used as the error function. In an actual application example, the error function can be determined according to the following formula:
  • I i the i-th training sample
  • W 0 (i) the initial weight value corresponding to the i-th training sample
  • G (I i ) the i-th training sample in the current round Training results
  • y i the standard result of the i-th training sample
  • m the number of training samples
  • the SVM model after this round of training can be regarded as a weak classifier.
  • the weight value of each training sample can be adjusted according to the above-mentioned error function, so that the adjusted The weighted training samples are used for the next training process.
  • a weight value can also be set for the weak classifier after the training round.
  • the weight value of the weak classifier can characterize the influence of the weak classifier on the final classification result. Specifically, the weight value of the weak classifier corresponding to the current round can be determined according to the following formula:
  • represents the weight value of the weak classifier corresponding to the current round.
  • each training sample can be assigned a new weight value according to the following formula:
  • W 1 (i) represents the new weight value assigned to the i-th training sample
  • Z represents the normalization factor, which can make the sum of the assigned new weight values always be 1.
  • the weight value of the weak classifier corresponding to the current round can be determined, and based on the determined weight value of the weak classifier, a new weight value can be newly assigned to each training sample, so that Using the training samples assigned new weight values, the support vector machine model is trained for the next round.
  • the process of the next round of training is consistent with the content described above, except that the weight value of each training sample has changed.
  • the error function can also be calculated again and determined by the error function
  • the weight value of the weak classifier in the next round, and the weight value is reassigned for each training sample again, and so on, until the training process of all rounds is completed.
  • each of the weak classifiers can be combined into a final classifier by weighted summation.
  • the prediction phase can be entered.
  • the control center server when the control center server receives the new request task, it can collect the node information of each of the CDN nodes to be scheduled in real time, and construct a test sample according to the collected node information in the manner of constructing the information vector described above. Then, the test sample may be input into the final classifier, so as to obtain a node identifier characterizing the target CDN node.
  • the node identifier can be, for example, the label of the CDN node, and then the control center server can schedule the new request task to the target CDN node.
  • the prediction sample of this time can be used as a new training sample, and the actual result that should be scheduled as the standard result of the training sample, Train the final classifier again to improve the scheduling accuracy of the final classifier. It can be seen that, in the case of scheduling misjudgment in practical applications, the final classifier can be continuously trained through machine learning, thereby further improving the classification accuracy of the final classifier.
  • the SVM model may classify the training sample by classifying the hyperplane.
  • the expression of the SVM model can be expressed as:
  • f (x) greater than 0 or equal to 0 can indicate different classification results.
  • x represents the input training sample
  • ⁇ (x) represents a certain mapping calculation on the training sample
  • w and b represent two coefficients.
  • ⁇ i represents the relaxation variable
  • C represents the penalty factor
  • y i represents the standard result of the training sample
  • l represents the number of elements contained in the training sample.
  • a kernel function can be used to map the input training samples to a high-dimensional space.
  • the kernel function may be a radial basis kernel function, and the mathematical expression is:
  • x i represents the i-th element in the training sample
  • x j represents the j-th element in the training sample
  • is an adjustable preset constant
  • classification expression of the support vector machine model may be:
  • f (x) represents the classification expression of the support vector machine model
  • K (x i , x) represents the radial basis kernel function
  • x i represents the i-th element in the training sample
  • x represents the input training sample
  • B * represents the redundancy factor
  • a i represents the i-th element in the optimal Lagrange multiplier
  • l represents the total number of elements in the training sample
  • SV represents the support vector field
  • N nsv represents the number of support vectors
  • C represents the penalty factor
  • y i represents the standard of the i-th training sample
  • represents the real part of the relaxation variable.
  • suitable parameters ⁇ and C can be found in practical applications. Specifically, it can be determined by means of grid optimization. First, the initial search is performed with an accuracy of 0.1, and after the area with a higher accuracy rate is obtained, the further search is performed with an accuracy of 0.01, thereby obtaining the optimal parameters ⁇ and C .
  • the dispatch center server includes:
  • a training sample construction unit configured to receive the node information reported by the CDN node to be scheduled, and construct multiple training samples based on the node information
  • a support vector machine model creation unit used to create a support vector machine model, the support vector machine model includes a specified number of two classifiers, and the specified number is determined based on the total number of CDN nodes to be scheduled;
  • An iterative training unit configured to perform multiple rounds of training on the support vector machine model using the constructed multiple training samples, each round of training generates a corresponding weak classifier, and the weak classifier has a weight value;
  • a task scheduling unit configured to combine each of the weak classifiers into a final classifier based on the weight value of each of the weak classifiers, and use the final classifier to place the received new request task in the Scheduling in the CDN node.
  • the iterative training unit includes:
  • An initial weight allocation module configured to allocate an initial weight value to each of the training samples in advance, and use the training samples with the initial weight value to train the support vector machine model;
  • the error function determination module is used to determine the error function of the current round of training according to the difference between the training result and the standard result of the training sample;
  • the weight value reassignment module is used to determine the weight value of the weak classifier corresponding to the current round according to the error function, and reallocate a new one for each training sample based on the determined weight value of the weak classifier Weights;
  • the task scheduling unit includes:
  • the node information collection module is used to collect the node information of each CDN node to be scheduled when receiving a new request task, and construct a test sample according to the collected node information;
  • a node identification prediction module used to input the test sample into the final classifier to obtain a node identification that characterizes the target CDN node;
  • a scheduling module is used to schedule the new request task to the target CDN node.
  • the dispatch center server may further include more or fewer components than those shown in FIG. 6, for example, may also include other processing hardware, such as GPU (Graphics Processing Unit, image processor), or have the same as FIG. 6 Different configurations shown.
  • GPU Graphics Processing Unit, image processor
  • this application does not exclude other implementations, such as logic devices or a combination of software and hardware.
  • the processor may include a central processing unit (CPU) or a graphics processor (GPU), of course, it may also include other single-chip computers, logic gate circuits, integrated circuits, etc. with logic processing capabilities, or their appropriate combination.
  • the memory described in this embodiment may be a memory device for storing information.
  • the device that can save binary data can be a memory; in an integrated circuit, a circuit that does not have a physical form with a storage function can also be a memory, such as RAM, FIFO, etc .; in the system, has a physical form of storage
  • the device can also be called a memory, etc.
  • the storage can also be implemented in the form of cloud storage. The specific implementation is well limited in this specification.
  • the technical solution provided by the present application can train various node information of the CDN node through machine learning, thereby obtaining a classifier capable of scheduling the requested task.
  • the scheduling center in the CDN can receive various node information reported by multiple CDN nodes to be scheduled, and can construct multiple training samples based on the node information.
  • a support vector machine (Support Vector Machine, SVM) model containing multiple binary classifiers can be created, and then the SVM model can be trained in multiple rounds using the above training samples to generate multiple weak classifiers.
  • the purpose of including multiple binary classifiers in the SVM model is to convert a multiple classification problem into multiple binary classification problems, so that it can smoothly schedule more than two CDN nodes.
  • the weight values of each weak classifier can be determined separately, and according to the determined weight values, the multiple weak classifiers can be combined into a final classifier.
  • the weight value of each weak classifier may indicate the role of the weak classifier in the final classifier.
  • the node information of each CDN node to be scheduled can be analyzed through the final classifier to determine a target CDN node suitable for processing the new request task.
  • the technical solution provided by this application can comprehensively consider various aspects of node information through machine learning, and through continuous correction of training samples, a final classifier with high enough accuracy can be generated, thereby improving the scheduling of requested tasks Precision.
  • each embodiment can be implemented by means of software plus a necessary general hardware platform, and of course, can also be implemented by hardware.
  • the above-mentioned technical solutions can be embodied in the form of software products in essence or to contribute to the existing technology, and the computer software products can be stored in computer-readable storage media, such as ROM / RAM, magnetic Discs, optical discs, etc., include several instructions to enable a computer device (which may be a personal computer, server, or network device, etc.) to execute the methods described in the various embodiments or some parts of the embodiments.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Software Systems (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Medical Informatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Multimedia (AREA)
  • Databases & Information Systems (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

本发明公开了一种请求任务的调度方法及控制中心服务器,其中,所述方法包括:接收待调度的CDN节点上报的节点信息,并基于所述节点信息,构建多个训练样本;创建支持向量机模型,所述支持向量机模型中包括指定数量的二分类器,并且所述指定数量基于所述待调度的CDN节点的总数量确定;利用构建的所述多个训练样本对所述支持向量机模型进行多轮训练,每轮训练后均生成对应的弱分类器,并且所述弱分类器具备权重值;基于各个所述弱分类器的权重值,将各个所述弱分类器组合为最终分类器,并通过所述最终分类器将接收到的新的请求任务在所述待调度的CDN节点中进行调度。本申请提供的技术方案,能够提高请求任务的调度精度。

Description

一种请求任务的调度方法及调度中心服务器 技术领域
本发明涉及互联网技术领域,特别涉及一种请求任务的调度方法及调度中心服务器。
背景技术
随着直播行业的不断兴起,为了给用户提供流畅的视频直播体验,直播服务商通常会选用CDN(Content Delivery Network,内容分发网络)分担直播流,并且对直播流进行加速。
目前,CDN的控制中心在接收到用户发来的用于加载直播流的请求任务时,可以根据当前网络中各个CDN节点的负载数、卡顿率等参数,确定出适合处理该请求任务的CDN节点,并将该请求任务调度至确定出的CDN节点处。
然而,目前对于请求任务的调度方法,对于CDN节点的判断依据过于单一,因此可能会造成误判,从而导致请求任务分配不均的情况。而一旦增加判断依据,会使得判断过程过于复杂,目前还没有合适的方案能够应对复杂的判断过程。
发明内容
本申请的目的在于提供一种请求任务的调度方法及调度中心服务器,能够提高请求任务的调度精度。
为实现上述目的,本申请一方面提供一种请求任务的调度方法,所述方法包括:接收待调度的CDN节点上报的节点信息,并基于所述节点信息,构建多个训练样本;创建支持向量机模型,所述支持向量机模型中包括指定数量的二分类器,并且所述指定数量基于所述待调度的CDN节点的总数量确定;利用构建的所述多个训练样本对所述支持向量机模型进行多轮训练,每轮训练后均生成对应的弱分类器,并且所述弱分类器具备权重值;基于各个所述弱分类器的权重值,将各个所述弱分类器组合为最终分类器,并通过所述最终分类器将接 收到的新的请求任务在所述待调度的CDN节点中进行调度。
为实现上述目的,本申请另一方面还提供一种调度中心服务器,所述调度中心服务器包括:训练样本构建单元,用于接收待调度的CDN节点上报的节点信息,并基于所述节点信息,构建多个训练样本;支持向量机模型创建单元,用于创建支持向量机模型,所述支持向量机模型中包括指定数量的二分类器,并且所述指定数量基于所述待调度的CDN节点的总数量确定;迭代训练单元,用于利用构建的所述多个训练样本对所述支持向量机模型进行多轮训练,每轮训练后均生成对应的弱分类器,并且所述弱分类器具备权重值;任务调度单元,用于基于各个所述弱分类器的权重值,将各个所述弱分类器组合为最终分类器,并通过所述最终分类器将接收到的新的请求任务在所述待调度的CDN节点中进行调度。
为实现上述目的,本申请另一方面还提供一种调度中心服务器,所述调度中心服务器包括存储器和处理器,所述存储器用于存储计算机程序,所述计算机程序被所述处理器执行时,实现上述的请求任务的调度方法。
由上可见,本申请提供的技术方案,可以通过机器学习的方式,对CDN节点的各项节点信息进行训练,从而得到能够进行请求任务调度的分类器。具体地,CDN中的调度中心可以接收待调度的多个CDN节点上报的各项节点信息,并可以基于这些节点信息构建多个训练样本。后续,可以创建包含多个二分类器的支持向量机(Support Vector Machine,SVM)模型,然后利用上述的训练样本对该SVM模型进行多轮训练,从而生成多个弱分类器。其中,SVM模型中包含多个二分类器的目的是,可以将一个多分类的问题转换为多个二分类的问题,从而能够顺利地对两个以上的CDN节点进行调度。在得到多个弱分类器之后,可以分别确定各个弱分类器的权重值,并根据确定的权重值,将多个弱分类器组合为最终分类器。各个弱分类器的权重值大小可以表示弱分类器在最终分类器中所起的作用大小。这样,通过大量训练样本对SVM模型进行训练后,便可以得到精准的最终分类器。后续当接收到新的请求任务时,可以通过该最终分类器对各个待调度的CDN节点的节点信息进行分析,从而确定出适合处理该新的请求任务的目标CDN节点。这样,本申请提供的技术方案,通过机器学习的方式,能够对多方面的节点信息进行综合考量,并且通过训练样本的不断校正,能够产生精度足够高的最终分类器,从而提高请求任务的调度精度。
附图说明
为了更清楚地说明本发明实施例中的技术方案,下面将对实施例描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。
图1是本发明实施方式中调度中心服务器与CDN节点的结构示意图;
图2是本发明实施方式中请求任务的调度方法流程图;
图3是本发明实施方式中DAG模型的示意图;
图4是本发明实施方式中多轮训练的流程示意图;
图5是本发明实施方式中调度中心服务器的功能模块示意图;
图6是本发明实施方式中调度中心服务器的结构示意图。
具体实施方式
为使本发明的目的、技术方案和优点更加清楚,下面将结合附图对本发明实施方式作进一步地详细描述。
本申请提供一种请求任务的调度方法,所述方法可以应用于CDN的调度中心服务器中。请参阅图1,所述调度中心服务器可以接收用户客户端发来的拉取直流视频流的请求任务,然后可以对当前待调度的多个CDN节点的节点信息进行分析,从而确定出应当将该请求任务调度至哪一个CDN节点处。
具体地,请参阅图2,本申请提供的请求任务的调度方法,可以包括以下步骤。
S1:接收待调度的CDN节点上报的节点信息,并基于所述节点信息,构建多个训练样本。
本申请提供的技术方案,可以包括训练阶段和预测阶段。针对训练阶段,可以预先构建多个训练样本。具体地,待调度的多个CDN节点可以采集各自的节点信息,并将采集的节点信息上传至调度中心服务器。其中,所述节点信息可以包括性能参数、负载参数、剩余带宽量、网络延时以及直播流的码率等多方面的信息。针对不同的信息,可以通过不同的物理量来表示。具体地,性能参数可以用A i表示、负载参数可以用L i表示、剩余带宽量可以用B i表示、网络延 时可以用P i表示以及直播流的码率可以用BR i表示。其中,下标i可以表示第i个待调度的CDN节点的节点信息。
在本实施方式中,待调度的CDN节点可以按照指定的时间间隔,定期向控制中心服务器上报不同时刻的节点信息。这样,调度中心服务器可以收集所述待调度的CDN节点在不同时刻上报的节点信息,并将同一时刻各个所述待调度的CDN节点上报的节点信息构建为一个信息向量。所述信息向量可以通过以下方式表示:
I=[A 1,L 1,B 1,P 1,BR 1,A 2,L 2,B 2,P 2,BR 2...A n,L n,B n,P n,BR n];
其中,I可以表示当前时刻的信息向量,该信息向量中可以包括n个待调度的CDN节点在当前时刻发来的各项节点信息。这样,每个不同的时刻均可以生成如上所示的一个信息向量,不同时刻构建的各个信息向量,便可以作为构建的所述多个训练样本。例如,在本实施方式中,可以针对m个不同时刻上报的节点信息构建m个训练样本,每个训练样本均可以通过上述的信息向量的形式来表示。
S3:创建支持向量机模型,所述支持向量机模型中包括指定数量的二分类器,并且所述指定数量基于所述待调度的CDN节点的总数量确定。
在本实施方式中,为了完成机器学习的过程,可以创建一个SVM模型,该SVM模型可以针对多个待调度的CDN节点的节点信息进行分析,从而确定出目前应当将请求任务调度至哪一个CDN节点。因此,创建的该SVM模型实际上就是一个多分类器,该多分类器所支持的分类数量,与待调度的CDN节点的数量一致。例如,当前共计有5个CDN节点参与调度,那么该SVM模型就需要从这5个CDN节点中选取一个合适的CDN节点,因此,该SVM模型就相当于一个5分类器。
在本实施方式中,具备多分类功能的SVM模型可以通过多个二分类器构建。具体地,可以通过DAG(Directed Acyclic Graph,有向无环图)模型,将一个多分类的问题拆解为多个二分类的问题。请参阅图3,假设现在要从5个CDN节点中挑选出一个CDN节点,这5个CDN节点的标号可以是从1至5,那么在图3中,可以将五选一的问题,拆解为10个二选一的问题,其中,1-5表示从标号为1和标号为5的CDN节点中挑选一个CDN节点,其它圆圈内的成对数字可以以此类推。这样,在当前层完成二选一的问题之后,可以根据选择的结果,进 入到下一层的二选一问题,最终可以得到5个标号中的一个标号,最终得到的该标号对应的CDN节点便可以作为请求任务应当调度至的CDN节点。
在本实施方式中,SVM模型中可以包含指定数量的二分类器,而该指定数量与待调度的CDN节点的总数量具备一定的关联。具体地,首先可以根据所述待调度的CDN节点的总数量,分别确定等差数列中的首项和末项。其中,所述等差数列的首项可以是1,末项可以是q-1,其中,q可以表示所述待调度的CDN节点的总数量。然后可以基于确定的所述首项和所述末项,计算所述等差数列的和,并将计算的所述等差数列的和作为所述指定数量。在一个实际应用示例中,所述指定数量可以按照以下公式确定:
Figure PCTCN2018120101-appb-000001
其中,P表示所述指定数量,q表示所述待调度的CDN节点的总数量。
这样,在实际应用中,可以预先统计参与调度的CDN节点的总数量,然后再按照上式计算出SVM模型中应当包含的二分类器的数量,最终便可以按照DAG模型构建出这些二分类器的逐层排布情况。
S5:利用构建的所述多个训练样本对所述支持向量机模型进行多轮训练,每轮训练后均生成对应的弱分类器,并且所述弱分类器具备权重值。
在本实施方式中,在构建出所述SVM模型时,便可以利用所述多个训练样本对所述SVM模型进行多轮训练。在利用训练样本进行训练时,首先需要为各个训练样本设置对应的权重值。在第一次训练时,可以为每个所述训练样本分配相同的初始权重值。具体地,该初始权重值例如可以是1/m,其中,m表示训练样本的总个数。这样,这些训练样本的权重值之和可以保证为1。
在本实施方式中,在为各个训练样本分配了初始权重值之后,便可以利用具备所述初始权重值的训练样本对所述支持向量机模型进行训练,训练的目的是,让SVM模型对训练样本的训练结果,能够逼近训练样本的标准结果。其中,所述训练样本的标准结果可以是调度中心服务器在接收节点信息时一并获取的。具体地,在构建训练样本时所采用的节点信息,都可以是已经完成请求任务调度的历史信息。举例来说,在某个历史时刻,调度中心服务器接收到用户客户端发来的请求任务,并经过一系列计算之后,将该请求任务调度至标号为5的CDN节点中,并且本次调度后续被证明是正确的调度。那么,调度中心服务器可以从CDN节点上报的节点信息中,查询出该历史时刻对应的节点信息,并 按照查询出的节点信息构建上述的信息向量,从而得到该历史时刻的训练样本。同时,该训练样本已经具备了标准结果:请求任务应当被调度至标号为5的CDN节点。因此,在利用该训练样本对SVM模型进行训练后,如果SVM模型输出的训练结果也是标号为5的CDN节点,则表示针对该训练样本的训练结果与标准结果是一致的。若SVM模型输出的训练结果不是标号为5的CDN节点,则表示针对该训练样本的训练结果与标准结果不一致,在这种情况下就需要对训练过程进行校正,以使得校正之后再次进行训练时,使得训练结果能够与标准结果一致。
鉴于此,在本实施方式中,在利用具备初始权重值的训练样本对所述SVM模型进行训练后,可以根据训练结果与所述训练样本的标准结果之间的差异,确定本轮训练的误差函数。具体地,针对所述多个训练样本中的当前训练样本,可以确定所述当前训练样本的判定数值。其中,若所述当前训练样本的训练结果与标准结果相同,可以将所述判定数值置为0,若所述当前训练样本的训练结果与标准结果不同,可以将所述判定数值置为1。然后,可以计算所述当前训练样本的初始权重值与所述判定数值的乘积,从而得到所述当前训练样本对应的误差贡献值。按照上述方式,便可以计算得到各个训练样本各自对应的误差贡献值。最终,可以将各个所述训练样本对应的误差贡献值之和作为所述误差函数。在一个实际应用示例中,可以按照以下公式确定所述误差函数:
Figure PCTCN2018120101-appb-000002
其中,er表示所述误差函数,I i表示第i个训练样本,W 0(i)表示第i个训练样本对应的初始权重值,G(I i)表示第i个训练样本在本轮的训练结果,y i表示第i个训练样本的标准结果,m表示所述训练样本的个数,f(*)表示若*成立,f(*)=1,若*不成立,f(*)=0。
由上可见,本轮训练结束后,对应的误差函数实际上就是被错误分类的训练样本的权重值之和。
在本实施方式中,本轮训练后的SVM模型可以视为一个弱分类器,在本轮训练结束之后,可以根据上述的误差函数,对各个训练样本的权重值进行调节,从而可以利用调节过权重值的训练样本进行下一轮的训练过程。同时,还可以为本轮训练后的弱分类器设置一个权重值,弱分类器的权重值可以表征该弱分类器对于最终分类结果的影响。具体地,可以按照以下公式确定本轮对应的弱 分类器的权重值:
Figure PCTCN2018120101-appb-000003
其中,α表示本轮对应的弱分类器的权重值。
另外,可以按照以下公式为每个所述训练样本分配新的权重值:
Figure PCTCN2018120101-appb-000004
其中,W 1(i)表示为第i个训练样本分配的新的权重值,Z表示归一化因子,该归一化因子可以使得分配的各个新的权重值之和始终为1。
这样,根据所述误差函数,可以确定本轮对应的弱分类器的权重值,并基于确定的所述弱分类器的权重值,重新为每个所述训练样本分配新的权重值,从而可以利用分配了新的权重值的训练样本,对所述支持向量机模型进行下一轮训练。
请参阅图4,下一轮训练的过程与上文描述的内容一致,只不过各个训练样本的权重值发生了改变,在下一轮训练结束之后,同样可以再次计算误差函数,并通过误差函数确定下一轮弱分类器的权重值,并且再次为各个训练样本重新分配权重值,以此类推,直至完成所有轮的训练过程。
S7:基于各个所述弱分类器的权重值,将各个所述弱分类器组合为最终分类器,并通过所述最终分类器将接收到的新的请求任务在所述待调度的CDN节点中进行调度。
在本实施方式中,所有轮的训练过程结束之后,便可以通过加权求和的方式,将各个所述弱分类器组合为最终分类器。
在本实施方式中,完成训练阶段之后,便可以进入预测阶段。具体地,当控制中心服务器接收到新的请求任务时,可以实时采集各个所述待调度的CDN节点的节点信息,并按照上述构建信息向量的方式,根据采集的所述节点信息构建测试样本。然后,可以将所述测试样本输入所述最终分类器,从而得到表征目标CDN节点的节点标识。该节点标识例如可以是CDN节点的标号,那么控制中心服务器便可以将所述新的请求任务调度至所述目标CDN节点处。
当然,如果后续发现本次的预测结果与真实的应当调度的结果存在偏差,则可以将本次的预测样本作为新的训练样本,并将真实的应当调度的结果作为该训练样本的标准结果,再次对最终分类器进行训练,从而完善最终分类器的 调度精度。由此可见,针对实际应用中存在调度误判的情况,可以通过机器学习的方式继续对最终分类器进行训练,从而进一步地提高最终分类器的分类精度。
在一个实施方式中,SVM模型接收到训练样本后,可以通过分类超平面来对该训练样本进行分类。具体地,所述SVM模型的表达式可以表示为:
f(x)=w·φ(x)+b;
其中,f(x)=0时可以表示分类超平面,而f(x)大于0或者等于0则可以表示不同的分类结果。其中,x表示输入的训练样本,φ(x)表示对训练样本进行一定的映射计算,w和b则表示两个系数。
在实际应用中,为了求解上式,可以将上述的表达式转换为以下带约束条件的解:
Figure PCTCN2018120101-appb-000005
其中,ξ i表示松弛变量,C表示惩罚因子,y i表示训练样本的标准结果,l表示训练样本中包含的元素个数。
进一步地,可以利用核函数将输入的训练样本映射到高维空间。其中,所述核函数可以为径向基核函数,数学表示为:
Figure PCTCN2018120101-appb-000006
其中,x i表示训练样本中的第i个元素,x j表示训练样本中的第j个元素,σ为可调节的预设常数。
后续,可以引入拉格朗日乘子,将上式转化为对偶问题,得到如下的对偶形式:
Figure PCTCN2018120101-appb-000007
在最优解为a=[a 1,a 2,…,a l],
Figure PCTCN2018120101-appb-000008
的情况下,可以求得各项参数如下,其中N nsv为支持向量的个数:
Figure PCTCN2018120101-appb-000009
Figure PCTCN2018120101-appb-000010
最终,所述支持向量机模型的分类表达式可以为:
Figure PCTCN2018120101-appb-000011
其中,f(x)表示所述支持向量机模型的分类表达式,K(x i,x)表示径向基核函数,x i表示训练样本中的第i个元素,x表示输入的训练样本,b *表示冗余因子,a i表示最优拉格朗日乘子中的第i个元素,
Figure PCTCN2018120101-appb-000012
表示a i的共轭转置,l表示训练样本中元素的总个数, SV表示支持向量域,N nsv表示支持向量的个数,C表示惩罚因子,y i表示第i个训练样本的标准结果,ε表示松弛变量的实部。
需要说明的是,为了达到更好的分类效果,在实际应用中可以寻找合适的参数σ和C。具体地,可以采用网格寻优的方式确定,先以0.1的精度进行初步搜寻,得到准确率较高的区域后,再以0.01的精度进行进一步的搜寻,从而得到最优的参数σ和C。
请参阅图5,本申请还提供一种调度中心服务器,所述调度中心服务器包括:
训练样本构建单元,用于接收待调度的CDN节点上报的节点信息,并基于所述节点信息,构建多个训练样本;
支持向量机模型创建单元,用于创建支持向量机模型,所述支持向量机模型中包括指定数量的二分类器,并且所述指定数量基于所述待调度的CDN节点的总数量确定;
迭代训练单元,用于利用构建的所述多个训练样本对所述支持向量机模型进行多轮训练,每轮训练后均生成对应的弱分类器,并且所述弱分类器具备权 重值;
任务调度单元,用于基于各个所述弱分类器的权重值,将各个所述弱分类器组合为最终分类器,并通过所述最终分类器将接收到的新的请求任务在所述待调度的CDN节点中进行调度。
在一个实施方式中,所述迭代训练单元包括:
初始权重分配模块,用于预先为每个所述训练样本分配初始权重值,并利用具备所述初始权重值的训练样本对所述支持向量机模型进行训练;
误差函数确定模块,用于根据训练结果与所述训练样本的标准结果之间的差异,确定本轮训练的误差函数;
权重值重新分配模块,用于根据所述误差函数,确定本轮对应的弱分类器的权重值,并基于确定的所述弱分类器的权重值,重新为每个所述训练样本分配新的权重值;
继续训练模块,用于利用分配了新的权重值的训练样本,对所述支持向量机模型进行下一轮训练。
在一个实施方式中,所述任务调度单元包括:
节点信息采集模块,用于在接收到新的请求任务时,采集各个所述待调度的CDN节点的节点信息,并根据采集的所述节点信息构建测试样本;
节点标识预测模块,用于将所述测试样本输入所述最终分类器,得到表征目标CDN节点的节点标识;
调度模块,用于将所述新的请求任务调度至所述目标CDN节点处。
请参阅图6,本申请还提供一种调度中心服务器,所述调度中心服务器包括存储器和处理器,所述存储器用于存储计算机程序,所述计算机程序被所述处理器执行时,可以实现如上述的请求任务的调度方法。具体地,如图6所示,在硬件层面,该调度中心服务器可以包括处理器、内部总线和存储器。所述存储器可以包括内存以及非易失性存储器。处理器从非易失性存储器中读取对应的计算机程序到内存中然后运行。本领域普通技术人员可以理解,图6所示的结构仅为示意,其并不对上述识别装置的结构造成限定。例如,所述调度中心服务器还可包括比图6中所示更多或者更少的组件,例如还可以包括其他的处理硬件,如GPU(Graphics Processing Unit,图像处理器),或者具有与图6所示不同的配置。当然,除了软件实现方式之外,本申请并不排除其他实 现方式,比如逻辑器件抑或软硬件结合的方式等等。
本实施方式中,所述的处理器可以包括中央处理器(CPU)或图形处理器(GPU),当然也可以包括其他的具有逻辑处理能力的单片机、逻辑门电路、集成电路等,或其适当组合。本实施方式所述的存储器可以是用于保存信息的记忆设备。在数字系统中,能保存二进制数据的设备可以是存储器;在集成电路中,一个没有实物形式的具有存储功能的电路也可以为存储器,如RAM、FIFO等;在系统中,具有实物形式的存储设备也可以叫存储器等。实现的时候,该存储器也可以采用云存储器的方式实现,具体实现方式,本说明书不错限定。
需要说明的是,本说明书中的调度中心服务器,具体的实现方式可以参照方法实施方式的描述,在此不作一一赘述。
由上可见,本申请提供的技术方案,可以通过机器学习的方式,对CDN节点的各项节点信息进行训练,从而得到能够进行请求任务调度的分类器。具体地,CDN中的调度中心可以接收待调度的多个CDN节点上报的各项节点信息,并可以基于这些节点信息构建多个训练样本。后续,可以创建包含多个二分类器的支持向量机(Support Vector Machine,SVM)模型,然后利用上述的训练样本对该SVM模型进行多轮训练,从而生成多个弱分类器。其中,SVM模型中包含多个二分类器的目的是,可以将一个多分类的问题转换为多个二分类的问题,从而能够顺利地对两个以上的CDN节点进行调度。在得到多个弱分类器之后,可以分别确定各个弱分类器的权重值,并根据确定的权重值,将多个弱分类器组合为最终分类器。各个弱分类器的权重值大小可以表示弱分类器在最终分类器中所起的作用大小。这样,通过大量训练样本对SVM模型进行训练后,便可以得到精准的最终分类器。后续当接收到新的请求任务时,可以通过该最终分类器对各个待调度的CDN节点的节点信息进行分析,从而确定出适合处理该新的请求任务的目标CDN节点。这样,本申请提供的技术方案,通过机器学习的方式,能够对多方面的节点信息进行综合考量,并且通过训练样本的不断校正,能够产生精度足够高的最终分类器,从而提高请求任务的调度精度。
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到各实施方式可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件来实现。基于这样的理解,上述技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品可以存储在计算机可读 存储介质中,如ROM/RAM、磁碟、光盘等,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行各个实施例或者实施例的某些部分所述的方法。
以上所述仅为本发明的较佳实施例,并不用以限制本发明,凡在本发明的精神和原则之内,所作的任何修改、等同替换、改进等,均应包含在本发明的保护范围之内。

Claims (15)

  1. 一种请求任务的调度方法,其特征在于,所述方法包括:
    接收待调度的CDN节点上报的节点信息,并基于所述节点信息,构建多个训练样本;
    创建支持向量机模型,所述支持向量机模型中包括指定数量的二分类器,并且所述指定数量基于所述待调度的CDN节点的总数量确定;
    利用构建的所述多个训练样本对所述支持向量机模型进行多轮训练,每轮训练后均生成对应的弱分类器,并且所述弱分类器具备权重值;
    基于各个所述弱分类器的权重值,将各个所述弱分类器组合为最终分类器,并通过所述最终分类器将接收到的新的请求任务在所述待调度的CDN节点中进行调度。
  2. 根据权利要求1所述的方法,其特征在于,所述节点信息中包括性能参数、负载参数、剩余带宽量、网络延时以及直播流的码率中的至少一种;
    相应地,所述构建多个训练样本包括:
    收集所述待调度的CDN节点在不同时刻上报的节点信息,并将同一时刻各个所述待调度的CDN节点上报的节点信息构建为一个信息向量;
    将不同时刻构建的各个信息向量作为构建的所述多个训练样本。
  3. 根据权利要求1所述的方法,其特征在于,所述指定数量按照以下方式确定:
    根据所述待调度的CDN节点的总数量,分别确定等差数列中的首项和末项;
    基于确定的所述首项和所述末项,计算所述等差数列的和,并将计算的所述等差数列的和作为所述指定数量。
  4. 根据权利要求1所述的方法,其特征在于,所述利用构建的所述多个训练样本对所述支持向量机模型进行多轮训练包括:
    预先为每个所述训练样本分配初始权重值,并利用具备所述初始权重值的训练样本对所述支持向量机模型进行训练;
    根据训练结果与所述训练样本的标准结果之间的差异,确定本轮训练的误差函数;
    根据所述误差函数,确定本轮对应的弱分类器的权重值,并基于确定的所述弱分类器的权重值,重新为每个所述训练样本分配新的权重值;
    利用分配了新的权重值的训练样本,对所述支持向量机模型进行下一轮训练。
  5. 根据权利要求4所述的方法,其特征在于,所述误差函数按照以下方式确定:
    针对所述多个训练样本中的当前训练样本,确定所述当前训练样本的判定数值;其中,若所述当前训练样本的训练结果与标准结果相同,将所述判定数值置为0,若所述当前训练样本的训练结果与标准结果不同,将所述判定数值置为1;
    计算所述当前训练样本的初始权重值与所述判定数值的乘积,得到所述当前训练样本对应的误差贡献值;
    将各个所述训练样本对应的误差贡献值之和作为所述误差函数。
  6. 根据权利要求4或5所述的方法,其特征在于,按照以下公式确定所述误差函数:
    Figure PCTCN2018120101-appb-100001
    其中,er表示所述误差函数,I i表示第i个训练样本,W 0(i)表示第i个训练样本对应的初始权重值,G(I i)表示第i个训练样本在本轮的训练结果,y i表示第i个训练样本的标准结果,m表示所述训练样本的个数,f(*)表示若*成立,f(*)=1,若*不成立,f(*)=0。
  7. 根据权利要求6所述的方法,其特征在于,按照以下公式确定本轮对应的弱分类器的权重值:
    Figure PCTCN2018120101-appb-100002
    其中,α表示本轮对应的弱分类器的权重值;
    相应地,按照以下公式为每个所述训练样本分配新的权重值:
    Figure PCTCN2018120101-appb-100003
    其中,W 1(i)表示为第i个训练样本分配的新的权重值,Z表示归一化因子。
  8. 根据权利要求1所述的方法,其特征在于,所述通过所述最终分类器将接收到的新的请求任务在所述待调度的CDN节点中进行调度包括:
    在接收到新的请求任务时,采集各个所述待调度的CDN节点的节点信息,并根据采集的所述节点信息构建测试样本;
    将所述测试样本输入所述最终分类器,得到表征目标CDN节点的节点标识;
    将所述新的请求任务调度至所述目标CDN节点处。
  9. 根据权利要求1所述的方法,其特征在于,所述支持向量机模型的分类表达式为:
    Figure PCTCN2018120101-appb-100004
    其中,f(x)表示所述支持向量机模型的分类表达式,K(x i,x)表示径向基核函数,x i表示训练样本中的第i个元素,x表示输入的训练样本,b *表示冗余因子,a i表示最优拉格朗日乘子中的第i个元素,
    Figure PCTCN2018120101-appb-100005
    表示a i的共轭转置,l表示训练样本中元素的总个数。
  10. 根据权利要求9所述的方法,其特征在于,所述冗余因子按照以下公式表示:
    Figure PCTCN2018120101-appb-100006
    其中,SV表示支持向量域,N nsv表示支持向量的个数,C表示惩罚因子,y i表示第i个训练样本的标准结果,ε表示松弛变量的实部。
  11. 根据权利要求9所述的方法,其特征在于,所述径向基核函数按照以下公式表示:
    Figure PCTCN2018120101-appb-100007
    其中,σ为预设常数。
  12. 一种调度中心服务器,其特征在于,所述调度中心服务器包括:
    训练样本构建单元,用于接收待调度的CDN节点上报的节点信息,并基于所述节点信息,构建多个训练样本;
    支持向量机模型创建单元,用于创建支持向量机模型,所述支持向量机模型中包括指定数量的二分类器,并且所述指定数量基于所述待调度的CDN节点的总数量确定;
    迭代训练单元,用于利用构建的所述多个训练样本对所述支持向量机模型进行多轮训练,每轮训练后均生成对应的弱分类器,并且所述弱分类器具备权重值;
    任务调度单元,用于基于各个所述弱分类器的权重值,将各个所述弱分类器组合为最终分类器,并通过所述最终分类器将接收到的新的请求任务在所述待调度的CDN节点中进行调度。
  13. 根据权利要求12所述的调度中心服务器,其特征在于,所述迭代训练单元包括:
    初始权重分配模块,用于预先为每个所述训练样本分配初始权重值,并利用具备所述初始权重值的训练样本对所述支持向量机模型进行训练;
    误差函数确定模块,用于根据训练结果与所述训练样本的标准结果之间的差异,确定本轮训练的误差函数;
    权重值重新分配模块,用于根据所述误差函数,确定本轮对应的弱分类器的权重值,并基于确定的所述弱分类器的权重值,重新为每个所述训练样本分配新的权重值;
    继续训练模块,用于利用分配了新的权重值的训练样本,对所述支持向量机模型进行下一轮训练。
  14. 根据权利要求12所述的调度中心服务器,其特征在于,所述任务调度单元包括:
    节点信息采集模块,用于在接收到新的请求任务时,采集各个所述待调度 的CDN节点的节点信息,并根据采集的所述节点信息构建测试样本;
    节点标识预测模块,用于将所述测试样本输入所述最终分类器,得到表征目标CDN节点的节点标识;
    调度模块,用于将所述新的请求任务调度至所述目标CDN节点处。
  15. 一种调度中心服务器,其特征在于,所述调度中心服务器包括存储器和处理器,所述存储器用于存储计算机程序,所述计算机程序被所述处理器执行时,实现如权利要求1至11中任一所述的方法。
PCT/CN2018/120101 2018-11-12 2018-12-10 一种请求任务的调度方法及调度中心服务器 WO2020098030A1 (zh)

Priority Applications (2)

Application Number Priority Date Filing Date Title
EP18940424.7A EP3879786A1 (en) 2018-11-12 2018-12-10 Scheduling method for request task and scheduling center server
US16/966,900 US20210049424A1 (en) 2018-11-12 2018-12-10 Scheduling method of request task and scheduling center server

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
CN201811340787 2018-11-12
CN201811340787.0 2018-11-12
CN201811393124.5 2018-11-21
CN201811393124.5A CN109547546B (zh) 2018-11-12 2018-11-21 一种请求任务的调度方法及调度中心服务器

Publications (1)

Publication Number Publication Date
WO2020098030A1 true WO2020098030A1 (zh) 2020-05-22

Family

ID=65848932

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/120101 WO2020098030A1 (zh) 2018-11-12 2018-12-10 一种请求任务的调度方法及调度中心服务器

Country Status (4)

Country Link
US (1) US20210049424A1 (zh)
EP (1) EP3879786A1 (zh)
CN (1) CN109547546B (zh)
WO (1) WO2020098030A1 (zh)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111913791A (zh) * 2020-07-02 2020-11-10 北京和瑞精准医学检验实验室有限公司 任务调度方法、装置、设备和计算机可读存储介质
CN112101609A (zh) * 2020-07-24 2020-12-18 西安电子科技大学 关于用户还款及时性的预测系统、方法、装置及电子设备
CN114726749A (zh) * 2022-03-02 2022-07-08 阿里巴巴(中国)有限公司 数据异常检测模型获取方法、装置、设备、介质及产品

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109547546B (zh) * 2018-11-12 2020-06-05 网宿科技股份有限公司 一种请求任务的调度方法及调度中心服务器
WO2020227968A1 (en) * 2019-05-15 2020-11-19 Beijing Didi Infinity Technology And Development Co., Ltd. Adversarial multi-binary neural network for multi-class classification
CN110391955B (zh) * 2019-07-22 2022-04-12 平安科技(深圳)有限公司 网络数据预处理方法、装置、设备及可读存储介质
CN111144784A (zh) * 2019-12-31 2020-05-12 中国电子科技集团公司信息科学研究院 面向有人/无人协同编队系统的任务分配方法及系统
CN114064262A (zh) * 2020-08-07 2022-02-18 伊姆西Ip控股有限责任公司 管理存储系统中的计算资源的方法、设备和程序产品
KR20220036494A (ko) * 2020-09-16 2022-03-23 삼성전자주식회사 딥러닝 워크로드를 위한 하이브리드 스케줄링 방법과 이를 수행하는 컴퓨팅 장치
CN113268322B (zh) * 2021-05-17 2023-11-07 深圳番多拉信息科技有限公司 一种拥有资源能力计算的方法、系统、装置以及存储介质
CN115102779B (zh) * 2022-07-13 2023-11-07 中国电信股份有限公司 预测模型的训练、访问请求的决策方法、装置、介质
CN118632372A (zh) * 2024-08-12 2024-09-10 北京中网华通设计咨询有限公司 一种无线通信网络中带宽分配方法及系统

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101442561A (zh) * 2008-12-12 2009-05-27 南京邮电大学 一种基于支持向量机的网格监控方法
CN106375452A (zh) * 2016-09-05 2017-02-01 河海大学 一种加权分类Web服务QoS监控方法
CN106650806A (zh) * 2016-12-16 2017-05-10 北京大学深圳研究生院 一种用于行人检测的协同式深度网络模型方法
WO2018200112A1 (en) * 2017-04-26 2018-11-01 Elasticsearch B.V. Clustering and outlier detection in anomaly and causation detection for computing environments
CN109547546A (zh) * 2018-11-12 2019-03-29 网宿科技股份有限公司 一种请求任务的调度方法及调度中心服务器

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9258217B2 (en) * 2008-12-16 2016-02-09 At&T Intellectual Property I, L.P. Systems and methods for rule-based anomaly detection on IP network flow
CN102163239B (zh) * 2011-05-11 2014-04-23 中科院成都信息技术股份有限公司 一种基于浮动分类阈值的分类器集成方法
CN104317658B (zh) * 2014-10-17 2018-06-12 华中科技大学 一种基于MapReduce的负载自适应任务调度方法
CN107948004B (zh) * 2017-12-29 2021-06-22 北京奇艺世纪科技有限公司 一种视频cdn调取优化方法及装置

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101442561A (zh) * 2008-12-12 2009-05-27 南京邮电大学 一种基于支持向量机的网格监控方法
CN106375452A (zh) * 2016-09-05 2017-02-01 河海大学 一种加权分类Web服务QoS监控方法
CN106650806A (zh) * 2016-12-16 2017-05-10 北京大学深圳研究生院 一种用于行人检测的协同式深度网络模型方法
WO2018200112A1 (en) * 2017-04-26 2018-11-01 Elasticsearch B.V. Clustering and outlier detection in anomaly and causation detection for computing environments
CN109547546A (zh) * 2018-11-12 2019-03-29 网宿科技股份有限公司 一种请求任务的调度方法及调度中心服务器

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111913791A (zh) * 2020-07-02 2020-11-10 北京和瑞精准医学检验实验室有限公司 任务调度方法、装置、设备和计算机可读存储介质
CN111913791B (zh) * 2020-07-02 2023-10-13 北京和瑞精湛医学检验实验室有限公司 任务调度方法、装置、设备和计算机可读存储介质
CN112101609A (zh) * 2020-07-24 2020-12-18 西安电子科技大学 关于用户还款及时性的预测系统、方法、装置及电子设备
CN112101609B (zh) * 2020-07-24 2023-08-01 西安电子科技大学 关于用户还款及时性的预测系统、方法、装置及电子设备
CN114726749A (zh) * 2022-03-02 2022-07-08 阿里巴巴(中国)有限公司 数据异常检测模型获取方法、装置、设备、介质及产品
CN114726749B (zh) * 2022-03-02 2023-10-31 阿里巴巴(中国)有限公司 数据异常检测模型获取方法、装置、设备及介质

Also Published As

Publication number Publication date
CN109547546A (zh) 2019-03-29
EP3879786A1 (en) 2021-09-15
CN109547546B (zh) 2020-06-05
US20210049424A1 (en) 2021-02-18

Similar Documents

Publication Publication Date Title
WO2020098030A1 (zh) 一种请求任务的调度方法及调度中心服务器
Heydari et al. Softadapt: Techniques for adaptive loss weighting of neural networks with multi-part loss functions
Somasundaram et al. CLOUDRB: A framework for scheduling and managing High-Performance Computing (HPC) applications in science cloud
CN106446959B (zh) 一种云计算资源动态匹配方法及装置
CN107203789B (zh) 分配模型建立方法、分配方法以及相关装置
Junaid et al. Modeling an optimized approach for load balancing in cloud
Junaid et al. A hybrid model for load balancing in cloud using file type formatting
Jatoth et al. Optimal fitness aware cloud service composition using an adaptive genotypes evolution based genetic algorithm
CN110390345A (zh) 一种基于云平台的大数据集群自适应资源调度方法
CN110796513B (zh) 多任务学习方法、装置、电子设备及存储介质
US20220383036A1 (en) Clustering data using neural networks based on normalized cuts
Muthusamy et al. Cluster-based task scheduling using K-means clustering for load balancing in cloud datacenters
CN108833302A (zh) 云环境下基于模糊聚类及严格双边匹配的资源分配方法
Jacob et al. Resource scheduling in cloud using bacterial foraging optimization algorithm
Devadas et al. Interdependency aware QUBIT and BROWNBOOST rank for large scale requirement prioritization
Zhou et al. Load balancing prediction method of cloud storage based on analytic hierarchy process and hybrid hierarchical genetic algorithm
KR101580202B1 (ko) 대규모 데이터를 위한 서포트 벡터 머신 이진 트리 구조 갱신 장치 및 방법
Akay et al. Predicting the performance measures of a 2-dimensional message passing multiprocessor architecture by using machine learning methods
CN110866605B (zh) 数据模型训练方法、装置、电子设备及可读介质
CN116976461A (zh) 联邦学习方法、装置、设备及介质
Thilak et al. Meta-heuristic Algorithms to Optimize Two-Stage Task Scheduling in the Cloud
Zhang et al. Learning driven parallelization for large-scale video workload in hybrid CPU-GPU cluster
JP6047190B2 (ja) リレーショナルモデルを決定するプログラムと装置
Wang et al. DualMatch: Robust Semi-supervised Learning with Dual-Level Interaction
Mohammed et al. Location-aware deep learning-based framework for optimizing cloud consumer quality of service-based service composition

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18940424

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 2018940424

Country of ref document: EP

Effective date: 20210607