WO2022100370A1 - 一种基于svm的流处理框架的自动调优方法 - Google Patents
一种基于svm的流处理框架的自动调优方法 Download PDFInfo
- Publication number
- WO2022100370A1 WO2022100370A1 PCT/CN2021/124402 CN2021124402W WO2022100370A1 WO 2022100370 A1 WO2022100370 A1 WO 2022100370A1 CN 2021124402 W CN2021124402 W CN 2021124402W WO 2022100370 A1 WO2022100370 A1 WO 2022100370A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- configuration parameters
- stream processing
- performance
- individual
- fitness
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 45
- 238000005457 optimization Methods 0.000 title abstract description 26
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 31
- 238000012549 training Methods 0.000 claims abstract description 24
- 230000002068 genetic effect Effects 0.000 claims abstract description 16
- 108090000623 proteins and genes Proteins 0.000 claims abstract description 4
- 238000012545 processing Methods 0.000 claims description 81
- 238000003860 storage Methods 0.000 claims description 16
- 238000004590 computer program Methods 0.000 claims description 7
- 230000035772 mutation Effects 0.000 claims description 6
- 238000010586 diagram Methods 0.000 description 15
- 230000000694 effects Effects 0.000 description 10
- 230000006870 function Effects 0.000 description 7
- 238000010801 machine learning Methods 0.000 description 5
- 238000013528 artificial neural network Methods 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 4
- 238000007637 random forest analysis Methods 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 238000010845 search algorithm Methods 0.000 description 3
- 238000003491 array Methods 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 238000013480 data collection Methods 0.000 description 2
- 239000000835 fiber Substances 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000001902 propagating effect Effects 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 238000012706 support-vector machine Methods 0.000 description 2
- 238000012795 verification Methods 0.000 description 2
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 229910052802 copper Inorganic materials 0.000 description 1
- 239000010949 copper Substances 0.000 description 1
- 238000013136 deep learning model Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 230000014509 gene expression Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000012821 model calculation Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000001052 transient effect Effects 0.000 description 1
- 239000002699 waste material Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F8/00—Arrangements for software engineering
- G06F8/20—Software design
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
- G06N20/10—Machine learning using kernel methods, e.g. support vector machines [SVM]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/12—Computing arrangements based on biological models using genetic models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/12—Computing arrangements based on biological models using genetic models
- G06N3/126—Evolutionary algorithms, e.g. genetic algorithms or genetic programming
Definitions
- the present invention relates to the technical field of big data processing, and more particularly, to an automatic tuning method of a stream processing framework based on SVM.
- Structured Streaming provides technologies such as incremental query, more advanced programming interface, unified batch streaming combined with programming model, etc., which realizes lower processing delay, simple and more advanced business implementation logic, and realizes high efficiency by combining with Spark SQL Stream processing and other high-quality features make Structured Streaming used by more and more enterprises as the first choice for real-time computing.
- Structured Streaming will be affected by configuration parameters during operation, and unreasonable configuration will seriously slow down task execution. Spark officially recommends a set of default configuration parameters. However, in actual stream processing tasks, the default configuration parameters cannot adapt to the real-time changes of the task scenario, and cannot be adapted according to different system resources, resulting in the performance of Structured Streaming being limited and a large number of waste of system resources. A large number of configuration parameters in Structured Streaming need to be set reasonably according to different application scenarios, and the existing manual parameter adjustment is difficult and expensive.
- the existing Structured Streaming automatic configuration parameter tuning method is not deep enough. It only considers the characteristics of Structured Streaming's batch-stream mixing, and automatically optimizes the parameters related to batch processing and stream processing.
- the underlying computing engine of Structured Streaming is Spark, which also needs to be optimized and adjusted. Therefore, it is far from enough to optimize parameters related to batch/stream processing.
- the machine learning algorithms used in existing methods cannot effectively search for the most optimal configuration parameters.
- the purpose of the present invention is to overcome the above-mentioned defects of the prior art, and to provide an automatic tuning method of a stream processing framework, which is based on SVM (support vector machine algorithm) and genetic algorithm to search for optimal configuration parameters suitable for specific application scenarios. new technology solutions.
- SVM support vector machine algorithm
- the invention provides an automatic tuning method of a stream processing framework based on SVM.
- the method includes the following steps:
- each sample data contains a set of configuration parameters and the corresponding relationship between the execution performance of the stream processing framework
- a set of configuration parameters is regarded as an individual, each parameter in a set of configuration parameters is regarded as a gene in the individual, and the output performance of the performance prediction model corresponding to a set of configuration parameters is used to measure the individual Fitness, using a genetic algorithm to search for the optimal configuration parameters of the stream processing framework.
- the present invention has the advantage that an automatic optimization method for the stream processing framework is designed based on SVM and a meta-heuristic algorithm (genetic algorithm), which realizes automatic optimization from the bottom layer to the upper layer, and combines with better machines. Learning algorithms can more efficiently search for better configuration parameters.
- FIG. 1 is a flowchart of an automatic tuning method for an SVM-based stream processing framework according to an embodiment of the present invention
- FIG. 2 is a schematic process diagram of an automatic tuning method of an SVM-based stream processing framework according to an embodiment of the present invention
- Fig. 3 is the effect comparison diagram of three kinds of algorithms of the prior art and the method of the present invention.
- FIG. 6 is an effect diagram of the ratio of data processing delay to data processing throughput according to an embodiment of the present invention.
- the automatic tuning method of the SVM-based stream processing framework includes the following steps.
- step S110 a training data set is constructed, and each training sample is used to represent the corresponding relationship between the execution performance of the stream processing framework and the used combination of configuration parameters.
- a parameter generator (Conf Generator) is included, which first selects parameters that significantly affect the performance of Structured Streaming and the underlying Spark, and then automatically generates and assigns parameters for the operation of the program to be optimized according to the selected parameters.
- the data processing latency and throughput of Structured Streaming at runtime combined with the parameters used are collected as a sample data in the training data set. In this way, a training dataset consisting of multiple sample data is obtained after multiple runs. Each training sample is used to characterize the correspondence between data processing latency and throughput and the combination of parameters used.
- the training data set is represented as ⁇ Pv 1 , Pv 2 , ..., Pv n ⁇
- the first training sample Pv 1 contains ⁇ t 1 , I i , conf i1 , ..., conf i23 ⁇
- t 1 represents the data Processing delay
- I 1 represents data throughput
- conf i1 , ..., conf i23 is a combination of configuration parameters.
- Each group of configuration parameters includes the upper-level parameters of the stream processing framework and the underlying parameters of Spark.
- Step S120 using the training data set, taking the configuration parameter combination as the input, and taking the stream processing framework execution performance as the output, train the SVM model to obtain the performance prediction model of the stream processing framework.
- This step is the modeling stage of the performance prediction model.
- the obtained training data set is used for modeling based on machine learning algorithms.
- the purpose is to build a performance prediction model that can reflect the impact of different configuration parameters on the delay and throughput of Structured Streaming.
- each set of configuration parameters is used as input, and the execution performance of the corresponding stream processing framework is used as output to train an SVM model to obtain a performance prediction model, which is used for delay and throughput prediction for different configuration parameters.
- the SVM model can quickly and accurately predict the execution performance of the stream processing framework.
- Step S130 in the search space of configuration parameters, use the performance prediction model combined with the genetic algorithm to perform an optimal configuration search.
- the genetic algorithm is used to perform an iterative search based on the performance prediction model, and the optimal configuration parameters are finally screened.
- the present invention preferably uses genetic algorithm as the basic search algorithm, and optimizes the genetic algorithm and structured streaming.
- Genetic algorithm As the basic search algorithm, and optimizes the genetic algorithm and structured streaming.
- the search process for optimal configuration parameters includes:
- Step S131 randomly input a group of structured streaming parameters and obtain the initialized individual fitness A standard through performance model calculation;
- Step S132 randomly select n groups of configuration parameters (for example, n is greater than 1/5 of the number of training sets) from the training data set obtained in the data collection stage as the initialization population P, and perform random crossover operation and mutation rate for each individual in P is a mutation operation of 0.02;
- Step S133 Use the performance prediction model to calculate the fitness of the population P and its descendants, and screen out individuals whose fitness is higher than A to form a new population P', and use the fitness A' of the individual with the highest fitness as the new fitness standard A';
- Step S134 repeating S132 and S133 until no better individual can be generated, then the current optimal individual is the optimal configuration parameter searched.
- the genetic algorithm is used to search for the optimal configuration based on the performance prediction model obtained in the modeling stage.
- the purpose is to use the crossover mutation characteristics of the genetic algorithm to avoid the search from falling into local optimum, while ensuring excellent search performance.
- the performance prediction model is used to predict the performance of different configuration parameters generated by the genetic algorithm in Structured Streaming, so as to achieve efficient search, and the search can finally obtain the optimal configuration parameters and directly use them in Structured Streaming.
- a recursive random search algorithm, a pattern search algorithm, etc. are used to search for the optimal configuration, it is easy to fall into the local optimum, which leads to the problem that the global optimal configuration cannot be found.
- Structured Streaming test program officially provided by Spark, including StructuredKafkaWordCount (referred to as KafkaWC), StructuredNetworkWordCount (referred to as NetworkWC), StructuredNetworkWordCountWindows (referred to as NetworkWCW) and StructuredSessionization (Sessionization for short), which automatically optimizes the configuration parameters of the Structured Streaming framework.
- KafkaWC StructuredKafkaWordCount
- NetworkWC StructuredNetworkWordCount
- NetworkWCW StructuredNetworkWordCountWindows
- StructuredSessionization StructuredSessionization for short
- Fig. 3 is three kinds of algorithms (RS, ANN, RF) commonly used at present and the method used in the present invention (marked as support vector machine algorithm in the figure), for the modeling effect comparison of four different Structured Streaming programs selected. It can be clearly seen from FIG. 3 that the modeling accuracy of the method of the present invention (the far right side of each item indicates the present invention) is higher than that of the other three algorithms under different programs. Specifically, the modeling accuracy of the method of the present invention is on average 8% higher than that of the RS algorithm, 9.7% higher than that of the ANN algorithm, and 6.7% higher than that of the RF algorithm.
- Fig. 4 is the optimization effect of the present invention on the operating throughput of Structured Streaming. Since the optimization method of the present invention automatically configures reasonable parameters for different programs, compared with the official default configuration (right side), the optimization method of the present invention is significantly improved.
- the data processing throughput of Structured Streaming under different programs is increased by an average of 2.29 times and a maximum increase of 2.52 times.
- Fig. 5 is the optimization effect of the present invention on reducing the runtime delay of Structured Streaming. Since the optimization method of the present invention automatically configures reasonable parameters for different programs, compared with the official default configuration (right side), the optimization method of the present invention is significantly Reduced the data processing delay of Structured Streaming under different programs, with an average reduction of 3.08 times and a maximum reduction of 3.96 times.
- Figure 6 shows the optimization effect of the present invention on the ratio of data processing delay and data processing throughput.
- Lower data processing delay and higher data processing throughput are the goals of stream processing system performance optimization. Low means that while achieving lower data processing latency, greater data throughput is achieved, which is a more comprehensive optimization evaluation criterion.
- the optimization method of the present invention significantly reduces the ratio of delay to throughput by an average of 5.95 times and a maximum of 8.36 times
- the experimental results show that the optimization method of the present invention realizes the automatic parameter adjustment and optimization of Structured Streaming, and the optimization performance is better than that of the prior art, and the data processing delay is significantly reduced compared with the official default configuration under different current program loads. 2.52 times, while improving the data processing throughput by up to 3.96 times.
- the existing automatic tuning method of Structured Streaming configuration parameters does not consider the optimization of Spark, the underlying computing engine of Structured Streaming.
- the invention realizes the overall optimization from the bottom layer Spark to the upper layer Structured Streaming, the optimization is more in-depth and the effect is better.
- the existing machine learning algorithms have poor performance and do not fit the optimization characteristics of Structured Streaming.
- the invention combines the SVM with the genetic algorithm, designs a technical scheme more in line with Structured Streaming optimization, and realizes high-performance automatic parameter tuning optimization.
- the present invention may be a system, method and/or computer program product.
- the computer program product may include a computer-readable storage medium having computer-readable program instructions loaded thereon for causing a processor to implement various aspects of the present invention.
- a computer-readable storage medium may be a tangible device that can hold and store instructions for use by the instruction execution device.
- the computer-readable storage medium may be, for example, but not limited to, an electrical storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
- Non-exhaustive list of computer readable storage media include: portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM) or flash memory), static random access memory (SRAM), portable compact disk read only memory (CD-ROM), digital versatile disk (DVD), memory sticks, floppy disks, mechanically coded devices, such as printers with instructions stored thereon Hole cards or raised structures in grooves, and any suitable combination of the above.
- RAM random access memory
- ROM read only memory
- EPROM erasable programmable read only memory
- flash memory static random access memory
- SRAM static random access memory
- CD-ROM compact disk read only memory
- DVD digital versatile disk
- memory sticks floppy disks
- mechanically coded devices such as printers with instructions stored thereon Hole cards or raised structures in grooves, and any suitable combination of the above.
- Computer-readable storage media are not to be construed as transient signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through waveguides or other transmission media (eg, light pulses through fiber optic cables), or through electrical wires transmitted electrical signals.
- the computer readable program instructions described herein may be downloaded to various computing/processing devices from a computer readable storage medium, or to an external computer or external storage device over a network such as the Internet, a local area network, a wide area network, and/or a wireless network.
- the network may include copper transmission cables, fiber optic transmission, wireless transmission, routers, firewalls, switches, gateway computers, and/or edge servers.
- a network adapter card or network interface in each computing/processing device receives computer-readable program instructions from a network and forwards the computer-readable program instructions for storage in a computer-readable storage medium in each computing/processing device .
- the computer program instructions for carrying out the operations of the present invention may be assembly instructions, instruction set architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state setting data, or instructions in one or more programming languages.
- Source or object code written in any combination, including object-oriented programming languages, such as Smalltalk, C++, etc., and conventional procedural programming languages, such as the "C" language or similar programming languages.
- the computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer, or entirely on the remote computer or server implement.
- the remote computer may be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or may be connected to an external computer (eg, using an Internet service provider through the Internet connect).
- LAN local area network
- WAN wide area network
- custom electronic circuits such as programmable logic circuits, field programmable gate arrays (FPGAs), or programmable logic arrays (PLAs)
- FPGAs field programmable gate arrays
- PDAs programmable logic arrays
- Computer readable program instructions are executed to implement various aspects of the present invention.
- These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer or other programmable data processing apparatus to produce a machine that causes the instructions when executed by the processor of the computer or other programmable data processing apparatus , resulting in means for implementing the functions/acts specified in one or more blocks of the flowchart and/or block diagrams.
- These computer readable program instructions can also be stored in a computer readable storage medium, these instructions cause a computer, programmable data processing apparatus and/or other equipment to operate in a specific manner, so that the computer readable medium on which the instructions are stored includes An article of manufacture comprising instructions for implementing various aspects of the functions/acts specified in one or more blocks of the flowchart and/or block diagrams.
- Computer readable program instructions can also be loaded onto a computer, other programmable data processing apparatus, or other equipment to cause a series of operational steps to be performed on the computer, other programmable data processing apparatus, or other equipment to produce a computer-implemented process , thereby causing instructions executing on a computer, other programmable data processing apparatus, or other device to implement the functions/acts specified in one or more blocks of the flowcharts and/or block diagrams.
- each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more functions for implementing the specified logical function(s) executable instructions.
- the functions noted in the blocks may occur out of the order noted in the figures. For example, two blocks in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
- each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations can be implemented in dedicated hardware-based systems that perform the specified functions or actions , or can be implemented in a combination of dedicated hardware and computer instructions. It is well known to those skilled in the art that implementation in hardware, implementation in software, and implementation in a combination of software and hardware are all equivalent.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Software Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Biophysics (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Mathematical Physics (AREA)
- Computing Systems (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Medical Informatics (AREA)
- Genetics & Genomics (AREA)
- Biomedical Technology (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Physiology (AREA)
- Complex Calculations (AREA)
- Stored Programmes (AREA)
Abstract
Description
Claims (8)
- 一种基于SVM的流处理框架的自动调优方法,包括以下步骤:构建训练数据集,其中,每条样本数据包含一组配置参数与流处理框架执行性能之间的对应关系;基于所述训练数据集,以各组配置参数作为输入,以对应的流处理框架执行性能作为输出,训练SVM模型,获得性能预测模型;在配置参数的搜索空间中,将一组配置参数作为一个个体,将一组配置参数中的各个参数作为个体中的基因,并利用一组配置参数所对应的性能预测模型的输出性能来衡量个体适应度,利用遗传算法搜索流处理框架的最优配置参数。
- 根据权利要求1所述的方法,其中,所述利用遗传算法搜索流处理框架的最优配置参数包括:随机输入一组配置参数并通过所述性能预测模型计算得到初始化的个体适应度标准;从所述训练数据集中随机选择n组配置参数作为初始化种群P,对P中每个个体的进行随机的交叉运算以及设定变异率的变异运算;利用所述性能模型对种群P及其后代进行适应度计算,并筛选出适应度高于所述初始化的个体适应度标准的个体组成新种群P’,将适应度最高的个体的适应度作为新的适应度标准,通过迭代运算找出适应度最高的个体,该个体对应流处理框架的最优配置参数。
- 根据权利要求2所述的方法,其中,对于所选择的n组配置参数,n大于所述训练数据集中训练样本数量的1/5,所述变异率设置为0.02。
- 根据权利要求1所述的方法,其中,所述流处理框架的执行性能包括数据处理延迟和数据吞吐量,所述个体适应度是数据处理延迟与吞吐量的比值。
- 根据权利要求1所述的方法,其中,所述流处理框架包括结构化流处理框架、Flink框架或Storm框架。
- 根据权利要求1所述的方法,其中,所述流处理框架是结构化流处理框架,所述构建训练数据集包括:根据对执行性能的影响程度选择出显著影响上层结构化流处理框架与底层Spark性能的参数;根据所选择的参数为待优化程序的运行自动生成并分配参数;在每次程序运行结束后,收集运行时所述结构化流处理框架的数据处理延迟和吞吐量与所使用的参数组合作为所述训练数据集中的一条样本数据。
- 一种计算机可读存储介质,其上存储有计算机程序,其中,该程序被处理器执行时实现根据权利要求1至6中任一项所述方法的步骤。
- 一种计算机设备,包括存储器和处理器,在所述存储器上存储有能够在处理器上运行的计算机程序,其特征在于,所述处理器执行所述程序时实现权利要求1至6中任一项所述的方法的步骤。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011261446.1 | 2020-11-12 | ||
CN202011261446.1A CN114489574B (zh) | 2020-11-12 | 2020-11-12 | 一种基于svm的流处理框架的自动调优方法 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022100370A1 true WO2022100370A1 (zh) | 2022-05-19 |
Family
ID=81490256
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/124402 WO2022100370A1 (zh) | 2020-11-12 | 2021-10-18 | 一种基于svm的流处理框架的自动调优方法 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN114489574B (zh) |
WO (1) | WO2022100370A1 (zh) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117252114A (zh) * | 2023-11-17 | 2023-12-19 | 湖南华菱线缆股份有限公司 | 一种基于遗传算法的电缆耐扭转实验方法 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130198119A1 (en) * | 2012-01-09 | 2013-08-01 | DecisionQ Corporation | Application of machine learned bayesian networks to detection of anomalies in complex systems |
CN106648654A (zh) * | 2016-12-20 | 2017-05-10 | 深圳先进技术研究院 | 一种数据感知的Spark配置参数自动优化方法 |
CN110086731A (zh) * | 2019-04-25 | 2019-08-02 | 北京计算机技术及应用研究所 | 一种云架构下网络数据稳定采集方法 |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108121999A (zh) * | 2017-12-10 | 2018-06-05 | 北京工业大学 | 基于混合蝙蝠算法的支持向量机参数选择方法 |
US11544621B2 (en) * | 2019-03-26 | 2023-01-03 | International Business Machines Corporation | Cognitive model tuning with rich deep learning knowledge |
CN111612528A (zh) * | 2020-04-30 | 2020-09-01 | 中国移动通信集团江苏有限公司 | 用户分类模型的确定方法、装置、设备及存储介质 |
-
2020
- 2020-11-12 CN CN202011261446.1A patent/CN114489574B/zh active Active
-
2021
- 2021-10-18 WO PCT/CN2021/124402 patent/WO2022100370A1/zh active Application Filing
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130198119A1 (en) * | 2012-01-09 | 2013-08-01 | DecisionQ Corporation | Application of machine learned bayesian networks to detection of anomalies in complex systems |
CN106648654A (zh) * | 2016-12-20 | 2017-05-10 | 深圳先进技术研究院 | 一种数据感知的Spark配置参数自动优化方法 |
CN110086731A (zh) * | 2019-04-25 | 2019-08-02 | 北京计算机技术及应用研究所 | 一种云架构下网络数据稳定采集方法 |
Non-Patent Citations (1)
Title |
---|
WANG DONG: "Parallel Analysis on Power Grid Equipment Monitoring Big Data Based on Storm Framework", MASTER THESIS, TIANJIN POLYTECHNIC UNIVERSITY, CN, no. 1, 15 January 2020 (2020-01-15), CN , XP055930023, ISSN: 1674-0246 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117252114A (zh) * | 2023-11-17 | 2023-12-19 | 湖南华菱线缆股份有限公司 | 一种基于遗传算法的电缆耐扭转实验方法 |
CN117252114B (zh) * | 2023-11-17 | 2024-02-13 | 湖南华菱线缆股份有限公司 | 一种基于遗传算法的电缆耐扭转实验方法 |
Also Published As
Publication number | Publication date |
---|---|
CN114489574A (zh) | 2022-05-13 |
CN114489574B (zh) | 2022-10-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7170779B2 (ja) | 自動的な意図のマイニング、分類、及び配置のための方法及びシステム | |
Nguyen et al. | Pay-as-you-go reconciliation in schema matching networks | |
WO2022111125A1 (zh) | 一种基于随机森林的图数据处理框架自动调优方法 | |
JP7392668B2 (ja) | データ処理方法および電子機器 | |
US11595415B2 (en) | Root cause analysis in multivariate unsupervised anomaly detection | |
CN105868334B (zh) | 一种基于特征递增型的电影个性化推荐方法及系统 | |
US20160110657A1 (en) | Configurable Machine Learning Method Selection and Parameter Optimization System and Method | |
WO2023124029A1 (zh) | 深度学习模型的训练方法、内容推荐方法和装置 | |
US10373071B2 (en) | Automated intelligent data navigation and prediction tool | |
US20200089832A1 (en) | Application- or algorithm-specific quantum circuit design | |
US9852390B2 (en) | Methods and systems for intelligent evolutionary optimization of workflows using big data infrastructure | |
US11429623B2 (en) | System for rapid interactive exploration of big data | |
US10762166B1 (en) | Adaptive accelerated yield analysis | |
US9582586B2 (en) | Massive rule-based classification engine | |
US20230342359A1 (en) | System and method for machine learning for system deployments without performance regressions | |
CN116057518A (zh) | 使用机器学习模型的自动查询谓词选择性预测 | |
Nagesh et al. | High performance computation of big data: performance optimization approach towards a parallel frequent item set mining algorithm for transaction data based on hadoop MapReduce framework | |
WO2022011553A1 (en) | Feature interaction via edge search | |
WO2022100370A1 (zh) | 一种基于svm的流处理框架的自动调优方法 | |
WO2023174189A1 (zh) | 图网络模型节点分类方法、装置、设备及存储介质 | |
CN107679107A (zh) | 一种基于图数据库的电网设备可达性查询方法及系统 | |
US20230186074A1 (en) | Fabricating data using constraints translated from trained machine learning models | |
US9928327B2 (en) | Efficient deployment of table lookup (TLU) in an enterprise-level scalable circuit simulation architecture | |
Oo et al. | Hyperparameters optimization in scalable random forest for big data analytics | |
JP2022189805A (ja) | コンピュータ実装方法、情報処理システム、コンピュータプログラム(it環境向けのアノマリ検出ドメインにおけるパフォーマンスモニタリング) |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21890897 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21890897 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21890897 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 111223) |