CN115062402A - A data-driven method for extracting train level acceleration - Google Patents
A data-driven method for extracting train level acceleration Download PDFInfo
- Publication number
- CN115062402A CN115062402A CN202210536133.5A CN202210536133A CN115062402A CN 115062402 A CN115062402 A CN 115062402A CN 202210536133 A CN202210536133 A CN 202210536133A CN 115062402 A CN115062402 A CN 115062402A
- Authority
- CN
- China
- Prior art keywords
- learning
- sampling
- train
- acceleration
- data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 50
- 230000001133 acceleration Effects 0.000 title claims abstract description 39
- 238000005070 sampling Methods 0.000 claims abstract description 33
- 238000012549 training Methods 0.000 claims abstract description 31
- 238000000605 extraction Methods 0.000 claims abstract description 27
- 230000000694 effects Effects 0.000 claims abstract description 17
- 230000006870 function Effects 0.000 claims abstract description 13
- 238000005457 optimization Methods 0.000 claims abstract description 11
- 230000008569 process Effects 0.000 claims abstract description 9
- 230000008859 change Effects 0.000 claims abstract description 7
- 238000013507 mapping Methods 0.000 claims abstract description 7
- 238000005065 mining Methods 0.000 claims abstract 2
- 238000012937 correction Methods 0.000 claims description 3
- 238000005265 energy consumption Methods 0.000 abstract description 5
- 238000004422 calculation algorithm Methods 0.000 description 12
- 238000002474 experimental method Methods 0.000 description 11
- 238000004088 simulation Methods 0.000 description 10
- 238000012360 testing method Methods 0.000 description 10
- 238000004364 calculation method Methods 0.000 description 4
- 238000011156 evaluation Methods 0.000 description 4
- 238000001404 Nemenyi test Methods 0.000 description 3
- 230000009286 beneficial effect Effects 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 238000007477 logistic regression Methods 0.000 description 2
- 230000003044 adaptive effect Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000013145 classification model Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 238000009795 derivation Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 239000006185 dispersion Substances 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000007786 learning performance Effects 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- PLAIAIKZKCZEQF-UHFFFAOYSA-N methyl 6-chloro-2-oxo-3h-1,2$l^{4},3-benzodithiazole-4-carboxylate Chemical compound COC(=O)C1=CC(Cl)=CC2=C1NS(=O)S2 PLAIAIKZKCZEQF-UHFFFAOYSA-N 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000010149 post-hoc-test Methods 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 238000012216 screening Methods 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 238000011426 transformation method Methods 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F30/00—Computer-aided design [CAD]
- G06F30/10—Geometric CAD
- G06F30/15—Vehicle, aircraft or watercraft design
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F30/00—Computer-aided design [CAD]
- G06F30/20—Design optimisation, verification or simulation
- G06F30/27—Design optimisation, verification or simulation using machine learning, e.g. artificial intelligence, neural networks, support vector machines [SVM] or training a model
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2119/00—Details relating to the type or aim of the analysis or the optimisation
- G06F2119/10—Noise analysis or noise optimisation
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Geometry (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Computer Hardware Design (AREA)
- Medical Informatics (AREA)
- Software Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Artificial Intelligence (AREA)
- Automation & Control Theory (AREA)
- Aviation & Aerospace Engineering (AREA)
- Computational Mathematics (AREA)
- Mathematical Analysis (AREA)
- Mathematical Optimization (AREA)
- Pure & Applied Mathematics (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
Description
技术领域technical field
本发明涉及数据驱动的列车级位加速度提取方法,属于计算机与信息科学技术领域。The invention relates to a data-driven train level acceleration extraction method, and belongs to the technical field of computer and information science.
背景技术Background technique
发展轨道交通是解决现代大城市居民出行问题的有效途径,也是建设绿色城市、智能城市的重要方式。常见的轨道交通有传统铁路(国家铁路、城际铁路和市域铁路)、地铁、轻轨和有轨电车,新型轨道交通有磁悬浮轨道系统、单轨系统(跨座式轨道系统和悬挂式轨道系统)和旅客自动捷运系统等。随着火车和铁路技术的多元化发展,轨道交通呈现出越来越多的类型,不仅遍布于长距离的陆地运输,也广泛运用于中短距离的城市公共交通中,对列车进行性能抽取、监控其实时运行状态、保障其安全运营具有重要意义。The development of rail transit is an effective way to solve the travel problem of residents in modern big cities, and it is also an important way to build a green city and a smart city. Common rail transits include traditional railways (national railways, intercity railways and municipal railways), subways, light rails and trams, and new types of rail transits include maglev rail systems, monorail systems (straddle rail systems and suspended rail systems) and Passenger automatic rapid transit system, etc. With the diversified development of trains and railway technology, rail transit has shown more and more types, not only in long-distance land transportation, but also widely used in short- and medium-distance urban public transportation. It is of great significance to monitor its real-time operating status and ensure its safe operation.
列车运行数据主要包含车载系统采集的速度、推荐速度、级位,以及CBTC反馈的位置、坡度、弯道及目标点位置等信息,逐周期以格式化日志形式存储。运动状态估计问题中,将对相邻采样点的速度进行差值计算获取加速度,作为有监督学习标签挖掘当前速度、坡度及级位影响下的列车运动状态变化模式。The train operation data mainly includes the speed, recommended speed and level collected by the on-board system, as well as the position, slope, curve and target point position fed back by the CBTC, which are stored in formatted logs cycle by cycle. In the motion state estimation problem, the speed difference between adjacent sampling points is calculated to obtain the acceleration, which is used as a supervised learning label to mine the change pattern of the train motion state under the influence of the current speed, slope and level.
车辆运行时,受传感器误差、通信丢包、处理延时等多种因素影响,采样数据易出现波动、跳点、缺失等情况,导致计算加速度时产生错误,引入标签噪声。推荐速度是满足限速、时刻表等约束,面向列车运行效率、能耗、舒适度等指标优化生成的理想运行速度。现有推荐速度优化方法一般基于连续或离散假设,重点关注控制目标的能耗水平。针对列车常用的级位控速系统,离散假设可一定程度上考虑列车延时、加速度离散等问题,但现有方法未能对列车性能进行分析与抽取,优化过程仍与受控对象相独立,推荐加速度可能与列车加速度可行域不匹配,导致运行时级位切换频繁,产生预期之外的能耗。When the vehicle is running, affected by various factors such as sensor error, communication packet loss, processing delay, etc., the sampled data is prone to fluctuations, jumps, missing, etc., resulting in errors in the calculation of acceleration and the introduction of label noise. The recommended speed is the ideal running speed that satisfies constraints such as speed limit and timetable, and is optimized for train operation efficiency, energy consumption, comfort and other indicators. Existing recommendation speed optimization methods are generally based on continuous or discrete assumptions, focusing on the energy consumption level of the control target. For the stage speed control system commonly used by trains, the discrete assumption can consider the train delay, acceleration dispersion and other issues to a certain extent, but the existing methods fail to analyze and extract the train performance, and the optimization process is still independent of the controlled object. The recommended acceleration may not match the feasible region of the train acceleration, resulting in frequent level switching during operation, resulting in unexpected energy consumption.
发明内容SUMMARY OF THE INVENTION
本发明的目的是针对现有方法未考虑推荐加速度与列车级位加速度的匹配,使推荐速度的跟踪难度大,控车级位频繁切换并产生额外能耗问题。提出了数据驱动的列车级位加速度提取方法。The purpose of the present invention is that the existing method does not consider the matching of the recommended acceleration and the train level acceleration, so that the tracking of the recommended speed is difficult, the control level is frequently switched, and additional energy consumption is generated. A data-driven extraction method of train level acceleration is proposed.
本发明的设计原理为:首先利用噪声样本与正常样本在训练过程中表现出的不同损失变化模式,对样本可靠性、学习难度进行估计,优化样本学习顺序。然后将自采样学习分别引入AdaBoost和梯度提升学习提出SSAdaBoost和SSGB稳健学习方法,建立目标列车性能估计模型,挖掘含噪运行数据中的专车性能知识,拟合加速度与速度、级位序列(隐含延时特性)、坡度等特征间的映射函数。最后提出基于特征控制的关联关系提取方法,利用“查询样本”控制级位序列、延时、坡度等特征影响,依次查询不同速度区间下各级位对应加速度值,建立列车专属性能表,实现对目标特征与标签间量化关系抽取。The design principle of the present invention is as follows: firstly, the reliability and learning difficulty of the samples are estimated by using the different loss change patterns exhibited by the noise samples and the normal samples in the training process, and the learning sequence of the samples is optimized. Then, the self-sampling learning is introduced into AdaBoost and gradient boosting learning respectively, and SSAdaBoost and SSGB robust learning methods are proposed to establish the target train performance estimation model, mine the special car performance knowledge in the noisy running data, and fit the acceleration, speed, and level sequence (implicitly). The mapping function between features such as delay characteristics), slope and so on. Finally, a correlation extraction method based on feature control is proposed. The "query sample" is used to control the influence of features such as stage sequence, delay, and slope, and the corresponding acceleration values of the lower stages in different speed ranges are inquired in turn to establish a train-specific performance table. Quantitative relationship extraction between target features and labels.
本发明的技术方案是通过如下步骤实现的:The technical scheme of the present invention is achieved through the following steps:
步骤1,面向含噪表格型数据,提出自采样学习方法SSL。
步骤1.1,利用含噪声样本与正常样本在提升学习模型训练过程中损失函数值的变化模式,包括绝对值、变化方向及变化速度等,对样本可靠性、学习难度进行估计。Step 1.1, using the variation pattern of the loss function value during the training process of the boosting learning model, including the absolute value, the variation direction and the variation speed, of the noise-containing samples and the normal samples, to estimate the reliability of the samples and the learning difficulty.
步骤1.2,调整每次迭代时的训练样本子集,在排除噪声点的同时优化样本学习顺序。Step 1.2, adjust the training sample subset at each iteration, and optimize the sample learning order while eliminating noise points.
步骤2,将自采样学习引入AdaBoost,建立SSAdaBoost和SSGB方法。
步骤2.1,建立目标列车性能估计模型,挖掘含噪运行数据中的专车性能知识,拟合加速度与速度、级位序列(隐含延时特性)、坡度等特征间的映射函数。Step 2.1, establish the target train performance estimation model, mine the special car performance knowledge in the noisy running data, and fit the mapping function between acceleration and speed, level sequence (implicit delay characteristic), gradient and other features.
步骤3,利用“查询样本”控制级位序列、延时、坡度等特征影响,依次查询不同速度区间下各级位对应加速度值。Step 3: Use the "query sample" to control the influence of features such as level sequence, delay, and gradient, and sequentially query the acceleration values corresponding to the lower levels in different speed ranges.
步骤3.1,建立列车专属性能表,实现对目标特征与标签间量化关系抽取。Step 3.1: Establish a train-specific performance table to extract the quantitative relationship between target features and labels.
有益效果beneficial effect
与原始AdaBoost相比,SSAdaBoost的样本权重调整由变化为viwi,有效限制了权重增长。通过对训练集进行筛选,学习器可在更稳定、低噪的样本上训练,相比SPLBoost,自采样学习可避免对非噪声困难样本的误判与抛弃,且无需样本学习顺序的先验知识。方法与原有AdaBoost相比,增加了两个采样控制超参数,计算成本(M×O(3kn+3n))与原方法(M×O(3kn+2n))接近。Compared with the original AdaBoost, the sample weight adjustment of SSAdaBoost is given by Changed to viwi, which effectively limits the weight growth. By screening the training set, the learner can be trained on more stable and low-noise samples. Compared with SPLBoost, self-sampling learning can avoid the misjudgment and abandonment of non-noise difficult samples, and does not require prior knowledge of the sample learning order. . Compared with the original AdaBoost, the method adds two sampling control hyperparameters, and the computational cost (M×O(3kn+3n)) is close to the original method (M×O(3kn+2n)).
随列车运行速度的增加,相同级位的实际加速度可能出现改变,基本规律为:1)牵引级位对列车产生的实际加速度随速度的上升而减小,尤其在高速阶段减小明显;2)惰行级位对列车产生的实际减速度随速度上升而增大,主要受与速度相关的阻力影响;3)制动级位对列车产生的实际减速度随速度上升而增大,变化幅度相对较小。同型号列车上线前,会对代表性车辆进行首车动调,在可控测试线路上,测试各速度区间下,级位实际加速度,绘制完整的性能表。With the increase of train running speed, the actual acceleration of the same level may change. The basic rules are: 1) The actual acceleration of the train generated by the traction level decreases with the increase of speed, especially in the high-speed stage; 2) The actual deceleration of the train caused by the coasting stage increases with the speed increase, which is mainly affected by the resistance related to the speed; 3) The actual deceleration of the train caused by the braking stage increases with the speed increase, and the variation range is relatively large. Small. Before the trains of the same model go online, the first car will be adjusted for representative vehicles. On the controllable test line, the actual acceleration of each speed range will be tested, and a complete performance table will be drawn.
附图说明Description of drawings
图1为本发明数据驱动的列车级位加速度提取方法原理图。FIG. 1 is a schematic diagram of a data-driven method for extracting train level acceleration according to the present invention.
图2为自采样提升学习UCI任务模型错误率统计。Figure 2 shows the error rate statistics of the self-sampling boosting learning UCI task model.
图3为自采样提升学习大规模真实任务实验结果图。Figure 3 shows the experimental results of self-sampling boosting learning on large-scale real tasks.
图4为知识抽取性能表仿真效果对比。Figure 4 shows the comparison of the simulation effects of the knowledge extraction performance table.
具体实施方式Detailed ways
为了更好的说明本发明的目的和优点,下面结合实例对本发明方法的实施方式做进一步详细说明。In order to better illustrate the purpose and advantages of the present invention, the embodiments of the method of the present invention will be described in further detail below with reference to examples.
实验主要验证SSAdaBoost和SSGB自采样提升学习算法的稳健性及建模性能和基于“查询样本”进行特征控制的模型解释方法在列车性能抽取任务中的应用效果。The experiments mainly verify the robustness and modeling performance of SSAdaBoost and SSGB self-sampling learning algorithms and the application effect of the model interpretation method based on "query samples" for feature control in the task of train performance extraction.
(1)采用70组不同噪声水平的UCI公开数据集及3组大规模真实任务数据集对所提SSAdaBoost和SSGB自采样提升学习性能进行评价,对比算法包括AdaBoost、LogitBoost、GentleBoost、RBoost、CB-AdaBoost、GradientBoost等6种高性能提升学习算法,涵盖现有先进稳健提升学习算法。所提算法的稳健性将直接影响模型对含噪运行数据中列车性能知识的识别效果,是实现准确性能抽取及性能自适应推荐速度优化的前提。(1) 70 sets of UCI public data sets with different noise levels and 3 sets of large-scale real task data sets are used to evaluate the learning performance of the proposed SSAdaBoost and SSGB self-sampling improvement. The comparison algorithms include AdaBoost, LogitBoost, GentleBoost, RBoost, CB- Six high-performance boosting learning algorithms, including AdaBoost and GradientBoost, cover the existing advanced robust boosting learning algorithms. The robustness of the proposed algorithm will directly affect the recognition effect of the model on the knowledge of train performance in noisy running data, which is the premise of accurate performance extraction and performance-adaptive recommended speed optimization.
实验中按照每5%一个区间进行划分,噪声水平设置范围从0%到30%,针对每个任务建立7个子数据集,故UCI数据建模实验的总数据集数为70个。噪声仅添加于训练集及验证集,各测试集均不添加噪声,确保测试结果的可靠性。将560组实验结果按任务划分,绘制不同算法随噪声加剧的建模错误率变化曲线以便于观察。In the experiment, each 5% interval is divided, the noise level is set from 0% to 30%, and 7 sub-data sets are established for each task, so the total number of data sets in the UCI data modeling experiment is 70. Noise is only added to the training set and validation set, and no noise is added to each test set to ensure the reliability of the test results. The 560 sets of experimental results are divided according to tasks, and the curve of the modeling error rate of different algorithms with increasing noise is drawn for easy observation.
实验以有监督二分类任务为场景,由于部分任务数据存在标签不平衡问题,评价指标采用错误率(Error Rate)、准确率(Accuracy)、F1分数(F1Score)以及用于检验算法间性能显著性的Nemenyi检验法(Nemenyi Test),计算方法如公式1~4所示:The experiment takes the supervised two-classification task as the scene. Due to the label imbalance problem in some task data, the evaluation indicators use Error Rate, Accuracy, F1 Score and are used to test the performance significance between algorithms. The Nemenyi Test (Nemenyi Test), the calculation method is shown in
矩阵中,TP(True Positive)代表将正例预测为正例的数量;FN(False Negative)代表将正例预测为负例的数量;FP(False Positive)代表将负例预测为正例的数量;TN(True Negative)代表将负例预测为负例的数量。根据上述符号定义,实验的评价指标计算方式如下:In the matrix, TP (True Positive) represents the number of positive examples predicted as positive examples; FN (False Negative) represents the number of positive examples predicted as negative examples; FP (False Positive) represents the number of negative examples predicted as positive examples ; TN (True Negative) represents the number of negative examples predicted as negative examples. According to the above symbol definitions, the calculation method of the evaluation index of the experiment is as follows:
Nemenyi检验法属于后续检验(Post-Hoc Test)方法,用于对算法性能差异的显著性进行两两检验,检验指标为错误率,具备显著性则说明两算法间具备明显性能差异,不具备则代表性能相近。其中,临界差异值(Critical Difference,CD)由算法数量K及单个算法的实验组数N决定:The Nemenyi test method belongs to the post-Hoc Test method, which is used to test the significance of the algorithm performance difference in pairs. represent similar performance. Among them, the critical difference (Critical Difference, CD) is determined by the number of algorithms K and the number of experimental groups N of a single algorithm:
其中,临界值qα基于学生化极差(Studentized Range)统计量除以获得,详细的参数设置方法参见实验过程。where the critical value q α is based on the Studentized Range statistic divided by obtained, and the detailed parameter setting method can be found in the experimental procedure.
实验的过程中设置弱学习器树深度为2,弱学习器数量为50,SSAdaBoost的超参数为衰退系数δ的调整范围为[0.9,1),采样比例μ的调整范围为[0,0.4]。During the experiment, the depth of the weak learner tree is set to 2, the number of weak learners is 50, the hyperparameter of SSAdaBoost is the decay coefficient δ, the adjustment range is [0.9, 1), and the adjustment range of the sampling ratio μ is [0, 0.4] .
(2)验证所提基于“查询样本”进行特征控制的模型解释方法在列车性能抽取任务中的应用效果。由于列车各速度区间下不同级位的实际加速度不易测量,无法获取真实性能表与抽取性能表进行对比,采用“转换法”(Conversion Method)进行实验。基于物理仿真模型,采用抽取性能表作为列车级位加速度参数,分析运行仿真效果,仿真结果与实际运行结果越相似则代表抽取性能与实际性能越相近。同时采用列车出厂性能表建立仿真模型,用于对比分析抽取性能表的准确性。实验数据为西安地铁机场及合肥地铁三号线列车运行数据。(2) Verify the application effect of the proposed model interpretation method based on "query samples" for feature control in the task of train performance extraction. Because the actual acceleration at different levels in each speed range of the train is not easy to measure, it is impossible to obtain the real performance table and compare the extracted performance table, and the "Conversion Method" is used for the experiment. Based on the physical simulation model, the extraction performance table is used as the train stage acceleration parameter to analyze the operation simulation effect. The more similar the simulation results are to the actual operation results, the closer the extraction performance is to the actual performance. At the same time, a simulation model is established by using the train ex-factory performance table, which is used to compare and analyze the accuracy of the extraction performance table. The experimental data is the train operation data of Xi'an Metro Airport and
实验采用“转换法”对基于抽取性能表的物理模型仿真精度进行分析,等效评价性能抽取效果。性能抽取效果受列车性能估计模型影响,需对SSGB所建加速度估计模型进行评价。模型输出为连续变量,采用具有较高灵敏度的回归任务典型指标均方误差(MeanSquared Error,MSE)进行评价:In the experiment, the "transformation method" is used to analyze the simulation accuracy of the physical model based on the extraction performance table, and the performance extraction effect is equivalently evaluated. The performance extraction effect is affected by the train performance estimation model, and the acceleration estimation model built by SSGB needs to be evaluated. The output of the model is a continuous variable, and the mean squared error (MSE), a typical indicator of regression tasks with high sensitivity, is used for evaluation:
式中,yi代表第i个测试样本的真实标签值,为模型输出的估计标签值,n为测试样本数量。MSE可对模型在测试集上的估计误差进行衡量,结果可代表平均绝对误差、均方根误差等相似指标的评价效果。In the formula, yi represents the true label value of the ith test sample, is the estimated label value output by the model, and n is the number of test samples. MSE can measure the estimation error of the model on the test set, and the result can represent the evaluation effect of similar indicators such as mean absolute error and root mean square error.
通过比较基于抽取性能所得仿真曲线与真实曲线的相似度,对性能抽取效果进行分析,采用曲线面积误差率AUCerr、终点速度误差verr、终点位置误差serr等指标:By comparing the similarity between the simulation curve obtained based on the extraction performance and the real curve, the performance extraction effect is analyzed, and the curve area error rate AUC err , the end point speed error verr , the end point position error s err and other indicators are used:
实验同时绘制曲线定性分析曲线相似度,比价对象包括真实速度-位置曲线、基于抽取性能表的仿真曲线、基于出厂性能表的仿真曲线。At the same time, the curve is drawn qualitatively to analyze the similarity of the curve, and the comparison objects include the real speed-position curve, the simulation curve based on the extraction performance table, and the simulation curve based on the factory performance table.
本次实验在硬件环境为MSI Prestige台式计算机,CPU型号为Intel Core i7-10700K八核十六线程处理器,CPU主频3.8GHz,物理内存32G,内存频率2400MHz,显卡为GeForce RTX 2080SUPER,配备8GB独立显存,操作系统是windows 10,64位。The hardware environment of this experiment is MSI Prestige desktop computer, the CPU model is Intel Core i7-10700K octa-core 16-thread processor, the CPU frequency is 3.8GHz, the physical memory is 32G, the memory frequency is 2400MHz, the graphics card is GeForce RTX 2080SUPER, equipped with 8GB Independent video memory, the operating system is
本次实验的具体流程为:The specific process of this experiment is as follows:
步骤1,面向含噪表格型数据,提出自采样学习方法(Self-sampling learning,SSL)。Step 1: A self-sampling learning (SSL) method is proposed for noisy tabular data.
步骤1.1,利用含噪声样本与正常样本在提升学习模型训练过程中损失函数值的变化模式,包括绝对值、变化方向及变化速度等,对样本可靠性、学习难度进行估计。Step 1.1, using the variation pattern of the loss function value during the training process of the boosting learning model, including the absolute value, the variation direction and the variation speed, of the noise-containing samples and the normal samples, to estimate the reliability of the samples and the learning difficulty.
具体来说,以二分类任务为例,假设训练集包含n个样本(x1,y1),…,(xn,yn),其中是第i个样本的特征向量,yi∈{-1,1}是第i个样本的标签。一般的有监督学习过程,将在该训练集上通过优化以下问题构建分类模型:Specifically, taking the binary classification task as an example, suppose the training set contains n samples (x 1 , y 1 ),...,(x n ,y n ), where is the feature vector of the ith sample, and y i ∈ {-1,1} is the label of the ith sample. A general supervised learning process will build a classification model on this training set by optimizing the following problems:
其中L(yi,f(xi,Θ))代表损失算法对应的损失函数,提升学习中强学习器由多个弱学习器组合而成,即为简化描述,下文将Fm(xi,Θ)写作Fm。自采样学习关注样本训练效果及训练速度,其形式为:where L(y i , f( xi , Θ)) represents the loss function corresponding to the loss algorithm, and the strong learner in boost learning is composed of multiple weak learners, namely To simplify the description, F m ( xi ,Θ) is hereinafter written as F m . Self-sampling learning focuses on the sample training effect and training speed, and its form is:
其中λ是自采样系数,决定下次迭代的训练样本规模,α为平衡系数,控制两种正则的作用强度。Among them, λ is the self-sampling coefficient, which determines the size of the training samples for the next iteration, and α is the balance coefficient, which controls the strength of the two regularities.
步骤1.2,调整每次迭代时的训练样本子集,在排除噪声点的同时优化样本学习顺序。Step 1.2, adjust the training sample subset at each iteration, and optimize the sample learning order while eliminating noise points.
步骤2,将自采样学习引入AdaBoost,建立SSAdaBoost和SSGB方法。
将自采样学习引入AdaBoost,建立SSAdaBoost方法。AdaBoost基于指数损失建立增量logistic回归模型,通过优化以下问题在迭代中建立新的弱学习器f(x):The self-sampling learning is introduced into AdaBoost, and the SSAdaBoost method is established. AdaBoost builds an incremental logistic regression model based on exponential loss, and builds a new weak learner f(x) in iterations by optimizing the following problems:
SSAdaBoost训练目标加入自采样正则,约束弱学习器训练样本集:The SSAdaBoost training target adds self-sampling regularization to constrain the weak learner training sample set:
其中,α是平衡系数,λ是采样率系数,可通过交替更新F(x)和自采样权重v对上式进行求解。方法将强学习器作为优化对象,更新F(x)意味着训练一个新的弱学习器f(x),并加权组合入强学习器中。Among them, α is the balance coefficient, λ is the sampling rate coefficient, and the above formula can be solved by alternately updating F(x) and the self-sampling weight v. The method takes the strong learner as the optimization object, and updating F(x) means training a new weak learner f(x), which is weighted and combined into the strong learner.
每次迭代中,首先固定v,使优化目标的子成分变为一个常数值,此时优化问题(11)变为加权指数损失最小化问题:In each iteration, v is first fixed so that the subcomponents of the optimization objective are becomes a constant value, and the optimization problem (11) becomes a weighted exponential loss minimization problem:
步骤2.1,建立目标列车性能估计模型,挖掘含噪运行数据中的专车性能知识,拟合加速度与速度、级位序列(隐含延时特性)、坡度等特征间的映射函数。Step 2.1, establish the target train performance estimation model, mine the special car performance knowledge in the noisy running data, and fit the mapping function between acceleration and speed, level sequence (implicit delay characteristic), gradient and other features.
该优化问题与AdaBoost原函数(10)相似,可采用相同的求解方法,通过拟牛顿法建立增量逻辑回归模型最小化E(ve-yF(x))。将上式进行二阶展开:This optimization problem is similar to the original function of AdaBoost (10), and the same solution method can be used to establish an incremental logistic regression model to minimize E(ve -yF(x) ) by the quasi-Newton method. Expand the above formula to the second order:
关于f(x)∈{-1,1}且c>0,对上式逐点进行最小化可得:Regarding f(x)∈{-1,1} and c>0, the point-by-point minimization of the above formula can be obtained:
其中viwi为样本权重,上式的二阶近似最小值仍是对f(xi)∈{-1,1}的加权最小二乘。可维持原算法中的计算模式在样本权重的基础上训练新的弱学习器f,而目标函数L(F+cf)关于c是凸的,可直接令其导数为0获得弱学习器权重c:where v i w i is the sample weight, and the second-order approximate minimum value of the above formula is still the weighted least squares of f(x i )∈{-1,1}. The calculation mode in the original algorithm can be maintained to train a new weak learner f based on the sample weight, and the objective function L(F+cf) is convex with respect to c, and the weak learner weight c can be obtained directly by setting its derivative to 0 :
其中如果弱学习器的判别准确率低于50%,则c可能为负值。in If the discriminative accuracy of the weak learner is lower than 50%, c may be negative.
随后固定c和f,对问题(11)中的v进行求解:Then fix c and f, and solve v in problem (11):
从自采样权重解的形式可以看出,研究所提方法可基于样本以表征的学习效果及以表征的学习速率来完成样本可靠性估计与选择,提升训练集质量。From the form of the self-sampling weight solution, it can be seen that the proposed method can be The learning effect of representation and The learning rate of the representation is used to complete the estimation and selection of sample reliability and improve the quality of the training set.
将自采样学习方法引入梯度提升学习,提出SSGB方法。梯度提升学习的特点是通过类似梯度下降的方式最小化目标损失,弱学习器训练模式及权重为:The self-sampling learning method is introduced into the gradient boosting learning, and the SSGB method is proposed. The characteristic of gradient boosting learning is to minimize the target loss in a similar way to gradient descent. The weak learner training mode and weight are:
SSGB的推导及优化过程与SSAdaBoost十分相似,其超参数也可采用相同的策略确定。The derivation and optimization process of SSGB is very similar to SSAdaBoost, and its hyperparameters can also be determined by the same strategy.
步骤3,利用“查询样本”控制级位序列、延时、坡度等特征影响,依次查询不同速度区间下各级位对应加速度值。Step 3: Use the "query sample" to control the influence of features such as level sequence, delay, and gradient, and sequentially query the acceleration values corresponding to the lower levels in different speed ranges.
假设模型拟合了特征(xa,xb,xc)与标签y的函数映射,且已知xb=cb,xc=cc时,对标签y无影响(例如xb代表校正坡度,cb=0,y代表加速度,校正坡度非零时与加速度的具体关系未知,但xb=0时对其它特征或加速度均无影响),则定义(qa,cb,cc)为查询样本,可用于抽取xa=qa时对应的标签y值。Assuming that the model fits the function mapping between the feature (x a , x b , x c ) and the label y, and it is known that x b = c b , x c = c c , it has no effect on the label y (for example, x b represents the correction Slope, c b = 0, y represents acceleration, the specific relationship between the correction slope and acceleration is unknown, but x b = 0 has no effect on other features or acceleration), then define (q a , c b , c c ) is a query sample, which can be used to extract the corresponding label y value when x a =q a .
步骤3.1,建立列车专属性能表,实现对目标特征与标签间量化关系抽取。Step 3.1: Establish a train-specific performance table to extract the quantitative relationship between target features and labels.
测试结果:基于西安机场线真实数据,结合自动驾驶系统及所提元提升学习准确运动仿真,测试性能自适应推荐速度优化效果。结果表明,SSAdaBoost相比现有方法显著提升了模型的稳健性,可面向延时含噪运行数据准确建模,方法抽取所得列车性能与实际相近。Test results: Based on the real data of Xi'an Airport Line, combined with the automatic driving system and the proposed element to improve the learning and accurate motion simulation, the test performance is adaptive to recommend the speed optimization effect. The results show that SSAdaBoost significantly improves the robustness of the model compared with the existing methods, and can accurately model the delayed and noisy running data, and the performance of the train extracted by the method is similar to the actual one.
以上所述的具体描述,对发明的目的、技术方案和有益效果进行了进一步详细说明,所应理解的是,以上所述仅为本发明的具体实施例而已,并不用于限定本发明的保护范围,凡在本发明的精神和原则之内,所做的任何修改、等同替换、改进等,均应包含在本发明的保护范围之内。The above-mentioned specific descriptions further describe the purpose, technical solutions and beneficial effects of the present invention in detail. It should be understood that the above-mentioned descriptions are only specific embodiments of the present invention, and are not intended to limit the protection of the present invention. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention shall be included within the protection scope of the present invention.
Claims (6)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210536133.5A CN115062402A (en) | 2022-05-17 | 2022-05-17 | A data-driven method for extracting train level acceleration |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210536133.5A CN115062402A (en) | 2022-05-17 | 2022-05-17 | A data-driven method for extracting train level acceleration |
Publications (1)
Publication Number | Publication Date |
---|---|
CN115062402A true CN115062402A (en) | 2022-09-16 |
Family
ID=83197945
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210536133.5A Pending CN115062402A (en) | 2022-05-17 | 2022-05-17 | A data-driven method for extracting train level acceleration |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115062402A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115600044A (en) * | 2022-11-28 | 2023-01-13 | 湖南大学(Cn) | Calculation method, device, equipment and storage medium of river section flow |
-
2022
- 2022-05-17 CN CN202210536133.5A patent/CN115062402A/en active Pending
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115600044A (en) * | 2022-11-28 | 2023-01-13 | 湖南大学(Cn) | Calculation method, device, equipment and storage medium of river section flow |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111930110B (en) | An intention trajectory prediction method combined with social generative adversarial network | |
Lao et al. | Intelligent fault diagnosis for rail transit switch machine based on adaptive feature selection and improved LightGBM | |
WO2020220439A1 (en) | Highway traffic flow state recognition method based on deep neural network | |
CN106844949B (en) | Training method of bidirectional LSTM model for realizing energy-saving control of locomotive | |
Yin et al. | Data-driven models for train control dynamics in high-speed railways: LAG-LSTM for train trajectory prediction | |
CN112650204B (en) | A method and system for identifying fault genes of intelligent rail unmanned vehicles | |
CN110378869A (en) | A kind of rail fastening method for detecting abnormality of sample automatic marking | |
CN109703606B (en) | Intelligent driving control method of high-speed train based on historical operation data | |
CN113044064B (en) | Meta-reinforcement learning-based vehicle adaptive autonomous driving decision-making method and system | |
Yin et al. | Smart train operation algorithms based on expert knowledge and ensemble CART for the electric locomotive | |
CN111598325A (en) | Traffic Speed Prediction Method Based on Hierarchical Clustering and Hierarchical Attention Mechanism | |
Zhang et al. | Data-driven train operation models based on data mining and driving experience for the diesel-electric locomotive | |
CN112308336A (en) | High-speed railway high wind speed limit dynamic disposal method based on multi-step time sequence prediction | |
CN113313941B (en) | Vehicle track prediction method based on memory network and encoder-decoder model | |
CN113610188A (en) | Bow net contact force non-section abnormity identification method and device | |
US20220126894A1 (en) | Deep learning - based stop control method and system for high-speed train | |
CN110084185A (en) | A kind of bullet train slightly crawls the rapid extracting method of operation characteristic | |
CN114882069A (en) | Taxi track abnormity detection method based on LSTM network and attention mechanism | |
CN116310542A (en) | An Image Classification Method Based on Improved Cross-Entropy Loss Function | |
Xu et al. | Robustness analysis of discrete state-based reinforcement learning models in traffic signal control | |
Zhou et al. | Fault classification for on-board equipment of high-speed railway based on attention capsule network | |
CN115062402A (en) | A data-driven method for extracting train level acceleration | |
CN107423761A (en) | Feature based selects and the rail locomotive energy saving optimizing method of operating of machine learning | |
CN114973665A (en) | Short-term traffic flow prediction method combining data decomposition and deep learning | |
CN113076235A (en) | Time sequence abnormity detection method based on state fusion |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |