CN115062402A - A data-driven method for extracting train level acceleration - Google Patents

A data-driven method for extracting train level acceleration Download PDF

Info

Publication number
CN115062402A
CN115062402A CN202210536133.5A CN202210536133A CN115062402A CN 115062402 A CN115062402 A CN 115062402A CN 202210536133 A CN202210536133 A CN 202210536133A CN 115062402 A CN115062402 A CN 115062402A
Authority
CN
China
Prior art keywords
learning
sampling
train
acceleration
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210536133.5A
Other languages
Chinese (zh)
Inventor
罗森林
周瑾洁
潘丽敏
刘晓双
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Institute of Technology BIT
Original Assignee
Beijing Institute of Technology BIT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Institute of Technology BIT filed Critical Beijing Institute of Technology BIT
Priority to CN202210536133.5A priority Critical patent/CN115062402A/en
Publication of CN115062402A publication Critical patent/CN115062402A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F30/00Computer-aided design [CAD]
    • G06F30/10Geometric CAD
    • G06F30/15Vehicle, aircraft or watercraft design
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F30/00Computer-aided design [CAD]
    • G06F30/20Design optimisation, verification or simulation
    • G06F30/27Design optimisation, verification or simulation using machine learning, e.g. artificial intelligence, neural networks, support vector machines [SVM] or training a model
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2119/00Details relating to the type or aim of the analysis or the optimisation
    • G06F2119/10Noise analysis or noise optimisation

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Geometry (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Hardware Design (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • Automation & Control Theory (AREA)
  • Aviation & Aerospace Engineering (AREA)
  • Computational Mathematics (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Pure & Applied Mathematics (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention relates to a data-driven train level acceleration extraction method, and belongs to the technical field of computers and information science. The method firstly optimizes the learning sequence of the samples by utilizing different loss change modes of the noise and the normal samples in the training process. Then introducing self-sampling learning into AdaBoost and gradient boosting learning respectively to provide an SSAdaBoost and SSGB steady learning method, establishing a target train performance estimation model, mining special train performance knowledge in noisy operation data, and fitting a mapping function between acceleration and characteristics such as speed, a rank sequence (implicit delay characteristic) and gradient. And finally, controlling the characteristic influence of a level sequence, time delay and the like by using a 'query sample', establishing a performance table, and realizing the extraction of the quantitative relation between the target characteristic and the label. The problems that the tracking difficulty of the recommended speed is high and the extra energy consumption is caused by frequent switching of the vehicle control level in the existing method are solved. The extraction performance is similar to the actual performance, and the method can be used for establishing performance constraint matched with the controlled train and improving the optimization effect of the recommendation speed.

Description

数据驱动的列车级位加速度提取方法A data-driven method for extracting train level acceleration

技术领域technical field

本发明涉及数据驱动的列车级位加速度提取方法,属于计算机与信息科学技术领域。The invention relates to a data-driven train level acceleration extraction method, and belongs to the technical field of computer and information science.

背景技术Background technique

发展轨道交通是解决现代大城市居民出行问题的有效途径,也是建设绿色城市、智能城市的重要方式。常见的轨道交通有传统铁路(国家铁路、城际铁路和市域铁路)、地铁、轻轨和有轨电车,新型轨道交通有磁悬浮轨道系统、单轨系统(跨座式轨道系统和悬挂式轨道系统)和旅客自动捷运系统等。随着火车和铁路技术的多元化发展,轨道交通呈现出越来越多的类型,不仅遍布于长距离的陆地运输,也广泛运用于中短距离的城市公共交通中,对列车进行性能抽取、监控其实时运行状态、保障其安全运营具有重要意义。The development of rail transit is an effective way to solve the travel problem of residents in modern big cities, and it is also an important way to build a green city and a smart city. Common rail transits include traditional railways (national railways, intercity railways and municipal railways), subways, light rails and trams, and new types of rail transits include maglev rail systems, monorail systems (straddle rail systems and suspended rail systems) and Passenger automatic rapid transit system, etc. With the diversified development of trains and railway technology, rail transit has shown more and more types, not only in long-distance land transportation, but also widely used in short- and medium-distance urban public transportation. It is of great significance to monitor its real-time operating status and ensure its safe operation.

列车运行数据主要包含车载系统采集的速度、推荐速度、级位,以及CBTC反馈的位置、坡度、弯道及目标点位置等信息,逐周期以格式化日志形式存储。运动状态估计问题中,将对相邻采样点的速度进行差值计算获取加速度,作为有监督学习标签挖掘当前速度、坡度及级位影响下的列车运动状态变化模式。The train operation data mainly includes the speed, recommended speed and level collected by the on-board system, as well as the position, slope, curve and target point position fed back by the CBTC, which are stored in formatted logs cycle by cycle. In the motion state estimation problem, the speed difference between adjacent sampling points is calculated to obtain the acceleration, which is used as a supervised learning label to mine the change pattern of the train motion state under the influence of the current speed, slope and level.

车辆运行时,受传感器误差、通信丢包、处理延时等多种因素影响,采样数据易出现波动、跳点、缺失等情况,导致计算加速度时产生错误,引入标签噪声。推荐速度是满足限速、时刻表等约束,面向列车运行效率、能耗、舒适度等指标优化生成的理想运行速度。现有推荐速度优化方法一般基于连续或离散假设,重点关注控制目标的能耗水平。针对列车常用的级位控速系统,离散假设可一定程度上考虑列车延时、加速度离散等问题,但现有方法未能对列车性能进行分析与抽取,优化过程仍与受控对象相独立,推荐加速度可能与列车加速度可行域不匹配,导致运行时级位切换频繁,产生预期之外的能耗。When the vehicle is running, affected by various factors such as sensor error, communication packet loss, processing delay, etc., the sampled data is prone to fluctuations, jumps, missing, etc., resulting in errors in the calculation of acceleration and the introduction of label noise. The recommended speed is the ideal running speed that satisfies constraints such as speed limit and timetable, and is optimized for train operation efficiency, energy consumption, comfort and other indicators. Existing recommendation speed optimization methods are generally based on continuous or discrete assumptions, focusing on the energy consumption level of the control target. For the stage speed control system commonly used by trains, the discrete assumption can consider the train delay, acceleration dispersion and other issues to a certain extent, but the existing methods fail to analyze and extract the train performance, and the optimization process is still independent of the controlled object. The recommended acceleration may not match the feasible region of the train acceleration, resulting in frequent level switching during operation, resulting in unexpected energy consumption.

发明内容SUMMARY OF THE INVENTION

本发明的目的是针对现有方法未考虑推荐加速度与列车级位加速度的匹配,使推荐速度的跟踪难度大,控车级位频繁切换并产生额外能耗问题。提出了数据驱动的列车级位加速度提取方法。The purpose of the present invention is that the existing method does not consider the matching of the recommended acceleration and the train level acceleration, so that the tracking of the recommended speed is difficult, the control level is frequently switched, and additional energy consumption is generated. A data-driven extraction method of train level acceleration is proposed.

本发明的设计原理为:首先利用噪声样本与正常样本在训练过程中表现出的不同损失变化模式,对样本可靠性、学习难度进行估计,优化样本学习顺序。然后将自采样学习分别引入AdaBoost和梯度提升学习提出SSAdaBoost和SSGB稳健学习方法,建立目标列车性能估计模型,挖掘含噪运行数据中的专车性能知识,拟合加速度与速度、级位序列(隐含延时特性)、坡度等特征间的映射函数。最后提出基于特征控制的关联关系提取方法,利用“查询样本”控制级位序列、延时、坡度等特征影响,依次查询不同速度区间下各级位对应加速度值,建立列车专属性能表,实现对目标特征与标签间量化关系抽取。The design principle of the present invention is as follows: firstly, the reliability and learning difficulty of the samples are estimated by using the different loss change patterns exhibited by the noise samples and the normal samples in the training process, and the learning sequence of the samples is optimized. Then, the self-sampling learning is introduced into AdaBoost and gradient boosting learning respectively, and SSAdaBoost and SSGB robust learning methods are proposed to establish the target train performance estimation model, mine the special car performance knowledge in the noisy running data, and fit the acceleration, speed, and level sequence (implicitly). The mapping function between features such as delay characteristics), slope and so on. Finally, a correlation extraction method based on feature control is proposed. The "query sample" is used to control the influence of features such as stage sequence, delay, and slope, and the corresponding acceleration values of the lower stages in different speed ranges are inquired in turn to establish a train-specific performance table. Quantitative relationship extraction between target features and labels.

本发明的技术方案是通过如下步骤实现的:The technical scheme of the present invention is achieved through the following steps:

步骤1,面向含噪表格型数据,提出自采样学习方法SSL。Step 1, for noisy tabular data, a self-sampling learning method SSL is proposed.

步骤1.1,利用含噪声样本与正常样本在提升学习模型训练过程中损失函数值的变化模式,包括绝对值、变化方向及变化速度等,对样本可靠性、学习难度进行估计。Step 1.1, using the variation pattern of the loss function value during the training process of the boosting learning model, including the absolute value, the variation direction and the variation speed, of the noise-containing samples and the normal samples, to estimate the reliability of the samples and the learning difficulty.

步骤1.2,调整每次迭代时的训练样本子集,在排除噪声点的同时优化样本学习顺序。Step 1.2, adjust the training sample subset at each iteration, and optimize the sample learning order while eliminating noise points.

步骤2,将自采样学习引入AdaBoost,建立SSAdaBoost和SSGB方法。Step 2, introduce self-sampling learning into AdaBoost, and establish SSAdaBoost and SSGB methods.

步骤2.1,建立目标列车性能估计模型,挖掘含噪运行数据中的专车性能知识,拟合加速度与速度、级位序列(隐含延时特性)、坡度等特征间的映射函数。Step 2.1, establish the target train performance estimation model, mine the special car performance knowledge in the noisy running data, and fit the mapping function between acceleration and speed, level sequence (implicit delay characteristic), gradient and other features.

步骤3,利用“查询样本”控制级位序列、延时、坡度等特征影响,依次查询不同速度区间下各级位对应加速度值。Step 3: Use the "query sample" to control the influence of features such as level sequence, delay, and gradient, and sequentially query the acceleration values corresponding to the lower levels in different speed ranges.

步骤3.1,建立列车专属性能表,实现对目标特征与标签间量化关系抽取。Step 3.1: Establish a train-specific performance table to extract the quantitative relationship between target features and labels.

有益效果beneficial effect

与原始AdaBoost相比,SSAdaBoost的样本权重调整由

Figure BDA0003648275850000021
变化为viwi,有效限制了权重增长。通过对训练集进行筛选,学习器可在更稳定、低噪的样本上训练,相比SPLBoost,自采样学习可避免对非噪声困难样本的误判与抛弃,且无需样本学习顺序的先验知识。方法与原有AdaBoost相比,增加了两个采样控制超参数,计算成本(M×O(3kn+3n))与原方法(M×O(3kn+2n))接近。Compared with the original AdaBoost, the sample weight adjustment of SSAdaBoost is given by
Figure BDA0003648275850000021
Changed to viwi, which effectively limits the weight growth. By screening the training set, the learner can be trained on more stable and low-noise samples. Compared with SPLBoost, self-sampling learning can avoid the misjudgment and abandonment of non-noise difficult samples, and does not require prior knowledge of the sample learning order. . Compared with the original AdaBoost, the method adds two sampling control hyperparameters, and the computational cost (M×O(3kn+3n)) is close to the original method (M×O(3kn+2n)).

随列车运行速度的增加,相同级位的实际加速度可能出现改变,基本规律为:1)牵引级位对列车产生的实际加速度随速度的上升而减小,尤其在高速阶段减小明显;2)惰行级位对列车产生的实际减速度随速度上升而增大,主要受与速度相关的阻力影响;3)制动级位对列车产生的实际减速度随速度上升而增大,变化幅度相对较小。同型号列车上线前,会对代表性车辆进行首车动调,在可控测试线路上,测试各速度区间下,级位实际加速度,绘制完整的性能表。With the increase of train running speed, the actual acceleration of the same level may change. The basic rules are: 1) The actual acceleration of the train generated by the traction level decreases with the increase of speed, especially in the high-speed stage; 2) The actual deceleration of the train caused by the coasting stage increases with the speed increase, which is mainly affected by the resistance related to the speed; 3) The actual deceleration of the train caused by the braking stage increases with the speed increase, and the variation range is relatively large. Small. Before the trains of the same model go online, the first car will be adjusted for representative vehicles. On the controllable test line, the actual acceleration of each speed range will be tested, and a complete performance table will be drawn.

附图说明Description of drawings

图1为本发明数据驱动的列车级位加速度提取方法原理图。FIG. 1 is a schematic diagram of a data-driven method for extracting train level acceleration according to the present invention.

图2为自采样提升学习UCI任务模型错误率统计。Figure 2 shows the error rate statistics of the self-sampling boosting learning UCI task model.

图3为自采样提升学习大规模真实任务实验结果图。Figure 3 shows the experimental results of self-sampling boosting learning on large-scale real tasks.

图4为知识抽取性能表仿真效果对比。Figure 4 shows the comparison of the simulation effects of the knowledge extraction performance table.

具体实施方式Detailed ways

为了更好的说明本发明的目的和优点,下面结合实例对本发明方法的实施方式做进一步详细说明。In order to better illustrate the purpose and advantages of the present invention, the embodiments of the method of the present invention will be described in further detail below with reference to examples.

实验主要验证SSAdaBoost和SSGB自采样提升学习算法的稳健性及建模性能和基于“查询样本”进行特征控制的模型解释方法在列车性能抽取任务中的应用效果。The experiments mainly verify the robustness and modeling performance of SSAdaBoost and SSGB self-sampling learning algorithms and the application effect of the model interpretation method based on "query samples" for feature control in the task of train performance extraction.

(1)采用70组不同噪声水平的UCI公开数据集及3组大规模真实任务数据集对所提SSAdaBoost和SSGB自采样提升学习性能进行评价,对比算法包括AdaBoost、LogitBoost、GentleBoost、RBoost、CB-AdaBoost、GradientBoost等6种高性能提升学习算法,涵盖现有先进稳健提升学习算法。所提算法的稳健性将直接影响模型对含噪运行数据中列车性能知识的识别效果,是实现准确性能抽取及性能自适应推荐速度优化的前提。(1) 70 sets of UCI public data sets with different noise levels and 3 sets of large-scale real task data sets are used to evaluate the learning performance of the proposed SSAdaBoost and SSGB self-sampling improvement. The comparison algorithms include AdaBoost, LogitBoost, GentleBoost, RBoost, CB- Six high-performance boosting learning algorithms, including AdaBoost and GradientBoost, cover the existing advanced robust boosting learning algorithms. The robustness of the proposed algorithm will directly affect the recognition effect of the model on the knowledge of train performance in noisy running data, which is the premise of accurate performance extraction and performance-adaptive recommended speed optimization.

实验中按照每5%一个区间进行划分,噪声水平设置范围从0%到30%,针对每个任务建立7个子数据集,故UCI数据建模实验的总数据集数为70个。噪声仅添加于训练集及验证集,各测试集均不添加噪声,确保测试结果的可靠性。将560组实验结果按任务划分,绘制不同算法随噪声加剧的建模错误率变化曲线以便于观察。In the experiment, each 5% interval is divided, the noise level is set from 0% to 30%, and 7 sub-data sets are established for each task, so the total number of data sets in the UCI data modeling experiment is 70. Noise is only added to the training set and validation set, and no noise is added to each test set to ensure the reliability of the test results. The 560 sets of experimental results are divided according to tasks, and the curve of the modeling error rate of different algorithms with increasing noise is drawn for easy observation.

实验以有监督二分类任务为场景,由于部分任务数据存在标签不平衡问题,评价指标采用错误率(Error Rate)、准确率(Accuracy)、F1分数(F1Score)以及用于检验算法间性能显著性的Nemenyi检验法(Nemenyi Test),计算方法如公式1~4所示:The experiment takes the supervised two-classification task as the scene. Due to the label imbalance problem in some task data, the evaluation indicators use Error Rate, Accuracy, F1 Score and are used to test the performance significance between algorithms. The Nemenyi Test (Nemenyi Test), the calculation method is shown in formulas 1 to 4:

矩阵中,TP(True Positive)代表将正例预测为正例的数量;FN(False Negative)代表将正例预测为负例的数量;FP(False Positive)代表将负例预测为正例的数量;TN(True Negative)代表将负例预测为负例的数量。根据上述符号定义,实验的评价指标计算方式如下:In the matrix, TP (True Positive) represents the number of positive examples predicted as positive examples; FN (False Negative) represents the number of positive examples predicted as negative examples; FP (False Positive) represents the number of negative examples predicted as positive examples ; TN (True Negative) represents the number of negative examples predicted as negative examples. According to the above symbol definitions, the calculation method of the evaluation index of the experiment is as follows:

Figure BDA0003648275850000041
Figure BDA0003648275850000041

Figure BDA0003648275850000042
Figure BDA0003648275850000042

Figure BDA0003648275850000043
Figure BDA0003648275850000043

Nemenyi检验法属于后续检验(Post-Hoc Test)方法,用于对算法性能差异的显著性进行两两检验,检验指标为错误率,具备显著性则说明两算法间具备明显性能差异,不具备则代表性能相近。其中,临界差异值(Critical Difference,CD)由算法数量K及单个算法的实验组数N决定:The Nemenyi test method belongs to the post-Hoc Test method, which is used to test the significance of the algorithm performance difference in pairs. represent similar performance. Among them, the critical difference (Critical Difference, CD) is determined by the number of algorithms K and the number of experimental groups N of a single algorithm:

Figure BDA0003648275850000044
Figure BDA0003648275850000044

其中,临界值qα基于学生化极差(Studentized Range)统计量除以

Figure BDA0003648275850000047
获得,详细的参数设置方法参见实验过程。where the critical value q α is based on the Studentized Range statistic divided by
Figure BDA0003648275850000047
obtained, and the detailed parameter setting method can be found in the experimental procedure.

实验的过程中设置弱学习器树深度为2,弱学习器数量为50,SSAdaBoost的超参数为衰退系数δ的调整范围为[0.9,1),采样比例μ的调整范围为[0,0.4]。During the experiment, the depth of the weak learner tree is set to 2, the number of weak learners is 50, the hyperparameter of SSAdaBoost is the decay coefficient δ, the adjustment range is [0.9, 1), and the adjustment range of the sampling ratio μ is [0, 0.4] .

(2)验证所提基于“查询样本”进行特征控制的模型解释方法在列车性能抽取任务中的应用效果。由于列车各速度区间下不同级位的实际加速度不易测量,无法获取真实性能表与抽取性能表进行对比,采用“转换法”(Conversion Method)进行实验。基于物理仿真模型,采用抽取性能表作为列车级位加速度参数,分析运行仿真效果,仿真结果与实际运行结果越相似则代表抽取性能与实际性能越相近。同时采用列车出厂性能表建立仿真模型,用于对比分析抽取性能表的准确性。实验数据为西安地铁机场及合肥地铁三号线列车运行数据。(2) Verify the application effect of the proposed model interpretation method based on "query samples" for feature control in the task of train performance extraction. Because the actual acceleration at different levels in each speed range of the train is not easy to measure, it is impossible to obtain the real performance table and compare the extracted performance table, and the "Conversion Method" is used for the experiment. Based on the physical simulation model, the extraction performance table is used as the train stage acceleration parameter to analyze the operation simulation effect. The more similar the simulation results are to the actual operation results, the closer the extraction performance is to the actual performance. At the same time, a simulation model is established by using the train ex-factory performance table, which is used to compare and analyze the accuracy of the extraction performance table. The experimental data is the train operation data of Xi'an Metro Airport and Hefei Metro Line 3.

实验采用“转换法”对基于抽取性能表的物理模型仿真精度进行分析,等效评价性能抽取效果。性能抽取效果受列车性能估计模型影响,需对SSGB所建加速度估计模型进行评价。模型输出为连续变量,采用具有较高灵敏度的回归任务典型指标均方误差(MeanSquared Error,MSE)进行评价:In the experiment, the "transformation method" is used to analyze the simulation accuracy of the physical model based on the extraction performance table, and the performance extraction effect is equivalently evaluated. The performance extraction effect is affected by the train performance estimation model, and the acceleration estimation model built by SSGB needs to be evaluated. The output of the model is a continuous variable, and the mean squared error (MSE), a typical indicator of regression tasks with high sensitivity, is used for evaluation:

Figure BDA0003648275850000045
Figure BDA0003648275850000045

式中,yi代表第i个测试样本的真实标签值,

Figure BDA0003648275850000046
为模型输出的估计标签值,n为测试样本数量。MSE可对模型在测试集上的估计误差进行衡量,结果可代表平均绝对误差、均方根误差等相似指标的评价效果。In the formula, yi represents the true label value of the ith test sample,
Figure BDA0003648275850000046
is the estimated label value output by the model, and n is the number of test samples. MSE can measure the estimation error of the model on the test set, and the result can represent the evaluation effect of similar indicators such as mean absolute error and root mean square error.

通过比较基于抽取性能所得仿真曲线与真实曲线的相似度,对性能抽取效果进行分析,采用曲线面积误差率AUCerr、终点速度误差verr、终点位置误差serr等指标:By comparing the similarity between the simulation curve obtained based on the extraction performance and the real curve, the performance extraction effect is analyzed, and the curve area error rate AUC err , the end point speed error verr , the end point position error s err and other indicators are used:

Figure BDA0003648275850000051
Figure BDA0003648275850000051

Figure BDA0003648275850000052
Figure BDA0003648275850000052

实验同时绘制曲线定性分析曲线相似度,比价对象包括真实速度-位置曲线、基于抽取性能表的仿真曲线、基于出厂性能表的仿真曲线。At the same time, the curve is drawn qualitatively to analyze the similarity of the curve, and the comparison objects include the real speed-position curve, the simulation curve based on the extraction performance table, and the simulation curve based on the factory performance table.

本次实验在硬件环境为MSI Prestige台式计算机,CPU型号为Intel Core i7-10700K八核十六线程处理器,CPU主频3.8GHz,物理内存32G,内存频率2400MHz,显卡为GeForce RTX 2080SUPER,配备8GB独立显存,操作系统是windows 10,64位。The hardware environment of this experiment is MSI Prestige desktop computer, the CPU model is Intel Core i7-10700K octa-core 16-thread processor, the CPU frequency is 3.8GHz, the physical memory is 32G, the memory frequency is 2400MHz, the graphics card is GeForce RTX 2080SUPER, equipped with 8GB Independent video memory, the operating system is windows 10, 64 bit.

本次实验的具体流程为:The specific process of this experiment is as follows:

步骤1,面向含噪表格型数据,提出自采样学习方法(Self-sampling learning,SSL)。Step 1: A self-sampling learning (SSL) method is proposed for noisy tabular data.

步骤1.1,利用含噪声样本与正常样本在提升学习模型训练过程中损失函数值的变化模式,包括绝对值、变化方向及变化速度等,对样本可靠性、学习难度进行估计。Step 1.1, using the variation pattern of the loss function value during the training process of the boosting learning model, including the absolute value, the variation direction and the variation speed, of the noise-containing samples and the normal samples, to estimate the reliability of the samples and the learning difficulty.

具体来说,以二分类任务为例,假设训练集包含n个样本(x1,y1),…,(xn,yn),其中

Figure BDA0003648275850000053
是第i个样本的特征向量,yi∈{-1,1}是第i个样本的标签。一般的有监督学习过程,将在该训练集上通过优化以下问题构建分类模型:Specifically, taking the binary classification task as an example, suppose the training set contains n samples (x 1 , y 1 ),...,(x n ,y n ), where
Figure BDA0003648275850000053
is the feature vector of the ith sample, and y i ∈ {-1,1} is the label of the ith sample. A general supervised learning process will build a classification model on this training set by optimizing the following problems:

Figure BDA0003648275850000054
Figure BDA0003648275850000054

其中L(yi,f(xi,Θ))代表损失算法对应的损失函数,提升学习中强学习器由多个弱学习器组合而成,即

Figure BDA0003648275850000055
为简化描述,下文将Fm(xi,Θ)写作Fm。自采样学习关注样本训练效果及训练速度,其形式为:where L(y i , f( xi , Θ)) represents the loss function corresponding to the loss algorithm, and the strong learner in boost learning is composed of multiple weak learners, namely
Figure BDA0003648275850000055
To simplify the description, F m ( xi ,Θ) is hereinafter written as F m . Self-sampling learning focuses on the sample training effect and training speed, and its form is:

Figure BDA0003648275850000056
Figure BDA0003648275850000056

其中λ是自采样系数,决定下次迭代的训练样本规模,α为平衡系数,控制两种正则的作用强度。Among them, λ is the self-sampling coefficient, which determines the size of the training samples for the next iteration, and α is the balance coefficient, which controls the strength of the two regularities.

步骤1.2,调整每次迭代时的训练样本子集,在排除噪声点的同时优化样本学习顺序。Step 1.2, adjust the training sample subset at each iteration, and optimize the sample learning order while eliminating noise points.

步骤2,将自采样学习引入AdaBoost,建立SSAdaBoost和SSGB方法。Step 2, introduce self-sampling learning into AdaBoost, and establish SSAdaBoost and SSGB methods.

将自采样学习引入AdaBoost,建立SSAdaBoost方法。AdaBoost基于指数损失建立增量logistic回归模型,通过优化以下问题在迭代中建立新的弱学习器f(x):The self-sampling learning is introduced into AdaBoost, and the SSAdaBoost method is established. AdaBoost builds an incremental logistic regression model based on exponential loss, and builds a new weak learner f(x) in iterations by optimizing the following problems:

Figure BDA0003648275850000061
Figure BDA0003648275850000061

SSAdaBoost训练目标加入自采样正则,约束弱学习器训练样本集:The SSAdaBoost training target adds self-sampling regularization to constrain the weak learner training sample set:

Figure BDA0003648275850000062
Figure BDA0003648275850000062

其中,α是平衡系数,λ是采样率系数,可通过交替更新F(x)和自采样权重v对上式进行求解。方法将强学习器作为优化对象,更新F(x)意味着训练一个新的弱学习器f(x),并加权组合入强学习器中。Among them, α is the balance coefficient, λ is the sampling rate coefficient, and the above formula can be solved by alternately updating F(x) and the self-sampling weight v. The method takes the strong learner as the optimization object, and updating F(x) means training a new weak learner f(x), which is weighted and combined into the strong learner.

每次迭代中,首先固定v,使优化目标的子成分

Figure BDA0003648275850000063
变为一个常数值,此时优化问题(11)变为加权指数损失最小化问题:In each iteration, v is first fixed so that the subcomponents of the optimization objective are
Figure BDA0003648275850000063
becomes a constant value, and the optimization problem (11) becomes a weighted exponential loss minimization problem:

Figure BDA0003648275850000064
Figure BDA0003648275850000064

步骤2.1,建立目标列车性能估计模型,挖掘含噪运行数据中的专车性能知识,拟合加速度与速度、级位序列(隐含延时特性)、坡度等特征间的映射函数。Step 2.1, establish the target train performance estimation model, mine the special car performance knowledge in the noisy running data, and fit the mapping function between acceleration and speed, level sequence (implicit delay characteristic), gradient and other features.

该优化问题与AdaBoost原函数(10)相似,可采用相同的求解方法,通过拟牛顿法建立增量逻辑回归模型最小化E(ve-yF(x))。将上式进行二阶展开:This optimization problem is similar to the original function of AdaBoost (10), and the same solution method can be used to establish an incremental logistic regression model to minimize E(ve -yF(x) ) by the quasi-Newton method. Expand the above formula to the second order:

Figure BDA0003648275850000065
Figure BDA0003648275850000065

关于f(x)∈{-1,1}且c>0,对上式逐点进行最小化可得:Regarding f(x)∈{-1,1} and c>0, the point-by-point minimization of the above formula can be obtained:

Figure BDA0003648275850000066
Figure BDA0003648275850000066

其中viwi为样本权重,上式的二阶近似最小值仍是对f(xi)∈{-1,1}的加权最小二乘。可维持原算法中的计算模式在样本权重的基础上训练新的弱学习器f,而目标函数L(F+cf)关于c是凸的,可直接令其导数为0获得弱学习器权重c:where v i w i is the sample weight, and the second-order approximate minimum value of the above formula is still the weighted least squares of f(x i )∈{-1,1}. The calculation mode in the original algorithm can be maintained to train a new weak learner f based on the sample weight, and the objective function L(F+cf) is convex with respect to c, and the weak learner weight c can be obtained directly by setting its derivative to 0 :

Figure BDA0003648275850000071
Figure BDA0003648275850000071

其中

Figure BDA0003648275850000072
如果弱学习器的判别准确率低于50%,则c可能为负值。in
Figure BDA0003648275850000072
If the discriminative accuracy of the weak learner is lower than 50%, c may be negative.

随后固定c和f,对问题(11)中的v进行求解:Then fix c and f, and solve v in problem (11):

Figure BDA0003648275850000073
Figure BDA0003648275850000073

从自采样权重解的形式可以看出,研究所提方法可基于样本以

Figure BDA0003648275850000074
表征的学习效果及以
Figure BDA0003648275850000075
表征的学习速率来完成样本可靠性估计与选择,提升训练集质量。From the form of the self-sampling weight solution, it can be seen that the proposed method can be
Figure BDA0003648275850000074
The learning effect of representation and
Figure BDA0003648275850000075
The learning rate of the representation is used to complete the estimation and selection of sample reliability and improve the quality of the training set.

将自采样学习方法引入梯度提升学习,提出SSGB方法。梯度提升学习的特点是通过类似梯度下降的方式最小化目标损失,弱学习器训练模式及权重为:The self-sampling learning method is introduced into the gradient boosting learning, and the SSGB method is proposed. The characteristic of gradient boosting learning is to minimize the target loss in a similar way to gradient descent. The weak learner training mode and weight are:

Figure BDA0003648275850000076
Figure BDA0003648275850000076

SSGB的推导及优化过程与SSAdaBoost十分相似,其超参数也可采用相同的策略确定。The derivation and optimization process of SSGB is very similar to SSAdaBoost, and its hyperparameters can also be determined by the same strategy.

步骤3,利用“查询样本”控制级位序列、延时、坡度等特征影响,依次查询不同速度区间下各级位对应加速度值。Step 3: Use the "query sample" to control the influence of features such as level sequence, delay, and gradient, and sequentially query the acceleration values corresponding to the lower levels in different speed ranges.

假设模型拟合了特征(xa,xb,xc)与标签y的函数映射,且已知xb=cb,xc=cc时,对标签y无影响(例如xb代表校正坡度,cb=0,y代表加速度,校正坡度非零时与加速度的具体关系未知,但xb=0时对其它特征或加速度均无影响),则定义(qa,cb,cc)为查询样本,可用于抽取xa=qa时对应的标签y值。Assuming that the model fits the function mapping between the feature (x a , x b , x c ) and the label y, and it is known that x b = c b , x c = c c , it has no effect on the label y (for example, x b represents the correction Slope, c b = 0, y represents acceleration, the specific relationship between the correction slope and acceleration is unknown, but x b = 0 has no effect on other features or acceleration), then define (q a , c b , c c ) is a query sample, which can be used to extract the corresponding label y value when x a =q a .

步骤3.1,建立列车专属性能表,实现对目标特征与标签间量化关系抽取。Step 3.1: Establish a train-specific performance table to extract the quantitative relationship between target features and labels.

测试结果:基于西安机场线真实数据,结合自动驾驶系统及所提元提升学习准确运动仿真,测试性能自适应推荐速度优化效果。结果表明,SSAdaBoost相比现有方法显著提升了模型的稳健性,可面向延时含噪运行数据准确建模,方法抽取所得列车性能与实际相近。Test results: Based on the real data of Xi'an Airport Line, combined with the automatic driving system and the proposed element to improve the learning and accurate motion simulation, the test performance is adaptive to recommend the speed optimization effect. The results show that SSAdaBoost significantly improves the robustness of the model compared with the existing methods, and can accurately model the delayed and noisy running data, and the performance of the train extracted by the method is similar to the actual one.

以上所述的具体描述,对发明的目的、技术方案和有益效果进行了进一步详细说明,所应理解的是,以上所述仅为本发明的具体实施例而已,并不用于限定本发明的保护范围,凡在本发明的精神和原则之内,所做的任何修改、等同替换、改进等,均应包含在本发明的保护范围之内。The above-mentioned specific descriptions further describe the purpose, technical solutions and beneficial effects of the present invention in detail. It should be understood that the above-mentioned descriptions are only specific embodiments of the present invention, and are not intended to limit the protection of the present invention. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention shall be included within the protection scope of the present invention.

Claims (6)

1. The data-driven train level acceleration extraction method is characterized by comprising the following steps of:
step 1, aiming at noisy table type data, a self-sampling learning method SSL is provided: estimating the reliability and the learning difficulty of the samples by using the change modes of the loss function values of the noise-containing samples and the normal samples in the training process of the learning model, including the absolute values, the change directions, the change speeds and the like, adjusting the training sample subset in each iteration, and optimizing the learning sequence of the samples while eliminating noise points;
step 2, introducing self-sampling learning into AdaBoost, establishing an SSAdaBoost and SSGB method, establishing a target train performance estimation model, mining special train performance knowledge in noisy operation data, and fitting a mapping function between acceleration and characteristics such as speed, a level sequence (implicit delay characteristic) and gradient;
and 3, controlling the characteristic influences of the level sequence, the time delay, the gradient and the like by using the 'query sample', sequentially querying the corresponding acceleration values of each level in different speed intervals, establishing a train exclusive performance table, and realizing the extraction of the quantitative relation between the target characteristic and the label.
2. The data-driven train-level acceleration extraction method of claim 1, characterized in that: in step 1, the present invention is directed to noisy tabular data.
3. The data-driven train-level acceleration extraction method of claim 1, characterized in that: in step 2, the method for establishing the SSAdaBoost and the SSGB is realized by introducing self-sampling learning into the AdaBoost and gradient boosting learning, the self-sampling learning focuses on the training effect and the training speed of the sample, and the form is as follows:
Figure FDA0003648275840000011
wherein, lambda is a self-sampling coefficient, the scale of the training sample of the next iteration is determined, alpha is a balance coefficient, and the action intensity of two types of regularization is controlled.
4. The data-driven train-level acceleration extraction method of claim 1, characterized in that: in step 2, adding a self-sampling regular into an SSAdaBoost training target, and constraining a weak learner training sample set:
Figure FDA0003648275840000012
wherein alpha is a balance coefficient, lambda is a sampling rate coefficient, the above formula can be solved by alternately updating F (x) and self-sampling weight v, the method takes a strong learner as an optimization object, updating F (x) means training a new weak learner f (x) and performing weighted combination into the strong learning, and the input of the method is a training sample set { (x) 1 ,y 1 ),…,(x n ,y n ) Training iteration times M, sampling proportion mu, balance coefficient alpha being 0.5, decay coefficient delta and outputting as a strong classifier F M (x)。
5. The data-driven train-level acceleration extraction method of claim 1, characterized in that: in step 3, a self-sampling learning method is introduced into gradient boosting learning, and an SSGB method is provided, wherein the gradient boosting learning is characterized in that target loss is minimized in a gradient descending similar mode, and a weak learner training mode and weight are as follows:
Figure FDA0003648275840000021
Figure FDA0003648275840000022
the input of the method is a training sample set { (x) 1 ,y 1 ),…,(x n ,y n ) The training iteration times M, the sampling proportion mu, the balance coefficient alpha equal to 0.5 and the decay coefficient delta, and the strong classifier F is output M (x)。
6. The data-driven train-level acceleration extraction method of claim 1, characterized in that: in step 3, the hypothesis model fits the features (x) a ,x b ,x c ) Mapping with a function of tag y, and knowing x b =c b ,x c =c c When it is time, it has no effect on tag y (e.g., x) b Representing the correction gradient, c b With 0, y representing acceleration, the specific relationship to acceleration when correcting for non-zero slope is unknown, but x b When 0 has no influence on other characteristics or acceleration), then (q) is defined a ,c b ,c c ) For query samples, it can be used to extract x a =q a The corresponding tag y value.
CN202210536133.5A 2022-05-17 2022-05-17 A data-driven method for extracting train level acceleration Pending CN115062402A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210536133.5A CN115062402A (en) 2022-05-17 2022-05-17 A data-driven method for extracting train level acceleration

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210536133.5A CN115062402A (en) 2022-05-17 2022-05-17 A data-driven method for extracting train level acceleration

Publications (1)

Publication Number Publication Date
CN115062402A true CN115062402A (en) 2022-09-16

Family

ID=83197945

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210536133.5A Pending CN115062402A (en) 2022-05-17 2022-05-17 A data-driven method for extracting train level acceleration

Country Status (1)

Country Link
CN (1) CN115062402A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115600044A (en) * 2022-11-28 2023-01-13 湖南大学(Cn) Calculation method, device, equipment and storage medium of river section flow

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115600044A (en) * 2022-11-28 2023-01-13 湖南大学(Cn) Calculation method, device, equipment and storage medium of river section flow

Similar Documents

Publication Publication Date Title
CN111930110B (en) An intention trajectory prediction method combined with social generative adversarial network
Lao et al. Intelligent fault diagnosis for rail transit switch machine based on adaptive feature selection and improved LightGBM
WO2020220439A1 (en) Highway traffic flow state recognition method based on deep neural network
CN106844949B (en) Training method of bidirectional LSTM model for realizing energy-saving control of locomotive
Yin et al. Data-driven models for train control dynamics in high-speed railways: LAG-LSTM for train trajectory prediction
CN112650204B (en) A method and system for identifying fault genes of intelligent rail unmanned vehicles
CN110378869A (en) A kind of rail fastening method for detecting abnormality of sample automatic marking
CN109703606B (en) Intelligent driving control method of high-speed train based on historical operation data
CN113044064B (en) Meta-reinforcement learning-based vehicle adaptive autonomous driving decision-making method and system
Yin et al. Smart train operation algorithms based on expert knowledge and ensemble CART for the electric locomotive
CN111598325A (en) Traffic Speed Prediction Method Based on Hierarchical Clustering and Hierarchical Attention Mechanism
Zhang et al. Data-driven train operation models based on data mining and driving experience for the diesel-electric locomotive
CN112308336A (en) High-speed railway high wind speed limit dynamic disposal method based on multi-step time sequence prediction
CN113313941B (en) Vehicle track prediction method based on memory network and encoder-decoder model
CN113610188A (en) Bow net contact force non-section abnormity identification method and device
US20220126894A1 (en) Deep learning - based stop control method and system for high-speed train
CN110084185A (en) A kind of bullet train slightly crawls the rapid extracting method of operation characteristic
CN114882069A (en) Taxi track abnormity detection method based on LSTM network and attention mechanism
CN116310542A (en) An Image Classification Method Based on Improved Cross-Entropy Loss Function
Xu et al. Robustness analysis of discrete state-based reinforcement learning models in traffic signal control
Zhou et al. Fault classification for on-board equipment of high-speed railway based on attention capsule network
CN115062402A (en) A data-driven method for extracting train level acceleration
CN107423761A (en) Feature based selects and the rail locomotive energy saving optimizing method of operating of machine learning
CN114973665A (en) Short-term traffic flow prediction method combining data decomposition and deep learning
CN113076235A (en) Time sequence abnormity detection method based on state fusion

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination