CN109146007B - Solid waste intelligent treatment method based on dynamic deep belief network - Google Patents

Solid waste intelligent treatment method based on dynamic deep belief network Download PDF

Info

Publication number
CN109146007B
CN109146007B CN201810768405.8A CN201810768405A CN109146007B CN 109146007 B CN109146007 B CN 109146007B CN 201810768405 A CN201810768405 A CN 201810768405A CN 109146007 B CN109146007 B CN 109146007B
Authority
CN
China
Prior art keywords
ddbn
rbm
solid waste
training
neuron
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810768405.8A
Other languages
Chinese (zh)
Other versions
CN109146007A (en
Inventor
宋威
张士昱
王晨妮
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiangnan University
Original Assignee
Jiangnan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiangnan University filed Critical Jiangnan University
Priority to CN201810768405.8A priority Critical patent/CN109146007B/en
Publication of CN109146007A publication Critical patent/CN109146007A/en
Application granted granted Critical
Publication of CN109146007B publication Critical patent/CN109146007B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/19Recognition using electronic means
    • G06V30/192Recognition using electronic means using simultaneous comparisons or correlations of the image signals with a plurality of references
    • G06V30/194References adjustable by an adaptive method, e.g. learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/06Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
    • G06N3/061Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using biological neurons, e.g. biological neurons connected to an integrated circuit
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Molecular Biology (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • Neurology (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Databases & Information Systems (AREA)
  • Microelectronics & Electronic Packaging (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

本发明提出一种基于动态深度置信网络的固体废弃物智能处理方法,属于深度学习、固体废弃物智能处理领域。该方法首先提出一种使用动态增减枝算法的DDBN,使DDBN在训练过程中根据当前训练情况增加隐藏层神经元和隐藏层,以及移除冗余神经元,有效的优化DDBN的网络结构。然后,利用DDBN能有效提取原始数据主要特征的优势,使用DDBN对固体废弃物随机、离散、非线性的特征向量进行有效的状态描述,使时间序列的状态特征更加易于鉴别,并确保不丢失原始数据的主要信息。同时,根据提取到的固体废弃物的状态描述,利用DDBN预测适合其状态的优化燃烧行为,减少了盲目燃烧行为对资源的浪费,实现对固体废弃物的智能处理。

Figure 201810768405

The invention proposes a solid waste intelligent processing method based on a dynamic deep belief network, which belongs to the field of deep learning and solid waste intelligent processing. This method firstly proposes a DDBN using dynamic branch increase and decrease algorithm, which enables DDBN to increase hidden layer neurons and hidden layers according to the current training situation during the training process, and remove redundant neurons to effectively optimize the network structure of DDBN. Then, using DDBN to effectively extract the main features of the original data, use DDBN to effectively describe the state of the random, discrete, and nonlinear feature vectors of solid waste, so that the state features of the time series are easier to identify and ensure that the original data is not lost. The main information of the data. At the same time, according to the state description of the extracted solid waste, DDBN is used to predict the optimal combustion behavior suitable for its state, which reduces the waste of resources caused by blind combustion behavior and realizes the intelligent treatment of solid waste.

Figure 201810768405

Description

一种基于动态深度置信网络的固体废弃物智能处理方法An intelligent treatment method of solid waste based on dynamic deep belief network

技术领域technical field

本发明属于深度学习、固体废弃物智能处理领域,提出一种使用动态增减枝算法的动态深度置信网络(Dynamic Deep Belief Network,DDBN)模型,可以有效的优化深度置信网络的网络结构,以此来解决轻工行业大宗固体废弃物智能处理问题。The invention belongs to the fields of deep learning and intelligent processing of solid waste, and proposes a dynamic deep belief network (Dynamic Deep Belief Network, DDBN) model using a dynamic branch increase and decrement algorithm, which can effectively optimize the network structure of the deep belief network. To solve the problem of intelligent processing of bulk solid waste in the light industry.

背景技术Background technique

轻工行业的发展目前面临极大的环保压力和艰巨的污染减排处理任务要求。随着国民经济的发展,发酵和造纸产品的市场需求大幅增长,尽管行业近年来单位产品的污染排放强度明显降低,但由于生产能力的快速放大,行业固体废弃物排放总量仍在增大。为实现行业节能减排的有关目标,需要研究废弃物处理的新方法,通过新方法的应用提高生产企业的污染控制和治理水平,支撑行业的污染减排行动。The development of light industry is currently facing great pressure on environmental protection and arduous task requirements for pollution reduction and emission reduction. With the development of the national economy, the market demand for fermentation and papermaking products has grown substantially. Although the pollution emission intensity per unit product of the industry has been significantly reduced in recent years, the total amount of solid waste emissions in the industry is still increasing due to the rapid expansion of production capacity. In order to achieve the relevant goals of energy conservation and emission reduction in the industry, it is necessary to study new methods of waste disposal, improve the pollution control and governance level of production enterprises through the application of new methods, and support the pollution reduction actions of the industry.

近年来,深度学习迅速发展,Hinton等人于2006年提出了深度置信网络(DeepBelief Network,DBN)以及无监督贪婪逐层训练算法,解决了深度神经网络易陷入局部最优的问题,引发了深度学习在学术界的新浪潮。DBN通过多层次的特征变换得到对于原始数据的抽象表示,从而提高分类和预测等任务的准确性,因DBN具有自动学习特征和数据降维的优势,已经成为深度学习应用最广泛的网络结构。In recent years, deep learning has developed rapidly. Hinton et al. proposed the Deep Belief Network (DBN) and the unsupervised greedy layer-by-layer training algorithm in 2006, which solved the problem that deep neural networks are prone to fall into local optimality, causing deep Learn about the new wave of learning in academia. DBN obtains an abstract representation of the original data through multi-level feature transformation, thereby improving the accuracy of tasks such as classification and prediction. Because DBN has the advantages of automatic learning of features and data dimensionality reduction, it has become the most widely used deep learning network structure.

利用DBN可以有效的提取原始数据主要特征的优势,可以使用DBN对固体废弃物随机、离散、非线性的特征向量进行有效的状态描述,使时间序列的状态特征更加易于鉴别,并确保不丢失原始数据的主要信息。同时,根据提取到的固体废弃物的状态描述,利用DBN预测适合其状态的优化燃烧行为,大大地减少盲目燃烧行为对资源的浪费,实现对固体废弃物的智能处理。Taking advantage of the advantage that DBN can effectively extract the main features of the original data, DBN can be used to effectively describe the state of the random, discrete, and nonlinear feature vectors of solid waste, so that the state features of the time series can be more easily identified and ensure that the original data is not lost. The main information of the data. At the same time, according to the state description of the extracted solid waste, DBN is used to predict the optimal combustion behavior suitable for its state, which greatly reduces the waste of resources caused by blind combustion behavior and realizes the intelligent treatment of solid waste.

但是,如果DBN要解决一个高复杂度的问题,例如上述问题,那么DBN就需要适当地增加隐藏层神经元和隐藏层。但是,隐藏层神经元和隐藏层的数量目前仍需要通过人工实验来选择,并且在训练过程中网络结构是固定不变的。这样不仅误差较大,而且计算成本较高,效率较低。因此,需要提出一种新的DBN结构设计方法,使DBN在训练过程中可以根据当前训练情况动态增枝和减枝,优化网络结构,使其更好的解决固体废弃物智能处理问题。However, if the DBN is to solve a high-complexity problem, such as the above problem, then the DBN needs to increase the hidden layer neurons and hidden layers appropriately. However, the number of hidden layer neurons and hidden layers still needs to be selected by manual experiments, and the network structure is fixed during the training process. In this way, not only the error is larger, but also the calculation cost is higher and the efficiency is lower. Therefore, it is necessary to propose a new DBN structure design method, so that the DBN can dynamically increase and decrease branches according to the current training situation during the training process, optimize the network structure, and make it better solve the problem of intelligent disposal of solid waste.

发明内容SUMMARY OF THE INVENTION

为了解决现有技术存在的问题,本发明提出一种基于动态深度置信网络(DynamicDeep Belief Network,DDBN)的固体废弃物智能处理方法。In order to solve the problems existing in the prior art, the present invention proposes an intelligent processing method for solid waste based on a dynamic deep belief network (Dynamic Deep Belief Network, DDBN).

本发明的技术方案:Technical scheme of the present invention:

一种基于动态深度置信网络的固体废弃物智能处理方法,步骤如下:An intelligent processing method for solid waste based on dynamic deep belief network, the steps are as follows:

步骤1、对固体废弃物进行测量,得到固废数据集,对固废数据集进行预处理,并划分得到训练数据集和测试数据集。Step 1. Measure the solid waste to obtain a solid waste data set, preprocess the solid waste data set, and divide it into a training data set and a test data set.

所述的预处理为:把固废数据集归一化到[0,1]之间,归一化公式为:The preprocessing is: normalize the solid waste data set to [0, 1], and the normalization formula is:

Figure BDA0001729614950000021
Figure BDA0001729614950000021

其中,

Figure BDA0001729614950000022
为固废数据集的特征值,xmax和xmin分别为固废数据集所有特征的最大值和最小值,x是经过归一化后的固废数据集。in,
Figure BDA0001729614950000022
is the feature value of the solid waste data set, x max and x min are the maximum and minimum values of all the features of the solid waste data set, respectively, and x is the normalized solid waste data set.

步骤2、将步骤1预处理后得到的训练数据集输入到DDBN模型中,使用对比散度算法(Contrastive Divergence,CD)无监督地自底向上单独训练每一层受限玻尔兹曼机(Restricted Boltzmann Machine,RBM),并且在训练过程中,通过动态增减枝算法,优化当前RBM的网络结构,通过迭代训练,得到每个RBM的网络结构和参数值,并最终得到训练数据集的高层次特征;所述参数值即权值和偏置。具体操作如下:Step 2. Input the training data set obtained after preprocessing in Step 1 into the DDBN model, and use the Contrastive Divergence (CD) algorithm to train each layer of restricted Boltzmann machine ( Restricted Boltzmann Machine, RBM), and in the training process, the network structure of the current RBM is optimized through the dynamic branch increase and decrease algorithm, and the network structure and parameter values of each RBM are obtained through iterative training, and finally the high value of the training data set is obtained. Hierarchical features; the parameter values are weights and biases. The specific operations are as follows:

步骤2.1,构建DDBN网络模型,设置DDBN的各参数值:可视层神经元、初始隐藏层神经元和隐藏层层数、学习率、迭代次数和微调次数。其中,可视层神经元个数为训练数据集的特征维数。Step 2.1, build a DDBN network model, and set the parameter values of DDBN: visual layer neurons, initial hidden layer neurons and hidden layer layers, learning rate, number of iterations and number of fine-tuning. Among them, the number of neurons in the visual layer is the feature dimension of the training data set.

步骤2.2,将预处理后得到的训练数据集输入到第一层RBM中,使用CD算法对RBM进行预训练,并且在训练过程中,通过动态增减枝算法,优化当前RBM的网络结构。In step 2.2, the training data set obtained after preprocessing is input into the first layer of RBM, and the CD algorithm is used to pre-train the RBM, and during the training process, the network structure of the current RBM is optimized through the dynamic branch increase and decrease algorithm.

(1)RBM的能量函数E(v,h;θ)和可视层和隐藏层神经元的联合概率分布P(v,h;θ)为:(1) The energy function E(v, h; θ) of the RBM and the joint probability distribution P(v, h; θ) of the neurons in the visible layer and the hidden layer are:

Figure BDA0001729614950000023
Figure BDA0001729614950000023

Figure BDA0001729614950000024
Figure BDA0001729614950000024

其中,vi(1≤i≤I)和hj(1≤j≤J)分别表示可视层神经元和隐藏层神经元,w是可视层和隐藏层之间的权重矩阵,b和c分别是可视层神经元和隐藏层神经元的偏置,θ表示模型中的参数,即θ={w,b,c};Z是对所有可能的可视层和隐藏层神经元对求和。Among them, v i (1≤i≤I) and h j (1≤j≤J) represent the neurons in the visible layer and the neurons in the hidden layer, respectively, w is the weight matrix between the visible layer and the hidden layer, b and c are the biases of the visual layer neurons and the hidden layer neurons, respectively, θ represents the parameters in the model, that is, θ={w,b,c}; Z is the pair of all possible visual layer and hidden layer neurons beg for peace.

利用贝叶斯公式的原理,根据公式(3)求出可视层神经元v和隐藏层神经元h的边缘概率分布:Using the principle of Bayesian formula, the marginal probability distribution of the visible layer neuron v and the hidden layer neuron h is obtained according to formula (3):

Figure BDA0001729614950000025
Figure BDA0001729614950000025

Figure BDA0001729614950000026
Figure BDA0001729614950000026

利用贝叶斯公式,推导出可视层神经元v和隐藏层神经元h的条件概率分布:Using the Bayesian formula, the conditional probability distribution of the visible layer neuron v and the hidden layer neuron h is derived:

Figure BDA0001729614950000031
Figure BDA0001729614950000031

Figure BDA0001729614950000032
Figure BDA0001729614950000032

利用公式(6)和公式(7),使用对比散度算法经过一步吉布斯采样得到训练样本的近似重构P(v;θ),然后根据重构误差更新网络参数θ={w,b,c}。Using formula (6) and formula (7), the approximate reconstruction P(v; θ) of the training sample is obtained by one-step Gibbs sampling using the contrastive divergence algorithm, and then the network parameter θ={w,b is updated according to the reconstruction error ,c}.

(2)在训练过程中,根据当前训练情况,通过动态增减枝算法,优化RBM的网络结构。(2) During the training process, according to the current training situation, the network structure of the RBM is optimized through the dynamic branch increase and decrease algorithm.

利用WD(Weight Distance,权重距离)方法来监测权重w的变化:Use the WD (Weight Distance, weight distance) method to monitor the change of the weight w:

WDj[m]=Met(wj[m],wj[m-1]) (8)WD j [m]=Met(w j [m],w j [m-1]) (8)

其中,wj[m]是隐藏层神经元j在经过m次迭代后的权重向量,Met表示度量函数,如欧氏距离。WD的值反映了两次迭代中隐藏层神经元j的权重向量的变化。Among them, w j [m] is the weight vector of hidden layer neuron j after m iterations, and Met represents a metric function, such as Euclidean distance. The value of WD reflects the change in the weight vector of hidden layer neuron j in two iterations.

从局部和全局两个方面考虑增枝条件。The branching conditions are considered from both local and global aspects.

局部条件定义为:The local condition is defined as:

Figure BDA0001729614950000033
Figure BDA0001729614950000033

其中

Figure BDA0001729614950000034
是在第m次迭代中,第j个隐藏神经元对第n个输入样本的WD值,j=1,2,3,..,J,J是隐藏神经元的个数,max(·)是最大值函数。in
Figure BDA0001729614950000034
is the WD value of the jth hidden neuron for the nth input sample in the mth iteration, j=1,2,3,..,J, J is the number of hidden neurons, max( ) is the maximum function.

全局条件定义为:Global conditions are defined as:

Figure BDA0001729614950000035
Figure BDA0001729614950000035

其中N是训练数据集中样本的个数,N'是与上一次迭代相比使第j个神经元的WD值增大的样本个数,即

Figure BDA0001729614950000036
where N is the number of samples in the training dataset, and N' is the number of samples that increase the WD value of the jth neuron compared to the previous iteration, namely
Figure BDA0001729614950000036

局部和全局条件分别是隐藏层神经元j对单个输入样本和所有输入样本考虑的。将这两个条件相乘即得到增枝条件:The local and global conditions are considered by the hidden layer neuron j for a single input sample and all input samples, respectively. Multiplying these two conditions gives the branching condition:

max_WDj[m]*iratioj[m]>y(m) (11)max_WD j [m]*iratio j [m]>y(m) (11)

其中y(m)是一条曲线,用来作为可变阈值,其定义为:where y(m) is a curve used as a variable threshold, which is defined as:

Figure BDA0001729614950000037
Figure BDA0001729614950000037

其中m是当前迭代次数,numepoches为最大迭代次数,u表示曲线的曲率,ymax和ymin分别是曲线的最大值和最小值。当第j个神经元满足公式(11)时,则该神经元将被分成两个神经元,并且新神经元的各参数都为0。where m is the current number of iterations, numepoches is the maximum number of iterations, u is the curvature of the curve, and y max and y min are the maximum and minimum values of the curve, respectively. When the jth neuron satisfies the formula (11), the neuron will be divided into two neurons, and each parameter of the new neuron is 0.

当RBM训练完成后,开始减枝:使用隐藏层神经元对所有样本激活概率的标准差作为减枝条件,标准差公式为:When the RBM training is completed, start pruning: use the standard deviation of the activation probability of the hidden layer neurons for all samples as the pruning condition. The standard deviation formula is:

Figure BDA0001729614950000041
Figure BDA0001729614950000041

其中n=1,2,3,…,N,N是输入训练数据集中样本的个数,j表示第j个隐藏层神经元,P(n,j)表示第j个神经元对第n个输入样本的激活概率,μj表示第j个神经元对所有输入样本的平均激活概率。where n=1,2,3,...,N,N is the number of samples in the input training data set, j represents the jth hidden layer neuron, P(n,j) represents the jth neuron pair nth The activation probability of the input sample, μ j represents the average activation probability of the jth neuron for all input samples.

减枝条件为:The pruning conditions are:

σ(j)<θA (14)σ(j) < θ A (14)

其中θA是一个阈值。当第j个神经元满足公式(14)时,则移除该神经元及其所有参数。同时,作一条关于减枝率与预测准确率之间的权衡曲线,根据此曲线选择θA的值,使移除更多的冗余神经元的同时保留原始准确率。where θ A is a threshold. When the jth neuron satisfies the formula (14), the neuron and all its parameters are removed. At the same time, a trade-off curve between the pruning rate and the prediction accuracy is made, and the value of θ A is selected according to this curve, so that more redundant neurons are removed while retaining the original accuracy.

在减枝后,重新训练RBM,使剩余的神经元能够补偿被移除的神经元,减枝后再重训练为一次迭代。每次迭代我们都更新阈值θAAfter pruning, the RBM is retrained so that the remaining neurons can compensate for the removed neurons, and then retrained for one iteration after pruning. Each iteration we update the threshold θ A :

θA←θA+δ[iter] (15)θ A ←θ A +δ[iter] (15)

通过δ[iter]来更新每次迭代减枝中的阈值以移除更多的神经元。每次迭代都是一次贪心搜索,根据每次减枝中的权衡曲线,可以在不损失准确率的情况下找到最佳减枝率,因此δ[iter]被设置为使θA满足此次迭代减枝所需的减枝率。The threshold in each iteration of pruning is updated by δ[iter] to remove more neurons. Each iteration is a greedy search, and according to the trade-off curve in each pruning, the optimal pruning rate can be found without loss of accuracy, so δ[iter] is set so that θ A satisfies this iteration The pruning rate required for pruning.

步骤2.3,当前RBM确定网络结构后,使用能量函数作为增加新RBM的条件:Step 2.3, after the current RBM determines the network structure, use the energy function as the condition for adding a new RBM:

Figure BDA0001729614950000042
Figure BDA0001729614950000042

其中El是第l层RBM的总能量,由公式(2)求得,l=1,2…L,L是DDBN当前的层数,mean(·)是平均函数,θL是一个阈值。当能量函数满足公式(16)时,则增加一层新的RBM,并且新RBM各参数的初始化与第一层相同。之后把当前RBM的输出作为新增加的RBM的输入。where E l is the total energy of the lth layer of RBM, obtained by formula (2), l=1, 2...L, L is the current number of layers of DDBN, mean( ) is the average function, and θ L is a threshold. When the energy function satisfies the formula (16), a new layer of RBM is added, and the initialization of each parameter of the new RBM is the same as that of the first layer. Then use the output of the current RBM as the input of the newly added RBM.

步骤2.4,按照步骤2.2和2.3循环训练网络,得到DDBN的网络结构。Step 2.4, follow steps 2.2 and 2.3 to train the network cyclically to obtain the network structure of DDBN.

步骤3、将步骤2得到的DDBN网络结构和参数值作为微调阶段的初始值,使用自顶向下的反向传播算法微调整个DDBN网络,得到最终的DDBN网络模型。具体操作如下:Step 3. Use the DDBN network structure and parameter values obtained in step 2 as the initial values of the fine-tuning stage, and use the top-down backpropagation algorithm to fine-tune the entire DDBN network to obtain the final DDBN network model. The specific operations are as follows:

步骤3.1,将步骤2中训练好的的DDBN网络结构和参数值θ作为微调阶段的初始值,并在最后一层RBM后增加一层输出层,用来预测适合该训练数据集样本的燃烧行为,包括温度、压强和燃气流量,输入训练数据集开始微调整个DDBN网络。Step 3.1, take the DDBN network structure and parameter value θ trained in step 2 as the initial value of the fine-tuning stage, and add an output layer after the last layer of RBM to predict the combustion behavior suitable for the training data set samples , including temperature, pressure, and gas flow, input the training dataset to start fine-tuning the entire DDBN network.

步骤3.2,利用前向传播算法计算每个隐藏层神经元的激活概率。Step 3.2, use the forward propagation algorithm to calculate the activation probability of each hidden layer neuron.

步骤3.3,计算出训练样本前向传播得到的预测结果,与实际结果作对比得到损失函数:Step 3.3, calculate the prediction result obtained by the forward propagation of the training sample, and compare it with the actual result to obtain the loss function:

Figure BDA0001729614950000051
Figure BDA0001729614950000051

其中,t是当前微调次数,N是训练数据集中样本的个数,yn和y′n分别为第n个训练样本的实际结果和预测结果。将实际结果和预测结果的误差反向传播,使用梯度下降法按照公式(18)、(19)对权重w和偏置c进行更新:Among them, t is the current number of fine-tuning, N is the number of samples in the training data set, and y n and y' n are the actual and predicted results of the nth training sample, respectively. The error of the actual result and the predicted result is back-propagated, and the weight w and the bias c are updated according to the formulas (18) and (19) using the gradient descent method:

Figure BDA0001729614950000052
Figure BDA0001729614950000052

Figure BDA0001729614950000053
Figure BDA0001729614950000053

其中,α是学习率。where α is the learning rate.

迭代使用梯度下降法,自顶向下的微调整个DDBN网络,来减小J(t)的值,直到达到最大微调次数获得最终的DDBN网络模型。Iteratively uses the gradient descent method to fine-tune the entire DDBN network top-down to reduce the value of J(t) until the maximum number of fine-tuning times is reached to obtain the final DDBN network model.

步骤4、将测试数据集输入到步骤3所得最终的DDBN网络模型中,最后输出预测结果。具体操作如下:Step 4. Input the test data set into the final DDBN network model obtained in step 3, and finally output the prediction result. The specific operations are as follows:

步骤4.1,将经过预处理的测试数据集输入到步骤3微调好的DDBN网络模型中,通过RBM提取固体废弃物的主要特征。Step 4.1, input the preprocessed test data set into the fine-tuned DDBN network model in step 3, and extract the main features of solid waste through RBM.

步骤4.2,将测试样本的主要特征输入到最后一层输出层中,预测出适合该训练样本的燃烧行为,包括温度、压强和燃气流量。Step 4.2, input the main features of the test sample into the last output layer, and predict the combustion behavior suitable for the training sample, including temperature, pressure and gas flow.

本发明的有益效果:为了使模型更具有提取特征和预测能力,提出一种使用动态增减枝算法的DDBN模型,使DDBN在训练过程中可以根据当前训练情况更改网络结构,包括增加隐藏层神经元和隐藏层、移除冗余神经元,可以有效的优化DDBN的网络结构,取代了人工实验,克服了网络结构设计的困难。然后,利用DDBN可以有效的提取原始数据主要特征的优势,使用DDBN对固体废弃物随机、离散、非线性的特征向量进行有效的状态描述,使时间序列的状态特征更加易于鉴别,并确保不丢失原始数据的主要信息。同时,根据提取到的固体废弃物的状态描述,利用DDBN预测适合其状态的优化燃烧行为,包括温度、压强和燃气流量,大大地减少盲目燃烧行为对资源的浪费,实现对固体废弃物的智能处理。Beneficial effects of the present invention: In order to make the model more capable of extracting features and predicting, a DDBN model using a dynamic branch increase and decrease algorithm is proposed, so that the DDBN can change the network structure according to the current training situation during the training process, including adding hidden layer neurons. Elements and hidden layers and removing redundant neurons can effectively optimize the network structure of DDBN, replacing manual experiments and overcoming the difficulty of network structure design. Then, using DDBN to effectively extract the main features of the original data, use DDBN to effectively describe the state of the random, discrete, and nonlinear feature vectors of solid waste, so that the state features of the time series are easier to identify and ensure that they are not lost. The main information of the raw data. At the same time, according to the state description of the extracted solid waste, DDBN is used to predict the optimal combustion behavior suitable for its state, including temperature, pressure and gas flow, which greatly reduces the waste of resources caused by blind combustion behavior and realizes the intelligentization of solid waste. deal with.

附图说明Description of drawings

图1为本发明中增加隐藏层神经元操作示意图;1 is a schematic diagram of the operation of adding hidden layer neurons in the present invention;

图2为本发明中移除冗余神经元操作示意图;2 is a schematic diagram of the operation of removing redundant neurons in the present invention;

图3为本发明中增加隐藏层操作示意图;3 is a schematic diagram of an operation of adding a hidden layer in the present invention;

图4为本发明中DDBN模型的训练过程流程图;Fig. 4 is the training process flow chart of DDBN model in the present invention;

图5为本发明中固体废弃物经智能处理后的效果图;Fig. 5 is the effect diagram after the solid waste is processed intelligently in the present invention;

具体实施方式Detailed ways

以下结合附图和技术方案,进一步说明本发明的具体实施方式。The specific embodiments of the present invention will be further described below with reference to the accompanying drawings and technical solutions.

如图4所示,一种基于动态深度置信网络的固体废弃物智能处理方法,具体步骤如下:As shown in Figure 4, a solid waste intelligent processing method based on dynamic deep belief network, the specific steps are as follows:

步骤1、对固体废弃物的GDP、危险物、固体废弃物量、冶炼废渣、炉煤灰、炉渣、尾渣等数据进行测量,得到固废数据集,并进行数据预处理,得到训练和测试数据集。Step 1. Measure the GDP of solid waste, hazardous substances, solid waste volume, smelting waste residue, coal ash, slag, tailings and other data to obtain a solid waste data set, and perform data preprocessing to obtain training and test data set.

由于固体废弃物的各个数据往往不在同一个数量级,因此需要把固废数据集归一化到[0,1]之间,这样有利于提高网络的训练速度。归一化公式为:Since the data of solid waste are often not in the same order of magnitude, it is necessary to normalize the solid waste data set to [0, 1], which is beneficial to improve the training speed of the network. The normalization formula is:

Figure BDA0001729614950000061
Figure BDA0001729614950000061

其中,

Figure BDA0001729614950000062
为固废数据集的特征值,xmax和xmin分别为固废数据集所有特征的最大值和最小值,x是经过归一化后的固废数据集。in,
Figure BDA0001729614950000062
is the feature value of the solid waste data set, x max and x min are the maximum and minimum values of all the features of the solid waste data set, respectively, and x is the normalized solid waste data set.

步骤2、将预处理后得到的训练数据集输入到DDBN模型中进行预训练,利用对比散度算法(Contrastive Divergence,CD)无监督地自底向上单独训练每一层受限玻尔兹曼机(Restricted Boltzmann Machine,RBM)。并且在训练过程中,通过动态增减枝算法,优化当前RBM的网络结构,包括增加新的隐藏层神经元和移除冗余神经元。当前RBM训练结束后,若DDBN满足层数生成条件,则增加一层新的RBM,并把当前RBM的输出作为新增加的RBM的输入,并且新RBM各参数的初始化与第一层相同。通过多次迭代,得到每个RBM的网络结构以及相应的权重和偏置,最终得到数据的高层次特征。具体操作如下:Step 2. Input the training data set obtained after preprocessing into the DDBN model for pre-training, and use the Contrastive Divergence (CD) algorithm to train each layer of restricted Boltzmann machines independently from the bottom to the top. (Restricted Boltzmann Machine, RBM). And in the training process, the network structure of the current RBM is optimized through the dynamic branch increase and decrease algorithm, including adding new hidden layer neurons and removing redundant neurons. After the current RBM training is completed, if the DDBN meets the layer generation conditions, a new RBM is added, and the output of the current RBM is used as the input of the newly added RBM, and the initialization of the parameters of the new RBM is the same as the first layer. Through multiple iterations, the network structure and corresponding weights and biases of each RBM are obtained, and finally the high-level features of the data are obtained. The specific operations are as follows:

步骤2.1,构建DDBN网络模型,设置DDBN的各参数值:可视层神经元个数为训练数据集的特征维数,初始隐藏层神经元和隐藏层层数分别设置为10和1,学习率设置为0.1,预训练迭代次数为100,微调次数为100。Step 2.1, build a DDBN network model, and set the parameter values of DDBN: the number of neurons in the visual layer is the feature dimension of the training data set, the initial number of neurons in the hidden layer and the number of hidden layers are set to 10 and 1, respectively, and the learning rate Set to 0.1, the number of pre-training iterations is 100, and the number of fine-tuning is 100.

步骤2.2,将预处理后得到的训练数据集作为第一层RBM的输入,使用CD算法对RBM进行预训练,并且在训练过程中,通过动态增减枝算法,优化当前RBM的网络结构,包括增加新的隐藏层神经元和移除冗余神经元。In step 2.2, the training data set obtained after preprocessing is used as the input of the first-layer RBM, and the CD algorithm is used to pre-train the RBM, and during the training process, the network structure of the current RBM is optimized through the dynamic branch increase and decrease algorithm, including Add new hidden layer neurons and remove redundant neurons.

(1)RBM是一种基于能量模型的随机网络,每一组可视层神经元v和隐藏层神经元h的取值都有一个对应的能量值。根据能量函数的定义及热力学统计的原理,可以得到v和h的联合概率分布。RBM的能量函数E(v,h;θ)和v和h的联合概率分布P(v,h;θ)为:(1) RBM is a random network based on an energy model. Each group of neurons in the visible layer v and neurons in the hidden layer has a corresponding energy value. According to the definition of energy function and the principle of thermodynamic statistics, the joint probability distribution of v and h can be obtained. The energy function E(v, h; θ) of the RBM and the joint probability distribution P(v, h; θ) of v and h are:

Figure BDA0001729614950000071
Figure BDA0001729614950000071

Figure BDA0001729614950000072
Figure BDA0001729614950000072

其中,vi(1≤i≤I)和hj(1≤j≤J)分别表示可视层神经元和隐藏层神经元,w是可视层和隐藏层之间的权重矩阵,b和c分别是可视层神经元和隐藏层神经元的偏置,θ表示模型中的参数,即θ={w,b,c},Z是对所有可能的可视层和隐藏层神经元对求和。Among them, v i (1≤i≤I) and h j (1≤j≤J) represent the neurons in the visible layer and the neurons in the hidden layer, respectively, w is the weight matrix between the visible layer and the hidden layer, b and c are the biases of the neurons in the visible layer and the neurons in the hidden layer, respectively, θ represents the parameters in the model, ie θ={w,b,c}, and Z is the pair of neurons in the visible layer and the hidden layer for all possible pairs beg for peace.

利用贝叶斯公式的原理,根据公式(3)可以求出可视层神经元v和隐藏层神经元h的边缘概率分布:Using the principle of Bayesian formula, the marginal probability distribution of the visible layer neuron v and the hidden layer neuron h can be obtained according to formula (3):

Figure BDA0001729614950000073
Figure BDA0001729614950000073

Figure BDA0001729614950000074
Figure BDA0001729614950000074

RBM网络训练的目标就是求解θ={w,b,c},使得在该参数下RBM能够极大地拟合输入样本,使得P(v;θ)最大,即求解输入样本的极大似然估计。然而,为了获得最大似然估计,需要计算所有可能情况,计算量是指数增长的,因此RBM使用对比散度算法进行估算。The goal of RBM network training is to solve θ={w,b,c}, so that RBM can fit the input sample greatly under this parameter, so that P(v; θ) is the largest, that is, to solve the maximum likelihood estimation of the input sample . However, in order to obtain a maximum likelihood estimate, all possible cases need to be calculated, and the amount of computation grows exponentially, so RBM uses a contrastive divergence algorithm for estimation.

利用贝叶斯公式,推导出可视层神经元v和隐藏层神经元h的条件概率分布:Using the Bayesian formula, the conditional probability distribution of the visible layer neuron v and the hidden layer neuron h is derived:

Figure BDA0001729614950000075
Figure BDA0001729614950000075

Figure BDA0001729614950000076
Figure BDA0001729614950000076

根据公式(6)和公式(7),使用对比散度算法经过一步吉布斯采样得到训练样本的近似重构P(v;θ),然后根据重构误差更新网络参数θ={w,b,c}。According to formula (6) and formula (7), the approximate reconstruction P(v; θ) of the training sample is obtained through one-step Gibbs sampling using the contrastive divergence algorithm, and then the network parameter θ={w,b is updated according to the reconstruction error ,c}.

(2)在训练过程中,根据当前训练情况,通过动态增减枝算法,优化RBM的网络结构。(2) During the training process, according to the current training situation, the network structure of the RBM is optimized through the dynamic branch increase and decrease algorithm.

在DBN中,权重w在网络训练中起着决定性的作用。因此,本发明提出了一种称为WD(Weight Distance,权重距离)的方法来监测权重w的变化:In DBN, the weight w plays a decisive role in network training. Therefore, the present invention proposes a method called WD (Weight Distance, weight distance) to monitor the change of the weight w:

WDj[m]=Met(wj[m],wj[m-1]) (8)WD j [m]=Met(w j [m],w j [m-1]) (8)

其中,wj[m]是隐藏层神经元j在经过m次迭代后的权重向量,Met是一个度量函数,如欧氏距离。WD的值反映了两次迭代中隐藏层神经元j的权重向量的变化。一般而言,神经元j的权重向量在训练一段时间后会收敛,也就是说,WD的值会越来越小。如果经过长时间迭代后,某些权重向量波动幅度较大,应该考虑到这是缺少隐藏层神经元来映射输入样本导致的。在这种情况下,需要增加神经元的数量来提高网络的性能。本发明从局部和全局两个方面考虑增枝条件。局部条件定义为:where w j [m] is the weight vector of hidden layer neuron j after m iterations, and Met is a metric function such as Euclidean distance. The value of WD reflects the change in the weight vector of hidden layer neuron j in two iterations. In general, the weight vector of neuron j will converge after a period of training, that is, the value of WD will become smaller and smaller. If some weight vectors fluctuate greatly after a long iteration, it should be considered that this is caused by the lack of hidden layer neurons to map input samples. In this case, the number of neurons needs to be increased to improve the performance of the network. The present invention considers branching conditions from both local and global aspects. The local condition is defined as:

Figure BDA0001729614950000081
Figure BDA0001729614950000081

其中

Figure BDA0001729614950000082
是在第m次迭代中,第j个隐藏神经元对第n个输入样本的WD值,j=1,2,3,..,J,J是隐藏神经元的个数,max(·)是最大值函数。全局条件定义为:in
Figure BDA0001729614950000082
is the WD value of the jth hidden neuron for the nth input sample in the mth iteration, j=1,2,3,..,J, J is the number of hidden neurons, max( ) is the maximum function. Global conditions are defined as:

Figure BDA0001729614950000083
Figure BDA0001729614950000083

其中N是训练样本的个数,N'是与上一次迭代相比使第j个神经元的WD值增大的样本个数,即

Figure BDA0001729614950000084
where N is the number of training samples, and N' is the number of samples that increase the WD value of the jth neuron compared with the previous iteration, that is,
Figure BDA0001729614950000084

局部和全局条件分别是隐藏层神经元j对单个输入样本和所有输入样本考虑的。找到神经元j对某个样本的最大WD值max_WDj[m]和使WD值增大的样本比率iratioj(m)。然后将这两个条件相乘即得到增枝条件:The local and global conditions are considered by the hidden layer neuron j for a single input sample and all input samples, respectively. Find the maximum WD value max_WD j [m] of neuron j for a certain sample and the sample ratio iratio j (m) that increases the WD value. Then multiply these two conditions to get the branching condition:

max_WDj[m]*iratioj[m]>y(m) (11)max_WD j [m]*iratio j [m]>y(m) (11)

其中y(m)是一条曲线,用来作为可变阈值,其定义为:where y(m) is a curve used as a variable threshold, which is defined as:

Figure BDA0001729614950000085
Figure BDA0001729614950000085

其中m是当前迭代次数,numepoches为最大迭代次数,u表示曲线的曲率,ymax和ymin分别是曲线的最大值和最小值。在训练过程中,如果网络向好的方向发展,则max_WD和iratio的值会越来越小,所以使用一条曲线y(m)作为增枝条件的可变阈值,并且当h>0时,y(m)是一条单调递减的曲线。如果第j个神经元满足公式(11),则该神经元将被分成两个神经元,并且新神经元的各参数都为0。where m is the current number of iterations, numepoches is the maximum number of iterations, u is the curvature of the curve, and y max and y min are the maximum and minimum values of the curve, respectively. During the training process, if the network develops in a good direction, the values of max_WD and iratio will become smaller and smaller, so a curve y(m) is used as the variable threshold of the branching condition, and when h>0, y (m) is a monotonically decreasing curve. If the jth neuron satisfies formula (11), the neuron will be divided into two neurons, and each parameter of the new neuron will be 0.

当RBM训练完成后,开始减枝。RBM的目的是提取输入样本的主要特征,即隐藏层神经元的激活概率。这些特征都是有鉴别性的,方便对数据做进一步的应用研究。如果某个神经元的激活概率对所有样本都接近平均值,则说明该神经元提取的特征不具有鉴别性,即冗余神经元。为了得到一个更加简洁紧凑的网络结构,需要移除这些冗余神经元。本发明使用标准差测量同一隐藏层神经元对所有样本激活概率的离散程度,标准差公式为:When the RBM training is complete, start pruning. The purpose of RBM is to extract the main feature of the input sample, that is, the activation probability of the hidden layer neurons. These features are discriminative and facilitate further applied research on the data. If the activation probability of a neuron is close to the average value for all samples, it means that the features extracted by this neuron are not discriminative, that is, redundant neurons. In order to obtain a more concise and compact network structure, these redundant neurons need to be removed. The present invention uses the standard deviation to measure the dispersion degree of the activation probability of the same hidden layer neuron to all samples, and the standard deviation formula is:

Figure BDA0001729614950000091
Figure BDA0001729614950000091

其中n=1,2,3,…,N,N是输入样本的个数,j表示第j个隐藏层神经元,P(n,j)表示第j个神经元对第n个输入样本的激活概率,μj表示第j个神经元对所有输入样本的平均激活概率。一个较小的标准差意味着这些值接近平均值,即这个神经元提取的特征不具有鉴别性,因此需要移除这个冗余神经元。减枝条件为:Where n=1,2,3,...,N,N is the number of input samples, j represents the jth hidden layer neuron, P(n,j) represents the jth neuron's response to the nth input sample The activation probability, μ j represents the average activation probability of the jth neuron for all input samples. A small standard deviation means that the values are close to the mean, i.e. the features extracted by this neuron are not discriminative, so this redundant neuron needs to be removed. The pruning conditions are:

σ(j)<θA (14)σ(j) < θ A (14)

其中θA是一个阈值。如果第j个神经元满足公式(14),则移除该神经元及其所有参数。同时,本发明做了一条关于减枝率与预测准确率之间的权衡曲线,根据此曲线选择θA的值,使移除更多的冗余神经元的同时保留原始准确率。此外,在减枝后,重新训练当前的RBM使剩余的神经元能够补偿被移除的神经元。这一步至关重要,减枝后再重训练为一次迭代。迭代减枝每次都移除较少的神经元,并且进行了多次重训练以进行补偿。经过多次这样的迭代,可以找到一个更高的减枝率并且不损失准确率。每次迭代都更新阈值θAwhere θ A is a threshold. If the jth neuron satisfies Equation (14), then remove that neuron and all its parameters. At the same time, the present invention makes a trade-off curve between the pruning rate and the prediction accuracy, and selects the value of θ A according to this curve, so as to remove more redundant neurons while retaining the original accuracy. Furthermore, after pruning, retraining the current RBM enables the remaining neurons to compensate for the removed neurons. This step is crucial, pruning and then retraining as one iteration. Iterative pruning removes fewer neurons each time, and retrains multiple times to compensate. After many such iterations, a higher pruning rate can be found without losing accuracy. The threshold θ A is updated at each iteration:

θA←θA+δ[iter] (15)θ A ←θ A +δ[iter] (15)

通过δ[iter]来更新每次迭代减枝中的阈值以移除更多的神经元。每次迭代都是一次贪心搜索,根据每次减枝中的权衡曲线,可以在不损失准确率的情况下找到最佳减枝率,因此δ[iter]被设置为使θA满足此次迭代减枝所需的减枝率。The threshold in each iteration of pruning is updated by δ[iter] to remove more neurons. Each iteration is a greedy search, and according to the trade-off curve in each pruning, the optimal pruning rate can be found without loss of accuracy, so δ[iter] is set so that θ A satisfies this iteration The pruning rate required for pruning.

步骤2.3,当前RBM确定网络结构后,开始考虑隐藏层的增长。根据公式(4)可以发现P(v;θ)与E(v,h;θ)成反比。所以如果想最大化P(v;θ),那么能量函数E(v,h;θ)应该尽可能的小。如果DBN的总能量大于一个阈值,则表明DBN缺乏数据表示能力,此时就需要增加一层新的RBM。因此本发明使用能量函数作为增加新RBM的条件:Step 2.3, after the current RBM determines the network structure, it starts to consider the growth of the hidden layer. According to formula (4), it can be found that P(v; θ) is inversely proportional to E(v, h; θ). So if you want to maximize P(v; θ), then the energy function E(v, h; θ) should be as small as possible. If the total energy of the DBN is greater than a threshold, it indicates that the DBN lacks the ability to represent data, and a new layer of RBM needs to be added at this time. Therefore, the present invention uses the energy function as a condition for adding new RBMs:

Figure BDA0001729614950000101
Figure BDA0001729614950000101

其中El是第l层RBM的总能量,由公式(2)求得,l=1,2…L,L是DDBN当前的层数,mean(·)是平均函数,θL是一个阈值。如果能量函数满足公式(16),则增加一层新的RBM,并且新RBM各参数的初始化与第一层相同,之后把当前RBM的输出作为新增加的RBM的输入。where E l is the total energy of the lth layer of RBM, obtained by formula (2), l=1, 2...L, L is the current number of layers of DDBN, mean( ) is the average function, and θ L is a threshold. If the energy function satisfies formula (16), a new layer of RBM is added, and the initialization of each parameter of the new RBM is the same as that of the first layer, and then the output of the current RBM is used as the input of the newly added RBM.

步骤2.4,按照步骤2.2和2.3循环训练网络,可以学习到一个深层次的DDBN网络结构。In step 2.4, according to steps 2.2 and 2.3, the network is cyclically trained, and a deep DDBN network structure can be learned.

步骤3、使用微调进一步优化DDBN。将预训练阶段得到的网络结构和参数值作为微调阶段的初始值,对整个DDBN网络进行微调。本发明使用反向传播算法微调整个网络,即将训练误差自顶向下反向传播,对网络进行优化,得到最终的DDBN网络模型。具体操作如下:Step 3. Use fine-tuning to further optimize the DDBN. The network structure and parameter values obtained in the pre-training stage are used as the initial values of the fine-tuning stage to fine-tune the entire DDBN network. The invention uses the back-propagation algorithm to fine-tune the entire network, that is, the training error is back-propagated from top to bottom, and the network is optimized to obtain the final DDBN network model. The specific operations are as follows:

步骤3.1,将预训练阶段训练好的的网络结构和参数值θ作为微调阶段的初始值,并在最后一层RBM后增加一层输出层。该输出层有3个神经元,输出分别代表温度、压强和燃气流量,用来预测适合该训练样本的燃烧行为。输入训练数据集到微调阶段的网络中进行优化。Step 3.1, take the network structure and parameter value θ trained in the pre-training stage as the initial value of the fine-tuning stage, and add an output layer after the last layer of RBM. The output layer has 3 neurons, the outputs represent temperature, pressure and gas flow, respectively, and are used to predict the combustion behavior suitable for the training sample. The input training data set is optimized into the network in the fine-tuning stage.

步骤3.2,利用前向传播算法计算每个隐藏层神经元的激活概率。Step 3.2, use the forward propagation algorithm to calculate the activation probability of each hidden layer neuron.

步骤3.3,计算出训练样本前向传播得到的预测结果,与实际结果作对比得到损失函数:Step 3.3, calculate the prediction result obtained by the forward propagation of the training sample, and compare it with the actual result to obtain the loss function:

Figure BDA0001729614950000102
Figure BDA0001729614950000102

其中,t是当前微调次数,N是训练样本的个数,yn和y′n分别为第n个训练样本的实际结果和预测结果。将实际结果和预测结果的误差反向传播,使用梯度下降法按照公式(18)、(19)对权重w和偏置c进行更新:Among them, t is the current number of fine-tuning, N is the number of training samples, y n and y' n are the actual and predicted results of the nth training sample, respectively. The error of the actual result and the predicted result is back-propagated, and the weight w and the bias c are updated according to the formulas (18) and (19) using the gradient descent method:

Figure BDA0001729614950000103
Figure BDA0001729614950000103

Figure BDA0001729614950000104
Figure BDA0001729614950000104

其中,α是学习率。迭代使用梯度下降法,自顶向下的微调整个DDBN网络,来减小J(t)的值,直到达到最大微调次数获得最终的DDBN网络模型。where α is the learning rate. Iteratively uses the gradient descent method to fine-tune the entire DDBN network top-down to reduce the value of J(t) until the maximum number of fine-tuning times is reached to obtain the final DDBN network model.

步骤4、将经过预处理的测试数据集输入到微调阶段得到的DDBN网络模型中,通过RBM提取测试样本的主要特征,然后将这些主要特征输入到最后一层输出层中,输出的值分别代表温度、压强和燃气流量的值,即预测出适合该训练样本的燃烧行为。Step 4. Input the preprocessed test data set into the DDBN network model obtained in the fine-tuning stage, extract the main features of the test samples through RBM, and then input these main features into the last output layer, and the output values represent The values of temperature, pressure and gas flow rate predict the combustion behavior suitable for the training sample.

下面通过本发明提供的方法对采集到的固废数据集进行检测。该数据集包括1000个样本,其中训练样本800个,测试样本200个。每个样本有7个特征,所以可视层神经元个数设置为7;每个样本有3个输出,即温度、压强和燃气流量,对应其燃烧行为。The collected solid waste data set is detected below by the method provided by the present invention. The dataset includes 1000 samples, including 800 training samples and 200 testing samples. Each sample has 7 features, so the number of neurons in the visual layer is set to 7; each sample has 3 outputs, namely temperature, pressure and gas flow, corresponding to its combustion behavior.

检测结果表明基于动态深度置信网络的固体废弃物智能处理方法比传统的人工控制方法节约了30%的处理时间,并且处理效果也达到了国家规定的固体废弃物处理指标。因此,本发明提出的方法可以有效的处理固体废弃物,并且节约时间和成本,实现高效的智能化处理。The test results show that the intelligent treatment method of solid waste based on dynamic deep confidence network saves 30% of the treatment time compared with the traditional manual control method, and the treatment effect also reaches the solid waste treatment index stipulated by the state. Therefore, the method proposed in the present invention can effectively treat solid waste, save time and cost, and realize efficient and intelligent treatment.

Claims (5)

1. A solid waste intelligent treatment method based on a dynamic deep belief network is characterized by comprising the following steps:
step 1, measuring solid waste to obtain a solid waste data set, preprocessing the solid waste data set, and dividing to obtain a training data set and a testing data set;
step 2, inputting the training data set obtained after the preprocessing in the step 1 into a DDBN model, using a contrast divergence algorithm to train each layer of restricted Boltzmann machine RBM from bottom to top independently without supervision, optimizing the network structure of the current RBM through a dynamic increase and decrease branch algorithm in the training process, obtaining the network structure and parameter values of each RBM through iterative training, and finally obtaining the high-level characteristics of the training data set; the parameter values are weight and bias;
step 3, taking the DDBN network structure and the parameter values obtained in the step 2 as initial values of a fine tuning stage, and fine tuning the whole DDBN network by using a top-down back propagation algorithm to obtain a final DDBN network model; the specific process is as follows:
step 3.1, taking the DDBN network structure and the parameter value theta trained in the step 2 as initial values of a fine tuning stage, adding an output layer after the last layer of RBM for predicting combustion behaviors including temperature, pressure and gas flow suitable for the training data set sample, and inputting the training data set to start fine tuning the whole DDBN network;
step 3.2, calculating the activation probability of each hidden layer neuron by using a forward propagation algorithm;
step 3.3, calculating a prediction result obtained by forward propagation of the training sample, and comparing the prediction result with an actual result to obtain a loss function:
Figure FDA0003148918320000011
where t is the current number of fine-tuning, N is the number of samples in the training dataset, ynAnd y'nRespectively obtaining an actual result and a predicted result of the nth training sample; and (3) reversely propagating errors of the actual result and the predicted result, and updating the weight w and the offset c according to the formulas (2) and (3) by using a gradient descent method:
Figure FDA0003148918320000012
Figure FDA0003148918320000013
wherein α is the learning rate;
iteratively using a gradient descent method to finely adjust the whole DDBN network from top to bottom to reduce the value of J (t) until the maximum fine adjustment times are reached to obtain a final DDBN network model;
and 4, inputting the test data set into the final DDBN network model obtained in the step 3, and finally outputting a prediction result.
2. The method for intelligently processing the solid waste based on the dynamic deep belief network as claimed in claim 1, wherein the preprocessing in the step 1 is: normalizing the solid waste data set to be between [0 and 1], wherein the normalization formula is as follows:
Figure FDA0003148918320000014
wherein,
Figure FDA0003148918320000015
characteristic value, x, of solid waste data setmaxAnd xminThe maximum value and the minimum value of all characteristics of the solid waste data set are respectively, and x is the solid waste data set after normalization.
3. The method for intelligently processing the solid waste based on the dynamic deep belief network as claimed in claim 1 or 2, wherein the specific process of the step 2 is as follows:
step 2.1, constructing a DDBN network model, and setting the parameter values of DDBN: visual layer neurons, initial hidden layer neurons, hidden layer numbers, learning rate, iteration times and fine tuning times; wherein, the number of neurons in the visual layer is the feature dimension of the training data set;
step 2.2, inputting the training data set obtained after preprocessing into a first layer RBM, pre-training the RBM by using a CD algorithm, and optimizing the network structure of the current RBM by using a dynamic branch-increasing and-decreasing algorithm in the training process;
(1) the energy function E (v, h; θ) of the RBM and the joint probability distribution P (v, h; θ) of the visible and hidden layer neurons are:
Figure FDA0003148918320000021
Figure FDA0003148918320000022
wherein v isi(1. ltoreq. I. ltoreq.I) and hj(J is more than or equal to 1 and less than or equal to J) respectively represents a visual layer neuron and a hidden layer neuron, w is a weight matrix between the visual layer and the hidden layer, b and c are the bias of the visual layer neuron and the hidden layer neuron respectively, and theta represents a parameter in the model, namely theta is { w, b, c }; z is for all possible visible and hidden layersSumming the channel pairs;
and (3) solving the edge probability distribution of the visual layer neuron v and the hidden layer neuron h according to a formula (6) by utilizing the principle of a Bayes formula:
Figure FDA0003148918320000023
Figure FDA0003148918320000024
and deducing the conditional probability distribution of the visual layer neuron v and the hidden layer neuron h by using a Bayesian formula:
Figure FDA0003148918320000025
Figure FDA0003148918320000026
obtaining approximate reconstruction P (v; theta) of a training sample by using a contrast divergence algorithm through one-step Gibbs sampling by using a formula (9) and a formula (10), and then updating a network parameter theta to be { w, b, c } according to a reconstruction error;
(2) in the training process, according to the current training condition, optimizing the network structure of the RBM through a dynamic branch increasing and decreasing algorithm;
the change in weight w is monitored using the weight distance WD method:
WDj[m]=Met(wj[m],wj[m-1]) (11)
wherein, wj[m]The weight vector of the hidden layer neuron j after m iterations, wherein Met represents a measurement function, such as Euclidean distance; the value of WD reflects the change in the weight vector of hidden layer neuron j in two iterations;
the branch increasing condition is considered from the aspects of local and global;
the local conditions are defined as:
Figure FDA0003148918320000027
wherein
Figure FDA0003148918320000028
In the mth iteration, the WD value of the jth hidden neuron to the nth input sample, J is 1,2,3, J is the number of hidden neurons, and max (·) is a maximum function;
the global condition is defined as:
Figure FDA0003148918320000031
where N is the number of samples in the training data set and N' is the number of samples that increased the WD value of the jth neuron compared to the last iteration, i.e., the number of samples
Figure FDA0003148918320000032
The local and global conditions are considered for a single input sample and all input samples by the hidden layer neuron j, respectively; multiplying the two conditions to obtain the propagation condition:
max_WDj[m]*iratioj(m)>y(m) (14)
where y (m) is a curve, which is used as a variable threshold, defined as:
Figure FDA0003148918320000033
where m is the current iteration number, numepochs is the maximum iteration number, u represents the curvature of the curve, ymaxAnd yminMaximum and minimum values of the curve, respectively; when the jth neuron satisfies equation (14), then the neuron will be divided into two neurons, and each parameter of the new neuron will beThe numbers are all 0;
when the RBM training is completed, branch reduction is started: and (3) using the standard deviation of the hidden layer neurons on the activation probability of all samples as a branch-reducing condition, wherein the standard deviation formula is as follows:
Figure FDA0003148918320000034
where N is 1,2,3, …, N is the number of samples in the input training dataset, j represents the jth hidden layer neuron, P (N, j) represents the activation probability of the jth neuron on the nth input sample, μjRepresenting the average activation probability of the jth neuron on all input samples;
the branch reducing conditions are as follows:
σ(j)<θA (17)
wherein theta isAIs a threshold value; when the jth neuron satisfies formula (17), removing the neuron and all parameters thereof; at the same time, a trade-off curve relating to the ratio of branch reduction and the prediction accuracy is made, and theta is selected according to the curveAA value of (a) such that the original accuracy is preserved while removing more redundant neurons;
after branch reduction, retraining the RBM to enable the remaining neurons to compensate the removed neurons, and retraining after branch reduction into one iteration; at each iteration we update the threshold θA
θA←θA+δ[iter] (18)
By delta [ iter ]]To update the threshold in each iteration of the pruning to remove more neurons; each iteration is a greedy search, and according to a balance curve in each branch reduction, the optimal branch reduction rate can be found without losing the accuracy rate, so that the delta [ iter ]]Is set so thatAThe branch reducing rate required by the iterative branch reducing is met;
step 2.3, after the current RBM determines the network structure, using an energy function as a condition for adding a new RBM:
Figure FDA0003148918320000041
wherein ElIs the total energy of the L-th layer RBM, which is found by equation (5), L is 1,2 … L, L is the current layer number of the DDBN, mean () is the average function, θLIs a threshold value; when the energy function meets the formula (19), a new layer of RBM is added, and the initialization of each parameter of the new RBM is the same as that of the first layer; then taking the output of the current RBM as the input of the newly added RBM;
and 2.4, training the network circularly according to the steps 2.2 and 2.3 to obtain the network structure of the DDBN.
4. The method for intelligently processing the solid waste based on the dynamic deep belief network as claimed in claim 1 or 2, wherein the specific process of the step 4 is as follows:
step 4.1, inputting the preprocessed test data set into the DDBN network model finely adjusted in the step 3, and extracting the main characteristics of the solid waste through RBM;
and 4.2, inputting the main characteristics of the test sample into the last output layer, and predicting the combustion behavior suitable for the training sample, including temperature, pressure and gas flow.
5. The method for intelligently processing the solid waste based on the dynamic deep belief network as claimed in claim 3, wherein the specific process of the step 4 is as follows:
step 4.1, inputting the preprocessed test data set into the DDBN network model finely adjusted in the step 3, and extracting the main characteristics of the solid waste through RBM;
and 4.2, inputting the main characteristics of the test sample into the last output layer, and predicting the combustion behavior suitable for the training sample, including temperature, pressure and gas flow.
CN201810768405.8A 2018-07-13 2018-07-13 Solid waste intelligent treatment method based on dynamic deep belief network Active CN109146007B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810768405.8A CN109146007B (en) 2018-07-13 2018-07-13 Solid waste intelligent treatment method based on dynamic deep belief network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810768405.8A CN109146007B (en) 2018-07-13 2018-07-13 Solid waste intelligent treatment method based on dynamic deep belief network

Publications (2)

Publication Number Publication Date
CN109146007A CN109146007A (en) 2019-01-04
CN109146007B true CN109146007B (en) 2021-08-27

Family

ID=64800535

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810768405.8A Active CN109146007B (en) 2018-07-13 2018-07-13 Solid waste intelligent treatment method based on dynamic deep belief network

Country Status (1)

Country Link
CN (1) CN109146007B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110108672B (en) * 2019-04-12 2021-07-06 南京信息工程大学 An inversion method of aerosol extinction coefficient based on deep belief network
CN111366123B (en) * 2020-03-06 2021-03-26 大连理工大学 Part surface roughness and cutter wear prediction method based on multi-task learning

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107269335A (en) * 2017-07-28 2017-10-20 浙江大学 The rubbish and gas combustion-gas vapor combined cycle system of a kind of use combustion gas garbage drying
CN107729988A (en) * 2017-09-30 2018-02-23 北京工商大学 Cyanobacterial bloom prediction method based on dynamic deep belief network

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107247260B (en) * 2017-07-06 2019-12-03 合肥工业大学 A kind of RFID localization method based on adaptive depth confidence network
CN108197427B (en) * 2018-01-02 2020-09-04 山东师范大学 Protein subcellular localization method and device based on deep convolutional neural network

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107269335A (en) * 2017-07-28 2017-10-20 浙江大学 The rubbish and gas combustion-gas vapor combined cycle system of a kind of use combustion gas garbage drying
CN107729988A (en) * 2017-09-30 2018-02-23 北京工商大学 Cyanobacterial bloom prediction method based on dynamic deep belief network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
An adaptive learning method of Deep Belief Network by layer generation;Shin Kamada 等;《IEEE》;20171231;第2967-2970页 *
一种动态构建深度信念网络模型方法;吴强 等;《中国计量大学学报》;20180315;第64-69页 *

Also Published As

Publication number Publication date
CN109146007A (en) 2019-01-04

Similar Documents

Publication Publication Date Title
Xie et al. Graph neural network approach for anomaly detection
Ni et al. Streamflow forecasting using extreme gradient boosting model coupled with Gaussian mixture model
CN109034205B (en) Image classification method based on transductive semi-supervised deep learning
He et al. A novel ensemble method for credit scoring: Adaption of different imbalance ratios
CN114218872B (en) Remaining service life prediction method based on DBN-LSTM semi-supervised joint model
CN107729716B (en) Coal mine water inrush prediction method based on long-time and short-time memory neural network
CN102831269B (en) Method for determining technological parameters in flow industrial process
CN108445752B (en) An ensemble modeling method of random weight neural network for adaptive selection of deep features
CN106709640A (en) Method for predicting ship energy consumption based on deep learning and support vector regression
CN110580496A (en) A deep transfer learning system and method based on entropy minimization
Yao et al. Pre-training the deep generative models with adaptive hyperparameter optimization
CN110045356B (en) A bidirectional long-short-term memory neural network water surface target recognition method
Xiang et al. A concise self-adapting deep learning network for machine remaining useful life prediction
CN115828754A (en) Cutter wear state monitoring method based on multi-scale space-time fusion network model
CN111985825A (en) A method for evaluating crystal plane quality for tumbler orientator
CN109146007B (en) Solid waste intelligent treatment method based on dynamic deep belief network
CN105913078A (en) Multi-mode soft measurement method for improving adaptive affine propagation clustering
US20210216872A1 (en) System and method of training a neural network
Jamaleddyn et al. An improved approach to Arabic news classification based on hyperparameter tuning of machine learning algorithms
CN108062566A (en) A kind of intelligent integrated flexible measurement method based on the potential feature extraction of multinuclear
CN113052261A (en) Image classification loss function design method based on cosine space optimization
CN117606801A (en) A cross-domain bearing fault diagnosis method based on multi-representation adaptive network
CN109740481A (en) Atrial fibrillation signal classification method based on skip connection combining CNN and LSTM
CN109214513B (en) An intelligent coupling treatment method for solid-liquid waste based on adaptive deep belief network
Wang Data preprocessing for soft sensor using generative adversarial networks

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant