CN109063765B - An Image Classification Method Based on Gated Neural Network Information Fusion - Google Patents

An Image Classification Method Based on Gated Neural Network Information Fusion Download PDF

Info

Publication number
CN109063765B
CN109063765B CN201810835370.5A CN201810835370A CN109063765B CN 109063765 B CN109063765 B CN 109063765B CN 201810835370 A CN201810835370 A CN 201810835370A CN 109063765 B CN109063765 B CN 109063765B
Authority
CN
China
Prior art keywords
fusion
neural network
input
evidence
weight
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201810835370.5A
Other languages
Chinese (zh)
Other versions
CN109063765A (en
Inventor
庞彦伟
孙汉卿
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tianjin University
Original Assignee
Tianjin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin University filed Critical Tianjin University
Priority to CN201810835370.5A priority Critical patent/CN109063765B/en
Publication of CN109063765A publication Critical patent/CN109063765A/en
Application granted granted Critical
Publication of CN109063765B publication Critical patent/CN109063765B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/257Belief theory, e.g. Dempster-Shafer

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

本发明涉及一种基于门控的神经网络信息融合方法,包括下列步骤:给定需要进行融合的神经网络特征张量,称这些张量为融合输入;确定输出张量的维度,记输出张量为O;对每个输入进行包括变换、激活在内的神经网络计算,使之与输出张量为O具有相同维度;选取合适的融合证据输入:融合证据是指控制每个融合输入的、用于计算融合权重的特征张量,记第i个输入的融合证据为Ei;对第i路融合,对Ei进行神经网络计算;对计算后的融合证据Ei进行激活,得到融合权重αi;将融合权重αi与输入Ii相乘;将各路线性或非线性组合成为输出张量O。

Figure 201810835370

The invention relates to a neural network information fusion method based on gate control, which comprises the following steps: given neural network feature tensors to be fused, these tensors are called fusion inputs; the dimensions of output tensors are determined, and output tensors are recorded. is O; perform neural network calculations including transformation and activation on each input, so that it has the same dimension as the output tensor O; select the appropriate fusion evidence input: fusion evidence refers to controlling each fusion input, using In order to calculate the feature tensor of the fusion weight, record the fusion evidence of the i-th input as E i ; for the i-th fusion, perform neural network calculation on E i ; Activate the calculated fusion evidence E i to obtain the fusion weight α i ; multiply the fusion weight α i with the input I i ; combine the linearity or nonlinearity into an output tensor O.

Figure 201810835370

Description

Image classification method based on gated neural network information fusion
Technical Field
The invention belongs to the field of machine learning and neural networks, and particularly relates to a gated information fusion method for a neural network.
Background
Neural Networks (NN) have achieved good results in many fields such as speech recognition, natural language recognition, image processing, and pattern recognition.
Multitasking, multi-branch neural networks are becoming more popular, and several popular neural network models, such as ResNet (He et, 2016), densnet (huang et, 2017), GRU (Cho et, 2014), etc., all introduce the operation of fusing two-branch information into one branch. On the comprehensive system of a robot, an unmanned aerial vehicle and an automatic driving system, a plurality of branches, a plurality of tasks and other complex neural network models are more and more common, and the information fusion of the neural network models is particularly important in the applications. Most of the existing information fusion methods use splicing (Concatenation) or weighted average as a fusion strategy: the use of splicing can cause the feature dimension to be greatly increased, and a large amount of computing resources are needed; weighted averaging, as a simple linear combination method, cannot fit a nonlinear fusion function.
HeK,ZhangX,Ren S,etal.DeepResidualLearningforImageRecognition[A].IEEEConferenceon ComputerVisionandPatternRecognition[C].LasVegas,NV,UnitedStates:IEEE,2016:770–778.
HuangG,LiuZ,van derMaatenL,etal.DenselyConnectedConvolutionalNetworks[A].IEEE ConferenceonComputerVisionandPatternRecognition[C].Honolulu,HI,USA:IEEE,2017:2261–2269.
ChoK,vanMerrienboerB,BahdanauD,etal.OntheProperties ofNeuralMachineTranslation:Encoder-DecoderApproaches[A].Workshop on Syntax,SemanticsandStructure in StatisticalTranslation[C].Doha,Qatar:2014.
Disclosure of Invention
The invention aims to provide a neural network information fusion method with small calculated amount and strong fitting capability. The technical scheme of the invention is as follows:
a gated neural network information fusion method comprises the following steps:
1) giving the neural network feature tensor I which needs to be fused1,I2,…,InN in total, and the tensors are called fusion input;
2) determining the dimensionality of the output tensor, and recording the output tensor as O;
3) calculating a neural network including transformation and activation for each input, so that the input has the same dimensionality as the output tensor O;
4) selecting a proper fusion evidence to input: the fusion evidence is a feature tensor for calculating fusion weight for controlling each fusion input, and the fusion evidence of the ith input is recorded as EiThen there is a pair input I1,I2,…,InHaving E of1,E2,…,En
5) For the fusion of the ith path, pair EiPerforming neural network calculation;
6) for calculated fused evidence EiActivation is performed to obtain fusion weight αi
7) Fuse weights αiAnd input IiMultiplying;
8) the paths are combined linearly or non-linearly into an output tensor O.
The invention has the substantial characteristics that: by introducing fusion evidence and nonlinear operation, the nonlinear intelligent fusion method capable of fusing any multi-path information and heterogeneous multi-source information is provided, fine-grained fusion of any characteristic tensor can be robustly carried out, and the method can be used for improving any neural network model and some non-neural network models. The beneficial effects are as follows:
1. it is applicable to all neural networks and some non-neural network methods.
2. Compared with the existing fusion method, the invention achieves better fusion performance and provides a fusion strategy of multipath, nonlinear and fine-grained control.
3. The method is simple to realize and has little influence on the existing structure.
Drawings
FIG. 1 Structure of the invention
FIG. 2 embodiment of the fusion construct
Detailed Description
The method of the invention can perform the feature tensor fusion element by element, the size and the quantity of the input and output tensors are flexible and variable, the fusion strategy is self-learning and has strong expression capability, and the method is not limited to a certain neural network and a certain network structure,
has stronger universality and practicability. In order to solve the above problems and achieve the above object, the technical solution of the present invention is as follows:
1) given any number of neural network feature tensors to be fused, these inputs I are recorded1,I2,…,InA total of n (yellow arrows in fig. 1), these inputs are referred to as fusion inputs;
2) determining the dimensionality of the output tensor, and recording the output tensor as O;
3) calculating (including transforming, activating and the like) each input by using a neural network method so that the input has the same dimension as the output tensor O;
4) selecting a proper fusion evidence to input: fused evidence refers to controlThe feature tensor for calculating the fusion weight for this fusion input (green arrow in FIG. 1) is recorded as the fusion evidence for the ith input as EiThen there is a pair input I1,I2,…,InHaving E of1,E2,…,En
5) For the fusion of the ith way, optionally, for EiPerforming neural network calculations ("arbitrary functions" in fig. 1);
6) for calculated fused evidence EiActivating (e.g. Sigmoid, tanh, ReLU, etc.) (the "activation function" in FIG. 1) to obtain fusion weight αi
7) Fuse weights αiAnd input IiMultiplication (symbol "x" in fig. 1);
8) the paths are combined linearly or non-linearly into an output tensor O (the "+" sign in figure 1).
9) In particular, when the data to be fused has only two paths, the fusion evidence of the two paths can be shared and α and 1- α can be used as the weight of the two data to be fused (as shown in fig. 1).
This section will be based on the inclusion-v 4 network architecture featuring multiple parallel branches as proposed by szegdy et al, 2016. it is clear that the invention is not limited to an infrastructure, which is only one example.
SZEGEDY C,IOFFE S,VANHOUCKE V,et al.Inception-v4,Inception-ResNet andthe Impact ofResidual Connections on Learning[C]//AAAI Conference onArtificial Intelligence.San Francisco,CA,USA:AAAI,2017.
(1) Suitable training data is prepared, the training data of the present example including training images and class labels.
(2) And establishing an inclusion-v 4 basic network.
(3) The data to be fused is determined (fig. 2), in this example a three-input fusion unit is used. Specifically, parallel branches of each unit in the inclusion-v 4 are used as fusion inputs.
(4) And selecting fusion evidence. In particular, in this example, the output of the previous unit in the inclusion-v 4 is taken as the fusion evidence of the unit, and a convolution and linear rectification unit (ReLU) is added in sequence as a fusion control branch.
(5) In order to obtain different receptive fields, the convolutions added by the convolution branches with large, medium and small scales in the inclusion-v 4 are respectively 3 × 3 convolution, 3 × 3 dilated convolution with a dilation rate of 2 and 3 × 3 dilated convolution with a dilation rate of 4.
(6) Evidence of fusion is activated. At the end of the fusion control branch of the previous step, tanh is added as an activation function.
(7) Each way fused input is multiplied by a corresponding fusion weight.
(8) And adding the obtained data of each path.
(9) Inputting the training data obtained in the step 1 into the obtained neural network, using an optimization method of mini-batch stochastic gradient descent (mini-batch SGD), selecting the sum of cross entropy loss and weight attenuation loss as a loss item, setting each batch of 32 training images and a weight attenuation coefficient of 0.01, and training until a loss function value is converged by descending in an exponential form of 0.95 power every 1 generation from 0.001.
(10) And (4) storing the neural network weights obtained by training in the step 9.
(11) Inputting the image to be detected into the neural network model obtained in the step 10, and obtaining a prediction result, namely a classification result of the image to be detected.

Claims (1)

1.一种基于门控的神经网络信息融合的图像分类方法,包括下列步骤:1. an image classification method based on gated neural network information fusion, comprising the following steps: (1)准备训练数据,训练数据包括训练图像和类别标注;(1) Prepare training data, which includes training images and category labels; (2)搭建Inception-v4基础网络;(2) Build the Inception-v4 basic network; (3)将Inception-v4中各个单元的并行支路作为融合输入;(3) Use the parallel branches of each unit in Inception-v4 as the fusion input; (4)选取融合证据;将Inception-v4中前一个单元的输出作为本单元的融合证据,并依次添加卷积、线性整流单元(ReLU),作为融合控制支路;(4) Select fusion evidence; take the output of the previous unit in Inception-v4 as the fusion evidence of this unit, and add convolution and linear rectification units (ReLU) in turn as fusion control branches; (5)为了获得不同的感受野,为Inception-v4中大中小三个尺度的卷积支路添加的卷积分别为3×3卷积、膨胀率为2的3×3膨胀卷积、膨胀率为4的3×3膨胀卷积;(5) In order to obtain different receptive fields, the convolutions added to the convolution branches of large, medium and small scales in Inception-v4 are 3×3 convolution, 3×3 dilated convolution with dilation rate of 2, dilation 3×3 dilated convolution with rate 4; (6)激活融合证据:在步骤(4)的融合控制支路结尾,添加tanh作为激活函数;(6) Activating fusion evidence: at the end of the fusion control branch in step (4), add tanh as an activation function; (7)将每一路融合输入与对应的融合权重相乘;(7) Multiply each fusion input by the corresponding fusion weight; (8)将得到的各路数据相加;(8) Add the obtained data of each channel; (9)将步骤1中的训练数据输入所得到的神经网络,使用迷你批量随机梯度下降(mini-batch SGD)的优化方法,选用交叉熵损失和权重衰减损失的和作为损失项,并设置每批含32张训练图像、权重衰减系数0.01,学习率从0.001开始每隔1代以0.95次幂的指数形式下降,训练至损失函数值收敛;(9) Input the training data in step 1 into the obtained neural network, use the mini-batch stochastic gradient descent (mini-batch SGD) optimization method, select the sum of the cross entropy loss and the weight decay loss as the loss term, and set each The batch contains 32 training images, the weight decay coefficient is 0.01, the learning rate starts from 0.001 and decreases exponentially with the power of 0.95 every 1 generation, and the training is until the loss function value converges; (10)将步骤(9)训练得到的神经网络权重保存;(10) save the neural network weights obtained by training in step (9); (11)将待测图像输入步骤(10)得到的神经网络模型中,得到的预测结果即为待测图像的分类结果;(11) Input the image to be tested into the neural network model obtained in step (10), and the obtained prediction result is the classification result of the image to be tested; 上述图像分类方法,采用的是被称之为基于门控的神经网络信息融合方法,此方法为:The above image classification method adopts the so-called gated-based neural network information fusion method, which is as follows: 1)给定需要进行融合的神经网络特征张量I1,I2,…,In共n个,称这些张量为融合输入;1) Given a total of n neural network feature tensors I 1 , I 2 ,..., In that need to be fused, these tensors are called fusion inputs; 2)确定输出张量的维度,记输出张量为O;2) Determine the dimension of the output tensor, and record the output tensor as O; 3)对每个输入进行包括变换、激活在内的神经网络计算,使之与输出张量为O具有相同维度;3) Perform neural network calculations including transformation and activation on each input, so that it has the same dimension as the output tensor of O; 4)选取合适的融合证据输入:融合证据是指控制每个融合输入的、用于计算融合权重的特征张量,记第i个输入的融合证据为Ei,则有对输入I1,I2,…,In,有E1,E2,…,En4) Select the appropriate fusion evidence input: fusion evidence refers to the feature tensor that controls each fusion input and is used to calculate fusion weights. Denote the fusion evidence of the i-th input as E i , then there are pairs of inputs I 1 , I 2 ,..., In , there are E 1 ,E 2 ,...,E n ; 5)对第i路融合,对Ei进行神经网络计算;5) Integrate the i-th road, and perform neural network calculation on E i ; 6)对计算后的融合证据Ei进行激活,得到融合权重αi6) Activating the calculated fusion evidence E i to obtain a fusion weight α i ; 7)将融合权重αi与输入Ii相乘;7) multiply the fusion weight α i with the input I i ; 8)将各路线性或非线性组合成为输出张量O。8) Combine each linear or nonlinear into an output tensor O.
CN201810835370.5A 2018-07-26 2018-07-26 An Image Classification Method Based on Gated Neural Network Information Fusion Expired - Fee Related CN109063765B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810835370.5A CN109063765B (en) 2018-07-26 2018-07-26 An Image Classification Method Based on Gated Neural Network Information Fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810835370.5A CN109063765B (en) 2018-07-26 2018-07-26 An Image Classification Method Based on Gated Neural Network Information Fusion

Publications (2)

Publication Number Publication Date
CN109063765A CN109063765A (en) 2018-12-21
CN109063765B true CN109063765B (en) 2020-03-27

Family

ID=64836583

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810835370.5A Expired - Fee Related CN109063765B (en) 2018-07-26 2018-07-26 An Image Classification Method Based on Gated Neural Network Information Fusion

Country Status (1)

Country Link
CN (1) CN109063765B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114548489A (en) * 2022-01-11 2022-05-27 山东锋士信息技术有限公司 Crop pest and disease damage prediction method and system

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107622275A (en) * 2017-08-21 2018-01-23 西安电子科技大学 A Method of Information Fusion Target Recognition Based on Evidence Synthesis
CN107729991A (en) * 2017-10-19 2018-02-23 天津大学 The neutral net neuron selectivity Activiation method that a kind of position can learn

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107622275A (en) * 2017-08-21 2018-01-23 西安电子科技大学 A Method of Information Fusion Target Recognition Based on Evidence Synthesis
CN107729991A (en) * 2017-10-19 2018-02-23 天津大学 The neutral net neuron selectivity Activiation method that a kind of position can learn

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
Deep Residual Learning for Image Recognition;Kaiming He等;《arXiv.org》;20151231;第1-12页 *
Densely Connected Convolutional Networks;Gao Huang等;《arXiv.org》;20180128;第1-9页 *
Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning;Christian Szegedy等;《Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence》;20171231;第4278-4284页 *
基于D-S证据理论和神经网络的信息融合方法及应用;张佑春等;《Proceedings of the 27th Chinese Control Conference》;20080718;第623-626页 *
基于二叉树型卷积神经网络信息融合的人脸验证;杨子文等;《计算机应用》;20171231;第37卷(第S2期);第155-159页 *

Also Published As

Publication number Publication date
CN109063765A (en) 2018-12-21

Similar Documents

Publication Publication Date Title
CN109816095B (en) Network flow prediction method based on improved gated cyclic neural network
CN108764460A (en) A kind of Time Series Forecasting Methods based on time convolution sum LSTM
CN112949828B (en) A graph convolutional neural network traffic prediction method and system based on graph learning
CN109325547A (en) Non-motor vehicle image multi-tag classification method, system, equipment and storage medium
CN111062395B (en) A real-time video semantic segmentation method
CN110909926A (en) TCN-LSTM-based solar photovoltaic power generation prediction method
CN115359654B (en) Updating method and device of flow prediction system
CN111445498A (en) Target tracking method adopting Bi-L STM neural network
CN112766603B (en) Traffic flow prediction method, system, computer equipment and storage medium
Liu et al. Telecommunication traffic prediction based on improved LSSVM
CN114117229B (en) An item recommendation method based on graph neural network with directed and undirected structural information
Chang et al. Differentiable architecture search with ensemble gumbel-softmax
CN110490324A (en) A kind of gradient decline width learning system implementation method
CN112785056B (en) Short-term load prediction method based on fusion of Catboost and LSTM models
CN111461353A (en) Model training method and system
CN117240384A (en) Intelligent spatial spectrum anti-interference method based on deep learning network
CN110852417A (en) Robustness improvement method of single deep neural network model for IoT applications
CN109063765B (en) An Image Classification Method Based on Gated Neural Network Information Fusion
Zand et al. Flow-based spatio-temporal structured prediction of motion dynamics
US10984320B2 (en) Highly trainable neural network configuration
Jing et al. Task transfer by preference-based cost learning
Lin et al. TAG: Teacher-advice mechanism with Gaussian process for reinforcement learning
CN109697511A (en) Data reasoning method, apparatus and computer equipment
CN112699848B (en) Counting method and system for dense crowd of image
CN114564568A (en) Dialogue state tracking method and system based on knowledge enhancement and context awareness

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20200327

Termination date: 20210726