CN110532545A - A kind of data information abstracting method based on complex neural network modeling - Google Patents

A kind of data information abstracting method based on complex neural network modeling Download PDF

Info

Publication number
CN110532545A
CN110532545A CN201910665705.8A CN201910665705A CN110532545A CN 110532545 A CN110532545 A CN 110532545A CN 201910665705 A CN201910665705 A CN 201910665705A CN 110532545 A CN110532545 A CN 110532545A
Authority
CN
China
Prior art keywords
neural network
module
complex neural
data information
output
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910665705.8A
Other languages
Chinese (zh)
Inventor
肖清林
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xiamen Useear Information Technology Co ltd
Original Assignee
Fujian Singularity Space-Time Digital Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fujian Singularity Space-Time Digital Technology Co Ltd filed Critical Fujian Singularity Space-Time Digital Technology Co Ltd
Priority to CN201910665705.8A priority Critical patent/CN110532545A/en
Publication of CN110532545A publication Critical patent/CN110532545A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)

Abstract

A kind of data information abstracting method based on complex neural network modeling, method and step includes: to establish training module and complex neural network model;Complex neural network model is trained using training module;Data information is collected, database module is constructed;Information extraction is carried out to database module using the complex neural network model of optimization.The present invention extracts information on the basis of the complex neural network model of BP and CNN, the cooperation, complementation of the two, so that information extraction accuracy is high, speed is fast, meets currently the needs of to information extraction.

Description

A kind of data information abstracting method based on complex neural network modeling
Technical field
The present invention relates to data informations to extract field more particularly to a kind of data information based on complex neural network modeling Abstracting method.
Background technique
The research hotspot that artificial intelligence field rises since artificial neural network is the 1980s.It is from information Reason angle is abstracted human brain neuroid, establishes certain naive model, different nets is formed by different connection types Network.Neural network or neural network are also often directly referred to as in engineering and academia.Neural network is a kind of operational model, by Composition is coupled to each other between a large amount of node (or neuron).A kind of each specific output function of node on behalf referred to as swashs Encourage function.Connection between every two node all represents a weighted value for passing through the connection signal, referred to as weight, this phase When in the memory of artificial neural network.The output of network then according to the connection type of network, the difference of weighted value and excitation function and It is different.Recently during the last ten years, the research work of artificial neural network deepens continuously, and very big progress is had been achieved for, in mould The fields such as formula identification, intelligent robot, automatic control, predictive estimation, biology, medicine, economy have successfully solved many existing For the insoluble practical problem of computer, good intelligent characteristic is shown, the type of artificial neural network is also gradually more Sample, wherein BP and CNN are most widely used.
Wherein BP is a kind of Multi-layered Feedforward Networks by error back propagation training, its basic thought is gradient decline Method, using gradient search technology, to make the real output value of network and the error mean square difference minimum of desired output;CNN Be it is a kind of copy the visual perception mechanism construction of biology comprising convolutional calculation and with the feedforward neural network of depth structure, can be with Exercise supervision study and unsupervised learning, and the sparsity that the convolution kernel parameter sharing in hidden layer is connected with interlayer makes CNN Feature can be revealed with lesser calculation amount plaid matching.But when two kinds of neural networks are used alone, still there is certain limitation, Such as when extracting to data information, individually a kind of Accuracy and high efficiency of neural network model is no longer satisfied people Demand.
To solve the above problems, proposing a kind of data information extraction side based on complex neural network modeling in the application Method.
Summary of the invention
(1) goal of the invention
To solve technical problem present in background technique, the present invention proposes a kind of number based on complex neural network modeling According to information extraction method, the present invention extracts information, the two on the basis of the complex neural network model of BP and CNN Cooperation, complementation so that information extraction accuracy is high, speed is fast, meets currently the needs of to information extraction.
(2) technical solution
To solve the above problems, the present invention provides a kind of data information extraction sides based on complex neural network modeling Method includes the following steps:
S1, training module and complex neural network model are established, wherein complex neural network model is compound by CNN and BP It constitutes;
S2, complex neural network model is trained using training module, the complex neural network model optimized; The complex neural network model wherein optimized includes input module, hidden module and output module;
S3, data information is collected, constructs database module;
S4, information extraction is carried out to database module using the complex neural network model of optimization;Extraction process are as follows:
A1, command signal is assigned to the complex neural network model of optimization by input module;
A2, signal are delivered to hidden module, and transmit step by step between neuron along positive direction;
If a3, obtaining matched data information in transmittance process, output module recalls data information, power cut-off Journey;
If a4, signal are transferred to the bottom end of neural network, matched data information is not obtained yet, then is transferred to backpropagation, is believed Number along original route return;In return course, the weight and threshold value of neurons at different levels are modified, reduces signal errors;
A5, signal modify the weight of neurons at different levels constantly along both forward and reverse directions back and forth between neurons at different levels simultaneously And threshold value, until obtaining matched data information, output module recalls data information, power cut-off process.
Preferably, in S2, the forward direction including signal is trained to complex neural network model using training module and is passed Broadcast two processes of backpropagation with error.
Preferably, it is carried out when calculating error output by from the direction for being input to output, and adjusts weight and threshold value then from defeated The direction for arriving input out carries out;When forward-propagating, input module input signal acts on output node by hidden module, passes through Nonlinear transformation generates output signal by output module and is transferred to the anti-of error if reality output is not consistent with desired output To communication process;Error-duration model is by hidden module by output error to the layer-by-layer anti-pass of input module, and by error distribution to each All neurons of layer, the error signal obtained from each layer neuron adjust the weight and threshold value of each neuron.By adjusting input The linking intensity and threshold value of Module nodes and hidden Module nodes decline error along gradient direction, train by repetition learning, Determine network parameter (weight and threshold value) corresponding with minimal error, training stops stopping.
Preferably, convolutional layer is provided in hidden module.
Preferably, for carrying out the convolutional layer of feature extraction to input data, internal includes multiple convolution kernels;Convolution kernel At work, it does matrix element multiplication to input feature vector to sum and be superimposed departure, calculating function is
Wherein b is departure, ZlAnd Zl+1Indicate that l+1 layers of convolution is output and input, also referred to as characteristic pattern, Ll+1For Zl+1Size, it is assumed here that feature Figure length and width are identical, and the pixel of Z (i, j) character pair figure, K is characterized the port number of figure, f, s0It is convolution layer parameter with p, it is corresponding Convolution kernel size, convolution step-length and the filling number of plies.
Above-mentioned technical proposal of the invention has following beneficial technical effect:
The present invention extracts information on the basis of the complex neural network model of BP and CNN, and wherein BP is a kind of By the Multi-layered Feedforward Networks of error back propagation training, its basic thought is gradient descent method, using gradient search technology, with Phase makes the real output value of network and the error mean square difference minimum of desired output;CNN is a kind of comprising convolutional calculation and tool There is the feedforward neural network of depth structure, copy the visual perception mechanism construction of biology, can exercise supervision study and non-supervisory It practises, the sparsity that the convolution kernel parameter sharing in hidden layer is connected with interlayer enables CNN with lesser calculation amount plaid matching Reveal feature;The cooperation, complementation of the two, so that information extraction accuracy is high, speed is fast, meets currently to the need of information extraction It asks.
Detailed description of the invention
Fig. 1 is a kind of flow chart of data information abstracting method based on complex neural network modeling proposed by the present invention.
Specific embodiment
In order to make the objectives, technical solutions and advantages of the present invention clearer, With reference to embodiment and join According to attached drawing, the present invention is described in more detail.It should be understood that these descriptions are merely illustrative, and it is not intended to limit this hair Bright range.In addition, in the following description, descriptions of well-known structures and technologies are omitted, to avoid this is unnecessarily obscured The concept of invention.
As shown in Figure 1, a kind of data information abstracting method based on complex neural network modeling proposed by the present invention, including Following steps:
S1, training module and complex neural network model are established, wherein complex neural network model is compound by CNN and BP It constitutes;
S2, complex neural network model is trained using training module, the complex neural network model optimized; The complex neural network model wherein optimized includes input module, hidden module and output module;
S3, data information is collected, constructs database module;
S4, information extraction is carried out to database module using the complex neural network model of optimization;Extraction process are as follows:
A1, command signal is assigned to the complex neural network model of optimization by input module;
A2, signal are delivered to hidden module, and transmit step by step between neuron along positive direction;
If a3, obtaining matched data information in transmittance process, output module recalls data information, power cut-off Journey;
If a4, signal are transferred to the bottom end of neural network, matched data information is not obtained yet, then is transferred to backpropagation, is believed Number along original route return;In return course, the weight and threshold value of neurons at different levels are modified, reduces signal errors;
A5, signal modify the weight of neurons at different levels constantly along both forward and reverse directions back and forth between neurons at different levels simultaneously And threshold value, until obtaining matched data information, output module recalls data information, power cut-off process.
In an alternative embodiment, in S2, packet is trained to complex neural network model using training module Include two processes of backpropagation of the propagated forward and error of signal.
In an alternative embodiment, it is carried out when calculating error output by from the direction for being input to output, and adjusts power Value and threshold value are then carried out from the direction for being output to input;When forward-propagating, input module input signal is acted on by hidden module Output node generates output signal by output module by nonlinear transformation, if reality output is not consistent with desired output, Then it is transferred to the back-propagation process of error;Error-duration model is by hidden module by output error to the layer-by-layer anti-pass of input module, and Give error distribution to all neurons of each layer, the error signal obtained from each layer neuron adjusts the weight and threshold of each neuron Value.By adjusting the linking intensity and threshold value of input module node and hidden Module nodes, decline error along gradient direction, warp Repetition learning training is crossed, determines network parameter (weight and threshold value) corresponding with minimal error, training stops stopping.
In an alternative embodiment, convolutional layer is provided in hidden module.
In an alternative embodiment, internal comprising more for carrying out the convolutional layer of feature extraction to input data A convolution kernel;Convolution kernel at work, does matrix element multiplication to input feature vector and sums and be superimposed departure, calculating function is
Wherein b is departure, ZlAnd Zl+1Indicate that l+1 layers of convolution is output and input, also referred to as characteristic pattern, Ll+1For Zl+1Size, it is assumed here that feature Figure length and width are identical, and the pixel of Z (i, j) character pair figure, K is characterized the port number of figure, f, s0It is convolution layer parameter with p, it is corresponding Convolution kernel size, convolution step-length and the filling number of plies.
In the present invention, on the basis of the complex neural network model of BP and CNN, information is extracted, wherein BP is A kind of Multi-layered Feedforward Networks by error back propagation training, its basic thought is gradient descent method, utilizes gradient search skill Art, to make the real output value of network and the error mean square difference minimum of desired output;CNN is a kind of comprising convolution meter Calculate and with depth structure feedforward neural network, copy biology visual perception mechanism construction, can exercise supervision study and it is non- Supervised learning, the sparsity that the convolution kernel parameter sharing in hidden layer is connected with interlayer enable CNN with lesser calculating Amount plaid matching reveals feature;The cooperation, complementation of the two, so that information extraction accuracy is high, speed is fast, meets and currently takes out to information The demand taken.
It should be understood that above-mentioned specific embodiment of the invention is used only for exemplary illustration or explains of the invention Principle, but not to limit the present invention.Therefore, that is done without departing from the spirit and scope of the present invention is any Modification, equivalent replacement, improvement etc., should all be included in the protection scope of the present invention.In addition, appended claims purport of the present invention Covering the whole variations fallen into attached claim scope and boundary or this range and the equivalent form on boundary and is repairing Change example.

Claims (5)

1. a kind of data information abstracting method based on complex neural network modeling, which comprises the steps of:
S1, training module and complex neural network model are established, wherein complex neural network model is constituted by CNN and BP are compound;
S2, complex neural network model is trained using training module, the complex neural network model optimized;Wherein The complex neural network model of optimization includes input module, hidden module and output module;
S3, data information is collected, constructs database module;
S4, information extraction is carried out to database module using the complex neural network model of optimization;Extraction process are as follows:
A1, command signal is assigned to the complex neural network model of optimization by input module;
A2, signal are delivered to hidden module, and transmit step by step between neuron along positive direction;
If a3, obtaining matched data information in transmittance process, output module recalls data information, power cut-off process;
If a4, signal are transferred to the bottom end of neural network, matched data information is not obtained yet, then is transferred to backpropagation, signal edge Original route returns;In return course, the weight and threshold value of neurons at different levels are modified, reduces signal errors;
A5, signal modify the weight and threshold of neurons at different levels constantly along both forward and reverse directions back and forth between neurons at different levels simultaneously Value, until obtaining matched data information, output module recalls data information, power cut-off process.
2. a kind of data information abstracting method based on complex neural network modeling according to claim 1, feature exist In being trained the anti-of propagated forward including signal and error to complex neural network model using training module in S2 To two processes of propagation.
3. a kind of data information abstracting method based on complex neural network modeling according to claim 2, feature exist In, it is carried out when calculating error output by from the direction for being input to output, and weight and threshold value are adjusted then from the side for being output to input To progress;When forward-propagating, input module input signal acts on output node by hidden module, by nonlinear transformation, leads to It crosses output module generation output signal and is transferred to the back-propagation process of error if reality output is not consistent with desired output;Accidentally Poor anti-pass is by hidden module by output error to the layer-by-layer anti-pass of input module, and gives error distribution to all neurons of each layer, The error signal obtained from each layer neuron, adjusts the weight and threshold value of each neuron.
4. a kind of data information abstracting method based on complex neural network modeling according to claim 1, feature exist In being provided with convolutional layer in hidden module.
5. a kind of data information abstracting method based on complex neural network modeling according to claim 4, feature exist In for carrying out the convolutional layer of feature extraction to input data, internal includes multiple convolution kernels;Convolution kernel is at work, right Input feature vector does matrix element multiplication and sums and be superimposed departure, calculates function and is
Wherein b is departure, ZlWith Zl+1Indicate that l+1 layers of convolution is output and input, also referred to as characteristic pattern, Ll+1For Zl+1Size, it is assumed here that characteristic pattern is long Width is identical, and the pixel of Z (i, j) character pair figure, K is characterized the port number of figure, f, s0It is convolution layer parameter with p, corresponding convolution Core size, convolution step-length and the filling number of plies.
CN201910665705.8A 2019-07-23 2019-07-23 A kind of data information abstracting method based on complex neural network modeling Pending CN110532545A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910665705.8A CN110532545A (en) 2019-07-23 2019-07-23 A kind of data information abstracting method based on complex neural network modeling

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910665705.8A CN110532545A (en) 2019-07-23 2019-07-23 A kind of data information abstracting method based on complex neural network modeling

Publications (1)

Publication Number Publication Date
CN110532545A true CN110532545A (en) 2019-12-03

Family

ID=68661714

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910665705.8A Pending CN110532545A (en) 2019-07-23 2019-07-23 A kind of data information abstracting method based on complex neural network modeling

Country Status (1)

Country Link
CN (1) CN110532545A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111831657A (en) * 2020-06-28 2020-10-27 苏州市测绘院有限责任公司 Leveling data real-time processing method
CN111881034A (en) * 2020-07-23 2020-11-03 深圳慕智科技有限公司 Confrontation sample generation method based on distance
CN113177997A (en) * 2021-04-25 2021-07-27 上海雷鸣文化传播有限公司 Method for realizing ink and wash effect rendering based on neural network graphic algorithm

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7069256B1 (en) * 2002-05-23 2006-06-27 Oracle International Corporation Neural network module for data mining
CN106845478A (en) * 2016-12-30 2017-06-13 同观科技(深圳)有限公司 The secondary licence plate recognition method and device of a kind of character confidence level
CN108460742A (en) * 2018-03-14 2018-08-28 日照职业技术学院 A kind of image recovery method based on BP neural network

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7069256B1 (en) * 2002-05-23 2006-06-27 Oracle International Corporation Neural network module for data mining
CN106845478A (en) * 2016-12-30 2017-06-13 同观科技(深圳)有限公司 The secondary licence plate recognition method and device of a kind of character confidence level
CN108460742A (en) * 2018-03-14 2018-08-28 日照职业技术学院 A kind of image recovery method based on BP neural network

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
姚红革等: "基于小波分析和BP神经网络的图像特征提取", 《西安工业大学学报》 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111831657A (en) * 2020-06-28 2020-10-27 苏州市测绘院有限责任公司 Leveling data real-time processing method
CN111881034A (en) * 2020-07-23 2020-11-03 深圳慕智科技有限公司 Confrontation sample generation method based on distance
CN113177997A (en) * 2021-04-25 2021-07-27 上海雷鸣文化传播有限公司 Method for realizing ink and wash effect rendering based on neural network graphic algorithm

Similar Documents

Publication Publication Date Title
Li et al. Prediction for tourism flow based on LSTM neural network
CN107247989A (en) A kind of neural network training method and device
CN104978580B (en) A kind of insulator recognition methods for unmanned plane inspection transmission line of electricity
CN110532545A (en) A kind of data information abstracting method based on complex neural network modeling
CN105160400B (en) The method of lifting convolutional neural networks generalization ability based on L21 norms
CN110060475A (en) A kind of multi-intersection signal lamp cooperative control method based on deeply study
CN109241291A (en) Knowledge mapping optimal path inquiry system and method based on deeply study
CN110223517A (en) Short-term traffic flow forecast method based on temporal correlation
CN106558058A (en) Parted pattern training method, lane segmentation method, control method for vehicle and device
CN110334705A (en) A kind of Language Identification of the scene text image of the global and local information of combination
CN109492556A (en) Synthetic aperture radar target identification method towards the study of small sample residual error
CN104217214A (en) Configurable convolutional neural network based red green blue-distance (RGB-D) figure behavior identification method
CN109063569A (en) A kind of semantic class change detecting method based on remote sensing image
CN108021947A (en) A kind of layering extreme learning machine target identification method of view-based access control model
CN109558902A (en) A kind of fast target detection method
CN113205595B (en) Construction method and application of 3D human body posture estimation model
CN108710906A (en) Real-time point cloud model sorting technique based on lightweight network LightPointNet
CN109241995A (en) A kind of image-recognizing method based on modified ArcFace loss function
CN106650918B (en) The method and apparatus for constructing system model
Qu et al. Visual cross-image fusion using deep neural networks for image edge detection
Pérez et al. Physics-Informed Long-Short Term Memory Neural Network Performance on Holloman High-Speed Test Track Sled Study
CN106896724A (en) For the tracking system and tracking of sun tracker
CN110503148A (en) A kind of point cloud object identifying method with scale invariability
CN110188621A (en) A kind of three-dimensional face expression recognition methods based on SSF-IL-CNN
CN109543748A (en) A kind of fruit recognition methods based on multi-model fusion

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right

Effective date of registration: 20220913

Address after: 361000 units 1702 and 1703, No. 59, Chengyi North Street, phase III, software park, Xiamen, Fujian

Applicant after: XIAMEN USEEAR INFORMATION TECHNOLOGY Co.,Ltd.

Address before: Unit 1701, unit 1704, No. 59, Chengyi North Street, phase III, software park, Xiamen City, Fujian Province, 361000

Applicant before: FUJIAN QIDIAN SPACE-TIME DIGITAL TECHNOLOGY Co.,Ltd.

TA01 Transfer of patent application right
RJ01 Rejection of invention patent application after publication

Application publication date: 20191203

RJ01 Rejection of invention patent application after publication