CN107609566A - Dual network deep learning method based on a small amount of personalized sample - Google Patents

Dual network deep learning method based on a small amount of personalized sample Download PDF

Info

Publication number
CN107609566A
CN107609566A CN201610546022.7A CN201610546022A CN107609566A CN 107609566 A CN107609566 A CN 107609566A CN 201610546022 A CN201610546022 A CN 201610546022A CN 107609566 A CN107609566 A CN 107609566A
Authority
CN
China
Prior art keywords
network
mrow
sample
msub
personalized
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201610546022.7A
Other languages
Chinese (zh)
Inventor
盛益强
赵震宇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Acoustics CAS
Shanghai 3Ntv Network Technology Co Ltd
Original Assignee
Institute of Acoustics CAS
Shanghai 3Ntv Network Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Acoustics CAS, Shanghai 3Ntv Network Technology Co Ltd filed Critical Institute of Acoustics CAS
Priority to CN201610546022.7A priority Critical patent/CN107609566A/en
Publication of CN107609566A publication Critical patent/CN107609566A/en
Pending legal-status Critical Current

Links

Landscapes

  • Machine Translation (AREA)

Abstract

The present invention relates to a kind of dual network deep learning method based on a small amount of personalized sample, the dual network includes reconstructed network and depth network, and this method includes:Gather a small amount of personalized sample;Wherein, the personalized sample includes sample data and its label;The reconstructed network is trained using the sample data in personalized sample, then the label in personalized sample is inputted in the reconstructed network after training, generates new reconstruct data and its label;Based on the new reconstruct data from reconstructed network and its label training depth network;By data input to be tested into trained depth network, the output of sigmoid function is obtained, then response of the result as personalization is selected from the output of the sigmoid function by soft maximum returns.

Description

Dual network deep learning method based on a small amount of personalized sample
Technical field
The present invention relates to communications and data process field, more particularly to the dual network depth based on a small amount of personalized sample Learning method.
Background technology
With the fast development of network technology, capacity, personalization and the diversity of data quickly increase, and processing data Algorithm complex is but difficult to improve, and relies on personal experience and manual operations to describe data, labeled data, selection feature, extraction Feature, the method for processing data, be difficult to meet the needs of individuation data rapid growth, how efficient process personalization number According to having become a urgent problem.
The research of deep learning method is broken through, and a direction for being worth exploring is specified to solve data processing problem.It is deep Degree study can automatically extract feature from big data, and obtain good treatment effect by the sample training of magnanimity.It is actual On, the research of the rapid growth and deep learning of big data is complementary, and the rapid growth of one side big data needs one The method of kind efficient process mass data, the training of another aspect deep learning system need the sample data of magnanimity.
But individuation data, mainly as caused by the user of network edge, current method is first to these data It is collected, then these data is analyzed and handled on server beyond the clouds again.This can produce substantial amounts of data transfer, Meanwhile for terminal user, the response results obtained from server end are that the model trained by global data responds Arrive, it is relatively low for the response accuracy of personalized user.If the structure of deep learning network is carried out in terminal, due to terminal The data volume of personalized user very little, can not be trained to deep learning network, and here it is the big number for being present in current network According to awkward antinomy.
The Hadoop framework of Apache exploitations at present, the parameter server framework of Baidu's exploitation, the Mariana of Tengxun's exploitation Framework etc., some trials made both for distributed extensive, deep learning system, by being split to model, Data are carried out to tear grading mode open, these systems solve the process problem of large-scale data to a certain extent.More than however, These distributed systems be to be handled large-scale model and data on multimachine, substantial amounts of data between multimachine be present Communication, this results in these distributed systems and can be only applied to LAN at present.And terminal user goes for response and also must First it must transfer data in center cluster, then the response of generation is sent to terminal user, therefore user by a center group of planes again End delay response is very high.
In Wide Area Network, the limitation of network bandwidth be present, big data face a difficult choice antinomy, individual demand the problems such as, if The problem of communication cost in distributed big data real time processing system is high, edge customer data are lacked is can solve the problem that, then can be had Distributed big data real time processing system is applied to Wide Area Network by effect ground, is increased network utilization, and improves Consumer's Experience.
Therefore, for a small amount of personalized sample, it is necessary to provide at a kind of distributed data more suitable for Wide Area Network Reason method, it is big to solve the transmission quantity of training data and model parameter, and when distributed data has customized information, place The problem of reason precision is difficult to lifting.
The content of the invention
It is an object of the invention to overcome existing distributed data processing method due to edge customer sample data is few, from And processing accuracy is difficult to the defects of lifting, so as to provide a kind of dual network deep learning method based on a small amount of personalized sample.
To achieve these goals, the invention provides a kind of dual network deep learning side based on a small amount of personalized sample Method, the dual network include reconstructed network and depth network, and this method includes:
Step 1), a small amount of personalized sample of collection;Wherein, the personalized sample includes sample data and its label;
Step 2), using the sample data in personalized sample the reconstructed network is trained, then by personalized sample Label input in reconstructed network after training, generate new reconstruct data and its label;
Step 3), based on the new reconstruct data and its label training depth from reconstructed network obtained by step 2) Spend network;
Step 4), by data input to be tested into trained depth network, obtain the output of sigmoid function, then Response of the result as personalization is selected from the output of the sigmoid function by soft maximum returns.
In above-mentioned technical proposal, the number of a small amount of personalized sample is designated as m, and m ≈ M/J, M therein are to use big data The total sample number during depth network of same precision is trained, J is the number by the sample that property sample reconstructs one by one.
In above-mentioned technical proposal, in step 2), the new reconstruct data and its set expression of label that are generated are {xr k,yr k};Wherein,
xr kRepresent any one reconstruct data, yr kFor label corresponding to the reconstruct data, k value is 1≤k≤m × J, M is personalized sample number, and J is the number of the sample reconstructed by any personalized sample;J and personalized sample number m Relation it is as follows:M values are smaller, and J value is bigger, to ensure to have enough reconstructed sample m × J to train depth network.
In above-mentioned technical proposal, the reconstructed network is including Bayesian network and without any including leveled neural net A kind of learning network with random reconstruct great amount of samples function.
In above-mentioned technical proposal, the reconstructed network is Bayes's programmed instruction programmed learning network with disposable learning functionality.
In above-mentioned technical proposal, the depth network is to include depth feedforward network, depth confidence net, depth Boltzmann Any one including machine, depth convolutional network has high accuracy deep learning network.
In above-mentioned technical proposal, the step 4) further comprises:
Step 4-1), calculate by the hidden layer of depth network the output of each neuron, its calculation formula is:
Wherein, yjIt is the output of j-th of neuron, f is a given activation primitive, wijBe from i-th of neuron to The connection weight of j-th of neuron, xiIt is the input from i-th of neuron, bjIt is biasing weight;
Step 4-2), calculate sigmoid function output;Wherein, the expression formula of sigmoid function is:
The output of resulting sigmoid function is the vector (o of a P dimension1,o2,…,oj,…,oP), vector in each Element oj=sigmoid (yj) all it is real number between 0 and 1, P here is the sum of personalized labels;
Step 4-3), by soft maximum return that to select a result be used as from the output of the sigmoid function personalized Response;Wherein, the formula of soft maximum recurrence is as follows:
Soft maximum return is represented respectively with (softmax (o1),softmax(o2),…,softmax(oj),…, softmax(oP)) probability sigmoid function output (o1,o2,…,oj,…,oP) in choose an ojResponded as personalization.
The advantage of the invention is that:
The present invention is converted a small amount of personalized sample by introducing an auxiliary reconstructed network for supporting an inquiry learning For mass data, so as to solve the problems, such as that a small amount of personalized sample can not directly train a high accuracy depth network.Together When, method of the invention is effectively prevented from the data transfer between center and peripheral, and it is carried out to existing deep learning system Extension, can be applied it in Wide Area Network, and it is not only solved when distributed data has customized information, processing Precision is difficult to the problem of lifting, and improves the transmission cost of existing distributed big data real time processing system.
Brief description of the drawings
Fig. 1 is the flow chart of the dual network deep learning method based on a small amount of personalized sample of the present invention;
Fig. 2 is the structural representation of dual network involved in the present invention.
Embodiment
In conjunction with accompanying drawing, the invention will be further described.
As depicted in figs. 1 and 2, the dual network deep learning method of the invention based on a small amount of personalized sample, including:
Step 1), a small amount of personalized sample of collection, the input as dual network system;Wherein, a small amount of personalized sample Sum is designated as m, and span is m >=1;Preferably, m ≈ M/J, M therein are the depth net that same precision is trained with big data Total sample number during network, J are the number by the sample that property sample reconstructs one by one.Sample in the personalized sample gathered The set of notebook data can be designated asMiddle i span is 1≤i≤m, and superscript p represents personalized (personalized)。
In this step, the personalized sample that is gathered except sample data in itself in addition to, in addition to corresponding to the sample Label.Personalized sample label refers to for arbitrary dataCorrect outputThe set of personalized sample label can be designated asThe personalized sample that this step is gathered is the set of sample data and label, can be designated as
Step 2), the sample data of gathered personalized sample is utilized to train the reconstructed network in dual network (to be designated asAn inquiry learning of a small amount of sample is completed, and the label of personalized sample (is designated as) be input to it is trained Reconstructed network, to generate a large amount of new reconstruct data and its label, the set of new reconstruct data and its label is designated asWhereinAny one reconstruct data,For label corresponding to the reconstruct data, k value for 1≤k≤m × J, m are personalized sample number, and J is the number of the sample reconstructed by any personalized sample, J and personalized sample number M relation is as follows:M values are smaller, and J value is bigger, to ensure to have enough reconstructed sample m × J to train depth network.
Wherein, the reconstructed network E in dual networka (i)Can be any with random reconstruct great amount of samples function Practise network, including Bayesian network and without leveled neural net including;Reconstructed network Ea (i)Preferably there is an inquiry learning work( Bayes's programmed instruction programmed learning network of energy;The priori that described Bayes's programmed instruction programmed learning network first obtains according to background knowledge storehouse Input data is split, input data is split into small part and tuple, then enters these small parts and tuple Row reconfigures, and generates new reconstruct dataAgain from multiple reconstruct data of generation, choose and former data similarity highest Output of the J data as module;Wherein J is positive integer, for representing by property sample reconstructs one by one sample Number.
Step 3), the set based on a large amount of new reconstruct data and its labels from reconstructed networkWith now There is technology to train the depth network in dual network (to be designated as Ed (i));
Wherein, the depth network E in dual networkd (i)Can be that any one has high accuracy deep learning network, it is main To include depth feedforward network, depth confidence net, depth Boltzmann machine, depth convolutional network.
Step 4), trained depth net will be input to for the personalized sample of test or from outside new data Network Ed (i)In, the output of sigmoid function is obtained, then a result is selected from the output of sigmoid function by soft maximum returns, Response as personalization.
Wherein, the expression formula of sigmoid function is:
Wherein, yjIt is the output of j-th of neuron, its value is calculated by the hidden layer of depth network, specific meter It is as follows to calculate formula:
Wherein, f is a given activation primitive, wijIt is the connection weight from i-th of neuron to j-th of neuron, xiIt is the input from i-th of neuron, bjIt is biasing weight.
The output of sigmoid function is the vector (o of a P dimension1,o2,…,oj,…,oP), vector in each element oj= sigmoid(yj) all it is real number between 0 and 1, P here is the sum of personalized labels.
Wherein, the formula of soft maximum recurrence is as follows:
Soft maximum return is represented respectively with (softmax (o1),softmax(o2),…,softmax(oj),…, softmax(oP)) probability in (o1,o2,…,oj,…,oP) choose an ojResponded as personalization.
It should be noted last that the above embodiments are merely illustrative of the technical solutions of the present invention and it is unrestricted.Although ginseng The present invention is described in detail according to embodiment, it will be understood by those within the art that, to the technical side of the present invention Case is modified or equivalent substitution, and without departure from the spirit and scope of technical solution of the present invention, it all should cover in the present invention Right among.

Claims (7)

1. a kind of dual network deep learning method based on a small amount of personalized sample, the dual network include reconstructed network and depth Network, this method include:
Step 1), a small amount of personalized sample of collection;Wherein, the personalized sample includes sample data and its label;
Step 2), using the sample data in personalized sample the reconstructed network is trained, then by the mark in personalized sample In reconstructed network of the label input after training, new reconstruct data and its label are generated;
Step 3), based on the new reconstruct data and its label training depth net from reconstructed network obtained by step 2) Network;
Step 4), by data input to be tested into trained depth network, obtain the output of sigmoid function, then pass through One soft maximum returns selects response of the result as personalization from the output of the sigmoid function.
2. the dual network deep learning method according to claim 1 based on a small amount of personalized sample, it is characterised in that institute The number for stating a small amount of personalized sample is designated as m, and m ≈ M/J, M therein are when training the depth network of same precision with big data Total sample number, J is by the number of the sample that property sample reconstructs one by one.
3. the dual network deep learning method according to claim 1 based on a small amount of personalized sample, it is characterised in that In step 2), the new reconstruct data and its set expression of label that are generated are { xr k,yr k};Wherein,
xr kRepresent any one reconstruct data, yr kFor label corresponding to the reconstruct data, k value is 1≤k≤m × J, and m is individual Property number of samples, J is the number of sample reconstructed by any personalized sample;J and personalized sample number m relation It is as follows:M values are smaller, and J value is bigger, to ensure to have enough reconstructed sample m × J to train depth network.
4. the dual network deep learning method according to claim 1 based on a small amount of personalized sample, it is characterised in that institute It is to have random reconstruct great amount of samples including Bayesian network and without any one including leveled neural net to state reconstructed network The learning network of function.
5. the dual network deep learning method according to claim 4 based on a small amount of personalized sample, it is characterised in that institute It is Bayes's programmed instruction programmed learning network with disposable learning functionality to state reconstructed network.
6. the dual network deep learning method according to claim 1 based on a small amount of personalized sample, it is characterised in that institute It is any including depth feedforward network, depth confidence net, depth Boltzmann machine, depth convolutional network to state depth network One kind has high accuracy deep learning network.
7. the dual network deep learning method according to claim 1 based on a small amount of personalized sample, it is characterised in that institute Step 4) is stated to further comprise:
Step 4-1), calculate by the hidden layer of depth network the output of each neuron, its calculation formula is:
<mrow> <msub> <mi>y</mi> <mi>j</mi> </msub> <mo>=</mo> <mi>f</mi> <mrow> <mo>(</mo> <munder> <mo>&amp;Sigma;</mo> <mi>i</mi> </munder> <msub> <mi>w</mi> <mrow> <mi>i</mi> <mi>j</mi> </mrow> </msub> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>+</mo> <msub> <mi>b</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> </mrow>
Wherein, yjIt is the output of j-th of neuron, f is a given activation primitive, wijIt is from i-th of neuron to j-th The connection weight of neuron, xiIt is the input from i-th of neuron, bjIt is biasing weight;
Step 4-2), calculate sigmoid function output;Wherein, the expression formula of sigmoid function is:
<mrow> <mi>s</mi> <mi>i</mi> <mi>g</mi> <mi>m</mi> <mi>o</mi> <mi>i</mi> <mi>d</mi> <mrow> <mo>(</mo> <msub> <mi>y</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mn>1</mn> <mrow> <mn>1</mn> <mo>+</mo> <mi>exp</mi> <mrow> <mo>(</mo> <msub> <mi>y</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> </mrow> </mfrac> <mo>;</mo> </mrow>
The output of resulting sigmoid function is the vector (o of a P dimension1,o2,…,oj,…,oP), vector in each element oj=sigmoid (yj) all it is real number between 0 and 1, P here is the sum of personalized labels;
Step 4-3), pass through soft maximum return and select a result be used as personalized response from the output of the sigmoid function; Wherein, the formula of soft maximum recurrence is as follows:
<mrow> <mi>s</mi> <mi>o</mi> <mi>f</mi> <mi>t</mi> <mi> </mi> <mi>m</mi> <mi>a</mi> <mi>x</mi> <mrow> <mo>(</mo> <msub> <mi>o</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mrow> <mi>exp</mi> <mrow> <mo>(</mo> <msub> <mi>o</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> </mrow> <mrow> <munder> <mo>&amp;Sigma;</mo> <mi>i</mi> </munder> <mi>exp</mi> <mrow> <mo>(</mo> <msub> <mi>o</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> </mrow> </mfrac> </mrow>
Soft maximum return is represented respectively with (softmax (o1),softmax(o2),…,softmax(oj),…,softmax (oP)) probability sigmoid function output (o1,o2,…,oj,…,oP) in choose an ojResponded as personalization.
CN201610546022.7A 2016-07-12 2016-07-12 Dual network deep learning method based on a small amount of personalized sample Pending CN107609566A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610546022.7A CN107609566A (en) 2016-07-12 2016-07-12 Dual network deep learning method based on a small amount of personalized sample

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610546022.7A CN107609566A (en) 2016-07-12 2016-07-12 Dual network deep learning method based on a small amount of personalized sample

Publications (1)

Publication Number Publication Date
CN107609566A true CN107609566A (en) 2018-01-19

Family

ID=61055027

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610546022.7A Pending CN107609566A (en) 2016-07-12 2016-07-12 Dual network deep learning method based on a small amount of personalized sample

Country Status (1)

Country Link
CN (1) CN107609566A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110223785A (en) * 2019-05-28 2019-09-10 北京师范大学 A kind of infectious disease transmission network reconstruction method based on deep learning

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110223785A (en) * 2019-05-28 2019-09-10 北京师范大学 A kind of infectious disease transmission network reconstruction method based on deep learning

Similar Documents

Publication Publication Date Title
Lim et al. Time-series forecasting with deep learning: a survey
Wu et al. Evolving RBF neural networks for rainfall prediction using hybrid particle swarm optimization and genetic algorithm
Makridakis et al. Statistical, machine learning and deep learning forecasting methods: Comparisons and ways forward
Gong et al. Multi-objective parameter optimization of common land model using adaptive surrogate modeling
Chen et al. Dictionary learning over distributed models
Corchado et al. Ibr retrieval method based on topology preserving mappings
CN107944559A (en) A kind of entity relationship automatic identifying method and system
CN103942749B (en) A kind of based on revising cluster hypothesis and the EO-1 hyperion terrain classification method of semi-supervised very fast learning machine
Konomi et al. Bayesian treed multivariate gaussian process with adaptive design: Application to a carbon capture unit
Chauhan et al. Approximating parameters of photovoltaic models using an amended reptile search algorithm
Said et al. AI-based solar energy forecasting for smart grid integration
Wang et al. An approach of recursive timing deep belief network for algal bloom forecasting
CN109670927A (en) The method of adjustment and its device of credit line, equipment, storage medium
Sahoo et al. Space–time forecasting of groundwater level using a hybrid soft computing model
Bakhashwain Prediction of global solar radiation using support vector machines
Zhang et al. Wind speed forecasting based on wavelet decomposition and wavelet neural networks optimized by the Cuckoo search algorithm
Wen et al. A solar forecasting framework based on federated learning and distributed computing
Fang et al. A novel application of artificial neural network for wind speed estimation
Abed et al. Data mining in web personalization using the blended deep learning model
Dunbar et al. Ensemble‐based experimental design for targeting data acquisition to inform climate models
Zhao et al. Short‐term load forecasting of multi‐scale recurrent neural networks based on residual structure
Zhang et al. Digital twin empowered PV power prediction
Al-Hajj et al. Estimating solar irradiance using genetic programming technique and meteorological records.
Gouthamkumar et al. Hybridized gravitational search algorithm for short-term hydrothermal scheduling
Semero et al. A GA-PSO hybrid algorithm based neural network modeling technique for short-term wind power forecasting

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20180119