CN106649203A - Method for improving processing quality of big data - Google Patents

Method for improving processing quality of big data Download PDF

Info

Publication number
CN106649203A
CN106649203A CN201611232063.5A CN201611232063A CN106649203A CN 106649203 A CN106649203 A CN 106649203A CN 201611232063 A CN201611232063 A CN 201611232063A CN 106649203 A CN106649203 A CN 106649203A
Authority
CN
China
Prior art keywords
function
powerball
value
big data
gradient
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201611232063.5A
Other languages
Chinese (zh)
Inventor
袁烨
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huazhong University of Science and Technology
Original Assignee
Huazhong University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huazhong University of Science and Technology filed Critical Huazhong University of Science and Technology
Priority to CN201611232063.5A priority Critical patent/CN106649203A/en
Publication of CN106649203A publication Critical patent/CN106649203A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/16Matrix or vector computation, e.g. matrix-matrix or matrix-vector multiplication, matrix factorization

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Pure & Applied Mathematics (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Computational Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computing Systems (AREA)
  • Algebra (AREA)
  • Databases & Information Systems (AREA)
  • Software Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a method for improving the processing quality of big data. A big data processing and analysis problem is transformed into a problem of solving the minimum value, based on big data processing characteristics, an existing numerical optimization algorithm is improved, and a numerical optimization method for accelerating the convergence rate of an algorithm is obtained. On the basis of an existing iterative algorithm, the Powerball function is introduced, to be specific, the nonlinear Powerball function about a target function gradient term is introduced in an iterative formula, and the function contains a dynamic coefficient so as to improve the convergence rate of the algorithm. The method has the advantages of being high in convergence rate, short in operation time and the like, the problem about the big data analysis speed and quality can be well solved, and the running time of a computer can be significantly reduced.

Description

It is a kind of to improve the method that big data processes quality
Technical field
The invention belongs to big data is processed and data mining problem field, more particularly, to a class data analysis is improved The data processing technique of speed and data analysis quality.
Background technology
Present society is the society of a high speed development, and science and technology is flourishing, information flow, and the exchange between people is increasingly Closely, life also increasingly facilitates, and big data is exactly the product of this cyberage.The development of computer realm, Yi Jiguo The appearance of border internet, the optimization problem to solve large-scale data provides convenient.Into 21 century since, big rule have been solved The algorithm design of mould optimization problem and theory innovation are by the extensive concern of each research field expert.How by big data Analyzing and process problem is converted into and solves Numerical Optimization in large scale, and how to solve large-scale optimization problem is Instantly one of emphasis that each area research person pays close attention to.And the previous work with regard to solving-optimizing problem is focused primarily upon and changed For the selection of parameter in algorithm, some more classical numerical optimisation algorithms are worth to by choosing different parameters, so far still What is be widely used has the methods such as famous gradient descent method, Newton method.But in numerical optimisation algorithms, convergence of algorithm speed Rate is to weigh the good and bad important indicator of the algorithm, therefore how to accelerate convergence of algorithm speed and be particularly important.Substantially, when When solving large-scale optimization problem, primary iteration calculate constraint it is critical that
Big data has the features such as data scale of construction is huge, data type is various, value density is low, processing speed is fast, research Solve having with the efficient algorithm that calculating is easy, storage demand amount is little for large-scale data problem analysis highly important Theoretical and practical significance.How quickly and efficiently from the data of various type, quickly obtain valuable information, be big One of primary goal of data processing.And the essence of data processing, it is equivalent to not only quick but also efficiently solves optimized problem.
In fact, it is ubiquitous to optimize this concept, therefore as a kind of optimization of the means being optimal Method, it should be and be really also with countless changes.The problem overwhelming majority handled in operational research is all optimization problem. Method for solving these problems, such as Mathematical Planning, queueing theory, Analysis of Policy Making, analogue technique etc., naturally also just belong to Optimal method this category.In addition, optimizing also includes Engineering Control, optimum control, systematic science etc..It is wherein optimum Control is mainly used in the optimization to various control systems, and for example, the optimum control of guided missile system can guarantee that and be completed with minimum fuel Aerial mission, with the shortest time target is reached;The works such as the optimum control of aircraft, ship, power system etc. for another example, chemical industry, metallurgy The control of the optimum condition of factory.The further development of computer interface unit constantly improve and optimization method, also exists for computer Line production control creates advantage.The object of optimum control is also by from the control to hardware systems such as machinery, electric, chemical industry Turn to ecological, environment so that the control of social economic system.
Optimal method discussion be decision problem optimal selection characteristic, construction seeks the computational methods of optimal solution, Study the theory property and Practical Calculation performance of these computational methods.Along with the high speed development and optimized calculation method of computer Progress, fairly large optimization problem is resolved.Because optimization problem extensively sees Economic planning, engineering design, life The key areas such as management, communications and transportation, national defence are produced, it is paid much attention to by government department, scientific research institution and branch of industry. In the face of data scale it is huge the features such as, the numerical algorithm of existing solving-optimizing problem either convergence rate, or during operation Between and the aspect such as running memory, can not meet the requirement of big data process.
The existing algorithm for processing optimization problem has a lot, but main still more classical gradient descent method, Newton method And quasi-Newton method, the hereinafter referred to as numerical method containing gradient terms is gradient class method.
I) gradient descent method:
Gradient descent method is a kind of optimization algorithm, be to solve for that unconstrained optimization problem is most simple and most ancient method it One, although no longer there is stronger practicality now, but many efficient algorithms be all be improved based on it and Obtained from amendment.It with negative gradient direction is the direction of search that steepest descent method is, steepest descent method is closer to desired value, step Length is less, advances slower.
Ii) Newton method:
Newton method is to solve for optimization problem (1.1) ancient and effective method, asks without constraint compared to other solutions The method of topic, the method needs less iterations, functional value calculation times when optimum point is found.Classic Newton method One significant advantage is its Local quadratic convergence, but the successful key of Newton method is to make use of sea plug Hesse matrixes to carry For curvature information, Newton method requires the second dervative of calculating target function, and when solution of the iteration point away from problem, function The Hesse matrixes of f may not positive definite or even unusual, now Newton method failure.
Iii) quasi-Newton method:
The workload for calculating Hesse matrixes by above-mentioned Newton method is larger, and the Hesse matrixes of the object function having are difficult Calculate, in addition it is bad obtain, quasi-Newton method be exactly construct object function curvature it is approximate, and do not need obvious form Hesse matrixes, while with fast convergence rate a little.
Consider following most widely used quasi-Newton method:L-BFGS methods
First, it is assumed that object function f (x) be Second Order Continuous can be micro-, using Taylor launch can obtain
Wherein X (k+1)=X (k)+αkdk, dk=-Hk▽f(X(k)).Construction ▽2F's (X (k)) suitably approaches matrix BkSo that following equation is set up
Bk+1Sk=Yk,
Wherein, Sk=X (k+1)-X (k), Yk=▽ f (X (k+1))-▽ f (X (k)).The B of next step iterationk+1By as follows Correction formula is obtained:
In order to avoid asking in each iterationOrderCan calculate next by following correction formula accordingly H needed for secondary iterationk+1,
K values are 0,1,2 in formula ..., when k=0 takes at the beginning initial matrix H0For unit matrix.
Quasi-Newton method is built upon the optimized algorithm on the basis of Newton method, and the method is mainly using target function value and single order The information of derivative is iterated calculating, the characteristics of it has fast convergence rate, and avoids the calculating of object function second dervative. But when the dimension of problem is very big, the method needs very big memory space.
Above-mentioned these algorithms are faced with that convergence rate is slow, precision is relatively low, Yi Jiji when the related problem of big data is processed The problems such as calculation amount is greatly and the requirement to internal memory is larger, therefore be not suitable for solving some optimization problems related to big data, And the aspect such as the development and application to information included in data.
The content of the invention
For the problems referred to above, the invention provides a kind of method for improving data processing quality, for solution and big data Some optimization problems of correlation or the problem of solution object function minimum of a value, method provided by the present invention solves prior art and deposits Convergence rate it is slow, relatively low and computationally intensive and to internal memory the requirement of precision is larger the problems such as
Method proposed by the present invention comprises the steps:
(1) according to the data characteristic collected, data are processed:If whether data processing problem is to solve for function Minimum of a value optimization problem:It is to go to step (2);Otherwise, solution minimum of a value optimization problem is converted into, is gone to step (2);
(2) minimum of a value Optimized model is set upWherein RnFor the n-dimensional vector of real number field, f (X) is object function It is the nonlinear function of a twice continuously differentiable, X is n-dimensional vector, its initial value is X (0);
(3) gradient class optimization method is chosen, methods described includes gradient descent method, Newton method and L-BFGS methods;According to The optimization method of selection, introduces Powerball functions, sets up Powerball iterative formulas, is iterated;It is described Powerball function expressions are σγ(z)=sign (z) | z |γ, γ ∈ (0, it is 1) Power coefficients, z ∈ R;
For gradient descent method, corresponding Powerball iterative formulas are:
For Newton method, corresponding Powerball iterative formulas are:
X (k+1)=X (k)-(▽2f(X(k)))-1σγ(▽f(X(k)));
For L-BFGS methods, corresponding Powerball iterative formulas are:
Wherein,It is that the Hesse matrixes of object function approach matrix, therefore with Hesse matrixes have identical dimension;Sk=X (k+1)-X (k), is the vector for having identical dimension with X (k);Yk=▽ f (X (k+1))-▽ f (X (k)),I.e.
Here, BkIt is that object function Hesse matrixes approach matrix, has same dimension with Hesse matrixes;
In formula, ▽ f (X) are the gradient of object function f (X);▽2F (X) is the Hesse matrixes of object function f (X);K is Iterations, value is 0,1,2 ..., αkStep-length during iteration secondary for kth, X (k) approaches value for what kth time iteration was obtained;When During k=0, BkInitial value is taken as unit matrix, and the initial value of X (k) can arbitrarily be chosen;σγ(·):R → R is Powerball function σγIt is right The nonlinear transformation of target function gradient is Powerball conversion, to arbitrary vector X=(x1,...,xn)T, Jing Powerball function σγNonlinear transformation, become σγ(X)=(σγ(x1),...,σγ(xn))T
(4) convergence is judged, concrete determination methods are as follows:
When object function is strong convex function, and its gradient meets L-Lipschitz conditions, that is, meet Lipschitz condition, And its Lipchitz coefficient is when being L, then differentiate that whether iterations is more than N;It is that iteration terminates, output optimal value X (k+1); Otherwise continue iteration;
When object function is not strong convex function, or its gradient is when being unsatisfactory for L-Lipschitz conditions, then judge | | X (k+ 1)-X (k) | | whether < ε set up, and are, iteration terminates, output optimal value X (k+1);Otherwise continue iteration;ε is error precision, Weighed according to required precision and amount of calculation;
Wherein, N is the default iterations upper limit.
Further, the L-Lipschitz conditions in the step (4) are:
To any X, Y ∈ Rn, function f (X) meets following formula:
||▽f(Y)-▽f(X)||≤L||Y-X||;| | | | any norm of vector is represented, L > 0 are Lipschitz Constant, is chosen for a upper bound of the norm of target function gradient;
The strong convex function refers to function f (X) to any X ∈ Rn,
It is convex function;
In above formula, | | | |2Then represent and take 2- norms;For the norms of vectorial X bis- square, m is normal more than zero Number.
Further, in the step (4),Can minimum iterations obtain Optimal value X (k);WhereinFor Liapunov Lyapunov letters Number;
Further, in the step (3), (0,1) value size is true according to error precision ε for Power coefficient gammas ∈ It is fixed;Error precision ε is bigger, and the value of γ is less, and convergence of algorithm speed is faster;
Further, the γ is selected according to iterations adaptive mode:
Wherein γ0, γ1The respectively initial value of γ and final value, its value mode is initial value γ0Better, final value γ closer to 01 Better the closer to 1, optimal value takes γ0=0.1, γ1=0.9, N are the iterations upper limit;Primary iteration and later stage iteration have Convergence rate faster.
Further, step (3) alternative manner is as follows:
(1) for an One-bit gradient descent method, Powerball function σ are madeγ(z)=sign (z) | z |γIn take γ =0;Now σγZ ()=sign (z), z ∈ R, iterative formula is:
Now, the symbol of each element is only needed in gradient calculation, to communication bandwidth when greatly reducing data conversion Requirement, so as to reduce communications cost optimized to strong convex function;
(2) when big data problem analysis is converted into into solving-optimizing problem, object function is sometimes some functions The form of sum, it is as follows
Wherein fi(X), j=1,2 ..., l be meet L-Lipschitz gradients, Second Order Continuous can be micro- strong convex function, So now consider the Powerball methods for stochastic variable, set up Powerball iterative formulas as follows;
WhereinRandomly select during iteration each time.
Further, in the step (4), m elects the absolute value of the minimal eigenvalue of the Hesse matrixes of object function as.
With regard to Lipschitz conditions, as long as giving a function meets Lipschitz conditions, it is possible to get Lipschitz constant L.L>0 is Lipschitz constants, is generally chosen for a upper bound of the norm of target function gradient.
Compared with prior art, the present invention in big data analysis for improving data analyzing speed and data analysis quality Data processing technique, solves emphatically the convergence of algorithm speed during some optimization problems related to big data are solved Upgrade Problem, it is considered to find the minimum of a value of object function, the initial value application alternative manner to giving is solved.By calculating in optimization The nonlinear Powerball functions with regard to gradient terms are introduced in method, Powerball is iterative for construction.Because Powerball changes For the primary iteration fast convergence rate of method, under big data application background, limited by effective computing resource, optimized algorithm Primary iteration efficiency it is just particularly important, Powerball methods are to improve the effective ways of big data optimization processing quality.For Traditional optimized algorithm, with the increase of data volume, optimized algorithm tends not to obtain its last convergence.Present invention utilizes calculating The strategy of method convergence in finite iteration step, it is middle in former optimization process to increase by one with regard to gradient terms containing the coefficient of impact Nonlinear function, in an iterative process the optional coefficient of impact be certain determine constant can also according to actual conditions be chosen for The variable element of iterations change, not only realizes and the effect of Fast Convergent is reached in finite iteration number of times, and gives tool The iterations upper limit of body, improves convergence of algorithm speed, it is to avoid existing algorithm unpredictable calculating time and amount of calculation Problem.Meanwhile, by carrying out deforming the Powerball methods for obtaining, such as stochastic gradient to other standards method Powerball methods, L-BFGS Powerball methods etc..Have the advantages that fast convergence rate, run time are short, can be fine Solution data analysis speed and quality.
Description of the drawings
Fig. 1 is the schematic flow sheet of the optimized algorithm method that the present invention improves rate of convergence;
Power coefficient gammas optimum value when Fig. 2 is assigned error precision ε;
Fig. 3 is to solve three data sets in example with gradient Powerball method (γ < 1) and gradient method (γ=1) Optimization problem result schematic diagram;
Fig. 4 is to solve three data in example with L-BGFS Powerball methods (γ < 1) and L-BGFS (γ=1) The optimization problem result schematic diagram of collection.
Specific embodiment
With reference to Figure of description and specific embodiment, the present invention is further elaborated.
Detailed description below is an instantiation of the present invention in terms of big data process, not to limit The use range of the present invention, every concrete case that can be converted into optimization problem can adopt this method.First according to above-mentioned Convergence is verified described in the content of the invention;Can be asked using this method if convergence.
(1) example is the non-immediate optimization problem of data processing, according to Fig. 1, can be micro- with second order What Logistic was returned- Regularization function as object function, according to given data pairThe target of this problem It is to solve for following minimum problems:
Now, rememberIts gradient meets Lipschitz conditions, It is also twice differentiable, then there are Solve problems i.e.And f (w) is apparent from for strong convex function, wherein w is to meet certain bar The parameter vector of part, for example, meet certain normal distribution.
Three data sets are chosen in embodiment 1, regularization is carried out to data set using formula (1.4), wherein choosing it respectively In to data set KDD10 and CTR choose λ=1, to data set RCV1 choose λ=0, standard basis as shown in table 1 below can be obtained Data set is classified:
Table 1
Wherein, RCV1 is Reuter's news category data set, and Reuter is first three big multimedia Inter-Press Service of the world, All kinds of news and finance data are provided.Can so utilize the data collected by it to be analyzed, such as it is all kinds of by analyzing News data, can be classified and be issued accurately, in time, efficiently to news, improve the promptness and height of the issue of news Effect property.KDD10 is to find and the data chosen in data mining contest that international intellectual finds and data from international intellectual in 2010 It is the most influential race of current Data Mining to excavate contest.The match is converged simultaneously towards business circles and academia World data is excavated top expert, scholar, engineer, the student on boundary etc. and is participated in, and by contest, is data mining practitioner The ideal place that shows there is provided an academic exchange and achievement in research.Match topic over the years takes from different excavation applications, And have a very strong application background.CTR is the set obtained by being sampled to ad click rate data, and ad click rate is estimated Critically important role is played in accurate advertisement launch process, income, gray receipts of the accuracy estimated to advertiser The friendly experience of benefit and user has great impact, therefore suffers from the extensive concern of Internet enterprises.
(2) by given initial value, here the initial value is to randomly select, and according to following classical gradient descent method, is obtained To the iterative formula for solving the optimization problem in (1)
Powerball function σ are introduced in above formulaγ(z)=sign (z) | z |γ, obtain such as Gradient Powerball side Method,
And it is 1,0.7,0.4,0.1 to take the value of selection γ respectively, wherein when γ=1, as former iterative formula itself is entered Row iteration solves the optimization problem in (1).In solution procedure, first, with the backtracking linear search method of standard step-length is chosen; Secondly as w is parameter vector, the initial value w of the selection weight coefficient w random as k=00, initial vector w0In component can Elect the stochastic variable of Normal Distribution N (0,0.01) as;Finally, repetition test 10 times and the average of experimental result is taken as most Whole result of the test;Obtain result as shown in figure 3, wherein the corresponding data set of left hand view be RCV1, the corresponding data of middle graph Integrate as KDD10, the corresponding data set of right part of flg is CTR.Assigned error precision ε=10-3When, changing the value of γ, such as γ divides Do not take 1,0.7,0.4,0.1, wherein γ=1 when, be the gradient method of standard.From the figure 3, it may be seen that introducing nonlinear factor Powerball methods accelerate the convergence rate of conventional method;And when γ gets over hour, Powerball convergence of algorithm speed is got over Hurry up;
(3) it is same by any given initial value, choose following L-BFGS methods
Wherein, BkSk=Yk Sk=w (k+1)-w (k), Yk=▽ f (w (k+1))-▽ f (w (k)),Hk+1= Hk+ΔHk.Powerball functions are introduced in above formula, following corresponding L-BFGS Powerball methods are obtained
Wherein
Solution is iterated to the optimization problem in (1) using said method.In solution procedure, first, with standard Backtracking linear search method choose step-length;Secondly, according to normal distribution N (0,0.01) weight coefficient w is initialized;Most Afterwards, repetition test 10 times and the average of experimental result is taken as final result of the test;Result is obtained as shown in figure 4, wherein left Figure corresponding data set in side is RCV1, and the corresponding data set of middle graph is News20, and the corresponding data set of right part of flg is CTR.Give Determine error precision ε=10-3When, change the value of γ, such as γ takes respectively 1,0.7,0.4,0.1, wherein γ=1 when, be standard L-BFGS methods.As shown in Figure 4, the Powerball methods for introducing nonlinear factor accelerate the convergence rate of conventional method;And When γ gets over hour, Powerball convergence of algorithm speed is faster.
It can be seen that, method proposed by the present invention can be realized restraining faster in finite iteration number of times, and this is in big data process In have stronger advantage.On the other hand, this method can in advance set the iterations for meeting required precision, can save operation Time, saving memory space.
As it will be easily appreciated by one skilled in the art that the foregoing is only presently preferred embodiments of the present invention, not to The present invention, all any modification, equivalent and improvement made within the spirit and principles in the present invention etc. are limited, all should be included Within protection scope of the present invention.

Claims (7)

1. it is a kind of to improve the method that big data processes quality, it is characterised in that to comprise the steps:
(1) according to the data characteristic collected, it is analyzed to optimize data:If whether data processing problem is to solve for letter Several minimum of a value optimization problems:It is to go to step (2);Otherwise, by the Regularization to data, it is converted into solution minimum of a value Optimization problem, goes to step (2);
(2) minimum of a value Optimized model is set upWherein RnFor the n-dimensional vector of real number field, f (X) is object function, is one The nonlinear function of individual twice continuously differentiable, X is n-dimensional vector;
(3) gradient class optimization method is chosen, methods described includes gradient descent method, Newton method and L-BFGS methods;With specific reference to The optimization method of selection, introduces Powerball functions, sets up Powerball iterative formulas, is iterated;The Powerball Function expression σγ(z)=sign (z) | z |γ, γ ∈ (0, it is 1) Power coefficients, z ∈ R;
For gradient descent method, corresponding Powerball iterative formulas are:
For Newton method, corresponding Powerball iterative formulas are:
X (k+1)=X (k)-(▽2f(X(k)))-1σγ(▽f(X(k)));
For L-BFGS methods, corresponding Powerball iterative formulas are:
Wherein,It is that the Hesse matrixes of object function approach matrix, with Hesse Matrix has identical dimension;Sk=X (k+1)-X (k), is the vector for having identical dimension with X (k);X (k+1)=X (k)+αkdk,Yk=▽ f (X (k+1))-▽ f (X (k)),I.e.
In formula, ▽ f (X) are the gradient of object function f (X);▽2F (X) is the Hesse matrixes of object function f (X);K is iteration time Number, value is 0,1,2 ... ..., αkStep-length during iteration secondary for kth, X (k) approaches value for what kth time iteration was obtained;Work as k=0 When, BkInitial value is taken as unit matrix, and the initial value X (0) of X (k) can arbitrarily choose;σγ(·):R → R is Powerball function σγIt is right The nonlinear transformation of target function gradient is Powerball conversion;
(4) convergence is judged, concrete determination methods are as follows:
When object function is strong convex function, and its gradient is when meeting L-Lipschitz conditions, then differentiate whether iterations is more than N;It is that iteration terminates, output optimal value X (k+1);Otherwise continue iteration;
When object function is not strong convex function, or its gradient is when being unsatisfactory for L-Lipschitz conditions, then judge | | X (k+1)-X (k) | | whether < ε set up, and are, iteration terminates, output optimal value X (k+1);Otherwise continue iteration;ε is error precision, according to Required precision and amount of calculation are weighed;
Wherein, N is the default iterations upper limit.
2. the method that big data processes quality is improved according to claim 1, it is characterised in that the L- in the step (4) Lipschitz conditions are:
To any X, Y ∈ Rn, function f (X) meets following formula:
||▽f(Y)-▽f(X)||≤L||Y-X||;| | | | any norm of vector is represented, L > 0 are Lipschitz constants, It is chosen for a upper bound of the norm of target function gradient;
The strong convex function refers to function f (X) to any X ∈ Rn,
It is convex function;
In above formula, | | | |2Then represent and take 2- norms;For the norms of vectorial X bis- square, m is the constant more than zero.
3. the method that big data processes quality is improved according to claim 2, it is characterised in that described In step (4),Can minimum iterations obtain optimal value X (k);WhereinFor Lyapunov functions.
4. the method that big data processes quality is improved according to claim 1, it is characterised in that described in the step (3) (0,1) value size determines Power coefficient gammas ∈ according to error precision ε;Error precision ε is bigger, and the value of γ is less, algorithm Convergence rate it is faster.
5. the method that big data processes quality is improved according to claim 1, it is characterised in that the γ is according to iterations Adaptive mode is selected:
Wherein γ0, γ1The respectively initial value of γ and final value, optimal value takes γ0=0.1, γ1=0.9, N are the iterations upper limit.
6. the method that big data processes quality is improved according to claim 1, it is characterised in that step (3) alternative manner is such as Under:
(1) for One-bit gradient descent methods, Powerball function σ are madeγ(z)=sign (z) | z |γIn take γ=0;Now σγZ ()=sign (z), z ∈ R, iterative formula is:
(2) for the Powerball methods of stochastic variable, Powerball iterative formulas are set up as follows;
In formula,Wherein fj(X), j=1,2 ..., l is to meet L-Lipschitz gradients, Second Order Continuous Strong convex function that can be micro-, whereinRandomly select during iteration each time.
7. the method that big data processes quality is improved according to Claims 2 or 3, it is characterised in that in the step (4), m Elect the absolute value of the minimal eigenvalue of the Hesse matrixes of object function as.
CN201611232063.5A 2016-12-28 2016-12-28 Method for improving processing quality of big data Pending CN106649203A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201611232063.5A CN106649203A (en) 2016-12-28 2016-12-28 Method for improving processing quality of big data

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201611232063.5A CN106649203A (en) 2016-12-28 2016-12-28 Method for improving processing quality of big data

Publications (1)

Publication Number Publication Date
CN106649203A true CN106649203A (en) 2017-05-10

Family

ID=58831912

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201611232063.5A Pending CN106649203A (en) 2016-12-28 2016-12-28 Method for improving processing quality of big data

Country Status (1)

Country Link
CN (1) CN106649203A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108510118A (en) * 2018-04-02 2018-09-07 张龙 A kind of building heating energy forecast analysis terminal based on Internet of Things
CN109033021A (en) * 2018-07-20 2018-12-18 华南理工大学 A kind of linear equation solver design method for joining convergence neural network based on change

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108510118A (en) * 2018-04-02 2018-09-07 张龙 A kind of building heating energy forecast analysis terminal based on Internet of Things
CN109033021A (en) * 2018-07-20 2018-12-18 华南理工大学 A kind of linear equation solver design method for joining convergence neural network based on change
CN109033021B (en) * 2018-07-20 2021-07-20 华南理工大学 Design method of linear equation solver based on variable parameter convergence neural network

Similar Documents

Publication Publication Date Title
CN111242282B (en) Deep learning model training acceleration method based on end edge cloud cooperation
Vicini et al. Inverse and direct airfoil design using a multiobjective genetic algorithm
CN111696345A (en) Intelligent coupled large-scale data flow width learning rapid prediction algorithm based on network community detection and GCN
CN107886161A (en) A kind of global sensitivity analysis method for improving Complex Information System efficiency
Zhou et al. Movie recommendation system employing the user-based cf in cloud computing
Davis et al. Grids versus graphs: Partitioning space for improved taxi demand-supply forecasts
Dieu et al. Augmented Lagrange Hopfield network initialized by quadratic programming for economic dispatch with piecewise quadratic cost functions and prohibited zones
Ardil Military Combat Aircraft Selection Using Trapezoidal Fuzzy Numbers with the Technique for Order of Preference by Similarity to Ideal Solution (TOPSIS)
CN105389291B (en) Data processing method based on the convex local Non-negative Matrix Factorization of increment type
Liu et al. EACP: An effective automatic channel pruning for neural networks
CN111898316A (en) Construction method and application of super-surface structure design model
Wang et al. A data‐driven probabilistic power flow method based on convolutional neural networks
CN108459993A (en) Based on the complicated High Dimensional Systems optimization method for quickly chasing after peak sampling
Sun et al. Damped Anderson mixing for deep reinforcement learning: Acceleration, convergence, and stabilization
CN106649203A (en) Method for improving processing quality of big data
CN114564787A (en) Bayesian optimization method, device and storage medium for target-related airfoil design
Guo et al. Skewed normal cloud modified whale optimization algorithm for degree reduction of S-λ curves
CN116502779A (en) Traveling merchant problem generation type solving method based on local attention mechanism
CN116054144A (en) Distribution network reconstruction method, system and storage medium for distributed photovoltaic access
CN112991402B (en) Wen Wudian cloud registration method and system based on improved differential evolution algorithm
Wang et al. COSA: Co-Operative Systolic Arrays for Multi-head Attention Mechanism in Neural Network using Hybrid Data Reuse and Fusion Methodologies
Liu et al. Learning-assisted intelligent risk assessment of highway project investment
Xu et al. A time series combined forecasting model based on prophet-lgbm
Li et al. Finding the Nash equilibria of $ n $-person noncooperative games via solving the system of equations
Held et al. Multiobjective Metamodel-Based Design Optimization—A Review and Classification Approach Using the Example of Engine Development

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20170510