CN113077853B - Global optimization method and system for mechanical parameters of double loss value network deep reinforcement learning KVFD model - Google Patents

Global optimization method and system for mechanical parameters of double loss value network deep reinforcement learning KVFD model Download PDF

Info

Publication number
CN113077853B
CN113077853B CN202110368257.2A CN202110368257A CN113077853B CN 113077853 B CN113077853 B CN 113077853B CN 202110368257 A CN202110368257 A CN 202110368257A CN 113077853 B CN113077853 B CN 113077853B
Authority
CN
China
Prior art keywords
parameter
value
curve
network
global
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110368257.2A
Other languages
Chinese (zh)
Other versions
CN113077853A (en
Inventor
张红梅
周衍
王凯
李文彬
张可浩
王炯
万明习
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian Jiaotong University
Original Assignee
Xian Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian Jiaotong University filed Critical Xian Jiaotong University
Priority to CN202110368257.2A priority Critical patent/CN113077853B/en
Publication of CN113077853A publication Critical patent/CN113077853A/en
Application granted granted Critical
Publication of CN113077853B publication Critical patent/CN113077853B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16CCOMPUTATIONAL CHEMISTRY; CHEMOINFORMATICS; COMPUTATIONAL MATERIALS SCIENCE
    • G16C60/00Computational materials science, i.e. ICT specially adapted for investigating the physical or chemical properties of materials or phenomena associated with their design, synthesis, processing, characterisation or utilisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F30/00Computer-aided design [CAD]
    • G06F30/20Design optimisation, verification or simulation
    • G06F30/27Design optimisation, verification or simulation using machine learning, e.g. artificial intelligence, neural networks, support vector machines [SVM] or training a model
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2111/00Details relating to CAD techniques
    • G06F2111/14Details relating to CAD techniques related to nanotechnology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2119/00Details relating to the type or aim of the analysis or the optimisation
    • G06F2119/14Force analysis or force optimisation, e.g. static or dynamic forces

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • Evolutionary Computation (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • Molecular Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Medical Informatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Geometry (AREA)
  • Computer Hardware Design (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The application discloses a global optimization method and a global optimization system for mechanical parameters of a double loss value network deep reinforcement learning KVFD model, wherein the method comprises the following steps: s1, inputting a pre-acquired nano indentation measurement curve into a trained predicted value acquisition network to obtain a parameter predicted value of the nano indentation measurement curve; s2, iterating the parameter predicted value as an iteration initial value of a deep reinforcement learning algorithm to obtain an approximation of a global parameter solution of the pre-acquired nano indentation measurement curve; and when the approximation of the global parameter solution reaches a preset convergence condition, outputting the approximation of the global parameter solution as a mechanical parameter of the KJFD model. According to the method, the parameter predicted value is introduced into iteration to conduct parameter guidance, and the global optimal solution can be well approximated.

Description

Global optimization method and system for mechanical parameters of double loss value network deep reinforcement learning KVFD model
Technical Field
The application belongs to the technical field of mechanical parameters of nano indentation measurement data, relates to the field of KVFD model multi-parameter function fitting and global parameter approximation, and particularly relates to a method and a system for global optimization of mechanical parameters of a double loss value network deep reinforcement learning KVFD model.
Background
At present, in the process of acquiring mechanical parameters of a measured material through nano indentation measurement data, a least square method is mostly adopted for simple function fitting, and the function parameters are sequentially and iteratively adjusted to reduce the minimum mean square error between a fitted curve and a real curve. The method is fast and effective for fitting simple functions, but often does not perform well for fitting complex functions and multi-parameter functions.
For multi-parameter optimization of a KVFD model complex function, a common greedy algorithm, a gradient descent algorithm and a simulated annealing algorithm can not obtain a better global optimal solution; the greedy algorithm and the gradient descent algorithm can find a local optimal solution near a given initial parameter, have certain applicability to complex function multi-parameter optimization, but have a great relation with the given initial value if a global parameter is found, and are difficult to approach the global parameter; the simulated annealing algorithm adopts a certain probability to accept a new parameter solution, has the capability of jumping out of a local optimal trap, shows better optimizing capability in the optimizing problem of a plurality of local optimal solutions, has a certain probability to find a solution near a global parameter solution, but the probability characteristic of the simulated annealing algorithm also causes that the simulated annealing algorithm cannot approach the global parameter solution every time, and has poor reliability.
In summary, it is difficult to effectively approach the global optimal solution to the problem of multi-parameter optimization of the complex function of the kfd model at present.
Disclosure of Invention
The application aims to provide a global optimization method and a global optimization system for mechanical parameters of a double loss value network deep reinforcement learning KVFD model, which are used for solving one or more of the technical problems. According to the method, the parameter predicted value is introduced into iteration to conduct parameter guidance, and the global optimal solution can be well approximated.
In order to achieve the above purpose, the application adopts the following technical scheme:
the application discloses a global optimization method for mechanical parameters of a double loss value network deep reinforcement learning KVFD model, which comprises the following steps:
s1, inputting a pre-acquired nano indentation measurement curve into a trained predicted value acquisition network to obtain a parameter predicted value of the nano indentation measurement curve; the trained predicted value acquisition network is a cyclic neural network based on an LSTM hidden layer, and LOSS function values used in the cyclic neural network are calculated by curve-curve corresponding parameters of an input network and parameter-parameter corresponding curves of network output;
s2, iterating the parameter predicted value as an iteration initial value of a deep reinforcement learning algorithm to obtain an approximation of a global parameter solution of the pre-acquired nano indentation measurement curve; the reward value prediction network of the deep reinforcement learning algorithm gives out the reward value when the current parameter changes to different directions through the difference value of the corresponding curve and the real curve of the current iteration parameter, and guides the parameter to approach to the global parameter;
and when the approximation of the global parameter solution reaches a preset convergence condition, outputting the approximation of the global parameter solution as a mechanical parameter of the KJFD model.
A further development of the application consists in that in step S1, the pre-acquired nanoindentation measurement profile comprises a time sequence, a stress sequence and an indentation depth sequence.
A further improvement of the present application is that, in step S1, the predicted value acquisition network includes: a plurality of LSTM hidden layers and a DNN network;
the unit numbers of each of the LSTM hidden layers are fixed and consistent, and each LSTM hidden layer is connected in a point-to-point mode; the first LSTM hidden layer inputs a pre-acquired nanoindentation measurement curve, and the last LSTM hidden layer output value enters a DNN network;
the DNN network comprises a plurality of full-connection layers and convolution layers with different dimensions, and the full-connection layers and the convolution layers are used for converting the value output by the last LSTM hidden layer into a parameter predicted value to be output.
A further improvement of the application is that, in step S1, the expression of the LOSS function value is calculated as,
wherein L is p Partial calculation of tag parameter θ train And network output parametersBetween loss values, L d Partial calculation curve D train Curve corresponding to network output parameters ∈ ->Between loss values, w p 、w d Respectively is L p 、L d Weight of the two parts.
A further improvement of the present application is that, in step S2, the reward value prediction network of the deep reinforcement learning algorithm includes: a plurality of LSTM hidden layers and a DNN network;
the unit numbers of each of the LSTM hidden layers are fixed and consistent, and each LSTM hidden layer is connected in a point-to-point mode; the first LSTM hidden layer inputs a difference value obtained by subtracting a pre-acquired nanoindentation measurement curve from a current iteration curve, and the last LSTM hidden layer output value enters a DNN network;
the DNN network comprises a plurality of fully connected layers and convolution layers with different dimensions, and the fully connected layers and the convolution layers are used for converting the value output by the last LSTM hidden layer into rewards prediction for actions in all directions.
A further improvement of the present application is that in step S2, the LOSS function used in the network training is the sum of absolute errors of the label rewards value vector and the rewards value vector output by the network.
In step S2, the iteration is performed by using the parameter predicted value obtained in step S1 as an iteration initial value of the deep reinforcement learning algorithm, and in the process of obtaining an approximation of the global parameter solution of the pre-obtained nanoindentation measurement curve, the specific steps of each iteration include:
(1) Respectively predicting the reward value of the candidate parameter set of the current iteration parameter by using a reward evaluation rule and the reward value prediction network, and taking the weighted sum of the reward evaluation rule and the reward value prediction network as the reward evaluation of the candidate parameter set of the current iteration parameter;
the reward evaluation rule is that for evaluating a certain alternative parameter, firstly calculating the curve difference delta between the corresponding curve of the alternative parameter and the pre-acquired nano indentation measurement curve, and then calculating the absolute average value of the curve difference
The evaluation formula of the prize value r is expressed as:
(2) And (3) calculating a new row of the Q table according to the reward evaluation obtained in the step (1) and the content of the current row of the Q table in the deep reinforcement learning algorithm, finding the maximum value in the new row of the Q table, and taking the corresponding alternative parameter as the current iteration result parameter.
The application further improves that in the step S2, the specific step of judging whether the approximation of the global parameter solution reaches the preset convergence condition comprises the following steps: stopping iteration when the error between the corresponding curve of the current iteration result parameter and the pre-acquired nanoindentation measurement curve is smaller than a certain preset value; or stopping iteration when the iteration times reach a preset value.
The application discloses a mechanical parameter extraction system of a nano indentation measurement curve based on a KVFD model, which comprises the following components:
the parameter prediction value acquisition module is used for inputting the pre-acquired nano indentation measurement curve into a trained prediction value acquisition network to acquire a parameter prediction value of the nano indentation measurement curve; the trained predicted value acquisition network is a cyclic neural network based on an LSTM hidden layer, and LOSS function values used in the cyclic neural network are calculated by curve-curve corresponding parameters of an input network and parameter-parameter corresponding curves of network output;
the depth reinforcement learning iteration output module is used for iterating the obtained parameter predicted value as an iteration initial value of a depth reinforcement learning algorithm to obtain an approximation of a global parameter solution of the pre-obtained nano indentation measurement curve; the reward value prediction network of the deep reinforcement learning algorithm gives out the reward value when the current parameter changes to different directions through the difference value of the corresponding curve and the real curve of the current iteration parameter, and guides the parameter to approach to the global parameter; and when the approximation of the global parameter solution reaches a preset convergence condition, outputting the approximation of the global parameter solution as a mechanical parameter of the KJFD model.
Compared with the prior art, the application has the following beneficial effects:
the application provides a hierarchical depth reinforcement learning strategy for multi-parameter optimization of a complex function, which can solve the problem that the existing multi-parameter optimization of the complex function is difficult to effectively approach to a global parameter solution. The method can more stably find the approach global parameter solution in the multi-parameter optimization of the complex function.
In the application, an integral algorithm is constructed based on deep reinforcement learning (DQN), iteration initial parameters are given by a deep-loop neural network (I-LSTM), and the network is also participated in guiding training by global parameters during training; the reward evaluation of the iteration parameter is given by another deep-loop neural network (R-LSTM) participation, and the network participates in guiding training by the global parameter during training; the design enables the nano indentation measurement curve based on the KVFD model to introduce parameter guidance in the fitting process, rather than guiding adjustment of fitting parameters and curves by curve errors, so that the method can better approximate to a global parameter solution in the complex function multi-parameter optimization problem.
In the application, the guidance of the global parameter on the multi-parameter optimization of the complex function is introduced by designing the neural network (namely R-LSTM) in the deep reinforcement learning (DQN), so that the application has better capability of approaching the global parameter than the prior art. By adding a neural network (i.e. I-LSTM) for initial parameter guidance and adding calculation of parameter errors in a loss function, initial iteration parameters of deep reinforcement learning (DQN) can be close to a global parameter solution, the capacity of the strategy of the application for approaching the global parameters in the complex function multi-parameter optimization problem is enhanced, the iteration times of the deep reinforcement learning (DQN) can be effectively reduced, and the operation speed is increased.
In the application, by adjusting the weight w p 、w d The importance degree of the I-LSTM on the parameter or the curve can be adjusted; by adjusting L p 、L d The specific calculation mode of the method can adapt the prediction capability of the I-LSTM to curve equations with different characteristics, and a good prediction effect is achieved.
Drawings
In order to more clearly illustrate the embodiments of the present application or the technical solutions in the prior art, the following description of the embodiments or the drawings used in the description of the prior art will make a brief description; it will be apparent to those of ordinary skill in the art that the drawings in the following description are of some embodiments of the application and that other drawings may be derived from them without undue effort.
FIG. 1 is a schematic flow chart of a method for extracting mechanical parameters of nano indentation measurement data based on a KVFD model according to an embodiment of the present application;
FIG. 2 is a schematic diagram of a Initial LSTM Network (I-LSTM) structure for providing initial parameter prediction for a QL portion in an embodiment of the application;
FIG. 3 is a schematic diagram of a Reward LSTM Network (R-LSTM) structure for providing reward evaluation prediction for the QL portion and a method of generating its training data set in an embodiment of the application;
FIG. 4 is a schematic diagram of a mechanical parameter extraction system based on nano-indentation measurement data of a KJFD model according to an embodiment of the present application;
FIG. 5 is a graph showing the results of experimental curve fitting in an embodiment of the present application;
FIG. 6 is a schematic diagram showing the comparison of simulation parameters and fitting parameters in an embodiment of the present application; wherein (a) in fig. 6 is a comparative schematic diagram of parameters representing elasticity, (b) in fig. 6 is a comparative schematic diagram of parameters representing fluidity, and (c) in fig. 6 is a comparative schematic diagram of parameters representing tackiness;
FIG. 7 is a graph showing experimental curve fitting error distribution in an embodiment of the present application; fig. 7 (a) is a schematic diagram of a scatter point of the error distribution, and fig. 7 (b) is a schematic diagram of a histogram of the error distribution.
Detailed Description
In order to make the purposes, technical effects and technical solutions of the embodiments of the present application more clear, the technical solutions of the embodiments of the present application are clearly and completely described below with reference to the accompanying drawings in the embodiments of the present application; it will be apparent that the described embodiments are some of the embodiments of the present application. Other embodiments, which may be made by those of ordinary skill in the art based on the disclosed embodiments without undue burden, are within the scope of the present application.
Based on the existing complex function multi-parameter optimization strategy, the global parameter solution is difficult to be found or approximated effectively and stably. The application discloses a strategy for fitting a multi-parameter complex function and approaching a global parameter based on Deep Q Network (DQN). The application uses deep reinforcement learning (DQN) to iterate optimization, and the Reward value of the DQN is predicted by a custom Reward evaluation rule (Reward Evaluation Rules, RER) and a Reward LSTM network (Reward-LSTM, R-LSTM); the Initial parameters of the first iteration are given by a custom Initial-LSTM network (I-LSTM) prediction. The LSTM is adopted to build a network, and aims to process time sequence data with a large number of points, namely a multi-parameter complex function curve.
Referring to FIG. 1, the present application designs an algorithm (Curve-fit-DQN, CF-DQN) for complex function multi-parameter Fitting, finding a solution to the global parameter approximation. The design framework of CF-DQN is shown in FIG. 1, and is input as curve D of the solution of the parameters to be found real Output as predicted result parameter θ out The method takes a Q-Learning algorithm (QL) as a skeleton, predicts initial parameters for the Q-Learning algorithm by 1 LSTM network (Initial LSTM Network, I-LSTM), and evaluates prize values for alternative parameters in the Q-Learning algorithm by another LSTM network (Reward LSTM Network, R-LSTM). AlgorithmAnd the R-LSTM in the model gives out a reward value when the current parameter changes to different directions through the difference value between the corresponding curve and the real curve of the current iteration parameter, and guides the parameter to approach to the global parameter.
In the embodiment of the application, for convenience of description, a multi-parameter complex function f is defined, and a curve D of parameters to be extracted is defined real ,D real The corresponding ideal global parameter is theta real D is real =f(θ real ) The CF-DQN output result is θ out It corresponds to curve D out Let the output of I-LSTM be θ I-LSTM The action set of Q-Learning is A= { a 1 ,a 2 ,......,a n The ith row vector of the Q table isThe algorithm operation mainly comprises 2 stages, namely a parameter guiding stage and an iterative optimizing stage:
(1) The parameter boot phase is completed by the I-LSTM network: curve D of the parameter to be extracted real Inputting I-LSTM network to obtain a group of parameters theta I-LSTM And inputs it into QL as an initial parameter of QL;
(2) The skeleton of the iterative optimization stage is QL, and meanwhile, the R-LSTM network participates in the evaluation rewarding link in the QL: at the time of iteration to the ith step, the current parameter theta is jointly determined by adopting a reward value evaluation rule (Reward Evaluation Rules, RER) and R-LSTM i And evaluating the rewarding value by the parameter set obtained after the action set A is executed. Based on the obtained prize value and the current row content Q of the Q table i Calculate a new line Q i+1 Selecting the maximum valueAction a of kth item of action set A is then taken k The parameter is defined by theta i Updated to theta i+1 And (5) completing one-time iterative optimization. Stopping iteration when the maximum number of iterative optimization or curve error is smaller than a preset value, wherein the obtained parameter is the final parameter theta out
Referring to FIG. 2, in an embodiment of the present application, the CF-DQN algorithmThe first step is to provide an initial iteration value for Q-Learning via Initial LSTM Network (I-LSTM). The input of the I-LSTM is a one-dimensional curve D of the parameters to be extracted, and the number of curve points is fixed to be m; the number of units of each layer is fixed and consistent, the number of units is similar to the number m of curve points when the number of units is selected, the first LSTM hidden layer receives input data D, so that the input dimension of the first LSTM layer is m, the output dimension is a set number of units, the input dimension and the output dimension of each subsequent LSTM layer are set numbers of units, and each LSTM layer is connected in a point-to-point mode; followed by a DNN network which may contain multiple fully-connected layers and convolutional layers of different dimensions for converting the output of the LSTM layer into an output parameter θ I-LSTM
In the embodiment of the application, in order to introduce the guidance of global parameters on the global parameters, a loss function L is designed I-LSTM From L p 、L d The two parts are respectively weighted by w p 、w d The weight of the composition,
wherein L is p Partial calculation of tag parameter θ train And network output parametersBetween loss values, L d Partial calculation curve D train Curve corresponding to network output parameters ∈ ->The value of loss between the two is adjusted by adjusting the weight w p 、w d We can adjust the importance of I-LSTM to parameters or curves; by adjusting L p 、L d The specific calculation mode of the method can adapt the prediction capability of the I-LSTM to curve equations with different characteristics, and a good prediction effect is achieved.
Referring to fig. 3, in the embodiment of the present application, the network structure of R-LSTM is similar to that of I-LSTM, and LSTM hidden layer and DNN network are adopted, the input is the point-by-point difference result of two curves, and the output is the reward prediction for motion in each direction.
For convenience of description, the following symbols are defined: the parameter change optional action set is a= { a 1 ,a 2 ,......,a n The curve of the parameter to be extracted is D real Corresponding real parameter is theta real The current reference parameter is theta now The corresponding curve is D now . Parameter θ now The parameter set obtained after the action set A isTheir corresponding curve sets areCurve D now And D real Is delta as the difference of train Curve set->Respectively with D real Is { delta } 1 ,Δ 2 ,......,Δ n }. Using Reward Evaluation Rules (RER) pair parameter setsThe rewards obtained by each parameter evaluation are respectively r 1 ,r 2 ,... rn, let R train =[r 1 ,r 2 ,......,r n ]. Will delta train The result obtained by inputting R-LSTM is +.>
The algorithm flow for the R-LSTM training phase is shown in FIG. 3. According to the real parameter theta real For the current reference parameter theta now The parameter set obtained after the action set A isAnd evaluating the rewards. The network memorizes the alternative parameters corresponding to the current reference parameters and the difference between the current reference parameter corresponding curve and the real curveThe relation between the prize rating of the number sets ". The trained R-LSTM has indirect guidance of global parameters on parameter optimization, and the global parameter guidance is introduced into the algorithm, so that the algorithm can have the capacity of approaching the global parameters.
Each set of data in the training set of R-LSTM comprises: a vector delta of length m train Is the corresponding curve D of the current reference parameter now And the parameter curve D to be extracted real A result of making a difference point by point; vector R of length n train Representing the current reference parameter theta now After executing the corresponding action set, the parameters can obtain the rewarding value; unlike I-LSTM, the loss function L employed in R-LSTM training R-LSTM For the sum of absolute errors of vectors of prize values, i.e
When Q-Learning iterates, the reward evaluation of the candidate parameter set comes from 2 aspects, one is a reward evaluation rule (Reward Evaluation Rules, RER), the other is a reward value predicted by R-LSTM, and the reward evaluation of the candidate parameter set is obtained by weighting the two, so that the parameter iteration is guided. And ending the iteration when the iteration is performed for a certain number of times or the error between the fitted curve and the real curve is smaller than a certain value.
In the application, the guidance of the global parameter on the multi-parameter optimization of the complex function is introduced by designing the neural network (namely R-LSTM) in the deep reinforcement learning (DQN), so that the application has better capability of approaching the global parameter than the prior art. By adding a neural network (i.e. I-LSTM) for initial parameter guidance and adding calculation of parameter errors in a loss function, initial iteration parameters of deep reinforcement learning (DQN) can be close to a global parameter solution, the capacity of the strategy of the application for approaching the global parameters in the complex function multi-parameter optimization problem is enhanced, the iteration times of the deep reinforcement learning (DQN) can be effectively reduced, and the operation speed is increased.
Referring to fig. 4, a mechanical parameter extraction system of a nanoindentation measurement curve based on a kfd model according to an embodiment of the present application includes:
the parameter prediction value acquisition module is used for inputting the pre-acquired nano indentation measurement curve into a trained prediction value acquisition network to acquire a parameter prediction value of the nano indentation measurement curve; the trained predicted value acquisition network is a cyclic neural network based on an LSTM hidden layer, and LOSS function values used in the cyclic neural network are calculated by curve-curve corresponding parameters of an input network and parameter-parameter corresponding curves of network output;
the depth reinforcement learning iteration output module is used for iterating the obtained parameter predicted value as an iteration initial value of a depth reinforcement learning algorithm to obtain an approximation of a global parameter solution of the pre-obtained nano indentation measurement curve; the reward value prediction network of the deep reinforcement learning algorithm gives out the reward value when the current parameter changes to different directions through the difference value of the corresponding curve and the real curve of the current iteration parameter, and guides the parameter to approach to the global parameter; and when the approximation of the global parameter solution reaches a preset convergence condition, outputting the approximation of the global parameter solution as a mechanical parameter of the KJFD model.
DETAILED DESCRIPTION OF EMBODIMENT (S) OF INVENTION
In the embodiment of the application, a round probe nanoindentation equation (hereinafter referred to as KJFD equation) of a KJFD model is selected, and the strategy of the application is implemented specifically. The equation is divided into three loading protocols (hereinafter, relaxation, load-unloading and creep) of Ramp-relaxation, load-unloading and Ramp-creep. The system of equations is shown in table 1.
Table 1. Nanoindentation equation of kvfd model under circular probe
Wherein the mechanical parameter to be optimized is [ E ] 0 ,α,τ]. R is the radius of the probe, v is the increasing speed of the indentation depth during loading, k is the increasing speed of the pressure during loading, T r For turning time, Γ (·) is a gamma function,for the complete beta function +.>As an incomplete beta function ++>Is a Mittag-Leffer (M-L) function.
Construction of CF-DQN algorithm example for KVFD equation 3 loading protocols. Assuming the KVFD equation, the probe radius r=8.5 μm=8.5×10 -6 m. Other conditions under the 3 loading protocols in the formula are set as follows:
(1) Under the relationship loading protocol: turning time T r Time of retention T =2s hold =3s, maximum indentation depth of 5 μm=5×10 -6 m, the indentation depth increasing speed v=2.5 μm/s=2.5×10 in the pressing stage -6 m/s;
(2) Under the load-unload loading protocol: turning time T r =25 s, the maximum indentation depth is 5 μm=5×10 - 6 m, the indentation depth increasing speed v=0.2 μm/s=0.2×10 in the pressing stage -6 m/s;
(3) Under the creep loading protocol: turning time T r Time of retention T =2s hold =3s, maximum pressure of 5 μn=5×10 -6 N, the pressure increase rate in the pressing stage k=2.5 μn/s=2.5×10 -6 N/s。
When the CF-DQN algorithm is implemented, a set of parameters is set as θ= [ E ] 0 ,α,τ]The parameter range is E 0 ∈[10,100000],α∈[0.01,0.99],τ∈[1,1000]The method comprises the steps of carrying out a first treatment on the surface of the Setting the point number of the curve D corresponding to the parameter theta to be m=250; the time starting point of the curve is set to t=0s, the sampling rate under the correlation loading protocol is 0.02 seconds, the sampling rate under the load-unload loading protocol is 0.1 seconds, and the sampling rate under the creep loading protocol is 0.02 seconds. Setting the equations corresponding to the relaxation, load-load and creep loading protocols as f respectively r 、f u 、f c Wherein f corresponding to the relationship loading protocol r The following are listed below
load-unload load protocol corresponds to f u T is not less than 0 but not more than T r Not using T > T r Is provided with an unloading part of (a)
F corresponding to creep loading protocol c The following are listed below
Note that for 3 loading protocols, it is precisely for 3 different complex curve equations, so that 3 different CF-DQN instances need to be implemented separately.
Action set a= { a of parameter change 1 ,a 2 ,......,a n (ii) contains 8 elements, i.e. n=8, see table 2 for details. The symbol "≡" represents that the parameter item increases in steps, and the symbol "≡" represents that the parameter item decreases in steps. The step size of each parameter item is 1% of the current parameter value.
TABLE 2 parameter Change action set A set for CF-DQN Algorithm implementation
The number of the 3-layer LSTM hidden layer nodes in the I-LSTM is set to 256 nodes, the last layer LSTM hidden layer is followed by a layer of 256-node fully-connected layer, the output of the fully-connected layer is 3 nodes, and the fully-connected layer corresponds to standardized E respectively 0 α, τ parameters. In training I-LSTM, we use a training dataset that contains one million.
The number of nodes of a 3-layer LSTM hidden layer in the R-LSTM is set to 256 nodes, a full-connection layer of 256 nodes is arranged behind the last layer LSTM hidden layer, the output of the full-connection layer is 8 nodes, and the full-connection layer corresponds to standardized rewards of corresponding parameters after 8 actions are executed. In training R-LSTM, we use a training dataset that contains one million.
Setting the maximum iteration number of the QL part as 1000 times, and ending the iteration when the iteration number reaches the maximum number. At the same time, the algorithm keeps track of the minimum curve error encountered in the iterative process during each iterationCorresponding result parameter theta k And the iteration sequence number k, once the current iteration sequence number i is larger than k+20, considering that the fitting effect of multiple iterations on the curve is not improved, terminating the iteration and outputting a final result theta out =θ k
After constructing the CF-DQN algorithm for 3 loading protocols, we tested the algorithm using simulation data. The simulation data is generated according to the above settings of the data, so we know the generated curve D real Corresponding real parameter theta real The result theta of the CF-DQN algorithm is convenient for us to examine out Relationship with the real parameters. Meanwhile, a noise signal can be added to the generated smooth curve to simulate data obtained in a real nanoindentation experiment.
The present application uses simulation data to perform 2 tests on CF-DQN, the following is a detailed description of 2 tests:
(1) Curve fitting effect test:
using the parameter theta real =[20000,0.2,50]5 curves of relaxation, load-unlock and creep are generated. For each of the 5 curves of the loading protocol, the following 5 treatments were performed: noise is not added, and the average value is 10- 7 Gaussian noise, uniform noise, rayleigh noise, exponential noise. The above 15 curves were fitted using the CF-DQN algorithm to see the curve fitting effect and the resulting parameters.
(2) Fitting 10000 randomly generated relatives:
in the range ofInternally randomly generating 10000 sets of parameters and using f r Generating 10000 corresponding relatives, and adding 10-percent average value to all the relatives 7 N gaussian random noise. The CF-DQN is used for fitting the 10000 curves and extracting result parameters, and the analysis result parameters and the real parameters are compared to verify the robustness of the CF-DQN.
The following are the results of tests performed on examples of the CF-DQN algorithm customized to the KJFD equation.
Curve fitting effect test:
parameter θ using CF-DQN algorithm real =[20000,0.2,50]The fitting effect of the generated KVFD equation curves with different noise added is shown in fig. 5. FIG. 5 shows the CF-DQN algorithm versus the real parameter θ real =[E 0 ,α,τ]=[20000,0.2,50]The loading protocol is relaxation, creep, load-unloading, and fitting results of 15 simulation curve pairs without adding noise, gaussian noise, uniform noise, rayleigh noise and exponential noise are respectively made. The curves from column 1 to column 3 in FIG. 5 are corresponding curves of relaxation, load-unlock and creep loading protocols, respectively; line 1 is a curve without noise, line 2 is a curve with gaussian noise, line 3 is a curve with uniform noise, line 4 is a curve with gareli noise, line 5 is a curve with exponential noise, the average of the added noise is 10 -7 . Generated simulation curve D real Fitting the result curve D to the solid line, CF-DQN out =f(θ out ) Represented by the hollow origin.
Table 3 shows the fitting result parameters θ for each curve out Absolute error of resultant parameter and true parameter |theta outreal Mean Absolute Error (MAE) of the results curves and simulation curves
TABLE 3 CF-DQN Algorithm versus real parameters θ real =[E 0 ,α,τ]=[20000,0.2,50]
In the absence of added noise (FIG. 5, line 1), the fitted result curves for the 3 loading protocols are almost identical to the simulation curves, with the curve MAE at 10- 9 The order of magnitude (see Table 3). The parameters of the fitting result of the correlation are [19339,0.1953, 70 ]]Load-unload is [19320,0.1951, 70]Creep is [18410,0.1935, 105 ]]Fitting the result parameter θ out And the true parameter theta real Proximity.
Under the condition of adding noise (line 2 to line 5 of fig. 5), the fitting result curves of the 3 loading protocols are also consistent with the skeleton of the simulation curve, the fitting is basically not interfered by noise burrs, and the curve MAE is 10- 8 The order of magnitude (see Table 3). From the image, the minimum disturbance to the curve is uniform noise (line 3 of fig. 5) in 4 kinds of noise with the same mean value, and the MAE of the three loading modes relaxation, creep, load-unloading are respectively 2.64E-8, 3.15E-8 and 2.65E-8; secondly, rayleigh noise (line 4 of FIG. 5) is respectively 5.47E-8, 5.74E-8 and 5.48E-8 for the relaxation, creep, load-unloading three loading modes MAE; the maximum disturbance to the curve morphology caused by gaussian noise (line 2 of fig. 5) and exponential noise (line 5 of fig. 5) is around 8E-8 for the 6 curves. Exponential noise also gives rise to many upward spike projections to the curve, making the curve appear to shift upward, but the fitted result curve can still match the bottom skeleton of the upper simulation curve. Fitting result parameter theta of 3 kinds of loading protocols out Still with the real parameter theta real Close, there is no significant change due to the addition of noise.
In the embodiment of the application, 10000 randomly generated correlation curve fitting results are obtained: randomly generating 10000 groups of parameters theta real Generating a corresponding relativity curve D real And add to the average value 10 -7 N Gaussian random noise, then using CF-DQN algorithm to extract parameters for 10000 curves, and extracting result parameters theta out And a true parameter theta real And (5) performing comparative analysis.
Respectively to E 0 The Pearson correlation analysis was performed for α, τ, and the results are shown in table 4.θ real E of (2) 0 Parameters and θ out E of (2) 0 The correlation coefficient between the parameters is r=0.8288 (p < 0.001), θ real Alpha parameter and theta of (2) out The correlation coefficient between the alpha parameters of (a) is r= 0.9963 (p < 0.001), θ real τ parameter and θ of (2) out The correlation coefficient between τ parameters is r= 0.2802 (p < 0.001), and a significant correlation exists between the output parameters and the real parameters. The correlation coefficient of the alpha parameter is closest to 1, E 0 The correlation coefficient of the parameter is also closer to 1, and the correlation coefficient of the τ parameter is farther from 1. Description of the CF-DQN algorithm example designed by us extracting 3 parameters, the alpha parameter is very close to the real parameter of the curve, E 0 The parameter is close to the curve real parameter, and the difference between the tau parameter and the curve real parameter is larger.
TABLE 4 real parameters θ for 10000 simulation curves real And the result parameter theta out Results of Pearson correlation analysis
At theta real The parameter term in (a) is the abscissa, and theta is taken as out The parameter item in (a) is the ordinate, E for 10000 pieces of data 0 The α, τ parameters are plotted in scatter plots, and the correlation line (red straight line) and 45 ° diagonal line (black straight line) are plotted, respectively, as shown in fig. 6. FIG. 6 shows a plot of θ for 10000 simulation curves real The parameter term in (a) is the abscissa, and theta is taken as out The parameter items in (a) are ordinate and the drawn scatter diagram. In the figure, the solid line is a correlation line, the dotted line is a 45-degree diagonal line, each small point is a piece of data, r is a correlation coefficient, and p is significance. (a) Parameter E 0 The method comprises the steps of carrying out a first treatment on the surface of the (b) parameter α; (c) a parameter τ. FIG. 6 (a) is E 0 Point-to-point plot of parameters, θ real E of (2) 0 And theta out E of (2) 0 The constituent pairs of points are distributed approximately about a 45 diagonal and follow θ real E in (E) 0 Is more diffuse, the correlation line is closer to the 45 diagonal, and the slope of the correlation line is smaller than the 45 diagonal. FIG. 6 (b) is a point of the alpha parameterFor the graph, θ real Alpha and theta of (2) out The pairs of points of alpha of (a) are closely spaced around the 45 diagonal with only a small number of pairs of points being offset from the 45 diagonal, the associated line being substantially coincident with the 45 diagonal. FIG. 6 (c) is a point-to-point plot of the τ parameter, θ real τ and θ of (2) out The pairs of points of τ of (c) spread over the entire parameter range, exhibiting a relatively weak positive correlation, the correlation line being farther from the 45 diagonal. The CF-DQN algorithm has the best fitting effect on alpha parameter and E 0 The fitting effect of the parameters is also good, and the fitting effect of the tau parameter is the worst.
Referring to FIG. 7, FIG. 7 shows MAE fitted to 10000 simulation curvesAnd the distribution thereof. (a) fitting error for each curve; (b) error distribution. 10000 fitting result curves D out And simulation curve D real Absolute average error>And MAE distribution is shown in FIG. 7. FIG. 7 (a) shows the log of MAE for 10000 data, showing that MAE is distributed at 10 -8 ~10 -4 Within the range. FIG. 7 (b) more intuitively reflects MAE distribution, and statistically, MEA's with 4550 curves are less than 10 -7 MEA with 3654 curves at 10 -7 ~10 -6 Between which an MEA with 3654 curves is at 10 -7 ~10 -6 Between which an MEA having 1293 curves is at 10 -6 ~10 -5 Between which an MEA having 503 curves is at 10 -5 ~10 -4 Between them. The curve fitting effect of the CF-DQN algorithm is good.
In summary, the embodiment of the application discloses a hierarchical Deep-reinforcement learning strategy (CF-DQN) for complex function multi-parameter optimization. The method is used for function fitting, solves the fitting problem of complex functions, and has good effect on the global parameter approximation of multi-parameter functions, particularly non-convex functions. Based on deep reinforcement learning (DQN), 2 special long-short-time memory neural networks (Long Short Term Mermory network, LSTM) are established and are respectively used for DQN winning prediction and DQN iterative initial parameter prediction. The prize value of the DQN is predicted by a custom prize evaluation rule (Reward Evaluation Rules, RER) and custom prize LSTM (Reward-LSTM, R-LSTM), and the Initial parameters of the first iteration are predicted by custom Initial LSTM (I-LSTM).
It will be appreciated by those skilled in the art that embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flowchart illustrations and/or block diagrams, and combinations of flows and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
The above embodiments are only for illustrating the technical solution of the present application and not for limiting the same, and although the present application has been described in detail with reference to the above embodiments, one skilled in the art may make modifications and equivalents to the specific embodiments of the present application, and any modifications and equivalents not departing from the spirit and scope of the present application are within the scope of the claims of the present application.

Claims (7)

1. The global optimization method for the mechanical parameters of the double loss value network deep reinforcement learning KVFD model is characterized by comprising the following steps:
s1, inputting a pre-acquired nano indentation measurement curve into a trained predicted value acquisition network to obtain a parameter predicted value of the nano indentation measurement curve; the trained predicted value acquisition network is a cyclic neural network based on an LSTM hidden layer, and LOSS function values used in the cyclic neural network are calculated by curve-curve corresponding parameters of an input network and parameter-parameter corresponding curves of network output;
s2, iterating the parameter predicted value as an iteration initial value of a deep reinforcement learning algorithm to obtain an approximation of a global parameter solution of the pre-acquired nano indentation measurement curve; the reward value prediction network of the deep reinforcement learning algorithm gives out the reward value when the current parameter changes to different directions through the difference value of the corresponding curve and the real curve of the current iteration parameter, and guides the parameter to approach to the global parameter;
when the approximation of the global parameter solution reaches a preset convergence condition, outputting the approximation of the global parameter solution as a mechanical parameter of the KJFD model;
in step S1, the pre-acquired nano indentation measurement curve includes a time sequence, a stress sequence and an indentation depth sequence;
in step S2, the iteration is performed by using the parameter predicted value obtained in step S1 as an iteration initial value of the deep reinforcement learning algorithm, and in the process of obtaining an approximation of the global parameter solution of the pre-obtained nanoindentation measurement curve, the specific steps of each iteration include:
(1) Respectively predicting the reward value of the candidate parameter set of the current iteration parameter by using a reward evaluation rule and the reward value prediction network, and taking the weighted sum of the reward evaluation rule and the reward value prediction network as the reward evaluation of the candidate parameter set of the current iteration parameter;
the reward evaluation rule is that for evaluating a certain alternative parameter, firstly calculating the curve difference delta between the corresponding curve of the alternative parameter and the pre-acquired nano indentation measurement curve, and then calculating the absolute average value of the curve difference
The evaluation formula of the prize value r is expressed as:
(2) And (3) calculating a new row of the Q table according to the reward evaluation obtained in the step (1) and the content of the current row of the Q table in the deep reinforcement learning algorithm, finding the maximum value in the new row of the Q table, and taking the corresponding alternative parameter as the current iteration result parameter.
2. The method according to claim 1, wherein in step S1, the predicted value obtaining network comprises: a plurality of LSTM hidden layers and a DNN network;
the unit numbers of each of the LSTM hidden layers are fixed and consistent, and each LSTM hidden layer is connected in a point-to-point mode; the first LSTM hidden layer inputs a pre-acquired nanoindentation measurement curve, and the last LSTM hidden layer output value enters a DNN network;
the DNN network comprises a plurality of full-connection layers and convolution layers with different dimensions, and the full-connection layers and the convolution layers are used for converting the value output by the last LSTM hidden layer into a parameter predicted value to be output.
3. The global optimization method for mechanical parameters of a kfd model according to claim 1, wherein in step S1, the calculation expression of the LOSS function value is,
wherein L is p Partial calculation of tag parameter θ train And network output parametersBetween loss values, L d Partial calculation curve D train Curve corresponding to network output parameters ∈ ->Between loss values, w p 、w d Respectively is L p 、L d Weight of the two parts.
4. The method according to claim 1, wherein in step S2, the reward value prediction network of the deep reinforcement learning algorithm comprises: a plurality of LSTM hidden layers and a DNN network;
the unit numbers of each of the LSTM hidden layers are fixed and consistent, and each LSTM hidden layer is connected in a point-to-point mode; the first LSTM hidden layer inputs a difference value obtained by subtracting a pre-acquired nanoindentation measurement curve from a current iteration curve, and the last LSTM hidden layer output value enters a DNN network;
the DNN network comprises a plurality of fully connected layers and convolution layers with different dimensions, and the fully connected layers and the convolution layers are used for converting the value output by the last LSTM hidden layer into rewards prediction for actions in all directions.
5. The global optimization method of mechanical parameters of a kfd model according to claim 4, wherein in step S2, the LOSS function used in the training of the prize value prediction network is the sum of absolute errors of the label prize value vector and the prize value vector output by the network.
6. The method for global optimization of mechanical parameters of a kfd model according to claim 1, wherein in step S2, the specific step of determining whether the approximation of the global parameter solution reaches a preset convergence condition includes: stopping iteration when the error between the corresponding curve of the current iteration result parameter and the pre-acquired nanoindentation measurement curve is smaller than a certain preset value; or stopping iteration when the iteration times reach a preset value.
7. The utility model provides a global optimization system of double loss value network deep reinforcement learning KVFD model mechanical parameters which is characterized in that the system comprises:
the parameter prediction value acquisition module is used for inputting the pre-acquired nano indentation measurement curve into a trained prediction value acquisition network to acquire a parameter prediction value of the nano indentation measurement curve; the trained predicted value acquisition network is a cyclic neural network based on an LSTM hidden layer, and LOSS function values used in the cyclic neural network are calculated by curve-curve corresponding parameters of an input network and parameter-parameter corresponding curves of network output;
the depth reinforcement learning iteration output module is used for iterating the obtained parameter predicted value as an iteration initial value of a depth reinforcement learning algorithm to obtain an approximation of a global parameter solution of the pre-obtained nano indentation measurement curve; the reward value prediction network of the deep reinforcement learning algorithm gives out the reward value when the current parameter changes to different directions through the difference value of the corresponding curve and the real curve of the current iteration parameter, and guides the parameter to approach to the global parameter; when the approximation of the global parameter solution reaches a preset convergence condition, outputting the approximation of the global parameter solution as a mechanical parameter of the KJFD model;
in the step parameter predicted value acquisition module, the pre-acquired nano indentation measurement curve comprises a time sequence, a stress sequence and an indentation depth sequence;
in the step of deep reinforcement learning iteration output module, the obtained parameter predicted value is iterated as an iteration initial value of a deep reinforcement learning algorithm, and in the process of obtaining the approximation of the global parameter solution of the pre-obtained nano indentation measurement curve, the specific steps of each iteration include:
(1) Respectively predicting the reward value of the candidate parameter set of the current iteration parameter by using a reward evaluation rule and the reward value prediction network, and taking the weighted sum of the reward evaluation rule and the reward value prediction network as the reward evaluation of the candidate parameter set of the current iteration parameter;
the reward evaluation rule is that for evaluating a certain alternative parameter, firstly calculating the curve difference delta between the corresponding curve of the alternative parameter and the pre-acquired nano indentation measurement curve, and then calculating the absolute average value of the curve difference
The evaluation formula of the prize value r is expressed as:
(2) And (3) calculating a new row of the Q table according to the reward evaluation obtained in the step (1) and the content of the current row of the Q table in the deep reinforcement learning algorithm, finding the maximum value in the new row of the Q table, and taking the corresponding alternative parameter as the current iteration result parameter.
CN202110368257.2A 2021-04-06 2021-04-06 Global optimization method and system for mechanical parameters of double loss value network deep reinforcement learning KVFD model Active CN113077853B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110368257.2A CN113077853B (en) 2021-04-06 2021-04-06 Global optimization method and system for mechanical parameters of double loss value network deep reinforcement learning KVFD model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110368257.2A CN113077853B (en) 2021-04-06 2021-04-06 Global optimization method and system for mechanical parameters of double loss value network deep reinforcement learning KVFD model

Publications (2)

Publication Number Publication Date
CN113077853A CN113077853A (en) 2021-07-06
CN113077853B true CN113077853B (en) 2023-08-18

Family

ID=76615137

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110368257.2A Active CN113077853B (en) 2021-04-06 2021-04-06 Global optimization method and system for mechanical parameters of double loss value network deep reinforcement learning KVFD model

Country Status (1)

Country Link
CN (1) CN113077853B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017004626A1 (en) * 2015-07-01 2017-01-05 The Board Of Trustees Of The Leland Stanford Junior University Systems and methods for providing reinforcement learning in a deep learning system
CN109002942A (en) * 2018-09-28 2018-12-14 河南理工大学 A kind of short-term load forecasting method based on stochastic neural net
CN110443364A (en) * 2019-06-21 2019-11-12 深圳大学 A kind of deep neural network multitask hyperparameter optimization method and device
WO2020220191A1 (en) * 2019-04-29 2020-11-05 Huawei Technologies Co., Ltd. Method and apparatus for training and applying a neural network
CN111914474A (en) * 2020-06-28 2020-11-10 西安交通大学 Fractional-order KVFD multi-parameter machine learning optimization method for viscoelastic mechanical characterization of soft substances

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017004626A1 (en) * 2015-07-01 2017-01-05 The Board Of Trustees Of The Leland Stanford Junior University Systems and methods for providing reinforcement learning in a deep learning system
CN109002942A (en) * 2018-09-28 2018-12-14 河南理工大学 A kind of short-term load forecasting method based on stochastic neural net
WO2020220191A1 (en) * 2019-04-29 2020-11-05 Huawei Technologies Co., Ltd. Method and apparatus for training and applying a neural network
CN110443364A (en) * 2019-06-21 2019-11-12 深圳大学 A kind of deep neural network multitask hyperparameter optimization method and device
CN111914474A (en) * 2020-06-28 2020-11-10 西安交通大学 Fractional-order KVFD multi-parameter machine learning optimization method for viscoelastic mechanical characterization of soft substances

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于深度学习长时预测的非线性系统建模方法研究;李聪;时宏伟;;现代计算机(第15期);全文 *

Also Published As

Publication number Publication date
CN113077853A (en) 2021-07-06

Similar Documents

Publication Publication Date Title
CN109931678B (en) Air conditioner fault diagnosis method based on deep learning LSTM
Zhang et al. A fault diagnosis method for wind turbines gearbox based on adaptive loss weighted meta-ResNet under noisy labels
CN111310915B (en) Data anomaly detection defense method oriented to reinforcement learning
CN108665058B (en) Method for generating countermeasure network based on segment loss
US20220147877A1 (en) System and method for automatic building of learning machines using learning machines
CN109961172B (en) CPS rare event probability prediction method based on statistical model test
Kang et al. A virtual sample generation method based on differential evolution algorithm for overall trend of small sample data: Used for lithium-ion battery capacity degradation data
CN112699045A (en) Software test case generation method based on multi-population genetic algorithm
CN112365033B (en) Wind power interval prediction method, system and storage medium
CN114298851A (en) Network user social behavior analysis method and device based on graph sign learning and storage medium
TW202123098A (en) Method and electronic device for selecting neural network hyperparameters
CN111123894A (en) Chemical process fault diagnosis method based on combination of LSTM and MLP
CN114166509A (en) Motor bearing fault prediction method
CN109951327B (en) Network fault data synthesis method based on Bayesian hybrid model
CN108665001B (en) Cross-tested idle state detection method based on deep belief network
CN113077853B (en) Global optimization method and system for mechanical parameters of double loss value network deep reinforcement learning KVFD model
CN112257348B (en) Method for predicting long-term degradation trend of lithium battery
CN111612022A (en) Method, apparatus, and computer storage medium for analyzing data
CN111783930A (en) Neural network test sufficiency evaluation method based on path state
CN113360772B (en) Interpretable recommendation model training method and device
CN114140736A (en) Image anomaly detection method based on high-frequency and low-frequency reconstruction
CN111061711B (en) Big data stream unloading method and device based on data processing behavior
Eidenbenz et al. Boosting exploratory testing of industrial automation systems with ai
CN115565611B (en) Biological regression prediction method, device, equipment and storage medium
CN114021632B (en) Laboratory electricity consumption measurement equipment test strategy evaluation method based on Markov chain, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant