CN112947078A - Servo motor intelligent optimization control method based on value iteration - Google Patents
Servo motor intelligent optimization control method based on value iteration Download PDFInfo
- Publication number
- CN112947078A CN112947078A CN202110148138.6A CN202110148138A CN112947078A CN 112947078 A CN112947078 A CN 112947078A CN 202110148138 A CN202110148138 A CN 202110148138A CN 112947078 A CN112947078 A CN 112947078A
- Authority
- CN
- China
- Prior art keywords
- strategy
- state
- optimal
- control
- value
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 37
- 238000005457 optimization Methods 0.000 title claims abstract description 12
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 16
- 238000005070 sampling Methods 0.000 claims abstract description 14
- 238000011217 control strategy Methods 0.000 claims abstract description 10
- 239000011159 matrix material Substances 0.000 claims description 27
- 238000013528 artificial neural network Methods 0.000 claims description 6
- 238000012854 evaluation process Methods 0.000 claims description 3
- 230000003044 adaptive effect Effects 0.000 abstract description 7
- 238000011156 evaluation Methods 0.000 abstract description 7
- 238000004364 calculation method Methods 0.000 abstract description 2
- 230000008859 change Effects 0.000 description 5
- 230000000007 visual effect Effects 0.000 description 3
- 230000000875 corresponding effect Effects 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 230000018109 developmental process Effects 0.000 description 2
- 238000002474 experimental method Methods 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000005200 bud stage Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000005094 computer simulation Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000012804 iterative process Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B13/00—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
- G05B13/02—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
- G05B13/04—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric involving the use of models or simulators
- G05B13/042—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric involving the use of models or simulators in which a parameter or coefficient is automatically adjusted to optimise the performance
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Computation (AREA)
- Medical Informatics (AREA)
- Software Systems (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Automation & Control Theory (AREA)
- Feedback Control In General (AREA)
Abstract
A servo motor intelligent optimization control method based on value iteration is characterized in that a system is initialized, and system control is determined; and sampling the system, performing strategy evaluation by using a least square method on-line calculation value function, and updating the strategy by using a greedy algorithm when an optimal value function is obtained, so as to finally obtain the optimal strategy. The method realizes intelligent optimization control on the system by solving the optimal control strategy through value iterative adaptive control, and compared with the prior art, the method does not need to identify the system under the condition that the model parameters of part of the system are unknown, but realizes the optimal control on the system on line based on the value iterative adaptive control method.
Description
Technical Field
The invention belongs to the technical field of control, and particularly relates to a servo motor intelligent optimization control method based on value iteration, which realizes the optimal control of a servo motor under the condition of unknown system model.
Background
In the past control problem, for solving the optimal control problem of the continuous linear time-invariant system, the solution of the generation Raccati equation is usually calculated, such as an eigenvector-based algorithm and a numerical simulation method, LQR method, but all of these methods and their numerically advantageous variables ARE off-line programs, and have been proved to be able to converge to the desired ARE solution. They either operate on hamiltonian matrices associated with ARE (eigenvector and matrix sign based algorithms) or require solving Lyapunov equations (newton's method). In all cases, a model of the system is required and a prior identification process is always required. In practice, however, the system modeling is difficult and requires a great deal of time and effort. Furthermore, even if a model is available, the state feedback controller derived based on the model is only optimal for model approximation of the actual system dynamics.
When the global model of the controlled system is unknown, application data driven control theory and method are considered to solve the actual control problem. The data-driven concept means that on-line and off-line data of a controlled system are utilized, various expected functions of the system based on data forecasting, evaluation, scheduling, monitoring, diagnosis, decision-making, optimization and the like are realized under the condition that a model is not completely known, the data-driven concept is originally derived from the field of computer science, and is still in the bud stage in the control field in recent years. At present, for the control research of a servo motor, more than all traditional control methods based on models are adopted, the control of a system cannot be realized on line by utilizing a data driving idea, and the models need to be completely known, so that a large amount of time is needed for modeling, and meanwhile, the problems of model mismatching, unmodeled dynamics and the like can exist.
Disclosure of Invention
In order to overcome the defects of the existing method, the invention provides a servo motor intelligent optimization control method based on value iteration, which provides a self-adaptive value iteration algorithm, combines the concepts of ADP and an intelligent optimization control system theory, and provides a new ADP technology, and can solve the problem of continuous time infinite time domain optimal control of a linear system with partially unknown dynamics (namely, the internal dynamics specified by a system matrix A) in a time forward mode. The controller parameters are updated based on a sequence of signals measuring the controller performance, using an iterative process of updating the control strategy and value function estimates to bring them closer to the optimal control strategy and corresponding optimal value function. Each iteration step includes an update of the value function estimate based on the current control strategy and then the control strategy is updated based on the new value function estimate.
An entity object (networked multi-axis motion control system) with physical characteristics considered by the invention is shown in fig. 1 and mainly comprises an upper computer, an ARM microprocessor, an AC servo system, a servo motor, a power supply, a switch and a CAN bus. The algorithm programming is based on a Visual Studio C + + environment, the Visual Studio is a set of powerful IDE (integrated development environment) developed by Microsoft (Microsoft) as a platform mainly based on Windows, and the Visual Studio supports the development of various languages such as C #, F #, VB, C/C + +, and the like. The main work of the upper computer is to receive data from the ARM microprocessor through a TCP/IP protocol, operate an embedded control algorithm and then send a control instruction to the ARM microprocessor. The ARM microprocessor is used as a data transfer station, obtains information such as speed, position, torque and the like of the servo motor from the servo system through the CAN bus, transmits the information to the upper computer, and simultaneously receives a control command of the upper computer and sends the control command to the servo system. The AC servo system is a high-performance communication type servo driver of Taida ASDA-A2 series, and is used for responding to a control command of a PC in real time and driving a servo motor to execute corresponding actions, and a power supply and a switch are responsible for power-on and power-off of the system.
The technical scheme adopted by the invention for solving the technical problems is as follows:
a servo motor intelligent optimization control method based on value iteration comprises the following steps:
step 1) establishing a motor system state space equation as follows:
wherein A is a system state matrix, and part of parameters are notKnowing that the information therein is obtained from the current state x (T) and the next time x (T + T), B is the input matrix, x (T) e Rn,u(t)=RmAnd (a, B) is controllable, for optimal control problems:
the problem of the optimal control in the finite field is as follows:
selecting Q ═ 1, R ═ 1, (A, B) can be controlled, the solving of controller is confirmed by Bellman optimum principle, u ═ Kx, where K ═ R-1BTP, and P satisfies the algebraic Riccati equation:
ATP+PA-PBR-1BTP+Q=0 (4);
step 2) initializing the system, comprising the following steps:
2.1) selecting a basis function: for continuous time LQR, the value is quadratic in the state,therefore, the basis functions of the actor neural network in equation (5) are selectedA quadratic polynomial vector as a state component, where the number of states of the vector is n, and the basis function contains n (n +1)/2 components, and a weight vector W is composed of elements in the matrix P;
the system control is linear in state, u ═ Kx, with the basis functions of the actor neural network (6) as state components;
uk=h(xk)=UTxk (6)
2.2) initializing the system: selecting an initial state x0Calculating the initial value of the basis function and determining an initial strategy K0;
Step 3) sampling the system, and calculating by a least square method to obtain an optimal value function, namely a strategy evaluation process; to obtain in strategy KiThe value function V (x (T + T)) ═ x for each next stepT(t+T)Pix (T + T), using a parameter matrix PiAnd calculating, wherein the formula is changed into:
whereinIs a second order polynomial basis vector of Kronecker inner product and has the element of { xi(t)xj(t)}i=1,n;j=i,n,Acting on the nxn matrix as a vector value matrix function, by superimposing the elements of the symmetric matrix into a vector, the off-diagonal elements are summed to Pij+PjiThereby obtaining a column vector;
in each iteration step, the same control strategy K is usediAfter collecting enough state track points, solving V function parameter by using minimum multiplicationThereby obtaining Pi+1Finding the minimum of the parameter vector P in the least-squares sense by minimizing the error between the objective functions, N in the state space>N (N +1) points xi are evaluated to obtain a least squares solution:
wherein X ═ X (T) -X (T + T)]T (9)
The state at discrete times of time T and T + T is measured, and the reward observed during the sampling interval:
and 4) updating the optimal parameters through a greedy algorithm according to the obtained optimal value function:
ui=-R-1BTPix=Kix (12)
when the least square method converges, the strategy is not updated any more, and the optimal strategy is obtained.
In the present invention, the continuous-time ADP algorithm consists of iterations between (11) and (12). However, using (12) the update control strategy (operator) requires a B matrix, which makes the tuning algorithm only partially model-free.
The working principle of the invention is as follows: initializing a system and determining system control; and sampling the system, performing strategy evaluation by using a least square method on-line calculation value function, and updating the strategy by using a greedy algorithm when an optimal value function is obtained, so as to finally obtain the optimal strategy.
The invention has the beneficial effects that: compared with the prior art, under the condition that partial model parameters of the system are unknown, the system is not required to be identified, and the optimal control is realized on line by the adaptive control method based on value iteration.
Drawings
FIG. 1 is a flow chart of a design method operating and implemented on a platform;
FIG. 2 is a system input u variation curve based on value iterative adaptive control;
FIG. 3 is a diagram of system state change based on value iterative adaptive control;
FIG. 4 is a graph of the change in state of a system at any given strategy value for a known system dynamics model;
FIG. 5 is a graph of a variation of the transition matrix P based on value iterative adaptive control;
FIG. 6 is a graph of performance index change based on value iterative adaptive control and any given policy K.
Detailed Description
In order to make the technical features, purposes and advantages of the present invention clearer and clearer, the technical scheme of the present invention is further described below with reference to the accompanying drawings and practical experiments.
Referring to fig. 1-6, a servo motor intelligent optimization control method based on value iteration, first initializing a target system, and selecting a proper basis function; sampling the system, calculating the state of the next moment according to the state of the current moment, and calculating the optimal value function on line; after the optimal value function is obtained, the strategy is updated by using a greedy algorithm, the strategy is optimal when converging, and the strategy is not updated any more, so that the optimal control on the system is realized.
The optimal control method of the servo motor based on value iteration comprises the following steps:
step 1) establishing a motor system state space equation as follows:
wherein, A is a system state matrix, part of parameters are unknown, the information is obtained from the current state x (T) and the next time x (T + T), B is an input matrix, x (T) epsilon Rn,u(t)=RmAnd (a, B) is controllable, for optimal control problems:
the problem of the optimal control in the finite field is as follows:
selecting Q ═ 1, R ═ 1, (A, B) can be controlled, the solving of controller is confirmed by Bellman optimum principle, u ═ Kx, where K ═ R-1BTP, and P satisfies the algebraic Riccati equation:
ATP+PA-PBR-1BTP+Q=0 (4);
step 2) initializing the system, comprising the following steps:
2.1) selecting a basis function: for continuous time LQR, the value is quadratic in the state,therefore, the basis functions of the actor neural network in equation (5) are selectedRn→RLA quadratic polynomial vector as a state component, where the number of states of the vector is n, and the basis function contains n (n +1)/2 components, and a weight vector W is composed of elements in the matrix P;
the system control is linear in state, u ═ Kx, with the basis functions of the actor neural network (6) as state components;
uk=h(xk)=UTxk (6)
2.2) initializing the system: selecting an initial state x0Calculating the initial value of the basis function and determining an initial strategy K0;
Step 3) sampling the system, calculating by a least square method, and obtaining an optimal value functionNamely a policy evaluation process; to obtain in strategy KiThe value function V (x (T + T)) ═ x for each next stepT(t+T)Pix (T + T), using a parameter matrix PiAnd calculating, wherein the formula is changed into:
whereinIs a second order polynomial basis vector of Kronecker inner product and has the element of { xi(t)xj(t)}i=1,n;j=i,n,Acting on the nxn matrix as a vector value matrix function, by superimposing the elements of the symmetric matrix into a vector, the off-diagonal elements are summed to Pij+PjiThereby obtaining a column vector;
in each iteration step, the same control strategy K is usediAfter collecting enough state track points, solving V function parameter by using minimum multiplicationThereby obtaining Pi+1Finding the minimum of the parameter vector P in the least-squares sense by minimizing the error between the objective functions, N in the state space>N (N +1) points xi are evaluated to obtain a least squares solution:
wherein X ═ X (T) -X (T + T)]T (9)
The state at discrete times of time T and T + T is measured, and the reward observed during the sampling interval:
and 4) updating the optimal parameters through a greedy algorithm according to the obtained optimal value function:
ui=-R-1BTPix=Kix (12)
when the least square method converges, the strategy is not updated any more, and the optimal strategy is obtained.
In this embodiment, in step 1), a second-order motor system is considered as follows:
in the step 2), the self-adaptive control algorithm based on value iteration is used for the experiment, the matrix A is only used for obtaining sampling data, and the evaluation and the updating of the strategy in the control algorithm do not involve the use of the matrix A. x is the number of1Finger motor position, x2Refers to the motor speed. Initializing the system and taking an initial state x0=[4 0.4]TSelecting basis functions
The above-mentionedIn step 3), arbitrarily giving a certain policy, performing policy evaluation on the system: at a given initial policy K0=[0 0]TIn the case of (1), the sampling time T is taken to be 0.05s, and the sampling time T is set to be 0.05s in a finite interval [ T, T + T ]]Sampling the system, updating the state x (T + T) at the next moment by using the state x (T) at the current moment, and calculating a value function by using a least square method;
in the step 4), strategy promotion is performed on the system: after strategy evaluation, an optimal value function is obtained, strategy updating is carried out by utilizing a greedy algorithm, when the strategy does not change along with time, an optimal strategy is obtained, control action is continuous time control, gain is updated at a sampling point, and the change is shown in figure 2.
From the experimental results of fig. 3, after the policy is updated 3 times, the policy convergence does not converge to 0 finally when the system state is updated. FIG. 4 illustrates that the final value of P, which is obtained by value iteration, isThe control effect can meet the expected requirement.
In comparison with the known dynamic model and the condition that any given strategy K is [ 30.3 ], fig. 5 and 6 illustrate that the system state convergence of the method is more gradual and rapid, and excessive overshoot does not occur in the process, which can be found in the performance index in fig. 6, and the method can better and more rapidly obtain the optimal performance index.
The invention provides a servo motor intelligent optimization control method based on value iteration, which uses a value iteration self-adaptive control method to realize on-line solving of the optimal control problem of a system through two steps of strategy evaluation and strategy promotion, and compared with the prior art, the invention has the practicability that: the system model parameters are not required to be identified, and system information can be acquired by acquiring system track data, so that an optimal control strategy is acquired.
The technical solution of the present invention is described in detail above with reference to the accompanying drawings but is not limited thereto, and various changes and modifications can be made within the knowledge of those skilled in the art based on the concept of the present invention.
Claims (1)
1. An intelligent optimization control method for a servo motor based on value iteration is characterized by comprising the following steps:
step 1) establishing a motor system state space equation as follows:
wherein, A is a system state matrix, part of parameters are unknown, the information is obtained from the current state x (T) and the next time x (T + T), B is an input matrix, x (T) epsilon Rn,u(t)=RmAnd (a, B) is controllable, for optimal control problems:
the problem of the optimal control in the finite field is as follows:
selecting Q ═ 1, R ═ 1, (A, B) can be controlled, the solving of controller is confirmed by Bellman optimum principle, u ═ Kx, where K ═ R-1BTP, and P satisfies the algebraic Riccati equation:
ATP+PA-PBR-1BTP+Q=0 (4);
step 2) initializing the system, comprising the following steps:
2.1) selecting a basis function: for continuous time LQR, the value is quadratic in the state,therefore, the basis functions of the actor neural network in equation (5) are selectedRn→RLA quadratic polynomial vector as a state component, where the number of states of the vector is n, and the basis function contains n (n +1)/2 components, and a weight vector W is composed of elements in the matrix P;
the system control is linear in state, u ═ Kx, with the basis functions of the actor neural network (6) as state components;
uk=h(xk)=UTxk (6)
2.2) initializing the system: selecting an initial state x0Calculating the initial value of the basis function and determining an initial strategy K0;
Step 3) sampling the system, and calculating by a least square method to obtain an optimal value function, namely a strategy evaluation process; to obtain in strategy KiThe value function V (x (T + T)) ═ x for each next stepT(t+T)Pix (T + T), using a parameter matrix PiAnd calculating, wherein the formula is changed into:
whereinIs a second order polynomial basis vector of Kronecker inner product and has the element of { xi(t)xj(t)}i=1,n;j=i,n,Acting on the nxn matrix as a vector-valued matrix function by superimposing the elements of the symmetric matrix into a vector, summing the off-diagonal elements to Pij+PjiThereby obtaining a column vector;
in each iteration step, the same control strategy K is usediAfter collecting enough state track points, solving V function parameter by least square methodThereby obtaining Pi+1The minimum of the parameter vector P is found by minimizing the error between the objective functions in the least squares sense, N in the state space>N (N +1) points xi are evaluated to obtain a least squares solution:
wherein X ═ X (T) -X (T + T)]T (9)
The state at discrete times of time T and T + T is measured, and the reward observed during the sampling interval:
and 4) updating the optimal parameters through a greedy algorithm according to the obtained optimal value function:
ui=-R-1BTPix=Kix (12)
when the least square method converges, the strategy is not updated any more, and the optimal strategy is obtained.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110148138.6A CN112947078A (en) | 2021-02-03 | 2021-02-03 | Servo motor intelligent optimization control method based on value iteration |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110148138.6A CN112947078A (en) | 2021-02-03 | 2021-02-03 | Servo motor intelligent optimization control method based on value iteration |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112947078A true CN112947078A (en) | 2021-06-11 |
Family
ID=76242039
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110148138.6A Pending CN112947078A (en) | 2021-02-03 | 2021-02-03 | Servo motor intelligent optimization control method based on value iteration |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112947078A (en) |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108181816A (en) * | 2018-01-05 | 2018-06-19 | 南京航空航天大学 | A kind of synchronization policy update method for optimally controlling based on online data |
CN110262235A (en) * | 2019-06-18 | 2019-09-20 | 北京理工大学 | A kind of model-free optimal switching method of switching system |
CN110782011A (en) * | 2019-10-21 | 2020-02-11 | 辽宁石油化工大学 | Networked multi-agent system distributed optimization control method based on reinforcement learning |
CN111722531A (en) * | 2020-05-12 | 2020-09-29 | 天津大学 | Online model-free optimal control method for switching linear system |
CN111740658A (en) * | 2020-04-29 | 2020-10-02 | 南京理工大学 | Optimal regulation control method of motor system based on strategy iteration |
CN112149361A (en) * | 2020-10-10 | 2020-12-29 | 中国科学技术大学 | Adaptive optimal control method and device for linear system |
CN112180730A (en) * | 2020-10-10 | 2021-01-05 | 中国科学技术大学 | Hierarchical optimal consistency control method and device for multi-agent system |
-
2021
- 2021-02-03 CN CN202110148138.6A patent/CN112947078A/en active Pending
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108181816A (en) * | 2018-01-05 | 2018-06-19 | 南京航空航天大学 | A kind of synchronization policy update method for optimally controlling based on online data |
CN110262235A (en) * | 2019-06-18 | 2019-09-20 | 北京理工大学 | A kind of model-free optimal switching method of switching system |
CN110782011A (en) * | 2019-10-21 | 2020-02-11 | 辽宁石油化工大学 | Networked multi-agent system distributed optimization control method based on reinforcement learning |
CN111740658A (en) * | 2020-04-29 | 2020-10-02 | 南京理工大学 | Optimal regulation control method of motor system based on strategy iteration |
CN111722531A (en) * | 2020-05-12 | 2020-09-29 | 天津大学 | Online model-free optimal control method for switching linear system |
CN112149361A (en) * | 2020-10-10 | 2020-12-29 | 中国科学技术大学 | Adaptive optimal control method and device for linear system |
CN112180730A (en) * | 2020-10-10 | 2021-01-05 | 中国科学技术大学 | Hierarchical optimal consistency control method and device for multi-agent system |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN113498523B (en) | Apparatus and method for controlling operation of machine object and storage medium | |
US20220326664A1 (en) | Improved machine learning for technical systems | |
JP3556956B2 (en) | System identification device and method | |
Rubies-Royo et al. | A classification-based approach for approximate reachability | |
CN105955206A (en) | Multi-shaft motion control method based on data driving and parameter mixing optimization | |
CN113874865A (en) | Method and device for determining model parameters of a control strategy of a technical system by means of a Bayesian optimization method | |
Solgi et al. | Variable structure fuzzy wavelet neural network controller for complex nonlinear systems | |
Balakrishna et al. | On-policy robot imitation learning from a converging supervisor | |
CN108972546A (en) | A kind of robot constant force curved surface tracking method based on intensified learning | |
Precup et al. | A survey on fuzzy control for mechatronics applications | |
Dong et al. | Friction modeling and compensation for haptic master manipulator based on deep Gaussian process | |
KR102382047B1 (en) | Automatic learning tuning system of motor controller using PSO | |
CN114063438B (en) | Data-driven multi-agent system PID control protocol self-learning method | |
Ma et al. | Active manipulation of elastic rods using optimization-based shape perception and sensorimotor model approximation | |
Hsu et al. | Recurrent fuzzy-neural approach for nonlinear control using dynamic structure learning scheme | |
Hager et al. | Adaptive Neural network control of a helicopter system with optimal observer and actor-critic design | |
Pham et al. | Fault-tolerant control for robotic systems using a wavelet type-2 fuzzy brain emotional learning controller and a topsis-based self-organizing algorithm | |
US20220067504A1 (en) | Training actor-critic algorithms in laboratory settings | |
Ennen et al. | Learning robust manipulation skills with guided policy search via generative motor reflexes | |
CN116382093A (en) | Optimal control method and equipment for nonlinear system with unknown model | |
CN112947078A (en) | Servo motor intelligent optimization control method based on value iteration | |
Toner et al. | Probabilistically safe mobile manipulation in an unmodeled environment with automated feedback tuning | |
Wang et al. | A learning-based tune-free control framework for large scale autonomous driving system deployment | |
CN117321511A (en) | Robust adaptive dynamic pattern decomposition for modeling, prediction and control of high-dimensional physical systems | |
Gukov et al. | Real-time Multi-Objective Trajectory Optimization |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20210611 |
|
RJ01 | Rejection of invention patent application after publication |