CN118133929A - Method and device for accelerating neural network training based on node freezing - Google Patents

Method and device for accelerating neural network training based on node freezing Download PDF

Info

Publication number
CN118133929A
CN118133929A CN202410545941.7A CN202410545941A CN118133929A CN 118133929 A CN118133929 A CN 118133929A CN 202410545941 A CN202410545941 A CN 202410545941A CN 118133929 A CN118133929 A CN 118133929A
Authority
CN
China
Prior art keywords
node
frozen
neural network
training
freezing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202410545941.7A
Other languages
Chinese (zh)
Other versions
CN118133929B (en
Inventor
陈翯
李石坚
张犁
潘纲
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN202410545941.7A priority Critical patent/CN118133929B/en
Priority claimed from CN202410545941.7A external-priority patent/CN118133929B/en
Publication of CN118133929A publication Critical patent/CN118133929A/en
Application granted granted Critical
Publication of CN118133929B publication Critical patent/CN118133929B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a method and a device for accelerating neural network training based on node freezing, wherein the method comprises the following steps: (1) acquiring a neural network model to be trained; (2) Initializing a neural network model, and performing the steps (2-1) - (2-4) to train the neural network model; (2-1) selecting a next node to be frozen from all candidate nodes of the neural network model; (2-2) generating a corresponding evaluation subgraph according to the node to be frozen, wherein the evaluation subgraph is used for evaluating whether the node to be frozen can be frozen or not; (2-3) if the convergence degree of the node to be frozen does not reach the threshold value, updating the weight of the neural network model, returning to the step (2-2), and if the convergence degree of the node to be frozen does not reach the threshold value, executing the step (2-4); (2-4) freezing the node with the convergence reaching the set threshold value, and skipping the back propagation and weight updating processes of the node; and (3) performing image classification application after model training is finished. The invention can accelerate model training under the condition of reducing the loss of training precision as much as possible.

Description

Method and device for accelerating neural network training based on node freezing
Technical Field
The invention belongs to the technical field of neural network training, and particularly relates to a method and a device for accelerating neural network training based on node freezing.
Background
Deep neural network (Deep-Learning Neural Network, DNN) is a multi-layer unsupervised neural network that obtains better feature expression on input data samples through layer-by-layer feature learning.
Modern DNNs consist of several tens to several hundreds of layers, each layer receiving an input tensor of a feature and outputting a corresponding activation value. Feature extraction and expression is achieved by training a DNN through multiple iterations and minimizing the loss function. The training iteration of DNN includes three steps of forward propagation, backward propagation, and parameter synchronization. In each iteration, the forward propagation receives a small batch of data and calculates the loss value relative to the target tag layer by layer through the model. Back propagation the gradient of the parameter is calculated sequentially from back to front according to the derivative chain law. After the back propagation is finished, the model parameters are updated by using an optimization algorithm. In data-parallel distributed training, gradients computed independently from all working nodes need to be aggregated to update the shared model parameters.
In recent years, as the scale and data volume of deep neural network models are continuously increased, the training cost of DNN is also increased. How to reduce the training cost and accelerate the model training becomes a problem to be solved urgently.
Many studies have focused on reducing the cost of communications in a multi-machine multi-card environment, increasing the parallelism of model training by overlapping the time of computation, communications, or employing different scheduling approaches. In addition, structural optimization of the computational graph is also a common approach. Subgraph replacement is a typical way of computational graph optimization, speeding up model training by replacing computational subgraphs with less costly equivalent subgraphs. The above-described methods can effectively reduce training time, but the overall calculation remains substantially unchanged.
In order to practically reduce the total calculation amount in the large-scale DNN model training, a learner proposes to accelerate the DNN training by adding a node freezing method in the model training process. The essence of node freezing is to stop the parameter updating of a given node before the model training iteration is completed, skip the back propagation and parameter updating processes, and thus reduce the calculation amount. In the training process of the model, the calculated amount of counter propagation is obviously larger than that of forward propagation, so that the training efficiency can be obviously improved due to node freezing, and the model training is accelerated.
However, existing node freezing schemes still have drawbacks. Firstly, the granularity of freezing is mostly single layers, the freezing is carried out layer by layer from shallow to deep according to the layer sequence, and the flexibility of the modern multi-branch network structure is relatively poor; second, there is currently no unified theorem of when each node of the model converges during the training process. Further research is still needed on how to freeze the layer without sacrificing training accuracy as much as possible.
Disclosure of Invention
The invention provides a method and a device for accelerating neural network training based on node freezing, which can accelerate model training under the condition of reducing training precision loss as much as possible and improve image classification efficiency.
A method for accelerating neural network training based on node freezing, comprising the following steps:
(1) Acquiring a neural network model to be trained, wherein the neural network model is a multi-branch structure model for image classification;
(2) Initializing the degree of entry of each node in the neural network model as the number of precursor nodes, and training the neural network model by cycling the following steps (2-1) - (2-4);
(2-1) selecting a next node to be frozen from all candidate nodes of the neural network model; the candidate nodes are nodes which meet the condition that the precursors of the candidate nodes are frozen and the candidate nodes are not frozen;
(2-2) generating a corresponding evaluation subgraph according to the node to be frozen, wherein the evaluation subgraph is used for evaluating whether the node to be frozen can be frozen or not;
Wherein, the evaluation subgraph is divided into a reference node and a similarity measurement node: the reference node is a copy of the node to be frozen; the intermediate activation layer of the node to be frozen and the reference node contains semantic information, the semantic information is used as the input of the similarity measurement node together, the similarity measurement node calculates the semantic similarity of the node to be frozen and the reference node, and then the convergence of the node to be frozen is obtained;
(2-3) if the convergence degree of the node to be frozen does not reach the set threshold value, updating the weight of the neural network model, returning to the step (2-2), and if the convergence degree of the node to be frozen reaches the set threshold value, executing the step (2-4);
(2-4) freezing the node with convergence reaching the set threshold value, and skipping the back propagation and weight updating processes of the node to realize acceleration of model training;
(3) And (3) finishing training the neural network model, and applying image classification by using the trained model.
Further, the specific process of the step (2-1) is as follows:
And adding all candidate nodes of the neural network model into a priority queue by adopting a priority queue mode based on a self-defined comparator, and selecting a queue head element each time as a next node to be frozen.
The custom comparator of the priority queue needs to comprehensively evaluate the depth of the candidate node and the calculated amount of the candidate node, and the candidate node with smaller depth and smaller calculated amount has higher priority.
In step (2-2), the reference node has the same input as the node to be frozen, and the weight of the reference node is periodically aligned with the node to be frozen.
In the step (2-2), a corresponding evaluation sub-graph is generated by an evaluation sub-graph generating module, and the evaluation sub-graph generating module is an extension on the neural network model to be trained.
In the step (2-3), the selection of the set threshold is related to the type and parameter setting of the node to be frozen, and the set thresholds of different nodes to be frozen are different.
The specific process of the step (2-4) is as follows:
Marking the node with the convergence reaching the set threshold as the frozen node, skipping the back propagation and the weight updating of the node, subtracting one from the ingress degree of all the subsequent nodes, and if the ingress degree of the subsequent nodes is reduced to zero, adding the subsequent nodes into the priority queue in the step (2-3).
The invention also provides a device for training the acceleration neural network based on the node freezing, which comprises a memory and one or more processors, wherein executable codes are stored in the memory, and the one or more processors are used for realizing the method for training the acceleration neural network based on the node freezing when executing the executable codes.
Compared with the prior art, the invention has the following beneficial effects:
1. Aiming at the modern multi-branch network structure, the invention adopts a freezing strategy with nodes as freezing units, and realizes the selection of the nodes to be frozen in the network based on the priority queue of the custom comparator. The structural characteristics of the network model are fully utilized, and a better solution is provided for the selection of the nodes in the original relative fuzzy freezing theory;
2. the invention expands and evaluates the sub-graph module on the basis of the training network, and evaluates the convergence degree of the node to be frozen according to the semantic information of the node to be frozen. The nodes reaching the threshold value are frozen, gradients are not calculated and parameters of the nodes are updated in the subsequent back propagation process, so that the complexity of the network training process is improved, the network training time is effectively reduced, and the training efficiency is improved.
Drawings
FIG. 1 is a flow chart of a method for accelerating neural network training based on node freezing.
Fig. 2 is a block diagram of the modules according to an embodiment of the present invention.
Fig. 3 is a schematic diagram of node selection to be frozen in an embodiment of the invention.
Fig. 4 is a schematic diagram of evaluation sub-graph generation in an embodiment of the present invention.
Fig. 5 is a schematic diagram of similarity metric node generation in an embodiment of the present invention.
Detailed Description
The invention will be described in further detail with reference to the drawings and examples, it being noted that the examples described below are intended to facilitate the understanding of the invention and are not intended to limit the invention in any way.
As shown in fig. 1, a method for accelerating neural network training based on node freezing comprises the following steps:
Step 1, a neural network model to be trained is obtained, and in this embodiment, the neural network model is a multi-branch structure model for image classification.
Specifically, a relevant training data set (image data set in this embodiment) is first loaded, and all data is normalized. And then loading the neural network model to be trained. Finally, setting reasonable training super parameters, and finishing initialization operation before training. The initializing further comprises: initializing the degree of entry of each node in the neural network model as the number of precursor nodes.
And 2, training the neural network to be trained, and cycling the following steps (2.1) - (2.4) until the training is finished.
Specifically, in the embodiment of the present invention, the connection relationship of each module is shown in fig. 2. And selecting candidate nodes to be frozen as the input of the node selection module according to the structure of the training network and the current freezing condition. The node selection module selects the candidate node which is most easy to converge from the candidate nodes as the next node to be frozen. And the evaluation subgraph generation module generates a corresponding evaluation subgraph according to the node to be frozen, and the corresponding evaluation subgraph is used as an extension of the training network. The convergence evaluation module takes the evaluation subgraph and an intermediate activation layer of the node to be frozen in the original training network as inputs together, evaluates the convergence of the current node to be frozen, and sends a frozen signal to the layer freezing module according to the evaluation result. After the layer freezing module receives the freezing signal, marking the node with convergence reaching the expected value as the frozen node, skipping the back propagation and parameter updating of the node, and starting a new round of node selection to be frozen.
(2.1) A node selection module selecting a next node to be frozen from the candidate nodes. Candidate nodes are nodes that satisfy that their precursors are both frozen and that themselves are not frozen. The node selection module is realized by a priority queue based on a custom comparator, all candidate nodes are added into the priority queue, and a queue head element is selected each time as a next node to be frozen.
As shown in fig. 3, the black nodes of the left graph represent frozen nodes, the gray nodes represent candidate nodes, and the candidate nodes in the left graph are used as inputs of the node selection module. The node selection module is realized by a priority queue based on a custom comparator, and the custom comparator comprehensively evaluates the depth of the candidate nodes and the calculated amount of the nodes, wherein the nodes with smaller depth and smaller calculated amount have higher priority. And selecting the candidate node which is most easy to converge as the next node to be frozen each time according to the depth and the calculated amount of the node.
Specifically, all nodes without precursor dependency are added to a priority queue in a node selection module at the beginning. Before training starts, taking out the head element in the queue as a first node to be frozen, and transmitting the first element into an evaluation subgraph generation module to generate a corresponding evaluation subgraph. After the training iteration starts, the convergence evaluation and the iterative training of the network model are synchronously carried out. And if the current node to be frozen reaches the set freezing threshold, the node freezing module freezes the current node to be frozen and reenters the node selection module to perform new-round node selection to be frozen. And after updating the candidate nodes, taking out the head element in the priority queue each time as the next node to be frozen.
(2.2) An evaluation sub-graph generating module for generating an evaluation sub-graph for evaluating whether the node to be frozen can be frozen. The evaluation subgraph is divided into a reference node and a similarity measurement node: the reference node is a copy of the node to be frozen, and the convergence of the node to be frozen is estimated through semantic information of the node to be frozen; the input of the similarity measure node is the intermediate active layer of the node to be frozen and the reference node.
As shown in fig. 4, black nodes represent frozen nodes, and light gray nodes represent nodes to be frozen. In the evaluation sub-graph generation module, the evaluation sub-graph is an extension on the model to be trained, and the connection relation between the evaluation sub-graph and the model to be trained is as follows: the reference node is a copy of the node to be frozen, which has the same input as the node to be frozen; the reference node and an intermediate activation layer of the node to be frozen are used as the input of the similarity measurement node together; the output of the similarity measure node is the output of the evaluation subgraph.
Specifically, the reference node is used for measuring the node to be frozen before a plurality of rounds of iteration, and the weight of the reference node is regularly aligned with the node to be frozen. The intermediate activation layer of the reference node and the node to be frozen contains semantic information, the semantic information is used as the input of the similarity measurement node, and the similarity measurement node evaluates the semantic similarity of the reference node and the node to be frozen, so that the convergence of the node to be frozen is obtained.
The application scene of the neural network model to be trained can be other scenes besides image classification. As shown in fig. 5, a plurality of reference tasks and similarity measure mode selection mechanisms for different application scenarios are constructed in advance. When the network model is transmitted, the optimal measurement mode is adaptively selected according to the application scene to which the model is applicable, and corresponding similarity measurement nodes are generated.
Specifically, the input of the similarity measurement node is an intermediate activation layer of the node to be frozen and the reference node, and the similarity measurement node analyzes the semantic similarity of the node to be frozen and the reference node in a specified measurement mode. And finally obtaining the convergence of the node to be frozen by calculating the semantic similarity of the two, and transmitting the convergence as input to a convergence evaluation module for further evaluation.
And (2.3) a convergence evaluation module, obtaining the convergence of the node to be frozen according to the evaluation subgraph, repeating the steps of the current module until the convergence reaches a set threshold value, and entering the next module.
Specifically, obtaining the convergence degree of the node to be frozen according to the evaluation subgraph, entering the next module if the convergence degree reaches a set threshold value, and otherwise repeating the current module step. The selection of the set threshold is related to the node type and the super parameter setting, and different thresholds are set for different nodes to be frozen.
And (2.4) a node freezing module freezes the node with convergence reaching the expected value, skips the back propagation and parameter updating processes of the node, and realizes acceleration of model training.
Specifically, the node with convergence reaching the expected value is marked as a frozen node, the back propagation and parameter updating of the node are skipped, the ingress of all the subsequent nodes is reduced by one, and if the ingress of the subsequent nodes is reduced to zero, the subsequent nodes are added into a priority queue in the convergence evaluation module.
And step 3, after the neural network model training is finished, returning a required result. And applying the image classification by using the trained model.
Based on the same inventive principles, the embodiment of the invention also provides a device based on the node freezing acceleration neural network training, which comprises a memory and one or more processors, wherein executable codes are stored in the memory, and the one or more processors are used for realizing the method based on the node freezing acceleration neural network training of the embodiment when executing the executable codes.
To verify the effect of the present invention, a classical multi-branch network structure GoogLeNet was used as an experimental network, and the operation results (including training time and accuracy) of the neural network with freezing and the original neural network were compared to obtain the comparison results shown in table 1.
TABLE 1
As can be seen from the graph, the training time length of each training time can be reduced by 15% by the method for accelerating the training of the neural network, and the training time length of the neural network is greatly shortened. Meanwhile, the accelerated neural network training basically has no loss of accuracy, and the average accuracy change on a test set is 0.06%. Therefore, the improved neural network greatly improves the operation speed of the neural network under the condition of basically not sacrificing the accuracy.
The foregoing embodiments have described in detail the technical solution and the advantages of the present invention, it should be understood that the foregoing embodiments are merely illustrative of the present invention and are not intended to limit the invention, and any modifications, additions and equivalents made within the scope of the principles of the present invention should be included in the scope of the invention.

Claims (8)

1. The method for accelerating the neural network training based on the node freezing is characterized by comprising the following steps of:
(1) Acquiring a neural network model to be trained, wherein the neural network model is a multi-branch structure model for image classification;
(2) Initializing the degree of entry of each node in the neural network model as the number of precursor nodes, and training the neural network model by cycling the following steps (2-1) - (2-4);
(2-1) selecting a next node to be frozen from all candidate nodes of the neural network model; the candidate nodes are nodes which meet the condition that the precursors of the candidate nodes are frozen and the candidate nodes are not frozen;
(2-2) generating a corresponding evaluation subgraph according to the node to be frozen, wherein the evaluation subgraph is used for evaluating whether the node to be frozen can be frozen or not;
Wherein, the evaluation subgraph is divided into a reference node and a similarity measurement node: the reference node is a copy of the node to be frozen; the intermediate activation layer of the node to be frozen and the reference node contains semantic information, the semantic information is used as the input of the similarity measurement node together, the similarity measurement node calculates the semantic similarity of the node to be frozen and the reference node, and then the convergence of the node to be frozen is obtained;
(2-3) if the convergence degree of the node to be frozen does not reach the set threshold value, updating the weight of the neural network model, returning to the step (2-2), and if the convergence degree of the node to be frozen reaches the set threshold value, executing the step (2-4);
(2-4) freezing the node with convergence reaching the set threshold value, and skipping the back propagation and weight updating processes of the node to realize acceleration of model training;
(3) And (3) finishing training the neural network model, and applying image classification by using the trained model.
2. The method for training a node freezing-based acceleration neural network according to claim 1, wherein the specific process of the step (2-1) is as follows:
And adding all candidate nodes of the neural network model into a priority queue by adopting a priority queue mode based on a self-defined comparator, and selecting a queue head element each time as a next node to be frozen.
3. The method for accelerating neural network training based on node freezing according to claim 2, wherein the custom comparator of the priority queue needs to comprehensively evaluate the depth of the candidate node and the calculation amount of the candidate node, and the candidate node with smaller depth and smaller calculation amount has higher priority.
4. The method of node-freeze accelerated neural network training of claim 1, wherein in step (2-2), the reference node has the same input as the node to be frozen, and the weight of the reference node is periodically aligned with the node to be frozen.
5. The method for accelerating neural network training based on node freezing according to claim 1, wherein in the step (2-2), the corresponding evaluation subgraph is generated by an evaluation subgraph generation module, and the evaluation subgraph generation module is an extension on the neural network model to be trained.
6. The method for accelerating neural network training based on node freezing according to claim 1, wherein in the step (2-3), the set threshold is selected in association with the type of the node to be frozen and the parameter setting, and the set threshold is different for different nodes to be frozen.
7. The method for training a neural network based on node freezing acceleration according to claim 1, wherein the specific process of the step (2-4) is as follows:
Marking the node with the convergence reaching the set threshold as the frozen node, skipping the back propagation and the weight updating of the node, subtracting one from the ingress degree of all the subsequent nodes, and if the ingress degree of the subsequent nodes is reduced to zero, adding the subsequent nodes into the priority queue in the step (2-3).
8. An apparatus for node-freezing-based accelerated neural network training, comprising a memory and one or more processors, the memory having executable code stored therein, the one or more processors, when executing the executable code, to implement the method for node-freezing-based accelerated neural network training of any of claims 1-7.
CN202410545941.7A 2024-05-06 Method and device for accelerating neural network training based on node freezing Active CN118133929B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202410545941.7A CN118133929B (en) 2024-05-06 Method and device for accelerating neural network training based on node freezing

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202410545941.7A CN118133929B (en) 2024-05-06 Method and device for accelerating neural network training based on node freezing

Publications (2)

Publication Number Publication Date
CN118133929A true CN118133929A (en) 2024-06-04
CN118133929B CN118133929B (en) 2024-08-02

Family

ID=

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109146000A (en) * 2018-09-07 2019-01-04 电子科技大学 A kind of method and device for improving convolutional neural networks based on frost weight
CN109635936A (en) * 2018-12-29 2019-04-16 杭州国芯科技股份有限公司 A kind of neural networks pruning quantization method based on retraining
US20190347538A1 (en) * 2018-05-11 2019-11-14 Qualcomm Incorporated Neural network with frozen nodes
US20200293870A1 (en) * 2019-09-06 2020-09-17 Intel Corporation Partially-frozen neural networks for efficient computer vision systems
CN113177639A (en) * 2020-01-27 2021-07-27 通用电气精准医疗有限责任公司 Freezing as a regularizer in training neural networks
CN116167413A (en) * 2023-04-20 2023-05-26 国网山东省电力公司济南供电公司 Method and system for quantized pruning joint optimization of deep convolutional neural network
US20230316729A1 (en) * 2022-04-01 2023-10-05 Deepmind Technologies Limited Training neural networks
CN117241376A (en) * 2023-09-14 2023-12-15 兰州理工大学 WSN node positioning method for multi-strategy improved BP neural network
US20240064045A1 (en) * 2020-12-30 2024-02-22 Orange Methods and devices for freezing or adapting parameters of a neural network which are used in a telecommunication network

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190347538A1 (en) * 2018-05-11 2019-11-14 Qualcomm Incorporated Neural network with frozen nodes
CN109146000A (en) * 2018-09-07 2019-01-04 电子科技大学 A kind of method and device for improving convolutional neural networks based on frost weight
CN109635936A (en) * 2018-12-29 2019-04-16 杭州国芯科技股份有限公司 A kind of neural networks pruning quantization method based on retraining
US20200293870A1 (en) * 2019-09-06 2020-09-17 Intel Corporation Partially-frozen neural networks for efficient computer vision systems
CN113177639A (en) * 2020-01-27 2021-07-27 通用电气精准医疗有限责任公司 Freezing as a regularizer in training neural networks
US20240064045A1 (en) * 2020-12-30 2024-02-22 Orange Methods and devices for freezing or adapting parameters of a neural network which are used in a telecommunication network
US20230316729A1 (en) * 2022-04-01 2023-10-05 Deepmind Technologies Limited Training neural networks
CN116167413A (en) * 2023-04-20 2023-05-26 国网山东省电力公司济南供电公司 Method and system for quantized pruning joint optimization of deep convolutional neural network
CN117241376A (en) * 2023-09-14 2023-12-15 兰州理工大学 WSN node positioning method for multi-strategy improved BP neural network

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
ZHANG LI: "Answering medical questions in Chinese using automatically mined knowledge and deep neural networks: an end-to-end solution", ARXIV, 22 March 2019 (2019-03-22) *
侯俊龙;潘文林;王璐;何翠玲;王翠;: "基于剪枝AlexNet的普米语孤立词识别", 云南民族大学学报(自然科学版), no. 04, 10 July 2020 (2020-07-10) *
孙荧;王荆;: "基于粒子群优化的神经网络算法的英语翻译", 科学技术与工程, no. 18, 28 June 2020 (2020-06-28) *
平嘉蓉;张正华;沈逸;陈豪;刘源;杨意;尤倩;苏权;: "基于轻量级神经网络的人群计数模型设计", 无线电工程, no. 06, 19 May 2020 (2020-05-19) *

Similar Documents

Publication Publication Date Title
Addanki et al. Placeto: Learning generalizable device placement algorithms for distributed machine learning
CN110533183B (en) Task placement method for heterogeneous network perception in pipeline distributed deep learning
CN111832627B (en) Image classification model training method, classification method and system for suppressing label noise
CN113191484A (en) Federal learning client intelligent selection method and system based on deep reinforcement learning
CN111027732B (en) Method and system for generating multi-wind power plant output scene
Zheng et al. Model compression based on differentiable network channel pruning
CN113469325A (en) Layered federated learning method, computer equipment and storage medium for edge aggregation interval adaptive control
CN109558898B (en) Multi-choice learning method with high confidence based on deep neural network
CN113971089A (en) Method and device for selecting equipment nodes of federal learning system
CN113159287B (en) Distributed deep learning method based on gradient sparsity
CN115361301B (en) Distributed computing network cooperative traffic scheduling system and method based on DQN
CN117150821B (en) Construction method of equipment efficiency evaluation data set based on intelligent simulation
Addanki et al. Placeto: Efficient progressive device placement optimization
WO2021055442A1 (en) Small and fast video processing networks via neural architecture search
CN114897274A (en) Method and system for improving time sequence prediction effect
CN118133929B (en) Method and device for accelerating neural network training based on node freezing
CN111913887B (en) Software behavior prediction method based on beta distribution and Bayesian estimation
CN116502774B (en) Time sequence prediction method based on time sequence decomposition and Legend projection
CN118133929A (en) Method and device for accelerating neural network training based on node freezing
CN115577797B (en) Federal learning optimization method and system based on local noise perception
Norouzi et al. A survey on proposed methods to address Adam optimizer deficiencies
CN115345303A (en) Convolutional neural network weight tuning method, device, storage medium and electronic equipment
CN115796006A (en) New energy storage power station wind speed prediction method based on WOA-ELM prediction model
CN114722490A (en) Agent model global optimization method based on mixed increase and interval reduction
CN115129471A (en) Distributed local random gradient descent method for large-scale GPU cluster

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant