CN117852588A - AI chip evaluation parameter determination method based on deep residual neural network - Google Patents

AI chip evaluation parameter determination method based on deep residual neural network Download PDF

Info

Publication number
CN117852588A
CN117852588A CN202311783926.8A CN202311783926A CN117852588A CN 117852588 A CN117852588 A CN 117852588A CN 202311783926 A CN202311783926 A CN 202311783926A CN 117852588 A CN117852588 A CN 117852588A
Authority
CN
China
Prior art keywords
neural network
layer
deep residual
residual neural
fdnet
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202311783926.8A
Other languages
Chinese (zh)
Inventor
梁培哲
祝名
李爽
张大宇
万旺
张松
王贺
焦美荣
王征
屈若媛
杨舒文
刘一帆
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Academy of Space Technology CAST
Original Assignee
China Academy of Space Technology CAST
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Academy of Space Technology CAST filed Critical China Academy of Space Technology CAST
Priority to CN202311783926.8A priority Critical patent/CN117852588A/en
Publication of CN117852588A publication Critical patent/CN117852588A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses an AI chip evaluation parameter determining method based on a deep residual neural network, which comprises the following steps: designing a deep residual neural network FDNet, training the deep residual neural network FDNet to obtain a floating point model, and obtaining an evaluation parameter under a target classification task; quantifying the floating point model to obtain a fixed point model, deploying the fixed point model on a hardware board, and obtaining an evaluation parameter under a target classification task through reasoning on the hardware board; the obtained evaluation parameters obtained by the floating point model and the fixed point model are used for evaluating the application performance of the AI chip when the deep residual neural network FDNet is used. The method and the device can be applied to various aerospace data sets, and have strong robustness.

Description

AI chip evaluation parameter determination method based on deep residual neural network
Technical Field
The invention belongs to the technical field of integrated circuit evaluation, and particularly relates to an AI chip evaluation parameter determination method based on a deep residual error neural network.
Background
Current artificial intelligence technology is vigorously developed, and big data, deep learning algorithms and hardware on which the technology depends are updated in a daily and monthly manner. In terms of algorithms, various brand-new neural network algorithms are proposed, and most of these algorithms require higher quality data requirements. The ability of hardware is often the bottleneck for such high data volume, high computation-volume tasks. The increasing interest in deep learning has stimulated the design of smart chips to a great extent, and various companies have proposed corresponding hardware designs. With the increasing number of AI chips in aerospace, how to systematically evaluate the suitability of aerospace intelligent algorithms of the chips becomes a research hot spot.
Along with the rapid development of the deep learning network model, the difference between models in different scenes is larger and larger, and the structure of the network model is gradually complicated and diversified from a convolutional neural network to a transfer model. At present, one of the fields with more AI chips is the target classification field, such as planetary detection, space debris removal, satellite on-orbit service, spacecraft intersection and docking and other tasks.
The algorithm models in the target classification are various, the operator differences among the models are larger and larger due to the complicated and diversified structures of the network models, the existing evaluation reference program is not applied to aerospace, the operator types of the aerospace typical algorithm can be contained are not more, and some operators are not commonly used in aerospace, so that the operator performance of the AI chip cannot be well evaluated by using the existing evaluation reference program. There is no unified standard for the use of network models in the current evaluation of the suitability of the intelligent algorithm of the AI chip, and most of the used models are still common models on the ground. Therefore, it is needed to design an AI chip intelligent algorithm suitable for aerospace applications, which is used in the evaluation work of the AI chip for aerospace applications.
Disclosure of Invention
The invention solves the technical problems that: the AI algorithm suitability of the chip for aerospace is evaluated by comparing the application performance, operator performance, quantization performance and the like of a floating point model and a quantization model.
The technical scheme of the invention is as follows: an AI chip evaluation parameter determining method based on deep residual neural network comprises the following steps:
designing a deep residual neural network FDNet, training the deep residual neural network FDNet to obtain a floating point model, and obtaining an evaluation parameter under a target classification task;
quantifying the floating point model to obtain a fixed point model, deploying the fixed point model on a hardware board, and obtaining an evaluation parameter under a target classification task through reasoning on the hardware board;
the obtained evaluation parameters obtained by the floating point model and the fixed point model are used for evaluating the application performance of the AI chip when the deep residual neural network FDNet is used.
The evaluation parameters include: accuracy Top-k, average accuracy mean mAP, and error rate indicator.
The operators used in the deep residual neural network FDNet are all common operators in the field of aerospace.
The deep residual neural network FDNet consists of 4 deep residual modules; the first layer and the third layer of the residual error module respectively use a convolution layer, a batch normalization layer and a rectification linear unit layer, the second layer consists of the convolution layer, the batch normalization layer, the rectification linear unit layer and an attention channel layer, and the fourth layer consists of the convolution layer and the batch normalization layer; the output part of the deep residual neural network FDNet comprises a convolution layer and a full connection layer; the classification layer of the deep residual neural network FDNet is a softmax layer.
And training the deep residual neural network FDNet network through the ImageNet data set to obtain a floating point model.
And when the floating point model is quantized to obtain the fixed point model, an AI chip self-contained software tool is adopted for quantization, and an int8 quantization mode is adopted.
Compared with the prior art, the invention has the beneficial effects that:
1. and the FDNet model is utilized, so that operator performance evaluation is more comprehensive in the evaluation of the space navigation AI chip. The FDNet model contains more space navigation common operators. Operator performance includes diversity of operators and support of operators.
2. The model is good in robustness, and can be suitable for different aerospace data sets and used for training and evaluating the model in the target classification task.
3. The target recognition accuracy is higher, and the FDNet model in the invention can obtain higher target recognition accuracy.
Drawings
Fig. 1 is an evaluation algorithm model of an AI chip for aerospace based on a deep residual neural network.
Fig. 2 is a flowchart of a method for determining an AI chip evaluation parameter based on a deep residual neural network.
Detailed Description
The intelligent algorithm suitability evaluation algorithm of the AI chip is designed in the field of target identification. Deep residual neural networks named FDNet (Feature Denoising Residual Network, FDNet) were designed.
Step one:
the algorithm network in the invention is named as FDNet, is an algorithm model based on a depth residual neural network, and extracts key features by utilizing a multi-layer residual neural network as shown in figure 1. FDNet consists of D blocks of depth residuals, where d=4, capturing the difference between the input and the ground truth value using elemental subtraction. FDNet adopts three-channel design, and related features are extracted from RGB channels of the picture respectively. In this architecture, advanced spatial features in the picture are captured from the input through 4 residual modules, helping to enhance the denoising ability of the model, and also ensuring the accuracy and depth of feature extraction.
The first and third layers of the residual block each use a combination operation including convolution (Conv), batch Normalization (BN), and rectifying linear units (ReLU). The batch normalization not only greatly improves the training speed of the model, but also enhances the stability of the model under various data conditions. The second layer is a channel attention layer, consisting of convolution (Conv), batch Normalization (BN), rectifying linear units (ReLU) and an attention channel layer. The fourth layer is composed of Conv layer and BN layer and is mainly responsible for generating a noise matrix, i.e. a noise matrix is generated before the subsequent element subtraction operation. This design ensures that not only noise is removed but also features are corrected and optimized during denoising, taking into account the various noise disturbances to which the image may be subjected.
In the design of FDNet, the output layer is a core part for realizing image recognition and plays a crucial role. Through the convolution and pooling operations of the previous layers, the local features of the image are extracted in a rich way. The design of the output layer is particularly critical in order to convert these extracted features into an explicit classification result. Specifically, the output portion of the network first includes a convolutional layer. This convolution layer further extracts advanced features of the image and prepares for the next full connection layer. Then, a full connection layer is introduced, and plays a role in the neural network, which can flatten the feature vector extracted by the front layer and map the feature vector to a feature space with a fixed length, so that preparation is made for final classification. In this feature space, each neuron can be considered as a weight corresponding to a particular class, and the result of multiplication with it is the score of that class. Finally, the output of the fully connected layer is translated into a probability distribution for each category using the softmax layer. The softmax layer can translate consecutive values into probability distributions, ensuring that the sum of probabilities for all classes is 1. Network training is performed using an ImageNet dataset, and first ImageNet is screened. To ensure training efficiency and generalization ability of the model, ten thousand representative images were selected. The training process is divided into the following stages:
1. pretreatment: to speed up training and improve network convergence, all training images are normalized, i.e., the mean is subtracted and divided by the standard deviation. And the data set pictures are subjected to random cutting, rotation and overturning so as to enhance the robustness of the model.
2. Training strategies: a random gradient descent (SGD) was used as an optimizer, and the initial learning rate was set to 0.1 and gradually decreased according to a predetermined strategy. To prevent overfitting, L2 regularization was added, and Momentum strategy with Momentum of 0.9 was used to accelerate convergence.
3. And (3) learning rate adjustment: the learning rate is reduced to 1/10 of the original learning rate after a certain number of training cycles.
4. And (3) verification and adjustment: at the end of each training period, the performance of the model is evaluated using the validation set. And adjusting the network structure and parameters according to the verification result.
The operator classes contained in the FDNet model are listed in table 1.
Table 1 operator species contained in FDNet model
The training, testing and verifying work of the FDNet model is completed by using a data set, wherein the adopted data set can be an open source data set or a training sample formed by actually measured pictures in a space environment. After the FDNet model is trained, three indexes of TOP1, TOP5 and error rate of the floating point model are obtained, wherein TOP1 is the highest output of the label of the picture and the model prediction is consistent, TOP5 is the label of the picture in the first five labels of the model prediction, and the error rate is the label of the picture is not in the first five results of the model output.
Step two:
after the floating point model is obtained, the model is deployed to a hardware board by using a software compiling tool chain. The hardware board card can be selected according to the needs, and the European bit jade yulon 810A board card is used for deployment in the patent. The FDNet model is generated through a PyTorch deep learning training framework, so that the network model after FDNet training is imported into YL-ACUITY through ONNX conversion, 8bit quantization is carried out on the network model, and an ovxlib-based C language code is generated through an export function. And testing on the board card to obtain three index results of TOP1, TOP5 and error rate under the actual measurement model.
The invention is further described below with reference to the accompanying drawings.
The invention discloses an AI chip evaluation parameter determination method based on a deep residual neural network, and the flow is shown in fig. 2.
Firstly, training, testing and verifying the FDNet model by using a data set, and obtaining three indexes of TOP1, TOP5 and error rate of the floating point model after the training of the FDNet model is completed. After the floating point model is obtained, the model is deployed to a hardware board by using a software compiling tool chain. And carrying out reasoning test on the board card to obtain three index results of TOP1, TOP5 and error rate under the actual measurement model. Table 2 shows the comparison results, comparing the test results of the FDNet model with the results of the Resnet50 model, and comparing the measured model with the results of the floating point model. It can be seen that the recognition accuracy of the FDNet model is higher than that of the Resnet50, and the target recognition task can be completed well. Meanwhile, the AI chip used for supporting all operators in the FDNet model can support more common operators in the astronavigation target recognition task. And compared with floating point and actual measurement model results, the 8bit quantization precision loss of the AI chip is smaller, and the model can be quantized through the chip, so that the complexity is reduced. The evaluation results of the AI chip show that the algorithm model designed by the invention can be applied to the evaluation work of the AI chip for aerospace.
TABLE 2 comparison of Floating Point model and measured model results
The present invention is not described in detail as being well known to those skilled in the art.

Claims (6)

1. An AI chip evaluation parameter determining method based on a deep residual neural network is characterized by comprising the following steps:
designing a deep residual neural network FDNet, training the deep residual neural network FDNet to obtain a floating point model, and obtaining an evaluation parameter under a target classification task;
quantifying the floating point model to obtain a fixed point model, deploying the fixed point model on a hardware board, and obtaining an evaluation parameter under a target classification task through reasoning on the hardware board;
the obtained evaluation parameters obtained by the floating point model and the fixed point model are used for evaluating the application performance of the AI chip when the deep residual neural network FDNet is used.
2. The AI chip evaluation parameter determination method based on the deep residual neural network of claim 1, wherein the evaluation parameter comprises: accuracy Top-k, average accuracy mean mAP, and error rate indicator.
3. The AI chip evaluation parameter determination method based on the deep residual neural network according to claim 1, wherein operators used in the deep residual neural network FDNet are all operators commonly used in the aerospace field.
4. The AI chip evaluation parameter determination method based on the deep residual neural network according to claim 1, wherein the deep residual neural network FDNet is composed of 4 deep residual modules; the first layer and the third layer of the residual error module respectively use a convolution layer, a batch normalization layer and a rectification linear unit layer, the second layer consists of the convolution layer, the batch normalization layer, the rectification linear unit layer and an attention channel layer, and the fourth layer consists of the convolution layer and the batch normalization layer; the output part of the deep residual neural network FDNet comprises a convolution layer and a full connection layer; the classification layer of the deep residual neural network FDNet is a softmax layer.
5. The AI chip evaluation parameter determination method based on the deep residual neural network according to claim 1, wherein the deep residual neural network FDNet network is trained through an ImageNet data set to obtain a floating point model.
6. The method for determining the evaluation parameters of the AI chip based on the deep residual neural network according to claim 1, wherein when the floating point model is quantized to obtain the fixed point model, an AI chip-carried software tool is adopted for quantization, and an int8 quantization mode is adopted.
CN202311783926.8A 2023-12-22 2023-12-22 AI chip evaluation parameter determination method based on deep residual neural network Pending CN117852588A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311783926.8A CN117852588A (en) 2023-12-22 2023-12-22 AI chip evaluation parameter determination method based on deep residual neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311783926.8A CN117852588A (en) 2023-12-22 2023-12-22 AI chip evaluation parameter determination method based on deep residual neural network

Publications (1)

Publication Number Publication Date
CN117852588A true CN117852588A (en) 2024-04-09

Family

ID=90537300

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311783926.8A Pending CN117852588A (en) 2023-12-22 2023-12-22 AI chip evaluation parameter determination method based on deep residual neural network

Country Status (1)

Country Link
CN (1) CN117852588A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN118393329A (en) * 2024-06-27 2024-07-26 中国信息通信研究院 System for testing AI chip in model training and reasoning performance

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN118393329A (en) * 2024-06-27 2024-07-26 中国信息通信研究院 System for testing AI chip in model training and reasoning performance

Similar Documents

Publication Publication Date Title
CN113641820B (en) Visual angle level text emotion classification method and system based on graph convolution neural network
CN113221687B (en) Training method of pressing plate state recognition model and pressing plate state recognition method
CN114743020B (en) Food identification method combining label semantic embedding and attention fusion
CN111966812B (en) Automatic question answering method based on dynamic word vector and storage medium
CN117852588A (en) AI chip evaluation parameter determination method based on deep residual neural network
CN115222950A (en) Lightweight target detection method for embedded platform
CN114821340A (en) Land utilization classification method and system
CN115019132B (en) Multi-target identification method for complex background ship
US20210056353A1 (en) Joint representation learning from images and text
CN114742211B (en) Convolutional neural network deployment and optimization method facing microcontroller
CN110991247B (en) Electronic component identification method based on deep learning and NCA fusion
CN114863407A (en) Multi-task cold start target detection method based on visual language depth fusion
CN117496124A (en) Large-area photovoltaic panel detection and extraction method based on deep convolutional neural network
Jang et al. Decision fusion approach for detecting unknown wafer bin map patterns based on a deep multitask learning model
CN117058716A (en) Cross-domain behavior recognition method and device based on image pre-fusion
CN111783688A (en) Remote sensing image scene classification method based on convolutional neural network
CN116883792A (en) Aurora ovum intensity image modeling method based on FA-GAN
CN109934248B (en) Multi-model random generation and dynamic self-adaptive combination method for transfer learning
CN115019183B (en) Remote sensing image model migration method based on knowledge distillation and image reconstruction
CN114357160B (en) Early rumor detection method and device based on generated propagation structural features
CN115358473A (en) Power load prediction method and prediction system based on deep learning
CN109543571A (en) A kind of intelligent recognition and search method of Complex Product abnormity machining feature
CN113743593A (en) Neural network quantization method, system, storage medium and terminal
CN116883709A (en) Carbonate fracture-cavity identification method and system based on channel attention mechanism
CN114282585A (en) Astronomical spectrum data-based outlier celestial body classification method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination