CN115565020A - Tooth surface damage identification method and device based on improved neural network - Google Patents

Tooth surface damage identification method and device based on improved neural network Download PDF

Info

Publication number
CN115565020A
CN115565020A CN202211414681.7A CN202211414681A CN115565020A CN 115565020 A CN115565020 A CN 115565020A CN 202211414681 A CN202211414681 A CN 202211414681A CN 115565020 A CN115565020 A CN 115565020A
Authority
CN
China
Prior art keywords
dimensional feature
feature map
carrying
tooth surface
dimensional
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211414681.7A
Other languages
Chinese (zh)
Inventor
吴鲁纪
秦佳音
杨林杰
曹延军
孙毅博
陈俊豪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zheng Ji Suo Zhengzhou Transmission Technology Co ltd
Original Assignee
Zhengzhou Research Institute of Mechanical Engineering Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhengzhou Research Institute of Mechanical Engineering Co Ltd filed Critical Zhengzhou Research Institute of Mechanical Engineering Co Ltd
Priority to CN202211414681.7A priority Critical patent/CN115565020A/en
Publication of CN115565020A publication Critical patent/CN115565020A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Image Analysis (AREA)

Abstract

A tooth surface damage identification method and a tooth surface damage identification device based on an improved neural network belong to the field of tooth surface damage identification of gears. According to the method, a neural network model is improved, the global average pooling is used for adapting to the tooth surface size and accelerating training, and the model is converged more quickly by applying transfer learning; and finally, loading the data set into an improved ResNet-34 to extract and train image features, classifying by using a classification function to finally obtain a tooth surface damage classification model, and inputting the gear surface image to be detected into the model, so that the precision of damage identification can be ensured, the model is effectively compressed, the training cost is reduced, and the identification efficiency is greatly improved.

Description

Tooth surface damage identification method and device based on improved neural network
Technical Field
The invention relates to the field of tooth surface damage identification of gears, in particular to a tooth surface damage identification method and device based on an improved neural network.
Background
Gears are key components of mechanical transmission systems, the quality of which directly affects the operating conditions and the service life of the transmission system. However, due to manufacturing and installation errors, motion impact, sliding friction, cyclic alternating bending stress, material properties and heat treatment processes, different failure modes of the gear may occur, and the failure modes are commonly represented by tooth surface pitting, tooth surface abrasion, tooth breakage and the like. The damage of the tooth surface can cause vibration and noise, reduce the service life and influence the motion precision and the working stability of the gear.
At present, the identification of the tooth surface damage type mainly depends on the visual detection of workshop workers and maintenance personnel, and has low efficiency, poor real-time performance and higher requirement on the workers. With the rapid development of computer technology and artificial intelligence, machine identification gradually replaces manual operation, but at present, no unified standard or method exists at home and abroad for identifying the tooth surface damage type of the gear.
Disclosure of Invention
The invention aims to provide a tooth surface damage identification method based on an improved neural network, which adopts a machine vision system to replace manual operation, and improves the accuracy and efficiency of gear tooth surface damage identification.
The technical scheme adopted by the invention for realizing the technical purpose is as follows: a tooth surface damage identification method based on an improved neural network is characterized in that the improved neural network is utilized to establish a tooth surface damage classification model, and then a gear surface image to be detected is input into the tooth surface damage classification model to obtain an analysis result of the gear surface damage; the steps of establishing the tooth surface damage classification model by utilizing the improved neural network are as follows:
firstly), acquiring a tooth surface image of a gear by using a machine vision system, classifying the acquired tooth surface image according to the characteristics of tooth surface damage, and marking a real label of each classification;
secondly), cutting the tooth surface images of each type to the same size, and performing image enhancement to form a classified tooth surface image data set;
thirdly), constructing a convolutional neural network by taking ResNet-34 as a primary network model, training the constructed convolutional neural network by utilizing a tooth surface image data set, and storing formed weights into the constructed convolutional neural network;
fourthly) carrying out classification processing on the output result of the convolutional neural network by using a classification function softmax to obtain probability vectors of each classification;
and fifthly), selecting the maximum value in the probability vector as a prediction type, and outputting the prediction type and the prediction probability as an analysis result of the gear surface damage.
As an optimization scheme of the tooth surface damage identification method based on the improved neural network, the step one) of classifying the acquired tooth surface images means that the acquired tooth surface images are classified according to four types of tooth surface pitting, tooth surface abrasion, gear tooth breakage and normal tooth surface, and the four types are used as real labels for corresponding classification.
As another optimization scheme of the tooth surface damage identification method based on the improved neural network, the image enhancement in the step two) is sequentially performed by horizontal turning, vertical turning, 180-degree rotation, zooming, shifting, blurring, brightening, darkening and salt and pepper noise adding.
As another optimization scheme of the tooth surface damage identification method based on the improved neural network, in the second step), a classified tooth surface image data set is formed, and the proportion of the training set to the test set is 9:1.
as another optimization scheme of the tooth surface damage identification method based on the improved neural network, the convolutional neural network in the step three) sequentially comprises an input layer, a first convolutional layer, a second convolutional layer, a third convolutional layer, a fourth convolutional layer, a fifth convolutional layer, a global average pooling layer and a full connection layer.
As another optimization scheme of the tooth surface damage identification method based on the improved neural network, the specific operation of constructing the convolutional neural network in the step three) is as follows:
1) The first winding layer
1.1 Passing the image with the dimension of (224, 64) through 64 convolution kernels of 7 by 7 and performing convolution operation with the step size of 2, and performing zero filling operation for 3 circles at the periphery of the obtained pixel matrix to obtain a three-dimensional characteristic map of (112, 64);
1.2 Carrying out batch normalization operation on the obtained three-dimensional characteristic graph, and then carrying out nonlinear mapping operation on the three-dimensional characteristic graph subjected to batch normalization through a ReLU activation function;
1.3 Maximum pooling operation of the signature output from step 1.2) by 3 × 3 convolution kernels and in 2 steps, followed by 1 cycle of zero padding operation at the periphery of the resulting pixel matrix to obtain a three-dimensional signature (56, 64);
2) The second convolution layer
2.1 Carrying out convolution on the three-dimensional feature map (56, 64) obtained in the previous step by 64 convolution kernels of 3 x 3 and with the step size of 1, and carrying out 1-turn zero padding operation on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map (56, 64);
2.2 Batch normalization operation is carried out on the three-dimensional feature maps (56, 64) obtained in the previous step, and then nonlinear mapping operation is carried out on the batch normalized three-dimensional feature maps through a ReLU activation function;
2.3 ) repeating the convolution and zero filling operation in the step 2.1) on the three-dimensional characteristic diagram obtained in the previous step, and carrying out batch normalization operation on the obtained three-dimensional characteristic diagram;
2.4 Adding the three-dimensional characteristic diagram input in the step 2.1) and the three-dimensional characteristic diagram output in the step 2.3), and then carrying out nonlinear mapping operation on the three-dimensional characteristic diagram obtained after addition through a ReLU activation function;
2.5 Repeating the step 2.1) to the step 2.4) twice, and outputting a three-dimensional feature map with dimensions (56, 64);
3) The third convolution layer
3.1 The three-dimensional characteristic map with the dimension of (56, 64) obtained in the previous step is convoluted by 128 convolution kernels of 3 x 3 and step size of 2, and zero padding operation is carried out for 1 circle at the periphery of the obtained pixel matrix to obtain a three-dimensional characteristic map of (28, 128);
3.2 Carrying out batch normalization operation on the three-dimensional feature maps (28, 128) obtained in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature maps after batch normalization through a ReLU activation function to obtain three-dimensional feature maps with the dimension of (28, 64);
3.3 Carrying out convolution on the three-dimensional feature map with the dimension of (28, 64) obtained in the previous step through 128 convolution kernels of 3 x 3 and the step size of 1, carrying out zero filling operation for 1 circle on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map with the dimension of (28, 128), and carrying out batch normalization operation on the three-dimensional feature map;
3.4 Carrying out convolution on the three-dimensional feature map input in the step 3.1) through 128 convolution kernels of 1 × 1 and with the step size of 2 to obtain a (28, 128) three-dimensional feature map, adding the three-dimensional feature map and the feature map obtained in the step 3.3), and then carrying out nonlinear mapping operation on the three-dimensional feature map obtained after addition through a ReLU activation function to obtain a three-dimensional feature map with the dimension of (28, 128);
3.5 Passing the three-dimensional feature map (28, 128) obtained in the previous step through 128 convolution kernels of 3 × 3 and performing convolution with a step size of 1, and performing zero padding operation for 1 circle on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map (28, 128);
3.6 Carrying out batch normalization operation on the three-dimensional characteristic diagram output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional characteristic diagram after batch normalization through a ReLU activation function;
3.7 Step 3.5) is repeated and the obtained three-dimensional characteristic diagram is subjected to batch normalization operation;
3.8 Adding the three-dimensional feature map output in the previous step with the three-dimensional feature map with the dimension (28, 128) input in the step 3.5), and then carrying out nonlinear mapping operation on the three-dimensional feature map obtained after addition through a ReLU activation function;
3.9 Repeating the step 3.5) to the step 3.8) twice, and outputting a three-dimensional feature map with the dimension of (28, 128);
4) The fourth convolution layer
4.1 Passing the three-dimensional feature map obtained in the previous step through 256 convolution kernels of 3 x 3 and performing convolution by 2 steps, and performing zero padding operation for 1 circle on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map (14, 256);
4.2 Carrying out batch normalization operation on the three-dimensional feature map (14, 256) obtained in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature map after batch normalization through a ReLU activation function to obtain a three-dimensional feature map with the dimension of (14, 256);
4.3 Passing the three-dimensional feature map obtained in the previous step through 256 convolution kernels of 3 x 3 and performing convolution by step size of 1, performing zero padding operation for 1 circle on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map (14, 256), and performing batch normalization operation on the three-dimensional feature map;
4.4 Carrying out convolution on the three-dimensional feature map input in the step 4.1) through 256 convolution kernels of 1 × 1 and in steps of 2, adding the three-dimensional feature map with the dimensionality (14, 256) and the three-dimensional feature map output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature map obtained after addition through a ReLU activation function;
4.5 Passing the three-dimensional feature map output in the previous step through 256 convolution kernels of 3 x 3 and performing convolution by step size of 1, and performing zero filling operation for 1 circle on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map (14, 256);
4.6 Carrying out batch normalization operation on the three-dimensional characteristic diagram output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional characteristic diagram after batch normalization through a ReLU activation function;
4.7 Step 4.5) is repeated and batch normalization operation is carried out on the obtained three-dimensional characteristic diagram;
4.8 Adding the three-dimensional characteristic diagram input in the step 4.5) with the three-dimensional characteristic diagram output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional characteristic diagram obtained after addition through a ReLU activation function;
4.9 Repeating the step 4.5) to the step 4.8) four times, and outputting a three-dimensional feature map with the dimension of (14, 256);
5) The fifth convolution layer
5.1 Carrying out convolution on the three-dimensional feature map obtained in the previous step by 512 convolution kernels of 3 x 3 and 2 steps, and carrying out 1-turn zero filling operation on the periphery of the obtained pixel matrix to obtain a (7, 512) three-dimensional feature map;
5.2 Carrying out batch normalization operation on the three-dimensional characteristic diagram obtained in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional characteristic diagram after batch normalization through a ReLU activation function to obtain a three-dimensional characteristic diagram with the dimensionality of (7, 512);
5.3 Carrying out convolution on the three-dimensional feature map obtained in the previous step by 512 convolution kernels of 3 x 3 and step length of 1, carrying out 1-turn zero padding operation on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map of (7, 512), and carrying out batch normalization operation on the three-dimensional feature map;
5.4 Carrying out convolution on the three-dimensional feature map input in the step 5.1) through 512 convolution kernels of 1 × 1 and with the step size of 2 to obtain a three-dimensional feature map with the dimensionality of (7, 512) and the three-dimensional feature map output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature map obtained after the addition through a ReLU activation function;
5.5 The three-dimensional feature map output in the previous step is convoluted by 512 convolution kernels of 3 x 3 and the step size of 1, and the zero filling operation is carried out for 1 circle at the periphery of the obtained pixel matrix, so as to obtain the three-dimensional feature map of (7, 512);
5.6 Carrying out batch normalization operation on the three-dimensional feature maps output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature maps subjected to batch normalization through a ReLU activation function;
5.7 Step 5.5) is repeated and batch normalization operation is carried out on the obtained three-dimensional characteristic diagram;
5.8 Adding the three-dimensional characteristic diagram input in the step 5.5) with the three-dimensional characteristic diagram output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional characteristic diagram obtained after addition through a ReLU activation function;
5.9 Repeating the step 5.5) to the step 5.8) once, and outputting a three-dimensional feature map with the dimension of (7, 512);
6) Global average pooling layer
Carrying out global average pooling operation on the three-dimensional feature map with the dimension (7, 512) output in the previous step, and outputting the three-dimensional feature map with the dimension (1, 512);
7) Full connection layer
And converting the three-dimensional feature map with the dimension of (1, 512) output in the previous step into a 4-dimensional feature vector.
As another optimization scheme of the tooth surface damage identification method based on the improved neural network, after batch normalization operations in step 1.2), step 2.2), step 2.3), step 3.2), step 3.3), step 3.6), step 3.7), step 4.2), step 4.3), step 4.6), step 4.7), step 5.2), step 5.3), step 5.6) and step 5.7), pruning operation with a pruning rate of 20% is performed on each batch normalization layer.
As another optimization scheme of the tooth surface damage identification method based on the improved neural network, the training of the constructed convolutional neural network and the saving of the formed weights in the constructed convolutional neural network comprise the following specific operations:
a) Inputting a training set in the tooth surface image data set into a built convolutional neural network for training, performing transfer learning by using official weight, and finally forming a plurality of output numerical values corresponding to the classification quantity;
b) Use of
Figure BDA0003939284910000061
A function, converting a plurality of output values into relative probabilities to obtain probability vectors corresponding to each classification, and selecting the classification with the highest probability as a predicted value of the output;
in the formula, i is a category index, C is a category number, and is the output of a preceding stage output unit of the classifier of the full connection layer, and e is a natural base number;
c) Inputting the obtained predicted value and the real label into a cross entropy function cross EntropyLoss to calculate an error value, calculating the gradient value of each parameter through an error back propagation function loss.
d) Inputting the test set into a convolutional neural network to verify the training effect, calculating an error value according to a cross entropy function, outputting the type with the maximum probability as a predicted value, judging that the prediction is correct if the predicted value is the same as the real label, and determining the ratio of the number of correct predictions to the number of test sets as the accuracy of the whole test set;
e) Training for 400 rounds, wherein the lowest error value on the set to be tested is 0.095, the accuracy rate reaches 97.7%, and the trained weight is stored in the convolutional neural network.
An electronic device comprising a processor and a memory, the memory storing computer readable instructions that, when executed by the processor, are capable of executing the above-described tooth flank damage identification method.
A readable storage medium, on which a computer program is stored which, when being executed by a processor, performs the steps of the above-mentioned tooth flank damage identification method.
Compared with the prior art, the invention has the following beneficial effects:
according to the method, a neural network model is improved, parameters and calculated amount are reduced by a channel pruning method within a range allowed by precision errors, global average pooling is used for adapting to tooth surface size and accelerating training, and the model is converged more quickly by applying transfer learning; and finally, loading the data set into an improved ResNet-34 to extract and train image features, classifying by using a classification function to finally obtain a tooth surface damage classification model, and inputting the gear surface image to be detected into the model, so that the precision of damage recognition can be ensured, the model is effectively compressed, the training cost is reduced, and the recognition efficiency is greatly improved.
Drawings
FIG. 1 is a schematic illustration of a tooth surface damage classification model processing a tooth surface image in accordance with the present invention;
Detailed Description
The technical solution of the present invention is further described in detail with reference to specific embodiments, and for example, for denoising, enhancing and edge detecting of an image, classification and identification by a classifier, batch normalization operation, nonlinear mapping by a ReLU activation function, pruning operation, and other operations used such as Softmax function, cross entropy function cross entropy, error back propagation function low, gradient descent function optimization step () are understood as the prior art known or should be known to those skilled in the art.
Example 1
A tooth surface damage identification method based on an improved neural network is disclosed, as shown in figure 1, the method utilizes the improved neural network to establish a tooth surface damage classification model, and then obtains an analysis result of the surface damage of a gear by inputting a gear surface image to be detected into the tooth surface damage classification model; the steps of establishing the tooth surface damage classification model by utilizing the improved neural network are as follows:
firstly), acquiring a tooth surface image of a gear by using a machine vision system, classifying the acquired tooth surface image according to the characteristics of tooth surface damage, and labeling a real label of each classification;
in practice, the damage identification and recognition can be carried out manually, and classification and real labels are marked;
classifying the acquired tooth surface images in the step means that the acquired tooth surface images are classified according to four types of tooth surface pitting, tooth surface abrasion, gear tooth breakage and normal tooth surface, and the four types are used as real labels of corresponding classification;
secondly), cutting the tooth surface images of each type to the same size, and performing image enhancement to form a classified tooth surface image data set;
in the step, the image enhancement method comprises the steps of horizontal turning, vertical turning, 180-degree rotation, zooming, shifting, blurring, brightening, darkening and adding salt and pepper noise in sequence;
in the tooth surface image data set for which classification is formed in this step, the ratio of the training set to the test set is 9:1;
thirdly), constructing a convolutional neural network by taking ResNet-34 as a primary network model, training the constructed convolutional neural network by utilizing a tooth surface image data set, and storing formed weights into the constructed convolutional neural network;
the convolutional neural network in the step sequentially comprises an input layer, a first convolutional layer, a second convolutional layer, a third convolutional layer, a fourth convolutional layer, a fifth convolutional layer, a global average pooling layer and a full-connection layer;
the specific operation of constructing the convolutional neural network in this step is, as shown in fig. 1, as follows:
1) The first winding layer
1.1 Passing the image with the dimension of (224, 64) through 64 convolution kernels of 7 by 7 and performing convolution operation with the step size of 2, and performing zero filling operation for 3 circles at the periphery of the obtained pixel matrix to obtain a three-dimensional characteristic map of (112, 64);
1.2 Carrying out batch normalization operation on the obtained three-dimensional characteristic diagram, and then carrying out nonlinear mapping operation on the batch normalized three-dimensional characteristic diagram through a ReLU activation function;
1.3 Maximum pooling operation of the signature output from step 1.2) by 3 × 3 convolution kernels and in 2 steps, followed by 1 cycle of zero padding operation at the periphery of the resulting pixel matrix to obtain a three-dimensional signature (56, 64);
2) A second convolution layer
2.1 The three-dimensional characteristic diagram (56, 64) obtained in the previous step is convoluted by 64 convolution kernels of 3 x 3 and with the step size of 1, and the zero padding operation is carried out for 1 circle on the periphery of the obtained pixel matrix to obtain the three-dimensional characteristic diagram (56, 64);
2.2 Carrying out batch normalization operation on the three-dimensional feature maps (56, 56 and 64) obtained in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature maps after batch normalization through a ReLU activation function;
2.3 ) repeating the convolution and zero filling operation in the step 2.1) on the three-dimensional characteristic diagram obtained in the previous step, and carrying out batch normalization operation on the obtained three-dimensional characteristic diagram;
2.4 Adding the three-dimensional characteristic diagram input in the step 2.1) and the three-dimensional characteristic diagram output in the step 2.3), and then carrying out nonlinear mapping operation on the three-dimensional characteristic diagram obtained after addition through a ReLU activation function;
2.5 Repeating the step 2.1) to the step 2.4) twice, and outputting a three-dimensional feature map with dimensions (56, 64);
3) The third convolution layer
3.1 Carrying out convolution on the three-dimensional feature map with the dimension (56, 64) obtained in the previous step through 128 convolution kernels with 3 x 3 and the step size of 2, and carrying out 1-turn zero padding operation on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map (28, 128);
3.2 Carrying out batch normalization operation on the three-dimensional feature maps (28, 128) obtained in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature maps after batch normalization through a ReLU activation function to obtain three-dimensional feature maps with the dimension of (28, 64);
3.3 Carrying out convolution on the three-dimensional feature map with the dimensionality of (28, 64) obtained in the previous step through 128 convolution kernels with 3 x 3 and the step size of 1, carrying out 1-turn zero filling operation on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map with (28, 128), and carrying out batch normalization operation on the three-dimensional feature map;
3.4 Carrying out convolution on the three-dimensional feature map input in the step 3.1) through 128 convolution kernels of 1 × 1 and with the step size of 2 to obtain a (28, 128) three-dimensional feature map, adding the three-dimensional feature map and the feature map obtained in the step 3.3), and then carrying out nonlinear mapping operation on the three-dimensional feature map obtained after addition through a ReLU activation function to obtain a three-dimensional feature map with the dimension of (28, 128);
3.5 Passing the three-dimensional feature map (28, 128) obtained in the previous step through 128 convolution kernels of 3 × 3 and performing convolution with a step size of 1, and performing zero padding operation for 1 circle on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map (28, 128);
3.6 Carrying out batch normalization operation on the three-dimensional characteristic diagram output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional characteristic diagram after batch normalization through a ReLU activation function;
3.7 Step 3.5) is repeated and batch normalization operation is carried out on the obtained three-dimensional characteristic diagram;
3.8 Adding the three-dimensional feature map output in the previous step with the three-dimensional feature map with the dimension (28, 128) input in the step 3.5), and then carrying out nonlinear mapping operation on the three-dimensional feature map obtained after addition through a ReLU activation function;
3.9 Repeating the step 3.5) to the step 3.8) twice, and outputting a three-dimensional feature map with the dimension of (28, 128);
4) The fourth convolution layer
4.1 Passing the three-dimensional feature map obtained in the previous step through 256 convolution kernels of 3 × 3 and performing convolution with 2 steps, and performing 1-turn zero padding operation on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map of (14, 256);
4.2 Carrying out batch normalization operation on the three-dimensional feature map (14, 256) obtained in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature map after batch normalization through a ReLU activation function to obtain a three-dimensional feature map with the dimension of (14, 256);
4.3 Passing the three-dimensional feature map obtained in the previous step through 256 convolution kernels of 3 x 3 and performing convolution by step size of 1, performing zero padding operation for 1 circle on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map (14, 256), and performing batch normalization operation on the three-dimensional feature map;
4.4 Carrying out convolution on the three-dimensional feature map input in the step 4.1) through 256 convolution kernels of 1 × 1 and in steps of 2, adding the three-dimensional feature map with the dimensionality (14, 256) and the three-dimensional feature map output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature map obtained after addition through a ReLU activation function;
4.5 Passing the three-dimensional feature map output in the previous step through 256 convolution kernels of 3 x 3 and performing convolution by step size of 1, and performing zero filling operation for 1 circle on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map (14, 256);
4.6 Carrying out batch normalization operation on the three-dimensional characteristic diagram output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional characteristic diagram after batch normalization through a ReLU activation function;
4.7 Step 4.5) is repeated and batch normalization operation is carried out on the obtained three-dimensional characteristic diagram;
4.8 Adding the three-dimensional feature map input in the step 4.5) with the three-dimensional feature map output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature map obtained after addition through a ReLU activation function;
4.9 Step 4.5) -step 4.8) are repeated four times, and a three-dimensional feature map with the dimensionality (14, 256) is output;
5) The fifth convolution layer
5.1 Passing the three-dimensional feature map obtained in the previous step through 512 convolution kernels of 3 x 3 and performing convolution by step size of 2, and performing zero padding operation for 1 circle on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map of (7, 512);
5.2 Carrying out batch normalization operation on the three-dimensional characteristic diagram obtained in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional characteristic diagram after batch normalization through a ReLU activation function to obtain a three-dimensional characteristic diagram with the dimensionality of (7, 512);
5.3 The three-dimensional feature map obtained in the previous step is convoluted by 512 convolution kernels of 3 x 3 and the step length of 1, and the zero padding operation is carried out for 1 circle on the periphery of the obtained pixel matrix to obtain the three-dimensional feature map of (7, 512), and the three-dimensional feature map is subjected to batch normalization operation;
5.4 Carrying out convolution on the three-dimensional feature map input in the step 5.1) through 512 convolution kernels of 1 × 1 and 2 steps to obtain a three-dimensional feature map with the dimension of (7, 512), adding the three-dimensional feature map with the dimension of (7, 512) with the three-dimensional feature map output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature map obtained after addition through a ReLU activation function;
5.5 The three-dimensional feature map output in the previous step is convoluted by 512 convolution kernels of 3 x 3 and the step size of 1, and the zero filling operation is carried out for 1 circle at the periphery of the obtained pixel matrix, so as to obtain the three-dimensional feature map of (7, 512);
5.6 Carrying out batch normalization operation on the three-dimensional characteristic diagram output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional characteristic diagram after batch normalization through a ReLU activation function;
5.7 Step 5.5) is repeated and the obtained three-dimensional characteristic diagram is subjected to batch normalization operation;
5.8 Adding the three-dimensional feature map input in the step 5.5) with the three-dimensional feature map output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature map obtained after addition through a ReLU activation function;
5.9 Repeating the step 5.5) to the step 5.8) once, and outputting a three-dimensional feature map with the dimension of (7, 512);
6) Global average pooling layer
Performing global average pooling operation on the three-dimensional feature map with the dimension of (7, 512) output in the previous step, and outputting the three-dimensional feature map with the dimension of (1, 512);
7) Full connection layer fc
Converting the three-dimensional feature map with the dimensionality of (1, 512) output in the previous step into a 4-dimensional feature vector;
fourthly) classifying the output result of the convolutional neural network by using a classification function softmax to obtain probability vectors of each classification;
and fifthly), selecting the maximum value in the probability vector as a prediction type, and outputting the prediction type and the prediction probability as an analysis result of the gear surface damage.
In this embodiment, after the batch normalization operations in step 1.2), step 2.2), step 2.3), step 3.2), step 3.3), step 3.6), step 3.7), step 4.2), step 4.3), step 4.6), step 4.7), step 5.2), step 5.3), step 5.6) and step 5.7), the pruning operation with the pruning rate of 20% is performed on each batch normalization layer.
In this embodiment, the specific operations of training the constructed convolutional neural network and storing the formed weights in the constructed convolutional neural network are as follows:
a) Inputting a training set in the tooth surface image data set into a built convolutional neural network for training, and performing transfer learning by using official weight to finally form a plurality of output numerical values corresponding to the classification quantity;
b) Use of
Figure BDA0003939284910000121
A function, wherein i is a category index, C is the number of categories and is the output of a preceding-stage output unit of a classifier of a full connection layer, e is a natural base number, a plurality of output values are converted into relative probabilities to obtain probability vectors corresponding to the categories, and the category with the highest probability is selected as a predicted value of the output;
c) Inputting the obtained predicted value and the real label into a cross entropy function CrossEntropyLoss to calculate an error value, calculating a gradient value of each parameter through an error back propagation function loss.
d) Inputting the test set into a convolutional neural network to verify the training effect, calculating an error value according to a cross entropy function, outputting the type with the maximum probability as a predicted value, and if the predicted value is the same as the real label, determining that the prediction is correct, wherein the ratio of the number of the correct prediction to the number of the test set is the accuracy of the whole test set;
e) Training for 400 times, wherein the lowest error value on the set to be tested is 0.095, the accuracy rate reaches 97.7%, and the trained weight is stored in the convolutional neural network.
Example 2
An electronic device comprising a processor and a memory, the memory storing computer readable instructions that, when executed by the processor, are capable of performing the optimization method of embodiment 1.
In this embodiment, the electronic apparatus includes a modeling device, a processor, a memory, a storage controller, a peripheral interface, an input-output unit, an audio unit, a display unit, and the like.
Specifically, the memory controller, the processor, the peripheral interface, the input/output unit, the audio unit, and the display unit are electrically connected to each other directly or indirectly to implement data transmission or interaction. The modelling means includes at least one software functional module which may be stored in said memory in the form of software or firmware (firmware) or fixed in an Operating System (OS) of the modelling means. The processor is used to execute executable modules stored in the memory, including software functional modules or computer programs.
The memory may be, but is not limited to, a Random Access Memory (RAM), a Read Only Memory (ROM), a Programmable Read Only Memory (PROM), an erasable read only memory (EPROM), an electrically erasable read only memory (EEPROM), and the like. The memory is used for storing programs, and the processor executes corresponding programs after receiving execution instructions.
The processor may be an integrated circuit chip having signal processing capabilities. The processor may be a general-purpose processor including a Central Processing Unit (CPU), a Network Processor (NP), etc.; but may also be a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), an off-the-shelf programmable gate array (FPGA) or other programmable logic device, discrete gate or transistor logic, discrete hardware components.
The peripheral interface couples various input/output devices to the processor and memory, which may be implemented in a single chip or, alternatively, may be implemented separately from a separate chip.
The input and output unit, the audio unit and the display unit are all the prior art, for example, the input and output unit is used for providing input data for a user to realize the interaction between the user and the server (or the local terminal), and can be a mouse, a keyboard and the like; such as an audio unit, which may include one or more microphones, one or more speakers, and audio circuitry; such as a display unit, provides an interactive interface (e.g., a user interface) between the electronic device and a user or for displaying image data for reference by the user.
Example 3
A readable storage medium having stored thereon a computer program which, when executed by a processor, performs the steps of embodiment 1.

Claims (10)

1. A tooth surface damage identification method based on an improved neural network is characterized in that the improved neural network is utilized to establish a tooth surface damage classification model, and then a gear surface image to be detected is input into the tooth surface damage classification model to obtain an analysis result of the gear surface damage; the method is characterized in that the steps of establishing the tooth surface damage classification model by utilizing the improved neural network are as follows:
firstly), acquiring a tooth surface image of a gear by using a machine vision system, classifying the acquired tooth surface image according to the characteristics of tooth surface damage, and marking a real label of each classification;
secondly), cutting the tooth surface images of each type to the same size, and performing image enhancement to form a classified tooth surface image data set;
thirdly), constructing a convolutional neural network by taking ResNet-34 as a primary network model, training the constructed convolutional neural network by utilizing a tooth surface image data set, and storing formed weights into the constructed convolutional neural network;
fourthly) carrying out classification processing on the output result of the convolutional neural network by using a classification function softmax to obtain probability vectors of each classification;
and fifthly), selecting the maximum value in the probability vector as a prediction type, and outputting the prediction type and the prediction probability as an analysis result of the gear surface damage.
2. The method for identifying the tooth surface damage based on the improved neural network as claimed in claim 1, wherein: the step one) of classifying the acquired tooth surface images means that the acquired tooth surface images are classified according to four types of tooth surface pitting, tooth surface abrasion, gear tooth breakage and normal tooth surface, and the four types are used as correspondingly classified real labels.
3. The method for identifying the tooth surface damage based on the improved neural network as claimed in claim 1, wherein: and the image enhancement method in the step two) comprises the steps of horizontal turning, vertical turning, 180-degree rotation, zooming, shifting, blurring, brightening, darkening and adding salt and pepper noise in sequence.
4. The method for identifying the tooth surface damage based on the improved neural network as claimed in claim 1, wherein: forming a classified tooth surface image data set in the second step, wherein the proportion of a training set to a testing set is 9:1.
5. the method for identifying the tooth surface damage based on the improved neural network as claimed in claim 1, wherein: the convolutional neural network in the third step sequentially comprises an input layer, a first convolutional layer, a second convolutional layer, a third convolutional layer, a fourth convolutional layer, a fifth convolutional layer, a global average pooling layer and a full-link layer.
6. The method for identifying tooth surface damage based on the improved neural network as claimed in claim 5, wherein the specific operation of constructing the convolutional neural network in the step three) is as follows:
1) The first winding layer
1.1 Passing the image with the dimension of (224, 64) through 64 convolution kernels of 7 by 7 and performing convolution operation with the step size of 2, and performing zero filling operation for 3 circles at the periphery of the obtained pixel matrix to obtain a three-dimensional characteristic map of (112, 64);
1.2 Carrying out batch normalization operation on the obtained three-dimensional characteristic diagram, and then carrying out nonlinear mapping operation on the batch normalized three-dimensional characteristic diagram through a ReLU activation function;
1.3 Performing maximal pooling operation on the feature map output in step 1.2) by 3-by-3 convolution kernel with 2 steps, and performing 1-turn zero-filling operation on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map (56, 64);
2) The second convolution layer
2.1 The three-dimensional characteristic diagram (56, 64) obtained in the previous step is convoluted by 64 convolution kernels of 3 x 3 and with the step size of 1, and the zero padding operation is carried out for 1 circle on the periphery of the obtained pixel matrix to obtain the three-dimensional characteristic diagram (56, 64);
2.2 Batch normalization operation is carried out on the three-dimensional feature maps (56, 64) obtained in the previous step, and then nonlinear mapping operation is carried out on the batch normalized three-dimensional feature maps through a ReLU activation function;
2.3 ) repeating the convolution and zero filling operation in the step 2.1) on the three-dimensional characteristic diagram obtained in the previous step, and carrying out batch normalization operation on the obtained three-dimensional characteristic diagram;
2.4 Adding the three-dimensional feature map input in the step 2.1) and the three-dimensional feature map output in the step 2.3), and then carrying out nonlinear mapping operation on the three-dimensional feature map obtained after addition through a ReLU activation function;
2.5 Repeating the step 2.1) to the step 2.4) twice, and outputting a three-dimensional feature map with dimensions (56, 64);
3) The third convolution layer
3.1 Carrying out convolution on the three-dimensional feature map with the dimension (56, 64) obtained in the previous step through 128 convolution kernels with 3 x 3 and the step size of 2, and carrying out 1-turn zero padding operation on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map (28, 128);
3.2 Carrying out batch normalization operation on the three-dimensional feature maps (28, 128) obtained in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature maps after batch normalization through a ReLU activation function to obtain three-dimensional feature maps with the dimension of (28, 64);
3.3 Carrying out convolution on the three-dimensional feature map with the dimension of (28, 64) obtained in the previous step through 128 convolution kernels of 3 x 3 and the step size of 1, carrying out zero filling operation for 1 circle on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map with the dimension of (28, 128), and carrying out batch normalization operation on the three-dimensional feature map;
3.4 Carrying out convolution on the three-dimensional feature map input in the step 3.1) through 128 convolution kernels of 1 × 1 and with the step size of 2 to obtain a three-dimensional feature map (28, 128), adding the three-dimensional feature map to the feature map obtained in the step 3.3), and then carrying out nonlinear mapping operation on the three-dimensional feature map obtained after addition through a ReLU activation function to obtain a three-dimensional feature map with the dimension of (28, 128);
3.5 Passing the three-dimensional feature map (28, 128) obtained in the previous step through 128 convolution kernels of 3 × 3 and performing convolution by step size of 1, and performing zero padding operation for 1 circle on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map (28, 128);
3.6 Carrying out batch normalization operation on the three-dimensional feature maps output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature maps subjected to batch normalization through a ReLU activation function;
3.7 Step 3.5) is repeated and batch normalization operation is carried out on the obtained three-dimensional characteristic diagram;
3.8 Adding the three-dimensional feature map output in the previous step with the three-dimensional feature map with the dimension (28, 128) input in the step 3.5), and then carrying out nonlinear mapping operation on the three-dimensional feature map obtained after addition through a ReLU activation function;
3.9 Repeating the step 3.5) to the step 3.8) twice, and outputting a three-dimensional feature map with the dimension of (28, 128);
4) The fourth convolution layer
4.1 Passing the three-dimensional feature map obtained in the previous step through 256 convolution kernels of 3 × 3 and performing convolution with 2 steps, and performing 1-turn zero padding operation on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map of (14, 256);
4.2 Carrying out batch normalization operation on the three-dimensional feature map (14, 256) obtained in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature map after batch normalization through a ReLU activation function to obtain a three-dimensional feature map with the dimension of (14, 256);
4.3 Passing the three-dimensional feature map obtained in the previous step through 256 convolution kernels of 3 x 3 and performing convolution by step size of 1, performing zero padding operation for 1 circle on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map (14, 256), and performing batch normalization operation on the three-dimensional feature map;
4.4 Carrying out convolution on the three-dimensional feature map input in the step 4.1) through 256 convolution kernels of 1 × 1 and with the step size of 2 to obtain a three-dimensional feature map with the dimension (14, 256) which is added to the three-dimensional feature map output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature map obtained after addition through a ReLU activation function;
4.5 Passing the three-dimensional feature map output in the previous step through 256 convolution kernels of 3 x 3 and performing convolution by step size of 1, and performing zero filling operation for 1 circle on the periphery of the obtained pixel matrix to obtain a three-dimensional feature map (14, 256);
4.6 Carrying out batch normalization operation on the three-dimensional characteristic diagram output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional characteristic diagram after batch normalization through a ReLU activation function;
4.7 Step 4.5) is repeated and the obtained three-dimensional characteristic diagram is subjected to batch normalization operation;
4.8 Adding the three-dimensional feature map input in the step 4.5) with the three-dimensional feature map output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature map obtained after addition through a ReLU activation function;
4.9 Repeating the step 4.5) to the step 4.8) four times, and outputting a three-dimensional feature map with the dimension of (14, 256);
5) The fifth convolution layer
5.1 Carrying out convolution on the three-dimensional feature map obtained in the previous step by 512 convolution kernels of 3 x 3 and 2 steps, and carrying out 1-turn zero filling operation on the periphery of the obtained pixel matrix to obtain a (7, 512) three-dimensional feature map;
5.2 Carrying out batch normalization operation on the three-dimensional characteristic diagram obtained in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional characteristic diagram after batch normalization through a ReLU activation function to obtain a three-dimensional characteristic diagram with the dimensionality of (7, 512);
5.3 The three-dimensional feature map obtained in the previous step is convoluted by 512 convolution kernels of 3 x 3 and the step length of 1, and the zero padding operation is carried out for 1 circle on the periphery of the obtained pixel matrix to obtain the three-dimensional feature map of (7, 512), and the three-dimensional feature map is subjected to batch normalization operation;
5.4 Carrying out convolution on the three-dimensional feature map input in the step 5.1) through 512 convolution kernels of 1 × 1 and 2 steps to obtain a three-dimensional feature map with the dimension of (7, 512), adding the three-dimensional feature map with the dimension of (7, 512) with the three-dimensional feature map output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature map obtained after addition through a ReLU activation function;
5.5 Carrying out convolution on the three-dimensional feature map output in the previous step by 512 convolution kernels of 3 x 3 and step length of 1, and carrying out 1-turn zero filling operation on the periphery of the obtained pixel matrix to obtain a (7, 512) three-dimensional feature map;
5.6 Carrying out batch normalization operation on the three-dimensional feature maps output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional feature maps subjected to batch normalization through a ReLU activation function;
5.7 Step 5.5) is repeated and batch normalization operation is carried out on the obtained three-dimensional characteristic diagram;
5.8 Adding the three-dimensional characteristic diagram input in the step 5.5) with the three-dimensional characteristic diagram output in the previous step, and then carrying out nonlinear mapping operation on the three-dimensional characteristic diagram obtained after addition through a ReLU activation function;
5.9 Repeating the step 5.5) to the step 5.8) once, and outputting a three-dimensional feature map with the dimension of (7, 512);
6) Global average pooling layer
Performing global average pooling operation on the three-dimensional feature map with the dimension of (7, 512) output in the previous step, and outputting the three-dimensional feature map with the dimension of (1, 512);
7) Full connection layer
And converting the three-dimensional feature map with the dimension of (1, 512) output in the previous step into a 4-dimensional feature vector.
7. The method for tooth surface damage identification based on the improved neural network as claimed in claim 6, wherein: after the batch normalization operation in the step 1.2), the step 2.2), the step 2.3), the step 3.2), the step 3.3), the step 3.6), the step 3.7), the step 4.2), the step 4.3), the step 4.6), the step 4.7), the step 5.2), the step 5.3), the step 5.6) and the step 5.7), pruning with a pruning rate of 20% is performed on each batch normalization layer.
8. The method for identifying tooth surface damage based on the improved neural network as claimed in claim 1, wherein the specific operations of training the constructed convolutional neural network and saving the formed weights into the constructed convolutional neural network are as follows:
a) Inputting a training set in the tooth surface image data set into a built convolutional neural network for training, and performing transfer learning by using official weight to finally form a plurality of output numerical values corresponding to the classification quantity;
b) Make itBy using
Figure FDA0003939284900000071
A function, converting a plurality of output values into relative probabilities to obtain probability vectors corresponding to each classification, and selecting the classification with the highest probability as a predicted value of the output;
where i is the index of the class, C is the number of classes, V i The output of a preceding stage output unit of the classifier of the full connection layer is shown, and e is a natural base number;
c) Inputting the obtained predicted value and the real label into a cross entropy function CrossEntropyLoss to calculate an error value, calculating a gradient value of each parameter through an error back propagation function loss.
d) Inputting the test set into a convolutional neural network to verify the training effect, calculating an error value according to a cross entropy function, outputting the type with the maximum probability as a predicted value, judging that the prediction is correct if the predicted value is the same as the real label, and determining the ratio of the number of correct predictions to the number of test sets as the accuracy of the whole test set;
e) Training for 400 times, wherein the lowest error value on the set to be tested is 0.095, the accuracy rate reaches 97.7%, and the trained weight is stored in the convolutional neural network.
9. An electronic device comprising a processor and a memory, the memory storing computer readable instructions, the computer readable instructions, when executed by the processor, being capable of performing the tooth surface damage identification method of any one of claims 1-8.
10. A readable storage medium, on which a computer program is stored, which, when being executed by a processor, performs the steps of the method for tooth flank damage identification according to any one of claims 1-8.
CN202211414681.7A 2022-11-11 2022-11-11 Tooth surface damage identification method and device based on improved neural network Pending CN115565020A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211414681.7A CN115565020A (en) 2022-11-11 2022-11-11 Tooth surface damage identification method and device based on improved neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211414681.7A CN115565020A (en) 2022-11-11 2022-11-11 Tooth surface damage identification method and device based on improved neural network

Publications (1)

Publication Number Publication Date
CN115565020A true CN115565020A (en) 2023-01-03

Family

ID=84770434

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211414681.7A Pending CN115565020A (en) 2022-11-11 2022-11-11 Tooth surface damage identification method and device based on improved neural network

Country Status (1)

Country Link
CN (1) CN115565020A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116091445A (en) * 2023-01-05 2023-05-09 中国长江电力股份有限公司 Ship lift gear and gear surface damage identification method based on deep learning
CN117636057A (en) * 2023-12-13 2024-03-01 石家庄铁道大学 Train bearing damage classification and identification method based on multi-branch cross-space attention model

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116091445A (en) * 2023-01-05 2023-05-09 中国长江电力股份有限公司 Ship lift gear and gear surface damage identification method based on deep learning
CN116091445B (en) * 2023-01-05 2024-01-02 中国长江电力股份有限公司 Ship lift gear and gear surface damage identification method based on deep learning
CN117636057A (en) * 2023-12-13 2024-03-01 石家庄铁道大学 Train bearing damage classification and identification method based on multi-branch cross-space attention model
CN117636057B (en) * 2023-12-13 2024-06-11 石家庄铁道大学 Train bearing damage classification and identification method based on multi-branch cross-space attention model

Similar Documents

Publication Publication Date Title
CN115565020A (en) Tooth surface damage identification method and device based on improved neural network
CN110969175B (en) Wafer processing method and device, storage medium and electronic equipment
CN112561910A (en) Industrial surface defect detection method based on multi-scale feature fusion
CN116258707A (en) PCB surface defect detection method based on improved YOLOv5 algorithm
CN110751195B (en) Fine-grained image classification method based on improved YOLOv3
CN114140683A (en) Aerial image target detection method, equipment and medium
CN113095370A (en) Image recognition method and device, electronic equipment and storage medium
CN114743189A (en) Pointer instrument reading identification method and device, electronic equipment and storage medium
TWI812888B (en) Image recognition method and image recognition system
CN112364974A (en) Improved YOLOv3 algorithm based on activation function
CN114971375A (en) Examination data processing method, device, equipment and medium based on artificial intelligence
CN114048817A (en) Deep learning input set priority test method based on variation strategy
CN116630700A (en) Remote sensing image classification method based on introduction channel-space attention mechanism
CN112381177A (en) Dial digital character recognition method and system based on deep learning
CN110321867B (en) Shielded target detection method based on component constraint network
CN113763364B (en) Image defect detection method based on convolutional neural network
US11195265B2 (en) Server and method for recognizing image using deep learning
CN113541985A (en) Internet of things fault diagnosis method, training method of model and related device
CN105654138A (en) Orthogonal projection and dimensionality reduction classification method and system for multidimensional data
CN116363136B (en) On-line screening method and system for automatic production of motor vehicle parts
CN113421223A (en) Industrial product surface defect detection method based on deep learning and Gaussian mixture
CN111666872A (en) Efficient behavior identification method under data imbalance
CN114743023A (en) Wheat spider image detection method based on RetinaNet model
CN114861771A (en) Industrial CT image defect classification method based on feature extraction and deep learning
CN113971737A (en) Object recognition method for robot, electronic device, medium, and program product

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right

Effective date of registration: 20240703

Address after: 450001 No.149, science Avenue, high tech Zone, Zhengzhou City, Henan Province

Applicant after: Zheng Ji Suo (Zhengzhou) Transmission Technology Co.,Ltd.

Country or region after: China

Address before: 450001 149 science Avenue, Zhengzhou high tech Industrial Development Zone, Henan

Applicant before: ZHENGZHOU RESEARCH INSTITUTE OF MECHANICAL ENGINEERING Co.,Ltd.

Country or region before: China

TA01 Transfer of patent application right