CN116773534B - Detection method and device, electronic equipment and computer readable medium - Google Patents
Detection method and device, electronic equipment and computer readable medium Download PDFInfo
- Publication number
- CN116773534B CN116773534B CN202311021779.0A CN202311021779A CN116773534B CN 116773534 B CN116773534 B CN 116773534B CN 202311021779 A CN202311021779 A CN 202311021779A CN 116773534 B CN116773534 B CN 116773534B
- Authority
- CN
- China
- Prior art keywords
- module
- dimensional
- layer
- feature map
- decoding
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 38
- 238000000605 extraction Methods 0.000 claims abstract description 83
- 238000000034 method Methods 0.000 claims abstract description 30
- 230000007547 defect Effects 0.000 claims abstract description 21
- 230000003595 spectral effect Effects 0.000 claims abstract description 9
- 239000000284 extract Substances 0.000 claims abstract 3
- 230000006870 function Effects 0.000 claims description 32
- 238000010586 diagram Methods 0.000 claims description 24
- 238000003860 storage Methods 0.000 claims description 15
- 239000011159 matrix material Substances 0.000 claims description 11
- 239000013598 vector Substances 0.000 claims description 11
- 238000007781 pre-processing Methods 0.000 claims description 8
- 238000012545 processing Methods 0.000 claims description 8
- 238000011176 pooling Methods 0.000 claims description 6
- 230000004913 activation Effects 0.000 claims description 3
- 230000008569 process Effects 0.000 abstract description 7
- 238000004590 computer program Methods 0.000 description 11
- WHXSMMKQMYFTQS-UHFFFAOYSA-N Lithium Chemical compound [Li] WHXSMMKQMYFTQS-UHFFFAOYSA-N 0.000 description 6
- 229910052744 lithium Inorganic materials 0.000 description 6
- 238000010606 normalization Methods 0.000 description 6
- 238000013528 artificial neural network Methods 0.000 description 5
- 238000004891 communication Methods 0.000 description 4
- 230000003287 optical effect Effects 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 3
- 238000003466 welding Methods 0.000 description 3
- 238000011161 development Methods 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 238000004519 manufacturing process Methods 0.000 description 2
- 230000000644 propagated effect Effects 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 238000001228 spectrum Methods 0.000 description 2
- 241001270131 Agaricus moelleri Species 0.000 description 1
- 241000282414 Homo sapiens Species 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000002950 deficient Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 238000003908 quality control method Methods 0.000 description 1
- 230000017105 transposition Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N21/00—Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
- G01N21/84—Systems specially adapted for particular applications
- G01N21/88—Investigating the presence of flaws or contamination
- G01N21/8851—Scan or image signal processing specially adapted therefor, e.g. for scan signal adjustment, for detecting different kinds of defects, for compensating for structures, markings, edges
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
- G06N3/0455—Auto-encoder networks; Encoder-decoder networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4038—Image mosaicing, e.g. composing plane images from plane sub-images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0004—Industrial image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/7715—Feature extraction, e.g. by transforming the feature space, e.g. multi-dimensional scaling [MDS]; Mappings, e.g. subspace methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
- G06V10/806—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N21/00—Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
- G01N21/84—Systems specially adapted for particular applications
- G01N21/88—Investigating the presence of flaws or contamination
- G01N21/8851—Scan or image signal processing specially adapted therefor, e.g. for scan signal adjustment, for detecting different kinds of defects, for compensating for structures, markings, edges
- G01N2021/8887—Scan or image signal processing specially adapted therefor, e.g. for scan signal adjustment, for detecting different kinds of defects, for compensating for structures, markings, edges based on image processing techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Evolutionary Computation (AREA)
- Health & Medical Sciences (AREA)
- Software Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- General Engineering & Computer Science (AREA)
- Molecular Biology (AREA)
- Data Mining & Analysis (AREA)
- Computational Linguistics (AREA)
- Multimedia (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Signal Processing (AREA)
- Biochemistry (AREA)
- Quality & Reliability (AREA)
- Pathology (AREA)
- Immunology (AREA)
- Chemical & Material Sciences (AREA)
- Analytical Chemistry (AREA)
- Image Analysis (AREA)
Abstract
The application discloses a detection method and device, electronic equipment and a computer readable medium, wherein the detection method comprises the steps of obtaining an appearance picture of a battery module; inputting the appearance picture into a coding layer, and extracting the characteristics of the appearance picture through the coding layer; analyzing the characteristics extracted by the coding layer through the decoding layer, and outputting the result. The three-dimensional feature extraction method is provided to more comprehensively and effectively extract spatial and spectral features, the features of spatial and spectral information in the images are obtained in the 3D convolution process by remolding the images and creating structures similar to three-dimensional medical images, and meanwhile, the multi-dimensional spatial feature extraction module and the depth codec network based on the depth separable convolution can simply and effectively detect defects of the appearance of the product.
Description
Technical Field
The present disclosure relates to the field of image recognition technologies, and in particular, to a detection method and apparatus, an electronic device, and a computer readable medium.
Background
With the development of new energy industry, the welding process of the lithium battery module is gradually automated, but errors are easily generated in the welding process, the errors can lead to unqualified product quality, and how to find the errors in an intelligent factory becomes a problem to be solved urgently. Currently, researchers use appearance defect detection technology as a method for finding out the errors, and the appearance defect detection technology is an important part of quality control of products such as lithium battery modules in new energy factories. In the traditional manufacturing industry, the appearance defect detection is usually carried out by using a manual method, however, due to the conditions of fatigue, distraction and the like of human beings, the cost of the manual detection is high and the accuracy is low. With the development of artificial intelligence, a deep learning method is introduced into an intelligent factory at present to detect appearance defects of a lithium battery module in a high precision in the welding process. Three fundamental problems are encountered in appearance defect detection. Firstly, the types of appearance defects in the manufacturing process of the lithium battery are very diversified; secondly, the appearance defects have no specific measurement standard, and the defect sizes are greatly different; third, defects in appearance may be similar to the background of the factory, resulting in poor detection.
Disclosure of Invention
Purpose of (one) application
In view of the foregoing, an object of the present application is to provide a detection method and apparatus, an electronic device, and a computer readable medium, which are used for solving the above technical problems in the prior art by designing a depth separable convolution layer and a three-dimensional feature extraction module in an encoder portion and designing a multi-input attention module in a decoder portion, and combining the depth separable convolution layer to realize effective feature extraction of an input image and defect detection of product appearance such as a lithium battery module.
(II) technical scheme
According to the method, the device and the system, the multi-dimensional feature extraction module, the three-dimensional feature extraction module, the multi-input attention module and the transposed convolution decoding module are designed, so that the product appearance defect detection work in the external image is effectively achieved. The multi-dimensional feature extraction module is provided with a depth separable convolution layer to reduce convolution parameters, so that features in a two-dimensional image are extracted more simply and effectively; the three-dimensional feature extraction module converts the two-dimensional image into a three-dimensional image, creates a structure similar to the three-dimensional medical image, obtains the spectrum and the space feature of the image through three-dimensional convolution, and learns the weight information in the feature map by fusing the feature map obtained by the multi-dimensional feature extraction module and the three-dimensional feature extraction module; the transposed convolution decoding module further learns information in the appearance image by utilizing transposed convolution and a designed depth separable convolution layer, and finally the proposed method can realize appearance defect detection of products such as lithium battery modules in intelligent factories.
The detection method disclosed by the application comprises the following steps: obtaining an appearance picture of the battery module; inputting the appearance picture into a coding layer, and extracting the characteristics of the appearance picture through the coding layer; the coding layer comprises at least one three-dimensional feature extraction module which is arranged in sequence; analyzing the characteristics extracted by the coding layer through the decoding layer, and outputting a result; the decoding layer comprises a transposed convolution decoding module and an output module, wherein the transposed convolution decoding module comprises a first decoding module and at least one second decoding module which are arranged in sequence; the first decoding module comprises a separable convolution layer and a transposed convolution layer, and the second decoding module comprises a multi-input attention module, the separable convolution layer and the transposed convolution layer; the output module analyzes the learning of the decoding layer through the convolution layer and the softmax function, outputs a result and judges whether defects exist or not.
In one embodiment of the application, the coding layer comprises a data preprocessing module and a multidimensional feature extraction module; the data preprocessing module is used for normalizing the appearance picture into an input vector; the multidimensional feature extraction module is used for extracting a feature map from the input vector, wherein the feature map comprises a low-level feature map and/or a high-level feature map; the three-dimensional feature extraction module is used for converting the appearance picture into a three-dimensional image and extracting a three-dimensional feature image from the three-dimensional image, wherein the three-dimensional feature image comprises three-dimensional spectral features and/or three-dimensional spatial features;
in one embodiment of the application, the multi-dimensional feature extraction module is configured as a first extraction module and at least one second extraction module arranged in sequence, and the first extraction module inputs the extracted features to the first-ordered second extraction module; the second extraction module inputs the extracted features to the second extraction module of the next order.
In one embodiment of the present application, the first decoding module inputs the extracted features to a first-ordered second decoding module; the second decoding module inputs the extracted features to the second decoding module of the next order.
In one embodiment of the present application, the multidimensional feature extraction module, the three-dimensional feature extraction module, and the built-in module of the transposed convolutional decoding module are all ordered according to a processed size standard; and the characteristics extracted by the multidimensional characteristic extraction module and the built-in module of the three-dimensional characteristic extraction module are input to the built-in module of the transposed convolution decoding module according to the size standard.
In one embodiment of the present application, the multi-input attention module fuses the low-level feature map, the three-dimensional feature map, and the high-level feature map as follows:
splicing the low-level characteristic diagram, the three-dimensional characteristic diagram and the high-level characteristic diagram,
wherein,representing a splicing operation->Is output after splicing, is->Three dimensions of width, height and depth of the feature map are respectively corresponding to +.>Representing a low-level feature map obtained by the multi-dimensional feature extraction module, S 1 Depth of representation low-level feature map, +.>Representing the three-dimensional feature map obtained by the three-dimensional feature extraction module, S 2 Representing the depth of the three-dimensional feature map,representing the decoded advanced feature map, S 3 Representing the depth of the advanced feature map;
obtaining the average value of all feature graphs based on global pooling layer,/>Wherein->Representing a global pooling layer of the pool,
obtainingThen, obtaining a weight vector coefficient through two full-connection layers and a sigmoid function,
Wherein the method comprises the steps ofRepresenting a full connection layer, < >>Representing the activation function of the fully connected layer,
obtaining a weight matrix of the feature map through copying operation,
Wherein the method comprises the steps ofA function representing a copy operation;
finally, the weight matrix of the feature mapMultiplying the spliced output->Get the output of the multiple input attention module +.>,
Is a weight matrix of the feature map.
In one embodiment of the present application, there is provided a detection apparatus including: the input module is configured to acquire an appearance picture of the battery module; the coding layer module is configured to extract the characteristics of the appearance pictures to obtain image characteristics; the coding layer comprises at least one three-dimensional feature extraction module which is arranged in sequence; the decoding layer module is configured to analyze the image characteristics and output a result; the decoding layer module comprises a transposed convolution decoding module and an output module, wherein the transposed convolution decoding module comprises a first decoding module and at least one second decoding module which are arranged in sequence; the first decoding module comprises a separable convolution layer and a transposed convolution layer, and the second decoding module comprises a multi-input attention module, the separable convolution layer and the transposed convolution layer; the output module analyzes the learning of the decoding layer through the convolution layer and the softmax function, outputs a result and judges whether defects exist or not.
In one embodiment of the present application, there is provided an electronic device including: one or more processors; and a storage means for storing one or more programs which, when executed by the one or more processors, cause the electronic device to implement the detection method as described above.
In one embodiment of the present application, a computer readable medium having computer readable instructions stored thereon, which when executed by a processor of a computer, cause the computer to perform the detection method as described above is provided.
(III) beneficial effects
The method comprises the steps of reshaping an image and creating a structure similar to a three-dimensional medical image, then obtaining features of spatial and spectral information in the image in a 3D convolution process, and simply and effectively detecting product appearance defects through a multi-dimensional spatial feature extraction module of depth separable convolution and a depth codec network.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the application.
Drawings
The embodiments described below with reference to the drawings are exemplary and intended for the purpose of illustrating and explaining the present application and are not to be construed as limiting the scope of protection of the present application.
FIG. 1 is a method step diagram of the present application;
FIG. 2 is a system block diagram corresponding to the method of the present application;
FIG. 3 is a block diagram of a three-dimensional feature extraction module of the present application;
FIG. 4 is a method step diagram of the three-dimensional feature extraction module of the present application;
FIG. 5 is a block diagram of a multiple input attention module of the present application;
FIG. 6 is a method step diagram of the multiple input attention module of the present application;
FIG. 7 is a block diagram of the detection device of the present application;
fig. 8 is a schematic diagram of a computer system suitable for use in implementing embodiments of the present application.
Detailed Description
Reference will now be made in detail to exemplary embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, the same numbers in different drawings refer to the same or similar elements, unless otherwise indicated. The implementations described in the following exemplary examples are not representative of all implementations consistent with the present application. Rather, they are merely examples of apparatus and methods consistent with some aspects of the present application as detailed in the accompanying claims.
The block diagrams depicted in the figures are merely functional entities and do not necessarily correspond to physically separate entities. That is, the functional entities may be implemented in software, or in one or more hardware modules or integrated circuits, or in different networks and/or processor devices and/or microcontroller devices.
The flow diagrams depicted in the figures are exemplary only, and do not necessarily include all of the elements and operations/steps, nor must they be performed in the order described. For example, some operations/steps may be decomposed, and some operations/steps may be combined or partially combined, so that the order of actual execution may be changed according to actual situations.
Reference to "a plurality" in this application means two or more than two. "and/or" describes an association relationship of an association object, meaning that there may be three relationships, e.g., a and/or B may represent: a exists alone, A and B exist together, and B exists alone. The character "/" generally indicates that the context-dependent object is an "or" relationship.
The scheme provided by the embodiment of the application relates to the technical field of image recognition. As shown in fig. 1, the present application discloses a detection method, including: s100, obtaining an appearance picture of the battery module; s200, inputting the appearance picture into a coding layer, and extracting the characteristics of the appearance picture through the coding layer; s300, analyzing the characteristics extracted by the coding layer through the decoding layer, and outputting a result.
In one embodiment of the present application, as shown in fig. 2, the encoding layer includes a data preprocessing module, a multi-dimensional feature extraction module, and at least one three-dimensional feature extraction module arranged in sequence; the main function of the data preprocessing module is to read the image input and normalize the image input into vector input; the multidimensional feature extraction module is used for extracting feature graphs from the input vectors, the multidimensional space feature extraction module is designed into a depth separable convolution layer, and two types of feature graphs are obtained from the input images by using a standardized function and a ReLU function.
The feature images are low-level feature images respectively, and comprise low-level spatial features such as textures, gray scales and the like; a high-level feature map comprising deep semantic information; the three-dimensional feature extraction module is used for converting the appearance picture into a three-dimensional image and extracting a three-dimensional feature image from the three-dimensional image by utilizing three-dimensional convolution, wherein the three-dimensional feature image comprises three-dimensional spectral features and/or three-dimensional spatial features;
the purpose of the three-dimensional feature extraction module is to obtain efficient spectral and spatial features by reshaping the input image and correlating the far and near pixels, as shown in fig. 3.
The three-dimensional feature extraction module is reconstructedIs converted into a three-dimensional image, specifically as shown in figure 4,
s210, dividing the input image intoThe third dimension is added to obtain +.>Is a three-dimensional image of (a). In this way, unlike standard two-dimensional convolution, spectral and spatial features can also be obtained from the image without destroying the original image structure.
S220, after reconstructionThe three-dimensional image is subjected to three-dimensional convolution, wherein the convolution kernel size is 3x3x3, and the step size is 1x1, so that a +.>Wherein K is the number of filters.
S230, combining the third dimension and the fourth dimension of the four-dimensional feature matrix, and reconstructing the four-dimensional feature matrix intoWherein K is the number of filters.
The three-dimensional feature extraction module is used for multiple times in the whole method, and is parallel to the multi-dimensional feature extraction module, 4 layers are added, so that the N value in the module is set to be the same as the width and the height of the feature map output by the multi-dimensional feature extraction module, see the table 1, and meanwhile, the number of filters of the module is inversely proportional to the N value to avoid high calculation cost caused by three-dimensional convolution.
Layer number | Input device | Layer name | N value | Filter | Output dimension |
1 | 96 x 96 x 4 | Three-dimensional convolution | 2 | 8 | 96 x 96 x 32 |
2 | 48 x 48 x 16 | Three-dimensional convolution | 4 | 4 | 48 x 48 x 64 |
3 | 24 x 24 x 64 | Three-dimensional convolution | 8 | 2 | 24 x 24 x 128 |
4 | 12 x 12 x 256 | Three-dimensional convolution | 16 | 1 | 12 x 12 x 256 |
TABLE 1 three-dimensional feature extraction Module parameters
In one embodiment of the application, the multi-dimensional feature extraction module is configured as a first extraction module and at least one second extraction module arranged in sequence, and the first extraction module inputs the extracted features to the first-ordered second extraction module; the second extraction module inputs the extracted features to the second extraction module of the next order.
The multidimensional feature extraction module comprises 5 layers of neural networks, wherein the first layer is a 3x3 convolution layer and a separable convolution layer, and the second to five layers each comprise two separable convolution layers, as shown in the multidimensional feature extraction module shown in fig. 2, wherein the sizes of parameters (convolution step length, filter number and output dimension) of each layer are shown in table 2.
Coding layer number | Layer name | Step size | Filter | Output dimension |
1 | Convolution + normalization + ReLU | 2 x 2 | 32 | 96 x 96 x 32 |
1 | Depth separable convolution | 1 x 1 | 32 | 96 x 96 x 32 |
2 | Depth separable convolution | 2 x 2 | 64 | 48 x 48 x 64 |
2 | Depth separable convolution | 1 x 1 | 64 | 48 x 48 x 64 |
3 | Depth separable convolution | 2 x 2 | 128 | 24 x 24 x 128 |
3 | Depth separable convolution | 1 x 1 | 128 | 24 x 24 x 128 |
4 | Depth separable convolution | 2 x 2 | 256 | 12 x 12 x 256 |
4 | Depth separable convolution | 1 x 1 | 256 | 12 x 12 x 256 |
5 | Depth separable convolution | 2 x 2 | 512 | 6 x 6 x 512 |
5 | Depth separable convolution | 1 x 1 | 512 | 6 x 6 x 512 |
Table 2 multidimensional feature extraction module parameters
In particular, the depth separable convolution layer includes two steps, depth convolution and point convolution, as shown in fig. 3.
The depth convolution function is shown in equation 1.
Wherein (1)>Is the result of the deep convolution,/->Representing input image +.>Is->The filter of the channel is used for the filtering of the channel,representing input image +.>Is->A filter of the channel.
In the second step of the depth separable convolution layer, the point convolution will use the depth convolution resultsAs shown in formula 2
Wherein, the point convolution result represents the convolution kernel size 1x1.
The depth separable convolution layer is used multiple times throughout the method, and in order to reduce the size of the feature map, the step size will be set to 2x2 in the convolution operation, except for the first layer.
In one embodiment of the present application, as shown in fig. 2, the decoding layer includes a transpose convolution decoding module and an output module; the transposed convolutional decoding module comprises a first decoding module and at least one second decoding module arranged in sequence, the first decoding module comprises a separable convolutional layer and a transposed convolutional layer, and the second decoding module comprises a multi-input attention module, a separable convolutional layer and a transposed convolutional layer as shown in fig. 5; the output module analyzes the learning of the decoding layer through the convolution layer and the softmax function, outputs a result and judges whether defects exist or not.
The transposed convolutional decoding module includes a total of 5 layers of neural networks, each layer of neural network is composed of a two-dimensional transposed convolutional layer, a normalized layer, a ReLu function, a depth separable convolutional layer, and a multi-input attention module, as shown in fig. 2.
Specifically, each layer of neural network firstly increases the width and the height of the feature map through two-dimensional transposition convolution operation, then obtains output through a depth separable convolution layer, and fuses different feature maps (low-level features, high-level features, three-dimensional spectrum and spatial features) by utilizing a multi-input attention module, wherein parameters of each layer of neural network are shown in table 3.
Decoding layer number | Layer name | Step size | Filter | Output dimension |
1 | Two-dimensional transpose convolution + normalization + ReLU | 2 x 2 | 256 | 12 x 12 x 256 |
1 | Multiple input attention + depth separable convolution | 1 x 1 | 256 | 12 x 12 x 256 |
2 | Two-dimensional transpose convolution + normalization + ReLU | 2 x 2 | 128 | 24 x 24 x 128 |
2 | Multiple input attention + depth separable convolution | 1 x 1 | 128 | 24 x 24 x 128 |
3 | Two-dimensional transpose convolution + normalization + ReLU | 2 x 2 | 64 | 48 x 48 x 64 |
3 | Multiple input attention + depth separable convolution | 1 x 1 | 64 | 48 x 48 x 64 |
4 | Two-dimensional transpose convolution + normalization + ReLU | 2 x 2 | 32 | 96 x 96 x 32 |
4 | Multiple input attention + depth separable convolution | 1 x 1 | 32 | 96 x 96 x 32 |
5 | Two-dimensional transpose convolution + normalization + ReLU | 2 x 2 | 16 | 192 x 192 x 16 |
5 | Multiple input attention + depth separable convolution | 1 x 1 | 16 | 192 x 192 x 16 |
Table 3 transposed convolutional decoding module parameters
The present module will use the output of the transposed convolutional decoding module to derive if the image is defective. Specifically, the output is first passed through a convolution layer with a convolution kernel of 1x1, and then the result is obtained using a softmax function, as shown in equation 8.
Where P is the result of the final defect detection,representing convolution,/->Is the output of the transposed convolutional decoding block.
In one embodiment of the present application, the first decoding module inputs the extracted features to a first-ordered second decoding module; the second decoding module inputs the extracted features to the second decoding module of the next order.
In one embodiment of the present application, the multidimensional feature extraction module, the three-dimensional feature extraction module, and the built-in module of the transposed convolutional decoding module are all ordered according to a processed size standard; the size standard is width x height x channel; and the characteristics extracted by the multidimensional characteristic extraction module and the built-in module of the three-dimensional characteristic extraction module are input to the built-in module of the transposed convolution decoding module according to the size standard.
In one embodiment of the present application, as shown in fig. 6, the steps of the multi-input attention module fusing different feature maps are as follows:
s310, splicing the low-level characteristic diagram, the three-dimensional characteristic diagram and the high-level characteristic diagram,
wherein,representing a splicing operation->Is output after splicing, is->Three dimensions of width, height and depth of the feature map are respectively corresponding to +.>Representing a low-level feature map obtained by the multi-dimensional feature extraction module, S 1 Depth of representation low-level feature map, +.>Representing the three-dimensional feature map obtained by the three-dimensional feature extraction module, S 2 Depth of representation three-dimensional feature map, ++>Representing the decoded advanced feature map, S 3 Representing the depth of the advanced feature map;
s320, obtaining the average value of all feature graphs based on the global pooling layerWherein G represents a global pooling layer,
s330, obtainThen, obtaining a weight vector coefficient through two full-connection layers and a sigmoid function,
Wherein the method comprises the steps ofRepresenting a full connection layer, < >>Representing an activation function of the fully connected layer;
s340, obtaining a weight matrix of the feature map through copy operation,
Wherein Repeat represents a function of the copy operation;
s350, finally, the weight matrix of the feature mapMultiplying the spliced output->Get the output of the multiple input attention module +.>,
Is output after splicing, is->Is a weight matrix of the feature map.
In one embodiment of the present application, there is provided a detection apparatus including: the input module is configured to acquire an appearance picture of the battery module; the coding layer module is configured to extract the characteristics of the appearance pictures to obtain image characteristics; and the decoding layer module is configured to analyze the image characteristics and output a result.
In one embodiment of the present application, there is provided an electronic device including: one or more processors; and a storage means for storing one or more programs which, when executed by the one or more processors, cause the electronic device to implement the detection method as described above.
In one embodiment of the present application, a computer readable medium having computer readable instructions stored thereon, which when executed by a processor of a computer, cause the computer to perform the detection method as described above is provided.
In an exemplary embodiment, as shown in fig. 7, the detection apparatus further includes:
an input module 710 configured to obtain an appearance picture of the battery module;
the coding layer module 720 is configured to perform feature extraction on the appearance picture to obtain image features;
the coding layer comprises at least one three-dimensional feature extraction module which is arranged in sequence;
the decoding layer module 730 is configured to parse the image feature and output a result; the decoding layer module comprises a transposed convolution decoding module, wherein the transposed convolution decoding module comprises a first decoding module and at least one second decoding module which are arranged in sequence; the first decoding module comprises a separable convolution layer and a transposed convolution layer, and the second decoding module comprises a multi-input attention module, the separable convolution layer and the transposed convolution layer; the output module analyzes the learning of the decoding layer through the convolution layer and the softmax function, outputs a result and judges whether defects exist or not.
It should be noted that, the detection apparatus provided in the foregoing embodiment and the detection method provided in the foregoing embodiment belong to the same concept, and a specific manner in which each module and unit perform an operation has been described in detail in the method embodiment, which is not described herein again. In practical application, the detection device provided in the above embodiment may distribute the functions to different functional modules according to needs, that is, the internal structure of the device is divided into different functional modules to complete all or part of the functions described above, which is not limited herein.
The embodiment of the application also provides electronic equipment, which comprises: one or more processors; and a storage means for storing one or more programs which, when executed by the one or more processors, cause the electronic device to implement the detection methods provided in the respective embodiments described above.
Fig. 8 shows a schematic diagram of a computer system suitable for use in implementing the electronic device of the embodiments of the present application. It should be noted that, the computer system 900 of the electronic device shown in fig. 8 is only an example, and should not impose any limitation on the functions and the application scope of the embodiments of the present application.
As shown in fig. 8, the computer system 900 includes a central processing unit (Central Processing Unit, CPU) 901 which can perform various appropriate actions and processes according to a program stored in a Read-Only Memory (ROM) 902 or a program loaded from a storage portion 908 into a random access Memory (Random Access Memory, RAM) 903, for example, performing the method described in the above embodiment. In the random access memory (Random Access Memory, RAM) 903, various programs and data required for system operation are also stored. A central processing unit (Central Processing Unit, CPU) 901, a Read-Only Memory (ROM) 902, and a random access Memory (Random Access Memory, RAM) 903 are connected to each other through a bus 904. An Input/Output (I/O) interface 905 is also connected to bus 904.
The following components are connected to the I/O interface 905: an input section 906 including a keyboard, a mouse, and the like; an output section 907 including a speaker and the like, such as a Cathode Ray Tube (CRT), a liquid crystal display (Liquid Crystal Display, LCD), and the like; a storage portion 908 including a hard disk or the like; and a communication section 909 including a network interface card such as a LAN (Local Area Network ) card, a modem, or the like. The communication section 909 performs communication processing via a network such as the internet. The drive 910 is also connected to the I/O interface 905 as needed. Removable media 911 such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory, or the like is installed as needed on the drive 910 so that a computer program read out therefrom is installed as needed into the storage section 908.
In particular, according to embodiments of the present application, the processes described above with reference to flowcharts may be implemented as computer software programs. For example, embodiments of the present application include a computer program product comprising a computer program embodied on a computer readable medium, the computer program comprising a computer program for performing the method shown in the flowchart. In such an embodiment, the computer program may be downloaded and installed from the network via the communication portion 909 and/or installed from the removable medium 911. When the computer program is executed by the central processing unit (Central Processing Unit, CPU) 901, various functions defined in the system of the present application are performed.
It should be noted that, the computer readable medium shown in the embodiments of the present application may be a computer readable signal medium or a computer readable storage medium, or any combination of the two. The computer readable storage medium may be, for example, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination thereof. More specific examples of the computer-readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-Only Memory (ROM), an erasable programmable read-Only Memory (Erasable Programmable Read Only Memory, EPROM), flash Memory, an optical fiber, a portable compact disc read-Only Memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the present application, a computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave, with a computer-readable computer program embodied therein. Such a propagated data signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination of the foregoing. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. A computer program embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: wireless, wired, etc., or any suitable combination of the foregoing.
The flowcharts and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present application. Where each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams or flowchart illustration, and combinations of blocks in the block diagrams or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The units involved in the embodiments of the present application may be implemented by means of software, or may be implemented by means of hardware, and the described units may also be provided in a processor. Wherein the names of the units do not constitute a limitation of the units themselves in some cases.
Another aspect of the present application also provides a computer readable storage medium having stored thereon a computer program which, when executed by a processor, implements the detection method as described above. The computer-readable storage medium may be included in the electronic device described in the above embodiment or may exist alone without being incorporated in the electronic device.
The foregoing is merely a preferred exemplary embodiment of the present application and is not intended to limit the embodiments of the present application, and those skilled in the art may make various changes and modifications according to the main concept and spirit of the present application, so that the protection scope of the present application shall be subject to the protection scope of the claims.
Claims (8)
1. A method of detection comprising:
obtaining an appearance picture of the battery module;
inputting the appearance picture into a coding layer, and extracting the characteristics of the appearance picture through the coding layer;
the coding layer comprises at least one three-dimensional feature extraction module which is arranged in sequence; the coding layer also comprises a data preprocessing module and a multidimensional feature extraction module;
the data preprocessing module normalizes the appearance picture into an input vector;
the multi-dimensional feature extraction module extracts a feature map from the input vector, the feature map comprising a low-level feature map and/or a high-level feature map, the low-level feature map comprising low-level spatial features; the high-level feature map comprises deep semantic information; the multidimensional feature extraction module is designed into a depth separable convolution layer, and two types of feature graphs are obtained from an input image by using a standardized function and a ReLU function;
the three-dimensional feature extraction module converts the appearance picture into a three-dimensional image, and extracts a three-dimensional feature map from the three-dimensional image, wherein the three-dimensional feature map comprises three-dimensional spectral features and/or three-dimensional spatial features;
analyzing the characteristics extracted by the coding layer through the decoding layer, and outputting a result;
the decoding layer comprises a transposed convolution decoding module and an output module, wherein the transposed convolution decoding module comprises a first decoding module and at least one second decoding module which are arranged in sequence; the first decoding module comprises a separable convolution layer and a transposed convolution layer, and the second decoding module comprises a multi-input attention module, the separable convolution layer and the transposed convolution layer; the output module analyzes the learning of the decoding layer through the convolution layer and the softmax function, outputs a result and judges whether defects exist or not.
2. A detection method according to claim 1, wherein the multi-dimensional feature extraction modules are arranged as a first extraction module and at least one second extraction module arranged in sequence,
the first extraction module inputs the extracted features to a second extraction module for sorting the first features;
the second extraction module inputs the extracted features to the second extraction module of the next order.
3. A detection method according to claim 2, wherein the first decoding module inputs the extracted features to a first-ranked second decoding module;
the second decoding module inputs the extracted features to the second decoding module of the next order.
4. A detection method according to claim 3, wherein the multi-dimensional feature extraction module, the three-dimensional feature extraction module and the built-in module of the transposed convolutional decoding module are all ordered according to a processed size criterion;
and the characteristics extracted by the multidimensional characteristic extraction module and the built-in module of the three-dimensional characteristic extraction module are input to the built-in module of the transposed convolution decoding module according to the size standard.
5. The method according to claim 4, wherein the step of merging the low-level feature map, the three-dimensional feature map, and the high-level feature map by the multi-input attention module is as follows:
splicing the low-level characteristic diagram, the three-dimensional characteristic diagram and the high-level characteristic diagram,
;
wherein,representing a splicing operation->Is output after splicing, is->Three dimensions of width, height and depth of the feature map are respectively corresponding to +.>Representing a low-level feature map obtained by the multi-dimensional feature extraction module, S 1 Depth of representation low-level feature map, +.>Representing the three-dimensional feature map obtained by the three-dimensional feature extraction module, S 2 Representing the depth of the three-dimensional feature map,representing the decoded advanced feature map, S 3 Representing the depth of the advanced feature map;
obtaining the average value of all feature graphs based on global pooling layer,/>Wherein->Representing a global pooling layer of the pool,
obtainingAfter that, the processing unit is configured to,obtaining a weight vector coefficient ++through two full connection layers and a sigmoid function>,
;
Wherein the method comprises the steps ofRepresenting a full connection layer, < >>Representing the activation function of the fully connected layer,
obtaining a weight matrix of the feature map through copying operation,
;
Wherein Repeat represents a function of the copy operation;
finally, the weight matrix of the feature mapMultiplying the spliced output->Get the output of the multiple input attention module +.>,
;
Is a weight matrix of the feature map.
6. A detection apparatus, characterized by comprising:
the input module is configured to acquire an appearance picture of the battery module;
the coding layer module is configured to extract the characteristics of the appearance pictures to obtain image characteristics;
the coding layer module comprises at least one three-dimensional feature extraction module which is arranged in sequence; the coding layer module further comprises a data preprocessing module and a multidimensional feature extraction module;
the data preprocessing module is used for normalizing the appearance picture into an input vector;
the multidimensional feature extraction module is used for extracting a feature map from the input vector, wherein the feature map comprises a low-level feature map and/or a high-level feature map; the low-level feature map comprises low-level spatial features; the high-level feature map comprises deep semantic information; the multidimensional feature extraction module is designed into a depth separable convolution layer, and two types of feature graphs are obtained from an input image by using a standardized function and a ReLU function;
the three-dimensional feature extraction module is used for converting the appearance picture into a three-dimensional image and extracting a three-dimensional feature image from the three-dimensional image, wherein the three-dimensional feature image comprises three-dimensional spectral features and/or three-dimensional spatial features;
the decoding layer module is configured to analyze the image characteristics and output a result; the decoding layer module comprises a transposed convolution decoding module and an output module, wherein the transposed convolution decoding module comprises a first decoding module and at least one second decoding module which are arranged in sequence; the first decoding module comprises a separable convolution layer and a transposed convolution layer, and the second decoding module comprises a multi-input attention module, the separable convolution layer and the transposed convolution layer; the output module analyzes the learning of the decoding layer through the convolution layer and the softmax function, outputs a result and judges whether defects exist or not.
7. An electronic device, comprising:
one or more processors; storage means for storing one or more programs which, when executed by the one or more processors, cause the electronic device to implement the detection method of any of claims 1 to 5.
8. A computer readable medium having stored thereon computer readable instructions which, when executed by a processor of a computer, cause the computer to perform the detection method of any of claims 1 to 5.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311021779.0A CN116773534B (en) | 2023-08-15 | 2023-08-15 | Detection method and device, electronic equipment and computer readable medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311021779.0A CN116773534B (en) | 2023-08-15 | 2023-08-15 | Detection method and device, electronic equipment and computer readable medium |
Publications (2)
Publication Number | Publication Date |
---|---|
CN116773534A CN116773534A (en) | 2023-09-19 |
CN116773534B true CN116773534B (en) | 2024-03-05 |
Family
ID=88006647
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202311021779.0A Active CN116773534B (en) | 2023-08-15 | 2023-08-15 | Detection method and device, electronic equipment and computer readable medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116773534B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117173182B (en) * | 2023-11-03 | 2024-03-19 | 厦门微亚智能科技股份有限公司 | Defect detection method, system, equipment and medium based on coding and decoding network |
Citations (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110490863A (en) * | 2019-08-22 | 2019-11-22 | 北京红云智胜科技有限公司 | Whether there is or not the systems of total occlusion for detection coronary angiography based on deep learning |
CN111145186A (en) * | 2019-12-17 | 2020-05-12 | 中国科学院深圳先进技术研究院 | Neural network structure, image segmentation method, device and storage medium |
CN113393582A (en) * | 2021-05-24 | 2021-09-14 | 电子科技大学 | Three-dimensional object reconstruction algorithm based on deep learning |
CN113822207A (en) * | 2021-09-27 | 2021-12-21 | 海南长光卫星信息技术有限公司 | Hyperspectral remote sensing image identification method and device, electronic equipment and storage medium |
CN113902611A (en) * | 2021-10-09 | 2022-01-07 | Oppo广东移动通信有限公司 | Image beautifying processing method and device, storage medium and electronic equipment |
US11222217B1 (en) * | 2020-08-14 | 2022-01-11 | Tsinghua University | Detection method using fusion network based on attention mechanism, and terminal device |
WO2022012459A1 (en) * | 2020-07-17 | 2022-01-20 | 中国人民解放军军事科学院军事医学研究院 | Parasite detection method and system based on artificial intelligence, and terminal device |
CN114037833A (en) * | 2021-11-18 | 2022-02-11 | 桂林电子科技大学 | Semantic segmentation method for Miao-nationality clothing image |
CN114066913A (en) * | 2022-01-12 | 2022-02-18 | 广东工业大学 | Heart image segmentation method and system |
CN114491083A (en) * | 2022-04-01 | 2022-05-13 | 江苏智云天工科技有限公司 | Knowledge graph library construction method and knowledge graph library construction device in industrial detection |
CN115170502A (en) * | 2022-06-30 | 2022-10-11 | 安徽大学 | Femoral pulley width measuring method based on deep learning |
CN115526863A (en) * | 2022-09-30 | 2022-12-27 | 华侨大学 | Cylindrical lithium battery surface defect detection method and device |
CN115546570A (en) * | 2022-08-25 | 2022-12-30 | 西安交通大学医学院第二附属医院 | Blood vessel image segmentation method and system based on three-dimensional depth network |
CN115619738A (en) * | 2022-10-18 | 2023-01-17 | 宁德思客琦智能装备有限公司 | Detection method for module side seam welding after welding |
CN115690013A (en) * | 2022-10-18 | 2023-02-03 | 福建工程学院 | Method and equipment for detecting surface defects of power battery module |
WO2023015409A1 (en) * | 2021-08-09 | 2023-02-16 | 百果园技术(新加坡)有限公司 | Object pose detection method and apparatus, computer device, and storage medium |
CN116416244A (en) * | 2023-04-26 | 2023-07-11 | 山东大学 | Crack detection method and system based on deep learning |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112215223B (en) * | 2020-10-16 | 2024-03-19 | 清华大学 | Multidirectional scene character recognition method and system based on multi-element attention mechanism |
US20230035475A1 (en) * | 2021-07-16 | 2023-02-02 | Huawei Technologies Co., Ltd. | Methods and systems for semantic segmentation of a point cloud |
-
2023
- 2023-08-15 CN CN202311021779.0A patent/CN116773534B/en active Active
Patent Citations (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110490863A (en) * | 2019-08-22 | 2019-11-22 | 北京红云智胜科技有限公司 | Whether there is or not the systems of total occlusion for detection coronary angiography based on deep learning |
CN111145186A (en) * | 2019-12-17 | 2020-05-12 | 中国科学院深圳先进技术研究院 | Neural network structure, image segmentation method, device and storage medium |
WO2022012459A1 (en) * | 2020-07-17 | 2022-01-20 | 中国人民解放军军事科学院军事医学研究院 | Parasite detection method and system based on artificial intelligence, and terminal device |
US11222217B1 (en) * | 2020-08-14 | 2022-01-11 | Tsinghua University | Detection method using fusion network based on attention mechanism, and terminal device |
CN113393582A (en) * | 2021-05-24 | 2021-09-14 | 电子科技大学 | Three-dimensional object reconstruction algorithm based on deep learning |
WO2023015409A1 (en) * | 2021-08-09 | 2023-02-16 | 百果园技术(新加坡)有限公司 | Object pose detection method and apparatus, computer device, and storage medium |
CN113822207A (en) * | 2021-09-27 | 2021-12-21 | 海南长光卫星信息技术有限公司 | Hyperspectral remote sensing image identification method and device, electronic equipment and storage medium |
CN113902611A (en) * | 2021-10-09 | 2022-01-07 | Oppo广东移动通信有限公司 | Image beautifying processing method and device, storage medium and electronic equipment |
CN114037833A (en) * | 2021-11-18 | 2022-02-11 | 桂林电子科技大学 | Semantic segmentation method for Miao-nationality clothing image |
CN114066913A (en) * | 2022-01-12 | 2022-02-18 | 广东工业大学 | Heart image segmentation method and system |
CN114491083A (en) * | 2022-04-01 | 2022-05-13 | 江苏智云天工科技有限公司 | Knowledge graph library construction method and knowledge graph library construction device in industrial detection |
CN115170502A (en) * | 2022-06-30 | 2022-10-11 | 安徽大学 | Femoral pulley width measuring method based on deep learning |
CN115546570A (en) * | 2022-08-25 | 2022-12-30 | 西安交通大学医学院第二附属医院 | Blood vessel image segmentation method and system based on three-dimensional depth network |
CN115526863A (en) * | 2022-09-30 | 2022-12-27 | 华侨大学 | Cylindrical lithium battery surface defect detection method and device |
CN115619738A (en) * | 2022-10-18 | 2023-01-17 | 宁德思客琦智能装备有限公司 | Detection method for module side seam welding after welding |
CN115690013A (en) * | 2022-10-18 | 2023-02-03 | 福建工程学院 | Method and equipment for detecting surface defects of power battery module |
CN116416244A (en) * | 2023-04-26 | 2023-07-11 | 山东大学 | Crack detection method and system based on deep learning |
Non-Patent Citations (2)
Title |
---|
基于深度自编码网络的轴承故障诊断;袁文军 等;《噪声与振动控制》(第5期);全文 * |
机器学习在设计高性能锂电池正极材料与电解质中的应用;刘振东 等;《化学进展》;第35卷(第4期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN116773534A (en) | 2023-09-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110188765B (en) | Image semantic segmentation model generation method, device, equipment and storage medium | |
US20240233313A1 (en) | Model training method, image processing method, computing and processing device and non-transient computer-readable medium | |
CN116773534B (en) | Detection method and device, electronic equipment and computer readable medium | |
CN111914654B (en) | Text layout analysis method, device, equipment and medium | |
CN111444807B (en) | Target detection method, device, electronic equipment and computer readable medium | |
US20210056353A1 (en) | Joint representation learning from images and text | |
CN112446869A (en) | Unsupervised industrial product defect detection method and device based on deep learning | |
CN116958131B (en) | Image processing method, device, equipment and storage medium | |
CN114841974B (en) | Nondestructive testing method, nondestructive testing system, nondestructive testing electronic equipment and nondestructive testing medium for internal structure of fruit | |
CN113920208A (en) | Image processing method and device, computer readable storage medium and electronic device | |
US20230351558A1 (en) | Generating an inpainted image from a masked image using a patch-based encoder | |
CN117557775A (en) | Substation power equipment detection method and system based on infrared and visible light fusion | |
CN118097089A (en) | Night warehousing robot target detection method and system based on integral network | |
CN112465050B (en) | Image template selection method, device, equipment and storage medium | |
CN117173182B (en) | Defect detection method, system, equipment and medium based on coding and decoding network | |
CN112800851B (en) | Water body contour automatic extraction method and system based on full convolution neuron network | |
Li et al. | Multi-Scale Fusion Siamese Network Based on Three-Branch Attention Mechanism for High-Resolution Remote Sensing Image Change Detection | |
CN111145178A (en) | High-resolution remote sensing image multi-scale segmentation method | |
CN117710295A (en) | Image processing method, device, apparatus, medium, and program product | |
CN116309612A (en) | Semiconductor silicon wafer detection method, device and medium based on frequency decoupling supervision | |
CN116935240A (en) | Surface coverage classification system and method for multi-scale perception pyramid | |
CN115205650B (en) | Unsupervised abnormal positioning and detecting method and unsupervised abnormal positioning and detecting device based on multi-scale standardized flow | |
Mire et al. | Automated approach for splicing detection using first digit probability distribution features | |
CN116257852A (en) | Chip hardware Trojan horse detection method based on differential curvature | |
CN115761268A (en) | Pole tower key part defect identification method based on local texture enhancement network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |