CN111191583B - Space target recognition system and method based on convolutional neural network - Google Patents
Space target recognition system and method based on convolutional neural network Download PDFInfo
- Publication number
- CN111191583B CN111191583B CN201911388125.5A CN201911388125A CN111191583B CN 111191583 B CN111191583 B CN 111191583B CN 201911388125 A CN201911388125 A CN 201911388125A CN 111191583 B CN111191583 B CN 111191583B
- Authority
- CN
- China
- Prior art keywords
- target
- layer
- neural network
- image
- target recognition
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/41—Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/082—Learning methods modifying the architecture, e.g. adding, deleting or silencing nodes or connections
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/513—Sparse representations
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Computational Linguistics (AREA)
- Molecular Biology (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Evolutionary Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a space target recognition system and method based on a convolutional neural network, which are used for solving the technical problem that the existing deep learning cannot achieve both recognition accuracy and calculated amount. The method comprises model training and target recognition, wherein the model training comprises the steps of inputting positive and negative samples, extracting features by using a sparse convolutional neural network, and training a classifier to form a learning dictionary; the target recognition comprises input image preprocessing, input model matching and output result. The beneficial technical effects of the invention are as follows: on the premise of ensuring accuracy, the recognition efficiency is improved, the storage space is saved, and the more efficient hardware design is facilitated.
Description
Technical Field
The invention relates to the technical field of deep learning, in particular to a space recognition system and method based on a convolutional neural network.
Background
Spatial destination identification refers to the process by which a particular object (or type of object) is distinguished from other objects (or types of objects). It includes both the identification of two very similar objects and the identification of one type of object with the other. Spatial object recognition is how objects are extracted from sequential images. The space target detection and recognition is a popular direction of computer vision and digital image processing, and is widely applied to the fields of robot navigation, intelligent video monitoring, industrial detection, aerospace and the like.
In recent years, with the development of artificial intelligence technology, image recognition technology based on deep learning has been greatly advanced, and important fields such as medical treatment, automatic driving, fashion and the like are brought into the field of floor application around the technology. Currently, extraction of model features using deep learning tools has become a mainstream technique. The deep learning field also generates a plurality of models, such as classical convolutional neural networks, deep confidence networks, automatic encoders, generation countermeasure networks and the like, wherein the convolutional neural networks have great success in the fields of voice, images, videos and the like due to the strong capability of abstracting and extracting high-level semantic features, and are widely applied.
However, the conventional convolutional neural network needs to extract as many features as possible when the precise recognition result is desired, but too many features are unfavorable for the calculation speed, and the burden of hardware is increased, even the large calculation amount cannot be supported, so that the problem of how to simplify the calculation on the premise of ensuring the recognition precision is urgent to be solved.
Disclosure of Invention
The invention provides a space target recognition system and a space target recognition method based on a convolutional neural network, which are used for solving the technical problem that the conventional deep learning cannot achieve both recognition accuracy and calculation amount.
In order to solve the technical problems, the invention adopts the following technical scheme:
the space target recognition system based on the convolutional neural network comprises a calculation unit and a storage unit, wherein the calculation unit comprises a convolutional layer for extracting features, a pooling layer for reducing dimensionality and a full-connection layer for combining features, and the convolutional layer comprises a sparse convolutional module for separating a target and a background; the storage unit comprises a RAM module for storing weight values and bias values and a ROM module for storing classification labels.
Further, the convolution layer comprises a first data buffer module and a feature classification module, the pooling layer comprises a second data buffer module, and the full-connection layer comprises a multiplier, an accumulator, a comparator and a residual calculator.
Further, the sparse convolution module includes an inner product calculator of the sparse vector and the dense vector.
The method for constructing the space target recognition model comprises the following steps:
s1: inputting a training sample set, selecting positive and negative target samples with labels and background patterns, manufacturing an aerial target static and video image data set, and storing the air target static and video image data set in a hard disk of the equipment;
s2: establishing a sparse convolutional neural network, wherein the network adopts a pruning algorithm to divide the characteristics of a space target into a sparse characteristic matrix and a dense characteristic matrix, the sparse characteristic matrix represents the background, and the dense characteristic matrix represents the target;
s3: initializing parameters of the neural network, and then carrying out feature extraction, feature classification and sample labeling on samples in the training sample set one by one, wherein the spatial target recognition system updates the parameters of the neural network in the operation process; the processing of the features mainly comprises:
(1) Color characteristics and selections;
(2) Texture features and selections;
(3) Shape characteristics and selection;
(4) Spatial features and selection;
s4: after all the samples are identified, a space target identification model can be obtained.
Further, the training sample set comprises ten thousands of positive and negative target samples, wherein the positive target sample refers to an image sample containing the space target, and the negative target sample refers to other image samples without the space target.
Further, after the convolutional neural network performs convolutional operation on the image, the result is input into the pooling layer after passing through the activation function ReLU and the bias.
Further, the features of the spatial target also include color, texture, shape.
Further, the training sample set needs to be preprocessed before training: coding, setting a threshold, improving modes, normalizing and performing discrete mode operation.
The method for identifying the space target comprises the following steps:
i, inputting an image to be identified into a storage unit in a space target identification model;
II, extracting features of the image by using a sparse convolutional neural network;
and III, matching the parameters of the characteristics with parameters in the space target recognition model, if the parameters are completely matched, indicating that the image to be recognized contains a target, otherwise, judging that the image is a background.
And further, preprocessing the image in the step II before extracting the characteristics, and adjusting the image to be the same as the apparent characteristics of the training sample of the space target recognition model.
Compared with the prior art, the invention has the beneficial technical effects that:
1. according to the invention, a sparse convolutional neural network is utilized, an image segmentation and activation function optimization technology is adopted, a network depth and width and a link mode are effectively designed, a target feature model is extracted, task retrieval and accurate classification based on the features of the extracted target such as shape, color, texture, illumination, position relation and semantic expression are trained, and a target recognition classifier is formed, so that the recognition accuracy and the recognition efficiency of a space target are improved.
2. The invention realizes that coarse-granularity pruning can realize the similar sparse proportion as unstructured pruning, almost has no loss in precision, balances the sparse rule and the prediction precision through quantization, and solves the problem of maintaining the precision of the model while ensuring more structured sparsity.
3. The sparse structure saves the storage space due to the index, and the coarse-granularity pruning can achieve better compression rate than fine granularity under the same accuracy threshold, so that the efficiency of hardware acceleration design and the accuracy of a model are improved, and the method is more suitable for the purpose of hardware acceleration.
4. The sparse structure solves the problem that coarse-grain sparse and fine-grain sparse can save twice the storage space, and the storage cost is two orders of magnitude higher than the operation cost of arithmetic operation, so that the regularity of the sparse structure can lead to more efficient hardware design.
Drawings
FIG. 1 is a hardware architecture diagram of a convolutional neural network-based spatial target recognition system of the present invention.
FIG. 2 is a flowchart of a training method of a spatial target recognition model based on a convolutional neural network.
Fig. 3 is a schematic diagram of a convolutional neural network-based spatial target recognition method of the present invention.
Fig. 4 is a schematic diagram of signal connections of a conventional convolutional neural network.
Fig. 5 is a schematic diagram of signal connections of a convolutional neural network of the present invention.
Fig. 6 is a schematic diagram of a sparse convolutional neural network target recognition algorithm according to the present invention.
Detailed Description
The following examples are given to illustrate the invention in detail, but are not intended to limit the scope of the invention in any way.
Example 1: referring to fig. 1, the whole system hardware is designed according to a CNN network structure, each layer is designed into a single module by utilizing the parallel characteristic of an FPGA hardware circuit, the final classification layer finishes the classification result of input data by utilizing an integrated learning dictionary classifier, and then the weight to be updated is calculated by a back propagation algorithm. The system firstly initializes the weight value index address of each layer of convolution Kernel Kernel by the controller, and loads the weight value and the offset value from the RAM module according to the index address. And when in forward propagation, input data enters a data buffer area through an input signal, then convolution operation is completed according to the number of the output characteristic diagrams of each layer and a convolution kernel, and the operation result is subjected to an activation function ReLU and offset to complete the final output characteristic diagram of the current layer and is input into the next layer. Downsampling operations are performed as they pass through the pooling layer, thereby reducing the dimensionality of the feature map. Finally, the Softmax classifier searches the corresponding numerical value in the ROM according to the input data, and then completes the final output result after probability conversion. And when the back propagation is carried out, comparing the label in the ROM with the output result to obtain residual errors of each layer, storing the residual errors in the RAM, completing updating of the weight and the offset value of the convolution kernels in all the convolution layers according to the set learning rate after calculation is completed, and storing the updated weight into corresponding storage positions by the controller until all training data are input.
Example 2: referring to fig. 2, the model firstly extracts features for a sample, classifies the identified features, such as a car lamp is marked as a, a car wheel is marked as B, and the classified features are combined to form a classifier for identifying different car models, namely a classification dictionary, so that the model is constructed, and the method is as follows:
(1) Creating positive and negative samples
The positive sample refers to a sample of the object to be detected, in this embodiment, the object is an automobile, the negative sample refers to any other picture that does not include the object, such as a background, and all sample pictures are normalized to the same size, in this embodiment, 20×20. In order to make the classification detection accuracy better, training samples are set to 1 ten thousand space target static and video image data so as to ensure the accuracy requirement of training.
(2) Extracting target features
The feature extraction is to transform the original data to obtain the features which can reflect the classification essence most. Mainly comprises the following steps: color, texture, shape, and spatial features.
Color features describe the surface properties of the scene corresponding to the image or image region, and common color features include image patch features, color channel histogram features and the like; texture is generally defined as a local property of an image or a measure of the relationship between pixels in a local region, and the texture feature extraction of the present invention adopts a method based on a spatial correlation matrix of gray level, i.e. a co-occurrence matrix; the shape is one of the basic characteristics of the object, and the shape characteristics are represented by two types, namely a shape outline characteristic description and a shape area characteristic. The shape profile features are mainly: straight line segment description, spline fitting curve, boli She Miaoshu, interior angle histogram, gaussian parameter curve, etc., the shape region features mainly are: extraneous moment of shape, area of region, aspect ratio of shape, etc.; the spatial features refer to the mutual spatial position or relative direction relation among a plurality of targets segmented in an image, and have relative position information, such as upper, lower, left and right, absolute position information, and the basic idea of the method for extracting the spatial features is that after the image is segmented, the features are extracted, and then the features are indexed.
(3) Modeling
The object modeling is a geometric description of important features of the object, the representation of the model contains all relevant feature information reflecting the object, but without any redundant feature information, and the feature information is organized in such a way that different components of the object recognition system can easily access the feature information. The model object is a spatial structural relationship between features; the main selection criteria, one is whether the assumptions of the model apply to the current problem; and secondly, whether the computational complexity required by the model can be tolerated or whether there is an algorithm that is as efficient, accurate or approximate as possible. The object model expression method mainly comprises the following steps: the local statistical features, i.e. the generative model, or the interrelationship definition between the object features, such as the positional relationship, etc., i.e. the discontinative model, or a hybrid model of both. The method for modeling the target comprises the following steps: (1) Entity representation, solid geometry, point cloud, volume network, voxels; 2) Boundary representations such as surface meshes, parametric surfaces, subdivision surfaces, implicit surfaces. In the description model, each class of object is modeled, and then class information is obtained by using maximum likelihood or Bayesian inference.
(4) Matching
The matching is also called feature selection, which is to match the target feature according to the feature template of the target, and if the feature value is completely matched, the recognition result is obtained, namely the feature selection. The target matching process is to use a scanning sub-window to continuously shift and slide in the image to be detected, calculate the characteristics of the region every time the sub-window reaches a position, and then use a classifier trained by us to screen the characteristics to determine whether the region is a target. Then, because the size of the target image may be different from the sample picture size that you use when training the classifier, the sub-window for this scan needs to be larger or smaller, then slide in the image, and then match one pass.
(5) Positioning
And constructing a space target data set according to the training sample data set format standard, wherein each image in the data set corresponds to one label, the labels mark the image name, the type of the initial target in the image, the coordinates and width and height of the circumscribed rectangle of the target, and thus the model construction is completed.
Example 3: referring to fig. 2, the target recognition method mainly comprises spatial target image input, target preprocessing, sparse convolutional neural network, feature extraction and feature selection, wherein the extracted target features are matched and compared with a target feature integrated learning dictionary, and finally the recognized target is output. Specific:
1. pretreatment of
The method is characterized in that preprocessing operations such as image information acquisition, analog-to-digital conversion, filtering, fuzzy elimination, noise reduction, geometric distortion correction and the like are completed, and on the premise that essential information carried by the images is not changed, the apparent characteristics (such as object shape edges, color distribution, textures, overall brightness, size and the like) of each image are consistent as much as possible, so that the subsequent processing process is facilitated.
The preprocessing described above includes five operations, (1) encoding: the effective description of the mode is suitable for computer operation. (2) threshold or filter operation: some functions are selected as needed, and others are suppressed. (3) mode improvement: errors in the pattern are excluded or corrected, or unnecessary function values. (4) normalization: some parameter values are adapted to standard values, or standard value ranges. (5) discrete mode operation: special operations in discrete mode processing.
2. Sparse convolutional neural network
Referring to fig. 5, a sparse convolutional neural network is used to perform positive and negative sample processing, that is, object and background separation (pruning algorithm), to obtain object features, to form a two-part matrix with sparse features and dense features, the sparse matrix is characterized by the background, the dense matrix is characterized by the object, and the two-part matrix is respectively represented as W, I. Meanwhile, the background matrix W can be further divided into W1 and I1 to extract finer target features. Therefore, a multi-level fine target classifier (target feature learning dictionary) can be constructed through systematic training. The essence of the process of target identification is that the background separation is carried out on a space target by adopting sparse nerve convolution, the characteristic of the space target is extracted, and then the characteristic is compared with the characteristic target in the classifier on the basis, and the higher the similarity is, the identification result is obtained.
Compared with the traditional neural network, the traditional convolutional neural network adopts an input-output layer full-connection structure, the calculated amount is large, the sparse convolutional neural network adopts a pruning technology to separate a target and a background to form two sparse (coarse granularity) and dense (fine granularity) matrixes, and then convolutional calculation (characteristic extraction) is carried out to circularly form two finer sparse (coarse granularity) and dense (fine granularity) matrixes until the lost information energy in the process of extracting the characteristic value is minimum. The pruning algorithm is to prune the connections between two neurons in the neural network, see fig. 5, i.e. to set the corresponding coefficients to zero.
The feature extraction by using the sparse convolution network is to directly sparse convolve the space picture on the basis of pruning, and the data replication in the input feature matrix is removed by a direct sparse convolution algorithm, which expands the scale of the convolution kernel matrix to the same size of the input matrix. And developing the extended convolution kernel row to generate a vector Wm, wherein the length of the vector Wm is C, H and W. Assuming that the convolution layer has M convolution kernels, a weight matrix M (c×h×w) is obtained by extending each convolution kernel. For the input matrix under the batch task, a column vector I is formed in a row expansion manner, and the length of the column vector I is c×h×w. Then, in computing the convolution, the elements for the different sensing regions may be mapped to the correct local region by adjusting the starting pointer of vector I.
Feature extraction includes extension of the convolution kernel matrix and updating of the feature pointers. In the extension of the convolution kernel matrix, a sparse matrix with a weight of M (c×r×s) is expressed in CSR format, and for the j-th non-zero element (C, y, x) in the output channel M, there is:
x=col%s
y=(col/s)%R
C=col/(S*R)
where col=colidx [ j ], then the extended weight matrix size is M (c×h×w), and the CSR format column pointer of the same non-zero element (C, y, x) is updated as: colidx [ j ] = (c×h+y) ×w+x.
The update of the pointer can be expressed as: o (O) m,y,x =W m I y,w+x . The core is to realize the inner product operation of sparse vector and dense vector. For calculating the points (m, y, x) in the output matrix, the MAC operation that needs to be done depends on the sparse vector W m Is a non-zero element number of (c). Since the sparse vector Wm is constant for all points in the same output channel m, the MAC operands required to calculate these output nodes are equal. Matrix I in a direct sparse convolution algorithm virtual Is generated from vector I, wherein the element pointed to by the start pointer of each column vector is I [ y, W+x]. According to this feature, we simply reside the elements of vector I in memory, rather than storing the entire dense matrix.
Considering that the sparsity of all the convolution layers subjected to weight pruning in the convolution CNN model is different, the sparsity of the current convolution layer is calculated by the following method:
Sparsity=1.0-Nnonzero/M*kernel_size
where Nnonzero is the number of all non-zero elements of the current convolutional layer, M is the number of output channels of the current convolutional layer, and kernel_size is the size of the convolutional kernel. For convolutional layers of different sparsities, a threshold is set. The convolution layer with the sparsity larger than the threshold value adopts an optimized direct sparse convolution mode, and the convolution layer with the sparsity smaller than the threshold value still adopts a traditional dimension reduction mode.
3. Feature extraction and selection
And transforming the original data to obtain the characteristics which can reflect the classification essence most. Mainly comprises the following steps: (1) color characterization. Color features describe the surface properties of the scene corresponding to the image or image region, and common color features include image patch features, color channel histogram features and the like; (2) texture features. Texture is generally defined as some local property of an image or a measure of the relationship between pixels in a local area. One effective method for extracting texture features is based on a gray level space correlation matrix, namely a co-occurrence matrix, and other methods are also based on feature extraction of an image friend difference histogram and feature extraction of an image gray level co-occurrence matrix; (3) shape features. The shape is one of the basic characteristics of the object, and the shape characteristics are represented by two types, namely a shape outline characteristic description and a shape area characteristic. The shape profile features are mainly: straight line segment description, spline fitting curve, boli She Miaoshu, interior angle histogram, gaussian parameter curve, etc., the shape region features mainly are: the irrelevant moment of the shape, the area of the region, the aspect ratio of the shape, etc. (4) spatial features. The spatial features refer to the mutual spatial position or relative direction relation among a plurality of targets segmented in an image, and have relative position information, such as upper, lower, left and right, absolute position information, and the basic idea of a common method for extracting the spatial features is to segment the image, extract the features, and then index the features. Algorithm for feature extraction: including Haar features, LBP features, haar-like features.
The feature selection is to match the target feature according to the feature template of the target, and if the feature value is completely matched, the recognition result is obtained, namely the feature selection.
4. Outputting the result
And marking the images matched with the model features as 'targets', storing the images in a memory unit, marking the images which are not matched as 'no targets', and storing the images in the memory unit, so as to finish one-time target identification.
While the present invention has been described in detail with reference to the drawings and the embodiments, those skilled in the art will understand that various specific parameters in the above embodiments may be changed without departing from the spirit of the invention, and a plurality of specific embodiments are common variation ranges of the present invention, and will not be described in detail herein.
Claims (8)
1. The space target recognition system based on the convolutional neural network comprises a calculation unit and a storage unit, and is characterized in that the calculation unit comprises a convolutional layer for extracting features, a pooling layer for reducing dimensionality and a fully-connected layer for combining the features, and the convolutional layer comprises a sparse convolutional module for separating a target and a background; the storage unit comprises a RAM module for storing weight values and bias values and a ROM module for storing classification labels;
the whole system hardware is designed according to the CNN network structure, each layer is designed into a single module by utilizing the parallel characteristic of an FPGA hardware circuit, the last classification layer finishes the classification result of input data by utilizing an integrated learning dictionary classifier, and then the weight to be updated is calculated by a back propagation algorithm; the system firstly initializes the weight value index address of each layer of convolution Kernel Kernel by a controller, and loads weight values and offset values from the RAM module according to the index address; when in forward propagation, input data enters a data buffer area through an input signal, then convolution operation is completed according to the number of the output characteristic diagrams of each layer and a convolution kernel, and the operation result is subjected to an activation function ReLU and bias to complete the final output characteristic diagram of the current layer and is input into the next layer; performing downsampling operation when passing through the pooling layer, so as to reduce the dimension of the feature map; finally, the Softmax classifier searches the corresponding numerical value in the ROM according to the input data and completes a final output result after probability conversion; and when the back propagation is carried out, comparing the label in the ROM with the output result to obtain residual errors of each layer, storing the residual errors in the RAM, completing updating of the weight and the offset value of the convolution kernels in all the convolution layers according to the set learning rate after calculation is completed, and storing the updated weight into corresponding storage positions by the controller until all training data are input.
2. A method for constructing a space target recognition model, which relates to the space target recognition system based on a convolutional neural network as set forth in claim 1, and is characterized by comprising the following steps:
s1: inputting a training sample set;
s2: establishing a sparse convolutional neural network space target recognition system, wherein the system comprises a convolutional layer, a pooling layer and a full-connection layer, the convolutional layer adopts a pruning algorithm to divide the characteristics of a space target into a sparse characteristic matrix and a dense characteristic matrix, the sparse characteristic matrix represents a background, and the dense characteristic matrix represents a target;
s3: initializing parameters of the neural network, and then carrying out feature extraction, feature classification and sample labeling on samples in the training sample set one by one, wherein the spatial target recognition system updates the parameters of the neural network in the operation process;
s4: after all the samples are identified, a space target identification model can be obtained.
3. The method for constructing a space object recognition model according to claim 2, wherein the output result of the volume layer is input to the pooling layer after passing through an activation function ReLU and a bias.
4. The method of claim 2, wherein the training sample set comprises ten thousand positive and negative target samples, wherein a positive target sample refers to an image sample containing the spatial target, and a negative target sample refers to other image samples not containing the spatial target.
5. The method for constructing a spatial target recognition model according to claim 2, wherein the characteristics of the spatial target further include color, texture, shape.
6. The method for constructing a spatial target recognition model according to claim 2, wherein the training sample set needs to be preprocessed before training: coding, setting a threshold, improving modes, normalizing and performing discrete mode operation.
7. A method for identifying a space target, which relates to a space target identification model as claimed in claim 2, and is characterized by comprising the following steps:
i, inputting an image to be identified into a storage unit in the space target identification model;
II, extracting features of the image by using a sparse convolutional neural network;
and III, matching the parameters of the characteristics with parameters in the space target recognition model, if the parameters are completely matched, indicating that the image to be recognized contains a target, otherwise, judging that the image is a background.
8. The method according to claim 7, wherein the image in step ii is preprocessed before feature extraction, and the image is adjusted to have the same apparent characteristics as the training sample of the spatial target recognition model.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911388125.5A CN111191583B (en) | 2019-12-30 | 2019-12-30 | Space target recognition system and method based on convolutional neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911388125.5A CN111191583B (en) | 2019-12-30 | 2019-12-30 | Space target recognition system and method based on convolutional neural network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111191583A CN111191583A (en) | 2020-05-22 |
CN111191583B true CN111191583B (en) | 2023-08-25 |
Family
ID=70709514
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201911388125.5A Active CN111191583B (en) | 2019-12-30 | 2019-12-30 | Space target recognition system and method based on convolutional neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111191583B (en) |
Families Citing this family (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111709427B (en) * | 2020-05-26 | 2023-04-07 | 淮阴工学院 | Fruit segmentation method based on sparse convolution kernel |
CN111681318B (en) * | 2020-06-10 | 2021-06-15 | 上海城市地理信息系统发展有限公司 | Point cloud data modeling method and device and electronic equipment |
CN111721361A (en) * | 2020-06-29 | 2020-09-29 | 杭州鲁尔物联科技有限公司 | Embankment monitoring system, method and equipment |
CN113970922B (en) * | 2020-07-22 | 2024-06-14 | 商汤集团有限公司 | Point cloud data processing method, intelligent driving control method and device |
CN112365455A (en) * | 2020-10-29 | 2021-02-12 | 杭州富阳富创大数据产业创新研究院有限公司 | Transformer substation equipment detection method based on 3D point cloud and deep learning |
CN112365463A (en) * | 2020-11-09 | 2021-02-12 | 珠海市润鼎智能科技有限公司 | Real-time detection method for tiny objects in high-speed image |
CN112783650B (en) * | 2021-01-20 | 2024-01-16 | 之江实验室 | Multi-model parallel reasoning method based on AI chip |
CN113270156B (en) * | 2021-04-29 | 2022-11-15 | 甘肃路桥建设集团有限公司 | Detection modeling and detection method and system of machine-made sandstone powder based on image processing |
CN112990225B (en) * | 2021-05-17 | 2021-08-27 | 深圳市维度数据科技股份有限公司 | Image target identification method and device in complex environment |
CN113222288B (en) * | 2021-06-02 | 2022-05-17 | 山东建筑大学 | Classified evolution and prediction method of village and town community space development map |
CN113657215A (en) * | 2021-08-02 | 2021-11-16 | 大连理工大学 | Target tracking method applied to unmanned ship based on target detection |
CN114120406B (en) * | 2021-11-22 | 2024-06-07 | 四川轻化工大学 | Face feature extraction and classification method based on convolutional neural network |
CN115500342A (en) * | 2022-09-23 | 2022-12-23 | 国网河北省电力有限公司衡水供电分公司 | Bird repelling device, method, terminal and storage medium |
CN115393597B (en) * | 2022-10-31 | 2023-01-24 | 之江实验室 | Semantic segmentation method and device based on pulse neural network and laser radar point cloud |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108169745A (en) * | 2017-12-18 | 2018-06-15 | 电子科技大学 | A kind of borehole radar target identification method based on convolutional neural networks |
WO2018120740A1 (en) * | 2016-12-29 | 2018-07-05 | 深圳光启合众科技有限公司 | Picture classification method, device and robot |
CN108985454A (en) * | 2018-06-28 | 2018-12-11 | 西南电子技术研究所(中国电子科技集团公司第十研究所) | Airline carriers of passengers individual goal recognition methods |
CN110598731A (en) * | 2019-07-31 | 2019-12-20 | 浙江大学 | Efficient image classification method based on structured pruning |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107239824A (en) * | 2016-12-05 | 2017-10-10 | 北京深鉴智能科技有限公司 | Apparatus and method for realizing sparse convolution neutral net accelerator |
-
2019
- 2019-12-30 CN CN201911388125.5A patent/CN111191583B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018120740A1 (en) * | 2016-12-29 | 2018-07-05 | 深圳光启合众科技有限公司 | Picture classification method, device and robot |
CN108169745A (en) * | 2017-12-18 | 2018-06-15 | 电子科技大学 | A kind of borehole radar target identification method based on convolutional neural networks |
CN108985454A (en) * | 2018-06-28 | 2018-12-11 | 西南电子技术研究所(中国电子科技集团公司第十研究所) | Airline carriers of passengers individual goal recognition methods |
CN110598731A (en) * | 2019-07-31 | 2019-12-20 | 浙江大学 | Efficient image classification method based on structured pruning |
Non-Patent Citations (1)
Title |
---|
李思泉 ; 张轩雄 ; .基于卷积神经网络的人脸表情识别研究.软件导刊.2018,(01),全文. * |
Also Published As
Publication number | Publication date |
---|---|
CN111191583A (en) | 2020-05-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111191583B (en) | Space target recognition system and method based on convolutional neural network | |
CN108154118B (en) | A kind of target detection system and method based on adaptive combined filter and multistage detection | |
Xie et al. | Multilevel cloud detection in remote sensing images based on deep learning | |
CN108154192B (en) | High-resolution SAR terrain classification method based on multi-scale convolution and feature fusion | |
CN107066559B (en) | Three-dimensional model retrieval method based on deep learning | |
CN111027493B (en) | Pedestrian detection method based on deep learning multi-network soft fusion | |
CN110321967B (en) | Image classification improvement method based on convolutional neural network | |
CN109029363A (en) | A kind of target ranging method based on deep learning | |
CN110175615B (en) | Model training method, domain-adaptive visual position identification method and device | |
CN111460980B (en) | Multi-scale detection method for small-target pedestrian based on multi-semantic feature fusion | |
CN111126127B (en) | High-resolution remote sensing image classification method guided by multi-level spatial context characteristics | |
Tereikovskyi et al. | The method of semantic image segmentation using neural networks | |
CN114187450A (en) | Remote sensing image semantic segmentation method based on deep learning | |
CN106778768A (en) | Image scene classification method based on multi-feature fusion | |
CN111652273B (en) | Deep learning-based RGB-D image classification method | |
CN108230330B (en) | Method for quickly segmenting highway pavement and positioning camera | |
CN112101364B (en) | Semantic segmentation method based on parameter importance increment learning | |
CN110852327A (en) | Image processing method, image processing device, electronic equipment and storage medium | |
CN113269224A (en) | Scene image classification method, system and storage medium | |
CN115393631A (en) | Hyperspectral image classification method based on Bayesian layer graph convolution neural network | |
CN107423771B (en) | Two-time-phase remote sensing image change detection method | |
CN116863194A (en) | Foot ulcer image classification method, system, equipment and medium | |
CN112597919A (en) | Real-time medicine box detection method based on YOLOv3 pruning network and embedded development board | |
CN114492634B (en) | Fine granularity equipment picture classification and identification method and system | |
CN114648667A (en) | Bird image fine-granularity identification method based on lightweight bilinear CNN model |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |