CN110580503A - AI-based double-spectrum target automatic identification method - Google Patents
AI-based double-spectrum target automatic identification method Download PDFInfo
- Publication number
- CN110580503A CN110580503A CN201910780121.5A CN201910780121A CN110580503A CN 110580503 A CN110580503 A CN 110580503A CN 201910780121 A CN201910780121 A CN 201910780121A CN 110580503 A CN110580503 A CN 110580503A
- Authority
- CN
- China
- Prior art keywords
- image
- target
- cnn
- identified
- images
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 23
- 238000001228 spectrum Methods 0.000 title claims abstract description 13
- 238000000605 extraction Methods 0.000 claims abstract description 22
- 239000013598 vector Substances 0.000 claims abstract description 19
- 238000012549 training Methods 0.000 claims abstract description 15
- 238000013135 deep learning Methods 0.000 claims abstract description 14
- 238000013145 classification model Methods 0.000 claims abstract description 10
- 230000006870 function Effects 0.000 claims abstract description 8
- 238000013527 convolutional neural network Methods 0.000 claims description 52
- 238000003384 imaging method Methods 0.000 claims description 12
- 239000003086 colorant Substances 0.000 claims description 4
- 239000000203 mixture Substances 0.000 claims description 4
- 238000013528 artificial neural network Methods 0.000 description 5
- 238000006243 chemical reaction Methods 0.000 description 4
- 238000001514 detection method Methods 0.000 description 4
- 230000010365 information processing Effects 0.000 description 4
- 238000012545 processing Methods 0.000 description 4
- 238000011160 research Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 3
- 239000000284 extract Substances 0.000 description 3
- 238000003331 infrared imaging Methods 0.000 description 3
- 238000007781 pre-processing Methods 0.000 description 3
- 238000012706 support-vector machine Methods 0.000 description 3
- 230000000694 effects Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 210000002569 neuron Anatomy 0.000 description 2
- 238000011176 pooling Methods 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000007477 logistic regression Methods 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000003058 natural language processing Methods 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/211—Selection of the most significant subset of features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/211—Selection of the most significant subset of features
- G06F18/2115—Selection of the most significant subset of features by evaluating different subsets according to an optimisation criterion, e.g. class separability, forward selection or backward elimination
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Biophysics (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- General Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
the invention discloses an AI-based double-spectrum target automatic identification method, which comprises the following steps: step 1, constructing a target database; training a target database image classification model by a deep learning network; step 2, collecting an image to be identified through target collection equipment, wherein the image to be identified comprises: an image containing the object and an image not containing the object; step 3, CNN feature extraction is carried out on the image to be identified; step 4, extracting local CNN features with small dimension from the acquired image and converting the extracted local CNN features into binary codes; step 5, reducing the number of candidate target pictures by using binary coding; step 6, retrieving the CNN characteristics; step 7, generating new data by linear combination of the feature vectors, and calculating the probability of each category by the new data through a softmax function; and 8, selecting the value with the maximum probability as the category of the target.
Description
Technical Field
the invention belongs to the field of machine vision, and particularly relates to an AI-based double-spectrum target automatic identification method.
background
the eye mark is an emerging application direction in the field of computer vision, is one of research hotspots in the recent years, and develops rapidly. Early target recognition was mainly: object detection is performed using Scale-invariant feature transform (SIFT), and pedestrian detection is performed using Histogram of Oriented Gradients (HOG). The above methods are all to input the extracted features into the classifier for recognition, which is essentially a manually designed feature engineering, and for a certain kind of specific target to be recognized, people often do not know which kind of features are more useful for realizing the recognition task, which requires people to have professional knowledge in the field of the problem to be solved, and even if the knowledge in the professional field is used as a support, sometimes manual feature selection experiments cannot be avoided. The manual feature selection means that the data size of the recognition task cannot be too large, the models are specially designed for a certain task, and if another task needs to be completed, the feature engineering needs to be redesigned, so that the generalization performance of the models is poor. These have all inhibited further development of object recognition technology.
In recent years, deep learning has become the most popular topic in the field of computer vision. In the middle of the nineties of the twentieth century, the shortcomings of poor theoretical interpretability and overlong training time of a neural network are exposed due to the rise of statistical learning and a support vector machine, and the research of the neural network is in the low tide. However, deep learning is better at processing large-scale data than traditional Machine learning algorithms such as Support Vector Machine (SVM) and logistic regression. With the advent of the big data era, the performance of computers is rapidly refreshed, so that the training of more complicated network structures with more hidden layers becomes possible. The rapid growth of deep learning makes breakthrough in many fields such as speech recognition, image recognition, and natural language processing. The application of deep learning to target recognition successfully improves the defects that the prior target recognition needs manual feature extraction and wastes time and labor, and enables the model to have the capability of large-scale data, thereby improving the generalization performance of the model. Deep learning has made significant progress in the study of object recognition.
Disclosure of Invention
the target recognition problem is always a hot direction in the field of image processing, and the flow comprises four parts of image acquisition, preprocessing, feature extraction and recognition and classification, wherein the feature extraction is the most core part. In the field of traditional image recognition, features are basically extracted manually, time and labor are wasted, the effectiveness of the features cannot be guaranteed, and the problem is avoided by utilizing deep learning to extract the features.
the invention provides a double-spectrum target recognition method and a double-spectrum target recognition system based on AI, which are applied to visible light imaging and infrared imaging to automatically recognize targets in the visible light imaging and the infrared imaging, are different from the traditional target classification, detection and tracking, do not need to research a new algorithm according to different scenes and objects, and achieve the capability of accurate recognition after a computer system extracts, cognizes, perceives, learns and self-trains the objects in the real world.
In order to solve the technical problems, the invention adopts the following technical scheme.
The invention aims to provide a target identification method and a system based on deep learning, which can be used for automatically identifying an image target based on visible light/infrared imaging by training a classification model of a target database, additionally extracting a CNN characteristic with smaller dimension and converting the CNN characteristic into binary coding when using the CNN to extract the image characteristic, reducing the number of candidate target images through coding comparison, and searching the CNN characteristic in a reduced image range, thereby solving the problem that the existing CNN characteristic needs to obtain higher dimension for searching, so that the time required for searching is longer, and indirectly improving the identification speed.
Step 1, constructing a target database; training a target database image classification model by a deep learning network;
step 2, collecting an image to be identified through target collection equipment, wherein the image to be identified comprises: an image containing the object and an image not containing the object;
The target acquisition equipment is a visible light camera and an infrared camera;
step 3, performing CNN feature extraction on the image to be identified to obtain a CNN feature vector; CNN feature extraction enables automatic feature extraction. The CNN algorithm automatically extracts feature points, and a gradient back propagation algorithm is used: the parameters of the convolution kernel are initially initialized randomly, and then the values of the convolution kernel are adaptively adjusted through an optimization algorithm based on a gradient back propagation algorithm, thereby minimizing the error between the predicted value and the true value of the model. The convolution kernel parameters obtained in the way are not necessarily intuitive, but the features can be effectively extracted, so that the error between the model predicted value and the true value is minimized.
the input of the neuron is connected to the local receptor of the previous layer and the local features are extracted. After the feature is extracted, the positional relationship between the feature and the remaining features is determined. CNNs generally consist of convolutional layers, pooling layers, and fully-connected layers. Each layer is composed of a plurality of two-dimensional planes, each plane is composed of a plurality of independent neurons, and the result is finally output through a full-connection layer through a plurality of convolution layers and pooling layers.
Step 4, respectively extracting the local CNN features with small dimensions from the images in the target database and the images to be identified collected by the target collection equipment and converting the extracted local CNN features into binary codes;
The collected image firstly refers to the image in the database in the step 1, and local CNN characteristic vectors of the template image are obtained. And then referring to the newly acquired image in the step 2, and performing local CNN feature extraction on the newly acquired image.
The dimension is small: the local CNN feature extraction needs to be selected according to specific local features, and the feature quantity needed by different features is different. The partial image includes object colors, object composition, and images of the object at different angles. Since the present invention is a dual spectrum target automatic identification method based on AI, different targets cannot be identified by the color of the target, but the target automatic identification is achieved using a contour. The dimensions used in the present invention are 1 x 1 templates. Due to the fact that the template is too large, the calculation amount is large, real-time performance cannot be well met, and the effect is affected.
and (3) CNN feature extraction with small dimension: the CNN features with smaller dimension are connected with a convolution layer after ResNet (residual neural network), the size of the convolution kernel is 1 x 1, 256 channels are used for compressing the features to 256 dimensions, a dichotomy is adopted for each dimension, namely, each dimension of the 256 dimensions ranges from 0 to 1, and if the dimension is larger than 0.5, 1 is selected; if the dimension is less than or equal to 0.5, 0 is selected. The binary values of the 256-dimensional feature vectors of the local region are finally obtained.
step 5, reducing the number of images to be identified by using binary coding;
binary coding has the following advantages: the technology is simple to realize, and the computer is composed of a logic circuit, and the logic circuit generally has only two states, namely on and off of a switch, which can be just represented by '1' and '0'. Meanwhile, binary representation data has the advantages of strong anti-interference capability, high reliability and the like. Because each bit of data has only two states, high and low, it can still be reliably distinguished whether it is high or low when disturbed to some extent.
And solving a Hamming distance between a binary value of an image in the target database and a binary value of an image to be recognized acquired by the target acquisition equipment, wherein if the Hamming distance is less than a preset threshold value 51 (more than 80 percent of characteristic point repetition rate), the target image to be recognized and the template image are possibly of the same type. Otherwise, the image to be recognized is not the image of the known category, and if the Hamming distance value is larger than 51, the image to be recognized is deleted, so that the number of the images to be recognized is effectively reduced.
Step 6, retrieving the image to be identified to obtain a CNN characteristic vector of the image area containing the target;
The purpose of retrieval is as follows: retrieval measures the degree of similarity between two images by extracting the underlying features of the images and computationally comparing the distances between these features and the search objects.
How to retrieve: the invention searches the CNN characteristics mainly by extracting the pre-trained CNN model and the fine-tuned CNN model. The pre-trained CNN model enables coarse extraction of feature vectors. Because the image set of the target task and the pre-training image set have great difference in both the category number and the image style, in the retrieval task of the target image set to be identified, it is often difficult to achieve the optimal performance by directly using the pre-training CNN model to extract the visual features of the images. Fine adjustment is carried out on parameters of the pre-trained CNN model by using the image of the target image set, so that a good effect can be obtained.
And (3) retrieval conditions: the target image area to be identified is larger than the sub-block area image in the training process.
Step 7, generating new data by linearly combining the CNN characteristic vectors retrieved in the step 6, and calculating the probability of each category by the new data through a softmax function;
And 8, selecting the value with the maximum probability as the category of the target.
the step 1 comprises the following steps: and downloading target images from a website, wherein the images comprise images of targets with different angles, different colors and sizes and different angles, and the downloaded target images form a target database.
Step 4 comprises the following steps: respectively executing the following operations on the images in the target database and the images to be identified acquired by the target acquisition equipment: aiming at different target areas, selecting a proper local area on an image to obtain a local image, wherein the local image comprises a target color, a target composition and different imaging angles of the target, and performing CNN feature extraction on the local image, wherein the CNN feature with smaller dimension is connected with a convolution layer after ResNet (residual neural network), the size of the convolution kernel is 1 x 1, 256 channels compress the feature to 256 dimensions, a bisection method is adopted for each dimension, namely, each dimension of the 256 dimensions is 0-1, and if the dimension is greater than 0.5, 1 is selected; and if the dimension is less than or equal to 0.5, taking 0, and finally obtaining the binary value of the 256-dimensional feature vector of the local area, thereby obtaining the binary value of the image in the target database and the binary value of the image to be identified acquired by the target acquisition equipment.
The step 5 comprises the following steps: and (3) solving the Hamming distance between the binary value of the image in the target database and the binary value of the image to be recognized acquired by the target acquisition equipment, if the Hamming distance is less than a preset threshold value 51, indicating that the image to be recognized and the image in the target database are of one type, and reserving the image to be recognized for accurate recognition of later-stage images.
The step 6 comprises the following steps: and (3) retrieving an image area containing the target in the image to be identified in the step (2), wherein the retrieved image area is required to be larger than an area in the image classification model of the target database trained by the deep learning network in the step (1), and finally, a CNN (convolutional neural network) feature vector of the image area containing the target is obtained.
The step 7 comprises the following steps: and (4) combining 256 preset constants into a group, multiplying the group by the CNN eigenvector obtained in the step (6), and adding the items to obtain new data.
In step 7, the softmax function is:
Wherein p (i) represents the probability that the input feature is the ith feature,A value representing the ith characteristic of the input,Represents the value of the input kth feature, k representing the dimension of the input feature.
the invention also provides a target recognition system based on deep learning. The method comprises the following steps: imaging system, preprocessing system, information processing system and display system:
An imaging system: used for imaging the object and storing the object as a picture or a video;
A pretreatment system: the image is mainly preprocessed, such as noise reduction, contrast enhancement and brightness enhancement.
an information processing system: and the database classification module is used for correspondingly processing the input images, classifying the processed images through the database classification model and outputting a classification result.
A display system: for presenting the classification results to the user.
Has the advantages that:
(1) the AI-based dual-spectrum target automatic identification can solve the problem of long retrieval time.
(2) the invention can respectively carry out target recognition on the visible light images and can also carry out target recognition on the infrared images.
(3) The method is different from the traditional target classification, detection and tracking, does not need to research a new algorithm according to different scenes and objects, and achieves the capability of accurate identification after a computer system extracts, cognizes, perceives, learns and trains objects in the real world.
drawings
The foregoing and other advantages of the invention will become more apparent from the following detailed description of the invention when taken in conjunction with the accompanying drawings.
FIG. 1 is a diagram of steps of an AI-based dual-spectrum target automatic identification method of the present invention
Fig. 2 is a schematic structural diagram of an AI-based dual-spectrum target recognition system according to the present invention.
FIG. 3 is a block diagram of an object imaging and recognition system of the present invention.
Detailed Description
The invention is further explained below with reference to the drawings and the embodiments.
as shown in fig. 1, an AI-based dual-spectrum target automatic identification method of the present invention includes the following steps:
1. constructing a target database, and downloading 10000 target images to be identified from corresponding websites, wherein the images comprise images of targets with different angles, different colors and sizes and different angles;
2. Training a target database image classification model by a deep learning network;
3. The target acquisition equipment acquires an image containing/not containing 1000 targets to be identified;
4. performing CNN feature extraction on an image to be identified, wherein the feature value is a 512-dimensional vector;
5. extracting local CNN features with small dimensions from the acquired images and converting the local CNN features into binary codes convenient for comparison, wherein the local CNN feature vectors are 256-dimensional, and the local binary codes of 1000 acquired images are obtained according to a preset conversion rule;
6. the CNN features and binary codes of the local images are extracted, the number of target pictures is reduced by using the codes, the Hamming distance between the binary value of the target image to be identified and the binary value of the template image is calculated, and if the Hamming distance is smaller than a preset threshold 51 (more than 80 percent of feature point repetition rate), the target image and the template image are possibly classified. Otherwise, the target image is not the known class image, and the target image is deleted. This effectively reduces the number of target images. Finally, 600 identification images are obtained.
7. The CNN features are retrieved within a reduced scope.
8. And generating new data by linear combination on the finally obtained characteristic vectors of the 600 identification images, and calculating the probability of each category through the softmax function by the new data.
The Softmax function can be expressed as:
9. And selecting the value with the maximum probability as the category of the target.
10. And the feature extraction uses a ResNet network structure to obtain a multidimensional feature vector of the picture feature.
The partial image adopted by the CNN feature extraction comprises target color, target composition and imaging of the target with different angles, wherein the CNN feature with smaller dimension is formed by connecting a convolution layer after ResNet, the convolution kernel size is 1 x 1, and 256 channels compress the feature to 256 dimensions. Adopting a dichotomy for each dimension, namely, each dimension of 256 dimensions ranges from 0 to 1, and if the dimension is greater than 0.5, taking 1; if the dimension is less than or equal to 0.5, 0 is selected.
As shown in fig. 2 and 3, the present invention also provides an AI-based target recognition system, including: imaging system, preprocessing system, information processing system and display system:
an imaging system: used for imaging the object and storing the object as a picture or a video;
A pretreatment system: the image is mainly preprocessed, such as noise reduction, contrast enhancement and brightness enhancement.
An information processing system: and the database classification module is used for correspondingly processing the input images, classifying the processed images through the database classification model and outputting a classification result.
(1) The system mainly comprises a CNN feature extraction module, a local feature extraction module, a binary conversion module, a target database, a model training module and a comparison and retrieval module.
(2) The CNN feature extraction module is used for performing CNN feature extraction on the image to be retrieved; the local feature extraction module is used for extracting features of local images needing to be retrieved; the binary conversion module is used for carrying out binary conversion on the local image with smaller dimension; the target database is used for providing a large number of target picture samples for the training model; the model training module is used for training a target database picture classification model through a deep neural network; the comparison module is used for comparing the binary code with a smaller dimension with the training model; the retrieval module is used for retrieving the characteristic picture to be identified from the rest target pictures after comparison.
A display system: for presenting the classification results to the user.
The present invention provides a dual-spectrum target automatic identification method based on AI, and a plurality of methods and approaches for implementing the technical solution, and the above description is only a preferred embodiment of the present invention, it should be noted that, for those skilled in the art, a plurality of improvements and modifications can be made without departing from the principle of the present invention, and these improvements and modifications should also be regarded as the protection scope of the present invention. All the components not specified in the present embodiment can be realized by the prior art.
Claims (7)
1. An AI-based dual-spectrum target automatic identification method is characterized by comprising the following steps:
Step 1, constructing a target database; training a target database image classification model by a deep learning network;
Step 2, collecting an image to be identified through target collection equipment, wherein the image to be identified comprises: an image containing the object and an image not containing the object;
step 3, performing CNN feature extraction on the image to be identified to obtain a CNN feature vector;
step 4, respectively extracting the local CNN features with small dimensions from the images in the target database and the images to be identified collected by the target collection equipment and converting the extracted local CNN features into binary codes;
step 5, reducing the number of images to be identified by using binary coding;
step 6, retrieving the image to be identified to obtain a CNN characteristic vector of the image area containing the target;
step 7, generating new data by linearly combining the CNN characteristic vectors retrieved in the step 6, and calculating the probability of each category by the new data through a softmax function;
And 8, selecting the value with the maximum probability as the category of the target.
2. the method of claim 1, wherein step 1 comprises: and downloading target images from a website, wherein the images comprise images of targets with different angles, different colors and sizes and different angles, and the downloaded target images form a target database.
3. The method of claim 2, wherein step 4 comprises: respectively executing the following operations on the images in the target database and the images to be identified acquired by the target acquisition equipment: selecting a proper local area on an image to obtain a local image aiming at different target areas, wherein the local image comprises a target color, a target composition and different imaging angles of the target, and performing CNN feature extraction on the local image, the CNN feature with smaller dimension is connected with a convolution layer after ResNet, the size of the convolution kernel is 1 x 1, 256 channels are used for compressing the feature to 256 dimensions, a dichotomy is adopted for each dimension, namely, each dimension of the 256 dimensions ranges from 0 to 1, and if the dimension is greater than 0.5, 1 is selected; and if the dimension is less than or equal to 0.5, taking 0, and finally obtaining the binary value of the 256-dimensional feature vector of the local area, thereby obtaining the binary value of the image in the target database and the binary value of the image to be identified acquired by the target acquisition equipment.
4. the method of claim 3, wherein step 5 comprises: and (3) solving the Hamming distance between the binary value of the image in the target database and the binary value of the image to be recognized acquired by the target acquisition equipment, if the Hamming distance is less than a preset threshold value 51, indicating that the image to be recognized and the image in the target database are of one type, and reserving the image to be recognized.
5. the method of claim 4, wherein step 6 comprises: and (3) retrieving an image area containing the target in the image to be identified in the step (2), wherein the retrieved image area is required to be larger than an area in the image classification model of the target database trained by the deep learning network in the step (1), and finally, a CNN (convolutional neural network) feature vector of the image area containing the target is obtained.
6. The method of claim 5, wherein step 7 comprises: and (3) combining 256 preset constants into a group, multiplying the group by the CNN characteristic vector obtained in the step (6), adding each product result to obtain new data, and calculating the probability of each category through the softmax function by the new data.
7. The method according to claim 6, characterized in that in step 7, the softmax function is:
Wherein p (i) represents the probability that the input feature is the ith feature,A value representing the ith characteristic of the input,Represents the value of the input kth feature, k representing the dimension of the input feature.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910780121.5A CN110580503A (en) | 2019-08-22 | 2019-08-22 | AI-based double-spectrum target automatic identification method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910780121.5A CN110580503A (en) | 2019-08-22 | 2019-08-22 | AI-based double-spectrum target automatic identification method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN110580503A true CN110580503A (en) | 2019-12-17 |
Family
ID=68811774
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910780121.5A Pending CN110580503A (en) | 2019-08-22 | 2019-08-22 | AI-based double-spectrum target automatic identification method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110580503A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112348016A (en) * | 2020-11-09 | 2021-02-09 | 青岛大东电子有限公司 | Smart picture LOGO identification method |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106682233A (en) * | 2017-01-16 | 2017-05-17 | 华侨大学 | Method for Hash image retrieval based on deep learning and local feature fusion |
CN106909924A (en) * | 2017-02-18 | 2017-06-30 | 北京工业大学 | A kind of remote sensing image method for quickly retrieving based on depth conspicuousness |
CN107589758A (en) * | 2017-08-30 | 2018-01-16 | 武汉大学 | A kind of intelligent field unmanned plane rescue method and system based on double source video analysis |
CN107609601A (en) * | 2017-09-28 | 2018-01-19 | 北京计算机技术及应用研究所 | A kind of ship seakeeping method based on multilayer convolutional neural networks |
CN108108657A (en) * | 2017-11-16 | 2018-06-01 | 浙江工业大学 | A kind of amendment local sensitivity Hash vehicle retrieval method based on multitask deep learning |
US20180210896A1 (en) * | 2015-07-22 | 2018-07-26 | Hangzhou Hikvision Digital Technology Co., Ltd. | Method and device for searching a target in an image |
CN108537120A (en) * | 2018-03-06 | 2018-09-14 | 安徽电科恒钛智能科技有限公司 | A kind of face identification method and system based on deep learning |
CN109308483A (en) * | 2018-07-11 | 2019-02-05 | 南京航空航天大学 | Double source image characteristics extraction and fusion identification method based on convolutional neural networks |
-
2019
- 2019-08-22 CN CN201910780121.5A patent/CN110580503A/en active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180210896A1 (en) * | 2015-07-22 | 2018-07-26 | Hangzhou Hikvision Digital Technology Co., Ltd. | Method and device for searching a target in an image |
CN106682233A (en) * | 2017-01-16 | 2017-05-17 | 华侨大学 | Method for Hash image retrieval based on deep learning and local feature fusion |
CN106909924A (en) * | 2017-02-18 | 2017-06-30 | 北京工业大学 | A kind of remote sensing image method for quickly retrieving based on depth conspicuousness |
CN107589758A (en) * | 2017-08-30 | 2018-01-16 | 武汉大学 | A kind of intelligent field unmanned plane rescue method and system based on double source video analysis |
CN107609601A (en) * | 2017-09-28 | 2018-01-19 | 北京计算机技术及应用研究所 | A kind of ship seakeeping method based on multilayer convolutional neural networks |
CN108108657A (en) * | 2017-11-16 | 2018-06-01 | 浙江工业大学 | A kind of amendment local sensitivity Hash vehicle retrieval method based on multitask deep learning |
CN108537120A (en) * | 2018-03-06 | 2018-09-14 | 安徽电科恒钛智能科技有限公司 | A kind of face identification method and system based on deep learning |
CN109308483A (en) * | 2018-07-11 | 2019-02-05 | 南京航空航天大学 | Double source image characteristics extraction and fusion identification method based on convolutional neural networks |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112348016A (en) * | 2020-11-09 | 2021-02-09 | 青岛大东电子有限公司 | Smart picture LOGO identification method |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107256246B (en) | printed fabric image retrieval method based on convolutional neural network | |
Lin et al. | RSCM: Region selection and concurrency model for multi-class weather recognition | |
CN111177446B (en) | Method for searching footprint image | |
CN108595636A (en) | The image search method of cartographical sketching based on depth cross-module state correlation study | |
CN108829677A (en) | A kind of image header automatic generation method based on multi-modal attention | |
CN104778242A (en) | Hand-drawn sketch image retrieval method and system on basis of image dynamic partitioning | |
Wang et al. | Multiscale deep alternative neural network for large-scale video classification | |
Zhang et al. | Automatic discrimination of text and non-text natural images | |
Zhang et al. | A small target detection method based on deep learning with considerate feature and effectively expanded sample size | |
Yu et al. | Pedestrian detection based on improved Faster RCNN algorithm | |
Lin et al. | Object detection algorithm based AdaBoost residual correction Fast R-CNN on network | |
CN115661754A (en) | Pedestrian re-identification method based on dimension fusion attention | |
CN111782852A (en) | High-level semantic image retrieval method based on deep learning | |
CN113192108B (en) | Man-in-loop training method and related device for vision tracking model | |
CN117994623A (en) | Image feature vector acquisition method | |
CN110472632A (en) | Character segmentation method, device and computer storage medium based on character feature | |
CN110580503A (en) | AI-based double-spectrum target automatic identification method | |
CN110674342B (en) | Method and device for inquiring target image | |
Özyurt et al. | A new method for classification of images using convolutional neural network based on Dwt-Svd perceptual hash function | |
CN116244464A (en) | Hand-drawing image real-time retrieval method based on multi-mode data fusion | |
CN110162654A (en) | It is a kind of that image retrieval algorithm is surveyed based on fusion feature and showing for search result optimization | |
CN115063831A (en) | High-performance pedestrian retrieval and re-identification method and device | |
CN112488116A (en) | Scene understanding semantic generation method based on multi-mode embedding | |
Yu et al. | Automatic image captioning system using integration of N-cut and color-based segmentation method | |
Yue et al. | Study on the deep neural network of intelligent image detection and the improvement of elastic momentum on image recognition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20191217 |