CN112597329B - Real-time image retrieval method based on improved semantic segmentation network - Google Patents
Real-time image retrieval method based on improved semantic segmentation network Download PDFInfo
- Publication number
- CN112597329B CN112597329B CN202011523748.1A CN202011523748A CN112597329B CN 112597329 B CN112597329 B CN 112597329B CN 202011523748 A CN202011523748 A CN 202011523748A CN 112597329 B CN112597329 B CN 112597329B
- Authority
- CN
- China
- Prior art keywords
- semantic
- image
- network
- feature
- pictures
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/583—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Library & Information Science (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Databases & Information Systems (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a real-time image retrieval method based on an improved semantic segmentation network, which comprises the following steps: s1, collecting retrieval images to form a data set to be trained, expanding the data set, and dividing the expanded data set into a training set, a testing set and a verification set; s2, constructing an improved real-time calculation semantic segmentation network; s3, inputting the pictures of the training set into an improved semantic segmentation network calculated in real time to obtain semantic vectors of the pictures, and storing the semantic vectors according to the category information of the pictures to obtain a semantic vector database for image retrieval; s4, sending the pictures to be detected in the test set into an improved semantic segmentation network calculated in real time to obtain semantic feature vectors, and comparing the semantic categories of the pictures to be searched with the images of the vectors in an image search semantic vector database to obtain a candidate picture group containing the semantic categories of the searched pictures; and S5, matching the semantic feature vector of the picture to be detected with the semantic feature vector of the candidate picture group.
Description
Technical Field
The invention relates to the technical field of image processing, in particular to a real-time image retrieval method based on an improved semantic segmentation network.
Background
The technology of finding out the image desired by the user from the database based on the basic visual features of the image is called content-based image retrieval (CBIR). Typically, the user's needs are given in the manner of querying the image. Briefly, it needs to analyze the content of the image to be queried by the user, extract the appropriate retrieval features, and retrieve the pictures in the image database that conform to the features to obtain the result. CBIR has been widely used in hundreds of degrees as an engineering requirement in daily life, and is also embedded in various e-commerce platforms such as kyoto and naobao for searching websites such as Google. Conventional image retrieval methods use a manually designed feature extraction operator to obtain low-level visual features of an image. Low-level visual features include color, edges, texture, location, and geometry, among others. Common extraction methods include Gabor filters, SIFT operators, SURF operators, and the like. However, the manually designed feature operators do not have generalization capability, and only have good extraction capability for a certain category and weak extraction capability for other categories. Image retrieval requires processing of hundreds of categories of images and there are great limitations to using low-level visual feature extraction operators (raney coma. Convolutional neural network-based image retrieval several key technical studies [ D ]. University of chejiang physics, 2020.).
In recent years, the development of deep learning, compared with the traditional image feature extraction method, the image processing technology based on deep learning has better effect. Semantic segmentation is a task of computer vision, and a good effect is achieved by using a deep learning technology to perform semantic segmentation. The semantic segmentation method applied to the image retrieval task at present still has the problems of low retrieval efficiency, low accuracy, incapability of providing semantic understanding and the like. Therefore, the method still has a larger promotion space on occasions with higher real-time performance, such as the closed-loop image detection problem of unmanned vehicles and the large-scale multi-class image retrieval problem (Jiang Si Yao. Semantic segmentation algorithm [ D ] based on a full convolution neural network model, the university of Liaoning engineering technology, 2020.).
Disclosure of Invention
The invention aims to solve the problems of precision ratio and rapidity in an image retrieval task, and provides a real-time image retrieval method based on an improved semantic segmentation network. The method has better effect when being applied to retrieval scenes with higher real-time requirements.
The invention is realized by at least one of the following technical schemes.
A real-time image retrieval method based on an improved semantic segmentation network comprises the following steps:
s1, collecting retrieval images to form a data set to be trained, expanding the data set, and dividing the expanded data set into a training set, a testing set and a verification set;
s2, constructing an improved real-time calculation semantic segmentation network;
s3, inputting the pictures of the training set into an improved semantic segmentation network calculated in real time to obtain semantic vectors of the pictures, and storing the semantic vectors according to the category information of the pictures to obtain a semantic vector database for image retrieval;
s4, sending the pictures to be detected in the test set into an improved semantic segmentation network calculated in real time to obtain semantic feature vectors, and comparing whether the semantic categories of the pictures to be retrieved are consistent with the category information of the images of the vectors in the image retrieval semantic vector database to obtain a candidate picture group containing the semantic categories with the retrieved pictures;
and S5, matching the semantic feature vector of the picture to be detected with the semantic feature vector of the candidate picture group to obtain a plurality of pictures closest to the unit value, namely the retrieval result.
Preferably, step S1 is to select a data set PASCAL (Statistical modeling and Computational Learning) VOC (Visual Object Classes) 2012 and a berkeley data set and construct an extended data set.
Preferably, the PSCAL VOC2012 data sets have twenty-one category, wherein the number of the images in the training set, the verification set and the test set is M, H and R, respectively, and the original data set is expanded by using the berkeley data set to finally obtain the training image, the verification image and the test image.
Preferably, the improved real-time computing semantic segmentation network adopts an image semantic segmentation and real-time retrieval deep learning network framework; the network framework adopts an encoder-decoder structure, the calculation result of the encoder is used for extracting the characteristic information vector of the object, and the decoder is used for extracting the category information of the object.
Preferably, the coding network of the encoder adopts a lightweight network MobileNet (Efficient Convolutional Neural network suitable for Mobile visual application) Efficient Convolutional Neural Networks for Mobile visual application as a main network, and a hole convolution is added to a convolution layer of an original network to maintain the resolution of a feature map, and simultaneously an Spatial Pyramid Pooling module (ASPP) with adjacent information is selected, and a global average Pooling module and a Pooling layer of a are added to an ASPP network structure to extract feature information with higher significance, and hole convolution with a scale of a is added to extract associated information of pixels.
Preferably, the decoding network of the decoder adopts a two-way Fusion mode, the decoding network uses a semantic Feature Fusion Module SFFM (semantic Feature Fusion Module), the SFFM combines the high-level image features sampled by the ASPP network structure and the bottom-level information features of the block3 grid in the backbone network, and the two-way input information is used for complementing each other to further reduce the number of parameters.
Preferably, the object feature information vector is extracted in the following two ways:
the method I comprises the following steps: adding a full connection layer after the convolution layer of the original encoder network to obtain a feature vector with fixed length, wherein x is as follows 1 To x n Taking the value of the corresponding position of the feature map, w 11 …w nn For full link layer parameters, a 1 To a n To be characterized byGraph derived feature vectors:
the second method comprises the following steps: constructing and constructing a feature vector for adding a global average pooling module, introducing a global pooling function to fuse high-dimensional features at each pixel point in a feature map, using an empirical weighted average method, and making weights the same to obtain an average value, wherein extracted features X belong to the following feature spaces, h and w are the height and width of the feature map respectively, and c is the number of feature channels:
X∈R h×w×c (2)
directly adding the feature vectors of all the spatial positions according to the feature dimension, and then normalizing by the total number of the spatial positions:
wherein X i Representing the feature vector, ω, at position i i The representation corresponds to X i The normalized image feature vector containing the object is obtained.
Preferably, all pictures in the training set are processed in the steps to construct an image retrieval database, and the pictures to be detected in the test set are sent to a semantic segmentation network to obtain semantic feature vectors of the pictures through semantic segmentation.
Preferably, the matching manner of step S5 is as follows:
wherein, X represents the semantic feature vector of the picture to be detected, and Y represents the semantic feature vector of the candidate picture group.
The system of the real-time image retrieval method based on the improved semantic segmentation network comprises an image acquisition system, a deep neural network system and an image processing system;
the image acquisition system is used for collecting and searching images to form a data set needing training and expanding the data set;
the deep neural network system comprises an improved semantic segmentation network calculated in real time, and is used for extracting characteristic information vectors of objects and extracting class information of the objects;
the image processing system is used for combining the object characteristic information vector and the object category information to obtain a semantic retrieval vector for image retrieval and perform similarity calculation and matching of images.
Compared with the prior art, the invention has the beneficial effects that: aiming at the defect that the existing general semantic segmentation network is low in computing speed, the improved semantic segmentation network capable of computing in real time is provided to improve the efficiency of the image retrieval process. Meanwhile, semantic category information is added in the process of retrieving the images, so that the information is richer in retrieving the images, and the accuracy of the retrieving process under complex conditions is ensured.
Drawings
FIG. 1 is an overall flow chart of the present invention;
FIG. 2 is a diagram of a real-time semantic segmentation network architecture of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, but not all embodiments of the present invention. All other embodiments, which can be obtained by a person skilled in the art without making any creative effort based on the embodiments in the present invention, belong to the protection scope of the present invention.
As shown in fig. 1 and fig. 2, a real-time image retrieval method based on an improved semantic segmentation network specifically includes the following steps:
(1) A data set is acquired. The PASCAL VOC2012 data set and the berkeley data set are downloaded and, because the data categories of the two data sets are the same, the two data sets are combined to construct an extended data set for the image retrieval task. There are twenty-one classes in PSCAL VOC 2012. The number of images in the training set, the verification set and the test set is 1464, 1449 and 1456 respectively. The expansion using the berkeley dataset finally yielded 10582 training images, 1229 validation images and 1456 test images. The expanded data set is obtained according to the proportion of a training set, a test set and a verification set as 6:2:2, grouping;
(2) Constructing an improved real-time computing semantic segmentation network; the conventional semantic segmentation network deep lab v3 (encoder-decoder network with separable convolutions) was modified to make it suitable for real-time computation. The improved encoder network uses MobileNet v2 (Efficient Neural Networks for Mobile Vision Applications), adds a feature extraction module to an encoding layer to extract features, adds a semantic feature fusion module to a decoding layer to perform category calculation, and trains images in a data set through the improved network;
the improved real-time computing semantic segmentation network adopts an image semantic segmentation and real-time retrieval deep learning network framework. The network framework adopts an encoder-decoder structure, the calculation result of the encoder is used for extracting the characteristic information vector of the object, and the decoder is used for extracting the class information of the object.
The coding network of the coder adopts a lightweight network MobileNet v2 as a backbone network, a cavity convolution is added in a convolution layer of an original network to keep the resolution ratio of a feature map, meanwhile, a space Pyramid Pooling module (ASPP) with adjacent information is selected, the ASPP network structure obtains context information of an image by aggregating sense field information with different scales and extracting high-level semantic information, a global average Pooling module and a 3 × 3 Pooling layer are added in the ASPP network structure to extract feature information with higher significance, the cavity convolution with the scale of 3 × 3 is added to extract associated information of pixels, and a coding result is used for extracting object feature information.
The decoding network of the decoder adopts a two-way Fusion mode, the decoding network uses a semantic Feature Fusion Module SFFM (semantic Feature Fusion Module), the SFFM combines the high-level image features sampled by an ASPP network structure and the bottom-level information features of a block3 grid in a backbone network, the two ways of input information are mutually supplemented to further reduce the number of parameters, and the decoding result is used for image category extraction.
After a feature map of a convolutional layer with optimal expression features is extracted through an improved encoder of a semantic segmentation network calculated in real time, feature vectors of the image are generated through methods such as global average pooling. The encoder processing result can obtain the category information contained in the image through a decoding network. And adding the class information of the object in the image to the tail part of the normalized feature vector to construct an image semantic feature vector.
The feature vector of the image is obtained by encoding the convolution layer of the network in the following two ways:
the method I comprises the following steps: after the convolution layer of the original encoder network, a full connection layer is added to obtain a feature vector with a fixed length. The formula is shown below, where x 1 To x n Taking the value of the corresponding position of the feature map, w 11 …w nn For full link layer parameters, a 1 To a n For the feature vectors derived from the feature map:
the second method comprises the following steps: a feature vector is constructed for the add global average pooling module construction. And introducing a global pooling function to fuse the high-dimensional features at each pixel point in the feature map, using an empirical weighted average method, and obtaining an average value by keeping the weights the same. The extracted feature X belongs to a feature space where h and w are the height and width of the feature map, respectively, and c is the number of feature channels.
X∈R h×w×c (2)
The feature vectors of each space position are directly added according to feature dimensions and then normalized by the total number of the space positions to obtain the following expression
Wherein X i Representing the feature vector at position i. Omega i The representation corresponds to X i The empirical weights of (1). And setting the experience weights to be 1 to obtain the normalized image feature vector containing the object.
And finally obtaining a segmentation result containing the object semantic category through a decoding network.
And combining the feature vector obtained by the coding layer of the improved real-time semantic segmentation network with the category information vector contained in the image obtained by the decoding network. And adding the category information of the object in the image to the tail part of the normalized feature vector to construct a picture semantic feature vector.
S3, inputting the pictures of the training set into an improved semantic segmentation network calculated in real time to obtain semantic vectors of the pictures, and storing the semantic vectors according to the category information of the pictures to obtain an image retrieval semantic vector database;
s4, sending the pictures to be detected in the test set into an improved semantic segmentation network calculated in real time to obtain semantic feature vectors, and comparing whether the semantic categories of the pictures to be searched are consistent with the category information of the images of the vectors in the image search semantic vector database to obtain a candidate picture group containing the semantic categories of the pictures to be searched;
and S5, calculating the cosine of the semantic feature vector of the picture to be detected and the semantic feature vector of the candidate picture group to obtain a plurality of pictures which are closest to the unit value, namely the retrieval result.
The matching method is that a cosine calculation method is used to calculate matching values, and K vectors closest to a unit value in a result are extracted as a result value of image retrieval. The value of K can be set by the searcher as desired.
Wherein X represents the semantic feature vector of the picture to be detected, and Y represents the semantic feature vector of the candidate picture group.
The embodiment provides a real-time image retrieval method based on an improved semantic segmentation network, which comprises the steps of improving the real-time performance through the improved semantic segmentation network, expanding a PASCAL VOC data set by using a Berkeley data set to perform semantic segmentation training, extracting category vectors after semantic segmentation results are obtained, performing full-connection or global average pooling according to a feature map obtained by an encoding layer to obtain feature vectors, and finally combining the feature vectors to obtain the semantic vectors of a picture to be retrieved. During retrieval, firstly, rough matching is carried out to obtain groups of pictures with consistent categories, and then K pictures meeting requirements are obtained through cosine calculation and serve as results. According to the technical scheme, the speed of semantic segmentation is effectively improved on the premise of not losing precision, and a better picture retrieval effect is obtained. The method has good adaptability under the conditions of large-scale image retrieval and more picture categories.
The invention also provides an image retrieval system based on deep learning and semantic segmentation, which comprises the following steps: the system comprises an image acquisition system, a deep neural network system and an image processing system; the image acquisition system is used for collecting and searching images to form a data set needing training and expanding the data set; the deep neural network system comprises an improved semantic segmentation network calculated in real time, and is used for extracting characteristic information vectors of objects and extracting class information of the objects; the image processing system is used for combining the object characteristic information vector and the object category information to obtain a semantic retrieval vector for image retrieval and perform similarity calculation and matching of the images.
The image retrieval system based on deep learning and semantic segmentation stores a computer program therein. After program initialization, the network parameters are trained by the semantic segmentation data set. The model parameters are greatly reduced due to the use of an improved semantic segmentation network. When a retrieval image is input, the system starts two thread running programs, wherein one thread inputs image data to perform semantic segmentation calculation, and after a calculation result is obtained, a classification result is transmitted to a second thread. The new thread retrieves the same category from the database as the input image resulting in a category grouping. And then carrying out vector similarity calculation on the original image and the images in the category grouping. And obtaining K images with the similarity sequence as input results.
The foregoing is only a preferred embodiment of the present invention, and it should be noted that, for those skilled in the art, various modifications and amendments can be made without departing from the principle of the present invention, and these modifications and amendments should also be considered as the protection scope of the present invention.
Claims (6)
1. A real-time image retrieval method based on an improved semantic segmentation network is characterized by comprising the following steps:
s1, collecting retrieval images to form a data set to be trained, expanding the data set, and dividing the expanded data set into a training set, a testing set and a verification set;
s2, constructing an improved real-time computing semantic segmentation network; extracting a feature map of a convolutional layer with optimal expression features through an improved encoder of a semantic segmentation network calculated in real time, and generating a feature vector of the image through a global average pooling method; performing full connection or global average pooling on the processing result of the encoder to obtain category information contained in the image; merging the feature vector with a category information vector contained in the image obtained by the encoder, and adding the category information of the object in the image to the tail part of the normalized feature vector to construct an image semantic feature vector;
the improved real-time computing semantic segmentation network adopts an image semantic segmentation and real-time retrieval deep learning network framework; the network framework adopts an encoder-decoder structure, the calculation result of the encoder is used for extracting the characteristic information vector of the object, and the decoder is used for extracting the class information of the object;
the coding network of the encoder adopts a lightweight network MobileNet as a backbone network, and adds a hole convolution to the convolution layer of the original network to keep the resolution of a characteristic diagram, selects a spatial pyramid pooling module ASPP with adjacent information at the same time, adds a global average pooling module and a pooling layer of a to extract characteristic information with higher significance in an ASPP network structure, and adds a hole convolution with the scale of a to extract associated information of pixels;
the decoding network of the decoder adopts a two-way fusion mode, the decoding network uses a semantic feature fusion module SFFM, the SFFM combines the high-level image features sampled by an ASPP network structure and the bottom-level information features of block3 grids in a backbone network, and the two paths of input information are used for supplementing each other to further reduce the parameter number; the object feature information vector is extracted in the following two ways:
the first method is as follows: adding a full-link layer after the convolutional layer of the original encoder network to obtain a feature vector with a fixed length, wherein x is as follows 1 To x n Taking the value of the corresponding position of the feature map, w 11 …w nn For full link layer parameters, a 1 To a n For the feature vectors derived from the feature map:
the second method comprises the following steps: constructing a feature vector for adding a global average pooling module, introducing a global pooling function to fuse high-dimensional features at each pixel point in a feature map, using an empirical weighted average method, and making weights the same to obtain an average value, wherein extracted features X belong to the following feature space, h and w are the height and width of the feature map respectively, and c is the number of feature channels:
X∈R h×w×c (2)
directly adding the feature vectors of all the spatial positions according to the feature dimension, and then normalizing by the total number of the spatial positions:
wherein X i Representing the feature vector, ω, at position i i The representation corresponds to X i Obtaining the normalized image feature vector containing the object;
s3, inputting the pictures of the training set into an improved semantic segmentation network calculated in real time to obtain semantic vectors of the pictures, and storing the semantic vectors according to the category information of the pictures to obtain a semantic vector database for image retrieval;
s4, sending the pictures to be detected in the test set into an improved semantic segmentation network calculated in real time to obtain semantic feature vectors, and comparing whether the semantic categories of the pictures to be searched are consistent with the category information of the images of the vectors in the image search semantic vector database to obtain a candidate picture group containing the semantic categories of the pictures to be searched;
and S5, matching the semantic feature vector of the picture to be detected with the semantic feature vector of the candidate picture group to obtain a plurality of pictures closest to the unit value, namely the retrieval result.
2. The method for real-time image retrieval based on improved semantic segmentation network as claimed in claim 1, wherein step S1 is to select the PSCAL VOC2012 data set and the berkeley data set and construct the augmented data set.
3. The method as claimed in claim 2, wherein the PSCAL VOC2012 data sets have twenty-one categories, wherein the numbers of the images in the training set, the verification set, and the test set are M, H, and R, respectively, and the berkeley data set is used to expand the original data set to obtain the training image, the verification image, and the test image.
4. The real-time image retrieval method based on the improved semantic segmentation network as claimed in claim 3, wherein the processing of step S2 is performed on all the pictures in the training set to construct an image retrieval database, and the pictures to be detected in the test set are sent to the semantic segmentation network to obtain semantic feature vectors of the pictures through semantic segmentation.
5. The method for retrieving an image in real time based on an improved semantic segmentation network according to claim 4, wherein the matching manner in step S5 is as follows:
wherein, x represents the semantic feature vector of the picture to be detected, and Y represents the semantic feature vector of the candidate picture group.
6. The system of the real-time image retrieval method based on the improved semantic segmentation network is characterized by comprising an image acquisition system, a deep neural network system and an image processing system;
the image acquisition system is used for collecting retrieval images as image information input of image retrieval;
the deep neural network system comprises an improved semantic segmentation network calculated in real time, and is used for extracting characteristic information vectors of objects and extracting class information of the objects;
the image processing system is used for combining the object characteristic information vector and the object category information to obtain a semantic retrieval vector for image retrieval, and performing similarity calculation and matching of images.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011523748.1A CN112597329B (en) | 2020-12-21 | 2020-12-21 | Real-time image retrieval method based on improved semantic segmentation network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011523748.1A CN112597329B (en) | 2020-12-21 | 2020-12-21 | Real-time image retrieval method based on improved semantic segmentation network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112597329A CN112597329A (en) | 2021-04-02 |
CN112597329B true CN112597329B (en) | 2022-12-16 |
Family
ID=75199836
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011523748.1A Active CN112597329B (en) | 2020-12-21 | 2020-12-21 | Real-time image retrieval method based on improved semantic segmentation network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112597329B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114494696B (en) * | 2022-01-26 | 2024-07-16 | 安徽理工大学 | Method, system and device for rapidly detecting multi-scale coal gangue images |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110909642A (en) * | 2019-11-13 | 2020-03-24 | 南京理工大学 | Remote sensing image target detection method based on multi-scale semantic feature fusion |
CN111209808A (en) * | 2019-12-25 | 2020-05-29 | 北京航空航天大学杭州创新研究院 | Unmanned aerial vehicle image semantic segmentation and identification method based on hierarchical processing |
CN111489357A (en) * | 2019-01-29 | 2020-08-04 | 广州市百果园信息技术有限公司 | Image segmentation method, device, equipment and storage medium |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10635927B2 (en) * | 2017-03-06 | 2020-04-28 | Honda Motor Co., Ltd. | Systems for performing semantic segmentation and methods thereof |
-
2020
- 2020-12-21 CN CN202011523748.1A patent/CN112597329B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111489357A (en) * | 2019-01-29 | 2020-08-04 | 广州市百果园信息技术有限公司 | Image segmentation method, device, equipment and storage medium |
CN110909642A (en) * | 2019-11-13 | 2020-03-24 | 南京理工大学 | Remote sensing image target detection method based on multi-scale semantic feature fusion |
CN111209808A (en) * | 2019-12-25 | 2020-05-29 | 北京航空航天大学杭州创新研究院 | Unmanned aerial vehicle image semantic segmentation and identification method based on hierarchical processing |
Non-Patent Citations (3)
Title |
---|
A Semantic Segmentation Algorithm Based on Improved Attention Mechanism;Chen CY 等;《INTERNATIONAL SYMPOSIUM ON AUTONOMOUS SYSTEMS(ISAS)》;20201206;244-248 * |
结合注意力机制的双路径语义分割;翟鹏博等;《中国图像图形学报》;20200831;第25卷(第8期);1627-1636 * |
语义分割及其在图像检索中的应用;苏雯;《中国优秀博硕士学位论文全文数据库(博士)信息科技辑(月刊)》;20190615(第6期);I138-46 * |
Also Published As
Publication number | Publication date |
---|---|
CN112597329A (en) | 2021-04-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110866140B (en) | Image feature extraction model training method, image searching method and computer equipment | |
CN101587478B (en) | Methods and devices for training, automatically labeling and searching images | |
CN109658445A (en) | Network training method, increment build drawing method, localization method, device and equipment | |
CN115171165A (en) | Pedestrian re-identification method and device with global features and step-type local features fused | |
CN108875076B (en) | Rapid trademark image retrieval method based on Attention mechanism and convolutional neural network | |
CN104794219A (en) | Scene retrieval method based on geographical position information | |
CN110929080B (en) | Optical remote sensing image retrieval method based on attention and generation countermeasure network | |
CN112347284B (en) | Combined trademark image retrieval method | |
CN111339343A (en) | Image retrieval method, device, storage medium and equipment | |
Qian et al. | Image location inference by multisaliency enhancement | |
Pang et al. | Deep feature aggregation and image re-ranking with heat diffusion for image retrieval | |
CN107291825A (en) | With the search method and system of money commodity in a kind of video | |
CN108984642A (en) | A kind of PRINTED FABRIC image search method based on Hash coding | |
CN114694185B (en) | Cross-modal target re-identification method, device, equipment and medium | |
CN113066089B (en) | Real-time image semantic segmentation method based on attention guide mechanism | |
CN109919084B (en) | Pedestrian re-identification method based on depth multi-index hash | |
CN106355210B (en) | Insulator Infrared Image feature representation method based on depth neuron response modes | |
CN113032613A (en) | Three-dimensional model retrieval method based on interactive attention convolution neural network | |
CN115545166A (en) | Improved ConvNeXt convolutional neural network and remote sensing image classification method thereof | |
CN113988147A (en) | Multi-label classification method and device for remote sensing image scene based on graph network, and multi-label retrieval method and device | |
CN116091946A (en) | Yolov 5-based unmanned aerial vehicle aerial image target detection method | |
CN112597329B (en) | Real-time image retrieval method based on improved semantic segmentation network | |
CN110110120B (en) | Image retrieval method and device based on deep learning | |
CN117994623A (en) | Image feature vector acquisition method | |
CN114168773A (en) | Semi-supervised sketch image retrieval method based on pseudo label and reordering |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |