CN112528905B - Image processing method, device and computer storage medium - Google Patents
Image processing method, device and computer storage medium Download PDFInfo
- Publication number
- CN112528905B CN112528905B CN202011505344.XA CN202011505344A CN112528905B CN 112528905 B CN112528905 B CN 112528905B CN 202011505344 A CN202011505344 A CN 202011505344A CN 112528905 B CN112528905 B CN 112528905B
- Authority
- CN
- China
- Prior art keywords
- image
- vector
- queried
- visual feature
- visual
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000003672 processing method Methods 0.000 title claims abstract description 16
- 239000013598 vector Substances 0.000 claims abstract description 71
- 230000000007 visual effect Effects 0.000 claims abstract description 62
- 238000000605 extraction Methods 0.000 claims abstract description 31
- 238000012545 processing Methods 0.000 claims abstract description 9
- 238000007781 pre-processing Methods 0.000 claims abstract description 8
- 238000004364 calculation method Methods 0.000 claims abstract description 7
- 238000013507 mapping Methods 0.000 claims description 4
- 238000004891 communication Methods 0.000 claims description 3
- 230000003287 optical effect Effects 0.000 claims description 2
- 230000004927 fusion Effects 0.000 abstract description 7
- 238000000034 method Methods 0.000 abstract description 4
- 230000000694 effects Effects 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012015 optical character recognition Methods 0.000 description 2
- 238000007500 overflow downdraw method Methods 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 1
- 230000001174 ascending effect Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000012512 characterization method Methods 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000010219 correlation analysis Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000008451 emotion Effects 0.000 description 1
- 230000002708 enhancing effect Effects 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000009499 grossing Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 230000031836 visual learning Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/583—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/5838—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/583—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/5846—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using extracted text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/583—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/5862—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using texture
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/22—Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/50—Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/56—Extraction of image or video features relating to colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/46—Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
- G06V10/467—Encoded features or binary features, e.g. local binary patterns [LBP]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Library & Information Science (AREA)
- General Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Databases & Information Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Image Analysis (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The invention discloses an image processing method, an image processing device and a computer storage medium, wherein the method comprises the following steps: preprocessing an image to be queried; performing visual feature extraction on the preprocessed image to be queried to obtain a visual feature extraction result; obtaining a visual feature vector based on the visual feature extraction result; extracting text information from the preprocessed image to be queried to obtain a text feature vector; fusing the visual feature vector and the text feature vector to obtain a fused query image vector; and carrying out similarity calculation on the fused query image vector and the image vectors in the database, and obtaining a picture similar to the image to be queried. By applying the embodiment of the invention, the LBP of a single pixel channel is calculated, and the LBP of the single pixel channel is integrated into the maximum multi-channel LBP to acquire the inter-channel texture information of the color image, so that the number of channels of the color image is not depended, the problem of dimension disaster is solved, and the accuracy of retrieval is improved by using the fusion of the text information and visual characteristics of the image.
Description
Technical Field
The present invention relates to the field of image retrieval technologies, and in particular, to an image processing method, an image processing device, and a computer storage medium.
Background
The content information of the image mainly comprises text semantic information and visual content information, and the existing image technology is divided into two types according to the difference of retrieval objects: one is Text-based image retrieval (Text-Based Image Retrieval, TBIR), and the other is Content-Based Image Retrieval, CBIR, based on the Content of the image itself. The TBIR utilizes the file name of the image and the surrounding text to construct an index, so that the retrieval of the image is converted into the retrieval of text information, and the required image can be returned by matching the keyword input by the user with the image library index. However, the acquisition of the text information of the image is mostly dependent on manual annotation, and with the rapid increase of the number of the images, TBIR faces the problems of time and labor consumption, subjective difference, uncertainty and the like of manual annotation. To overcome the problems faced by TBIR, CBIR should occur. The CBIR utilizes the visual features (color, texture and shape features) of the image to establish feature vectors, avoids artificial subjectivity, greatly improves accuracy, and improves image retrieval precision by utilizing a similarity measurement algorithm. However, the visual features of the image cannot fully represent the information of the image, and cannot embody the understanding and perception of people on the image. When one image is observed, not only the visual information such as the color, texture and the like of the image is seen, but also the image is understood by using the visual learning ability of the user, and the semantics and emotion of the image are perceived. Which is not expressible by the low-level visual features. Thus resulting in an unavoidable bottleneck problem, namely a "semantic gap" between the low-level visual features and the high-level semantics.
For merchant sign image retrieval, a user uses a mobile phone to shoot a sign picture through specific software so as to retrieve a corresponding merchant and obtain online information and service of the merchant. Most of the signboard images are single in background for highlighting shop names, the background is possibly repeated, and the images are easy to be inclined in angle, uneven in light, high in ambiguity and the like when the user shoots through a mobile phone, so that certain difficulty is brought to image retrieval.
Disclosure of Invention
The invention aims to provide an image processing method, an image processing device and a computer storage medium, which aim to overcome the existing defects and achieve better retrieval effect through fusion of image visual characteristics and text information.
In order to achieve the above object, the present invention provides an image processing method including:
preprocessing an image to be queried;
and carrying out visual feature extraction on the preprocessed image to be queried to obtain a visual feature extraction result, wherein the visual feature extraction result comprises the following steps: color features and texture features;
obtaining a visual feature vector based on the visual feature extraction result;
extracting text information from the preprocessed image to be queried to obtain text feature vectors;
fusing the visual feature vector and the text feature vector to obtain a fused query image vector;
and carrying out similarity calculation on the fused query image vector and the image vectors in the database to obtain a picture similar to the image to be queried.
Optionally, the step of extracting visual features of the preprocessed image to be queried includes:
extracting each color channel of the preprocessed image to be processed, and generating adder mapping by using a local binary pattern of each channel pixel;
based on the adder map, a local binary pattern of the largest channel of each pixel in the image is calculated, forming a texture histogram.
In one implementation, the expression of the visual feature vector is:
wherein F is Feature Descriptors representing texture features of the image, F Color Descriptor, len (F Feature ) The length sum, len (F Color ) Representing the length of the color feature descriptor, imageSize (I) is the number of pixels in the input image.
Optionally, extracting text information from the preprocessed image to be queried to obtain text feature vectors:
identifying texts in the image to be queried through an open source optical character identification tool;
text feature vectors are obtained based on the identified text.
The invention also provides a step of acquiring the image in the database, which comprises the following steps:
preprocessing each image to be placed in the database;
and carrying out visual feature extraction on each preprocessed image to obtain a visual feature extraction result corresponding to each image, wherein the visual feature extraction result corresponding to each image comprises the following steps: color features and texture features;
obtaining a visual feature vector corresponding to a visual feature extraction result of each image;
extracting text information of each image line, and using text feature vectors corresponding to the images;
and fusing the corresponding visual feature vector and the text feature vector of each image to obtain a fused image vector corresponding to the image, and storing the fused image vector into a database.
Further, an image processing apparatus is disclosed, the apparatus comprising a processor, and a memory connected to the processor through a communication bus; wherein,
the memory is used for storing an image processing program;
the processor is configured to execute the image processing program to implement any one of the image processing methods.
And, the present invention also discloses a computer storage medium storing one or more programs executable by one or more processors to cause the one or more processors to perform the steps of any of the image processing methods.
The image processing method, the device and the computer storage medium provided by the embodiment of the invention have the following beneficial effects:
(1) The present invention calculates the LBP of a single channel of pixels and integrates them into a maximum multi-channel LBP (MMLBP) to acquire inter-channel texture information of a color image, independent of the number of channels of the color image and solving the dimension disaster problem.
(2) The invention fully utilizes the fusion of the text information and visual characteristics of the image, and improves the retrieval accuracy.
Drawings
Fig. 1 is a schematic flow chart of an image processing method according to an embodiment of the invention.
Fig. 2 is a schematic diagram of visual feature extraction according to an embodiment of the invention.
Detailed Description
Other advantages and effects of the present invention will become apparent to those skilled in the art from the following disclosure, which describes the embodiments of the present invention with reference to specific examples. The invention may be practiced or carried out in other embodiments that depart from the specific details, and the details of the present description may be modified or varied from the spirit and scope of the present invention.
Please refer to fig. 1-2. It should be noted that, the illustrations provided in the present embodiment merely illustrate the basic concept of the present invention by way of illustration, and only the components related to the present invention are shown in the drawings and are not drawn according to the number, shape and size of the components in actual implementation, and the form, number and proportion of the components in actual implementation may be arbitrarily changed, and the layout of the components may be more complex.
The present invention provides an image processing method as shown in fig. 1, comprising:
s110, preprocessing an image to be queried;
it should be noted that, before indexing and searching, preprocessing is required for the image to be queried, for example, histogram equalization, binary filtering and gamma conversion are adopted for the image to be queried to achieve the effects of increasing local contrast of the image, smoothing and eliminating dryness of the image and enhancing the image, so that feature extraction is conveniently performed in the later stage, and specific processing procedures are not repeated in the embodiments of the present invention.
S120, performing visual feature extraction on the preprocessed image to be queried to obtain a visual feature extraction result, wherein the visual feature extraction result comprises: color features and texture features;
it can be understood that the pre-processed image is subjected to visual feature extraction, and the color features and texture features of the image to be queried are respectively extracted in the whole visual feature extraction process.
Specifically, the color information of each pixel is extracted by quantizing the RGB color space, and the color histogram is calculated, and the RGB color space has the characteristics of low calculation complexity and strong robustness. In this color space, R, G, B channels are uniformly quantized to Q r 、Q G 、Q B Represented elements. Color characterization of a single image, which quantifies the entire color space from 0 to Q r *Q G *Q B Q of-1 r *Q G *Q B Color, defined as:
Q RGB =Q r *Q G *R+Q B *G+B
here, Q r 、Q G And Q B Is set to 4 in order to quantify the resulting color space with 64 different intensity values from 0 to 63.
Further, extracting each color channel of the image, generating an enlarged mapper by using a local binary pattern of each channel pixel, and calculating a local binary pattern LBP of a maximum multi-channel of each pixel in the image based on the mapping of the adder to form a texture histogram. Let I be a multichannel image of size p×q×c, where P, Q, C represents the number of rows, columns, and channels of the image, respectively. I c Is the c channel in I, c E [1, C]. Is provided with N neighborhoods and arbitrary pixels I c (a, b) equi-radially spaced, expressed as
LBP c The calculation formula of (a, b) is as follows:
wherein n is E [1, N],a∈[1,P],b∈[1,Q],And (3) calculating the generated LBP equivalent decimal weight function. The adder map is defined as:
based on the adder map, a maximum multi-channel local binary pattern (MMLBP) for each pixel in the image is calculated as follows:
after calculating the MMLBP of each image, the texture features of the image are represented as a histogram obtained by looking up the frequency of occurrence of MMLBP values in the image, defined as F Texture =h(r k )=n k Wherein r is k Is the MMLBP value of the pixel in the generated image, the value is [0,255 ]]In between, nk is the number of pixels with a value MMLBP. As in fig. 2.
S130, obtaining a visual feature vector based on the visual feature extraction result;
the color feature and the texture feature are extracted based on the visual feature respectively, and the formed color feature vector and the texture feature vector are fused into a single visual feature vector. The fusion formula is as follows:
wherein F is Feature Representation of the drawingsLike descriptors of texture features, F Color Descriptor, len (F Feature ) The length sum, len (F Color ) Representing the length of the color feature descriptor, imageSize (I) is the number of pixels in the input image.
S140, extracting text information from the preprocessed image to be queried to obtain text feature vectors;
it can be understood that the text in the image to be queried is recognized through the existing open-source optical character recognition OCR tool, and the recognized text word is used as a label and a keyword to index the image so as to form a text feature vector.
S150, fusing the visual feature vector and the text feature vector to obtain a fused query image vector;
it should be noted that feature fusion is a process of combining two feature vectors into one feature vector having more recognition power than a single input feature vector. Feature fusion can be implemented at a certain level, such as a matching level, a feature level, and a decision level. The present invention contemplates the use of feature level fusion methods because 2 data patterns, namely visual feature vectors and text feature vectors, are generated. According to this fusion method, features extracted from the input entities are first combined and then further processed as a single unit for final fusion analysis. A typical correlation analysis method (CCA) is used to handle the mutual representation between two random eigenvectors. The aim of CCA is to find two sets of projection directions, so that the correlation between the projected two eigenvectors is maximized. The two eigenvectors ultimately represent the result as:
wherein X and Y are respectively expressed as a visual feature vector and a text feature vector, W x And W is y Representing a pair of projection directions on X and Y.
S160, similarity calculation is carried out between the fused query image vector and the image vectors in the database so as to obtain a picture similar to the image to be queried.
The invention also provides a step of acquiring images in a database, comprising:
preprocessing each image to be placed in the database;
and carrying out visual feature extraction on each preprocessed image to obtain a visual feature extraction result corresponding to each image, wherein the visual feature extraction result corresponding to each image comprises the following steps: color features and texture features;
obtaining a visual feature vector corresponding to a visual feature extraction result of each image;
extracting text information of each image line, and using text feature vectors corresponding to the images;
and fusing the corresponding visual feature vector and the text feature vector of each image to obtain a fused image vector corresponding to the image, and storing the fused image vector into a database.
The image vector in the database is the same as the image vector to be queried in the processing steps, specifically, the processing steps of steps S110-S150, and the embodiments of the present invention are not described herein.
It should be noted that, the images stored in the database are fused with the corresponding visual feature vectors and text feature vectors, the distance between the image to be queried and the image stored in the database is calculated by using Euclidean distance, then the images which are most similar to the image to be queried are output according to ascending order of the obtained distance, and then the retrieval effect of the fused images is measured by common indexes such as precision, recall ratio and the like. The Euclidean distance calculation formula is as follows:
wherein,feature vectors representing database images, F q Representing to-be-queriedFeature vectors of the image.
It can be understood that the calculated euclidean distance represents the similarity between two images, and then a maximum value can be selected from the values of the euclidean distances, and the image corresponding to the maximum value can be used as the image with the highest similarity with the image to be queried.
In addition, a threshold value can be set, and when the Euclidean distance is larger than the threshold value, all corresponding images are obtained, and the images are the most similar to the images to be queried.
Further, an image processing apparatus is disclosed, the apparatus comprising a processor, and a memory connected to the processor through a communication bus; wherein,
the memory is used for storing an image processing program;
the processor is configured to execute the image processing program to implement any one of the image processing methods.
And, the present invention also discloses a computer storage medium storing one or more programs executable by one or more processors to cause the one or more processors to perform the steps of any of the image processing methods. The above embodiments are merely illustrative of the principles of the present invention and its effectiveness, and are not intended to limit the invention. Modifications and variations may be made to the above-described embodiments by those skilled in the art without departing from the spirit and scope of the invention. Accordingly, it is intended that all equivalent modifications and variations of the invention be covered by the claims, which are within the ordinary skill of the art, be within the spirit and scope of the present disclosure.
Claims (3)
1. An image processing method, comprising:
preprocessing an image to be queried;
and carrying out visual feature extraction on the preprocessed image to be queried to obtain a visual feature extraction result, wherein the visual feature extraction result comprises the following steps: color features and texture features;
obtaining a visual feature vector based on the visual feature extraction result;
extracting text information from the preprocessed image to be queried to obtain text feature vectors;
fusing the visual feature vector and the text feature vector to obtain a fused query image vector;
performing similarity calculation on the fused query image vector and the image vectors in the database to obtain a picture similar to the image to be queried;
the step of extracting the visual characteristics of the preprocessed image to be queried comprises the following steps:
extracting each color channel of the preprocessed image to be processed, and generating adder mapping by using a local binary pattern of each channel pixel;
based on the adder mapping, calculating a local binary pattern of a maximum channel of each pixel in the image to form a texture histogram;
the expression of the visual characteristic vector is as follows:
wherein F is Feature Descriptors representing texture features of the image, F Color Descriptor, len (F Feature ) The length sum, len (F Color ) Representing the length of the color feature descriptor, imageSize (I) is the number of pixels in the input image;
extracting text information from the preprocessed image to be queried to obtain text feature vectors:
identifying texts in the image to be queried through an open source optical character identification tool;
acquiring a text feature vector based on the identified text;
a step of acquiring an image vector in a database, comprising:
preprocessing each image to be placed in the database;
and carrying out visual feature extraction on each preprocessed image to obtain a visual feature extraction result corresponding to each image, wherein the visual feature extraction result corresponding to each image comprises the following steps: color features and texture features;
obtaining a visual feature vector corresponding to a visual feature extraction result of each image;
extracting text information of each image line, and using text feature vectors corresponding to the images;
and fusing the corresponding visual feature vector and the text feature vector of each image to obtain a fused image vector corresponding to the image, and storing the fused image vector into a database.
2. An image processing apparatus, comprising a processor, and a memory coupled to the processor via a communication bus; wherein,
the memory is used for storing an image processing program;
the processor is configured to execute the image processing program to implement the steps of the image processing method according to claim 1.
3. A computer storage medium storing one or more programs executable by one or more processors to cause the one or more processors to perform the steps of the image processing method of claim 1.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011505344.XA CN112528905B (en) | 2020-12-18 | 2020-12-18 | Image processing method, device and computer storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011505344.XA CN112528905B (en) | 2020-12-18 | 2020-12-18 | Image processing method, device and computer storage medium |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112528905A CN112528905A (en) | 2021-03-19 |
CN112528905B true CN112528905B (en) | 2024-04-05 |
Family
ID=75001505
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011505344.XA Active CN112528905B (en) | 2020-12-18 | 2020-12-18 | Image processing method, device and computer storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112528905B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116797533B (en) * | 2023-03-24 | 2024-01-23 | 东莞市冠锦电子科技有限公司 | Appearance defect detection method and system for power adapter |
CN116595064A (en) * | 2023-04-28 | 2023-08-15 | 华为技术有限公司 | Data mining system, method and device based on graphic and text information combination |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104376105A (en) * | 2014-11-26 | 2015-02-25 | 北京航空航天大学 | Feature fusing system and method for low-level visual features and text description information of images in social media |
CN106708943A (en) * | 2016-11-22 | 2017-05-24 | 安徽睿极智能科技有限公司 | Image retrieval reordering method and system based on arrangement fusion |
WO2018188240A1 (en) * | 2017-04-10 | 2018-10-18 | 北京大学深圳研究生院 | Cross-media retrieval method based on deep semantic space |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9075825B2 (en) * | 2011-09-26 | 2015-07-07 | The University Of Kansas | System and methods of integrating visual features with textual features for image searching |
-
2020
- 2020-12-18 CN CN202011505344.XA patent/CN112528905B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104376105A (en) * | 2014-11-26 | 2015-02-25 | 北京航空航天大学 | Feature fusing system and method for low-level visual features and text description information of images in social media |
CN106708943A (en) * | 2016-11-22 | 2017-05-24 | 安徽睿极智能科技有限公司 | Image retrieval reordering method and system based on arrangement fusion |
WO2018188240A1 (en) * | 2017-04-10 | 2018-10-18 | 北京大学深圳研究生院 | Cross-media retrieval method based on deep semantic space |
Non-Patent Citations (2)
Title |
---|
单强 ; 孙晓明 ; .多特征分层融合医疗设备图像检索方法.哈尔滨理工大学学报.2017,(02),全文. * |
张霞 ; 郑逢斌 ; .基于多层次视觉语义特征融合的图像检索算法.包装工程.2018,(19),全文. * |
Also Published As
Publication number | Publication date |
---|---|
CN112528905A (en) | 2021-03-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11657084B2 (en) | Correlating image annotations with foreground features | |
CN108694225B (en) | Image searching method, feature vector generating method and device and electronic equipment | |
Sajjad et al. | Integrating salient colors with rotational invariant texture features for image representation in retrieval systems | |
US20180129658A1 (en) | Color sketch image searching | |
CN112528905B (en) | Image processing method, device and computer storage medium | |
Rastegar et al. | Designing a new deep convolutional neural network for content-based image retrieval with relevance feedback | |
CN114972847A (en) | Image processing method and device | |
CN113269170A (en) | Intelligent portrait building block matching method and system based on feature similarity measurement | |
CN116562270A (en) | Natural language processing system supporting multi-mode input and method thereof | |
CN116244464A (en) | Hand-drawing image real-time retrieval method based on multi-mode data fusion | |
CN112561976A (en) | Image dominant color feature extraction method, image retrieval method, storage medium and device | |
CN105095468A (en) | Novel image retrieval method and system | |
CN111695507B (en) | Static gesture recognition method based on improved VGGNet network and PCA | |
CN103020630A (en) | Processing method and device of image features | |
Lizarraga-Morales et al. | Improving a rough set theory-based segmentation approach using adaptable threshold selection and perceptual color spaces | |
Parra et al. | Automatic gang graffiti recognition and interpretation | |
Varish et al. | A content based image retrieval using color and texture features | |
CN111754459B (en) | Dyeing fake image detection method based on statistical depth characteristics and electronic device | |
CN114842301A (en) | Semi-supervised training method of image annotation model | |
CN114329024A (en) | Icon searching method and system | |
Gupta et al. | A Framework for Semantic based Image Retrieval from Cyberspace by mapping low level features with high level semantics | |
Admile | A survey on different image retrieval techniques | |
CN113538214A (en) | Method and system for controlling makeup migration and storage medium | |
Khwildi et al. | A new indexing method of HDR images using color histograms | |
CN113129410B (en) | Sketch image conversion method and related product |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |