CN113094538A - Image retrieval method, device and computer-readable storage medium - Google Patents

Image retrieval method, device and computer-readable storage medium Download PDF

Info

Publication number
CN113094538A
CN113094538A CN201911334579.4A CN201911334579A CN113094538A CN 113094538 A CN113094538 A CN 113094538A CN 201911334579 A CN201911334579 A CN 201911334579A CN 113094538 A CN113094538 A CN 113094538A
Authority
CN
China
Prior art keywords
target image
determining
retrieval
image
module
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911334579.4A
Other languages
Chinese (zh)
Inventor
张超颖
项超
刘珮
贾丹
赵啸宇
孟维业
王建秀
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Telecom Corp Ltd
Original Assignee
China Telecom Corp Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Telecom Corp Ltd filed Critical China Telecom Corp Ltd
Priority to CN201911334579.4A priority Critical patent/CN113094538A/en
Publication of CN113094538A publication Critical patent/CN113094538A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/5846Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using extracted text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/5866Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, manually generated location and time information
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Abstract

The disclosure relates to a retrieval method and a retrieval device for images and a computer readable storage medium, and relates to the technical field of computers. Determining a description text of the target image by utilizing a first machine learning model according to the extracted feature vector of the target image; performing face recognition on the target image by using a second machine learning model to determine identity information of the target image; and matching the retrieval keywords provided by the user according to the description texts and the identity information of the target images, and determining corresponding retrieval results in the target images.

Description

Image retrieval method, device and computer-readable storage medium
Technical Field
The present disclosure relates to the field of computer technologies, and in particular, to a method and an apparatus for retrieving an image, and a computer-readable storage medium.
Background
With the rapid development of technologies such as digital imaging and computer storage, a large amount of image data is generated every day through various intelligent terminals. With the rapid increase of image data volume, a large-scale visual information image library is formed, so that the traditional image retrieval based on file names cannot meet the requirements of people. In the face of hundreds of millions of image big data, how to accurately and rapidly retrieve the image required by the user becomes an urgent problem to be solved.
In the related art, the image retrieval method mainly includes: TBIR (Text-based Image Retrieval technology), which performs keyword matching to complete a Retrieval task by indexing the labeled Image keywords in the database; CBIR (Content-based Image Retrieval) searches for low-level visual features such as color, texture, and object outline included in an Image itself as a basis.
Disclosure of Invention
The inventors of the present disclosure found that the following problems exist in the above-described related art: the method is greatly influenced by human factors, and deep information of the image cannot be mined, so that the image retrieval result is inaccurate.
In view of this, the present disclosure provides an image retrieval technique, which can achieve accuracy of an image retrieval result.
According to some embodiments of the present disclosure, there is provided an image retrieval method including: determining a description text of the target image by utilizing a first machine learning model according to the extracted feature vector of the target image; performing face recognition on the target image by using a second machine learning model to determine identity information of the target image; and matching the retrieval keywords provided by the user according to the description texts and the identity information of the target images, and determining corresponding retrieval results in the target images.
In some embodiments, determining the description text of the target image comprises: determining the weight of each feature vector by using an attention mechanism module of a first machine learning model according to the extracted feature vectors of different regions in the target image; and determining the description text of the target image by utilizing the first machine learning model according to the feature vectors and the weights thereof.
In some embodiments, the first machine learning model is a double layer LSTM model (Long Short-Term Memory network) that includes a first LSTM module and a second LSTM module, and the attention mechanism module is disposed between the first LSTM module and the second LSTM module.
In some embodiments, determining the weight for each feature vector comprises: the weight of the feature vector is determined based on the output of the first LSTM module and the feature vector.
In some embodiments, determining the description text of the target image comprises: determining the output of a second LSTM module according to the output of the first LSTM module and the weighted sum of the feature vectors; and determining the description text of the target image according to the output of the second LSTM module.
In some embodiments, determining the description text of the target image comprises: determining the probability of each candidate word by using a first machine learning model; determining a plurality of candidate description texts of the target image according to the probability of each candidate word; determining the probability of each candidate description text according to the probability of each candidate word in each candidate description text; and determining one or more candidate description texts with the highest probability as the description texts.
In some embodiments, the method further comprises: and in response to the target image being deleted, deleting the corresponding description text and the corresponding identity information of the target image.
According to other embodiments of the present disclosure, there is provided an image retrieval apparatus including: the text determining unit is used for determining a description text of the target image by utilizing the first machine learning model according to the extracted feature vector of the target image; the information determining unit is used for carrying out face recognition on the target image by utilizing the second machine learning model and determining the identity information of the target image; and the retrieval unit is used for matching the retrieval keywords provided by the user according to the description texts and the identity information of the target images and determining corresponding retrieval results in the target images.
In some embodiments, the text determination unit determines a weight of each feature vector by using an attention mechanism module of the first machine learning model according to the extracted feature vectors of different regions in the target image, and determines the description text of the target image by using the first machine learning model according to each feature vector and the weight thereof.
In some embodiments, the first machine learning model is a two-layer LSTM model, the LSTM model including a first LSTM module and a second LSTM module, the attention mechanism module disposed between the first LSTM module and the second LSTM module; the text determination unit determines a weight of the feature vector based on the output of the first LSTM module and the feature vector.
In some embodiments, the text determination unit determines an output of a second LSTM module based on the output of the first LSTM module and a weighted sum of the feature vectors, and determines a descriptive text of the target image based on the output of the second LSTM module.
In some embodiments, the text determination unit determines a probability of each candidate word using the first machine learning model, determines a plurality of candidate description texts of the target image according to the probability of each candidate word, determines a probability of each candidate description text according to the probability of each candidate word in each candidate description text, and determines one or more candidate description texts with the highest probability as the description text.
In some embodiments, the apparatus further comprises: and the deleting unit is used for deleting the corresponding description text and the corresponding identity information of the target image in response to the target image being deleted.
According to still further embodiments of the present disclosure, there is provided an image retrieval apparatus including: a memory; and a processor coupled to the memory, the processor configured to perform the method of retrieving an image in any of the above embodiments based on instructions stored in the memory device.
According to still further embodiments of the present disclosure, there is provided a computer-readable storage medium having stored thereon a computer program which, when executed by a processor, implements the retrieval method of an image in any of the above embodiments.
In the above embodiment, the image description and the face recognition result determined by the machine learning method are combined as the basis for image retrieval. Therefore, the deep information of the image can be mined, and the accuracy of image retrieval is improved.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments of the disclosure and together with the description, serve to explain the principles of the disclosure.
The present disclosure can be more clearly understood from the following detailed description with reference to the accompanying drawings, in which:
FIG. 1 illustrates a flow diagram of some embodiments of an image retrieval method of the present disclosure;
FIG. 2 illustrates a schematic diagram of some embodiments of an image retrieval method of the present disclosure;
FIG. 3 shows a schematic diagram of further embodiments of an image retrieval method of the present disclosure;
FIG. 4 illustrates a schematic diagram of some embodiments of an image retrieval device of the present disclosure;
FIG. 5 illustrates a flow diagram of further embodiments of an image retrieval method of the present disclosure;
FIG. 6 illustrates a flow diagram of further embodiments of the image retrieval method of the present disclosure;
FIG. 7 illustrates a block diagram of some embodiments of an image retrieval device of the present disclosure;
FIG. 8 shows a block diagram of further embodiments of an image retrieval device of the present disclosure;
fig. 9 shows a block diagram of further embodiments of an image retrieval device of the present disclosure.
Detailed Description
Various exemplary embodiments of the present disclosure will now be described in detail with reference to the accompanying drawings. It should be noted that: the relative arrangement of the components and steps, the numerical expressions, and numerical values set forth in these embodiments do not limit the scope of the present disclosure unless specifically stated otherwise.
Meanwhile, it should be understood that the sizes of the respective portions shown in the drawings are not drawn in an actual proportional relationship for the convenience of description.
The following description of at least one exemplary embodiment is merely illustrative in nature and is in no way intended to limit the disclosure, its application, or uses.
Techniques, methods, and apparatus known to those of ordinary skill in the relevant art may not be discussed in detail, but are intended to be part of the specification where appropriate.
In all examples shown and discussed herein, any particular value should be construed as merely illustrative, and not limiting. Thus, other examples of the exemplary embodiments may have different values.
It should be noted that: like reference numbers and letters refer to like items in the following figures, and thus, once an item is defined in one figure, further discussion thereof is not required in subsequent figures.
As mentioned above, the TBIR needs to be manually labeled in advance, which has the problems of high labeling cost and low efficiency, and the manual labeling mode is affected by personal subjective factors; CBIR utilizes the low-level visual features of the image, and semantic gaps exist between the low-level visual features and the high-level semantic features of the image understood by people, so that the retrieval effect is poor.
Moreover, the current search method for personal photos is only a single search method based on file name, date and classification. And most of the photo libraries store personal life photos, but a retrieval mode based on the names of people is lacked. Therefore, in the case of a picture library with more than thousands of data, the user searches for the required target picture based on the single modes, which is inefficient and affects the searching efficiency of the user.
In order to solve the problems, the visual feature mapping of the lower layer of the image is automatically converted into high-layer semantic information (namely, an image semantic description algorithm) which is known by people through a computer from the content information of the image. Therefore, the accurate searching function of the target image can be realized by searching semantic information contained in the image.
The method solves the problem that the emphasis and the details of the generated sentences are not prominent by optimizing the neural network model. The computer automatically generates Chinese and English description sentences of each image, thereby improving the accuracy of image description. And a face recognition algorithm is introduced, so that Chinese and English description sentences and character identity information can be automatically generated for the photos in the photo library. And constructing a database of the pictures by using the semantic information, and further realizing accurate retrieval based on Chinese and English keywords and sentences and names of people through the database. For example, it can be realized by the following embodiments.
Fig. 1 illustrates a flow diagram of some embodiments of an image retrieval method of the present disclosure.
As shown in fig. 1, the method includes: step 110, determining a description text; step 120, determining identity information; step 130, determining the retrieval result.
In step 110, a description text of the target image is determined by using the first machine learning model according to the extracted feature vector of the target image. For example, the feature vector of the target image may be extracted by a CNN (Convolutional Neural Networks) module. The first machine learning model may be a deep learning machine learning model.
In some embodiments, determining a weight of each feature vector by using an Attention mechanism (Attention) module of the first machine learning model according to the extracted feature vectors of different regions in the target image; and determining the description text of the target image by utilizing the first machine learning model according to the feature vectors and the weights thereof.
For example, the first machine learning model is a two-layer LSTM model that includes a first LSTM module and a second LSTM module. An attention mechanism module is disposed between the first LSTM module and the second LSTM module. The weight of the feature vector is determined based on the output of the first LSTM module and the feature vector.
In some embodiments, the output of the second LSTM module is determined from the weighted sum of the output of the first LSTM module and the respective feature vectors; and determining the description text of the target image according to the output of the second LSTM module.
In some embodiments, a probability of each candidate word is determined using a first machine learning model; determining a plurality of candidate description texts of the target image according to the probability of each candidate word; determining the probability of each candidate description text according to the probability of each candidate word in each candidate description text; and determining one or more candidate description texts with the highest probability as the description texts.
In step 120, the second machine learning model is used to perform face recognition on the target image, and identify information of the target image is determined. The first machine learning model may be a deep learning machine learning model. The second machine learning model may be a deep learning machine learning model.
In step 130, the search keywords provided by the user are matched according to the description texts and the identity information of the target images, and corresponding search results are determined in the target images.
In some embodiments, in response to a target image being deleted, the corresponding descriptive text and identity information for the target image is deleted.
In the above embodiment, the image description and the face recognition result determined by the machine learning method are combined as the basis for image retrieval. Therefore, the deep information of the image can be mined, and the accuracy of image retrieval is improved.
In the embodiment, starting from the content information of the image, the computer can automatically convert the visual feature mapping of the lower layer of the image into high-layer semantic information known by people by improving the image description algorithm and adding the face recognition algorithm. Thus, the retrieval function is realized by searching semantic information contained in the image itself. This may be achieved, for example, by the embodiment of fig. 2.
Fig. 2 shows a schematic diagram of some embodiments of an image retrieval method of the present disclosure.
As shown in fig. 2, in response to the image being stored in the image library, in terms of generating the chinese-english description sentence, a CNN (e.g., inclusion _ V4) model is used to extract a feature vector of the image; analyzing descriptive sentences (descriptive texts) of the images by using a double-layer LSTM model; and an improved Attention mechanism (Attention) is added so that the generated descriptive statement has corresponding weights for different regions of the image. In this way, the primary and secondary relationships of the image content can be highlighted, and the problem that the emphasis and the detail of the generated sentence are not highlighted is solved.
In response to the image being stored in the image repository, in generating the person identity information, a Face recognition algorithm (Face _ recognition) is introduced to build a Face database, automatically generating the person identity information in the photograph. Therefore, the semantic information of the photos can be enriched, and the user can better search the target photos.
And associating the generated description sentences and the identity information with corresponding images and storing the associated description sentences and the identity information into a database to serve as matching objects for keyword retrieval. Thus, the matching of keywords can be realized to search the image library.
Fig. 3 shows a schematic diagram of further embodiments of the image retrieval method of the present disclosure.
As shown in fig. 3, the feature extraction module of the input image may adopt an inclusion _ V4 network model to obtain high-level semantic information of the image; the descriptive sentence generating module may include a two-layer LSTM network model, and an Attention-improving mechanism (Attention) module for determining feature vector weight values for different regions of the image.
In some embodiments, a two-layer LSTM based loop attention mechanism weights the feature vectors of each image region such that each feature vector has a different weight value. When generating a Chinese output sequence, the double-layer LSTM loop attention mechanism model can select input sequence information with the maximum relevance to the word output at the current moment. For example, the model may also be learned by iteratively optimizing parameters of the network based on a loss function of the words.
In some embodiments, the two-layer LSTM model may include two layers of LSTM1 modules and LSTM2 modules.
Figure BDA0002330603840000071
And
Figure BDA0002330603840000072
respectively representing the outputs of the LSTM1 module and the LSTM2 module at time t, and so on at times t-1 and t + 1. v. ofiThe feature vector of the ith area of the image is an integer which is greater than or equal to 1 and less than or equal to k.
Figure BDA0002330603840000073
The feature vectors are weighted by the attention mechanism module. XtAnd Xt+1The inputs to the LSTM1 module at time t and time t +1, respectively.
In some embodiments, the attention mechanism module calculates a weight a of the ith eigenvector at time ti,tThe following were used:
Figure BDA0002330603840000081
αt=Softmax(at)
Wva、Whaand waIs a model parameter determined by training, phi (-) represents the cosine tangent function.
In some embodiments, the attention mechanism module weights each feature vector
Figure BDA0002330603840000082
The following were used:
Figure BDA0002330603840000083
in some embodiments, the model predicts the generated word S at time ttThe probability of (c) is:
Figure BDA0002330603840000084
S1to St-1Words generated from time 1 to time t-1. WpAnd bpAre model parameters obtained through training. In this case, for T moments, each word is represented byThe probabilities of the composed description text are:
Figure BDA0002330603840000085
in the above-described embodiments, the image description method is improved. An Attention mechanism is added, so that the generated descriptive statement has certain weight. Therefore, the primary and secondary relations of the image content are highlighted, and the function of automatically generating Chinese and English description sentences for the photos is realized. And a face recognition algorithm is added, and the multilayer semantic information of the photos is integrated. Thus, a semantic database of the personal photo system is constructed and information is automatically updated in real time according to the photo database.
In the embodiment, through the generated personal photo semantic database, the photo accurate retrieval based on Chinese and English key word groups is realized. Not only can show the information of the personal photo library in a visual form, but also can delete the photo library in a classified manner based on the key phrase. Therefore, an effective and quick method and system are provided for realizing personal photo retrieval.
In the embodiment, a technical scheme for retrieving the personal photos based on the Chinese and English descriptions of the images is designed, so that people can retrieve the needed photos by searching semantic information of the photos and also can retrieve the photos by identity information of people. In addition, the technical scheme of the disclosure also comprises multiple convenient functions of deleting and displaying the retrieval picture. Therefore, the method can have wide application space in the field of future photo retrieval.
Fig. 4 shows a schematic diagram of some embodiments of an image retrieval apparatus of the present disclosure.
As shown in fig. 4, the image retrieval device may be a personal photo retrieval system based on the chinese and english descriptions of the images. For example, the image retrieval device may include an image English-Chinese description subsystem and a personal photo retrieval subsystem.
In some embodiments, the image Chinese and English description subsystem may include an image description sentence generation module, a face recognition module and an image information database module.
For example, the image description sentence generation module (e.g., text determination unit) implements a function of automatically generating three Chinese and English description sentences with the highest probability for each photo in the personal computer photo library. When the addition, deletion or replacement of the photos in the photo library of the personal computer occurs, the corresponding photo information in the image information database module is correspondingly operated.
For example, a face recognition module (e.g., an information determination unit) implements a function of automatically recognizing name information of a person in a personal photo library. The photo information database module realizes real-time management of photo names, Chinese and English description sentences and character information appearing in the photos in the personal computer photo library.
In some embodiments, the personal photo retrieval subsystem may include a photo retrieval module, a photo deletion module, and a photo display module.
For example, the photo retrieval module (such as a retrieval unit) realizes the functions of inputting and deleting keywords and retrieving related photos from the image information database according to the keywords; the photo deleting module (such as a deleting unit) realizes the function of deleting the retrieved photos; the photo display module realizes visualization of the retrieved photos and the photo names.
Fig. 5 illustrates a flow diagram of further embodiments of the image retrieval method of the present disclosure.
As shown in fig. 5, in the step of setting a path of a photo library, when the photo information database is generated, a folder path of a pc photo library needs to be set to obtain related photo information.
In the step of initializing the database, the database is initialized, and the data storage format of the database is set.
And judging whether the photo library is newly added with photos. When detecting that the photos are added in the personal photo library, in the step of calling the command terminal, the system automatically calls the terminal command, generates Chinese and English description sentences and figure identity information and stores the Chinese and English description sentences and the figure identity information into a database; otherwise, judging whether the photo library deletes the photo.
When detecting that the photo in the personal photo library is deleted, deleting the corresponding photo and semantic information from the image information database, thereby realizing the dynamic management of the database. Thus, the work of generating the photo information database is completed.
Fig. 6 illustrates a flow diagram of further embodiments of the image retrieval method of the present disclosure.
As shown in fig. 6, when retrieving a photo from the database in the step of inputting a chinese and english keyword, a photo file name, chinese and english keyword information, and person identification information may be input.
In the search matching step, photo character information is retrieved from the image data information base according to the input keyword.
In the step of judging whether the matching is successful, the corresponding degree of the input information and the information in the database can be matched during retrieval.
When the matching is successful, all matched photo names may be returned. And the corresponding photo can be extracted from the personal computer photo library according to the photo name and displayed.
When the matching fails, a prompt window, such as "not find desired picture", etc., may pop up to prompt the user that the personal computer photo library does not have a photo corresponding to the keyword.
It may be determined whether to stop the retrieval. Under the condition of not stopping the retrieval, the user can input the keywords again for retrieval; otherwise, the retrieval task may be ended directly.
Fig. 7 illustrates a block diagram of some embodiments of an image retrieval device of the present disclosure.
As shown in fig. 7, the retrieval means 7 of the image includes a text determination unit 71, an information determination unit 72, and a retrieval unit 73.
The text determination unit 71 determines the description text of the target image using the first machine learning model based on the extracted feature vector of the target image.
In some embodiments, the text determination unit 71 determines the weight of each feature vector according to the feature vectors of different regions in the extracted target image by using the attention mechanism module of the first machine learning model; the text determination unit 71 determines the description text of the target image using the first machine learning model based on each feature vector and its weight.
In some embodiments, the first machine learning model is a two-layer LSTM model, the LSTM model including a first LSTM module and a second LSTM module; the attention mechanism module is arranged between the first LSTM module and the second LSTM module; the text determination unit 71 determines the weight of the feature vector based on the output of the first LSTM module and the feature vector.
In some embodiments, the text determination unit 71 determines the output of the second LSTM module from the weighted sum of the output of the first LSTM module and the respective feature vectors; the text determination unit 71 determines the description text of the target image based on the output of the second LSTM module.
In some embodiments, the text determination unit 71 determines the probability of each candidate word using a first machine learning model; the text determining unit 71 determines a plurality of candidate description texts of the target image according to the probability of each candidate word; the text determining unit 71 determines the probability of each candidate description text according to the probability of each candidate word in each candidate description text; the text determination unit 71 determines one or more candidate description texts having the highest probability as the description text.
The information determination unit 72 performs face recognition on the target image using the second machine learning model, and determines identity information of the target image.
The retrieval unit 73 matches the retrieval keywords provided by the user according to the description texts and the identity information of the plurality of target images, and determines corresponding retrieval results in the plurality of target images.
In some embodiments, the retrieving device 7 further includes a deleting unit 74, which is used for deleting the corresponding description text and the identity information of the target image in response to the target image being deleted.
In the above embodiment, the image description and the face recognition result determined by the machine learning method are combined as the basis for image retrieval. Therefore, the deep information of the image can be mined, and the accuracy of image retrieval is improved.
Fig. 8 shows a block diagram of further embodiments of an image retrieval apparatus of the present disclosure.
As shown in fig. 8, the image retrieval device 8 of this embodiment includes: a memory 81 and a processor 82 coupled to the memory 81, the processor 82 being configured to execute a retrieval method of an image in any one of the embodiments of the present disclosure based on instructions stored in the memory 81.
The memory 81 may include, for example, a system memory, a fixed nonvolatile storage medium, and the like. The system memory stores, for example, an operating system, application programs, a boot loader, a database, and other programs.
Fig. 9 shows a block diagram of further embodiments of an image retrieval device of the present disclosure.
As shown in fig. 9, the image retrieval device 9 of this embodiment includes: a memory 910 and a processor 920 coupled to the memory 910, wherein the processor 920 is configured to execute the image retrieval method in any of the embodiments based on instructions stored in the memory 910.
The memory 910 may include, for example, system memory, fixed non-volatile storage media, and the like. The system memory stores, for example, an operating system, an application program, a boot loader, and other programs.
The retrieval means 9 of the image may further include an input-output interface 930, a network interface 940, a storage interface 950, and the like. These interfaces 930, 940, 950 and the memory 910 and the processor 920 may be connected, for example, by a bus 960. The input/output interface 930 provides a connection interface for input/output devices such as a display, a mouse, a keyboard, and a touch screen. The network interface 940 provides a connection interface for various networking devices. The storage interface 950 provides a connection interface for external storage devices such as an SD card and a usb disk.
As will be appreciated by one skilled in the art, embodiments of the present disclosure may be provided as a method, system, or computer program product. Accordingly, the present disclosure may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present disclosure may take the form of a computer program product embodied on one or more computer-usable non-transitory storage media having computer-usable program code embodied therein.
So far, a retrieval method, an apparatus, and a computer-readable storage medium of an image according to the present disclosure have been described in detail. Some details that are well known in the art have not been described in order to avoid obscuring the concepts of the present disclosure. It will be fully apparent to those skilled in the art from the foregoing description how to practice the presently disclosed embodiments.
The method and system of the present disclosure may be implemented in a number of ways. For example, the methods and systems of the present disclosure may be implemented by software, hardware, firmware, or any combination of software, hardware, and firmware. The above-described order for the steps of the method is for illustration only, and the steps of the method of the present disclosure are not limited to the order specifically described above unless specifically stated otherwise. Further, in some embodiments, the present disclosure may also be embodied as programs recorded in a recording medium, the programs including machine-readable instructions for implementing the methods according to the present disclosure. Thus, the present disclosure also covers a recording medium storing a program for executing the method according to the present disclosure.
Although some specific embodiments of the present disclosure have been described in detail by way of example, it should be understood by those skilled in the art that the foregoing examples are for purposes of illustration only and are not intended to limit the scope of the present disclosure. It will be appreciated by those skilled in the art that modifications may be made to the above embodiments without departing from the scope and spirit of the present disclosure. The scope of the present disclosure is defined by the appended claims.

Claims (14)

1. An image retrieval method, comprising:
determining a description text of the target image by utilizing a first machine learning model according to the extracted feature vector of the target image;
performing face recognition on the target image by using a second machine learning model, and determining identity information of the target image;
and matching the retrieval keywords provided by the user according to the description texts and the identity information of the target images, and determining corresponding retrieval results in the target images.
2. The retrieval method of claim 1, wherein the determining the description text of the target image comprises:
determining the weight of each feature vector by using an attention mechanism module of the first machine learning model according to the extracted feature vectors of different regions in the target image;
and determining the description text of the target image by utilizing the first machine learning model according to the feature vectors and the weights thereof.
3. The retrieval method according to claim 2,
the first machine learning model is a double-layer long-short term memory (LSTM) model, the LSTM model comprises a first LSTM module and a second LSTM module, and the attention mechanism module is arranged between the first LSTM module and the second LSTM module;
the determining the weight of each feature vector comprises:
and determining the weight of the feature vector according to the output of the first LSTM module and the feature vector.
4. The retrieval method of claim 3, wherein the determining the description text of the target image comprises:
determining the output of the second LSTM module according to the output of the first LSTM module and the weighted sum of the feature vectors;
and determining the description text of the target image according to the output of the second LSTM module.
5. The retrieval method of claim 1, wherein the determining the description text of the target image comprises:
determining the probability of each candidate word by using the first machine learning model;
determining a plurality of candidate description texts of the target image according to the probability of each candidate word;
determining the probability of each candidate description text according to the probability of each candidate word in each candidate description text;
and determining one or more candidate description texts with the highest probability as the description texts.
6. The retrieval method of any one of claims 1-5, further comprising:
and in response to the target image being deleted, deleting the corresponding description text and the corresponding identity information of the target image.
7. An image retrieval apparatus comprising:
the text determining unit is used for determining a description text of the target image by utilizing a first machine learning model according to the extracted feature vector of the target image;
the information determining unit is used for carrying out face recognition on the target image by utilizing a second machine learning model and determining the identity information of the target image;
and the retrieval unit is used for matching the retrieval keywords provided by the user according to the description texts and the identity information of the target images and determining corresponding retrieval results in the target images.
8. The retrieval device of claim 7,
the text determining unit determines the weight of each feature vector by using an attention mechanism module of the first machine learning model according to the extracted feature vectors of different areas in the target image, and determines the description text of the target image by using the first machine learning model according to each feature vector and the weight thereof.
9. The retrieval device of claim 8,
the first machine learning model is a double-layer long-short term memory (LSTM) model, the LSTM model comprises a first LSTM module and a second LSTM module, and the attention mechanism module is arranged between the first LSTM module and the second LSTM module;
the text determination unit determines the weight of the feature vector according to the output of the first LSTM module and the feature vector.
10. The retrieval device of claim 9,
the text determining unit determines the output of the second LSTM module according to the output of the first LSTM module and the weighted sum of the feature vectors, and determines the description text of the target image according to the output of the second LSTM module.
11. The retrieval device of claim 7,
the text determination unit determines the probability of each candidate word by using the first machine learning model, determines a plurality of candidate description texts of the target image according to the probability of each candidate word, determines the probability of each candidate description text according to the probability of each candidate word in each candidate description text, and determines one or more candidate description texts with the highest probability as the description texts.
12. The retrieval device of any of claims 7-11, further comprising:
and the deleting unit is used for deleting the corresponding description text and the corresponding identity information of the target image in response to the target image being deleted.
13. An image retrieval apparatus comprising:
a memory; and
a processor coupled to the memory, the processor configured to perform the method of retrieving an image of any of claims 1-6 based on instructions stored in the memory.
14. A computer-readable storage medium, on which a computer program is stored which, when being executed by a processor, implements the image retrieval method according to any one of claims 1 to 6.
CN201911334579.4A 2019-12-23 2019-12-23 Image retrieval method, device and computer-readable storage medium Pending CN113094538A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911334579.4A CN113094538A (en) 2019-12-23 2019-12-23 Image retrieval method, device and computer-readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911334579.4A CN113094538A (en) 2019-12-23 2019-12-23 Image retrieval method, device and computer-readable storage medium

Publications (1)

Publication Number Publication Date
CN113094538A true CN113094538A (en) 2021-07-09

Family

ID=76662827

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911334579.4A Pending CN113094538A (en) 2019-12-23 2019-12-23 Image retrieval method, device and computer-readable storage medium

Country Status (1)

Country Link
CN (1) CN113094538A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113779297A (en) * 2021-09-01 2021-12-10 北京橙色云科技有限公司 Information searching method and device based on picture and storage medium
CN117173163A (en) * 2023-11-01 2023-12-05 浙江同花顺智能科技有限公司 Portrait quality assessment method, system, device and readable storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103793697A (en) * 2014-02-17 2014-05-14 北京旷视科技有限公司 Identity labeling method of face images and face identity recognition method of face images
CN104462212A (en) * 2014-11-04 2015-03-25 百度在线网络技术(北京)有限公司 Information exhibiting method and device
CN105808732A (en) * 2016-03-10 2016-07-27 北京大学 Integration target attribute identification and precise retrieval method based on depth measurement learning
US20190005069A1 (en) * 2017-06-28 2019-01-03 Google Inc. Image Retrieval with Deep Local Feature Descriptors and Attention-Based Keypoint Descriptors
CN110599557A (en) * 2017-08-30 2019-12-20 深圳市腾讯计算机系统有限公司 Image description generation method, model training method, device and storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103793697A (en) * 2014-02-17 2014-05-14 北京旷视科技有限公司 Identity labeling method of face images and face identity recognition method of face images
CN104462212A (en) * 2014-11-04 2015-03-25 百度在线网络技术(北京)有限公司 Information exhibiting method and device
CN105808732A (en) * 2016-03-10 2016-07-27 北京大学 Integration target attribute identification and precise retrieval method based on depth measurement learning
US20190005069A1 (en) * 2017-06-28 2019-01-03 Google Inc. Image Retrieval with Deep Local Feature Descriptors and Attention-Based Keypoint Descriptors
CN110599557A (en) * 2017-08-30 2019-12-20 深圳市腾讯计算机系统有限公司 Image description generation method, model training method, device and storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
FEN XIAO等: "DAA: Dual LSTMs with adaptive attention for image captioning", 《NEUROCOMPUTING》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113779297A (en) * 2021-09-01 2021-12-10 北京橙色云科技有限公司 Information searching method and device based on picture and storage medium
CN117173163A (en) * 2023-11-01 2023-12-05 浙江同花顺智能科技有限公司 Portrait quality assessment method, system, device and readable storage medium

Similar Documents

Publication Publication Date Title
CN111753060B (en) Information retrieval method, apparatus, device and computer readable storage medium
CN108959246B (en) Answer selection method and device based on improved attention mechanism and electronic equipment
EP2812883B1 (en) System and method for semantically annotating images
US10740545B2 (en) Information extraction from open-ended schema-less tables
US11954139B2 (en) Deep document processing with self-supervised learning
CN112084337A (en) Training method of text classification model, and text classification method and equipment
WO2010119615A1 (en) Learning-data generating device and named-entity-extraction system
CN107085583B (en) Electronic document management method and device based on content
CN109471944A (en) Training method, device and the readable storage medium storing program for executing of textual classification model
CN111190997A (en) Question-answering system implementation method using neural network and machine learning sequencing algorithm
EP2806336A1 (en) Text prediction in a text input associated with an image
Mueen et al. Automatic multilevel medical image annotation and retrieval
CN114461839A (en) Multi-mode pre-training-based similar picture retrieval method and device and electronic equipment
JP2011248596A (en) Searching system and searching method for picture-containing documents
CN113094538A (en) Image retrieval method, device and computer-readable storage medium
CN113434636A (en) Semantic-based approximate text search method and device, computer equipment and medium
CN111133429A (en) Extracting expressions for natural language processing
CN112632258A (en) Text data processing method and device, computer equipment and storage medium
Yurtsever et al. Figure search by text in large scale digital document collections
US20230138491A1 (en) Continuous learning for document processing and analysis
KR102215580B1 (en) Electronic device for selecting important keywords for documents based on style attributes and operating method thereof
US10628632B2 (en) Generating a structured document based on a machine readable document and artificial intelligence-generated annotations
US20200311059A1 (en) Multi-layer word search option
CN112149389A (en) Resume information structured processing method and device, computer equipment and storage medium
CN117216217B (en) Intelligent classification and retrieval method for files

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20210709