WO2013078307A1 - Image searching - Google Patents

Image searching Download PDF

Info

Publication number
WO2013078307A1
WO2013078307A1 PCT/US2012/066251 US2012066251W WO2013078307A1 WO 2013078307 A1 WO2013078307 A1 WO 2013078307A1 US 2012066251 W US2012066251 W US 2012066251W WO 2013078307 A1 WO2013078307 A1 WO 2013078307A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
label
search
images
description information
Prior art date
Application number
PCT/US2012/066251
Other languages
French (fr)
Inventor
Huaxing Jin
Original Assignee
Alibaba Group Holding Limited
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Alibaba Group Holding Limited filed Critical Alibaba Group Holding Limited
Priority to JP2014543557A priority Critical patent/JP6141305B2/en
Priority to EP12794621.8A priority patent/EP2783310A1/en
Publication of WO2013078307A1 publication Critical patent/WO2013078307A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/53Querying
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/5866Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, manually generated location and time information

Definitions

  • the present disclosure relates to the field of network search technology and, more specifically, to a method and an apparatus for searching images.
  • FIG. 1 illustrates a flowchart of an example method for establishing the visual feature based image searching database.
  • the feature extraction system defines visual feature templates to be extracted from images.
  • visual features are extracted from multiple images in the image searching database by utilizing the established feature extraction system.
  • An image may be represented by one or more visual features extracted from the image.
  • the visual features corresponding to each image are used as index keyword to establish an inverted index to generate the visual feature based image searching database.
  • the workflow of establishing the visual feature based image searching database is thus completed.
  • the visual feature based image searching database is established.
  • a server may perform image searching based on the visual feature based image searching database.
  • FIG. 2 illustrates a flowchart of an example method for searching images in the image searching database by the server.
  • the server receives an image searching request submitted by a client terminal.
  • the server may include the visual feature based image searching database or the server may access the visual feature based image searching database.
  • the image searching request includes an image for a search that a user requests.
  • the server extracts visual features from the image for the search received in the image searching request.
  • the method that the visual features are extracted from the image for search may be the same as the method that the visual features are extracted from the image when the visual feature based image searching data base is established.
  • the server based on the extracted visual features searches a set of images corresponding to the extracted visual feature from the visual feature based image searching database.
  • the server sends search results to the client terminal that sends the image searching request.
  • the operation that the server searches images is then complete.
  • the server Based on the operations from 204 to 208, the server returns a set of images requested by the user based on the visual feature based image searching database.
  • the above visual feature based image searching technology may be implemented through creating a new visual feature based searching engine instead of using text searching engines that servers with searching functions are normally equipped with. Accordingly, the server resources cannot be effectively utilized.
  • the present disclosure provides a method and an apparatus for image searching.
  • the present techniques may improve a utilization rate of server resources.
  • the present disclosure provides a method for image searching.
  • a respective label of a respective image is generated based on description information corresponding to the respective image.
  • a corresponding relationship between the generated respective label and the respective image is stored.
  • the method for searching image may include the following operations.
  • description information corresponding to an image for search in the image searching request is obtained.
  • the label of the image for search is generated.
  • the method for generating the label of the image for search is the same as or substantially similar to the method for generating the respective label of each image in the image searching database.
  • the present disclosure also provides an apparatus for image searching.
  • the apparatus may include a corresponding relationship generation unit, a description information retrieval unit, a label generation unit, an image determination unit, and an image supply unit.
  • the corresponding relationship generation unit with respect to each image in an image searching database, generates a respective label of a respective image based on description information corresponding to the respective image and stores a corresponding relationship between the generated respective label and the respective image.
  • the method for searching image may include the following operations.
  • the description information retrieval unit based on a received image searching request, obtains description information corresponding to an image for search in the image searching request.
  • the label generation unit based on the description information of the image for search obtained by the description information retrieval unit, generates the label of the image for search.
  • the method for generating the label of the image for search is the same as or substantially similar to the method for generating the respective label of each image in the image searching database.
  • the image determination unit determines one or more images corresponding to the label of the image for search generated by the label generation unit.
  • the image supply unit sends the determined one or more images to the client terminal that sends the image searching request.
  • At least one of the present techniques provided in the present disclosure may firstly, with respect to each image in the image searching database, generate the respective label of the respective image based on description information corresponding to the respective image and store a corresponding relationship between the generated respective label and the respective image.
  • the present techniques may, based on an image searching request, obtains description information corresponding to an image for search in the image searching request.
  • the present techniques may then, based on the description information of the image for search, generate the label of the image for search.
  • the method for generating the label of the image for search is the same as or substantially similar to the method for generating the label of each image in the image searching database.
  • the present techniques may, based on the stored corresponding relationship between the respective labels and the respective images, determine one or more images corresponding to the label of the image for search, and send the determined one or more images to the client terminal that sends the request for image searching.
  • the label of the corresponding image generated based on description information of the image may be in a form of text.
  • the label based image searching may utilize the existing text searching engine, thereby improving the utilization rate of the server resources.
  • FIG. 1 illustrates a flowchart of an example method for establishing a visual feature based image searching database.
  • FIG. 2 illustrates a flowchart of an example method for searching images in the image searching database by a server.
  • FIG. 3 illustrates a flowchart of an example method for generating a label for each image in the image searching database.
  • FIG. 4 illustrates a flowchart of another example method for generating the label for each image in the image searching database.
  • FIG. 5 illustrates a flowchart of another example method for generating the label corresponding to the image.
  • FIG. 6 illustrates a flowchart of an example method for generating the label corresponding to the image based on the keywords extracted from description information of the image.
  • FIG. 7 illustrates a flowchart of an example method for searching images performed by a server.
  • FIG. 8 illustrates a flowchart of another example method for searching images by the server.
  • FIG. 9 illustrates a flowchart of an example method for extract a text label of a product in accordance a second example embodiment of the present disclosure.
  • FIG. 10 illustrates a diagram of an example apparatus for searching images in accordance with a third example embodiment of the present disclosure.
  • FIG. 11 illustrates a diagram of another example apparatus for searching images in accordance with a third example embodiment of the present disclosure.
  • the present disclosure provides a method and an apparatus for searching images.
  • the present techniques may improve the utilization rate of server resources.
  • the present techniques may be applied in the process of image searching and be implemented either as a method or an apparatus.
  • the following is a detailed description of the present techniques by reference to the FIGs.
  • the described embodiments herein are example embodiments and should not be used to restrict the scope of the present disclosure.
  • the present disclosure provides a method for searching images in accordance with a first example embodiment of the present disclosure.
  • the method may be used in a server with information searching functionalities on the Internet to use a text searching engine of the server to conduct image searching, thereby improving the utilization rate of the server resources.
  • the method for searching images in the first example embodiment may pre-establish a label for each image in an image searching database as a basis for the server to search images.
  • the text searching engine in the server may thus be used to conduct image searching.
  • the following description describes two examples of establishing the label for each image.
  • FIG. 3 illustrates a flowchart of an example method for generating a label for each image in the image searching database.
  • a respective image is selected from the image searching database and description information of the respective image is obtained.
  • the process to obtain the description information of the respective image may be as follows. An entity corresponding to the respective image is determined and description information of the entity is used as the description information of the respective image. For instance, the description information of the respective image mainly describes attributes of the entity corresponding to the respective image.
  • the description information of the respective image may generally include the description information of the entity.
  • the entity corresponding to the respective image may include particular contents contained in the respective image.
  • the entity corresponding to the respective image may generally be a product.
  • the description information of the image may include various description information of the product, such as a product name, product attribute (such as actual color, actual size, etc.), product usage, product category, etc.
  • the respective label of the respective image is generated based on the description information of the obtained respective image.
  • the description information of the respective image is generally in a form of text.
  • the respective label generated based on the description information of the respective image is generally also in the form of text.
  • the respective label is mainly used to identify an attribute characteristic of the entity corresponding to the respective image. For instance, one label may correspond to different images.
  • the corresponding relationship between the generated respective labels and the respective images is stored. For instance, a method for indexing may be used to store the corresponding relationship to improve a speed of searching images.
  • the respective labels corresponding to the respective images in the image searching database may be used as an index.
  • An inverted indexing method may be used to store the corresponding relationship between the respective labels and the respective images. It should be understood that the methods described herein to store the respective labels and the respective image are just examples. Different storing methods may be flexibly set to store such corresponding relationship such as a table or a list and are not listed herein.
  • the determination whether there is an unlabeled image in the image searching database is to determine whether there is an unselected image in the image searching database.
  • the process to establish an index for images in the image searching database is thus completed.
  • the respective labels of corresponding respective images in the image searching database are generated and the corresponding relationship between the respective labels and the respective images is stored.
  • the following description describes a second example of establishing the label for each image.
  • one or more respective visual features in the respective image are further extracted.
  • a corresponding relationship between the respective visual features and the respective label is also stored.
  • the corresponding relationship between the respective label generated for the respective image, the respective image corresponding to the respective label, and the respective visual features extracted from the respective image is stored.
  • FIG. 4 illustrates a flowchart of another example method for generating a label for each image in the image searching database.
  • a respective image is selected from the image searching database and description information of the respective image is obtained.
  • the operations at 402 are substantially the same as or similar to the operations at 302 and are thus not detailed herein.
  • the label of the respective image is generated based on the description information of the obtained respective image.
  • the operations at 404 are substantially the same as or similar to the operations at 304 and are thus not detailed herein.
  • one or more image features of the respective image are extracted.
  • the image features of the respective image may be extracted based on a preset image feature template.
  • the image features may include visual features of the respective image or any other information representing the respective image such as a screenshot of the respective image.
  • the detailed extraction process may be implemented through image processing techniques.
  • the operations at 404 and 406 may have no sequence order such that the operations at 404 and 406 may be performed in parallel or the operations at 406 may be performed prior to the operations at 404.
  • the corresponding relationship between the respective label generated for the respective image, the respective one or more image features extracted from the respective image, and the respective image is stored.
  • the corresponding relationship between the respective label generated for the respective image, the respective one or more image features extracted from the respective image, and the respective image may be stored through indexing to increase the speed of searching images.
  • the respective labels corresponding to the respective images in the image searching database may be used as an index.
  • the inverted indexing method may be used to store the corresponding relationship between the respective label generated for the respective image, the respective one or more image features extracted from the respective image, and the respective image. It should be understood that the methods described herein to store the respective label and the respective image are just examples. Different storing methods may be flexibly set to store such corresponding relationship such as a table or a list and are not listed herein.
  • the respective label is generated and the respective image feature is extracted for each image in the image searching database, and the corresponding relationship between the respective generated label, the respective extracted image features and the respective image is stored.
  • the above first and second example methods for creating the respective label for each image in the image searching database may be implemented by the server with searching functionalities, the image searching database, or other entities with functionalities to access the image searching database.
  • the present disclosure further provides examples to implement the operations at 304 and operations at 404. Based on the obtained description information of the respective image, the respective label is generated for the respective image.
  • FIG. 5 illustrates a flowchart of another example method for generating the label corresponding to the image.
  • one or more keywords for a category are extracted from the description information of the respective image.
  • the keywords to be extracted from the obtained description information to generate the label may be set based on the categories of the respective entity corresponding to the respective image.
  • the description information of an entity with a category of product may include a product name, a product attribute, a product usage, a product classification, etc.
  • the product name and the product attribute may be set as keywords to be extracted from the description information of the respective image and be used to extract information corresponding to the product name and the product attribute. It should be understood that the above is just an example for extracting the above keywords for generation of the label from the description information corresponding to the respective image.
  • different keywords may be set to be extracted from the description information of the images, which are not detailed herein.
  • the respective label corresponding to the respective image is generated based on the extracted keywords.
  • an algorithm to generate the respective label of the respective image may be preset to generate the respective label based on the keywords extracted from the description information. In the applications, the algorithm may be flexibly set. The process to generate the label corresponding to the image is completed.
  • the first example embodiment of the present disclosure may provide more examples at 502 to generate the respective label corresponding to the respective image based on keywords extracted from the description information of the respective image.
  • FIG. 6 illustrates a flowchart of an example method for generating the respective label corresponding to the respective image based on the keywords extracted from the description information of the respective image.
  • a keyword is selected from the keywords extracted from the description information of the respective image.
  • the selected keyword may be randomly selected from all extracted keywords, or selected according to a preset order.
  • a category of the selected keyword is determined and a label vector corresponding to the keyword is generated based on an algorithm corresponding to the category of the keyword.
  • the algorithm corresponding to the category of the keyword may be flexibly set, such as a normalization of the extracted keyword, a sample processing of information corresponding to the keyword, etc.
  • the present disclosure does not impose restrictions.
  • each keyword extracted from the description information of the respective image is selected. If a result of the determination is positive, operations at 608 are performed. Otherwise, the process returns to operations at 602.
  • the respective label of the respective image is generated based on each label vector corresponding to each keyword and a weight value of each keyword's category. For example, when the respective label is generated, the label vector of each keyword is multiplied with its weight and products of respective label vectors with respective weights for all keywords are added to determine the respective label corresponding to the respective image. In another example, the respective label corresponding to the respective image is determined by the average value of respective label vectors of all keywords multiple with their respective weight.
  • the algorithm to generate the respective label corresponding to the respective image may be flexibly set based on the label vector corresponding to each keyword and the weight of each keyword's category. For instance, a respective weight of a respective category may be set as 1, which represents that the respective category does not require weighted processing.
  • FIG. 7 illustrates a flowchart of an example method for searching images performed by a server.
  • the server receives an image searching request from a client terminal.
  • the image searching request may generally include an image requested by a user for search (an image for search).
  • the server obtains corresponding information of the image for search included in the image searching request.
  • the operations at 704 that the server obtains the corresponding information of the image for search are generally the same as or similar to the operations at 302 to obtain the corresponding description information of the selected images. Thus details are not described herein.
  • a label of the image for search is generated based on the description information corresponding to the image for search.
  • the operations at 706 to generate the label of the image for search are generally the same as or similar to the operations at 304 to generate the labels of the images in the image searching database. Thus details are not described herein.
  • one or more images corresponding to the label of the image for search are determined based on the stored corresponding relationship between the labels and the images. For example, based on the label of the image for search, the stored corresponding relationship between the labels and the images is searched to find the one or more images corresponding to the label of the image for search.
  • the stored corresponding relationship between the labels and the images may be the stored corresponding relationship established in accordance with the first example of creating labels or the stored corresponding relationship established in accordance with the second example of creating labels.
  • the server sends the determined one or more images to the client terminal that sends the image searching request.
  • the server may process the determined images based on one or more preset rules.
  • the preset rules may be flexibly set up.
  • the preset rules may be set by the server or by the client terminal.
  • the preset rules may be a ranking of the determined images based on their resolutions from high to low or a ranking of the images based the attributes of the entities corresponding to the images from high to low. Details are not described herein.
  • the server may conduct searching images based on the label of the image for search. As the respective label is generated based on the description information corresponding to the respective image, the label is generally in the form of text.
  • the present text search engines of the server may be used to perform searching images, thereby improving the utilization rate of the server resources.
  • the corresponding relationship between the labels and the images for the image searching may be the corresponding relationship established in accordance with the first example method for creating labels or the corresponding relationship established in accordance with the second example method for example of creating labels.
  • the first example embodiment of the present disclosure provides a method for searching images.
  • the method for generating labels for the images in the image searching database may be based on the corresponding relationship generated in accordance with the second example of creating labels.
  • FIG. 8 illustrates a flowchart of another example method for searching images by the server.
  • the server receives an image searching request from a client terminal.
  • the image searching request may generally include an image for search.
  • the server obtains corresponding information of the image for search included in the image searching request.
  • the operations at 804 that the server obtains for the corresponding information of the image for search are generally the same as or similar to the operations at 302 to obtain the corresponding description information of the selected images. Thus details are not described herein.
  • a label of the image for search is generated based on the description information corresponding to the image for search.
  • the operations at 806 to generate the label of the image for search are generally the same as or similar to the operations at 304 to generate the labels of the images in the image searching database. Thus details are not described herein.
  • the server extracts one or more image features of the image for search.
  • the operations at 808 to extract the image features from the image for search are generally the same as or similar to the operations at 406 to generate the image features of the respective image. Thus details are not described herein. In addition, the operations at 808 may be performed prior to the operations at 804 or 806.
  • the server performs a similarity degree comparison between the extracted image features of the image for search and the image features of one or more determined images respectively.
  • the one or more image corresponding to the label of the image for search may be determined based on the stored corresponding relationship between the labels and the images.
  • the determined images are ranked from high to low based on their similarity degrees between the determined images and the image for search respectively.
  • the server sends the determined images to the client terminal that sends the image searching request.
  • the operations at 814 are generally the same as or similar to the operations at 710. Thus details are not described herein.
  • the server may conduct searching images based on the label of the image for search.
  • the label is generally in the form of text.
  • the present text search engines of the server may be used to perform searching images, thereby improving the utilization rate of the server resources.
  • the images most relevant to the image search may be represented to the client with priority.
  • a second example embodiment of the present disclosure provides an example application scenario to generate the labels.
  • An e-commerce website may have an information database with a large volume of product information. Each product may have description information in a form of text.
  • the text description information has a relatively strong correlation with the image of the product as they are both descriptions of the product.
  • the text description information of the product may be equivalent to the description information of the image of the product.
  • the text label may represent contents of the image.
  • the product image searching may search the relevant images based on the text labels and then rank the images based on their similarity degrees.
  • the text description information corresponds to the product and may also include labels manually added after human review.
  • This example embodiment of the present disclosure extracts keywords from such description information to generate text labels.
  • these text labels may be ranked according to their relativity degrees.
  • a set of labels including labels of different dimensions of the product form a vector space.
  • the dimensions may be the categories set at 502 as shown in FIG. 5 of the first example embodiment of the present disclosure. For the convenience of the following description, some symbols are defined as below.
  • T represents a text label of the product.
  • Tl represents a label generated based on a label manually added for the product.
  • T2 represents a label generated based on brief description in the product.
  • T3 represents a label generated based on a keyword of the product.
  • T4 represents a label generated based on a classification of the product.
  • the label manually added for the product, the brief description in the product, the keyword of the product, the classification of the product, and the attribute of the product correspond to the categories set up at 502.
  • FIG. 9 illustrates a flowchart of an example method for extract a text label of a product in accordance with the second example embodiment of the present disclosure.
  • a manual label is input.
  • the user's filled-in information for products on the e-commerce website are often informal and often cause errors by machine identification.
  • the manual label is input into the system as a correction to automatic labeling by machine.
  • a product may have multiple labels which are input based on their relativities from high to low.
  • the product ID may be used as the index key and the set or sequence of the labels may be used as the index values.
  • Such index keys and index values may be stored in a high performance searching data structure, such as a Trie database.
  • rank represents a ranking of a respective label according to relativities of the labels. Based on the above formula, the ranking is normalized to the label Tl .
  • the label T2 is generated based on the brief descriptions of the product.
  • the operations at 906 may correspond to the operations at 604.
  • the brief description of the product may be short.
  • Natural Language Processing (NPL) may be used to pre-process the brief description to obtain three kinds of words (core word, noun, and other words). Each word may be used as a label.
  • the sequence of importance of the words is: core word > noun > other word.
  • a weight value may be set for each type based on experiences.
  • Label T2 category weight * appearance of word frequency of respective word in such category
  • the label T2 is then obtained through normalization.
  • the label T3 is generated based on the keyword of the product.
  • the operations at 908 correspond to the operations at 604.
  • the user may fill in multiple keywords.
  • Each keyword may have a short length.
  • the sequence of filling-in is treated as an importance sequence from high to low based on the statistics.
  • the formula of calculation of importance may be as follows:
  • Label T3 ranking weight * category weight * appearance of word frequency of respective word.
  • the label T3 is then obtained through normalization.
  • n 1, 2 ...n, wherein n may be any integer. The smaller the value the more detailed the grain is.
  • the label T4 is obtained through such normalization.
  • the label T5 is generated based on the attribute of the product.
  • the operations at 912 may correspond to the operations at 604.
  • the product may have multiple attribute dimensions. Each attribute dimension corresponds to a respective label.
  • the importance score of each label may be the same, and normalized to obtain the label vector T5.
  • operations from 904 to 912 may be independent.
  • the operation number may be used to differentiate each other.
  • the corresponding relationship between the label T and the product is stored.
  • the label corresponding to the image is determined and the image feature is extracted, their corresponding relationship with the image may be saved. For example, an index may be established.
  • the text label and the image features may be included into the index.
  • the text label may be used as a key of an inverted index and the image content together with the product ID may be used as a value of the index.
  • the text label generation method as described in this example embodiment of the present disclosure may be applicable to a scenario where each entity corresponds to an image and text description information and the text description information and the image both describe a single concept.
  • a concept is a product.
  • the user may click an image.
  • the serve may respond by searching related text description information, generating text labels, using the text searching engine to search related products based on the text labels, and ranking the found images based on similarities.
  • FIG. 10 illustrates a diagram of an example apparatus for searching images in accordance with the third example embodiment of the present disclosure.
  • the apparatus may include software, hardware, or a combination thereof.
  • the apparatus 1000 in FIG. 10 may include one or more processor(s) 1002 and memory 1004.
  • the memory 1004 is an example of computer storage media.
  • Computer storage media includes volatile and non-volatile, removable and nonremovable media implemented in any method or technology for storage of information such as computer-executable instructions, data structures, program modules, or other data.
  • Examples of computer storage media includes, but is not limited to, phase change memory (PRAM), static random-access memory (SRAM), dynamic random-access memory (DRAM), other types of random-access memory (RAM), read-only memory (ROM), electrically erasable programmable read-only memory (EEPROM), flash memory or other memory technology, compact disk read-only memory (CD-ROM), digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other non-transmission medium that can be used to store information for access by a computing device.
  • computer storage media does not include transitory media such as modulated data signals and carrier waves.
  • the memory 1004 may store therein program units or modules and program data.
  • the memory 1004 may store therein a corresponding relationship generation unit 1006, a description information obtaining unit 1008, a label generation unit 1010, an image determination unit 1012, and an image supply unit 1014.
  • the description information obtaining unit 1008 based on an image searching request, obtains corresponding description information of an image for search included in the image searching request.
  • the label generation unit 1010 based on the description information corresponding to the image for search obtained by the description information obtaining unit 1008, generates a corresponding label of the image for search.
  • the method for generating the label of the image for search may be the same as the method for generating the labels of the images in the image searching database by the corresponding relationship generation unit 1006.
  • the image determination unit 1012 determines one or more images that correspond to the label of the image for search generated by the label generation unit 1010.
  • the image supply unit 1014 sends the one or more images determined by the image determination unit 1012 to a client terminal that sends the image searching request.
  • the label generation unit 1010 may, based on the extracted keyword of each defined category, use an algorithm corresponding to the defined category to generate a label vector corresponding to the keyword, and generate the label of the image for search based on the label vector corresponding to each keyword and a weight value of each category of the keyword.
  • the corresponding relationship generation unit 1006 may use the generated label of each image as the index and use an inverted index to store the corresponding relationship between the labels and the images.
  • the corresponding relationship generation unit 1006 may extract image features of each image in the image searching database and store the corresponding relationship between the stored label, the image corresponding to the label, and the image feature extracted from the image.
  • the apparatus 1000 may further include an image ranking unit 1016.
  • the image ranking unit 1016 before the image supply unit 1014 sends the determined images to the client terminal that sends the image searching request, based on the received image searching request, extracts image features of the image for search included in the image searching request.
  • the image ranking unit 1016 compares the image features of the image for search with image features corresponding to each of the determined images, and ranks the determined images according to their similarities degrees with the image for search from high to low.
  • apparatus for searching images as described in the third example embodiment of the present disclosure may further include multiple units to implement respective functionality as described in the above first example embodiment of the present disclosure. Such classifications of units are not described herein.
  • each flow and/or block and the combination of the flow and/or block of the flowchart and/or block diagram can be implemented by computer program instructions.
  • These computer program instructions can be provided to the general computers, specific computers, embedded processor or other programmable data processors to generate a machine, so that a device of implementing one or more flows of the flow chart and/or one or more blocks of the block diagram can be generated through the instructions operated by a computer or other programmable data processors.
  • These computer program instructions can also be loaded in a computer or other programmable data processors, so that the computer or other programmable data processors can operate a series of operation steps to generate the process implemented by a computer. Accordingly, the instructions operated in the computer or other programmable data processors can provides the steps for implementing the functions specified in one or more flows of the flow chart and/or one or more blocks of the block diagram.

Abstract

The present disclosure introduces a method and an apparatus for searching images. With respect to each image in an image searching database, respective labels of respective images are generated based on description information corresponding to the respective images. A corresponding relationship between the generated respective labels and the respective images is stored. Based on a received image searching request, description information corresponding to an image for search in the image searching request is obtained. Based on the description information of the image for search, the label of the image for search is generated. Based on the stored corresponding relationship between the respective labels and the respective images, one or more images corresponding to the label of the image for search are determined. The determined one or more images are sent to the client terminal that sends the image searching request.

Description

IMAGE SEARCHING
CROSS REFERENCE TO RELATED PATENT APPLICATIONS
This application claims foreign priority to Chinese Patent Application No. 201110380442.X filed on 25 November 2011, entitled "Image Searching Method and Apparatus," which is hereby incorporated by reference in its entirety.
TECHNICAL FIELD
The present disclosure relates to the field of network search technology and, more specifically, to a method and an apparatus for searching images.
BACKGROUND
With the application and development of image searching technology, the image searching technology based on visual feature is the main stream image searching technology. Implementation of the image searching technology based on visual features requires pre- establishing a visual feature based image searching database. Specifically, FIG. 1 illustrates a flowchart of an example method for establishing the visual feature based image searching database.
At 102, a feature extraction system is established. The feature extraction system defines visual feature templates to be extracted from images.
At 104, visual features are extracted from multiple images in the image searching database by utilizing the established feature extraction system. An image may be represented by one or more visual features extracted from the image.
At 106, the visual features corresponding to each image are used as index keyword to establish an inverted index to generate the visual feature based image searching database. The workflow of establishing the visual feature based image searching database is thus completed.
Based on the operations from 102 to 106, the visual feature based image searching database is established. A server may perform image searching based on the visual feature based image searching database. FIG. 2 illustrates a flowchart of an example method for searching images in the image searching database by the server.
At 202, the server receives an image searching request submitted by a client terminal. At 202, the server may include the visual feature based image searching database or the server may access the visual feature based image searching database. The image searching request includes an image for a search that a user requests.
At 204, the server extracts visual features from the image for the search received in the image searching request. At 204, the method that the visual features are extracted from the image for search may be the same as the method that the visual features are extracted from the image when the visual feature based image searching data base is established.
At 206, the server based on the extracted visual features, according to the established invert index, searches a set of images corresponding to the extracted visual feature from the visual feature based image searching database.
At 208, the server sends search results to the client terminal that sends the image searching request. The operation that the server searches images is then complete.
Based on the operations from 204 to 208, the server returns a set of images requested by the user based on the visual feature based image searching database.
In actual applications, the above visual feature based image searching technology may be implemented through creating a new visual feature based searching engine instead of using text searching engines that servers with searching functions are normally equipped with. Accordingly, the server resources cannot be effectively utilized. SUMMARY
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify all key features or essential features of the claimed subject matter, nor is it alone intended to be used as an aid in determining the scope of the claimed subject matter. The term "techniques," for instance, may refer to device(s), system(s), method(s) and/or computer-readable instructions as permitted by the context above and throughout the present disclosure.
The present disclosure provides a method and an apparatus for image searching. The present techniques may improve a utilization rate of server resources.
The present disclosure provides a method for image searching. With respect to each image in an image searching database, a respective label of a respective image is generated based on description information corresponding to the respective image. A corresponding relationship between the generated respective label and the respective image is stored. The method for searching image may include the following operations.
Based on a received image searching request, description information corresponding to an image for search in the image searching request is obtained.
Based on the description information of the image for search, the label of the image for search is generated. The method for generating the label of the image for search is the same as or substantially similar to the method for generating the respective label of each image in the image searching database.
Based on the stored corresponding relationship between the respective labels and the respective images, one or more images corresponding to the label of the image for search are determined. The determined one or more images are sent to the client terminal that sends the image searching request. The present disclosure also provides an apparatus for image searching. The apparatus may include a corresponding relationship generation unit, a description information retrieval unit, a label generation unit, an image determination unit, and an image supply unit.
The corresponding relationship generation unit, with respect to each image in an image searching database, generates a respective label of a respective image based on description information corresponding to the respective image and stores a corresponding relationship between the generated respective label and the respective image. The method for searching image may include the following operations.
The description information retrieval unit, based on a received image searching request, obtains description information corresponding to an image for search in the image searching request.
The label generation unit, based on the description information of the image for search obtained by the description information retrieval unit, generates the label of the image for search. The method for generating the label of the image for search is the same as or substantially similar to the method for generating the respective label of each image in the image searching database.
The image determination unit, based on the stored corresponding relationship between the respective labels and the respective images, determines one or more images corresponding to the label of the image for search generated by the label generation unit. The image supply unit sends the determined one or more images to the client terminal that sends the image searching request.
At least one of the present techniques provided in the present disclosure may firstly, with respect to each image in the image searching database, generate the respective label of the respective image based on description information corresponding to the respective image and store a corresponding relationship between the generated respective label and the respective image. The present techniques may, based on an image searching request, obtains description information corresponding to an image for search in the image searching request. The present techniques may then, based on the description information of the image for search, generate the label of the image for search. The method for generating the label of the image for search is the same as or substantially similar to the method for generating the label of each image in the image searching database.
Further, the present techniques may, based on the stored corresponding relationship between the respective labels and the respective images, determine one or more images corresponding to the label of the image for search, and send the determined one or more images to the client terminal that sends the request for image searching.
In the present techniques, the label of the corresponding image generated based on description information of the image may be in a form of text. Thus, the label based image searching may utilize the existing text searching engine, thereby improving the utilization rate of the server resources.
BRIEF DESCRIPTION OF THE DRAWINGS
To better illustrate embodiments of the present disclosure, the following is a brief introduction of figures to be used in descriptions of the embodiments. It is apparent that the following figures only relate to some embodiments of the present disclosure and shall not be used to restrict the present disclosure. A person of ordinary skill in the art can obtain other figures according to the figures in the present disclosure without creative efforts.
FIG. 1 illustrates a flowchart of an example method for establishing a visual feature based image searching database.
FIG. 2 illustrates a flowchart of an example method for searching images in the image searching database by a server.
FIG. 3 illustrates a flowchart of an example method for generating a label for each image in the image searching database.
FIG. 4 illustrates a flowchart of another example method for generating the label for each image in the image searching database.
FIG. 5 illustrates a flowchart of another example method for generating the label corresponding to the image.
FIG. 6 illustrates a flowchart of an example method for generating the label corresponding to the image based on the keywords extracted from description information of the image.
FIG. 7 illustrates a flowchart of an example method for searching images performed by a server.
FIG. 8 illustrates a flowchart of another example method for searching images by the server.
FIG. 9 illustrates a flowchart of an example method for extract a text label of a product in accordance a second example embodiment of the present disclosure. FIG. 10 illustrates a diagram of an example apparatus for searching images in accordance with a third example embodiment of the present disclosure.
FIG. 11 illustrates a diagram of another example apparatus for searching images in accordance with a third example embodiment of the present disclosure.
DETAILED DESCRIPTION
The present disclosure provides a method and an apparatus for searching images. The present techniques may improve the utilization rate of server resources. The present techniques may be applied in the process of image searching and be implemented either as a method or an apparatus. The following is a detailed description of the present techniques by reference to the FIGs. The described embodiments herein are example embodiments and should not be used to restrict the scope of the present disclosure.
The present disclosure provides a method for searching images in accordance with a first example embodiment of the present disclosure. The method may be used in a server with information searching functionalities on the Internet to use a text searching engine of the server to conduct image searching, thereby improving the utilization rate of the server resources.
The method for searching images in the first example embodiment may pre-establish a label for each image in an image searching database as a basis for the server to search images. The text searching engine in the server may thus be used to conduct image searching. The following description describes two examples of establishing the label for each image.
In one example of establishing the label for each image, a respective label is established for each image in the image searching database and a corresponding relationship between respective labels and respective image is stored. FIG. 3 illustrates a flowchart of an example method for generating a label for each image in the image searching database. At 302, a respective image is selected from the image searching database and description information of the respective image is obtained. The process to obtain the description information of the respective image may be as follows. An entity corresponding to the respective image is determined and description information of the entity is used as the description information of the respective image. For instance, the description information of the respective image mainly describes attributes of the entity corresponding to the respective image. The description information of the respective image may generally include the description information of the entity. The entity corresponding to the respective image may include particular contents contained in the respective image. For instance, in the e- commerce field, the entity corresponding to the respective image may generally be a product. Correspondingly, the description information of the image may include various description information of the product, such as a product name, product attribute (such as actual color, actual size, etc.), product usage, product category, etc.
At 304, the respective label of the respective image is generated based on the description information of the obtained respective image. At 302, the description information of the respective image is generally in a form of text. Correspondingly, the respective label generated based on the description information of the respective image is generally also in the form of text. The respective label is mainly used to identify an attribute characteristic of the entity corresponding to the respective image. For instance, one label may correspond to different images.
At 306, the corresponding relationship between the generated respective labels and the respective images is stored. For instance, a method for indexing may be used to store the corresponding relationship to improve a speed of searching images. The respective labels corresponding to the respective images in the image searching database may be used as an index. An inverted indexing method may be used to store the corresponding relationship between the respective labels and the respective images. It should be understood that the methods described herein to store the respective labels and the respective image are just examples. Different storing methods may be flexibly set to store such corresponding relationship such as a table or a list and are not listed herein. At 308, it is determined whether there is an image without a generated label in the image searching database. If a result of determination is positive, the process returns to operations at 302. Otherwise, the process is terminated.
At 308, the determination whether there is an unlabeled image in the image searching database is to determine whether there is an unselected image in the image searching database. The process to establish an index for images in the image searching database is thus completed.
Through the operations of 302 to 308, the respective labels of corresponding respective images in the image searching database are generated and the corresponding relationship between the respective labels and the respective images is stored.
The following description describes a second example of establishing the label for each image. In addition to generating the label for each image in the image searching database, one or more respective visual features in the respective image are further extracted. When the corresponding relationship between the respective label and the respective image is stored, a corresponding relationship between the respective visual features and the respective label is also stored. The corresponding relationship between the respective label generated for the respective image, the respective image corresponding to the respective label, and the respective visual features extracted from the respective image is stored.
FIG. 4 illustrates a flowchart of another example method for generating a label for each image in the image searching database.
At 402, a respective image is selected from the image searching database and description information of the respective image is obtained. The operations at 402 are substantially the same as or similar to the operations at 302 and are thus not detailed herein.
At 404, the label of the respective image is generated based on the description information of the obtained respective image. The operations at 404 are substantially the same as or similar to the operations at 304 and are thus not detailed herein.
At 406, one or more image features of the respective image are extracted. The image features of the respective image may be extracted based on a preset image feature template. The image features may include visual features of the respective image or any other information representing the respective image such as a screenshot of the respective image. The detailed extraction process may be implemented through image processing techniques.
In one example, the operations at 404 and 406 may have no sequence order such that the operations at 404 and 406 may be performed in parallel or the operations at 406 may be performed prior to the operations at 404.
At 408, the corresponding relationship between the respective label generated for the respective image, the respective one or more image features extracted from the respective image, and the respective image is stored. At 408, the corresponding relationship between the respective label generated for the respective image, the respective one or more image features extracted from the respective image, and the respective image may be stored through indexing to increase the speed of searching images. For instance, the respective labels corresponding to the respective images in the image searching database may be used as an index. The inverted indexing method may be used to store the corresponding relationship between the respective label generated for the respective image, the respective one or more image features extracted from the respective image, and the respective image. It should be understood that the methods described herein to store the respective label and the respective image are just examples. Different storing methods may be flexibly set to store such corresponding relationship such as a table or a list and are not listed herein.
At 410, it is determined whether there is an image without a generated label in the image searching database. If a result of the determination is positive, the process returns to the operations at 402. Otherwise, the process is terminated. The process to establish the index for each image in the image searching database is completed.
Through the operations of 402 to 410, the respective label is generated and the respective image feature is extracted for each image in the image searching database, and the corresponding relationship between the respective generated label, the respective extracted image features and the respective image is stored.
The above first and second example methods for creating the respective label for each image in the image searching database may be implemented by the server with searching functionalities, the image searching database, or other entities with functionalities to access the image searching database.
The present disclosure further provides examples to implement the operations at 304 and operations at 404. Based on the obtained description information of the respective image, the respective label is generated for the respective image. FIG. 5 illustrates a flowchart of another example method for generating the label corresponding to the image.
At 502, one or more keywords for a category are extracted from the description information of the respective image. For example, the keywords to be extracted from the obtained description information to generate the label may be set based on the categories of the respective entity corresponding to the respective image. For instance, the description information of an entity with a category of product may include a product name, a product attribute, a product usage, a product classification, etc. The product name and the product attribute may be set as keywords to be extracted from the description information of the respective image and be used to extract information corresponding to the product name and the product attribute. It should be understood that the above is just an example for extracting the above keywords for generation of the label from the description information corresponding to the respective image. In some examples, based on a specific type of the respective entity corresponding to the respective image, different keywords may be set to be extracted from the description information of the images, which are not detailed herein.
At 504, the respective label corresponding to the respective image is generated based on the extracted keywords. At 504, an algorithm to generate the respective label of the respective image may be preset to generate the respective label based on the keywords extracted from the description information. In the applications, the algorithm may be flexibly set. The process to generate the label corresponding to the image is completed.
The first example embodiment of the present disclosure may provide more examples at 502 to generate the respective label corresponding to the respective image based on keywords extracted from the description information of the respective image. FIG. 6 illustrates a flowchart of an example method for generating the respective label corresponding to the respective image based on the keywords extracted from the description information of the respective image.
At 602, a keyword is selected from the keywords extracted from the description information of the respective image. For example, at 602, the selected keyword may be randomly selected from all extracted keywords, or selected according to a preset order.
At 604, a category of the selected keyword is determined and a label vector corresponding to the keyword is generated based on an algorithm corresponding to the category of the keyword. For example, the algorithm corresponding to the category of the keyword may be flexibly set, such as a normalization of the extracted keyword, a sample processing of information corresponding to the keyword, etc. The present disclosure does not impose restrictions.
At 606, it is determined whether each keyword extracted from the description information of the respective image is selected. If a result of the determination is positive, operations at 608 are performed. Otherwise, the process returns to operations at 602. At 608, the respective label of the respective image is generated based on each label vector corresponding to each keyword and a weight value of each keyword's category. For example, when the respective label is generated, the label vector of each keyword is multiplied with its weight and products of respective label vectors with respective weights for all keywords are added to determine the respective label corresponding to the respective image. In another example, the respective label corresponding to the respective image is determined by the average value of respective label vectors of all keywords multiple with their respective weight. In some other examples, the algorithm to generate the respective label corresponding to the respective image may be flexibly set based on the label vector corresponding to each keyword and the weight of each keyword's category. For instance, a respective weight of a respective category may be set as 1, which represents that the respective category does not require weighted processing.
The process to extract the keywords from the corresponding description information of the image to generate the label of the image is completed.
After the labels are generated for the images in the image searching database and the corresponding relationships between the labels and the images are stored in accordance with techniques in the above example embodiments, the server may perform image searching based on the corresponding relationships. FIG. 7 illustrates a flowchart of an example method for searching images performed by a server.
At 702, the server receives an image searching request from a client terminal. The image searching request may generally include an image requested by a user for search (an image for search).
At 704, based on the received image searching request the server obtains corresponding information of the image for search included in the image searching request. The operations at 704 that the server obtains the corresponding information of the image for search are generally the same as or similar to the operations at 302 to obtain the corresponding description information of the selected images. Thus details are not described herein.
At 706, a label of the image for search is generated based on the description information corresponding to the image for search. The operations at 706 to generate the label of the image for search are generally the same as or similar to the operations at 304 to generate the labels of the images in the image searching database. Thus details are not described herein.
At 708, one or more images corresponding to the label of the image for search are determined based on the stored corresponding relationship between the labels and the images. For example, based on the label of the image for search, the stored corresponding relationship between the labels and the images is searched to find the one or more images corresponding to the label of the image for search. The stored corresponding relationship between the labels and the images may be the stored corresponding relationship established in accordance with the first example of creating labels or the stored corresponding relationship established in accordance with the second example of creating labels.
At 710, the server sends the determined one or more images to the client terminal that sends the image searching request. Before the server sends the determined images to the client terminal at 710, the server may process the determined images based on one or more preset rules. The preset rules may be flexibly set up. For example, the preset rules may be set by the server or by the client terminal. The preset rules may be a ranking of the determined images based on their resolutions from high to low or a ranking of the images based the attributes of the entities corresponding to the images from high to low. Details are not described herein. Based on the above process, the server may conduct searching images based on the label of the image for search. As the respective label is generated based on the description information corresponding to the respective image, the label is generally in the form of text. Thus, the present text search engines of the server may be used to perform searching images, thereby improving the utilization rate of the server resources.
In the above image searching process, the corresponding relationship between the labels and the images for the image searching may be the corresponding relationship established in accordance with the first example method for creating labels or the corresponding relationship established in accordance with the second example method for example of creating labels.
The first example embodiment of the present disclosure provides a method for searching images. The method for generating labels for the images in the image searching database may be based on the corresponding relationship generated in accordance with the second example of creating labels. FIG. 8 illustrates a flowchart of another example method for searching images by the server.
At 802, the server receives an image searching request from a client terminal. The image searching request may generally include an image for search.
At 804, based on the received image searching request, the server obtains corresponding information of the image for search included in the image searching request. The operations at 804 that the server obtains for the corresponding information of the image for search are generally the same as or similar to the operations at 302 to obtain the corresponding description information of the selected images. Thus details are not described herein.
At 806, a label of the image for search is generated based on the description information corresponding to the image for search. The operations at 806 to generate the label of the image for search are generally the same as or similar to the operations at 304 to generate the labels of the images in the image searching database. Thus details are not described herein.
At 808, the server extracts one or more image features of the image for search. The operations at 808 to extract the image features from the image for search are generally the same as or similar to the operations at 406 to generate the image features of the respective image. Thus details are not described herein. In addition, the operations at 808 may be performed prior to the operations at 804 or 806.
At 810, the server performs a similarity degree comparison between the extracted image features of the image for search and the image features of one or more determined images respectively. For example, the one or more image corresponding to the label of the image for search may be determined based on the stored corresponding relationship between the labels and the images.
At 812, the determined images are ranked from high to low based on their similarity degrees between the determined images and the image for search respectively.
At 814, the server sends the determined images to the client terminal that sends the image searching request. The operations at 814 are generally the same as or similar to the operations at 710. Thus details are not described herein.
Based on the above process, the server may conduct searching images based on the label of the image for search. As the respective label is generated based on the description information corresponding to the respective image, the label is generally in the form of text. Thus, the present text search engines of the server may be used to perform searching images, thereby improving the utilization rate of the server resources. In addition, as the return result to the client terminal is processed based on the ranking of similarity degrees, the images most relevant to the image search may be represented to the client with priority. A second example embodiment of the present disclosure provides an example application scenario to generate the labels. An e-commerce website may have an information database with a large volume of product information. Each product may have description information in a form of text. The text description information has a relatively strong correlation with the image of the product as they are both descriptions of the product. The text description information of the product may be equivalent to the description information of the image of the product. Thus, the text label may represent contents of the image. The product image searching may search the relevant images based on the text labels and then rank the images based on their similarity degrees.
The text description information (such as the brief description, keyword, classification, attribute, etc.) corresponds to the product and may also include labels manually added after human review. This example embodiment of the present disclosure extracts keywords from such description information to generate text labels. In some example, these text labels may be ranked according to their relativity degrees. In one example, a set of labels including labels of different dimensions of the product form a vector space. The dimensions may be the categories set at 502 as shown in FIG. 5 of the first example embodiment of the present disclosure. For the convenience of the following description, some symbols are defined as below.
T represents a text label of the product.
Tl represents a label generated based on a label manually added for the product.
T2 represents a label generated based on brief description in the product.
T3 represents a label generated based on a keyword of the product.
T4 represents a label generated based on a classification of the product.
T5 represents a label generated based on an attribute of the product.
wl, w2, w3, w4, w5 represents a respective weigh of Tl, T2, T3, T4, and T5. The text label of the product may be calculated based on the following formula:
T=wl *T 1+W2*T2+W3 *T3+w4*T4+w5 *T5
The label manually added for the product, the brief description in the product, the keyword of the product, the classification of the product, and the attribute of the product correspond to the categories set up at 502.
FIG. 9 illustrates a flowchart of an example method for extract a text label of a product in accordance with the second example embodiment of the present disclosure.
At 902, a manual label is input. For various reasons, the user's filled-in information for products on the e-commerce website are often informal and often cause errors by machine identification. Thus, there are labels marked manually by humans after they review the product information. The manual label is input into the system as a correction to automatic labeling by machine. A product may have multiple labels which are input based on their relativities from high to low. For example, the product ID may be used as the index key and the set or sequence of the labels may be used as the index values. Such index keys and index values may be stored in a high performance searching data structure, such as a Trie database.
At 904, the manual label is searched according to the product ID and the label Tl is generated based on the manual label. The operations at 904 may correspond to operations at 604. For example, when the manual labels are input, the manual labels have been ranked from high to low according to their relativities. The calculation formula of Tl may be as follows:
Tl=l/log (rank + 1),
wherein rank represents a ranking of a respective label according to relativities of the labels. Based on the above formula, the ranking is normalized to the label Tl .
At 906, the label T2 is generated based on the brief descriptions of the product. The operations at 906 may correspond to the operations at 604. For example, the brief description of the product may be short. Natural Language Processing (NPL) may be used to pre-process the brief description to obtain three kinds of words (core word, noun, and other words). Each word may be used as a label. The sequence of importance of the words is: core word > noun > other word. A weight value may be set for each type based on experiences.
Label T2 = category weight * appearance of word frequency of respective word in such category
The label T2 is then obtained through normalization.
At 908, the label T3 is generated based on the keyword of the product. The operations at 908 correspond to the operations at 604. For example, the user may fill in multiple keywords. Each keyword may have a short length. The sequence of filling-in is treated as an importance sequence from high to low based on the statistics. For example, the formula of calculation of importance may be as follows:
1 /log (rank + 1)
Natural Language Processing (NPL) may be used to pre-process each keyword to obtain three kinds of words (core word, noun, and other words). Each word may be used as a label. The sequence of importance of the words is: core word > noun > other word. A weight value may be set for each type based on experiences.
Label T3 = ranking weight * category weight * appearance of word frequency of respective word.
The label T3 is then obtained through normalization.
At 910, the label T4 is generated based on the classification of the product.
The operations at 910 may correspond to the operations at 604. For example, each product belongs to a classification system. The classification system has a hierarchy. The respective labels may be extracted from the classification name of each level of hierarchy. The more detailed the classification the more important the respective label. For example, the calculation formula of importance level of the label in the hierarchy may be as follows:
1/log (grain + 1)
The grain = 1, 2 ...n, wherein n may be any integer. The smaller the value the more detailed the grain is.
Label T4 = importance score of a hierarchy level * appearance of word frequency of respective word.
The label T4 is obtained through such normalization.
At 912, the label T5 is generated based on the attribute of the product. The operations at 912 may correspond to the operations at 604. For example, the product may have multiple attribute dimensions. Each attribute dimension corresponds to a respective label. The importance score of each label may be the same, and normalized to obtain the label vector T5.
For example, in the above operations, except that operations at 902 and 904 may have sequence order, operations from 904 to 912 may be independent. The operation number may be used to differentiate each other.
At 914, the text label T of the product is generated based on various labels.
The operations at 914 may correspond to operations at 606. For example, the label T may be determined based on the above formula:
T=wl *T 1+W2*T2+W3 *T3+w4*T4+w5 *T5
After the label T of the product is determined, the corresponding relationship between the label T and the product is stored.
As shown in FIG. 9, the source of keywords that generate the labels Tl, T2, T3, T4, and T5 may come from description information of the product 916. The manual label input at 902 may be stored at a manual label database 918. Tl is generated based on one or more manual labels from the manual label database 918. In addition, based on the above processing, the image features may be further extracted from the product image. For example, the image processing techniques may be used to extract image features (such as a shape feature, a color feature, a portion feature, a distributive feature, etc.) The image processing techniques may be referenced to the conventional image processing techniques.
After the label corresponding to the image is determined and the image feature is extracted, their corresponding relationship with the image may be saved. For example, an index may be established. The text label and the image features may be included into the index. The text label may be used as a key of an inverted index and the image content together with the product ID may be used as a value of the index.
The text label generation method as described in this example embodiment of the present disclosure may be applicable to a scenario where each entity corresponds to an image and text description information and the text description information and the image both describe a single concept. In the example of an e-commerce file, such a concept is a product. In applications, for example, the user may click an image. The serve may respond by searching related text description information, generating text labels, using the text searching engine to search related products based on the text labels, and ranking the found images based on similarities.
A third example embodiment of the present disclosure provides an apparatus for searching images. For example, the apparatus for searching images may be located at a searching server. The apparatus for searching images at the searching server may improve the resource utilization rate of the server.
FIG. 10 illustrates a diagram of an example apparatus for searching images in accordance with the third example embodiment of the present disclosure. The apparatus may include software, hardware, or a combination thereof. In an example configuration, the apparatus 1000 in FIG. 10 may include one or more processor(s) 1002 and memory 1004. The memory 1004 is an example of computer storage media.
Computer storage media includes volatile and non-volatile, removable and nonremovable media implemented in any method or technology for storage of information such as computer-executable instructions, data structures, program modules, or other data. Examples of computer storage media includes, but is not limited to, phase change memory (PRAM), static random-access memory (SRAM), dynamic random-access memory (DRAM), other types of random-access memory (RAM), read-only memory (ROM), electrically erasable programmable read-only memory (EEPROM), flash memory or other memory technology, compact disk read-only memory (CD-ROM), digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other non-transmission medium that can be used to store information for access by a computing device. As defined herein, computer storage media does not include transitory media such as modulated data signals and carrier waves.
The memory 1004 may store therein program units or modules and program data. In the example of FIG. 10, the memory 1004 may store therein a corresponding relationship generation unit 1006, a description information obtaining unit 1008, a label generation unit 1010, an image determination unit 1012, and an image supply unit 1014.
The corresponding relationship generation unit 1006, with respect to each image in an image searching database, generates a respective label of a respective image based on description information of the respective image, stores a corresponding relationship between the respective label and the respective image.
The description information obtaining unit 1008, based on an image searching request, obtains corresponding description information of an image for search included in the image searching request. The label generation unit 1010, based on the description information corresponding to the image for search obtained by the description information obtaining unit 1008, generates a corresponding label of the image for search. The method for generating the label of the image for search may be the same as the method for generating the labels of the images in the image searching database by the corresponding relationship generation unit 1006.
The image determination unit 1012, based on the corresponding relationship between the respective label and the respective image that is stored by the corresponding relationship generation unit 1006, determines one or more images that correspond to the label of the image for search generated by the label generation unit 1010.
The image supply unit 1014 sends the one or more images determined by the image determination unit 1012 to a client terminal that sends the image searching request.
In an example, the description information obtaining unit 1008 may, based on the image for search included in the image searching request, determine an entity corresponding to the image for search and obtain description information of the entity as the description information of the image for search.
In an example, the label generation unit 1010 may extract keywords of defined categories from the description information of the image for search and generate the label of the image for search based on the extracted keywords.
In another example, the label generation unit 1010 may, based on the extracted keyword of each defined category, use an algorithm corresponding to the defined category to generate a label vector corresponding to the keyword, and generate the label of the image for search based on the label vector corresponding to each keyword and a weight value of each category of the keyword.
In an example, the corresponding relationship generation unit 1006 may use the generated label of each image as the index and use an inverted index to store the corresponding relationship between the labels and the images.
In another example, the corresponding relationship generation unit 1006 may extract image features of each image in the image searching database and store the corresponding relationship between the stored label, the image corresponding to the label, and the image feature extracted from the image.
As shown in FIG. 11, the apparatus 1000 may further include an image ranking unit 1016. The image ranking unit 1016, before the image supply unit 1014 sends the determined images to the client terminal that sends the image searching request, based on the received image searching request, extracts image features of the image for search included in the image searching request. The image ranking unit 1016 compares the image features of the image for search with image features corresponding to each of the determined images, and ranks the determined images according to their similarities degrees with the image for search from high to low.
The functionalities of each of the above units may correspond to relevant operations of image searching process in accordance with the first example embodiment of the present disclosure. Thus, details are not described herein.
The apparatus for searching images as described in the third example embodiment of the present disclosure may be implemented through one or more processors configured with computer-executable instructions. Persons skilled in the art should understand that the classification of the above units is only one of many unit classifications. If the units are classified otherwise or there is no classification of the units, as long as the apparatus for searching image has the above functionalities, such apparatus still falls within the protection scope of the present disclosure.
Further, the apparatus for searching images as described in the third example embodiment of the present disclosure may further include multiple units to implement respective functionality as described in the above first example embodiment of the present disclosure. Such classifications of units are not described herein.
At least one example provided by the present disclosure may re-use the existing text search engine, thereby improving the utilization rate of the server resources.
Persons skilled in the art should understand that the embodiments of the present disclosure can be methods, systems, or the programming products of computers. Therefore, the present disclosure can be implemented by hardware, software, or in combination of both. In addition, the present disclosure can be in a form of one or more computer programs containing the computer-executable or processor-executable instructions that can be implemented in the computer storage medium (including but not limited to disks, CD-ROM, optical disks, etc.).
The present disclosure is described by referring to the flow charts and/or block diagrams of the method, device (system) and computer program of the embodiments of the present disclosure. It should be understood that each flow and/or block and the combination of the flow and/or block of the flowchart and/or block diagram can be implemented by computer program instructions. These computer program instructions can be provided to the general computers, specific computers, embedded processor or other programmable data processors to generate a machine, so that a device of implementing one or more flows of the flow chart and/or one or more blocks of the block diagram can be generated through the instructions operated by a computer or other programmable data processors.
These computer program instructions can also be stored in other computer storage media which can instruct a computer or other programmable data processors to operate in a certain way, so that the instructions stored in the computer-readable storage generate a product containing the instruction device, wherein the instruction device implements the functions specified in one or more flows of the flow chart and/or one or more blocks of the block diagram.
These computer program instructions can also be loaded in a computer or other programmable data processors, so that the computer or other programmable data processors can operate a series of operation steps to generate the process implemented by a computer. Accordingly, the instructions operated in the computer or other programmable data processors can provides the steps for implementing the functions specified in one or more flows of the flow chart and/or one or more blocks of the block diagram.
The embodiments are merely for illustrating the present disclosure and are not intended to limit the scope of the present disclosure. It should be understood for persons in the technical field that certain modifications and improvements can be made and should be considered under the protection of the present disclosure without departing from the principles of the present disclosure.

Claims

CLAIMS What is claimed is:
1. A method performed by one or more processors configured with computer-executable instructions, the method comprising:
receiving an image searching request, the image searching request including an image for search;
obtaining description information of the image for search based on the image searching request;
generating a label of the image for search based on the description information of the image for search; and
determining one or more images in the image searching database that correspond to the label of the image for search; wherein the one or more images in the image searching database are preprocessed comprising:
generating one or more respective labels corresponding to one or more respective description informations of one or more respective images in an image searching database; storing a corresponding relationship between the respective labels and the respective images.
2. The method as recited in claim 1, further comprising:
sending the determined one or more images to a client terminal that sent the image searching request.
3. The method as recited in claim 1, wherein the obtaining description information of the image for search based on the image searching request comprises:
determining an entity corresponding to the image for search based on the image for search included in the image searching request; and
obtaining description information of the entity as the description information of the image for search.
4. The method as recited in claim 1, wherein the generating the label of the image for search based on the description information of the image for search comprises:
extracting one or more keywords of defined categories from the description information of the image for search; and
generating the label of the image for search based on the extracted one or more keywords.
5. The method as recited in claim 1, wherein the generating the label of the image for search based on the extracted one or more keywords comprises:
for each keyword with a respective defined category, generating a respective label vector corresponding to a respective keyword based on an algorithm corresponding to the respective category; and
generating the label of the image for search based on the respective label vector corresponding to the respective keyword and a respective weight of the respective category for each keyword.
6. The method as recited in claim 1, wherein the storing the corresponding relationship between the respective labels and the respective images comprises:
using the respective labels corresponding to the respective images as an index; and using an inverted indexing to store the corresponding relationship between the respective labels and the respective images.
7. The method as recited in claim 1, wherein the storing the corresponding relationship between the respective labels and the respective images comprises:
extracting image features of the respective images in the image searching database; and
storing a corresponding relationship between the respective labels, the respective images corresponding to the respective labels, and the extracted image features of the respective images.
8. The method as recited in claim 1 , further comprising:
extracting image features of the image for search included in the image searching request;
comparing the image features of the image for search with image features of the determined images; and
ranking the determined images based on their similarity degrees with the image features of the image for search.
9. An apparatus comprising:
a corresponding relationship generation unit that, with respect to each image in an image searching database, generates a respective label of a respective image based on description information of the respective image and stores a corresponding relationship between the respective label and the respective image;
a description information obtaining unit that, based on an image searching request, obtains corresponding description information of an image for search included in the image searching request; a label generation unit that, based on the description information corresponding to the image for search obtained by the description information obtaining unit, generates a corresponding label of the image for search; and
an image determination unit that, based on the corresponding relationship between the respective label and the respective image that is stored by the corresponding relationship generation unit, determines one or more images that correspond to the corresponding label of the image for search generated by the label generation unit.
10. The apparatus as recited in claim 9, further comprising:
an image supply unit that sends the one or more images determined by the image determination unit to a client terminal that sends the image searching request.
11. The apparatus as recited in claim 9, wherein the description information obtaining unit, based on the image for search included in the image searching request, determines an entity corresponding to the image for search and obtains description information of the entity as the description information of the image for search.
12. The apparatus as recited in claim 9, wherein the label generation unit extracts one or more keywords of one or more defined categories from the description information of the image for search and generates the label of the image for search based on the extracted keywords.
13. The apparatus as recited in claim 9, wherein the label generation unit, based on an extracted keyword of each defined category, uses an algorithm corresponding to the defined category to generate a label vector corresponding to the keyword, and generates the label of the image for search based on the label vector corresponding to each keyword and a weight value of the category corresponding to the keyword.
14. The apparatus as recited in claim 9, wherein the corresponding relationship generation unit uses the generated label of each image as the index and uses an inverted index to store the corresponding relationship between labels and images.
15. The apparatus as recited in claim 9, wherein the corresponding relationship generation unit extracts one or more image features of each image in the image searching database and stores the corresponding relationship between the respective stored label, the respective image corresponding to the respective label, and respective image feature extracted from the respective image.
16. The apparatus as recited in claim 9, further comprising an image ranking unit that, before the image supply unit sends the determined images to the client terminal that sends the image searching request, based on the received image searching request, extracts image features of the image for search included in the image searching request, compares the image features of the image for search with image features corresponding to each of the determined images, and ranks the determined images according to their similarities degrees with the image for search from high to low.
17. One or more computer storage media including processor-executable instructions that, when executed by one or more processors, direct the one or more processors to perform a method comprising:
receiving an image searching request, the image searching request including an image for search;
obtaining description information of the image for search based on the image searching request;
generating a label of the image for search based on the description information of the image for search; and
determining one or more images that correspond to the label of the image for search.
18. The one or more computer storage media as recited in claim 17, wherein the obtaining description information of the image for search based on the image searching request comprises:
determining an entity corresponding to the image for search based on the image for search included in the image searching request; and
obtaining description information of the entity as the description information of the image for search.
19. The one or more computer storage media as recited in claim 17, wherein the generating the label of the image for search based on the description information of the image for search comprises:
extracting one or more keywords of defined categories from the description information of the image for search; and
generating the label of the image for search based on the extracted one or more keywords.
20. The one or more computer storage media as recited in claim 17, wherein the generating the label of the image for search based on the extracted one or more keywords comprises:
for each keyword with a respective defined category, generating a respective label vector corresponding to a respective keyword based on an algorithm corresponding to the respective category; and
generating the label of the image for search based on the respective label vector corresponding to the respective keyword and a respective weight of the respective category for each keyword.
PCT/US2012/066251 2011-11-25 2012-11-21 Image searching WO2013078307A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
JP2014543557A JP6141305B2 (en) 2011-11-25 2012-11-21 Image search
EP12794621.8A EP2783310A1 (en) 2011-11-25 2012-11-21 Image searching

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201110380442XA CN103136228A (en) 2011-11-25 2011-11-25 Image search method and image search device
CN201110380442.X 2011-11-25

Publications (1)

Publication Number Publication Date
WO2013078307A1 true WO2013078307A1 (en) 2013-05-30

Family

ID=47263610

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2012/066251 WO2013078307A1 (en) 2011-11-25 2012-11-21 Image searching

Country Status (6)

Country Link
US (1) US9460117B2 (en)
EP (1) EP2783310A1 (en)
JP (1) JP6141305B2 (en)
CN (1) CN103136228A (en)
TW (1) TW201322021A (en)
WO (1) WO2013078307A1 (en)

Families Citing this family (60)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9111012B2 (en) * 2012-11-26 2015-08-18 Accenture Global Services Limited Data consistency management
US10223454B2 (en) * 2013-05-01 2019-03-05 Cloudsight, Inc. Image directed search
US10831820B2 (en) * 2013-05-01 2020-11-10 Cloudsight, Inc. Content based image management and selection
CN104239336B (en) * 2013-06-19 2018-03-16 华为技术有限公司 A kind of method for screening images, device and terminal
US11250203B2 (en) * 2013-08-12 2022-02-15 Microsoft Technology Licensing, Llc Browsing images via mined hyperlinked text snippets
CN104424257A (en) * 2013-08-28 2015-03-18 北大方正集团有限公司 Information indexing unit and information indexing method
CN103714125A (en) * 2013-12-11 2014-04-09 广州亿码科技有限公司 Searching method
CN103744860A (en) * 2013-12-12 2014-04-23 中国科学院深圳先进技术研究院 Method, device and mobile phone for searching for picture
CN103699679A (en) * 2013-12-31 2014-04-02 上海华勤通讯技术有限公司 Method and equipment for retrieving information of target objects
WO2015130185A1 (en) * 2014-02-26 2015-09-03 Андрей Юрьевич ЩЕРБАКОВ Method for indexing, comparing and searching for images in a computer system
CN103995848B (en) * 2014-05-06 2017-04-05 百度在线网络技术(北京)有限公司 Image searching method and device
US10013436B1 (en) * 2014-06-17 2018-07-03 Google Llc Image annotation based on label consensus
US20160139777A1 (en) * 2014-11-18 2016-05-19 Sony Corporation Screenshot based indication of supplemental information
CN104484438A (en) * 2014-12-23 2015-04-01 小米科技有限责任公司 Image processing method and device
CN104462590B (en) * 2014-12-30 2019-03-12 百度在线网络技术(北京)有限公司 Information search method and device
CN104504110B (en) 2014-12-30 2019-05-03 百度在线网络技术(北京)有限公司 Searching method and device
CN104504109B (en) * 2014-12-30 2017-03-15 百度在线网络技术(北京)有限公司 Image searching method and device
US9760792B2 (en) 2015-03-20 2017-09-12 Netra, Inc. Object detection and classification
US9922271B2 (en) 2015-03-20 2018-03-20 Netra, Inc. Object detection and classification
US9852131B2 (en) 2015-05-18 2017-12-26 Google Llc Techniques for providing visual translation cards including contextually relevant definitions and examples
CN104899744B (en) * 2015-06-29 2018-03-16 汪新淮 Coin deposit of stamp product transaction system and method for commerce
CN105138680A (en) * 2015-09-14 2015-12-09 郑州悉知信息科技股份有限公司 Keyword classification method and device and product search method and device
CN105138681A (en) * 2015-09-15 2015-12-09 成都品果科技有限公司 Picture search method and system based on space coordinate positioning
CN106570714B (en) * 2015-10-08 2021-01-15 阿里巴巴集团控股有限公司 Recommendation method for matching object picture, and mapping relation establishment method and device
CN105389333B (en) * 2015-10-13 2019-04-09 深圳市红坚果科技有限公司 A kind of searching system construction method and server architecture
CN105468716A (en) * 2015-11-19 2016-04-06 努比亚技术有限公司 Picture search device and method, and terminal
CN106779791B (en) * 2015-11-25 2021-01-15 阿里巴巴集团控股有限公司 Generation method and device for collocation object picture combination
CN105528428A (en) * 2015-12-09 2016-04-27 深圳市金立通信设备有限公司 Image display method and terminal
US10599731B2 (en) * 2016-04-26 2020-03-24 Baidu Usa Llc Method and system of determining categories associated with keywords using a trained model
CN106126647A (en) * 2016-06-24 2016-11-16 宇龙计算机通信科技(深圳)有限公司 Image searching method and device
CN106202316A (en) * 2016-07-01 2016-12-07 传线网络科技(上海)有限公司 Merchandise news acquisition methods based on video and device
CN106779304A (en) * 2016-11-24 2017-05-31 广州宝钢南方贸易有限公司 A kind of inside competition for orders system and method towards steel-trading
CN108268571A (en) * 2017-01-04 2018-07-10 腾讯科技(深圳)有限公司 A kind of Feature Selection method and device
CN107862239A (en) * 2017-09-15 2018-03-30 广州唯品会研究院有限公司 A kind of combination text carries out the method and its device of picture recognition with picture
CN107861970A (en) * 2017-09-15 2018-03-30 广州唯品会研究院有限公司 A kind of commodity picture searching method and device
CN107861971A (en) * 2017-09-15 2018-03-30 广州唯品会研究院有限公司 A kind of product search method and device
CN110019903A (en) * 2017-10-10 2019-07-16 阿里巴巴集团控股有限公司 Generation method, searching method and terminal, the system of image processing engine component
CN107741972A (en) * 2017-10-12 2018-02-27 广东欧珀移动通信有限公司 A kind of searching method of picture, terminal device and storage medium
CN107679208A (en) * 2017-10-16 2018-02-09 广东欧珀移动通信有限公司 A kind of searching method of picture, terminal device and storage medium
CN107590291A (en) * 2017-10-30 2018-01-16 广东欧珀移动通信有限公司 A kind of searching method of picture, terminal device and storage medium
CN107895050A (en) * 2017-12-07 2018-04-10 联想(北京)有限公司 Image searching method and system
CN108228826A (en) * 2018-01-02 2018-06-29 联想(北京)有限公司 A kind of information searching method and electronic equipment
CN108256010A (en) * 2018-01-03 2018-07-06 阿里巴巴集团控股有限公司 Content recommendation method and device
US10592782B2 (en) * 2018-01-22 2020-03-17 International Business Machines Corporation Image analysis enhanced related item decision
CN110309464A (en) * 2018-03-01 2019-10-08 北京京东尚科信息技术有限公司 Information display system, method and device
CN108596280A (en) * 2018-04-27 2018-09-28 佛山市日日圣科技有限公司 Image identification method for literature search
TWI693524B (en) * 2018-05-22 2020-05-11 正修學校財團法人正修科技大學 Optimization method for searching exclusive personalized pictures
CN108717464A (en) * 2018-05-31 2018-10-30 中国联合网络通信集团有限公司 photo processing method, device and terminal device
CN108848270B (en) * 2018-06-29 2021-05-11 维沃移动通信(深圳)有限公司 Method for processing screen shot image and mobile terminal
CN109635135A (en) * 2018-11-30 2019-04-16 Oppo广东移动通信有限公司 Image index generation method, device, terminal and storage medium
CN109840288A (en) * 2019-02-21 2019-06-04 北京一品智尚信息科技有限公司 Picture retrieval method, equipment and storage medium
CN111753181A (en) * 2019-03-28 2020-10-09 北京京东尚科信息技术有限公司 Image-based search method, apparatus, server, client, and medium
CN112131417B (en) * 2019-06-25 2024-04-02 北京百度网讯科技有限公司 Image tag generation method and device
CN112347289A (en) * 2019-08-06 2021-02-09 Tcl集团股份有限公司 Image management method and terminal
CN110598038A (en) * 2019-09-27 2019-12-20 京东方科技集团股份有限公司 Painting label generation method and electronic equipment
CN110765301B (en) * 2019-11-06 2022-02-25 腾讯科技(深圳)有限公司 Picture processing method, device, equipment and storage medium
CN111522967B (en) * 2020-04-27 2023-09-15 北京百度网讯科技有限公司 Knowledge graph construction method, device, equipment and storage medium
CN112860929A (en) * 2021-02-09 2021-05-28 北京百度网讯科技有限公司 Picture searching method and device, electronic equipment and storage medium
CN114741550B (en) * 2022-06-09 2023-02-10 腾讯科技(深圳)有限公司 Image searching method and device, electronic equipment and computer readable storage medium
CN115168637B (en) * 2022-06-17 2023-03-17 杭州晶彩数字科技有限公司 Method, system and storage medium for adding label to picture

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1103900A2 (en) * 1999-11-26 2001-05-30 Medison Co., Ltd. Ultrasonic image searching apparatus and ultrasonic image transmission and reception system adopting the same
US20030229628A1 (en) * 2002-06-10 2003-12-11 International Business Machines Corporation Method and apparatus for processing user input selecting images from a web page in a data processing system

Family Cites Families (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0947937B1 (en) * 1998-04-02 2010-11-03 Canon Kabushiki Kaisha Image search apparatus and method
US8032483B1 (en) * 2004-12-03 2011-10-04 Google Inc. Using game responses to gather data
US7660468B2 (en) * 2005-05-09 2010-02-09 Like.Com System and method for enabling image searching using manual enrichment, classification, and/or segmentation
JP2007134934A (en) * 2005-11-10 2007-05-31 Matsushita Electric Ind Co Ltd Video image information retrieval system and video image monitoring system
US8341112B2 (en) * 2006-05-19 2012-12-25 Microsoft Corporation Annotation by search
JP2008004081A (en) * 2006-05-23 2008-01-10 Hitachi High-Technologies Corp Image search apparatus, image search system, image search method, and program for executing image search method
JP5098253B2 (en) * 2006-08-25 2012-12-12 コニカミノルタエムジー株式会社 Database system, program, and report search method
US20080126191A1 (en) 2006-11-08 2008-05-29 Richard Schiavi System and method for tagging, searching for, and presenting items contained within video media assets
US20090024577A1 (en) 2007-07-20 2009-01-22 Gregory Prestel System and method for organizing, posting and searching information on a network
CN101458695A (en) * 2008-12-18 2009-06-17 西交利物浦大学 Mixed picture index construct and enquiry method based on key word and content characteristic and use thereof
CN101458711B (en) * 2008-12-30 2011-01-05 国家电网公司 Image describing and transforming method and system
CN101782904A (en) * 2009-01-21 2010-07-21 神达电脑股份有限公司 Personal navigation device and method for automatic adding label to photo
US8452794B2 (en) * 2009-02-11 2013-05-28 Microsoft Corporation Visual and textual query suggestion
WO2010101187A1 (en) 2009-03-04 2010-09-10 公立大学法人大阪府立大学 Method and program for creating image database, and method for retrieving image
CN102483745B (en) 2009-06-03 2014-05-14 谷歌公司 Co-selected image classification
US9384214B2 (en) * 2009-07-31 2016-07-05 Yahoo! Inc. Image similarity from disparate sources
CN101694657B (en) * 2009-09-18 2011-11-09 浙江大学 Picture retrieval clustering method facing to Web2.0 label picture shared space
US8503739B2 (en) 2009-09-18 2013-08-06 Adobe Systems Incorporated System and method for using contextual features to improve face recognition in digital images
US20110072047A1 (en) * 2009-09-21 2011-03-24 Microsoft Corporation Interest Learning from an Image Collection for Advertising
US9710491B2 (en) * 2009-11-02 2017-07-18 Microsoft Technology Licensing, Llc Content-based image search
US9323784B2 (en) 2009-12-09 2016-04-26 Google Inc. Image search using text-based elements within the contents of images
CN102110122B (en) 2009-12-24 2013-04-03 阿里巴巴集团控股有限公司 Method and device for establishing sample picture index table, method and device for filtering pictures and method and device for searching pictures
US8631004B2 (en) * 2009-12-28 2014-01-14 Yahoo! Inc. Search suggestion clustering and presentation
CN101826102B (en) * 2010-03-26 2012-07-25 浙江大学 Automatic book keyword generation method
US8401282B2 (en) 2010-03-26 2013-03-19 Mitsubishi Electric Research Laboratories, Inc. Method for training multi-class classifiers with active selection and binary feedback
CN101853297A (en) 2010-05-28 2010-10-06 英华达(南昌)科技有限公司 Method for fast obtaining expected image in electronic equipment
WO2012064893A2 (en) * 2010-11-10 2012-05-18 Google Inc. Automated product attribute selection

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1103900A2 (en) * 1999-11-26 2001-05-30 Medison Co., Ltd. Ultrasonic image searching apparatus and ultrasonic image transmission and reception system adopting the same
US20030229628A1 (en) * 2002-06-10 2003-12-11 International Business Machines Corporation Method and apparatus for processing user input selecting images from a web page in a data processing system

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
See also references of EP2783310A1 *
VLACHOS M ET AL: "Effective ways for querying images by content over the internet", ELECTROTECHNICAL CONFERENCE, 2000. MELECON 2000. 10TH MEDITERRANEAN MAY 29-31, 2000, PISCATAWAY, NJ, USA,IEEE, vol. 1, 29 May 2000 (2000-05-29), pages 337 - 340, XP010518887, ISBN: 978-0-7803-6290-1 *

Also Published As

Publication number Publication date
EP2783310A1 (en) 2014-10-01
JP6141305B2 (en) 2017-06-07
US9460117B2 (en) 2016-10-04
TW201322021A (en) 2013-06-01
JP2014533868A (en) 2014-12-15
US20130138636A1 (en) 2013-05-30
CN103136228A (en) 2013-06-05

Similar Documents

Publication Publication Date Title
US9460117B2 (en) Image searching
US10452691B2 (en) Method and apparatus for generating search results using inverted index
US9396413B2 (en) Choosing image labels
CN109885773B (en) Personalized article recommendation method, system, medium and equipment
US20170091175A1 (en) Question answering system using multilingual information sources
US20130339369A1 (en) Search Method and Apparatus
US11361030B2 (en) Positive/negative facet identification in similar documents to search context
US20200257679A1 (en) Natural language to structured query generation via paraphrasing
US20230252095A1 (en) Search engine using joint learning for multi-label classification
CN110909120B (en) Resume searching/delivering method, device and system and electronic equipment
CN111078776A (en) Data table standardization method, device, equipment and storage medium
CN112883030A (en) Data collection method and device, computer equipment and storage medium
JP2023516209A (en) METHOD, APPARATUS, APPARATUS AND COMPUTER-READABLE STORAGE MEDIUM FOR SEARCHING CONTENT
CN110210038B (en) Core entity determining method, system, server and computer readable medium thereof
WO2015084757A1 (en) Systems and methods for processing data stored in a database
CN107368489A (en) A kind of information data processing method and device
CN111666383A (en) Information processing method, information processing device, electronic equipment and computer readable storage medium
US11080249B2 (en) Establishing industry ground truth
CN114547257B (en) Class matching method and device, computer equipment and storage medium
CN112487181A (en) Keyword determination method and related equipment
US20180293325A1 (en) Visual leaf page identification and processing
CN111597368A (en) Data processing method and device
US11074591B2 (en) Recommendation system to support mapping between regulations and controls
US11954424B2 (en) Automatic domain annotation of structured data
US20230351101A1 (en) Automatic domain annotation of structured data

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12794621

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2014543557

Country of ref document: JP

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: 2012794621

Country of ref document: EP

NENP Non-entry into the national phase

Ref country code: DE