US20120203764A1 - Identifying particular images from a collection - Google Patents
Identifying particular images from a collection Download PDFInfo
- Publication number
- US20120203764A1 US20120203764A1 US13/021,188 US201113021188A US2012203764A1 US 20120203764 A1 US20120203764 A1 US 20120203764A1 US 201113021188 A US201113021188 A US 201113021188A US 2012203764 A1 US2012203764 A1 US 2012203764A1
- Authority
- US
- United States
- Prior art keywords
- image
- images
- keywords
- collection
- descriptors
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 claims abstract description 54
- 238000012545 processing Methods 0.000 claims description 27
- 230000002123 temporal effect Effects 0.000 claims description 13
- 239000003086 colorant Substances 0.000 claims description 10
- 230000007246 mechanism Effects 0.000 claims description 10
- 239000000463 material Substances 0.000 claims description 8
- 230000000694 effects Effects 0.000 claims description 7
- 230000000007 visual effect Effects 0.000 claims description 4
- 230000001419 dependent effect Effects 0.000 claims 1
- 238000004422 calculation algorithm Methods 0.000 description 17
- 238000013500 data storage Methods 0.000 description 15
- 230000008569 process Effects 0.000 description 14
- 239000011159 matrix material Substances 0.000 description 12
- 230000002093 peripheral effect Effects 0.000 description 9
- 238000001914 filtration Methods 0.000 description 8
- 230000015654 memory Effects 0.000 description 8
- 241001481824 Indri Species 0.000 description 6
- 241000283083 Sirenia Species 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 4
- 244000025254 Cannabis sativa Species 0.000 description 3
- 238000013459 approach Methods 0.000 description 3
- 239000011435 rock Substances 0.000 description 3
- 239000004576 sand Substances 0.000 description 3
- 230000006399 behavior Effects 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 230000001413 cellular effect Effects 0.000 description 2
- 238000007635 classification algorithm Methods 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 230000001815 facial effect Effects 0.000 description 2
- 238000002372 labelling Methods 0.000 description 2
- 238000013507 mapping Methods 0.000 description 2
- 238000003058 natural language processing Methods 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 239000002023 wood Substances 0.000 description 2
- 241001263092 Alchornea latifolia Species 0.000 description 1
- 241000282326 Felis catus Species 0.000 description 1
- 241000283973 Oryctolagus cuniculus Species 0.000 description 1
- 241000052209 Triplasis Species 0.000 description 1
- 239000003795 chemical substances by application Substances 0.000 description 1
- 238000004040 coloring Methods 0.000 description 1
- 238000009472 formulation Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 230000036651 mood Effects 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000009182 swimming Effects 0.000 description 1
- 239000013598 vector Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/953—Querying, e.g. by the use of web search engines
- G06F16/9535—Search customisation based on user profiles and personalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/24—Querying
- G06F16/245—Query processing
- G06F16/2457—Query processing with adaptation to user needs
- G06F16/24573—Query processing with adaptation to user needs using data annotations, e.g. user-defined metadata
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/53—Querying
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/583—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/5866—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, manually generated location and time information
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/60—Information retrieval; Database structures therefor; File system structures therefor of audio data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/907—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/953—Querying, e.g. by the use of web search engines
- G06F16/9538—Presentation of query results
Definitions
- the present invention relates to identifying one or more particular images from an image collection.
- An advantage of the present invention is that it enables users to search their personal image collections using arbitrarily complex and potentially obscure search terms, without requiring the user to provide annotations for the images and without requiring the system to incorporate indexers for recognizing the corresponding concepts.
- the present invention enables semantically complex concepts to be automatically associated with images for which only low-level semantic concepts are available.
- the present invention further improves the user's ability to interpret, navigate and view search results.
- FIG. 1 is a high-level diagram showing the components of a system for receiving and processing image queries according to an embodiment of the present invention
- FIG. 2 is a flow diagram illustrating the major steps in receiving and processing media queries according to an embodiment of the present invention
- FIG. 3 is a flow diagram illustrating the steps used to transform a user-provided query into a set of keywords that can be used as search terms;
- FIG. 4 is a flow diagram illustrating the steps used to compute a co-occurrency matrix
- FIG. 5 illustrates an example user interface which can be used to manually tag images as part of creating a tagged image collection
- FIG. 6 illustrates an example user interface for searching an image collection
- FIG. 7 illustrates an example user interface for viewing and navigating a group of search results.
- the present invention combines automatic labeling with domain-specific concept expansion to enable image retrieval from unannotated or minimally annotated image collections.
- This approach can permit a user to search collections using queries containing semantically complex concepts or keywords.
- a user can search their personal image collection for pictures of manatees, for example, and get appropriate results, even though none of the pictures were tagged as containing manatees nor does the system include an algorithm for identifying manatees. Instead, the system relies on combining low level or common concepts for which there are automated algorithms (although the user is also free to add such tags if desired).
- algorithms for identifying water are relatively robust, and can be combined with information obtained from reverse geocoding latitude and longitude information which is increasingly recorded as part of the image capture process. Such information can be applied to a picture of an object taken in a river in a Florida wildlife state park, permitting the system to conclude that this image is more likely to contain manatees than other images in the user's collection.
- identifying that the associated audio track contains instrumental music and combining that with knowledge that the video was captured at a music school in the evening can again permit the system to conclude that the given video is more likely than other videos to be a video of a piano recital.
- a feature of the present invention is that the returned results can be grouped according to logical groupings such as events or visual similarity, and presented according to such groupings. Such groups permit the representation of the results to be condensed, enabling the user to quickly identify any groupings that might contain the desired results.
- a grouping can be visually represented by picking a representative image from the set, such as a mini collage of a subset of the images, or by some other appropriate representation. The user can then readily expand the selected groupings to reveal the specific intended target, if the intended target was not itself originally displayed as one of the search results.
- the present invention enables successful concept-based image retrieval on large personal image collections, even in the absence of user-provided annotations.
- the present invention is written specific to retrieving image objects from an image collection, the concept described here is not limited to images.
- the present invention can be applied to search for music, video or document files.
- FIG. 1 illustrates a system 100 for identifying one or more particular images from an image collection, according to an embodiment of the present invention.
- the system 100 includes a data processing system 110 , a peripheral system 120 , a user interface system 130 , and a processor-accessible data storage system 140 .
- the processor-accessible data storage system 140 , the peripheral system 120 , and the user interface system 130 are communicatively connected to the data processing system 110 .
- the data processing system 110 includes one or more data processing devices that implement the processes of the various embodiments of the present invention, including the example processes of FIGS. 2-4 described herein.
- the phrases “data processing device” or “data processor” are intended to include any data processing device, such as a central processing unit (“CPU”), a desktop computer, a laptop computer, a mainframe computer, a personal digital assistant, a BlackberryTM, a digital camera, cellular phone, or any other device for processing data, managing data, or handling data, whether implemented with electrical, magnetic, optical, biological components, or otherwise.
- the processor-accessible data storage system 140 includes one or more processor-accessible memories configured to store information, including the information needed to execute the processes of the various embodiments of the present invention, including the example processes of FIGS. 2-4 described herein.
- the data storage system 140 includes an image collection 145 and a separate tagged image collection 155 .
- the processor-accessible data storage system 140 can be a distributed processor-accessible data storage system including multiple processor-accessible data storage systems communicatively connected to the data processing system 110 via a plurality of computers or devices.
- the processor-accessible data storage system 140 need not be a distributed processor-accessible data storage system and, consequently, can include one or more processor-accessible memories located within a single data processor or device.
- processor-accessible memory is intended to include any processor-accessible data storage device, whether volatile or nonvolatile, electronic, magnetic, optical, or otherwise, including but not limited to, floppy disks, hard disks, Compact Discs, DVDs, flash memories, ROMs, and RAMs.
- the phrase “communicatively connected” is intended to include any type of connection, whether wired or wireless, between devices, data processors, or programs in which data can be communicated. Further, the phrase “communicatively connected” is intended to include a connection between devices or programs within a single data processor, a connection between devices or programs located in different data processors, and a connection between devices not located in data processors at all.
- the processor-accessible data storage system 140 is shown separately from the data processing system 110 , one skilled in the art will appreciate that the processor-accessible data storage system 140 can be stored completely or partially within the data processing system 110 .
- the peripheral system 120 and the user interface system 130 are shown separately from the data processing system 110 , one skilled in the art will appreciate that one or both of such systems can be stored completely or partially within the data processing system 110 .
- the peripheral system 120 can include one or more devices configured to provide images to the data processing system 110 .
- the peripheral system 120 can include digital video cameras, cellular phones, regular digital cameras, or other data processors.
- the data processing system 110 upon receipt of images from a device in the peripheral system 120 , can store such images in the image collection 145 in the processor-accessible data storage system 140 .
- the user interface system 130 can include a mouse, a keyboard, another computer, or any device or combination of devices from which data is input to the data processing system 110 .
- the peripheral system 120 is shown separately from the user interface system 130 , the peripheral system 120 can be included as part of the user interface system 130 .
- the user interface system 130 also can include a display device, a processor-accessible memory, or any device or combination of devices to which data is output by the data processing system 110 .
- a display device e.g., a liquid crystal display
- a processor-accessible memory e.g., a liquid crystal display
- any device or combination of devices to which data is output by the data processing system 110 e.g., a liquid crystal display
- the user interface system 130 includes a processor-accessible memory, such memory can be part of the processor-accessible memory system 140 even though the user interface system 130 and the processor-accessible data storage system 140 are shown separately in FIG. 1 .
- FIG. 2 illustrates a method for identifying one or more particular images from an image collection, according to an embodiment of the present invention.
- This method can be implemented by the data processing system 110 in FIG. 1 .
- the system takes as its input an image collection 205 (equivalently, image collection 145 in FIG. 1 ) which is indexed in step 210 to produce a set of image descriptors 215 .
- the indexing process includes a variety of indexers, including temporal event clustering on groups of assets as well as temporal indexing based on individual assets; scene, activity, object and material classifiers; face-based age and gender estimators; color; camera capture metadata including, for example, Exif metadata; people recognition and reverse geocoding.
- These indexers cover the one or more of the usual who, what, when, and where vectors that can be used as the basis for a search for image objects.
- These indexers can all be implemented by the data processing system 110 in FIG. 1 .
- Individual images can be indexed based upon their capture date and time.
- the date and time information can be readily mapped to concepts such as seasons of the year or other temporal periods.
- concepts such as birthday, anniversary or even the names of specific personal events such as “Florida vacation” can be associated with assets.
- Appropriate civil and religious calendars can be used to further associate the names of holidays such as Christmas or Independence Day with individual assets or event-based clusters.
- Temporal event clustering of images in the image collection 205 is generated by automatically sorting, segmenting, and clustering an unorganized set of media assets into separate temporal events and sub-events, as described in detail in commonly assigned U.S. Pat. No. 6,606,411 entitled “A method for Automatically Classifying Images Into Events,” issued Aug. 12, 2003, and commonly assigned U.S. Pat. No. 6,351,556, entitled “A Method for Automatically Comparing Content of Images for Classification Into Events,” issued Feb. 26, 2002.
- an event classification algorithm can be employed as described in US Patent Application Publication 2010/0124378 entitled “Method for Event-Based Semantic Classification” to classify the event into one of four types: family, vacation, sports or family moment.
- Other event classification algorithms can be equally applied, resulting in potentially other event types.
- Scene classifiers identify or classify an image as representing one or more scene types such as mountain, beach or indoor.
- Material and texture classifiers identify or classify an image as containing one or more materials or texture types, such as rock, sand, grass, or blue sky.
- Object classifiers identify or classify an image as containing one or more objects, such as car, cat, or bicycle.
- Some example scene and material classification types include indoor, outdoor, natural, urban, sunset, beach, foliage, field, mountain, sky, grass, snow, water and sand. Operation of such classifiers are described in U.S. Pat. No. 6,282,317 entitled “Method for automatic determination of main subjects in photographic images”; U.S. Pat. No.
- Color classifiers identify or classify an image as containing one or more prominent colors. Such classifiers can compute a histogram of the different colors present in the image, potentially grouping together pixels of nearly identical coloring. The most frequently occurring colors can be mapped to their common color names, which can then be associated with the image as image descriptors.
- a face detector algorithm is used to find faces in image collections, and is described in U.S. Pat. No. 7,110,575 entitled “Method for Locating Faces in Digital Color Images”, issued Sep. 19, 2006; U.S. Pat. No. 6,940,545 entitled “Face Detecting Camera and Method” issued Sep. 6, 2005; U.S. Patent Application Publication 2004/0179719 entitled “Method and system for face detection in digital image assets” filed Mar. 12, 2003.
- the approximate age of an individual can be estimated as described in the work by A. Lanitis, C. Draganova, and C. Christodoulou, “Comparing different classifiers for automatic age estimation,” IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics, vol. 34, no. 1, pp. 621-628, February 2004.
- the gender of a person can be predicted based upon the facial shape as described in the literature.
- Other cues such as clothing, hair and social context can be further used to improve upon the age and gender estimation, as published in the literature, e.g., the Ph.D. dissertation by Andrew C. Gallagher entitled “A Framework for Using Context to Understand Images of People,” published by Carnegie Mellon University, May 2009.
- a high-level semantic label can be associated with the face or the containing image object. For example, given an estimated age of less than one year; the system can generate the label “baby.”Given a face estimated to be approximately age 29 and female, the label “ woman” can be generated. The presence of multiple faces in an image can further result in additional descriptive labels being generated. For example, the presence of multiple faces estimated to be age 16 or less can result in the label “children.” Other labels include “man,” “boy,” “girl,” and “teen.” As described in the previously cited dissertation by Andrew C. Gallagher, the presence of multiple faces also can be further used to refine the individual age and gender estimates of each individual in the image.
- the presence of multiple detected people can be used to tag or label an image as being a group shot; estimates of age and gender of the individuals in the group shot, possibly combined with other contextual cues, can provide additional estimates as to the relative relationships and nature of the group shot, such as a family photo. Such estimates can be used to generate additional semantic labels.
- Face recognition is the identification or classification of a face to an example of a person or a label associated with a person based on facial features as described in U.S. Patent Application Publication 2008/0112621 entitled “User interface for face recognition”; U.S. Pat. No. 7,711,145 entitled “Finding Images with Multiple People or Objects”; and U.S. Patent Application Publication 2007/0098303 entitled “Determining a Particular Person from a Collection”.
- Reverse geocoding is the process of mapping geospatial coordinates such as latitude and longitude to specific place names or geographic feature types. Any of a number of commonly available services for reverse geocoding, such as the service offered by geonames.org, can be employed to perform such mapping. Furthermore, such services can be used to obtain not only nearby feature points, but also geospatial feature types, such as lake, school, beach or park. Heuristics such as those described in the section on geospatial classification in the work by Mark D. Wood entitled “Matching Songs to Events in Image Collections,” IEEE International Conference on Semantic Computing, 2009. ICSC '09., pp. 95-102, 14-16 Sep. 2009, can be used to associate one or more feature types with an image.
- the technique described herein applies to other media object types.
- the described invention can also be applied with minor variations to any object type for which one might expect to find an independent set of tagged objects of possibly the same or a different type, where the tags are from a vocabulary that one might use as search terms.
- One type would be music or audio files.
- the descriptors can include musical genre, artist, album, recording location, and mood.
- Video is another type; most of the types associated with still image objects also apply to audio.
- Other types include special sound effects, olfactory recordings or food recipes.
- Other classifier types can be applicable for other data types and can be used as indexers in addition to or besides the types of classifiers for described here. For example, activity classifiers recognizing activities such as swimming, running, or singing can be employed for video objects.
- the set of descriptors for an image can include descriptors provided directly or indirectly by the user. For example, many users will rename the files corresponding to images of particular importance to them, giving the file a more descriptive name than the name originally used by the capture device. Likewise, users can organize images into folders or directories, giving the directory or folder a meaningful name. In some cases, users will also associate a caption with an image. Information can also be indirectly associated with an image.
- U.S. Pat. No. 7,025,253 entitled “Agent for integrated annotation and retrieval of images” describes a mechanism for automatically associating suggested annotations with an image.
- descriptors for a given image object are represented as a distinct XML document.
- Mechanisms for storing the descriptors within the image object can be used for container-style file formats, such as Exif or MOV files. Some image descriptors have definitions as part of standardized formats such as Exif. Other descriptors can be incorporated into an existing image file format using extension mechanisms.
- One popular way to include metadata within documents is to use the Extensible Metadata Platform (XMP). Generated metadata can also be stored directly in a database.
- XMP Extensible Metadata Platform
- Probabilities or numeric scores can also be stored with the metadata properties.
- the XML elements scenes and colors shown above might in one embodiment include numeric data representing the degree of belief, confidence, probability, or some other numeric score, as illustrated here:
- the system receives a query 225 from the user.
- This query 225 can be entered by the user using the user interface system 130 of FIG. 1 and can be entered using any of a variety of mechanisms, including the user typing the query 225 into a dialog box, speaking the query into a microphone, or indicating the query by selecting items from a checklist.
- the query 225 is expressed as a set of keywords, such as “camping Adirondacks.”
- the user can enter the query 225 as a phrase or sentence in either an imperative or descriptive form, such as, “find pictures of my camping trip in the Adirondacks” or “Michelle's pet rabbit.”
- the query 225 is processed in step 230 by the data processing system 110 of FIG. 1 to produce a set of keywords 235 to use as search terms.
- step 230 illustrates the steps used to transform a user-provided query 305 into a set of keywords 335 that can be used as search terms.
- step 310 the set of keywords 315 are extracted from the user-provided query 305 . If the user provided a phrase or sentence, commonly practiced natural language processing techniques are applied to extract keywords. If the user input is already in the form of a string of keywords, then the string must simply be parsed into a set of keywords 315 . Either way, the resulting set of keywords 315 is further processed in step 320 to eliminate any stop words or commonly used terms such as “picture,” producing a filtered set of keywords 325 . Depending on the system, stopwords can also be partially or completely eliminated as part of step 310 , especially if natural language processing techniques are applied to extract keywords 315 from natural language phrases. The filtered set of keywords 325 can then be expanded in step 330 .
- the expansion process 330 operates by considering each keyword 325 , and determining what other keywords 325 are most likely to co-occur with that keyword 325 in the tagged second collection, to produce an expanded set of keywords 335 (equivalently, set of search keywords 235 from FIG. 2 ). Given a suitable tagged image collection 155 from FIG. 1 , the computation of co-occurrent keywords can be computed by one skilled in the art.
- FIG. 4 illustrates the process of computing a co-occurrency matrix.
- a tagged image collection 405 (equivalently, 155 from FIG. 1 ) is processed by step 410 to compute a co-occurrency matrix 415 .
- the following pseudocode illustrates one algorithm implementing Step 410 for computing the co-occurrency matrix 415 of FIG. 4 .
- the co-occurrency matrix 415 is denoted by the variable Cooccurs.
- the value Cooccurs[a,b] is the frequency of times tag b co-occurs with tag a.
- this algorithm illustrates in lines 7 and 8 the optional steps used by some embodiments to maintain a dictionary of all terms known by the system, and the set of categories associated with each term.
- the Category(t) of the tag can be determined based upon the semantic classifier that generated the tag.
- the various XML element names such as colors, scenes and temporal can denote the corresponding categories.
- a given tag can potentially be generated by multiple classifiers, and therefore have multiple associated categories.
- FIG. 5 illustrates a simple user interface for tagging pictures.
- the user can enter a new tag into text box 510 and then click on button 520 labeled “Add new tag” to add the tag to the image.
- the list of tags currently associated with the image is shown in area 530 .
- the tagging application can be part of a desktop application, or it can be offered as part of some on-line photo sharing service such as Flickr, Photobucket, or the Kodak Gallery. Similar services such as Vimeo and YouTube provide tagging functionality for videos.
- the tagging system can be completely manual; alternatively, it can recommend or suggest tags to a user. Tags can be recommended by using their likelihood of co-occurrence with tags already associated with an image.
- Some systems automatically generate tags based upon capture metadata, user-provided information, or other information available to the system. Any of these embodiments enable tag information to be collected over a set of images. The images can all belong to a single user, or they can come from thousands or millions of users.
- the tagging system can further differentiate between tags originating from different users or different types of images, enabling different Cooccurs matrices to be computed. For example, the system can compute different Cooccurs matrices for tags originating from different user demographics, where the demographics might include language, age, or geographic region. This permits the system to account for different uses of tags depending upon language, age or geography. In such embodiments, the system in step 330 would pick the Cooccurs matrix most appropriate for the current user.
- the step 330 can be carried out using the following procedure:
- the set Cooccurs(k) is the set of co-occurrent keywords computed by step 330 .
- the set of co-occurrent keywords can be further filtered, to ensure that the computed keywords are mutually consistent with other search terms provided by the user. This check for consistency is carried out by a modified Step 330 , as follows:
- Some embodiments applying filtering can apply it selectively to keywords, depending on the keyword.
- the filtering step might only be applied to keywords corresponding to semantic concepts returned by image understanding algorithms such as event classification or scene recognition, but not filter keywords corresponding to concepts extracted from an image filename or path, or that were deduced based upon calendar information and the image capture date.
- the optional category information as described previously as part of Step 410
- some embodiments of Step 330 can incorporate a lookup operation to determine the category of a keyword, either provided as an original search term or obtained from a prior expansion, and adjust the filtering behavior accordingly.
- the performance of the system can be improved, both in the quality of the results and the required computation time.
- the expansion process 330 uses a single co-occurrence matrix to carry out both the expansion and filtering operations; those skilled in the art will readily recognize other improvements that can be readily made.
- a separate source of information can be used for the filtering operation.
- This information can take the form of a co-occurrence matrix computed from a third source of tagged information.
- Another source of information is to consider the co-occurrence of terms in knowledgebases such as Wikipedia, or even general Internet search results.
- the expansion process 330 can also be applied iteratively, so that each set of expanded keywords is further expanded using the same algorithm. The iteration can continue a bounded number of times, or until some other constraint is satisfied.
- step 230 can also incorporate other mechanisms not shown in FIG. 3 in producing the set of search keywords 235 .
- one technique applied in the query expansion literature is to use a lexical database such as WordNet to identify related words; these related words can be included in the set of expanded search keywords.
- Other sources of knowledge such as gazetteers, Cyc, or Wikipedia can also be used to provide related concepts, using the term or a derivative word of the term, such as the stemmed word form.
- step 230 can combine the mechanism illustrated in FIG. 3 with other expansion mechanisms in an interleaved or iterative fashion.
- the expanded search terms can further have weights associated with each term, indicating the strength of the expected significance of that term. For example, the weights can be proportional to the co-occurrency frequency. In addition, terms whose co-occurrency frequency is sufficiently low can be completely excluded from the set of expanded terms.
- step 240 the search keywords 235 are used to identify a set of candidate images 245 .
- step 240 is carried out by indexing the image descriptors 215 using an information retrieval system such as the Indri information retrieval system from the University of Massachusetts, which is described in T. Strohman, D. Metzler, H. Turtle, and W. B. Croft, “Indri: A language-model based search engine for complex queries (extended version),” Center for Intelligence Information Retrieval Technical Report, 2005.
- This information retrieval system supports complex queries including weighted search terms, and quickly returns matching documents.
- This embodiment assigns different weights to each search term in the expanded list of keywords 335 in FIG. 3 , depending on the origin.
- keywords 335 provided by the user are given a higher weight than terms obtained through the expansion process.
- the weight for expanded keywords 335 can be further adjusted based upon their frequency of co-occurrence, their category, or how many iterations of the expansion process were required to result in the word being added to the expansion list.
- Search terms obtained from alternate sources such as WordNet can have separate weights, with the weight based upon the expanded term's relationship to the term being expanded. For example synonyms can be assigned a higher weight than hypernyms and hyponyms.
- Indri Using the image descriptors 215 as the indexed documents, Indri will identify which documents satisfy the specified search terms expressed as an Indri query. By incorporating into each document a reference to the original image, the results from Indri can be easily mapped to the actual image objects. Of course, one skilled in the art will recognize that a search system also can be more directly implemented, including one that accommodates the storage of the image descriptors directly within the image objects. Indri combines the inference network model of information retrieval with a language model for estimating the probability that a given term is satisfied by a given document. However other models known to those skilled in the art of information retrieval can also be applied in step 240 to determine which candidate images satisfy the search keywords.
- Step 240 produces the set of candidate images 245 , which in some embodiments are ranked or ordered according to the estimated strength of their match.
- the candidate images are grouped to produce sets of grouped images 255 .
- the grouping process is based on some commonality in the candidate images with respect to one or more of the image descriptors. For example, in some embodiments, images are grouped by event, so that all images which were taken as part of the same event, using the previously mentioned temporal event clustering algorithm to determine event boundaries. In other embodiments, images can be grouped based upon location, based upon the people identified in the image, based upon visual similarity, or based upon scene, material or activity similarity. The user can have the ability to determine the mode of grouping the candidate images, or the system can automatically chose an appropriate grouping method by analyzing the images.
- a representative image is selected from each candidate grouping to produce a set of representative images 265 .
- the representative image 265 can be selected using any of a variety of algorithms.
- the representative image 265 can be the image from the group that scored the highest according to the search criteria.
- the representative image 265 can also be selected according to some other scoring function partially or completely independent of the search ranking
- the images in a group can be ranked according to their expected aesthetic appeal or information content.
- the representative image 265 for a grouping can be a new image, formed by compositing a plurality of the images in the candidate group.
- Step 270 the set of representative images is displayed to the user. These images can be displayed using the user interface system 130 of FIG. 1 .
- the user interface system 130 can further provide the user with a variety of choices for controlling the display. For example, the user can elect to have all results displayed, not just the representative images 265 chosen in step 260 . Or the user can request that a different mechanism be used to pick the representative images 265 , resulting in Step 260 being repeated using the user's specified criteria for selecting representative images 265 .
- FIG. 6 illustrates one embodiment for the user interface.
- query text box 610 the user entered the search terms “camping Adirondacks.”
- the system displayed the resulting representative images 265 in a scrollable region 640 , with the corresponding image pathnames and search score shown in scrollable area 630 .
- area 620 the user can see the actual query that was formed from the set of expanded search terms. Providing the user with the ability to see the query formed from the search term permits the sophisticated user the ability to further refine the search terms.
- Some embodiments can prefer to show the user the list of expanded search terms using an alternative user interface, perhaps in the form of checkboxes, permitting the user to deselect any inappropriate search terms. However, some embodiments would conceal that level of complexity from the average user.
- FIG. 7 illustrates the use of a film strip view 740 , where the user can click on a given image and see it displayed at a larger resolution in area 710 ; the user can also navigate forward and backwards in a grouping-appropriate manner, such as by chronological order, using a previous button 720 and a forward button 730 .
- the user can have the ability to confirm the appropriateness of a returned result, and the system can then automatically incorporate the user-provided search terms into the corresponding image descriptor; it can also incorporate a subset of the search terms, according to their category, into the image descriptors for other images related to that image, such as images from the same event or place.
- Such tags can also be directly added into the image metadata embedded within the image file itself.
- some embodiments of this invention can apply the concept expansion step directly to each image's descriptors to generate additional candidate concepts to apply to each image, apart from and without requiring the user to perform a search operation.
- Such concepts can be presented to the user as suggested tags or annotations, or otherwise be used to facilitate navigation and browsing the image collection.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Databases & Information Systems (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Library & Information Science (AREA)
- Multimedia (AREA)
- Computational Linguistics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
- The present invention relates to identifying one or more particular images from an image collection.
- Personal image collections can easily include thousands or tens of thousands of images. As image collections grow, retrieving individual images becomes increasingly difficult. Various image retrieval systems have been deployed in order to address this problem. A familiar paradigm for searching documents is one where the user provides a set of search terms and the system returns a list of documents satisfying those search terms, ranked in order of how well each document satisfies the specified search terms. This paradigm has been applied to searching for images in applications such as Google Image Search and Flickr, as well as in a variety of desktop applications. A limitation of such systems is that each image must somehow be annotated with terms that might be used as search terms in order for searches to return any results. It is very time consuming to manually apply such annotations, although some systems automatically annotate images using a limited number of concepts. Moreover, users can only search using search terms that match the terms used in the annotations; such vocabularies tend to be very limited and constrained. Work in the area of information retrieval and specifically query processing has considered the problem of query expansion, but it is difficult to expand queries using terms appropriate for the consumer imaging domain. Such systems tend to either return too few results to be useful, or too many results for the user to effectively consider.
- The above-described problems are addressed and a technical solution is achieved by a method of identifying one or more particular images from an image collection, comprising using a data processing system for:
- (a) indexing the image collection to provide image descriptors for each image in the image collection such that each image is described by one or more of the image descriptors;
- (b) receiving a query from a user specifying at least one keyword for an image search;
- (c) using the keyword(s) to search a second collection of tagged images to identify co-occurrence keywords;
- (d) using the identified co-occurrence keywords to provide an expanded list of keywords;
- (e) using the expanded list of keywords to search the image descriptors to identify a set of candidate images satisfying the keywords;
- (f) grouping the set of candidate images according to at least one of the image descriptors, and selecting one or more representative images from each grouping; and
- (g) displaying the representative images to the user.
- An advantage of the present invention is that it enables users to search their personal image collections using arbitrarily complex and potentially obscure search terms, without requiring the user to provide annotations for the images and without requiring the system to incorporate indexers for recognizing the corresponding concepts. The present invention enables semantically complex concepts to be automatically associated with images for which only low-level semantic concepts are available. The present invention further improves the user's ability to interpret, navigate and view search results.
- The present invention will be more readily understood from the detailed description of exemplary embodiments presented below considered in conjunction with the attached drawings, of which:
-
FIG. 1 is a high-level diagram showing the components of a system for receiving and processing image queries according to an embodiment of the present invention; -
FIG. 2 is a flow diagram illustrating the major steps in receiving and processing media queries according to an embodiment of the present invention; -
FIG. 3 is a flow diagram illustrating the steps used to transform a user-provided query into a set of keywords that can be used as search terms; -
FIG. 4 is a flow diagram illustrating the steps used to compute a co-occurrency matrix; -
FIG. 5 illustrates an example user interface which can be used to manually tag images as part of creating a tagged image collection; -
FIG. 6 illustrates an example user interface for searching an image collection; and -
FIG. 7 illustrates an example user interface for viewing and navigating a group of search results. - It is to be understood that the attached drawings are for purposes of illustrating the present invention.
- Searching for images in personal image collections is becoming increasingly difficult, especially as image collections grow to include thousands or tens of thousands of images. A user can have in mind some concepts, and wish to retrieve images associated with that concept. Or an individual can recall a specific event, scene or object and wish to retrieve any corresponding pictures they might have. However, unless the user has manually annotated the images in their collection according to some concept ontology, they typically are unable to search based on those concepts.
- Researchers have developed a limited number of algorithms for automatically labeling images with specific concepts from a limited vocabulary, using trained concept detectors. Once such algorithms have been used to label images, the images can then be searched using search terms from the corresponding concept vocabulary. However, the concept vocabulary is typically very small and often generic in meaning, limiting the expressiveness of searches. People using such systems must be familiar with the limited vocabulary in order to write searches. Often too many images will match a generic concept, limiting the effectiveness of the search. If the user wishes to search on a term not included in the vocabulary, the search will fail. Some researchers have investigated ways to take a prescribed set of concepts that form a query, and expand them using some sort of concept expansion algorithm. Concept expansion algorithms that are specific to the query domain are most effective.
- The present invention combines automatic labeling with domain-specific concept expansion to enable image retrieval from unannotated or minimally annotated image collections. This approach can permit a user to search collections using queries containing semantically complex concepts or keywords. Using this approach, a user can search their personal image collection for pictures of manatees, for example, and get appropriate results, even though none of the pictures were tagged as containing manatees nor does the system include an algorithm for identifying manatees. Instead, the system relies on combining low level or common concepts for which there are automated algorithms (although the user is also free to add such tags if desired). Continuing the example, algorithms for identifying water are relatively robust, and can be combined with information obtained from reverse geocoding latitude and longitude information which is increasingly recorded as part of the image capture process. Such information can be applied to a picture of an object taken in a river in a Florida wildlife state park, permitting the system to conclude that this image is more likely to contain manatees than other images in the user's collection. As another example, it can be difficult to automatically associate the term “piano recital” with a given video clip taken at a child's first piano recital. However, identifying that the associated audio track contains instrumental music and combining that with knowledge that the video was captured at a music school in the evening can again permit the system to conclude that the given video is more likely than other videos to be a video of a piano recital.
- Although this approach alone will often provide satisfactory results, in some cases it will return results that do not satisfy the search criteria directly, but are nevertheless related in some way to the correct results. A feature of the present invention is that the returned results can be grouped according to logical groupings such as events or visual similarity, and presented according to such groupings. Such groups permit the representation of the results to be condensed, enabling the user to quickly identify any groupings that might contain the desired results. A grouping can be visually represented by picking a representative image from the set, such as a mini collage of a subset of the images, or by some other appropriate representation. The user can then readily expand the selected groupings to reveal the specific intended target, if the intended target was not itself originally displayed as one of the search results. The present invention enables successful concept-based image retrieval on large personal image collections, even in the absence of user-provided annotations.
- Although the present invention is written specific to retrieving image objects from an image collection, the concept described here is not limited to images. For example, the present invention can be applied to search for music, video or document files.
-
FIG. 1 illustrates asystem 100 for identifying one or more particular images from an image collection, according to an embodiment of the present invention. Thesystem 100 includes adata processing system 110, aperipheral system 120, auser interface system 130, and a processor-accessibledata storage system 140. The processor-accessibledata storage system 140, theperipheral system 120, and theuser interface system 130 are communicatively connected to thedata processing system 110. - The
data processing system 110 includes one or more data processing devices that implement the processes of the various embodiments of the present invention, including the example processes ofFIGS. 2-4 described herein. The phrases “data processing device” or “data processor” are intended to include any data processing device, such as a central processing unit (“CPU”), a desktop computer, a laptop computer, a mainframe computer, a personal digital assistant, a Blackberry™, a digital camera, cellular phone, or any other device for processing data, managing data, or handling data, whether implemented with electrical, magnetic, optical, biological components, or otherwise. - The processor-accessible
data storage system 140 includes one or more processor-accessible memories configured to store information, including the information needed to execute the processes of the various embodiments of the present invention, including the example processes ofFIGS. 2-4 described herein. In particular, thedata storage system 140 includes animage collection 145 and a separate taggedimage collection 155. The processor-accessibledata storage system 140 can be a distributed processor-accessible data storage system including multiple processor-accessible data storage systems communicatively connected to thedata processing system 110 via a plurality of computers or devices. On the other hand, the processor-accessibledata storage system 140 need not be a distributed processor-accessible data storage system and, consequently, can include one or more processor-accessible memories located within a single data processor or device. - The phrase “processor-accessible memory” is intended to include any processor-accessible data storage device, whether volatile or nonvolatile, electronic, magnetic, optical, or otherwise, including but not limited to, floppy disks, hard disks, Compact Discs, DVDs, flash memories, ROMs, and RAMs.
- The phrase “communicatively connected” is intended to include any type of connection, whether wired or wireless, between devices, data processors, or programs in which data can be communicated. Further, the phrase “communicatively connected” is intended to include a connection between devices or programs within a single data processor, a connection between devices or programs located in different data processors, and a connection between devices not located in data processors at all. In this regard, although the processor-accessible
data storage system 140 is shown separately from thedata processing system 110, one skilled in the art will appreciate that the processor-accessibledata storage system 140 can be stored completely or partially within thedata processing system 110. Further in this regard, although theperipheral system 120 and theuser interface system 130 are shown separately from thedata processing system 110, one skilled in the art will appreciate that one or both of such systems can be stored completely or partially within thedata processing system 110. - The
peripheral system 120 can include one or more devices configured to provide images to thedata processing system 110. For example, theperipheral system 120 can include digital video cameras, cellular phones, regular digital cameras, or other data processors. Thedata processing system 110, upon receipt of images from a device in theperipheral system 120, can store such images in theimage collection 145 in the processor-accessibledata storage system 140. - The
user interface system 130 can include a mouse, a keyboard, another computer, or any device or combination of devices from which data is input to thedata processing system 110. In this regard, although theperipheral system 120 is shown separately from theuser interface system 130, theperipheral system 120 can be included as part of theuser interface system 130. - The
user interface system 130 also can include a display device, a processor-accessible memory, or any device or combination of devices to which data is output by thedata processing system 110. In this regard, if theuser interface system 130 includes a processor-accessible memory, such memory can be part of the processor-accessible memory system 140 even though theuser interface system 130 and the processor-accessibledata storage system 140 are shown separately inFIG. 1 . -
FIG. 2 illustrates a method for identifying one or more particular images from an image collection, according to an embodiment of the present invention. This method can be implemented by thedata processing system 110 inFIG. 1 . The system takes as its input an image collection 205 (equivalently,image collection 145 inFIG. 1 ) which is indexed instep 210 to produce a set ofimage descriptors 215. The indexing process includes a variety of indexers, including temporal event clustering on groups of assets as well as temporal indexing based on individual assets; scene, activity, object and material classifiers; face-based age and gender estimators; color; camera capture metadata including, for example, Exif metadata; people recognition and reverse geocoding. These indexers cover the one or more of the usual who, what, when, and where vectors that can be used as the basis for a search for image objects. These indexers can all be implemented by thedata processing system 110 inFIG. 1 . - Individual images can be indexed based upon their capture date and time. The date and time information can be readily mapped to concepts such as seasons of the year or other temporal periods. In addition, by referring to one or more personal calendars associated with individuals portrayed in the asset or who captured the asset, concepts such as birthday, anniversary or even the names of specific personal events such as “Florida vacation” can be associated with assets. Appropriate civil and religious calendars can be used to further associate the names of holidays such as Christmas or Independence Day with individual assets or event-based clusters. Although the present invention does not require knowledge of such specific events and in fact provides a way to search for such concepts without such tags having been associated with images, it nevertheless can benefit from such tags.
- Temporal event clustering of images in the
image collection 205 is generated by automatically sorting, segmenting, and clustering an unorganized set of media assets into separate temporal events and sub-events, as described in detail in commonly assigned U.S. Pat. No. 6,606,411 entitled “A method for Automatically Classifying Images Into Events,” issued Aug. 12, 2003, and commonly assigned U.S. Pat. No. 6,351,556, entitled “A Method for Automatically Comparing Content of Images for Classification Into Events,” issued Feb. 26, 2002. Given the division of images into events, an event classification algorithm can be employed as described in US Patent Application Publication 2010/0124378 entitled “Method for Event-Based Semantic Classification” to classify the event into one of four types: family, vacation, sports or family moment. Other event classification algorithms can be equally applied, resulting in potentially other event types. - Scene classifiers identify or classify an image as representing one or more scene types such as mountain, beach or indoor. Material and texture classifiers identify or classify an image as containing one or more materials or texture types, such as rock, sand, grass, or blue sky. Object classifiers identify or classify an image as containing one or more objects, such as car, cat, or bicycle. Some example scene and material classification types include indoor, outdoor, natural, urban, sunset, beach, foliage, field, mountain, sky, grass, snow, water and sand. Operation of such classifiers are described in U.S. Pat. No. 6,282,317 entitled “Method for automatic determination of main subjects in photographic images”; U.S. Pat. No. 6,697,502 entitled “Image processing method for detecting human figures in a digital image assets”; U.S. Pat. No. 6,504,951 entitled “Method for Detecting Sky in Images”; U.S. Patent Application Publication 2005/0105776 entitled “Method for Semantic Scene Classification Using Camera Metadata and Content-based Cues”; U.S. Patent Application Publication 2005/0105775 entitled “Method of Using Temporal Context for Image Classification”; and U.S. Patent Application Publication 2004/003746 entitled “Method for Detecting Objects in Digital image assets.” Research in image understanding algorithms continues to result in new and improved scene, material activity and object classifiers and recognizers, which can easily be incorporated into the invention.
- Color classifiers identify or classify an image as containing one or more prominent colors. Such classifiers can compute a histogram of the different colors present in the image, potentially grouping together pixels of nearly identical coloring. The most frequently occurring colors can be mapped to their common color names, which can then be associated with the image as image descriptors.
- A face detector algorithm is used to find faces in image collections, and is described in U.S. Pat. No. 7,110,575 entitled “Method for Locating Faces in Digital Color Images”, issued Sep. 19, 2006; U.S. Pat. No. 6,940,545 entitled “Face Detecting Camera and Method” issued Sep. 6, 2005; U.S. Patent Application Publication 2004/0179719 entitled “Method and system for face detection in digital image assets” filed Mar. 12, 2003.
- Given face detection, the approximate age of an individual can be estimated as described in the work by A. Lanitis, C. Draganova, and C. Christodoulou, “Comparing different classifiers for automatic age estimation,” IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics, vol. 34, no. 1, pp. 621-628, February 2004. Likewise, the gender of a person can be predicted based upon the facial shape as described in the literature. Other cues such as clothing, hair and social context can be further used to improve upon the age and gender estimation, as published in the literature, e.g., the Ph.D. dissertation by Andrew C. Gallagher entitled “A Framework for Using Context to Understand Images of People,” published by Carnegie Mellon University, May 2009. Given an estimated age and gender, a high-level semantic label can be associated with the face or the containing image object. For example, given an estimated age of less than one year; the system can generate the label “baby.”Given a face estimated to be approximately age 29 and female, the label “woman” can be generated. The presence of multiple faces in an image can further result in additional descriptive labels being generated. For example, the presence of multiple faces estimated to be age 16 or less can result in the label “children.” Other labels include “man,” “boy,” “girl,” and “teen.” As described in the previously cited dissertation by Andrew C. Gallagher, the presence of multiple faces also can be further used to refine the individual age and gender estimates of each individual in the image. The presence of multiple detected people can be used to tag or label an image as being a group shot; estimates of age and gender of the individuals in the group shot, possibly combined with other contextual cues, can provide additional estimates as to the relative relationships and nature of the group shot, such as a family photo. Such estimates can be used to generate additional semantic labels.
- Face recognition is the identification or classification of a face to an example of a person or a label associated with a person based on facial features as described in U.S. Patent Application Publication 2008/0112621 entitled “User interface for face recognition”; U.S. Pat. No. 7,711,145 entitled “Finding Images with Multiple People or Objects”; and U.S. Patent Application Publication 2007/0098303 entitled “Determining a Particular Person from a Collection”.
- Reverse geocoding is the process of mapping geospatial coordinates such as latitude and longitude to specific place names or geographic feature types. Any of a number of commonly available services for reverse geocoding, such as the service offered by geonames.org, can be employed to perform such mapping. Furthermore, such services can be used to obtain not only nearby feature points, but also geospatial feature types, such as lake, school, beach or park. Heuristics such as those described in the section on geospatial classification in the work by Mark D. Wood entitled “Matching Songs to Events in Image Collections,” IEEE International Conference on Semantic Computing, 2009. ICSC '09., pp. 95-102, 14-16 Sep. 2009, can be used to associate one or more feature types with an image.
- As noted previously, although the invention is described here in the context of searching for image objects, the technique described herein applies to other media object types. The described invention can also be applied with minor variations to any object type for which one might expect to find an independent set of tagged objects of possibly the same or a different type, where the tags are from a vocabulary that one might use as search terms. One type would be music or audio files. In the case of music objects, the descriptors can include musical genre, artist, album, recording location, and mood. Video is another type; most of the types associated with still image objects also apply to audio. Other types include special sound effects, olfactory recordings or food recipes. Other classifier types can be applicable for other data types and can be used as indexers in addition to or besides the types of classifiers for described here. For example, activity classifiers recognizing activities such as swimming, running, or singing can be employed for video objects.
- In addition to descriptors generated algorithmically, the set of descriptors for an image can include descriptors provided directly or indirectly by the user. For example, many users will rename the files corresponding to images of particular importance to them, giving the file a more descriptive name than the name originally used by the capture device. Likewise, users can organize images into folders or directories, giving the directory or folder a meaningful name. In some cases, users will also associate a caption with an image. Information can also be indirectly associated with an image. U.S. Pat. No. 7,025,253 entitled “Agent for integrated annotation and retrieval of images” describes a mechanism for automatically associating suggested annotations with an image.
- The following is an example of one embodiment of the descriptors, where the descriptors for a given image object are represented as a distinct XML document.
-
<DOC> <docno>urn:guid:fe5ffb3b-b305-4e02-85cd- 55f4c4cc9be2</docno> <title>C:\DATA\USERS\WOOD\My Pictures\GeoFamily\2002-2005\2005\Florida\100_0438.JPG</title> <author>urn:guid:0e9c24e1-0731-4dff-aa8e- 802567f3a9f7</author> <text> <nonprob> <path>C:\DATA\USERS\WOOD\My Pictures\GeoFamily\2002-2005\2005\Florida\100_0438.JPG</path> <label>Valentine's Day 2005</label> <temporal>February 14, 2005 winter afternoon</temporal> <subject/> <address>Homosassa, Florida, US</address> <featureClasses>park spring(s)</featureClasses> <nearbyFeatures>Homosassa Springs: spring(s) Homosassa Springs Wildlife State Park: park</nearbyFeatures> </nonprob> <prob> <type>Sports</type> <topType>Vacation</topType> <scenes>rock sand grass water sky outdoor </scenes> <colors>gray gray green blue </colors> </prob> </text> </DOC> - Numerous alternative representations can be used, both apart from and within the image object. Mechanisms for storing the descriptors within the image object can be used for container-style file formats, such as Exif or MOV files. Some image descriptors have definitions as part of standardized formats such as Exif. Other descriptors can be incorporated into an existing image file format using extension mechanisms. One popular way to include metadata within documents is to use the Extensible Metadata Platform (XMP). Generated metadata can also be stored directly in a database.
- Probabilities or numeric scores can also be stored with the metadata properties. For example, the XML elements scenes and colors shown above might in one embodiment include numeric data representing the degree of belief, confidence, probability, or some other numeric score, as illustrated here:
-
<scenes>rock 0.11014811 sand 0.12858073 grass 0.10345459 water 0.56204224 sky 0.44345093 outdoor 0.873681 </scenes> <colors>gray 0.193939 green 0.0938212 blue 0.460917 </colors> - Alternatively, other mechanisms can be used to represent different levels of confidence or probability, including discrete categories such as high, medium and low. Again such values can be represented in any of a number of ways familiar to those practiced in software engineering.
- Referring back to
FIG. 2 , instep 220, the system receives aquery 225 from the user. Thisquery 225 can be entered by the user using theuser interface system 130 ofFIG. 1 and can be entered using any of a variety of mechanisms, including the user typing thequery 225 into a dialog box, speaking the query into a microphone, or indicating the query by selecting items from a checklist. In one embodiment, thequery 225 is expressed as a set of keywords, such as “camping Adirondacks.” However, in other embodiments, the user can enter thequery 225 as a phrase or sentence in either an imperative or descriptive form, such as, “find pictures of my camping trip in the Adirondacks” or “Michelle's pet rabbit.” Thequery 225 is processed instep 230 by thedata processing system 110 ofFIG. 1 to produce a set ofkeywords 235 to use as search terms. - The mechanics of
step 230 are shown inFIG. 3 , which illustrates the steps used to transform a user-providedquery 305 into a set ofkeywords 335 that can be used as search terms. Instep 310, the set ofkeywords 315 are extracted from the user-providedquery 305. If the user provided a phrase or sentence, commonly practiced natural language processing techniques are applied to extract keywords. If the user input is already in the form of a string of keywords, then the string must simply be parsed into a set ofkeywords 315. Either way, the resulting set ofkeywords 315 is further processed instep 320 to eliminate any stop words or commonly used terms such as “picture,” producing a filtered set ofkeywords 325. Depending on the system, stopwords can also be partially or completely eliminated as part ofstep 310, especially if natural language processing techniques are applied to extractkeywords 315 from natural language phrases. The filtered set ofkeywords 325 can then be expanded instep 330. - The
expansion process 330 operates by considering eachkeyword 325, and determining whatother keywords 325 are most likely to co-occur with thatkeyword 325 in the tagged second collection, to produce an expanded set of keywords 335 (equivalently, set ofsearch keywords 235 fromFIG. 2 ). Given a suitable taggedimage collection 155 fromFIG. 1 , the computation of co-occurrent keywords can be computed by one skilled in the art.FIG. 4 illustrates the process of computing a co-occurrency matrix. A tagged image collection 405 (equivalently, 155 fromFIG. 1 ) is processed bystep 410 to compute aco-occurrency matrix 415. - The following pseudocode illustrates one
algorithm implementing Step 410 for computing theco-occurrency matrix 415 ofFIG. 4 . In this pseudocode, theco-occurrency matrix 415 is denoted by the variable Cooccurs. Upon completion of the algorithm, the value Cooccurs[a,b] is the frequency of times tag b co-occurs with tag a. In addition, this algorithm illustrates in lines 7 and 8 the optional steps used by some embodiments to maintain a dictionary of all terms known by the system, and the set of categories associated with each term. For a given tag t occurring in image Img, the Category(t) of the tag can be determined based upon the semantic classifier that generated the tag. For example, inFIG. 3 , the various XML element names such as colors, scenes and temporal can denote the corresponding categories. A given tag can potentially be generated by multiple classifiers, and therefore have multiple associated categories. -
1. Let S = the set of Tagged Images 2. For each image i in S do 3. Let T = the set of tags associated with image i 4. For each t in T do 5. For each t′ in T do 6. Increment CooccursCounts[t, t′] 7. If t′ is not in Dictionary, add t′ to Dictionary 8. If Category(t′) is not in the list of categories associated with t′ in Dictionary, add Category(t′) to the set of categories for t′ 9. For each tag r do 10. For each tag t do 11. Set Cooccurs[t,r] = CooccursCounts[t,r] / CooccursCounts[t,t] - Those skilled in the art will recognize that these values or functionally similar values can be computed in alternative ways. Likewise, alternative formulations can result in different definitions for co-occurrence, although still providing data useful for estimating the likelihood two tags or concepts are likely to co-occur.
-
FIG. 5 illustrates a simple user interface for tagging pictures. The user can enter a new tag intotext box 510 and then click onbutton 520 labeled “Add new tag” to add the tag to the image. The list of tags currently associated with the image is shown inarea 530. The tagging application can be part of a desktop application, or it can be offered as part of some on-line photo sharing service such as Flickr, Photobucket, or the Kodak Gallery. Similar services such as Vimeo and YouTube provide tagging functionality for videos. The tagging system can be completely manual; alternatively, it can recommend or suggest tags to a user. Tags can be recommended by using their likelihood of co-occurrence with tags already associated with an image. Some systems automatically generate tags based upon capture metadata, user-provided information, or other information available to the system. Any of these embodiments enable tag information to be collected over a set of images. The images can all belong to a single user, or they can come from thousands or millions of users. In some embodiments, the tagging system can further differentiate between tags originating from different users or different types of images, enabling different Cooccurs matrices to be computed. For example, the system can compute different Cooccurs matrices for tags originating from different user demographics, where the demographics might include language, age, or geographic region. This permits the system to account for different uses of tags depending upon language, age or geography. In such embodiments, the system instep 330 would pick the Cooccurs matrix most appropriate for the current user. - Given the Cooccurs matrix, the
step 330 can be carried out using the following procedure: - For each keyword k in the set of filtered
keywords 325 -
- Let Cooccurs(k) be the set of all keywords k′ such that Cooccurs(k, k′) is greater than some threshold t
- The set Cooccurs(k) is the set of co-occurrent keywords computed by
step 330. In some embodiments, the set of co-occurrent keywords can be further filtered, to ensure that the computed keywords are mutually consistent with other search terms provided by the user. This check for consistency is carried out by a modifiedStep 330, as follows: - For each keyword k in the set of filtered
keywords 325 -
- Let Cooccurs(k) be the set of all keywords k′ such that Cooccurs(k, k′) is greater than some threshold t and for each keyword k″ in the set of filtered keywords, k″ # k,
- Cooccurs(k′, k″) is greater than some threshold t′
- Let Cooccurs(k) be the set of all keywords k′ such that Cooccurs(k, k′) is greater than some threshold t and for each keyword k″ in the set of filtered keywords, k″ # k,
- For example, suppose the co-occurrence frequency for the terms sky, beach, clouds, architecture, London, snow and vacation is as follows:
-
archi- sky beach clouds tecture London snow vacation sky 1.00 0.08 0.37 0.05 0.01 0.04 0.03 beach 0.12 1.00 0.09 0.01 0.00 0.01 0.04 clouds 0.58 0.10 1.00 0.04 0.01 0.04 0.05 architecture 0.09 0.01 0.04 1.00 0.04 0.01 0.05 London 0.03 0.00 0.02 0.05 1.00 0.01 0.01 snow 0.09 0.01 0.06 0.02 0.01 1.00 0.17 vacation 0.16 0.13 0.15 0.15 0.03 0.34 1.00 - The behavior of the filtering algorithm can be illustrated by applying the above example co-occurrence matrix to the sample query “London vacation,” where the threshold t=0.13 and the threshold t′ is 0.04. Using the above example data, and taking the keyword vacation, the co-occurrence frequencies for sky, beach, clouds, architecture and snow are all greater than or equal to the threshold t=0.13. In particular, Cooccurs[vacation, sky]=0.16, Cooccurs[vacation,beach]=0.13, Cooccurs[vacation, clouds]=0.15, Cooccurs[vacation, architecture]=0.15, and Cooccurs[vacation, snow]=0.34. However, only Cooccurs[London, architecture], which equals 0.05 in the example matrix, is greater than threshold t′=0.04, resulting in Cooccurs(vacation) having the value of {vacation, architecture} in the context of the example query “London vacation.”
- Some embodiments applying filtering can apply it selectively to keywords, depending on the keyword. For example, the filtering step might only be applied to keywords corresponding to semantic concepts returned by image understanding algorithms such as event classification or scene recognition, but not filter keywords corresponding to concepts extracted from an image filename or path, or that were deduced based upon calendar information and the image capture date. By computing the optional category information as described previously as part of
Step 410, some embodiments ofStep 330 can incorporate a lookup operation to determine the category of a keyword, either provided as an original search term or obtained from a prior expansion, and adjust the filtering behavior accordingly. By selectively filtering the categories where the expansion is believed to be most “noisy,” the performance of the system can be improved, both in the quality of the results and the required computation time. - The
expansion process 330 uses a single co-occurrence matrix to carry out both the expansion and filtering operations; those skilled in the art will readily recognize other improvements that can be readily made. For example, a separate source of information can be used for the filtering operation. This information can take the form of a co-occurrence matrix computed from a third source of tagged information. Another source of information is to consider the co-occurrence of terms in knowledgebases such as Wikipedia, or even general Internet search results. Some embodiments can prefer to incorporate rules for filtering based upon the category. - The
expansion process 330 can also be applied iteratively, so that each set of expanded keywords is further expanded using the same algorithm. The iteration can continue a bounded number of times, or until some other constraint is satisfied. - Referring back to
FIG. 2 , step 230 can also incorporate other mechanisms not shown inFIG. 3 in producing the set ofsearch keywords 235. For example, one technique applied in the query expansion literature is to use a lexical database such as WordNet to identify related words; these related words can be included in the set of expanded search keywords. Other sources of knowledge such as gazetteers, Cyc, or Wikipedia can also be used to provide related concepts, using the term or a derivative word of the term, such as the stemmed word form. Furthermore, step 230 can combine the mechanism illustrated inFIG. 3 with other expansion mechanisms in an interleaved or iterative fashion. The expanded search terms can further have weights associated with each term, indicating the strength of the expected significance of that term. For example, the weights can be proportional to the co-occurrency frequency. In addition, terms whose co-occurrency frequency is sufficiently low can be completely excluded from the set of expanded terms. - In
step 240, thesearch keywords 235 are used to identify a set ofcandidate images 245. In one embodiment,step 240 is carried out by indexing theimage descriptors 215 using an information retrieval system such as the Indri information retrieval system from the University of Massachusetts, which is described in T. Strohman, D. Metzler, H. Turtle, and W. B. Croft, “Indri: A language-model based search engine for complex queries (extended version),” Center for Intelligence Information Retrieval Technical Report, 2005. This information retrieval system supports complex queries including weighted search terms, and quickly returns matching documents. This embodiment assigns different weights to each search term in the expanded list ofkeywords 335 inFIG. 3 , depending on the origin. For example,keywords 335 provided by the user are given a higher weight than terms obtained through the expansion process. The weight for expandedkeywords 335 can be further adjusted based upon their frequency of co-occurrence, their category, or how many iterations of the expansion process were required to result in the word being added to the expansion list. Search terms obtained from alternate sources such as WordNet can have separate weights, with the weight based upon the expanded term's relationship to the term being expanded. For example synonyms can be assigned a higher weight than hypernyms and hyponyms. - Using the
image descriptors 215 as the indexed documents, Indri will identify which documents satisfy the specified search terms expressed as an Indri query. By incorporating into each document a reference to the original image, the results from Indri can be easily mapped to the actual image objects. Of course, one skilled in the art will recognize that a search system also can be more directly implemented, including one that accommodates the storage of the image descriptors directly within the image objects. Indri combines the inference network model of information retrieval with a language model for estimating the probability that a given term is satisfied by a given document. However other models known to those skilled in the art of information retrieval can also be applied instep 240 to determine which candidate images satisfy the search keywords. - Step 240 produces the set of
candidate images 245, which in some embodiments are ranked or ordered according to the estimated strength of their match. Instep 250, the candidate images are grouped to produce sets of groupedimages 255. The grouping process is based on some commonality in the candidate images with respect to one or more of the image descriptors. For example, in some embodiments, images are grouped by event, so that all images which were taken as part of the same event, using the previously mentioned temporal event clustering algorithm to determine event boundaries. In other embodiments, images can be grouped based upon location, based upon the people identified in the image, based upon visual similarity, or based upon scene, material or activity similarity. The user can have the ability to determine the mode of grouping the candidate images, or the system can automatically chose an appropriate grouping method by analyzing the images. - In
Step 260, a representative image is selected from each candidate grouping to produce a set ofrepresentative images 265. Therepresentative image 265 can be selected using any of a variety of algorithms. For example, therepresentative image 265 can be the image from the group that scored the highest according to the search criteria. Therepresentative image 265 can also be selected according to some other scoring function partially or completely independent of the search ranking For example, the images in a group can be ranked according to their expected aesthetic appeal or information content. Alternatively, therepresentative image 265 for a grouping can be a new image, formed by compositing a plurality of the images in the candidate group. - In
Step 270, the set of representative images is displayed to the user. These images can be displayed using theuser interface system 130 ofFIG. 1 . Theuser interface system 130 can further provide the user with a variety of choices for controlling the display. For example, the user can elect to have all results displayed, not just therepresentative images 265 chosen instep 260. Or the user can request that a different mechanism be used to pick therepresentative images 265, resulting inStep 260 being repeated using the user's specified criteria for selectingrepresentative images 265. -
FIG. 6 illustrates one embodiment for the user interface. Inquery text box 610, the user entered the search terms “camping Adirondacks.” The system displayed the resultingrepresentative images 265 in ascrollable region 640, with the corresponding image pathnames and search score shown inscrollable area 630. Inarea 620, the user can see the actual query that was formed from the set of expanded search terms. Providing the user with the ability to see the query formed from the search term permits the sophisticated user the ability to further refine the search terms. Some embodiments can prefer to show the user the list of expanded search terms using an alternative user interface, perhaps in the form of checkboxes, permitting the user to deselect any inappropriate search terms. However, some embodiments would conceal that level of complexity from the average user. - In the user interface illustrated in
FIG. 6 , the user can click on an image shown inscrollable region 640 to see other images from the same candidate grouping. For example, the user can click onimage 645 to see other images in the same event, as illustrated inFIG. 7 .FIG. 7 illustrates the use of afilm strip view 740, where the user can click on a given image and see it displayed at a larger resolution inarea 710; the user can also navigate forward and backwards in a grouping-appropriate manner, such as by chronological order, using aprevious button 720 and aforward button 730. - In some embodiments, the user can have the ability to confirm the appropriateness of a returned result, and the system can then automatically incorporate the user-provided search terms into the corresponding image descriptor; it can also incorporate a subset of the search terms, according to their category, into the image descriptors for other images related to that image, such as images from the same event or place. Such tags can also be directly added into the image metadata embedded within the image file itself. Furthermore, some embodiments of this invention can apply the concept expansion step directly to each image's descriptors to generate additional candidate concepts to apply to each image, apart from and without requiring the user to perform a search operation. Such concepts can be presented to the user as suggested tags or annotations, or otherwise be used to facilitate navigation and browsing the image collection.
- It is to be understood that the exemplary embodiment(s) is/are merely illustrative of the present invention and that many variations of the above-described embodiment(s) can be devised by one skilled in the art without departing from the scope of the invention. It is therefore intended that all such variations be included within the scope of the following claims and their equivalents.
-
- 110 data processing system
- 120 peripheral system
- 130 user interface system
- 140 data storage system
- 145 image collection
- 155 tagged image collection
- 205 image collection
- 210 step
- 215 image descriptors
- 220 step
- 225 query
- 230 step
- 235 search keywords
- 240 step
- 245 candidate images
- 250 step
- 255 grouped images
- 260 step
- 265 representative images
- 270 step
- 305 user provided query
- 310 step
- 315 extracted keywords
- 320 step
- 325 filtered keywords
- 330 step
- 335 expanded set of keywords
- 405 tagged image collection
- 410 step
- 415 co-occurrency matrix
- 510 text box
- 520 button
- 530 area
- 610 query text box
- 620 area
- 640 scrollable region
- 645 image
- 710 area
- 720 previous button
- 730 forward button
Claims (18)
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/021,188 US8612441B2 (en) | 2011-02-04 | 2011-02-04 | Identifying particular images from a collection |
US14/079,787 US9037569B2 (en) | 2011-02-04 | 2013-11-14 | Identifying particular images from a collection |
US14/691,265 US9524349B2 (en) | 2011-02-04 | 2015-04-20 | Identifying particular images from a collection |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/021,188 US8612441B2 (en) | 2011-02-04 | 2011-02-04 | Identifying particular images from a collection |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/079,787 Continuation US9037569B2 (en) | 2011-02-04 | 2013-11-14 | Identifying particular images from a collection |
Publications (2)
Publication Number | Publication Date |
---|---|
US20120203764A1 true US20120203764A1 (en) | 2012-08-09 |
US8612441B2 US8612441B2 (en) | 2013-12-17 |
Family
ID=46601385
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/021,188 Active 2031-02-25 US8612441B2 (en) | 2011-02-04 | 2011-02-04 | Identifying particular images from a collection |
US14/079,787 Active US9037569B2 (en) | 2011-02-04 | 2013-11-14 | Identifying particular images from a collection |
US14/691,265 Active US9524349B2 (en) | 2011-02-04 | 2015-04-20 | Identifying particular images from a collection |
Family Applications After (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/079,787 Active US9037569B2 (en) | 2011-02-04 | 2013-11-14 | Identifying particular images from a collection |
US14/691,265 Active US9524349B2 (en) | 2011-02-04 | 2015-04-20 | Identifying particular images from a collection |
Country Status (1)
Country | Link |
---|---|
US (3) | US8612441B2 (en) |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120307094A1 (en) * | 2009-08-28 | 2012-12-06 | Nikon Corporation | Image file data structure, image file generation device, image file generation method, and electronic camera |
US20130051687A1 (en) * | 2011-08-25 | 2013-02-28 | Canon Kabushiki Kaisha | Image processing system and image processing method |
US20130144975A1 (en) * | 2011-12-05 | 2013-06-06 | Seiji Miyahara | Image information processing apparatus and image information processing method |
US20130251198A1 (en) * | 2012-03-26 | 2013-09-26 | Canon Kabushiki Kaisha | Information processing apparatus, control method thereof, and storage medium |
CN104751168A (en) * | 2013-12-30 | 2015-07-01 | 宏达国际电子股份有限公司 | Method for searching relevant images via active learning, electronic device using the same |
US20160132194A1 (en) * | 2014-11-06 | 2016-05-12 | Dropbox, Inc. | Searching digital content |
US20160283557A1 (en) * | 2015-03-27 | 2016-09-29 | Orange | Method and device for accessing a plurality of contents, corresponding terminal and computer program |
US9576203B2 (en) * | 2015-04-29 | 2017-02-21 | Canon Kabushiki Kaisha | Devices, systems, and methods for knowledge-based inference for material recognition |
CN106649610A (en) * | 2016-11-29 | 2017-05-10 | 北京智能管家科技有限公司 | Image labeling method and apparatus |
CN107766394A (en) * | 2016-08-23 | 2018-03-06 | 阿里巴巴集团控股有限公司 | Business data processing method and its system |
WO2018040059A1 (en) * | 2016-09-02 | 2018-03-08 | Microsoft Technology Licensing, Llc | Clip content categorization |
US10002375B1 (en) * | 2014-06-10 | 2018-06-19 | Amazon Technologies, Inc. | Hashtag shopping and rating |
US20180349467A1 (en) * | 2017-06-02 | 2018-12-06 | Apple Inc. | Systems and methods for grouping search results into dynamic categories based on query and result set |
US10162865B2 (en) * | 2015-10-08 | 2018-12-25 | Microsoft Technology Licensing, Llc | Generating image tags |
CN109088979A (en) * | 2018-06-28 | 2018-12-25 | 珠海格力电器股份有限公司 | Collection control method, device, storage medium and terminal |
US10268758B2 (en) * | 2013-09-29 | 2019-04-23 | Peking University Founder Group Co. Ltd. | Method and system of acquiring semantic information, keyword expansion and keyword search thereof |
US10929547B2 (en) | 2015-04-17 | 2021-02-23 | Dropbox, Inc. | Collection folder for collecting file submissions using email |
CN112534421A (en) * | 2018-12-20 | 2021-03-19 | 谷歌有限责任公司 | Map search recommendation system based on image content analysis-driven geographic semantic index |
CN113939813A (en) * | 2019-05-21 | 2022-01-14 | 微软技术许可有限责任公司 | Generating and applying object-level relational indexes for images |
US11403352B2 (en) * | 2015-09-28 | 2022-08-02 | Yahoo Assets Llc | Multi-touch gesture search |
WO2022268094A1 (en) * | 2021-06-22 | 2022-12-29 | Huawei Technologies Co., Ltd. | Methods, systems, and media for image searching |
EP4145305A1 (en) * | 2021-09-06 | 2023-03-08 | Bull SAS | Multifacet and modular search method within a collection of video sequences |
US11948473B2 (en) | 2015-12-31 | 2024-04-02 | Dropbox, Inc. | Assignments for classrooms |
Families Citing this family (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8106856B2 (en) | 2006-09-06 | 2012-01-31 | Apple Inc. | Portable electronic device for photo management |
EP3260969B1 (en) | 2009-09-22 | 2021-03-03 | Apple Inc. | Device, method, and graphical user interface for manipulating user interface objects |
US8780069B2 (en) | 2009-09-25 | 2014-07-15 | Apple Inc. | Device, method, and graphical user interface for manipulating user interface objects |
US8698762B2 (en) | 2010-01-06 | 2014-04-15 | Apple Inc. | Device, method, and graphical user interface for navigating and displaying content in context |
US9171018B2 (en) * | 2012-01-17 | 2015-10-27 | Google Inc. | System and method for associating images with semantic entities |
CN103294684B (en) * | 2012-02-24 | 2016-08-24 | 浙江易网科技股份有限公司 | Association lexical search system and method |
US20140019867A1 (en) * | 2012-07-12 | 2014-01-16 | Nokia Corporation | Method and apparatus for sharing and recommending content |
US9111547B2 (en) * | 2012-08-22 | 2015-08-18 | Kodak Alaris Inc. | Audio signal semantic concept classification method |
US9183849B2 (en) | 2012-12-21 | 2015-11-10 | The Nielsen Company (Us), Llc | Audio matching with semantic audio recognition and report generation |
US9195649B2 (en) * | 2012-12-21 | 2015-11-24 | The Nielsen Company (Us), Llc | Audio processing techniques for semantic audio recognition and report generation |
US9881023B2 (en) * | 2014-07-22 | 2018-01-30 | Microsoft Technology Licensing, Llc | Retrieving/storing images associated with events |
US10198498B2 (en) * | 2015-05-13 | 2019-02-05 | Rovi Guides, Inc. | Methods and systems for updating database tags for media content |
US10430476B2 (en) * | 2016-05-06 | 2019-10-01 | Google Llc | Annotation of videos using aggregated user session data |
EP3465478A1 (en) | 2016-06-02 | 2019-04-10 | Kodak Alaris Inc. | Method for providing one or more customized media centric products |
AU2017100670C4 (en) | 2016-06-12 | 2019-11-21 | Apple Inc. | User interfaces for retrieving contextually relevant media content |
DK201670608A1 (en) | 2016-06-12 | 2018-01-02 | Apple Inc | User interfaces for retrieving contextually relevant media content |
CN106372184A (en) * | 2016-08-31 | 2017-02-01 | 迈普通信技术股份有限公司 | Network equipment and method for preventing filtering escape of Chinese keywords |
TWI647580B (en) * | 2017-06-01 | 2019-01-11 | 正修學校財團法人正修科技大學 | Search filtering method that enhances the matching of text search results |
DK180171B1 (en) | 2018-05-07 | 2020-07-14 | Apple Inc | USER INTERFACES FOR SHARING CONTEXTUALLY RELEVANT MEDIA CONTENT |
DK201970535A1 (en) | 2019-05-06 | 2020-12-21 | Apple Inc | Media browsing user interface with intelligently selected representative media items |
CN110377724A (en) * | 2019-07-01 | 2019-10-25 | 厦门美域中央信息科技有限公司 | A kind of corpus keyword Automatic algorithm based on data mining |
DK202070612A1 (en) | 2020-02-14 | 2021-10-26 | Apple Inc | User interfaces for workout content |
US12072920B2 (en) | 2022-12-12 | 2024-08-27 | Motorola Solutions, Inc. | System and method for summarization of search results |
Citations (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5852823A (en) * | 1996-10-16 | 1998-12-22 | Microsoft | Image classification and retrieval system using a query-by-example paradigm |
US5899999A (en) * | 1996-10-16 | 1999-05-04 | Microsoft Corporation | Iterative convolution filter particularly suited for use in an image classification and retrieval system |
US20030174859A1 (en) * | 2002-03-14 | 2003-09-18 | Changick Kim | Method and apparatus for content-based image copy detection |
US20030198380A1 (en) * | 1999-05-17 | 2003-10-23 | Samsung Electronics Co., Ltd.., University Of , Calif. At Santa Barbara | Color image processing method |
US6751343B1 (en) * | 1999-09-20 | 2004-06-15 | Ut-Battelle, Llc | Method for indexing and retrieving manufacturing-specific digital imagery based on image content |
US6993180B2 (en) * | 2001-09-04 | 2006-01-31 | Eastman Kodak Company | Method and system for automated grouping of images |
US20060147094A1 (en) * | 2003-09-08 | 2006-07-06 | Woong-Tuk Yoo | Pupil detection method and shape descriptor extraction method for a iris recognition, iris feature extraction apparatus and method, and iris recognition system and method using its |
US20080089561A1 (en) * | 2006-10-11 | 2008-04-17 | Tong Zhang | Face-based image clustering |
US7379627B2 (en) * | 2003-10-20 | 2008-05-27 | Microsoft Corporation | Integrated solution to digital image similarity searching |
US20080208791A1 (en) * | 2007-02-27 | 2008-08-28 | Madirakshi Das | Retrieving images based on an example image |
US20080306995A1 (en) * | 2007-06-05 | 2008-12-11 | Newell Catherine D | Automatic story creation using semantic classifiers for images and associated meta data |
US20090161962A1 (en) * | 2007-12-20 | 2009-06-25 | Gallagher Andrew C | Grouping images by location |
US20090324026A1 (en) * | 2008-06-27 | 2009-12-31 | Palo Alto Research Center Incorporated | System and method for finding a picture image in an image collection using localized two-dimensional visual fingerprints |
US20100027895A1 (en) * | 2006-08-31 | 2010-02-04 | Osaka Prefecture University Public Corporation | Image recognition method, image recognition device, and image recognition program |
US7826661B2 (en) * | 2003-07-04 | 2010-11-02 | Mitsubishi Denki Kabushiki Kaisha | Method and apparatus for representing a group of images |
US8005831B2 (en) * | 2005-08-23 | 2011-08-23 | Ricoh Co., Ltd. | System and methods for creation and use of a mixed media environment with geographic location information |
US8027512B2 (en) * | 2005-09-30 | 2011-09-27 | Robert Bosch Gmbh | Method and software program for searching image information |
Family Cites Families (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6243713B1 (en) * | 1998-08-24 | 2001-06-05 | Excalibur Technologies Corp. | Multimedia document retrieval by application of multimedia queries to a unified index of multimedia data for a plurality of multimedia data types |
US6606411B1 (en) | 1998-09-30 | 2003-08-12 | Eastman Kodak Company | Method for automatically classifying images into events |
US6351556B1 (en) | 1998-11-20 | 2002-02-26 | Eastman Kodak Company | Method for automatically comparing content of images for classification into events |
US6470336B1 (en) * | 1999-08-25 | 2002-10-22 | Matsushita Electric Industrial Co., Ltd. | Document image search device and recording medium having document search program stored thereon |
US6504951B1 (en) | 1999-11-29 | 2003-01-07 | Eastman Kodak Company | Method for detecting sky in images |
US6940545B1 (en) | 2000-02-28 | 2005-09-06 | Eastman Kodak Company | Face detecting camera and method |
US6697502B2 (en) | 2000-12-14 | 2004-02-24 | Eastman Kodak Company | Image processing method for detecting human figures in a digital image |
US20040205116A1 (en) * | 2001-08-09 | 2004-10-14 | Greg Pulier | Computer-based multimedia creation, management, and deployment platform |
DE10209257A1 (en) | 2002-03-01 | 2003-09-11 | Philips Intellectual Property | Process for the elastographic examination of an object |
US7664830B2 (en) * | 2002-03-29 | 2010-02-16 | Sony Corporation | Method and system for utilizing embedded MPEG-7 content descriptions |
US7110575B2 (en) | 2002-08-02 | 2006-09-19 | Eastman Kodak Company | Method for locating faces in digital color images |
US7117453B2 (en) * | 2003-01-21 | 2006-10-03 | Microsoft Corporation | Media frame object visualization system |
US7508961B2 (en) | 2003-03-12 | 2009-03-24 | Eastman Kodak Company | Method and system for face detection in digital images |
US8589373B2 (en) * | 2003-09-14 | 2013-11-19 | Yaron Mayer | System and method for improved searching on the internet or similar networks and especially improved MetaNews and/or improved automatically generated newspapers |
US7555165B2 (en) | 2003-11-13 | 2009-06-30 | Eastman Kodak Company | Method for semantic scene classification using camera metadata and content-based cues |
US7680340B2 (en) | 2003-11-13 | 2010-03-16 | Eastman Kodak Company | Method of using temporal context for image classification |
US20060106793A1 (en) * | 2003-12-29 | 2006-05-18 | Ping Liang | Internet and computer information retrieval and mining with intelligent conceptual filtering, visualization and automation |
US20060047649A1 (en) | 2003-12-29 | 2006-03-02 | Ping Liang | Internet and computer information retrieval and mining with intelligent conceptual filtering, visualization and automation |
US7958063B2 (en) * | 2004-11-11 | 2011-06-07 | Trustees Of Columbia University In The City Of New York | Methods and systems for identifying and localizing objects based on features of the objects that are mapped to a vector |
US20070027901A1 (en) * | 2005-08-01 | 2007-02-01 | John Chan | Method and System for Developing and Managing A Computer-Based Marketing Campaign |
US20070098303A1 (en) | 2005-10-31 | 2007-05-03 | Eastman Kodak Company | Determining a particular person from a collection |
US7711145B2 (en) | 2006-01-27 | 2010-05-04 | Eastman Kodak Company | Finding images with multiple people or objects |
US8315463B2 (en) | 2006-11-14 | 2012-11-20 | Eastman Kodak Company | User interface for face recognition |
US8122356B2 (en) * | 2007-10-03 | 2012-02-21 | Eastman Kodak Company | Method for image animation using image value rules |
US8611677B2 (en) | 2008-11-19 | 2013-12-17 | Intellectual Ventures Fund 83 Llc | Method for event-based semantic classification |
US8200548B2 (en) * | 2009-08-31 | 2012-06-12 | Peter Wiedl | Recipe engine system and method |
US8429559B2 (en) * | 2009-10-19 | 2013-04-23 | Xerox Corporation | Elicitation method for custom image preferences using keywords |
US9009163B2 (en) * | 2009-12-08 | 2015-04-14 | Intellectual Ventures Fund 83 Llc | Lazy evaluation of semantic indexing |
-
2011
- 2011-02-04 US US13/021,188 patent/US8612441B2/en active Active
-
2013
- 2013-11-14 US US14/079,787 patent/US9037569B2/en active Active
-
2015
- 2015-04-20 US US14/691,265 patent/US9524349B2/en active Active
Patent Citations (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5852823A (en) * | 1996-10-16 | 1998-12-22 | Microsoft | Image classification and retrieval system using a query-by-example paradigm |
US5899999A (en) * | 1996-10-16 | 1999-05-04 | Microsoft Corporation | Iterative convolution filter particularly suited for use in an image classification and retrieval system |
US7245762B2 (en) * | 1999-05-17 | 2007-07-17 | Samsung Electronics Co., Ltd. | Color image processing method |
US20030198380A1 (en) * | 1999-05-17 | 2003-10-23 | Samsung Electronics Co., Ltd.., University Of , Calif. At Santa Barbara | Color image processing method |
US6751343B1 (en) * | 1999-09-20 | 2004-06-15 | Ut-Battelle, Llc | Method for indexing and retrieving manufacturing-specific digital imagery based on image content |
US6993180B2 (en) * | 2001-09-04 | 2006-01-31 | Eastman Kodak Company | Method and system for automated grouping of images |
US7167574B2 (en) * | 2002-03-14 | 2007-01-23 | Seiko Epson Corporation | Method and apparatus for content-based image copy detection |
US20030174859A1 (en) * | 2002-03-14 | 2003-09-18 | Changick Kim | Method and apparatus for content-based image copy detection |
US7826661B2 (en) * | 2003-07-04 | 2010-11-02 | Mitsubishi Denki Kabushiki Kaisha | Method and apparatus for representing a group of images |
US20060147094A1 (en) * | 2003-09-08 | 2006-07-06 | Woong-Tuk Yoo | Pupil detection method and shape descriptor extraction method for a iris recognition, iris feature extraction apparatus and method, and iris recognition system and method using its |
US7379627B2 (en) * | 2003-10-20 | 2008-05-27 | Microsoft Corporation | Integrated solution to digital image similarity searching |
US8005831B2 (en) * | 2005-08-23 | 2011-08-23 | Ricoh Co., Ltd. | System and methods for creation and use of a mixed media environment with geographic location information |
US8027512B2 (en) * | 2005-09-30 | 2011-09-27 | Robert Bosch Gmbh | Method and software program for searching image information |
US20100027895A1 (en) * | 2006-08-31 | 2010-02-04 | Osaka Prefecture University Public Corporation | Image recognition method, image recognition device, and image recognition program |
US20080089561A1 (en) * | 2006-10-11 | 2008-04-17 | Tong Zhang | Face-based image clustering |
US20080208791A1 (en) * | 2007-02-27 | 2008-08-28 | Madirakshi Das | Retrieving images based on an example image |
US20080306995A1 (en) * | 2007-06-05 | 2008-12-11 | Newell Catherine D | Automatic story creation using semantic classifiers for images and associated meta data |
US20090161962A1 (en) * | 2007-12-20 | 2009-06-25 | Gallagher Andrew C | Grouping images by location |
US20090324026A1 (en) * | 2008-06-27 | 2009-12-31 | Palo Alto Research Center Incorporated | System and method for finding a picture image in an image collection using localized two-dimensional visual fingerprints |
Cited By (43)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120307094A1 (en) * | 2009-08-28 | 2012-12-06 | Nikon Corporation | Image file data structure, image file generation device, image file generation method, and electronic camera |
US20130051687A1 (en) * | 2011-08-25 | 2013-02-28 | Canon Kabushiki Kaisha | Image processing system and image processing method |
US10459968B2 (en) | 2011-08-25 | 2019-10-29 | Canon Kabushiki Kaisha | Image processing system and image processing method |
US20130144975A1 (en) * | 2011-12-05 | 2013-06-06 | Seiji Miyahara | Image information processing apparatus and image information processing method |
US9639778B2 (en) | 2012-03-26 | 2017-05-02 | Canon Kabushiki Kaisha | Information processing apparatus, control method thereof, and storage medium |
US9087237B2 (en) * | 2012-03-26 | 2015-07-21 | Canon Kabushiki Kaisha | Information processing apparatus, control method thereof, and storage medium |
US20130251198A1 (en) * | 2012-03-26 | 2013-09-26 | Canon Kabushiki Kaisha | Information processing apparatus, control method thereof, and storage medium |
US10268758B2 (en) * | 2013-09-29 | 2019-04-23 | Peking University Founder Group Co. Ltd. | Method and system of acquiring semantic information, keyword expansion and keyword search thereof |
US20150186425A1 (en) * | 2013-12-30 | 2015-07-02 | Htc Corporation | Method for searching relevant images via active learning, electronic device using the same |
CN104751168A (en) * | 2013-12-30 | 2015-07-01 | 宏达国际电子股份有限公司 | Method for searching relevant images via active learning, electronic device using the same |
US10169702B2 (en) * | 2013-12-30 | 2019-01-01 | Htc Corporation | Method for searching relevant images via active learning, electronic device using the same |
US10002375B1 (en) * | 2014-06-10 | 2018-06-19 | Amazon Technologies, Inc. | Hashtag shopping and rating |
US11487405B2 (en) | 2014-11-06 | 2022-11-01 | Dropbox, Inc. | Searching digital content |
US10725614B2 (en) * | 2014-11-06 | 2020-07-28 | Dropbox, Inc. | Searching digital content |
US20160132194A1 (en) * | 2014-11-06 | 2016-05-12 | Dropbox, Inc. | Searching digital content |
US12039148B2 (en) | 2014-11-06 | 2024-07-16 | Dropbox, Inc. | Searching digital content |
US10955998B2 (en) | 2014-11-06 | 2021-03-23 | Dropbox, Inc. | Searching digital content |
US20160283557A1 (en) * | 2015-03-27 | 2016-09-29 | Orange | Method and device for accessing a plurality of contents, corresponding terminal and computer program |
US10467231B2 (en) * | 2015-03-27 | 2019-11-05 | Orange | Method and device for accessing a plurality of contents, corresponding terminal and computer program |
US11475144B2 (en) | 2015-04-17 | 2022-10-18 | Dropbox, Inc. | Collection folder for collecting file submissions |
US11157636B2 (en) | 2015-04-17 | 2021-10-26 | Dropbox, Inc. | Collection folder for collecting file submissions in response to a public file request |
US12086276B2 (en) | 2015-04-17 | 2024-09-10 | Dropbox, Inc. | Collection folder for collecting file submissions in response to a public file request |
US12079353B2 (en) | 2015-04-17 | 2024-09-03 | Dropbox, Inc. | Collection folder for collecting file submissions |
US11783059B2 (en) | 2015-04-17 | 2023-10-10 | Dropbox, Inc. | Collection folder for collecting file submissions |
US11270008B2 (en) * | 2015-04-17 | 2022-03-08 | Dropbox, Inc. | Collection folder for collecting file submissions |
US10929547B2 (en) | 2015-04-17 | 2021-02-23 | Dropbox, Inc. | Collection folder for collecting file submissions using email |
US11630905B2 (en) | 2015-04-17 | 2023-04-18 | Dropbox, Inc. | Collection folder for collecting file submissions in response to a public file request |
US11244062B2 (en) | 2015-04-17 | 2022-02-08 | Dropbox, Inc. | Collection folder for collecting file submissions |
US9576203B2 (en) * | 2015-04-29 | 2017-02-21 | Canon Kabushiki Kaisha | Devices, systems, and methods for knowledge-based inference for material recognition |
US11403352B2 (en) * | 2015-09-28 | 2022-08-02 | Yahoo Assets Llc | Multi-touch gesture search |
US10162865B2 (en) * | 2015-10-08 | 2018-12-25 | Microsoft Technology Licensing, Llc | Generating image tags |
US11948473B2 (en) | 2015-12-31 | 2024-04-02 | Dropbox, Inc. | Assignments for classrooms |
CN107766394A (en) * | 2016-08-23 | 2018-03-06 | 阿里巴巴集团控股有限公司 | Business data processing method and its system |
WO2018040059A1 (en) * | 2016-09-02 | 2018-03-08 | Microsoft Technology Licensing, Llc | Clip content categorization |
CN106649610A (en) * | 2016-11-29 | 2017-05-10 | 北京智能管家科技有限公司 | Image labeling method and apparatus |
US20180349467A1 (en) * | 2017-06-02 | 2018-12-06 | Apple Inc. | Systems and methods for grouping search results into dynamic categories based on query and result set |
US11669550B2 (en) | 2017-06-02 | 2023-06-06 | Apple Inc. | Systems and methods for grouping search results into dynamic categories based on query and result set |
CN109088979A (en) * | 2018-06-28 | 2018-12-25 | 珠海格力电器股份有限公司 | Collection control method, device, storage medium and terminal |
CN112534421A (en) * | 2018-12-20 | 2021-03-19 | 谷歌有限责任公司 | Map search recommendation system based on image content analysis-driven geographic semantic index |
CN113939813A (en) * | 2019-05-21 | 2022-01-14 | 微软技术许可有限责任公司 | Generating and applying object-level relational indexes for images |
US11954145B2 (en) | 2021-06-22 | 2024-04-09 | Huawei Technologies Co., Ltd. | Methods, systems, and media for image searching |
WO2022268094A1 (en) * | 2021-06-22 | 2022-12-29 | Huawei Technologies Co., Ltd. | Methods, systems, and media for image searching |
EP4145305A1 (en) * | 2021-09-06 | 2023-03-08 | Bull SAS | Multifacet and modular search method within a collection of video sequences |
Also Published As
Publication number | Publication date |
---|---|
US9037569B2 (en) | 2015-05-19 |
US8612441B2 (en) | 2013-12-17 |
US20140074825A1 (en) | 2014-03-13 |
US20150227634A1 (en) | 2015-08-13 |
US9524349B2 (en) | 2016-12-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9524349B2 (en) | Identifying particular images from a collection | |
US20220004573A1 (en) | Method for creating view-based representations from multimedia collections | |
US8611677B2 (en) | Method for event-based semantic classification | |
US9009163B2 (en) | Lazy evaluation of semantic indexing | |
RU2444072C2 (en) | System and method for using content features and metadata of digital images to find related audio accompaniment | |
US7076503B2 (en) | Managing media objects in a database | |
Larson et al. | Automatic tagging and geotagging in video collections and communities | |
US8934717B2 (en) | Automatic story creation using semantic classifiers for digital assets and associated metadata | |
US20080306995A1 (en) | Automatic story creation using semantic classifiers for images and associated meta data | |
US20060253491A1 (en) | System and method for enabling search and retrieval from image files based on recognized information | |
Iyengar et al. | Discriminative model fusion for semantic concept detection and annotation in video | |
US20100226582A1 (en) | Assigning labels to images in a collection | |
Smeaton et al. | The TREC-2002 video track report | |
Mills et al. | Shoebox: A digital photo management system | |
Cao et al. | Image annotation within the context of personal photo collections using hierarchical event and scene models | |
Sandhaus et al. | Semantic analysis and retrieval in personal and social photo collections | |
Kuo et al. | Building personal digital photograph libraries: An approach with ontology-based MPEG-7 dozen dimensional digital content architecture | |
Kuo et al. | MPEG-7 based dozen dimensional digital content architecture for semantic image retrieval services | |
Liu et al. | Semantic extraction and semantics-based annotation and retrieval for video databases | |
Gu et al. | Personal photo organization using event annotation | |
Magalhaes et al. | Exploring multimedia in a keyword space | |
Kim et al. | User‐Friendly Personal Photo Browsing for Mobile Devices | |
Mulhem et al. | Advances in digital home photo albums | |
US20070094252A1 (en) | ImageRank | |
Wood et al. | Searching consumer image collections using web-based concept expansion |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: EASTMAN KODAK COMPANY, NEW YORK Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WOOD, MARK D.;LOUI, ALEXANDER C.;SIGNING DATES FROM 20110202 TO 20110203;REEL/FRAME:025746/0488 |
|
AS | Assignment |
Owner name: CITICORP NORTH AMERICA, INC., AS AGENT, NEW YORK Free format text: SECURITY INTEREST;ASSIGNORS:EASTMAN KODAK COMPANY;PAKON, INC.;REEL/FRAME:028201/0420 Effective date: 20120215 |
|
AS | Assignment |
Owner name: WILMINGTON TRUST, NATIONAL ASSOCIATION, AS AGENT, MINNESOTA Free format text: PATENT SECURITY AGREEMENT;ASSIGNORS:EASTMAN KODAK COMPANY;PAKON, INC.;REEL/FRAME:030122/0235 Effective date: 20130322 Owner name: WILMINGTON TRUST, NATIONAL ASSOCIATION, AS AGENT, Free format text: PATENT SECURITY AGREEMENT;ASSIGNORS:EASTMAN KODAK COMPANY;PAKON, INC.;REEL/FRAME:030122/0235 Effective date: 20130322 |
|
AS | Assignment |
Owner name: PAKON, INC., NEW YORK Free format text: RELEASE OF SECURITY INTEREST IN PATENTS;ASSIGNORS:CITICORP NORTH AMERICA, INC., AS SENIOR DIP AGENT;WILMINGTON TRUST, NATIONAL ASSOCIATION, AS JUNIOR DIP AGENT;REEL/FRAME:031157/0451 Effective date: 20130903 Owner name: EASTMAN KODAK COMPANY, NEW YORK Free format text: RELEASE OF SECURITY INTEREST IN PATENTS;ASSIGNORS:CITICORP NORTH AMERICA, INC., AS SENIOR DIP AGENT;WILMINGTON TRUST, NATIONAL ASSOCIATION, AS JUNIOR DIP AGENT;REEL/FRAME:031157/0451 Effective date: 20130903 |
|
AS | Assignment |
Owner name: 111616 OPCO (DELAWARE) INC., NEW YORK Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:EASTMAN KODAK COMPANY;REEL/FRAME:031172/0025 Effective date: 20130903 |
|
AS | Assignment |
Owner name: KODAK ALARIS INC., NEW YORK Free format text: CHANGE OF NAME;ASSIGNOR:111616 OPCO (DELAWARE) INC.;REEL/FRAME:031394/0001 Effective date: 20130920 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
AS | Assignment |
Owner name: KPP (NO. 2) TRUSTEES LIMITED, NORTHERN IRELAND Free format text: SECURITY INTEREST;ASSIGNOR:KODAK ALARIS INC.;REEL/FRAME:053993/0454 Effective date: 20200930 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |
|
AS | Assignment |
Owner name: THE BOARD OF THE PENSION PROTECTION FUND, UNITED KINGDOM Free format text: ASSIGNMENT OF SECURITY INTEREST;ASSIGNOR:KPP (NO. 2) TRUSTEES LIMITED;REEL/FRAME:058175/0651 Effective date: 20211031 |
|
AS | Assignment |
Owner name: FGI WORLDWIDE LLC, NEW YORK Free format text: SECURITY AGREEMENT;ASSIGNOR:KODAK ALARIS INC.;REEL/FRAME:068325/0938 Effective date: 20240801 |
|
AS | Assignment |
Owner name: KODAK ALARIS INC., NEW YORK Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:THE BOARD OF THE PENSION PROTECTION FUND;REEL/FRAME:068481/0300 Effective date: 20240801 |