US20210365490A1 - Method for ranking and selecting events in media collections - Google Patents
Method for ranking and selecting events in media collections Download PDFInfo
- Publication number
- US20210365490A1 US20210365490A1 US17/391,957 US202117391957A US2021365490A1 US 20210365490 A1 US20210365490 A1 US 20210365490A1 US 202117391957 A US202117391957 A US 202117391957A US 2021365490 A1 US2021365490 A1 US 2021365490A1
- Authority
- US
- United States
- Prior art keywords
- events
- event
- ranking
- processor
- images
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 55
- 238000009826 distribution Methods 0.000 claims description 42
- 238000004458 analytical method Methods 0.000 claims description 4
- 230000002123 temporal effect Effects 0.000 description 5
- YHXISWVBGDMDLQ-UHFFFAOYSA-N moclobemide Chemical compound C1=CC(Cl)=CC=C1C(=O)NCCN1CCOCC1 YHXISWVBGDMDLQ-UHFFFAOYSA-N 0.000 description 4
- 230000008569 process Effects 0.000 description 4
- 230000003247 decreasing effect Effects 0.000 description 3
- 230000003997 social interaction Effects 0.000 description 3
- 238000001514 detection method Methods 0.000 description 2
- 230000001404 mediated effect Effects 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 230000008447 perception Effects 0.000 description 2
- 239000000654 additive Substances 0.000 description 1
- 230000000996 additive effect Effects 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 238000007418 data mining Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000004880 explosion Methods 0.000 description 1
- 238000009472 formulation Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- DETAHNVSLBCZAA-ARJGXJLFSA-N photo product Chemical compound C[C@@H]([C@]12O)[C@@H](OC(C)=O)[C@@]3(OC(C)=O)C(C)(C)C3[C@@H]2C2[C@]3(COC(C)=O)C[C@]4(O)[C@H]1C2[C@@]3(C)C4=O DETAHNVSLBCZAA-ARJGXJLFSA-N 0.000 description 1
- 230000002062 proliferating effect Effects 0.000 description 1
- 230000035755 proliferation Effects 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 238000012731 temporal analysis Methods 0.000 description 1
- 210000003813 thumb Anatomy 0.000 description 1
- 238000000700 time series analysis Methods 0.000 description 1
- 239000002023 wood Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/41—Indexing; Data structures therefor; Storage structures
Definitions
- the invention relates generally to the field of digital image processing, and in particular to methods and systems for ranking and selecting events in consumer media collections.
- the present invention also teaches how to select events from the ranked list of events based on a calculated target distribution, which can be computed using the distribution of one or more event attributes of the events derived from the media collection.
- event ranking and selection can also be tied to social networks, where different user input such as tags and comments, will be used for aid in the ranking and selection.
- the organization and retrieval of images and videos is a problem for the typical consumer. It is useful for the user to be able to browse an overview of important events in their collection.
- Technology disclosed in prior art allows the classification of images in a collection into events, but not the ability to ascertain the importance or ranking of such events. As a result, these include uninteresting or common day-to-day events that inflate the number of events to the point where it is difficult to find more important events even when browsing a list of events.
- This invention teaches a method and system for automatically ranking events that have been detected from a media collection. In addition, it also discloses how to select events from a ranked list of events based on a calculated target distribution, which can be computed using the distribution of one or more event attributes of the events derived from the media collection.
- a method and system for ranking events in media collections comprising designating a media collection, using a processor to cluster the media collection items into a hierarchical event structure, using the processor to identify and count visually similar sub-events within each event in the hierarchical event structure, using the processor to determine a ranking of events based on the count of sub-events within each event, and associating the determined ranking with each event in the media collection.
- a method for selecting events from media collections comprising designating a media collection, using a processor to cluster the media collection items into a hierarchical event structure, using the processor to identify and count visually similar sub-events within each event in the hierarchical event structure, using the processor to determine a ranked list of events based on the count of sub-events within each event, using the processor to calculate a target distribution that is based on the distribution of one or more event attributes of the events derived from the media collection, and selecting events from the ranked list of events based on the calculated target distribution.
- FIG. 1 is a block diagram of a system that practices the present invention
- FIG. 2 is an overall flowchart of the method practiced by system shown in FIG. 1 ;
- FIG. 3 shows the method for selecting events from a ranked list of events according to an aspect of the present invention
- FIG. 4 shows a model for interestingness of an image according to an aspect of the present invention
- FIG. 5 shows additional detail for Step 330 of FIG. 3 ;
- FIG. 6 shows additional detail for Step 535 of FIG. 5 .
- the present invention can be implemented in computer systems as will be well known to those skilled in the art.
- some embodiments of the present invention will be described as software programs. Those skilled in the art will readily recognize that the equivalent of such a method may also be constructed as hardware or software within the scope of the invention.
- the present invention can be implemented in computer hardware and computerized equipment.
- the method can be performed in a digital camera, a multimedia smart phone, a digital printer, on an internet server, on a kiosk, and on a personal computer.
- FIG. 1 there is illustrated a computer system for implementing the present invention.
- the computer system is shown for the purpose of illustrating a preferred embodiment, the present invention is not limited to the computer system shown, but can be used on any electronic processing system such as found in digital cameras, home computers, kiosks, or any other system for the processing of digital images.
- the computer 10 includes a microprocessor-based unit 20 (also referred to herein as a processor) for receiving and processing software programs and for performing other processing functions.
- a microprocessor-based unit 20 also referred to herein as a processor
- a memory unit 30 stores user-supplied and computer-generated data which may be accessed by the processor 20 when running a computer program.
- a display device (such as a monitor) 70 is electrically connected to the computer 10 for displaying information and data associated with the software, e.g., by means of a graphical user interface.
- a keyboard 60 is also connected to the computer. As an alternative to using the keyboard 60 for input, a mouse can be used for moving a selector on the display device 70 and for selecting an item on which the selector overlays, as is well known in the art.
- Input devices 50 such as compact disks (CD) and DVDs can be inserted into the computer 10 for inputting the software programs and other information to the computer 10 and the processor 20 .
- the computer 10 can be programmed, as is well known in the art, for storing the software program internally.
- media files (such as images, music and videos) can be transferred to the memory 30 of the computer 10 by means of input devices 50 such as memory cards, thumb drives, CDs and DVDs, or by connecting a capture device (such as camera, cell phone, video recorder) directly to the computer 10 as an input device.
- the computer 10 can have a network connection, such as a telephone line or wireless connection 80 , to an external network, such as a local area network or the Internet.
- Software programs and media files can be transferred to the computer from other computers or the Internet through the network connection.
- the present invention can be implemented in a combination of software or hardware and is not limited to devices which are physically connected or located within the same physical location.
- One or more of the devices illustrated in FIG. 1 can be located remotely and can be connected via a network.
- One or more of the devices can be connected wirelessly, such as by a radio-frequency link, either directly or via a network.
- FIG. 2 a user's digital image collection 105 resides in the memory 30 of a computer 10 .
- the other blocks in the figure are implemented by a software program and are executed by the processor 20 of the computer 10 .
- FIG. 2 shows the overall workflow of an aspect of the present invention and each component module will be described in detail below.
- a user's digital image collection 105 is grouped into an event representation by the event hierarchy generator 110 .
- Commonly assigned U.S. Pat. Nos. 6,606,411 and 6,351,556 disclose algorithms for clustering image content by temporal events and sub-events, the disclosures of which are incorporated herein. According to U.S. Pat. No. 6,606,411 events have consistent color distributions, and therefore, these pictures are likely to have been taken with the same backdrop. For each sub-event, a single color and texture representation is computed for all background areas taken together. The above two patents teach how to cluster images and videos in a digital image collection into temporal events and sub-events.
- a collection of images is classified into one or more events determining one or more largest time differences of the collection of images based on time and/or date clustering of the images and separating the plurality of images into the events based on having one or more boundaries between events where one or more boundaries correspond to the one or more largest time differences.
- sub-events can be determined (if any) by comparing the color histogram information of successive images as described in U.S. Pat. No.
- the events detected continue to be chronologically ordered in a timeline from earliest to latest.
- Using the method described above it is not possible to detect single events that span a long period of time (days) and encompass a variety of activities and settings (for example, a long vacation covering multiple destinations) or events that occur in distinct parts separated by some hours from each other (for example, a sporting event with many matches or a wedding). Gaps in photo-taking corresponding to the overnight period also cause breaks in event continuity. Further processing is needed to detect these super-events, defined as a grouping of multiple contiguous events that may span multiple days. Inter-event duration, defined as the time duration between the last image of one event and the first image of the next event on a continuous timeline, is computed for each event.
- the events are then treated as single points on a time axis, separated by the inter-event durations.
- a density-based clustering method is applied to these points (ref. Data Mining Concepts and Techniques by Han and Kamber, Elsevier, 2006, supra, pp. 418-420) to cluster events into super-events when they are separated by relatively small duration gaps (for example, less than 18 hours).
- the final three-level hierarchical event representation includes super-events, events and sub-events.
- the term “event” refers to the top-level of the hierarchical event representation—which can be a super-event or an event. Referring to FIG. 2 , the digital image collection 105 is grouped into temporal events, sub-events and super-events using the methods described above.
- significant events are detected in step 115 from the digital image collection 105 .
- a significant event detection algorithm using time-series analysis of the capture date/time information of the images is used to detect the significant events.
- the details of the algorithm can be found in U.S. Pat. No. 8,340,436, “Detecting significant events in consumer image collections,” by Das and Loui, issued on Dec. 25, 2012, the disclosure of which is incorporated herein by reference.
- the predicted output of the selected ARIMA model is compared with the image counts time-series that was used to generate the model. Residuals are computed as the difference between the predicted output of the model and the image counts time-series at each time step.
- the variance ( ⁇ ) of the residuals is computed and a threshold is determined based on this variance.
- the output of the event hierarchy generator 110 and the significant event detector 115 are fed into the event ranking module 120 .
- the events can be ranked by a number of different criteria.
- the number of sub-events in the event is used to rank events in descending order of importance. Since each sub-event extracted using the method disclosed in U.S. Pat. No. 6,606,411 has consistent color distribution as determined by block-level color histogram similarity; more sub-events in an event indicates that these pictures are likely to have been taken with diverse backgrounds that increase the scope of the event. This justifies a higher ranking when there are more sub-events.
- the significance score defined as the residual divided by the variance (a), is used to rank the events, with a higher score getting a higher rank. The significance score generated at the end of the significant event detection described earlier indicates how well the event fits into the estimated model, with a higher score indicating a lower fit, and therefore, the event is more likely to be something unusual and important in the collection.
- the interestingness of an event can be modeled as shown in FIG. 4 .
- the interestingness score of an event is initially high (close to 1.0) at the time of capture (start time of the event), and then falls rapidly as the first month passes by.
- the interestingness score again rises around the picture's one-year anniversary mark (because people are often interested in reliving the previous year's happenings, especially if the event is an annual recurring event such as a birthday).
- the interestingness score then plateaus to a higher level than the first year, and at each subsequent anniversary achieves a slightly higher level than the previous year.
- the events are ranked according to their interestingness score.
- the albums of images a user uploads for sharing to social networks are gathered, along with social interactions such as “likes”, comments, and tags associated with each image.
- the images in these albums are treated as a set of images that have no capture date-time information, but are in a list arranged by the time of upload.
- This list of images is merged into a user's personal image collection that resides on their private storage (which can be on a personal computer, mobile device or online storage) using the method described in U.S. Pat. No. 7,831,599 “Additive clustering of images lacking individualized date-time information” by Das et al issued Sep. 11, 2010.
- This patent describes a method that uses a dynamic programming-based formulation to merge images lacking capture time into an organized collection where events have already been computed and capture date-time information exists.
- the method computes image similarity scores based on image content, and ensures that the ordering of the incoming list of images is maintained.
- the number of social interactions (“likes”, comments and tags) derived from the shared images are counted for each event in the user collection that contains shared images from the merging process.
- the events are ranked in decreasing order of number of social interactions.
- the number of images that are marked by the user is counted for each event, and the events are ranked in decreasing order of the number of user markings
- the user markings can take different forms including being marked a “favorite” either at time of capture on the capture device itself, or later on the site of storage (computer or online storage); marked as to be used for sharing; or marked with a star rating system provided by the capture device or storage site with the maximum star rating allowed.
- steps 310 - 330 refer to additional steps performed with the ranked list of events (step 150 of FIG. 2 ) to select events from the ranked list.
- One or more event attributes 310 are computed for each of the events on the ranked list.
- the event attributes that can be computed include event class, event size, and media type of the event.
- Event class refers to the classification of the event into one of a finite set of event classes e.g., vacation, party, sports and family moments.
- a method for event classification is described in US Patent application US 2010/0124378A1 titled “Method for event-based semantic classification”. In this method, a classifier is trained to recognize events of different pre-specified classes.
- Event size refers to the number of assets (images or video) contained in the top-level event (i.e., a super-event or an event).
- the media type of an event refers to the ratio of videos to images in the event, discretized into a pre-specified number of bins. The media type indicates the mix of video and images in an event.
- step 320 determines a target distribution of the selected event attribute.
- the target distribution is initially computed to match the distribution of the attribute in the collection.
- a histogram is generated where each bin represents a category of the selected event attribute, and the value of the bin is the count of events in the collection with that category.
- the histogram is then normalized (each bin count is divided by the total number of events in the collection) so that the values are between 0.0 and 1.0.
- This normalized histogram represents the target distribution 320 .
- input from the user 315 can be incorporated to alter the target distribution at this point.
- the target distribution of the event class attribute is altered so that the “vacation” class is set to 1.0 and the rest of the classes are set to 0.0.
- the user input does not need to be binary—an interface could be provided that allows the user to indicate interest in a sliding scale e.g., more vacations, less sports, and these can be translated into corresponding changes in the target distribution.
- step 330 selects events from the ranked list in descending order while maintaining the target distribution.
- the output product type e.g., photobook, calendar or collage, determines the number (typically not an exact number, but a range) of images needed to create the output product type.
- the user or system in case of an auto-generated product
- the fraction is chosen to be 0.1 (i.e., 10% of the images from an event are typically expected to be used in the product).
- This output product-based requirement for the number of images, fraction of images from an event, and selected sub-set is provided in step 325 as input parameters to the selection step 330 .
- the number of images needed (or the number in the center of the range, if a range is provided) is used to proportionally allot number of images according to the target distribution 525 .
- the ranked list is filtered 530 by the selected sub-set e.g. if a time range is specified, only events in that time range are retained in the list, and the others are eliminated.
- the events are selected in step 535 that is shown in detail in FIG. 6 .
- each event encountered in the list is selected if the number of remaining images allotted to that event type is greater than zero.
- the number of allotted images for that event attribute is decreased by the product of the given fraction and the event size.
- the process is continued till there is no positive number of remaining images for any event attribute or when the list is exhausted (without meeting all the allotments).
- a second pass is performed through the remaining events in the list, after re-allotting the event types that did not have enough candidates, to other event types in proportion to the target distribution values for those event types. This process is continued till either the remaining allotments are all negative, or when there are no events left in the list.
- a method for ranking events in media collections comprises designating a media collection, using a processor to cluster the media collection items into a hierarchical event structure, using the processor to identify and count visually similar sub-events within each event in the hierarchical event structure, using the processor to determine a ranking of events based on the count of sub-events within each event, and associating the determined ranking with each event in the media collection.
- the ranking of events can be based on the significance score of the event, on a distribution that models the importance of an event over an elapsed time period, on a score or distribution that models the interestingness of an event over an elapsed time period, on metadata from social networks such as number of likes and comments, on metadata from social networks through the analysis of user tags and comments, or on the number of images in the event that have been marked by the user as being a favorite or to be used for sharing.
- a method for selecting events from media collections comprises designating a media collection, using a processor to cluster the media collection items into a hierarchical event structure, using the processor to identify and count visually similar sub-events within each event in the hierarchical event structure, using the processor to determine a ranked list of events based on the count of sub-events within each event, using the processor to calculate a target distribution that is based on the distribution of one or more event attributes of the events derived from the media collection, and selecting events from the ranked list of events based on the calculated target distribution.
- the event attribute used in the target distribution can be the event class, the event size, or the media type of the event.
- the ranking of events is based on the significance score of the event, on a distribution that models the importance of an event over an elapsed time period, on scores or a distribution that models the interestingness of an event over an elapsed time period, on metadata from social networks such as number of likes and comments, on metadata from social networks through the analysis of user tags and comments, or on the number of images in the event that have been marked by the user as being a favorite or to be used for sharing.
- a system for ranking events in media collections comprises a processor-accessible memory for storing a media collection, and a processor for clustering the media collection items into a hierarchical event structure, for identifying and visually counting similar sub-events within each event in the hierarchical event structure, for determining a ranking of events based on the count of sub-events within each event, and for associating the determined ranking with each event in the media collection.
- a system for selecting events from media collections comprises a processor-accessible memory for storing a media collection and a processor for clustering the media collection items into a hierarchical event structure, for identifying and visually counting similar sub-events within each event in the hierarchical event structure, for determining a ranked list of events based on the count of sub-events within each event, for calculating a target distribution based on the distribution of one or more event attributes of the events derived from the media collection, and for selecting events from the ranked list of events based on the calculated target distribution.
Abstract
Description
- The present application is a continuation of U.S. application Ser. No. 14/310,777, filed on Jun. 20, 2014, which claims priority to U.S. provisional application No. 61/840,031, filed on Jun. 27, 2013. The aforementioned applications are hereby incorporated by reference in their entirety.
- The invention relates generally to the field of digital image processing, and in particular to methods and systems for ranking and selecting events in consumer media collections.
- The proliferation of digital cameras and scanners has lead to an explosion of digital images, creating large personal image databases. Since taking digital pictures is easy and practically free, consumers no longer restrict picture-taking to important events and special occasions. Images are being captured frequently, and of day-to-day occurrences in the consumers' life. Since a typical user has already accumulated many years of digital images, browsing the collection to find images taken during important events is a time-consuming process for the consumer.
- There has been work in grouping images into events. U.S. Pat. No. 6,606,411, assigned to A. Loui and E. Pavie, entitled “A method for automatically classifying images into events,” issued Aug. 12, 2003 and U.S. Pat. No. 6,351,556, assigned to A. Loui, and E. Pavie, entitled “A method for automatically comparing content of images for classification into events,” issued Feb. 26, 2002, disclose algorithms for clustering image content by temporal events and sub-events. According to U.S. Pat. No. 6,606,411 events have consistent color distributions, and therefore, these pictures are likely to have been taken with the same backdrop. For each sub-event, a single color and texture representation is computed for all background areas taken together. The above two patents teach how to cluster images and videos in a digital image collection into temporal events and sub-events. The terms “event” and “sub-event” are used in an objective sense to indicate the products of a computer mediated procedure that attempts to match a user's subjective perceptions of specific occurrences (corresponding to events) and divisions of those occurrences (corresponding to sub-events). Another method of automatically organizing images into events is disclosed in U.S. Pat. No. 6,915,011, assigned to A. Loui, M. Jeanson, and Z. Sun, entitled “Event clustering of images using foreground and background segmentation” issued Jul. 5, 2005. The events detected are chronologically ordered in a timeline from earliest to latest.
- Using the above methods, it is possible to reduce the amount of browsing required by the user to locate a particular event by viewing representatives of the events along a timeline, instead of each image thumbnail. However, a typical user may still generate hundreds of such events over a few year period, and more prolific picture-takers can easily exceed a few thousands detected events. It will be a very tedious task for the user to browse through their collection to pick various events or sub-events to create a photo product such as a collage or photobook. Hence, there is a need for new methods and systems to automatically rank the events and to select the preferred set of events based on some relevant criteria. In addition, the present invention also teaches how to select events from the ranked list of events based on a calculated target distribution, which can be computed using the distribution of one or more event attributes of the events derived from the media collection. Further, event ranking and selection can also be tied to social networks, where different user input such as tags and comments, will be used for aid in the ranking and selection.
- There has been other work in event clustering using metadata. U.S. Pat. No. 7,860,866, assigned to Kim el at., entitled “Heuristic event clustering of media using metadata,” issued Dec. 28, 2010, discloses algorithms for clustering an media collection into event based on time difference and location difference between consecutive media files. However the above patent does not teach how to rank or select event from a media collection, which is the main idea in the present invention. The '866 patent only teaches how to cluster media files into separate events with no ranking information. There also has been work in identifying media assets using contextual information. U.S. Pat. No. 8,024,311, assigned to Wood and Hibino, entitled “Identifying media assets from contextual information,” issued on Sep. 20, 2011, discloses a method to select media assets by identifying an event using the received contextual information such as text data, gesture data, or audio data. The above patent clearly depends on a user to first provide some contextual information as input before it can identify the appropriate event, and the subsequent selection of the media assets. This is a different application as it requires user input and direction, whereas the present invention teaches how to automatically rank and select events without user input. Further, the '311 patent only identify one event (see
FIG. 2 ) based on the input contextual information, whereas the present invention will provide a rank for each of the events in the collection. - The organization and retrieval of images and videos is a problem for the typical consumer. It is useful for the user to be able to browse an overview of important events in their collection. Technology disclosed in prior art allows the classification of images in a collection into events, but not the ability to ascertain the importance or ranking of such events. As a result, these include uninteresting or common day-to-day events that inflate the number of events to the point where it is difficult to find more important events even when browsing a list of events. This invention teaches a method and system for automatically ranking events that have been detected from a media collection. In addition, it also discloses how to select events from a ranked list of events based on a calculated target distribution, which can be computed using the distribution of one or more event attributes of the events derived from the media collection.
- In accordance with the present invention, there is provided a method and system for ranking events in media collections comprising designating a media collection, using a processor to cluster the media collection items into a hierarchical event structure, using the processor to identify and count visually similar sub-events within each event in the hierarchical event structure, using the processor to determine a ranking of events based on the count of sub-events within each event, and associating the determined ranking with each event in the media collection.
- In another embodiment of the present invent, there is provide a method for selecting events from media collections comprising designating a media collection, using a processor to cluster the media collection items into a hierarchical event structure, using the processor to identify and count visually similar sub-events within each event in the hierarchical event structure, using the processor to determine a ranked list of events based on the count of sub-events within each event, using the processor to calculate a target distribution that is based on the distribution of one or more event attributes of the events derived from the media collection, and selecting events from the ranked list of events based on the calculated target distribution.
-
FIG. 1 is a block diagram of a system that practices the present invention; -
FIG. 2 is an overall flowchart of the method practiced by system shown inFIG. 1 ; -
FIG. 3 shows the method for selecting events from a ranked list of events according to an aspect of the present invention; -
FIG. 4 shows a model for interestingness of an image according to an aspect of the present invention; -
FIG. 5 shows additional detail for Step 330 ofFIG. 3 ; and -
FIG. 6 shows additional detail for Step 535 ofFIG. 5 . - The present invention can be implemented in computer systems as will be well known to those skilled in the art. In the following description, some embodiments of the present invention will be described as software programs. Those skilled in the art will readily recognize that the equivalent of such a method may also be constructed as hardware or software within the scope of the invention.
- Because image manipulation algorithms and systems are well known, the present description will be directed in particular to algorithms and systems forming part of, or cooperating more directly with, the method in accordance with the present invention. Other aspects of such algorithms and systems, and hardware or software for producing and otherwise processing the image signals involved therewith, not specifically shown or described herein can be selected from such systems, algorithms, components, and elements known in the art. Given the description as set forth in the following specification, all software implementation thereof is conventional and within the ordinary skill in such arts. Videos in a collection are included in the term “images” in the rest of the description.
- The present invention can be implemented in computer hardware and computerized equipment. For example, the method can be performed in a digital camera, a multimedia smart phone, a digital printer, on an internet server, on a kiosk, and on a personal computer. Referring to
FIG. 1 , there is illustrated a computer system for implementing the present invention. Although the computer system is shown for the purpose of illustrating a preferred embodiment, the present invention is not limited to the computer system shown, but can be used on any electronic processing system such as found in digital cameras, home computers, kiosks, or any other system for the processing of digital images. The computer 10 includes a microprocessor-based unit 20 (also referred to herein as a processor) for receiving and processing software programs and for performing other processing functions. Amemory unit 30 stores user-supplied and computer-generated data which may be accessed by theprocessor 20 when running a computer program. A display device (such as a monitor) 70 is electrically connected to the computer 10 for displaying information and data associated with the software, e.g., by means of a graphical user interface. A keyboard 60 is also connected to the computer. As an alternative to using the keyboard 60 for input, a mouse can be used for moving a selector on the display device 70 and for selecting an item on which the selector overlays, as is well known in the art. Input devices 50 such as compact disks (CD) and DVDs can be inserted into the computer 10 for inputting the software programs and other information to the computer 10 and theprocessor 20. Still further, the computer 10 can be programmed, as is well known in the art, for storing the software program internally. In addition, media files (such as images, music and videos) can be transferred to thememory 30 of the computer 10 by means of input devices 50 such as memory cards, thumb drives, CDs and DVDs, or by connecting a capture device (such as camera, cell phone, video recorder) directly to the computer 10 as an input device. The computer 10 can have a network connection, such as a telephone line or wireless connection 80, to an external network, such as a local area network or the Internet. Software programs and media files can be transferred to the computer from other computers or the Internet through the network connection. - It should also be noted that the present invention can be implemented in a combination of software or hardware and is not limited to devices which are physically connected or located within the same physical location. One or more of the devices illustrated in
FIG. 1 can be located remotely and can be connected via a network. One or more of the devices can be connected wirelessly, such as by a radio-frequency link, either directly or via a network. - Referring to
FIG. 2 , a user'sdigital image collection 105 resides in thememory 30 of a computer 10. The other blocks in the figure are implemented by a software program and are executed by theprocessor 20 of the computer 10.FIG. 2 shows the overall workflow of an aspect of the present invention and each component module will be described in detail below. - Referring to
FIG. 2 , a user'sdigital image collection 105 is grouped into an event representation by theevent hierarchy generator 110. Commonly assigned U.S. Pat. Nos. 6,606,411 and 6,351,556 disclose algorithms for clustering image content by temporal events and sub-events, the disclosures of which are incorporated herein. According to U.S. Pat. No. 6,606,411 events have consistent color distributions, and therefore, these pictures are likely to have been taken with the same backdrop. For each sub-event, a single color and texture representation is computed for all background areas taken together. The above two patents teach how to cluster images and videos in a digital image collection into temporal events and sub-events. The terms “event” and “sub-event” are used in an objective sense to indicate the products of a computer mediated procedure that attempts to match a user's subjective perceptions of specific occurrences (corresponding to events) and divisions of those occurrences (corresponding to sub-events). Briefly summarized, a collection of images is classified into one or more events determining one or more largest time differences of the collection of images based on time and/or date clustering of the images and separating the plurality of images into the events based on having one or more boundaries between events where one or more boundaries correspond to the one or more largest time differences. For each event, sub-events can be determined (if any) by comparing the color histogram information of successive images as described in U.S. Pat. No. 6,351,556. This is accomplished by dividing an image into a number of blocks and then computing the color histogram for each of the blocks. A block-based histogram correlation procedure is used as described in U.S. Pat. No. 6,351,556 to detect sub-event boundaries. Another method of automatically organizing images into events is disclosed in commonly assigned U.S. Pat. No. 6,915,011, which is herein incorporated by reference. - The events detected continue to be chronologically ordered in a timeline from earliest to latest. Using the method described above, it is not possible to detect single events that span a long period of time (days) and encompass a variety of activities and settings (for example, a long vacation covering multiple destinations) or events that occur in distinct parts separated by some hours from each other (for example, a sporting event with many matches or a wedding). Gaps in photo-taking corresponding to the overnight period also cause breaks in event continuity. Further processing is needed to detect these super-events, defined as a grouping of multiple contiguous events that may span multiple days. Inter-event duration, defined as the time duration between the last image of one event and the first image of the next event on a continuous timeline, is computed for each event. The events are then treated as single points on a time axis, separated by the inter-event durations. A density-based clustering method is applied to these points (ref. Data Mining Concepts and Techniques by Han and Kamber, Elsevier, 2006, supra, pp. 418-420) to cluster events into super-events when they are separated by relatively small duration gaps (for example, less than 18 hours). The final three-level hierarchical event representation includes super-events, events and sub-events. After this point, the term “event” refers to the top-level of the hierarchical event representation—which can be a super-event or an event. Referring to
FIG. 2 , thedigital image collection 105 is grouped into temporal events, sub-events and super-events using the methods described above. - Referring to
FIG. 2 , significant events are detected in step 115 from thedigital image collection 105. A significant event detection algorithm using time-series analysis of the capture date/time information of the images is used to detect the significant events. The details of the algorithm can be found in U.S. Pat. No. 8,340,436, “Detecting significant events in consumer image collections,” by Das and Loui, issued on Dec. 25, 2012, the disclosure of which is incorporated herein by reference. In U.S. Pat. No. 8,340,436, the predicted output of the selected ARIMA model is compared with the image counts time-series that was used to generate the model. Residuals are computed as the difference between the predicted output of the model and the image counts time-series at each time step. The variance (σ) of the residuals is computed and a threshold is determined based on this variance. Here, we compute an additional significance score defined as the residual divided by the variance (σ). - Referring to
FIG. 2 , the output of theevent hierarchy generator 110 and the significant event detector 115 are fed into theevent ranking module 120. The events can be ranked by a number of different criteria. - In one aspect of the present invention, the number of sub-events in the event is used to rank events in descending order of importance. Since each sub-event extracted using the method disclosed in U.S. Pat. No. 6,606,411 has consistent color distribution as determined by block-level color histogram similarity; more sub-events in an event indicates that these pictures are likely to have been taken with diverse backgrounds that increase the scope of the event. This justifies a higher ranking when there are more sub-events. In another embodiment, the significance score, defined as the residual divided by the variance (a), is used to rank the events, with a higher score getting a higher rank. The significance score generated at the end of the significant event detection described earlier indicates how well the event fits into the estimated model, with a higher score indicating a lower fit, and therefore, the event is more likely to be something unusual and important in the collection.
- In another aspect of the present invention, the interestingness of an event can be modeled as shown in
FIG. 4 . As represented inFIG. 4 , the interestingness score of an event is initially high (close to 1.0) at the time of capture (start time of the event), and then falls rapidly as the first month passes by. The interestingness score again rises around the picture's one-year anniversary mark (because people are often interested in reliving the previous year's happenings, especially if the event is an annual recurring event such as a birthday). The interestingness score then plateaus to a higher level than the first year, and at each subsequent anniversary achieves a slightly higher level than the previous year. The events are ranked according to their interestingness score. - In another aspect of the present invention, the albums of images a user uploads for sharing to social networks are gathered, along with social interactions such as “likes”, comments, and tags associated with each image. The images in these albums are treated as a set of images that have no capture date-time information, but are in a list arranged by the time of upload. This list of images is merged into a user's personal image collection that resides on their private storage (which can be on a personal computer, mobile device or online storage) using the method described in U.S. Pat. No. 7,831,599 “Additive clustering of images lacking individualized date-time information” by Das et al issued Sep. 11, 2010. This patent describes a method that uses a dynamic programming-based formulation to merge images lacking capture time into an organized collection where events have already been computed and capture date-time information exists. The method computes image similarity scores based on image content, and ensures that the ordering of the incoming list of images is maintained. After merging the shared images into the user's personal collection, the number of social interactions (“likes”, comments and tags) derived from the shared images are counted for each event in the user collection that contains shared images from the merging process. The events are ranked in decreasing order of number of social interactions.
- In another aspect of the present invention, the number of images that are marked by the user is counted for each event, and the events are ranked in decreasing order of the number of user markings, The user markings can take different forms including being marked a “favorite” either at time of capture on the capture device itself, or later on the site of storage (computer or online storage); marked as to be used for sharing; or marked with a star rating system provided by the capture device or storage site with the maximum star rating allowed.
- Referring to
FIG. 3 , steps 310-330 refer to additional steps performed with the ranked list of events (step 150 ofFIG. 2 ) to select events from the ranked list. One or more event attributes 310 are computed for each of the events on the ranked list. The event attributes that can be computed include event class, event size, and media type of the event. Event class refers to the classification of the event into one of a finite set of event classes e.g., vacation, party, sports and family moments. A method for event classification is described in US Patent application US 2010/0124378A1 titled “Method for event-based semantic classification”. In this method, a classifier is trained to recognize events of different pre-specified classes. - Event size refers to the number of assets (images or video) contained in the top-level event (i.e., a super-event or an event). The media type of an event refers to the ratio of videos to images in the event, discretized into a pre-specified number of bins. The media type indicates the mix of video and images in an event.
- Referring to
FIG. 3 ,step 320 determines a target distribution of the selected event attribute. The target distribution is initially computed to match the distribution of the attribute in the collection. For this purpose, a histogram is generated where each bin represents a category of the selected event attribute, and the value of the bin is the count of events in the collection with that category. The histogram is then normalized (each bin count is divided by the total number of events in the collection) so that the values are between 0.0 and 1.0. This normalized histogram represents thetarget distribution 320. As an optional step, input from the user 315 can be incorporated to alter the target distribution at this point. For example, if the user prefers a selection representing vacations in the collection, the target distribution of the event class attribute is altered so that the “vacation” class is set to 1.0 and the rest of the classes are set to 0.0. The user input does not need to be binary—an interface could be provided that allows the user to indicate interest in a sliding scale e.g., more vacations, less sports, and these can be translated into corresponding changes in the target distribution. - Referring to
FIG. 3 , step 330 selects events from the ranked list in descending order while maintaining the target distribution. The output product type e.g., photobook, calendar or collage, determines the number (typically not an exact number, but a range) of images needed to create the output product type. The user (or system in case of an auto-generated product) may also provide a sub-set of the whole collection to select from, where the sub-set may be specified by a time range, selected folders or a selection set (default is the whole collection). It is assumed that selecting a fraction of the images in an event provides sufficient representation for the event in an output product. The fraction is based on the type of output product (e.g. calendars may use fewer images than photobooks for the same event). In one embodiment, the fraction is chosen to be 0.1 (i.e., 10% of the images from an event are typically expected to be used in the product). This output product-based requirement for the number of images, fraction of images from an event, and selected sub-set is provided in step 325 as input parameters to the selection step 330. - Referring to
FIG. 5 , the number of images needed (or the number in the center of the range, if a range is provided) is used to proportionally allot number of images according to the target distribution 525. e.g. if the target distribution is based on the event class, the bin value is 0.3 for the “party” class, and 120 images are needed, then 0.3×120=36 images are allotted for images from the “party” event class. This is a rough estimate and need not be accurate, as the product generation system (with or without manual selection by user) can select more or less from any event. The ranked list is filtered 530 by the selected sub-set e.g. if a time range is specified, only events in that time range are retained in the list, and the others are eliminated. The events are selected in step 535 that is shown in detail inFIG. 6 . - Referring to
FIG. 6 , traversing the filtered, ranked event list in descending order, each event encountered in the list is selected if the number of remaining images allotted to that event type is greater than zero. After adding the event, the number of allotted images for that event attribute is decreased by the product of the given fraction and the event size. The process is continued till there is no positive number of remaining images for any event attribute or when the list is exhausted (without meeting all the allotments). In the latter instance, a second pass is performed through the remaining events in the list, after re-allotting the event types that did not have enough candidates, to other event types in proportion to the target distribution values for those event types. This process is continued till either the remaining allotments are all negative, or when there are no events left in the list. - A method for ranking events in media collections comprises designating a media collection, using a processor to cluster the media collection items into a hierarchical event structure, using the processor to identify and count visually similar sub-events within each event in the hierarchical event structure, using the processor to determine a ranking of events based on the count of sub-events within each event, and associating the determined ranking with each event in the media collection.
- The ranking of events can be based on the significance score of the event, on a distribution that models the importance of an event over an elapsed time period, on a score or distribution that models the interestingness of an event over an elapsed time period, on metadata from social networks such as number of likes and comments, on metadata from social networks through the analysis of user tags and comments, or on the number of images in the event that have been marked by the user as being a favorite or to be used for sharing.
- A method for selecting events from media collections comprises designating a media collection, using a processor to cluster the media collection items into a hierarchical event structure, using the processor to identify and count visually similar sub-events within each event in the hierarchical event structure, using the processor to determine a ranked list of events based on the count of sub-events within each event, using the processor to calculate a target distribution that is based on the distribution of one or more event attributes of the events derived from the media collection, and selecting events from the ranked list of events based on the calculated target distribution.
- The event attribute used in the target distribution can be the event class, the event size, or the media type of the event. The ranking of events is based on the significance score of the event, on a distribution that models the importance of an event over an elapsed time period, on scores or a distribution that models the interestingness of an event over an elapsed time period, on metadata from social networks such as number of likes and comments, on metadata from social networks through the analysis of user tags and comments, or on the number of images in the event that have been marked by the user as being a favorite or to be used for sharing.
- A system for ranking events in media collections comprises a processor-accessible memory for storing a media collection, and a processor for clustering the media collection items into a hierarchical event structure, for identifying and visually counting similar sub-events within each event in the hierarchical event structure, for determining a ranking of events based on the count of sub-events within each event, and for associating the determined ranking with each event in the media collection.
- A system for selecting events from media collections comprises a processor-accessible memory for storing a media collection and a processor for clustering the media collection items into a hierarchical event structure, for identifying and visually counting similar sub-events within each event in the hierarchical event structure, for determining a ranked list of events based on the count of sub-events within each event, for calculating a target distribution based on the distribution of one or more event attributes of the events derived from the media collection, and for selecting events from the ranked list of events based on the calculated target distribution. The invention has been described in detail with particular reference to certain preferred embodiments thereof, but it will be understood that variations and modifications can be effected within the spirit and scope of the invention.
-
-
- 10 Computer
- 20 Processor
- 30 Memory
- 50 Input device
- 60 Keyboard
- 70 Display Device
- 80 Network connection
- 105 Digital image collection
- 110 Time-series generator
- 115 Time-series modeling step
- 120 Significant event detector
- 205 Extract date/time step
- 215 Accumulators for different time units
- 225 Group of image counts time-series
- 305 Image counts time-series
- 310 Estimate initial parameters step
- 315 Fit ARIMA models step
- 320 Choose viable models step
- 325 Compute goodness-of-fit measures step
- 330 Choose best ARIMA model step
- 405 Image counts time-series
- 410 ARIMA model
- 415 Compute residuals step
- 420 Determine threshold step
- 430 Identify time steps of interest step
- 440 Identify significant events step
- 510 Significant events
- 520 Additional inputs
- 530 Time granularity selector
Claims (17)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/391,957 US20210365490A1 (en) | 2013-06-27 | 2021-08-02 | Method for ranking and selecting events in media collections |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361840031P | 2013-06-27 | 2013-06-27 | |
US14/310,777 US11080318B2 (en) | 2013-06-27 | 2014-06-20 | Method for ranking and selecting events in media collections |
US17/391,957 US20210365490A1 (en) | 2013-06-27 | 2021-08-02 | Method for ranking and selecting events in media collections |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/310,777 Continuation US11080318B2 (en) | 2013-06-27 | 2014-06-20 | Method for ranking and selecting events in media collections |
Publications (1)
Publication Number | Publication Date |
---|---|
US20210365490A1 true US20210365490A1 (en) | 2021-11-25 |
Family
ID=52116670
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/310,777 Active US11080318B2 (en) | 2013-06-27 | 2014-06-20 | Method for ranking and selecting events in media collections |
US17/391,957 Pending US20210365490A1 (en) | 2013-06-27 | 2021-08-02 | Method for ranking and selecting events in media collections |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/310,777 Active US11080318B2 (en) | 2013-06-27 | 2014-06-20 | Method for ranking and selecting events in media collections |
Country Status (1)
Country | Link |
---|---|
US (2) | US11080318B2 (en) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105447006B (en) | 2014-08-08 | 2019-08-16 | 阿里巴巴集团控股有限公司 | A kind of picture selection method and its device |
US10282384B2 (en) * | 2014-12-29 | 2019-05-07 | Facebook, Inc. | Systems and methods for throttling click bait |
US9268465B1 (en) * | 2015-03-31 | 2016-02-23 | Guguly Corporation | Social media system and methods for parents |
CN113254632B (en) * | 2021-04-22 | 2022-07-22 | 国家计算机网络与信息安全管理中心 | Timeline abstract automatic generation method based on event detection technology |
Citations (33)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5341142A (en) * | 1987-07-24 | 1994-08-23 | Northrop Grumman Corporation | Target acquisition and tracking system |
US5737487A (en) * | 1996-02-13 | 1998-04-07 | Apple Computer, Inc. | Speaker adaptation based on lateral tying for large-vocabulary continuous speech recognition |
US20040193870A1 (en) * | 2003-03-25 | 2004-09-30 | Digital Doors, Inc. | Method and system of quantifying risk |
US20050008193A1 (en) * | 2000-06-13 | 2005-01-13 | Microsoft Corporation | System and process for bootstrap initialization of nonparametric color models |
US20050119547A1 (en) * | 2001-12-13 | 2005-06-02 | Ananda Shastri | Systems and methods for detecting deception by measuring brain activity |
US20060025931A1 (en) * | 2004-07-30 | 2006-02-02 | Richard Rosen | Method and apparatus for real time predictive modeling for chronically ill patients |
US20060104520A1 (en) * | 2004-11-17 | 2006-05-18 | Eastman Kodak Company | Multi-tiered image clustering by event |
US20060120609A1 (en) * | 2004-12-06 | 2006-06-08 | Yuri Ivanov | Confidence weighted classifier combination for multi-modal identification |
US20060253491A1 (en) * | 2005-05-09 | 2006-11-09 | Gokturk Salih B | System and method for enabling search and retrieval from image files based on recognized information |
US20060269111A1 (en) * | 2005-05-27 | 2006-11-30 | Stoecker & Associates, A Subsidiary Of The Dermatology Center, Llc | Automatic detection of critical dermoscopy features for malignant melanoma diagnosis |
US20070098303A1 (en) * | 2005-10-31 | 2007-05-03 | Eastman Kodak Company | Determining a particular person from a collection |
US20080077570A1 (en) * | 2004-10-25 | 2008-03-27 | Infovell, Inc. | Full Text Query and Search Systems and Method of Use |
US20080205771A1 (en) * | 2007-02-28 | 2008-08-28 | Kraus Bryan D | Classifying complete and incomplete date-time information |
US20080292196A1 (en) * | 2007-04-27 | 2008-11-27 | Ramesh Jain | Photo classification using optical parameters of camera from exif metadata |
US20090091798A1 (en) * | 2007-10-05 | 2009-04-09 | Lawther Joel S | Apparel as event marker |
US20090148071A1 (en) * | 2007-12-11 | 2009-06-11 | Sony Corporation | Information processing apparatus, method, and program |
US20090185723A1 (en) * | 2008-01-21 | 2009-07-23 | Andrew Frederick Kurtz | Enabling persistent recognition of individuals in images |
US20090265338A1 (en) * | 2008-04-16 | 2009-10-22 | Reiner Kraft | Contextual ranking of keywords using click data |
US20090324026A1 (en) * | 2008-06-27 | 2009-12-31 | Palo Alto Research Center Incorporated | System and method for finding a picture image in an image collection using localized two-dimensional visual fingerprints |
US20100082301A1 (en) * | 2008-09-30 | 2010-04-01 | Sense Netwoks, Inc. | Event Identification In Sensor Analytics |
US20110085697A1 (en) * | 2009-10-09 | 2011-04-14 | Ric Clippard | Automatic method to generate product attributes based solely on product images |
US20110115943A1 (en) * | 2009-11-18 | 2011-05-19 | Sony Corporation | Information processing apparatus, information processing method, and program |
US20110122153A1 (en) * | 2009-11-26 | 2011-05-26 | Okamura Yuki | Information processing apparatus, information processing method, and program |
US20120124594A1 (en) * | 2009-07-23 | 2012-05-17 | Nec Corporation | Event processing system, distribution controller, event processing method, distribution control method, and program storage medium |
US20120150835A1 (en) * | 2005-06-27 | 2012-06-14 | Make Sence, Inc. | Knowledge correlation search engine |
US20120294511A1 (en) * | 2011-05-18 | 2012-11-22 | International Business Machines Corporation | Efficient retrieval of anomalous events with priority learning |
US20130053415A1 (en) * | 2010-02-05 | 2013-02-28 | Peter B. Rahl | Combination methods for treatment of disease |
US20130058535A1 (en) * | 2010-06-11 | 2013-03-07 | Technische Universitat Darmstadt | Detection of objects in an image using self similarities |
US20130103615A1 (en) * | 2009-02-11 | 2013-04-25 | Johnathan Mun | Project economics analysis tool |
US20130148852A1 (en) * | 2011-12-08 | 2013-06-13 | Canon Kabushiki Kaisha | Method, apparatus and system for tracking an object in a sequence of images |
US20130223714A1 (en) * | 2012-02-28 | 2013-08-29 | Albert Einstein College Of Medicine Of Yeshiva University | Methods for quantitative assessment of volumetric image from a subject and uses therof |
US20140348423A1 (en) * | 2011-07-08 | 2014-11-27 | Nikon Corporation | Image sorting method |
US20150131902A1 (en) * | 2012-06-18 | 2015-05-14 | St-Ericsson Sa | Digital Image Analysis |
Family Cites Families (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6460036B1 (en) * | 1994-11-29 | 2002-10-01 | Pinpoint Incorporated | System and method for providing customized electronic newspapers and target advertisements |
US6606411B1 (en) * | 1998-09-30 | 2003-08-12 | Eastman Kodak Company | Method for automatically classifying images into events |
US6351556B1 (en) | 1998-11-20 | 2002-02-26 | Eastman Kodak Company | Method for automatically comparing content of images for classification into events |
US8429026B1 (en) * | 1999-06-28 | 2013-04-23 | Dietfood Corp. | System and method for creating and submitting electronic shopping lists |
US7406436B1 (en) * | 2001-03-22 | 2008-07-29 | Richard Reisman | Method and apparatus for collecting, aggregating and providing post-sale market data for an item |
US6915011B2 (en) | 2001-03-28 | 2005-07-05 | Eastman Kodak Company | Event clustering of images using foreground/background segmentation |
US7519589B2 (en) * | 2003-02-04 | 2009-04-14 | Cataphora, Inc. | Method and apparatus for sociological data analysis |
US6946715B2 (en) * | 2003-02-19 | 2005-09-20 | Micron Technology, Inc. | CMOS image sensor and method of fabrication |
US6865297B2 (en) * | 2003-04-15 | 2005-03-08 | Eastman Kodak Company | Method for automatically classifying images into events in a multimedia authoring application |
US7831599B2 (en) | 2005-03-04 | 2010-11-09 | Eastman Kodak Company | Addition of new images to an image database by clustering according to date/time and image content and representative image comparison |
US20070100563A1 (en) * | 2005-10-27 | 2007-05-03 | Webb Peter G | Probe selection methods, computer program products, systems and arrays |
JP4624248B2 (en) * | 2005-12-06 | 2011-02-02 | 富士フイルム株式会社 | Image processing apparatus, skin color adjustment method, and program |
US7869658B2 (en) * | 2006-10-06 | 2011-01-11 | Eastman Kodak Company | Representative image selection based on hierarchical clustering |
JP5200015B2 (en) * | 2007-06-14 | 2013-05-15 | パナソニック株式会社 | Image recognition apparatus and image recognition method |
US8122356B2 (en) * | 2007-10-03 | 2012-02-21 | Eastman Kodak Company | Method for image animation using image value rules |
AU2007231756B2 (en) * | 2007-10-30 | 2011-04-07 | Canon Kabushiki Kaisha | A method of segmenting videos into a hierachy of segments |
US8150098B2 (en) * | 2007-12-20 | 2012-04-03 | Eastman Kodak Company | Grouping images by location |
US7860866B2 (en) | 2008-03-26 | 2010-12-28 | Microsoft Corporation | Heuristic event clustering of media using metadata |
US9569531B2 (en) * | 2008-06-23 | 2017-02-14 | Sri International | System and method for multi-agent event detection and recognition |
US8611677B2 (en) | 2008-11-19 | 2013-12-17 | Intellectual Ventures Fund 83 Llc | Method for event-based semantic classification |
US8024311B2 (en) | 2008-12-05 | 2011-09-20 | Eastman Kodak Company | Identifying media assets from contextual information |
US20110213655A1 (en) * | 2009-01-24 | 2011-09-01 | Kontera Technologies, Inc. | Hybrid contextual advertising and related content analysis and display techniques |
EP2224357A1 (en) * | 2009-02-27 | 2010-09-01 | BRITISH TELECOMMUNICATIONS public limited company | Video segmentation |
US8340436B2 (en) | 2009-06-19 | 2012-12-25 | Eastman Kodak Company | Detecting significant events in consumer image collections |
US8392430B2 (en) * | 2009-09-23 | 2013-03-05 | Microsoft Corp. | Concept-structured image search |
US8718386B2 (en) * | 2010-02-23 | 2014-05-06 | Intellectual Ventures Fund 83 Llc | Adaptive event timeline in consumer image collections |
US8345934B2 (en) * | 2010-07-19 | 2013-01-01 | Telefonica, S.A. | Method for automatic storytelling for photo albums using social network context |
ES2711196T3 (en) * | 2010-07-30 | 2019-04-30 | Fund D Anna Sommer Champalimaud E Dr Carlos Montez Champalimaud | Systems and procedures for the segmentation and processing of tissue images and extraction of tissue characteristics to treat, diagnose or predict medical conditions |
US8892605B2 (en) * | 2010-12-03 | 2014-11-18 | Relationship Capital Technologies, Inc. | Systems and methods for managing social networks based upon predetermined objectives |
US8823726B2 (en) * | 2011-02-16 | 2014-09-02 | Apple Inc. | Color balance |
GB201113633D0 (en) * | 2011-08-08 | 2011-09-21 | Vision Semantics Ltd | User-primed iterative discovery of object associations in distributed multi-source data |
US8929687B2 (en) * | 2011-08-29 | 2015-01-06 | Dst Technologies, Inc. | System and method for candidate sorting and clustering |
US8855430B1 (en) * | 2012-05-30 | 2014-10-07 | Google Inc. | Refining image annotations |
US8873851B2 (en) * | 2012-06-29 | 2014-10-28 | Intellectual Ventures Fund 83 Llc | System for presenting high-interest-level images |
-
2014
- 2014-06-20 US US14/310,777 patent/US11080318B2/en active Active
-
2021
- 2021-08-02 US US17/391,957 patent/US20210365490A1/en active Pending
Patent Citations (33)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5341142A (en) * | 1987-07-24 | 1994-08-23 | Northrop Grumman Corporation | Target acquisition and tracking system |
US5737487A (en) * | 1996-02-13 | 1998-04-07 | Apple Computer, Inc. | Speaker adaptation based on lateral tying for large-vocabulary continuous speech recognition |
US20050008193A1 (en) * | 2000-06-13 | 2005-01-13 | Microsoft Corporation | System and process for bootstrap initialization of nonparametric color models |
US20050119547A1 (en) * | 2001-12-13 | 2005-06-02 | Ananda Shastri | Systems and methods for detecting deception by measuring brain activity |
US20040193870A1 (en) * | 2003-03-25 | 2004-09-30 | Digital Doors, Inc. | Method and system of quantifying risk |
US20060025931A1 (en) * | 2004-07-30 | 2006-02-02 | Richard Rosen | Method and apparatus for real time predictive modeling for chronically ill patients |
US20080077570A1 (en) * | 2004-10-25 | 2008-03-27 | Infovell, Inc. | Full Text Query and Search Systems and Method of Use |
US20060104520A1 (en) * | 2004-11-17 | 2006-05-18 | Eastman Kodak Company | Multi-tiered image clustering by event |
US20060120609A1 (en) * | 2004-12-06 | 2006-06-08 | Yuri Ivanov | Confidence weighted classifier combination for multi-modal identification |
US20060253491A1 (en) * | 2005-05-09 | 2006-11-09 | Gokturk Salih B | System and method for enabling search and retrieval from image files based on recognized information |
US20060269111A1 (en) * | 2005-05-27 | 2006-11-30 | Stoecker & Associates, A Subsidiary Of The Dermatology Center, Llc | Automatic detection of critical dermoscopy features for malignant melanoma diagnosis |
US20120150835A1 (en) * | 2005-06-27 | 2012-06-14 | Make Sence, Inc. | Knowledge correlation search engine |
US20070098303A1 (en) * | 2005-10-31 | 2007-05-03 | Eastman Kodak Company | Determining a particular person from a collection |
US20080205771A1 (en) * | 2007-02-28 | 2008-08-28 | Kraus Bryan D | Classifying complete and incomplete date-time information |
US20080292196A1 (en) * | 2007-04-27 | 2008-11-27 | Ramesh Jain | Photo classification using optical parameters of camera from exif metadata |
US20090091798A1 (en) * | 2007-10-05 | 2009-04-09 | Lawther Joel S | Apparel as event marker |
US20090148071A1 (en) * | 2007-12-11 | 2009-06-11 | Sony Corporation | Information processing apparatus, method, and program |
US20090185723A1 (en) * | 2008-01-21 | 2009-07-23 | Andrew Frederick Kurtz | Enabling persistent recognition of individuals in images |
US20090265338A1 (en) * | 2008-04-16 | 2009-10-22 | Reiner Kraft | Contextual ranking of keywords using click data |
US20090324026A1 (en) * | 2008-06-27 | 2009-12-31 | Palo Alto Research Center Incorporated | System and method for finding a picture image in an image collection using localized two-dimensional visual fingerprints |
US20100082301A1 (en) * | 2008-09-30 | 2010-04-01 | Sense Netwoks, Inc. | Event Identification In Sensor Analytics |
US20130103615A1 (en) * | 2009-02-11 | 2013-04-25 | Johnathan Mun | Project economics analysis tool |
US20120124594A1 (en) * | 2009-07-23 | 2012-05-17 | Nec Corporation | Event processing system, distribution controller, event processing method, distribution control method, and program storage medium |
US20110085697A1 (en) * | 2009-10-09 | 2011-04-14 | Ric Clippard | Automatic method to generate product attributes based solely on product images |
US20110115943A1 (en) * | 2009-11-18 | 2011-05-19 | Sony Corporation | Information processing apparatus, information processing method, and program |
US20110122153A1 (en) * | 2009-11-26 | 2011-05-26 | Okamura Yuki | Information processing apparatus, information processing method, and program |
US20130053415A1 (en) * | 2010-02-05 | 2013-02-28 | Peter B. Rahl | Combination methods for treatment of disease |
US20130058535A1 (en) * | 2010-06-11 | 2013-03-07 | Technische Universitat Darmstadt | Detection of objects in an image using self similarities |
US20120294511A1 (en) * | 2011-05-18 | 2012-11-22 | International Business Machines Corporation | Efficient retrieval of anomalous events with priority learning |
US20140348423A1 (en) * | 2011-07-08 | 2014-11-27 | Nikon Corporation | Image sorting method |
US20130148852A1 (en) * | 2011-12-08 | 2013-06-13 | Canon Kabushiki Kaisha | Method, apparatus and system for tracking an object in a sequence of images |
US20130223714A1 (en) * | 2012-02-28 | 2013-08-29 | Albert Einstein College Of Medicine Of Yeshiva University | Methods for quantitative assessment of volumetric image from a subject and uses therof |
US20150131902A1 (en) * | 2012-06-18 | 2015-05-14 | St-Ericsson Sa | Digital Image Analysis |
Also Published As
Publication number | Publication date |
---|---|
US20150006523A1 (en) | 2015-01-01 |
US11080318B2 (en) | 2021-08-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20210365490A1 (en) | Method for ranking and selecting events in media collections | |
US20220004573A1 (en) | Method for creating view-based representations from multimedia collections | |
US20150006545A1 (en) | System for ranking and selecting events in media collections | |
US8718386B2 (en) | Adaptive event timeline in consumer image collections | |
US11500907B2 (en) | System and method for user-behavior based content recommendations | |
US8811755B2 (en) | Detecting recurring events in consumer image collections | |
US8340436B2 (en) | Detecting significant events in consumer image collections | |
JP2013520725A5 (en) | ||
US20140093174A1 (en) | Systems and methods for image management | |
US9229958B2 (en) | Retrieving visual media | |
CN102150163A (en) | Interactive image selection method | |
US9201900B2 (en) | Related image searching method and user interface controlling method | |
JP2014092955A (en) | Similar content search processing device, similar content search processing method and program | |
CN112099709A (en) | Method and device for arranging multimedia objects, electronic equipment and storage medium | |
Das et al. | Collaborative content synchronization through an event-based framework |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
AS | Assignment |
Owner name: KODAK ALARIS INC., NEW YORK Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:DAS, MADIRAKSHI;LOUI, ALEXANDER C.;REEL/FRAME:060701/0501 Effective date: 20150527 |
|
AS | Assignment |
Owner name: THE BOARD OF THE PENSION PROTECTION FUND, UNITED KINGDOM Free format text: IP SECURITY AGREEMENT SUPPLEMENT (FISCAL YEAR 2022);ASSIGNOR:KODAK ALARIS INC.;REEL/FRAME:061504/0900 Effective date: 20220906 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |