WO2012153744A1 - Information processing device, information processing method, and information processing program - Google Patents

Information processing device, information processing method, and information processing program Download PDF

Info

Publication number
WO2012153744A1
WO2012153744A1 PCT/JP2012/061788 JP2012061788W WO2012153744A1 WO 2012153744 A1 WO2012153744 A1 WO 2012153744A1 JP 2012061788 W JP2012061788 W JP 2012061788W WO 2012153744 A1 WO2012153744 A1 WO 2012153744A1
Authority
WO
WIPO (PCT)
Prior art keywords
still image
still
image pair
relationship
relevance
Prior art date
Application number
PCT/JP2012/061788
Other languages
French (fr)
Japanese (ja)
Inventor
真澄 石川
Original Assignee
日本電気株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 日本電気株式会社 filed Critical 日本電気株式会社
Publication of WO2012153744A1 publication Critical patent/WO2012153744A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation
    • G06T13/802D [Two Dimensional] animation, e.g. using sprites
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N1/00Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
    • H04N1/00127Connection or combination of a still picture apparatus with another apparatus, e.g. for storage, processing or transmission of still picture signals or of information associated with a still picture
    • H04N1/00129Connection or combination of a still picture apparatus with another apparatus, e.g. for storage, processing or transmission of still picture signals or of information associated with a still picture with a display device, e.g. CRT or LCD monitor
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N1/00Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
    • H04N1/0035User-machine interface; Control console
    • H04N1/00405Output means
    • H04N1/00408Display of information to the user, e.g. menus
    • H04N1/0044Display of information to the user, e.g. menus for image preview or review, e.g. to help the user position a sheet
    • H04N1/00458Sequential viewing of a plurality of images, e.g. browsing or scrolling
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N1/00Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
    • H04N1/21Intermediate information storage
    • H04N1/2104Intermediate information storage for one or a few pictures
    • H04N1/2112Intermediate information storage for one or a few pictures using still video cameras
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/61Control of cameras or camera modules based on recognised objects
    • H04N23/611Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/79Processing of colour television signals in connection with recording
    • H04N9/80Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
    • H04N9/82Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
    • H04N9/8205Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
    • H04N9/8227Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal the additional signal being at least another television signal
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N2101/00Still video cameras
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N2201/00Indexing scheme relating to scanning, transmission or reproduction of documents or the like, and to details thereof
    • H04N2201/0077Types of the still picture apparatus
    • H04N2201/0084Digital still camera

Definitions

  • the present invention relates to an information processing device, an information processing method, and an information processing program, and more particularly to an information processing device, an information processing method, and an information processing program for determining a still image presentation method.
  • Patent Document 1 is a slide show generation technique that realizes a presentation time according to the content of each still image. Specifically, the presentation time of the still image is to be controlled by the number of faces included in the still image.
  • the technique of Patent Document 2 controls the presentation time according to the still image shooting time when generating a slide show.
  • the presentation time of continuous still images in a slide show is determined based on the difference in shooting time of still images.
  • the technique of Patent Document 3 extracts the degree of association between images and reproduces a slide show while giving a visual effect based on the degree of association (paragraph 0024).
  • This technique acquires two images to be compared and extracts the degree of association for each image (paragraph 0063).
  • This technique provides a visual effect for images with low relevance when the relevance is weak and for images with high relevance when the relevance is strong (paragraph 0074).
  • Patent Literatures 1, 2, and 3 determine a still image presentation method using information of at least one still image. Therefore, these techniques represent changes in the contents of two consecutive still images by changes in the presentation method. However, these techniques do not change the presentation method when the relationship between two consecutive still images does not change.
  • the objective of this invention is providing the information processing apparatus which solves the above-mentioned subject.
  • an information processing apparatus in a still image group including at least three still images, has a first relationship between still images of a first still image pair included in the still image group.
  • the comparison means for comparing the second relevance between the still images of the second still image pair, and the presentation method of the first still image pair are specified based on the change from the first relevance to the second relevance
  • the presentation method determination means for determining the second information specifying the presentation method of the second still image pair from the first information
  • generating means for generating a slide show including the second still image pair.
  • an information processing method includes, in a still image group including at least three still images, a first association between still images of a first still image pair included in the still image group.
  • the first still image pair is compared with the second relationship between the still images of the second still image pair, and the presentation method of the first still image pair is specified based on the change from the first relationship to the second relationship.
  • an information processing program causes a computer to select a first still image of a pair of first still images included in a still image group in a still image group including at least three still images.
  • Comparison means for comparing the relevance and the second relevance between the still images of the second still image pair, and presentation of the first still image pair based on a change from the first relevance to the second relevance
  • the presentation method determining means for determining the second information specifying the presentation method of the second still image pair, and the presentation method determined in the determining step, the first still image It is operated as a generating means for generating a slide show including a pair and the second still image pair.
  • the information processing apparatus 100 is an apparatus that generates a slide show including at least three still images.
  • the information processing apparatus 100 includes an association comparison unit (comparison unit) 101, a presentation method determination unit 102, and a slide show generation unit (generation unit) 103.
  • the relevance comparison unit 101 includes a first relevance between still images of the first still image pair included in the still image group and a still image of the second still image pair in the still image group including at least three still images. The second relevance between each other is compared.
  • the presentation method determination unit 102 identifies the presentation method of the second still image pair from the first information that identifies the presentation method of the first still image pair based on the change from the first relevance to the second relevance. The second information is determined.
  • the slide show generation unit 103 generates a slide show including the first still image pair and the second still image pair based on the presentation method determined by the presentation method determination unit 102. According to the above configuration, it is possible to generate a slide show that presents still images according to a change in relevance between still images.
  • the information processing apparatus 200 includes an image input unit 210 that inputs image information from an imaging device 250 such as a digital camera or a digital video camera, a relevance comparison unit (comparison unit) 201 that compares relevance of input images, A relevance determination unit 204 that determines gender.
  • an imaging device 250 such as a digital camera or a digital video camera
  • a relevance comparison unit (comparison unit) 201 that compares relevance of input images
  • a relevance determination unit 204 that determines gender.
  • the information processing apparatus 200 also includes a presentation method determination unit 202 that determines the display length of each image included in the slide show, the effect at the time of switching images, the BGM (Background Music) being displayed, the jingle at the time of switching, and the like. Further, the information processing apparatus 200 includes a slide show generation unit (generation unit) 203 that generates a slide show by combining still images input by the presentation method determined by the presentation method determination unit 202. The information processing apparatus 200 is also connected to a display 260 for displaying a stationary slide show.
  • the image information input by the image input unit 210 includes an image ID (identifier) for identifying a still image, a presentation order in a slide show, and pixel information of the still image.
  • the image information may include meta information describing a subject, a shooting location, and a shooting time shown in a still image, and sensor information such as GPS (Global Positioning System).
  • the relevancy determination unit 204 determines the first relevance based on the commonality of objects represented in two still images included in the first still image pair.
  • the relevance determination unit 204 determines the second relevance based on the commonality of objects represented in two still images included in the second still image pair.
  • the first still image pair and the second still image pair are pairs of continuous still images, respectively.
  • the relevancy determination unit 204 inputs the image ID of the still image and the relevance flag to the relevance comparison unit 201 as the image relevance information.
  • the relevancy determination unit 204 may input pixel information in addition to the above as image relevance information.
  • the relevance flag is data representing a relevance type existing between a current still image and a still image presented thereafter among relevance types defined in advance. Alternatively, the relevance flag is data indicating that no relevance type exists between these images (no relevance).
  • a flag 1 is set to the relevance flags of all relevance types existing between a certain still image and a subsequent still image.
  • a flag 0 is set for the relationship flag of the relationship type that does not exist.
  • the relevance flag may be set with any numerical value that has meaning depending on the relevance type.
  • the relationship comparison unit 201 compares the first relationship between the still images of the first still image pair included in the still image group and the second relationship between the still images of the second still image pair. Specifically, the relevance comparison unit 201 determines the relevance between pairs of still images that are consecutive in the slide show.
  • the still image group includes continuous first, second, and third still images.
  • the first still image pair is a pair of a first still image and a second still image.
  • the second still image pair is a pair of the second still image and the third still image.
  • the presentation method determination unit 202 presents each still image based on the image relevance information input from the relevance comparison unit 201, the image information input from the image input unit 210, and a pre-registered presentation rule. Decide how. Then, the presentation method determination unit 202 inputs the presentation method information to the slide show generation unit 203.
  • the slide show generation unit 203 generates a slide show by combining still images based on the determined presentation method.
  • the relevancy determination unit 204 determines relevance based on the commonality (identity) of objects represented in two still images included in the still image pair.
  • the relationship 1 is the relationship determined in this way.
  • the identity can be determined by the commonality of the feature quantities derived from a plurality of still images included in the still image group.
  • the relevance 1 is a relevance indicating that the objects shown in the still image pairs continuous in the slide show are the same.
  • the relevance determination unit 204 sets 1 in the relevance flag for relevance 1 if it is the same, and 0 if it is not the same.
  • the relevancy determination unit 204 can determine the identity of the target based on the similarity of the target area images detected from the still images.
  • the target region is a region on a still image of a target having a certain image-like pattern such as a stationary object such as a tree or a building, or a moving body such as a human or an animal.
  • the target area may be a partial area in the still image.
  • the target area may be the entire still image.
  • the relevancy determination unit 204 detects a target region from a pair of still images that are continuous in the slide show.
  • the relevancy determination unit 204 determines the identity of the target based on the similarity of the target area.
  • the relevancy determination unit 204 may determine the identity based on the similarity between the target areas detected from all the still images included in the slide show.
  • the relevance determination unit 204 groups all detected target areas based on similarity.
  • the relevance determination unit 204 determines that the target areas are the same when the target areas detected from adjacent still image pairs belong to the same group.
  • the detection method of the target area is divided into a detection method for detecting a specific target registered in advance and a detection method for detecting a general target that is not registered.
  • the image data of each registered target may be used as a template.
  • the relevancy determination unit 204 may scan the input image with templates converted into various resolutions.
  • the relevancy determination unit 204 may detect a region having a small difference in pixel values at the same position as the template as a corresponding target region.
  • the relevancy determination unit 204 may extract an image feature amount expressing color, texture, and shape from each partial region of the input image.
  • the relevancy determination unit 204 may set a partial region having an image feature amount similar to the registered image feature amount of each target as a corresponding target region.
  • the specific target is a person, there is a method of using information obtained from the entire face.
  • a method of storing an image showing various faces as a template and determining that a face exists in the input image when the difference between the input image and the template is equal to or smaller than a threshold value it is conceivable to store a model combining color information such as skin color, edge direction and density in advance, and determine that a face exists when an area similar to the model is detected from an input frame.
  • the face detection method using the characteristics of the luminance distribution that the cheeks and forehead are bright and the eyes and mouth are low, and the face is detected using the face symmetry and skin color area and position.
  • the method of performing etc. is mentioned.
  • a method for statistically learning the feature value distribution obtained from a large number of human face and non-face learning samples and determining whether the feature value obtained from the input image belongs to the face or non-face distribution. examples include a method using a neural network, a support vector machine, an AdaBoost method, and the like.
  • detecting a general target for example, Normalized Cut, Saliency Map, Depth of Field (DoF), or the like may be used.
  • Normalized Cut is a method of dividing an image into a plurality of regions. For details, see Jianbo Shi and Jitendra Malik, “Normalized Cuts and Image Segmentation”, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 22, no. 8. August 2000 has a disclosure.
  • the relevancy determination unit 204 may detect a region located in the center of the screen among regions divided by Normalized Cut as a target region.
  • Saliency Map is a method of calculating an object region in an image from visual attention. For SaliencyMap, see L. Itti, C.I. Koch and E.M. Niebur, “A Model of Saliency-based Visual Attention for Rapid Scene Analysis,” IEEE Trans. Pattern Analysis and Machine Intelligence, Vol. 20, no. 11, pp.
  • the relevancy determination unit 204 may detect, as a target region, a region for which a high importance level has been calculated by the Saliency Map. Dof is a method based on the characteristic that the target edge existing within the depth of field is not blurred and the edge outside the depth of field is blurred. For details, see 3Du-Ming Tsai, Hu-Jong Wang, “Segmenting focused objects in complex visual images”, Pattern Recognition Letters, Vol. 19, pp. 929 940, 1998. There is a disclosure. The relevancy determination unit 204 may calculate the amount of blur based on the thickness of the edge, combine edges with less blur, and detect a focused area as a target area.
  • the relevancy determination unit 204 is a position in a still image or high visibility (lighting conditions, orientation, angle, position on the screen, hiding by another object, blur, facial expression (in the case of a person), etc.
  • One evaluation area may be detected for each still image based on an evaluation value indicating the quality of the reflection based on the above or the appearance frequency of a plurality of images.
  • the relevancy determination unit 204 may combine a plurality of detected target areas into one target area.
  • the relevancy determination unit 204 may extract the image feature amount, and calculate the similarity between the target regions based on a scale that calculates a higher value as the difference in the image feature amount is smaller.
  • the relevancy determination unit 204 can calculate an image feature amount based on image information such as a color, an edge, and a texture detected from the target region. Alternatively, the relevancy determination unit 204 may detect local feature points such as SIFT (Scale-Invariant Feature Transform) from each target image region. Then, the relevance determination unit 204 may associate feature points between image regions. The relevancy determination unit 204 may use a scale that calculates a higher value as the number of associated feature points is larger or the positional relationship between the associated feature points is similar between images.
  • SIFT Scale-Invariant Feature Transform
  • the presentation method determination unit 202 selects the second presentation method information from the first presentation method information so that the presentation method in the second still image pair changes, similarly to the change in the presentation method in the first still image pair.
  • Determine presentation method information For example, when there is no relevance change, the presentation method determination unit 202 determines the presentation method information so that the presentation time changes in the second still image pair, similarly to the change in the presentation time in the first still image pair.
  • the presentation method information is data indicating the presentation method of each still image.
  • the presentation method information includes an image ID and a presentation time.
  • the presentation method information may include an effect, BGM, audio jingle, and video jingle.
  • the presentation rule is a rule that defines a method of presenting a still image according to the relevance type. It is assumed that the presentation method determination unit 202 holds a parameter that defines each presentation time of continuous still image pairs as a presentation rule. In addition to the presentation time, the presentation method determination unit 202 may hold control parameters related to effects, BGM, and jingles (short video, music, and sound effects) inserted between still images. In addition, the presentation rule may define a presentation method in a case where no relevance type exists in a continuous still image pair. In the present embodiment, the presentation method determination unit 202 determines the presentation time of a still image pair based on the identity of objects included in successive still image pairs.
  • the presentation method determination unit 202 sets the presentation time of the still image to be presented first to the initial value Ts.
  • the presentation method determination unit 202 determines the presentation time of the subsequent still image based on Ts. If the objects included in the continuous still image pairs are not the same, the presentation method determination unit 202 determines the subsequent presentation time independently of the previous still image presentation time.
  • the presentation method determination unit 202 may set the subsequent presentation time to, for example, the initial value Ts.
  • the presentation method determination unit 202 may set the subsequent presentation time to a random value within a specified range. Note that the presentation method determination unit 202 may set Tp as the presentation time of a still image with high visibility among the group of still images obtained by photographing the same object.
  • the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Tp. In addition, the presentation method determination unit 202 may set the presentation time of the next still image in which the presentation time of the still image is equal to or less than Tq among the still image groups obtained by photographing the same target as the initial value Ts. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Ts. In addition, the presentation method determination unit 202 may set the presentation time of the still image presented last among the still image groups obtained by photographing the same target as the initial value Ts. The presentation method determination unit 202 may calculate the values of Ts and Tp according to the number of images to be presented in consideration of the preset presentation time of the entire slide show.
  • the presentation method determination unit 202 calculates the presentation time of the subsequent still image by multiplying the parameter a by the presentation time of a certain still image.
  • the presentation time of the subsequent still images 302 to 305 is expressed by the following equation (1).
  • the visibility evaluation value of the still image 303 facing the front is equal to or greater than the threshold, if the presentation time of the still image 303 is Tp, the presentation time of the subsequent still image is expressed by the following equation (2). .
  • the information processing apparatus 200 can generate a video in which the presentation time of consecutive images changes even if the consecutive images include the same target.
  • the presentation method determination unit 202 determines an effect, a BGM, and a jingle to be inserted between still image pairs based on the identity of objects included in successive still image pairs. For example, when the target included in the pair of still images is the same, the presentation method determination unit 202 has a special effect (such as dissolve or fade) registered in advance as an effect with little visual change when switching still images. Insert. If they are not the same, the presentation method determination unit 202 inserts special effects (DVE (Digital Video Effects) such as page turning and wipe) registered in advance as effects having a large visual change when switching still images.
  • DVE Digital Video Effects
  • the presentation method determination unit 202 gradually shortens (lengthens) the length of the effect jingle.
  • the presentation method determination unit 202 gradually decreases (increases) the volume of the BGM. Further, for example, when the targets included in the continuous still image pairs are the same, the presentation method determination unit 202 plays the same BGM during the presentation of the still image pairs. When the targets included in the continuous still image pairs are not the same, the presentation method determination unit 202 stops the BGM or switches to a different BGM when switching the still images. In addition, the presentation method determination unit 202 may insert jingles between images that do not have identity. Thereby, the still image group which image
  • the presentation rules for controlling the presentation method are rules based on the magnitude relationship or partial relationship for still image pairs with the same target area, and rules based on homogeneity for still image pairs with the same target area. .
  • the rules based on the magnitude relationship, the partial relationship, and the homogeneity will be described in detail in the third and subsequent embodiments.
  • the image input unit 210 inputs the image information of the still image 501 to the relevancy determination unit 204, and proceeds to step S403. If the input still image is a start image, the process returns from step S403 to step S401, and the image input unit 210 inputs the image information of the second still image 502 to the relevancy determination unit 204 (step S403).
  • step S ⁇ b> 405 the relevance determination unit 204 detects a target area from the still images 501 and 502. Assume that in the relevancy determination unit 204, buildings, flowers, and people are registered in advance as targets, and each model is learned. Then, the relevancy determination unit 204 detects a part surrounded by a solid line rectangle as a target area of the building from the still images 501 and 502. The relevancy determination unit 204 extracts image feature amounts from the pixel information of the target region 0 and the target region 1, and determines identity, magnitude relationship, partial relationship, and homogeneity based on the similarity between the regions. Since the target areas 0 and 1 are detected as the types of buildings, it is determined that there is a homogeneity.
  • a broken-line rectangular area on the still image 501 is detected as a common area of the target area 1 and the target area 0. It is determined that the target areas 1 and 0 have a magnitude relationship. Further, since there is no area other than the common area on the target area 0, it is determined that there is no partial relationship. Therefore, the relevance flags between the still image 501 and the still image 502 are 1, ⁇ 1, 0, 1 in the order of identity, magnitude relationship, partial relationship, and homogeneity (step S407).
  • the presentation method determination unit 202 determines the presentation method based on the image ID and the relevance flag as the image relevance information. Since the target areas of the still image 501 and the still image 502 are the same, a rule based on a magnitude relationship or a partial relationship is applied.
  • the presentation method determination unit 202 sets the presentation time of the still image 501 that is the start image to the initial value Ts.
  • the presentation method determination unit 202 sets the presentation time of the still image 502 to a * Ts because the magnitude relationship between the still images 501 and 502 is a small / large relationship. Since the still images 501 and 502 have a magnitude relationship, the presentation method determination unit 202 inserts a dissolve with little visual change as an effect of switching between the still images 501 and 502 (step S409).
  • the slide show generation unit 203 generates a slide show using the still images 501 and 502 with the determined presentation time / effect (step S411). The above steps are performed for all still images (step S413). In FIG.
  • the type of the target area detected from the still image is the target area 601.
  • the relevance flag for each relevance type is 602.
  • the presentation time length and the effect determined by the presentation method determination unit 202 are the presentation time length 603 and the effect 604.
  • the presentation method determination unit 202 controls the presentation method of continuous still images according to the relevance of each other.
  • even when consecutive still images include the same number of face images if there is no relevance in terms of content, it can be shown to the viewer that there is no relevance. .
  • the presentation method may be changed according to any one of the following relevance changes, and any one of the following presentation rules may be adopted: . (Relevance 2. Target size relationship)
  • the relevancy determination unit 204 may determine the relevance based on the magnitude relationship between the objects represented in the two still images included in the still image pair.
  • the relevancy determination unit 204 may determine relevance by changing the size of a region including a specific object registered in advance in two still images included in a still image pair.
  • the relevance 2 is the relevance determined in this way.
  • the “target size relationship” means that the targets included in the still image pairs that are continuous in the slide show are the same, and the area of the target region has a difference greater than a specified value. For example, there is a case where a target is introduced by generating a slide show by combining an image including the periphery of the target and an image obtained by photographing only the target.
  • the relevancy determination unit 204 can determine the magnitude relationship between objects based on the areas of partial areas common to the target areas determined to be the same or the distances between feature points included in the common partial areas.
  • the relevancy determination unit 204 can determine that the larger the distance between feature points is, the larger the object is photographed.
  • the relevancy determination unit 204 may determine between target areas determined to be the same between a pair of still images that are consecutive in a slide show. In this case, the relevancy determination unit 204 determines that the relevance flag for relevance 2 has a larger area of the target area in the next still image than the area of the target area in the still image. 1 is set when it is small, and -1 is set when there is no magnitude relationship.
  • the relevancy determination unit 204 compares the areas of the partial areas common to the target areas determined to be the same among the target areas detected from all the still images included in the slide show or the distance between the feature points, The magnitude relationship may be determined.
  • the relevancy determination unit 204 reduces the same target area smaller than (Smax + 2Smin) / 3 based on the maximum area Smax and the minimum area Smin of the partial areas common to the target areas determined to be the same. That's fine.
  • the relevancy determination unit 204 may set the same target region larger than (Smax + 2Smin) / 3 and smaller than (2Smax + Smin) / 3.
  • the relevancy determination unit 204 may increase the same target area larger than (2Smax + Smin) / 3. In this case, the relevancy determination unit 204 sets the relevance flag to 1 if the target area in a still image and the target area in the next still image have a small-medium or medium-large relationship. .
  • the relevancy determination unit 204 sets 2 if the target area in a certain still image and the target area in the next still image have a small and large relationship.
  • the relevance determination unit 204 sets ⁇ 1 in the relevance flag if the target area in a certain still image and the target area in the next still image have a large-medium or medium-small relationship.
  • the relevancy determination unit 204 sets ⁇ 2 if the target area in a certain still image and the target area in the next still image have a large and small relationship.
  • the relevancy determination unit 204 sets 0 in the relevance flag.
  • the presentation method determination unit 202 changes the presentation method in the same manner assuming that there is no change in relevance when the change of the object from large to small continues. For example, the presentation method determination unit 202 gradually shortens the presentation time at the same time interval. Specifically, the presentation method determination unit 202 controls the presentation method based on the following rules. [Rules according to the target size] (2-1) Rules regarding presentation time The presentation method determination unit 202 determines the presentation time of a still image pair based on the size relationship of objects included in successive still image pairs. For example, the presentation method determination unit 202 sets the presentation time of the first still image to be presented as the initial value Ts among the still image groups having a target size relationship.
  • the presentation method determination unit 202 determines the presentation time of the subsequent still image based on Ts.
  • the presentation method determination unit 202 may set Tp as the presentation time of a still image with high visibility among a group of still images having a target size relationship.
  • the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Tp.
  • the presentation method determination unit 202 may set, as the initial value Ts, the presentation time of the next still image in which the presentation time of the still image is equal to or less than Tq among the still image groups having a magnitude relationship.
  • the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Ts.
  • the presentation method determination unit 202 may set the presentation time of the still image presented last among the still image groups having a magnitude relationship to Ts.
  • the presentation method determination unit 202 may calculate the values of Ts and Tp according to the number of images to be presented from the preset presentation time of the entire slide show.
  • the presentation method determination unit 202 determines the presentation time of the subsequent still image independently of the presentation time of the previous still image. For example, the presentation method determination unit 202 may set the presentation time of the subsequent still image to the initial value Ts.
  • the presentation method determination unit 202 may set the presentation time of the subsequent still image to a random value within a specified range.
  • the relevancy determination unit 204 determines the magnitude relationship between successive still images by comparing the areas between target regions determined to be the same among target regions detected from all the still images included in the slide show. .
  • the presentation method determination unit 202 calculates the presentation time of the next still image by multiplying the presentation time of a certain still image by the relevance flag parameter a.
  • the presentation time of the first still image 701 is the initial value Ts
  • the still images 701 and 702 have a small and medium relationship
  • the still images 702 and 703 have a medium and large relationship
  • the presentation time of the still image 702 is a ⁇ Ts (multiplication of a).
  • the presentation time of the still image 703 is a ⁇ a ⁇ Ts (multiplication of a). Since the relevance flag of the still images 703 and 704 is ⁇ 2, the presentation time of 704 is Ts (a ⁇ a division).
  • the information processing apparatus 200 can generate a video in which the presentation time of consecutive images changes even if the consecutive images include the same target. For this reason, this embodiment has an effect that a slide show with a tempo that does not bore viewers can be generated. (2-2) Rules regarding effects, BGM, and jingles
  • the presentation method determination unit 202 determines an effect, a BGM, and a jingle to be inserted between a pair of still images based on the size relationship of objects included in successive still image pairs.
  • the presentation method determination unit 202 performs special effects (dissolve or fade) registered in advance as effects with little visual change when switching still images. Etc.).
  • special effects dissolve or fade
  • the presentation method determination unit 202 displays special effects (page turning, wipe, etc.) registered in advance as effects having a large visual change when switching still images. DVE) is inserted.
  • the presentation method determination unit 202 plays the same BGM during the presentation of the still image pairs.
  • the presentation method determination unit 202 stops the BGM or switches to a different BGM when switching the still images.
  • the presentation method determination unit 202 may insert jingles between images that do not have a magnitude relationship. Thereby, the still image group which image
  • the target included in a pair of still images does not have a large or small relationship, the image and sound change greatly, so that the viewer notices that the content has changed and concentrates on understanding the content of the slide show. Can do. (Relevance 3.
  • the relevancy determination unit 204 may determine the relevance based on a partial relationship of objects represented in two still images included in the still image pair. That is, the relevancy determination unit 204 may determine whether the object represented in the two still images included in the still image pair is in a relationship between the whole and the part.
  • the relationship 3 is the relationship determined in this way. “In a partial relationship of objects” represents a relationship in which the objects shown in consecutive still image pairs in the target slide show are the same and are images obtained by capturing different target parts. For example, when it is desired to shoot a wide landscape, a large object, or a long object, a case where a whole image is expressed by combining a still image obtained by capturing a part of the object and performing a slide show.
  • the relevancy determination unit 204 sets 1 in the relevance flag for the relevance 3.
  • the relevance determination unit 204 sets 0 in the relevance flag for relevance 3 when the partial relationship is not the target.
  • the relevancy determination unit 204 can determine the target partial relationship based on a partial area (common area) that is common to the target areas determined to be the same in consecutive still images in the slide show. For example, the relevancy determination unit 204 uses one of the target areas as a template. Then, the relevancy determination unit 204 scans the other target area, detects a position with a small difference, and sets the overlapping area as a common area.
  • the relevance determining unit 204 determines that the regions other than the common region of each target region are in a target partial relationship when both regions are equal to or larger than the specified area. Alternatively, the relevancy determination unit 204 may perform the determination based on the relative position of the target area determined to be the same from all the still images included in the slide show. When the change of the object from the whole to the part continues, the presentation method determination unit 202 assumes that there is no change in relevance, and gradually shortens the presentation time, for example, at the same time interval. Specifically, the presentation method determination unit 202 controls the presentation method based on the following rules.
  • the presentation method determination unit 202 determines the presentation time of a still image pair based on the target partial relationship included in the continuous still image pair. For example, the presentation method determination unit 202 sets the presentation time of the first still image to be presented as the initial value Ts among the still image groups in the target partial relationship. Then, the presentation method determination unit 202 determines the presentation time of the subsequent still image based on Ts. In addition, the presentation method determination unit 202 may set Tp as the presentation time of a still image with high visibility among the still image groups in the target partial relationship. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Tp.
  • the presentation method determination unit 202 may set, as an initial value Ts, the presentation time of the next still image in which the presentation time of the still image is equal to or less than Tq among the still image groups having a partial relationship. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Ts. In addition, the presentation method determination unit 202 may set the presentation time of the last presented image among the still image groups having a partial relationship to Ts. The presentation method determination unit 202 may calculate the values of Ts and Tp according to the number of images to be presented from the preset presentation time of the entire slide show.
  • the presentation method determination unit 202 determines the presentation time of the subsequent still image independently of the presentation time of the previous still image. For example, the presentation method determination unit 202 may set the presentation time of the subsequent still image to the initial value Ts. The presentation method determination unit 202 may set the presentation time of the subsequent still image to a random value within a specified range. A case where a still image obtained by photographing a landscape is reproduced will be described with reference to FIG.
  • the relevancy determination unit 204 includes a partial area that is common among target areas determined to be the same among target areas detected from all still images included in a slide show, and a target area between partial still images. Is determined based on the positional relationship between and.
  • the presentation method determination unit 202 calculates the presentation time of the next still image by multiplying the presentation time of a certain still image by a specified parameter.
  • the relevancy determination unit 204 sets the presentation time of the first still image 801 to the initial value Ts.
  • Still images 801 and 802 and 802 and 803 have a partial relationship, and still images 803 and 804 have no partial relationship.
  • the presentation time of the first still image 801 is the initial value Ts
  • the relevance flag of the still images 801 and 802 is 1, so the presentation time of the still image 802 is a ⁇ Ts.
  • the presentation time of the still image 803 is a 2 Ts.
  • the presentation method determination unit 202 returns the presentation time of the still image 804 to the initial value and sets it to Ts.
  • the parameter a is set to a value between 0 and 1 and the smaller the area of the matching partial region is, the still image 801 presented for the first time for the landscape is presented longer.
  • the other parts are presented at a presentation time corresponding to the amount of information overlapping with the previously presented image.
  • the information processing apparatus 200 can generate a video in which the presentation time of consecutive images changes even if the consecutive images include the same target.
  • the presentation method determination unit 202 determines an effect, a BGM, and a jingle to be inserted between a pair of still images based on a target partial relationship included in a pair of still images. For example, when the targets included in the continuous still image pairs are in a partial relationship, the presentation method determination unit 202 uses special effects (dissolve and fade) that are registered in advance as effects with little visual change when switching still images. Etc.).
  • the presentation method determination unit 202 displays a special effect (page turning, wipe, etc.) registered in advance as an effect having a large visual change when switching still images. DVE) is inserted.
  • the presentation method determination unit 202 plays the same BGM during presentation of still image pairs.
  • the presentation method determination unit 202 stops the BGM or switches to a different BGM when switching the still images.
  • the presentation method determination unit 202 may insert jingles between images that do not have a magnitude relationship.
  • the relevancy determination unit 204 may determine relevance depending on whether or not the objects represented in the two still images included in the still image pair are of the same type.
  • the relationship 4 is the relationship determined in this way. “The objects are of the same type” means that main objects appearing in a pair of still images that are consecutive in the slide show are objects of the same type.
  • the relevancy determination unit 204 determines that the relevance flag for relevance 4 is related to relevance 4 when the target area in a certain still image and the target area in the next still image have the same kind of relationship. 1 is set in the relevance flag. When the target area in a certain still image and the target area in the next still image are different from each other, the relevancy determination unit 204 sets 0 in the relevance flag for the relevance 4. Discrimination of the homogeneity of the object can be realized by a method based on machine learning based on the image data (registered data) of the object belonging to each type for which homogeneity is desired. First, the relevancy determination unit 204 extracts target image feature amounts belonging to various types from registered data.
  • the relevancy determination unit 204 may use a global feature such as a color histogram or an edge histogram as the image feature amount.
  • the relevancy determination unit 204 may use local feature amounts such as HoG (Histograms of Oriented Gradients) and SIFT as the image feature amounts.
  • the relevancy determination unit 204 may perform learning by using a global feature, such as SVM (Signal Value Decomposition), a neural network, or a GMM (Gaussian Mixture Model). Alternatively, the relevancy determination unit 204 may perform learning after converting the feature amount space from the local feature amount like BoW (Bag of Words).
  • BoW BoW
  • the relevance determining unit 204 determines between the image feature amount of each target area and various types of models obtained as a result of learning. Each seeks similarity. Then, the relevancy determination unit 204 determines that the target region is the closest model type that has obtained a similarity equal to or greater than a specified value. The relevancy determination unit 204 determines that the target areas determined to be the same type are the same type. The relevancy determination unit 204 may determine the homogeneity by a method other than the above. When three images including the same type of target are consecutive, the presentation method determination unit 202 assumes that there is no change in relevance and, for example, gradually shortens the presentation time at the same time interval.
  • the presentation method determination unit 202 controls the presentation method based on the following rules. [Rules according to target homogeneity] (4-1) Rules regarding presentation time
  • the presentation method determination unit 202 determines the presentation time of a still image pair based on the homogeneity of objects included in successive still image pairs. For example, the presentation method determination unit 202 sets the presentation time of the still image presented first among the still image groups including the same type of target to the initial value Ts. Then, the presentation method determination unit 202 determines the presentation time of the subsequent still image based on Ts. In addition, the presentation method determination unit 202 may set Tp as the presentation time of a still image with high visibility among a group of still images including the same type of target.
  • the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Tp. In addition, the presentation method determination unit 202 may set the presentation time of the next still image in which the presentation time of the still image is equal to or less than Tq among the still image group including the same type of target as the initial value Ts. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Ts. In addition, the presentation method determination unit 202 may set the presentation time of the last presented image among the still image groups including the same type of target to Ts. The presentation method determination unit 202 may calculate the values of Ts and Tp according to the number of images to be presented from the preset presentation time of the entire slide show.
  • the presentation method determination unit 202 determines the presentation time of the subsequent still image independently of the presentation time of the previous still image.
  • the presentation method determination unit 202 may set the presentation time of subsequent still images to, for example, the initial value Ts.
  • the presentation method determination unit 202 may set the presentation time of the subsequent still image to a random value within a specified range.
  • a still image obtained by photographing a flower is reproduced will be described with reference to FIG.
  • the relevancy determination unit 204 determines the homogeneity between consecutive still images by a method based on machine learning.
  • the presentation method determination unit 202 calculates the presentation time of the next still image by multiplying the presentation time of a certain still image by the parameter for the relevance flag.
  • the presentation method determination unit 202 sets the presentation time of the first still image 901 to the initial value Ts.
  • the still images 901 and 902 and the still images 902 and 903 have the same type of relationship, and the still images 903 and 904 have a different type of relationship.
  • the presentation time of the still image 902 is a ⁇ Ts.
  • the presentation time of the still image 903 is a 2 Ts.
  • the presentation method determination unit 202 returns the presentation time of the still image 904 to the initial value and sets it to Ts.
  • the parameter a is set between 0 and 1
  • a still image 901 presented for the first time among the still images including plants is presented for a long time.
  • the subsequent still image is presented in a presentation time that is shorter as the distance from 901 increases.
  • the information processing apparatus 200 can generate a video in which the presentation time of consecutive images changes even if the consecutive images include the same target.
  • this embodiment has an effect that it is possible to generate a slide show with a tempo that does not bore the viewer (this embodiment reproduces images of a plurality of flowers taken in a flower field in order of the same kind of subject in order. Can express that there were many subjects of this type).
  • (4-2) Rules regarding effects, BGM, and jingles The presentation method determination unit 202 determines an effect, a BGM, and a jingle to be inserted between a pair of still images based on the homogeneity of objects included in successive still image pairs. For example, when the target included in the pair of still images is the same type, the presentation method determination unit 202 uses special effects (such as dissolves and fades) registered in advance as effects with little visual change when switching still images.
  • the presentation method determination unit 202 when the target included in the continuous still image pair is different, special effects (such as page turning and wipe) registered in advance as effects having a large visual change when switching still images. DVE) is inserted.
  • special effects such as page turning and wipe
  • DVE digital image
  • the presentation method determination unit 202 plays the same BGM during presentation of the still image pairs.
  • the presentation method determination unit 202 stops BGM or switches to a different BGM when switching still images.
  • the presentation method determination unit 202 may insert jingles between different types of still images. Thereby, when the object contained in the continuous still image pair is the same type, the still image pair is smoothly connected without any change in image or sound.
  • the relevancy determination unit 204 may determine the relevance based on the commonality of the shooting locations of two still images included in the still image pair.
  • the relationship 5 is the relationship determined in this way. “The shooting locations are the same” means that the locations where the still image pairs that were consecutive in the slide show were shot are the same.
  • the relevancy determination unit 204 sets 1 in the relevance flag for relevance 5 when the shooting location of a certain still image is the same as the shooting location of the next still image.
  • the relevancy determination unit 204 sets 0 for the relevance flag for relevance 5.
  • the relevance determination unit 204 can determine the identity of the shooting location based on the similarity of an area (background area) other than the target area in the still image. For example, the relationship determination unit 204 may separate the target area and the background area from the still image. Then, the relevancy determination unit 204 may determine the same shooting location when the image feature values extracted from the background region are similar.
  • the relevancy determination unit 204 may determine the identity of the shooting location by a method other than the above.
  • the relevancy determination unit 204 may determine the similarity of the shooting location and the similarity of the background between still images consecutive in the slide show.
  • the relevancy determination unit 204 may determine the identity of the shooting location based on the identity of the background areas in all the still images included in the slide show.
  • the relevancy determination unit 204 may determine the identity of the shooting location by combining the shooting location as meta information and the GPS as sensor information in addition to the image information. For example, when three images taken at the same shooting location are consecutive, the presentation method determination unit 202 assumes that there is no change in relevance and, for example, gradually reduces the presentation time at the same time interval. Specifically, the presentation method determination unit 202 controls the presentation method based on the following rules.
  • the presentation method determination unit 202 determines the presentation time of still image pairs based on the identity of the shooting locations of consecutive still image pairs. For example, the presentation method determination unit 202 sets the presentation time of a still image presented first among the still images captured at the same place as the initial value Ts. Then, the presentation method determination unit 202 determines the presentation time of the subsequent still image based on Ts. In addition, the presentation method determination unit 202 may set Tp as the presentation time of a still image with high visibility among the still images captured at the same place. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Tp.
  • the presentation method determination unit 202 may set the presentation time of the next still image in which the presentation time of the still image is equal to or less than Tq in the still image group captured at the same place as the initial value Ts. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Ts. In addition, the presentation method determination unit 202 may set the presentation time of the last presented image in the group of still images captured at the same place as Ts. The presentation method determination unit 202 may calculate values of Ts and Tp according to the number of images to be presented from preset presentation time of the entire slide show. When successive still image pairs are taken at different locations, the presentation method determination unit 202 determines the presentation time of the subsequent still image independently of the presentation time of the previous still image.
  • the presentation method determination unit 202 may set the presentation time of the subsequent still image to the initial value Ts.
  • the presentation method determination unit 202 may set the presentation time of the subsequent still image to a random value within a specified range.
  • (5-2) Rules regarding effects, BGM, and jingles The presentation method determination unit 202 determines an effect, a BGM, and a jingle to be inserted between a pair of still images based on the identity of the shooting locations of consecutive still image pairs. For example, when consecutive still image pairs are shot at the same place, the presentation method determination unit 202 uses special effects (such as dissolves and fades) registered in advance as effects with little visual change when switching still images. Insert.
  • the presentation method determination unit 202 displays special effects (DVE such as page turning and wipe) that are registered in advance as effects having a large visual change when switching still images. ) Is inserted. Further, for example, when consecutive still image pairs are photographed at the same place, the presentation method determination unit 202 plays the same BGM during presentation of the still image pairs. When consecutive still image pairs are captured at different locations, the presentation method determination unit 202 stops BGM or switches to a different BGM when switching still images. In addition, the presentation method determination unit 202 may insert a jingle between still images taken at different places. Thereby, when continuous still image pairs are taken at the same place, the still image pairs are smoothly connected without any change in image or sound.
  • DVE page turning and wipe
  • the relevancy determination unit 204 may determine the relevance based on the commonality of the shooting time zones of two still images included in the still image pair.
  • the relationship 6 is the relationship determined in this way. “The shooting time zone is the same” means that the time zone in which a pair of still images is taken in the slide show is the same.
  • the relevancy determination unit 204 sets 1 in the relevance flag for relevance 6 when the shooting time zone of a certain still image is the same as the shooting time zone of the next still image.
  • the relevancy determination unit 204 sets a relevance flag for relevance 6 to 0 when the shooting time zone of a certain still image is different from the shooting time zone of the next still image.
  • the relevancy determination unit 204 can determine the identity of the shooting time period based on the color information of the background area in the still image. For example, the relevance determination unit 204 divides a day into a plurality of time zones, and holds the statistics of the color histogram of sunlight in each time zone.
  • the relevancy determination unit 204 determines that the still image is captured during that time zone.
  • the relevancy determination unit 204 estimates the shooting time period of each still image.
  • the relevance determination unit 204 determines that the shooting time zones are the same when the estimated time is the same.
  • the relevancy determination unit 204 may determine the identity of the shooting time period by a method other than the above.
  • the relevance determination unit 204 may determine the identity of the shooting time zones based on the similarity of the shooting time zones between consecutive still images in the slide show. Alternatively, the relevancy determination unit 204 may determine the identity of the shooting time zone based on the identity of the shooting time zones in all the still images included in the slide show.
  • the relevancy determination unit 204 may determine the identity of the shooting time period in combination with the shooting time that is meta information in addition to the image information.
  • the presentation method determination unit 202 assumes that there is no change in relevance when three images taken in the same shooting time period are consecutive, and for example, gradually reduces the presentation time at the same time interval.
  • the presentation method determination unit 202 controls the presentation method based on the following rules. [Rules according to the identity of the shooting period] (6-1) Rules regarding presentation time
  • the presentation method determination unit 202 determines the presentation time of still image pairs based on the identity of the shooting time zones of consecutive still image pairs. For example, the presentation method determination unit 202 sets the presentation time of the first still image to be presented as the initial value Ts in the group of still images taken in the same time zone.
  • the presentation method determination unit 202 determines the presentation time of the subsequent still image based on Ts.
  • the presentation method determination unit 202 may set the presentation time of a still image with high visibility among the still image groups taken in the same time zone as Tp.
  • the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Tp.
  • the presentation method determination unit 202 may set the presentation time of the next still image in which the presentation time of the still image is equal to or less than Tq among the still image groups photographed in the same time zone as the initial value Ts.
  • the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Ts.
  • the presentation method determination unit 202 may set the presentation time of the image to be presented last among the still image groups captured in the same time zone as Ts.
  • the presentation method determination unit 202 may calculate the values of Ts and Tp according to the number of images to be presented from the preset presentation time of the entire slide show. When successive still image pairs are photographed at different time periods, the presentation method determination unit 202 determines the presentation time of the subsequent still image independently of the presentation time of the previous still image.
  • the presentation method determination unit 202 may set the presentation time of subsequent still images to, for example, the initial value Ts.
  • the presentation method determination unit 202 may set the presentation time of the subsequent still image to a random value within a specified range.
  • the presentation method determination unit 202 determines an effect, a BGM, and a jingle to be inserted between a pair of still images based on the identity of the shooting time zones of consecutive still image pairs. For example, when a pair of still images is shot in the same time zone, the presentation method determination unit 202 can use special effects (dissolve or fade) that are registered in advance as effects with little visual change when switching still images. Etc.). The presentation method determination unit 202, when a pair of still images is shot at different time zones, special effects (such as page turning and wipe) registered in advance as effects having a large visual change when switching still images. DVE) is inserted.
  • special effects such as page turning and wipe
  • the presentation method determination unit 202 plays the same BGM during presentation of still image pairs.
  • the presentation method determination unit 202 stops BGM or switches to a different BGM when switching still images.
  • the presentation method determination unit 202 may insert jingles between still images in different time zones.
  • the slide show generation unit 203 generates a slide show based on the presentation method information input from the presentation method determination unit 202 and the image information input from the image input unit 210.
  • the relevancy determination unit 204 may group all target areas detected from all still images included in the slide show based on similarity. Then, the relevancy determination unit 204 may determine that the target areas are the same when the target areas detected from the adjacent still image pairs belong to the same group.
  • the detection method of the target area is divided into a detection method for detecting a specific target registered in advance and a detection method for detecting a general target that is not registered.
  • the relevance determination unit 204 may use the registered image data of each target as a template. Then, the relevancy determination unit 204 may scan the input image with templates converted into various resolutions. Then, the relevancy determination unit 204 may detect a region having a small difference in pixel values at the same position as the template as a corresponding target region.
  • the relevancy determination unit 204 determines the identity, the magnitude relationship, and the partial relationship as the relevance type.
  • the presentation rule determination method is the same as in the first embodiment.
  • the image input unit 210 inputs image information of all still images to the relevancy determination unit 204 (step S1001).
  • the relevancy determination unit 204 extracts image feature amounts from all still images.
  • the relevancy determination unit 204 groups the still images obtained by photographing the same object based on the similarity of the image feature amounts (step S1003). In the example of FIG.
  • still images 501, 502, 503, and 504 are classified into group A, and still images 505, 506, 507, 508, 509, and 510 are classified into group B, 511, 512, and 513.
  • the relevancy determination unit 204 determines the magnitude relationship and the partial relationship between still images belonging to the same group (S1005).
  • the relevancy determination unit 204 extracts local feature points such as SIFT from all still images and takes corresponding points with each other. Then, for group A, it can be seen that still images 502, 503, and 504 are included in the still image 501. It can be seen that the common areas of the still images 502, 503, and 504 are small.
  • the relevancy determination unit 204 determines that there is a magnitude relationship between the still image 501 and the still image 502. Then, the relationship comparison unit 201 determines that there is a partial relationship between the still image 502 and the still image 503, and the still image 503 and the still image 504. The relevancy determination unit 204 compares local feature amounts in the same manner, so that it can be understood that the still images 506, 507, 508, 509, and 510 are included in 505 for the group B. Further, it can be seen that the still images 506 and 507 are included in the still image 513.
  • the presentation method determination unit 202 determines the presentation time length 1101 and the effect 1102 as shown in FIG. 11 (S1007).
  • the slide show generating unit 203 generates a slide show using the determined presentation method (S1009).
  • the system or apparatus which combined the separate characteristic contained in each embodiment how was included in the category of this invention.
  • the present invention may be applied to a system composed of a plurality of devices, or may be applied to a single device.
  • the present invention can also be applied to a case where an information processing program that implements the functions of the embodiments is supplied directly or remotely to a system or apparatus.
  • each of the information processing apparatus 100 and the information processing apparatus 200 can be realized by a computer and a program for controlling the computer, dedicated hardware, or a combination of the computer and the program for controlling the computer and dedicated hardware.
  • the relevance comparison unit 101, the presentation method determination unit 102, the slide show generation unit 103, the relevance comparison unit 201, the presentation method determination unit 202, the slide show generation unit 203, and the relevance determination unit 204 are, for example, from a recording medium that stores a program.

Abstract

[Problem] To present still images in accordance with changes in the degrees of association therebetween. [Solution] This information processing device is characterized by the provision of a comparison means, a presentation-method determination means, and a generation means. Given a still-image collection containing at least three still images, the comparison means compares a first degree of association with a second degree of association. Said first degree of association indicates the degree of association between the still images constituting a first still-image pair that is part of the still-image collection, and the second degree of association indicates the degree of association between the still images constituting a second still-image pair. On the basis of the change from the first degree of association to the second degree of association, the presentation-method determination means uses first information to determine second information, where said first information specifies the presentation method for the first still-image pair and said second information specifies the presentation method for the second still-image pair. On the basis of the presentation method determined by the presentation-method determination means, the generation means generates a slideshow that contains the first still-image pair and the second still-image pair.

Description

情報処理装置、情報処理方法および情報処理プログラムInformation processing apparatus, information processing method, and information processing program
 本発明は、情報処理装置、情報処理方法および情報処理プログラムに関し、特に静止画像の提示方法を決定する情報処理装置、情報処理方法および情報処理プログラムに関する。 The present invention relates to an information processing device, an information processing method, and an information processing program, and more particularly to an information processing device, an information processing method, and an information processing program for determining a still image presentation method.
 上記の技術分野において、特許文献1に記載の技術が知られている。この技術は、各静止画像の内容に応じた提示時間を実現するスライドショー生成技術である。具体的には、静止画像に含まれる顔の個数によって、その静止画像の提示時間を制御しようとするものである。
 また、特許文献2の技術は、スライドショー生成にあたり、静止画像の撮影時刻によって提示時間を制御する。特許文献2の技術は、スライドショーの中で連続する静止画像の提示時間を、静止画像の撮影時刻の差によって決定する。
 また、特許文献3の技術は、画像間の関連度を抽出して、この関連度に基づいて視覚効果を与えながらスライドショーの再生を行う(段落0024)。この技術は、比較すべき2枚の画像を取得し、画像ごとの関連度を抽出する(段落0063)。この技術は、関連度が弱い場合には関連度の弱い画像用の、関連度が強い場合には関連度の強い画像用の、視覚効果を付与する(段落0074)。
In the above technical field, the technique described in Patent Document 1 is known. This technique is a slide show generation technique that realizes a presentation time according to the content of each still image. Specifically, the presentation time of the still image is to be controlled by the number of faces included in the still image.
The technique of Patent Document 2 controls the presentation time according to the still image shooting time when generating a slide show. In the technique of Patent Document 2, the presentation time of continuous still images in a slide show is determined based on the difference in shooting time of still images.
The technique of Patent Document 3 extracts the degree of association between images and reproduces a slide show while giving a visual effect based on the degree of association (paragraph 0024). This technique acquires two images to be compared and extracts the degree of association for each image (paragraph 0063). This technique provides a visual effect for images with low relevance when the relevance is weak and for images with high relevance when the relevance is strong (paragraph 0074).
特許第4593314号公報Japanese Patent No. 4593314 特開2006−41586号公報JP 2006-41586 A 特開2006−140559号公報JP 2006-140559 A
 特許文献1、2、3の技術は、少なくとも1つの静止画像の情報を用いて、静止画像の提示方法を決定している。したがって、これらの技術は、2つの連続する静止画像の内容の変化を、提示方法の変化によって表している。しかし、これらの技術は、連続する2つの静止画像の関連性が変わらない場合は、提示方法を変化させることはない。
 本発明の目的は、上述の課題を解決する情報処理装置を提供することにある。
The techniques of Patent Literatures 1, 2, and 3 determine a still image presentation method using information of at least one still image. Therefore, these techniques represent changes in the contents of two consecutive still images by changes in the presentation method. However, these techniques do not change the presentation method when the relationship between two consecutive still images does not change.
The objective of this invention is providing the information processing apparatus which solves the above-mentioned subject.
 上記目的を達成するため、本発明に係る情報処理装置は、少なくとも3つの静止画像を含む静止画像群において、前記静止画像群に含まれる第1静止画像対の静止画像同士の第1関連性と第2静止画像対の静止画像同士の第2関連性とを比較する比較手段と、前記第1関連性から第2関連性への変化に基づいて、前記第1静止画像対の提示方法を特定する第1情報から、前記第2静止画像対の提示方法を特定する第2情報を決定する提示方法決定手段と、前記提示方法決定手段で決定した提示方法に基づいて、前記第1静止画像対および前記第2静止画像対を含むスライドショーを生成する生成手段と、を備えたことを特徴とする。
 上記目的を達成するため、本発明に係る情報処理方法は、少なくとも3つの静止画像を含む静止画像群において、前記静止画像群に含まれる第1静止画像対の静止画像同士の第1関連性と第2静止画像対の静止画像同士の第2関連性とを比較し、前記第1関連性から第2関連性への変化に基づいて、前記第1静止画像対の提示方法を特定する第1情報から、前記第2静止画像対の提示方法を特定する第2情報を決定し、前記決定ステップで決定した提示方法に基づいて、前記第1静止画像対および前記第2静止画像対を含むスライドショーを生成することを特徴とする。
 上記目的を達成するため、本発明に係る情報処理プログラムは、コンピュータを、少なくとも3つの静止画像を含む静止画像群において、前記静止画像群に含まれる第1静止画像対の静止画像同士の第1関連性と第2静止画像対の静止画像同士の第2関連性とを比較する比較手段と、前記第1関連性から第2関連性への変化に基づいて、前記第1静止画像対の提示方法を特定する第1情報から、前記第2静止画像対の提示方法を特定する第2情報を決定する提示方法決定手段と、前記決定ステップで決定した提示方法に基づいて、前記第1静止画像対および前記第2静止画像対を含むスライドショーを生成する生成手段として動作させることを特徴とする。
In order to achieve the above object, an information processing apparatus according to the present invention, in a still image group including at least three still images, has a first relationship between still images of a first still image pair included in the still image group. The comparison means for comparing the second relevance between the still images of the second still image pair, and the presentation method of the first still image pair are specified based on the change from the first relevance to the second relevance Based on the presentation method determined by the presentation method determination means, the presentation method determination means for determining the second information specifying the presentation method of the second still image pair from the first information And generating means for generating a slide show including the second still image pair.
In order to achieve the above object, an information processing method according to the present invention includes, in a still image group including at least three still images, a first association between still images of a first still image pair included in the still image group. The first still image pair is compared with the second relationship between the still images of the second still image pair, and the presentation method of the first still image pair is specified based on the change from the first relationship to the second relationship. Slide information including the first still image pair and the second still image pair based on the presentation method determined in the determining step, determining second information specifying a presentation method of the second still image pair from the information; Is generated.
In order to achieve the above object, an information processing program according to the present invention causes a computer to select a first still image of a pair of first still images included in a still image group in a still image group including at least three still images. Comparison means for comparing the relevance and the second relevance between the still images of the second still image pair, and presentation of the first still image pair based on a change from the first relevance to the second relevance Based on the first information specifying the method, the presentation method determining means for determining the second information specifying the presentation method of the second still image pair, and the presentation method determined in the determining step, the first still image It is operated as a generating means for generating a slide show including a pair and the second still image pair.
 本発明によれば、静止画像間の関連性の変化に応じて静止画像を提示することができる。 According to the present invention, it is possible to present a still image according to a change in relevance between still images.
本発明の第1実施形態に係る情報処理装置の構成を示すブロック図である。It is a block diagram which shows the structure of the information processing apparatus which concerns on 1st Embodiment of this invention. 本発明の第2実施形態に係る情報処理装置の構成を示すブロック図である。It is a block diagram which shows the structure of the information processing apparatus which concerns on 2nd Embodiment of this invention. 本発明の第2実施形態に係る情報処理装置のスライドショー生成を説明する図である。It is a figure explaining the slide show production | generation of the information processing apparatus which concerns on 2nd Embodiment of this invention. 本発明の第2実施形態に係る情報処理装置の処理の流れを示すフローチャートである。It is a flowchart which shows the flow of a process of the information processing apparatus which concerns on 2nd Embodiment of this invention. 本発明の第2実施形態に係る情報処理装置のスライドショー生成を説明する図である。It is a figure explaining the slide show production | generation of the information processing apparatus which concerns on 2nd Embodiment of this invention. 本発明の第2実施形態に係る情報処理装置のスライドショー生成を説明する図である。It is a figure explaining the slide show production | generation of the information processing apparatus which concerns on 2nd Embodiment of this invention. 本発明の第3実施形態に係る情報処理装置のスライドショー生成を説明する図である。It is a figure explaining the slide show production | generation of the information processing apparatus which concerns on 3rd Embodiment of this invention. 本発明の第3実施形態に係る情報処理装置のスライドショー生成を説明する図である。It is a figure explaining the slide show production | generation of the information processing apparatus which concerns on 3rd Embodiment of this invention. 本発明の第3実施形態に係る情報処理装置のスライドショー生成を説明する図である。It is a figure explaining the slide show production | generation of the information processing apparatus which concerns on 3rd Embodiment of this invention. 本発明の第4実施形態に係る情報処理装置の処理の流れを示すフローチャートである。It is a flowchart which shows the flow of a process of the information processing apparatus which concerns on 4th Embodiment of this invention. 本発明の第4実施形態に係る情報処理装置のスライドショー生成を説明する図である。It is a figure explaining the slide show production | generation of the information processing apparatus which concerns on 4th Embodiment of this invention.
 以下に、図面を参照して、本発明の実施の形態について例示的に詳しく説明する。ただし、以下の実施の形態に記載されている構成要素はあくまで例示であり、本発明の技術範囲をそれらのみに限定する趣旨のものではない。
 [第1実施形態]
 本発明の第1実施形態としての情報処理装置100について、図1を用いて説明する。情報処理装置100は、少なくとも3つの静止画像を含むスライドショーを生成する装置である。
 情報処理装置100は、関連性比較部(比較部)101と提示方法決定部102とスライドショー生成部(生成部)103とを含む。
 関連性比較部101は、少なくとも3つの静止画像を含む静止画像群において、前記静止画像群に含まれる第1静止画像対の静止画像同士の第1関連性と、第2静止画像対の静止画像同士の第2関連性とを比較する。
 提示方法決定部102は、第1関連性から第2関連性への変化に基づいて、第1静止画像対の提示方法を特定する第1情報から、第2静止画像対の提示方法を特定する第2情報を決定する。
 スライドショー生成部103は、提示方法決定部102で決定した提示方法に基づいて、第1静止画像対および第2静止画像対を含むスライドショーを生成する。
 以上の構成によれば、静止画像間の関連性の変化に応じて静止画像を提示するスライドショーを生成することができる。
 [第2実施形態]
 次に本発明の第2実施形態に係る情報処理装置200について、図2を用いて説明する。図2は、本実施形態に係る情報処理装置200の構成を説明するためのブロック図である。
 (構成)
 情報処理装置200は、デジタルカメラやデジタルビデオカメラなどの撮像装置250から画像情報を入力する画像入力部210と、入力した画像の関連性を比較する関連性比較部(比較部)201と、関連性を決定する関連性決定部204とを含む。また、情報処理装置200は、スライドショーに含まれる各画像の表示長さ、画像切替時のエフェクト、表示中のBGM(Background Music)、切替時のジングルなどを決定する提示方法決定部202を含む。さらに情報処理装置200は、提示方法決定部202が決定した提示方法によって入力した静止画を組み合わせてスライドショーを生成するスライドショー生成部(生成部)203を含む。情報処理装置200は、また、静止したスライドショーを表示するためのディスプレイ260に接続されている。
 画像入力部210が入力する画像情報は、静止画像を識別する画像ID(identifier)、スライドショー内での提示順位、静止画像の画素情報を含む。また、画像情報は、静止画像内に映っている被写体、撮影場所、撮影時刻を記述したメタ情報や、GPS(Global Positioning System)等のセンサ情報を含んでもよい。
 関連性決定部204は、第1関連性を、第1静止画像対に含まれる2つの静止画像に表わされた対象の共通性によって決定する。関連性決定部204は、第2関連性を、第2静止画像対に含まれる2つの静止画像に表わされた対象の共通性によって決定する。第1静止画像対および第2静止画像対は、それぞれ連続する静止画像の対である。
 関連性決定部204は、画像関連性情報として、静止画像の画像IDと関連性フラグを関連性比較部201に入力する。関連性決定部204は、画像関連性情報として、上記に加えて画素情報を入力してもよい。関連性フラグは、あらかじめ規定された関連性種別のうち、現在の静止画像とその後に提示される静止画像との間に存在する関連性種別を表すデータである。あるいは、関連性フラグは、これらの画像の間にいずれの関連性種別も存在しない(関連性が無い)ことを示すデータである。ある静止画像とその後の静止画像との間に存在する全関連性種別の関連性フラグにはフラグ1が設定されている。存在しない関連種別の関連性フラグにはフラグ0を設定されている。あるいは、関連性フラグには、関連性種別に応じて意味を持つ任意の数値が設定されていてもよい。
 関連性比較部201は、静止画像群に含まれる第1静止画像対の静止画像同士の第1関連性と、第2静止画像対の静止画像同士の第2関連性とを比較する。具体的には、関連性比較部201は、スライドショー内で連続する静止画像対の間の関連性を判定する。より具体的には、静止画像群は連続する第1、第2、第3静止画像を含む。第1静止画像対は、第1静止画像と第2静止画像の対である。第2静止画像対は、第2静止画像と第3静止画像の対である。
 提示方法決定部202は、関連性比較部201から入力された画像関連性情報と、画像入力部210から入力される画像情報と、あらかじめ登録された提示ルールをもとに、各静止画像の提示方法を決定する。そして、提示方法決定部202は、提示方法情報をスライドショー生成部203に入力する。
 スライドショー生成部203は、決定された提示方法に基づいて、静止画像を組み合わせてスライドショーを生成する。
 (関連性決定方法)
 以下に、関連性決定部204における関連性の決定方法について説明する。関連性種別としては様々なものが考えられるが、本実施形態では、一例として、下記の対象の同一性について説明する。
 関連性1.対象の同一性
 関連性決定部204は、関連性を、静止画像対に含まれる2つの静止画像に表わされた対象の共通性(同一性)によって決定する。関連性1は、このように決定された関連性である。同一性は、静止画像群に含まれる複数の静止画像から導き出される特徴量の共通性によって判断できる。
 すなわち、関連性1は、スライドショー内で連続する静止画像対に映っている対象が同一であることを示す関連性である。関連性決定部204は、関連性1についての関連性フラグには、同一の場合に1を、同一でない場合には0を設定する。関連性決定部204は、対象の同一性を、静止画像から検出された対象領域画像の類似性をもとに判定することが可能である。対象領域は、木や建物のような静止物体、あるいは、人間や動物のような動物体のように、ある画像的なパターンを持った対象の静止画像上での領域である。対象領域は、静止画像中の部分領域であってもよい。対象領域は、静止画像全体であってもよい。
 関連性決定部204は、スライドショー内で連続する静止画像対から対象領域を検出する。そして、関連性決定部204は、対象領域の類似性で対象の同一性を判定する。あるいは、関連性決定部204は、スライドショーに含まれる全静止画像から検出した対象領域の間の類似性をもとに同一性を判定してもよい。この場合、関連性決定部204は、検出された全対象領域を類似性に基づいてグループ化する。そして、関連性決定部204は、隣接する静止画像対から検出された対象領域が同じグループに属する場合には同一の対象と判定する。
 対象領域の検出手法は、あらかじめ登録された特定対象を検出する場合の検出手法と、登録されていない一般対象を検出する場合の検出手法とに分けられる。関連性決定部204は、特定対象を検出する場合は、登録された各対象の画像データをテンプレートとすればよい。そして、関連性決定部204は、様々な解像度に変換したテンプレートで入力画像を走査すればよい。関連性決定部204は、テンプレートと同じ位置の画素値の差分が小さい領域を、対応する対象領域として検出してもよい。
 あるいは、関連性決定部204は、入力画像の各部分領域から色・テクスチャ・形状を表現する画像特徴量を抽出してもよい。そして、関連性決定部204は、登録された各対象の画像特徴量と類似した画像特徴量をもつ部分領域を、対応する対象領域としてもよい。また、特定対象が人物の場合、顔全体から得られる情報を利用する手法がある。例えば、様々な顔の映っている画像をテンプレートとして記憶し、入力画像とテンプレートの差分がある閾値以下のとき顔が入力画像中に存在すると判定する手法が挙げられる。また、肌色などの色情報や、エッジの方向や密度を組み合わせたモデルをあらかじめ記憶しておき、入力フレームからモデルに類似した領域が検出された場合に顔が存在すると判定する手法が考えられる。また、顔(頭部)の輪郭を楕円、目や口を細長の形状をしていることを利用して作成したテンプレートを用いて顔検出を行う手法もある。さらに、頬や額の部分は輝度が高く、目や口の部分の輝度は低いという輝度分布の特性を利用して顔検出手法や、顔の対称性や肌色領域と位置を利用して顔検出を行う手法などの手法が挙げられる。
 また、大量の人物顔と非顔の学習サンプルから得られた特徴量分布を統計的に学習し、入力画像から得られる特徴量が顔と非顔のどちらの分布に属するかを判定する手法がある。このような手法として、ニューラルネットやサポートベクターマシン、AdaBoost法などを利用した手法が挙げられる。また一般対象を検出する場合は、例えばNormalized Cutや、Saliency Mapや、Depth of Field(DoF)などを用いてもよい。Normalized Cutは、画像を複数の領域に分割する手法である。詳しくは、Jianbo Shi and Jitendra Malik、“Normalized Cuts and Image Segmentation”、IEEETransactions on Pattern Analysis and Machine Intelligence、vol.22、No.8、August2000に開示がある。関連性決定部204は、Normalized Cutにより分割された領域のうち、画面中央に位置する領域を対象領域として検出してもよい。
 また、Saliency Mapは画像中の物体領域を視覚注意から算出する方法である。SaliencyMapについては、L.Itti、C.Koch and E.Niebur、“A Model of Saliency−based VisualAttention for Rapid Scene Analysis、”IEEE Trans.Pattern Analysis and Machine Intelligence、Vol.20、No.11、pp.1254−1259、1998に開示がある。関連性決定部204は、Saliency Mapにより高い重要度が算出された領域を対象領域として検出してもよい。また、Dofは、被写界深度内に存在する対象のエッジにはボケがなく、被写界深度外のエッジにボケが発生する特性に基づく手法である。詳しくは、3Du−Ming Tsai、Hu−Jong Wang、“Segmenting focused objects in complex visual images”、Pattern Recognition Letters、Vol.19、pp.929 940、1998.に開示がある。関連性決定部204は、エッジの太さをもとにボケ量を算出し、ボケの少ないエッジを結合し、焦点が合っている領域を対象領域として検出してもよい。
 関連性決定部204は、静止画像中の位置または視認性の高さ(照明条件、向き、角度、画面上での位置、他の対象による隠れ、ボケ、(人物の場合には)表情、等に基づく映りの良さを示す評価値)または複数画像での出現頻度をもとに、静止画像につき1つの対象領域を検出してもよい。また、関連性決定部204は、検出された複数の対象領域を組み合わせて1つの対象領域としてもよい。関連性決定部204は、画像特徴量を抽出し、画像特徴量の差が小さいほど高い値を算出する尺度をもとに、対象領域間の類似度を算出してもよい。関連性決定部204は、対象領域から検出された色、エッジ、テクスチャ等の画像情報をもとに、画像特徴量を算出できる。あるいは、関連性決定部204は、各対象の画像領域からSIFT(Scale−Invaliant Feature Transform)等の局所特徴点を検出してもよい。そして、関連性決定部204は、画像領域間で特徴点を対応づければよい。関連性決定部204は、対応づけられた特徴点の個数が多い、もしくは、対応づけられた特徴点の位置関係が画像間で似ているほど高い値を算出する尺度を用いてもよい。
 (提示方法)
 提示方法決定部202は、関連性の変化が無い場合、第1静止画像対における提示方法の変化と同様に、第2静止画像対における提示方法が変化するように第1提示方法情報から第2提示方法情報を決定する。例えば、提示方法決定部202は、関連性変化が無い場合、第1静止画像対における提示時間の変化と同様に、第2静止画像対において提示時間が変化するように、提示方法情報を決定する。
 提示方法情報は、各静止画像の提示方法を示すデータである。提示方法情報は、画像IDと提示時間を含む。提示方法情報は、上記に加えてエフェクト、BGM、音響ジングル、映像ジングルを含んでいてもよい。提示ルールは、関連性種別に応じた静止画像の提示方法を規定するルールである。提示方法決定部202は、提示ルールとして、連続する静止画像対の各提示時間を規定するパラメータを保持するものとする。また、提示方法決定部202は、提示時間に加えて、静止画像の間に挿入するエフェクトやBGM、ジングル(短い映像や音楽、効果音)に関する制御パラメータを保持してもよい。また、提示ルールは、連続する静止画像対にいずれの関連性種別も存在しない場合の提示方法を規定してもよい。
 本実施形態では、提示方法決定部202は、連続する静止画像対に含まれる対象の同一性をもとに、静止画像対の提示時間を決定する。
 例えば、連続する静止画像対に含まれる対象が同一の場合には、提示方法決定部202は、はじめに提示される静止画像の提示時間を初期値Tsにする。提示方法決定部202は、Tsを基準に後続の静止画像の提示時間を決定する。連続する静止画像対に含まれる対象が同一でない場合には、提示方法決定部202は、前の静止画像の提示時間と独立に後続の提示時間を決定する。提示方法決定部202は、後続の提示時間を、例えば、初期値Tsに設定してもよい。提示方法決定部202は、後続の提示時間を、規定範囲内のランダムな値に設定してもよい。
 なお、提示方法決定部202は、同一の対象を撮影した静止画像群のうち、視認性の高い静止画像の提示時間をTpにしてもよい。そして、提示方法決定部202は、Tpを基準に後続の静止画像の提示時間を決定してもよい。また、提示方法決定部202は、同一の対象を撮影した静止画像群のうち、静止画像の提示時間がTq以下になった次の静止画像の提示時間を初期値Tsにしてもよい。そして、提示方法決定部202は、Tsを基準に後続の静止画像の提示時間を決定してもよい。また、提示方法決定部202は、同一の対象を撮影した静止画像群のうち、最後に提示される静止画像の提示時間を初期値Tsに設定してもよい。提示方法決定部202は、Ts、Tpの値を、あらかじめ設定されたスライドショー全体の提示時間を考慮して、提示する画像数に応じて算出してもよい。
 図3を用いて、様々な向きの人物Aを撮影した静止画像群を再生させる場合について説明する。提示方法決定部202は、ある静止画像の提示時間に対するパラメータaの乗算によって、後続の静止画像の提示時間を算出する。このとき、始めの静止画像301の提示時間が初期値Tsであるとすると、後続の静止画像302~305の提示時間は以下の式(1)で表わされる。
Figure JPOXMLDOC01-appb-M000001
 さらに、正面を向いた静止画像303で視認性の評価値が閾値以上の場合、静止画像303の提示時間がTpであれば、後続の静止画像の提示時間は以下の式(2)で表わされる。
Figure JPOXMLDOC01-appb-M000002
 パラメータaが0から1の間に設定されていると、人物Aを含む静止画像のうち、初めに提示された静止画像301と、人物Aの映りがよい静止画像303は長く提示される。その他のAを含む静止画像は、静止画像303から遠ざかるに従って徐々に短く提示される。これにより、本実施形態には、利用者が、対象が初めて登場した瞬間や映りのよい静止画像の内容を理解することができ、その他の画像は理解した内容とほぼ同様の内容が映っていると理解することができるという効果がある。また、本実施形態の情報処理装置200は、連続する画像が同じ対象を含む画像であっても、連続する画像の提示時間が変化する映像を生成できる。そのため、本実施形態には、視聴者を飽きさせないテンポ感あるスライドショーを生成できるという効果がある。
 提示方法決定部202は、また、連続する静止画像対に含まれる対象の同一性をもとに、静止画像対の間に挿入するエフェクト・BGM・ジングルを決定する。提示方法決定部202は、例えば、連続する静止画像対に含まれる対象が同一の場合には、静止画像の切り替え時に視覚的な変化の少ないエフェクトとしてあらかじめ登録された特殊効果(ディゾルブやフェード等)を挿入する。同一でない場合には、提示方法決定部202は、静止画像の切り替え時に視覚的な変化の大きいエフェクトとしてあらかじめ登録された特殊効果(ページめくり、ワイプ等のDVE(Digital Video Effects))を挿入する。提示方法決定部202は、エフェクト・ジングルの長さを徐々に短く(長く)する。提示方法決定部202は、BGMの音量を徐々に小さく(大きく)する。
 また例えば、連続する静止画像対に含まれる対象が同一の場合には、提示方法決定部202は、静止画像対の提示中同じBGMを流す。連続する静止画像対に含まれる対象が同一でない場合には、提示方法決定部202は、静止画像の切り替え時にBGMを止めるもしくは異なるBGMに切り替える。また、提示方法決定部202は、同一性が存在しない画像間に、ジングルを挿入してもよい。これにより、同一の対象を撮影した静止画像群は、画像や音響の変化がなく滑らかに接続される。そのため、視聴者は、静止画像が互いに変化がなくほぼ同じ内容であることを容易に理解できる。また、同一の対象でない場合には、画像や音響が大きく変化するために、視聴者は、内容に変化があったことに気づき、スライドショーの内容理解に集中することができる。
 (動作)
 次に、図4のフローチャートを参照して、本実施の形態の動作について詳細に説明する。ここでは一例として図5に示す静止画像501~513を昇順に再生するスライドショーを生成する場合を考える。このスライドショーは、建物の中にある温室で、花と人物とを撮影した出来事を伝えるものである。図5の矩形は、関連性決定部204によって各静止画像から検出される対象領域の例である。
 また、提示方法を制御する提示ルールは、対象領域が同一である静止画像対については大小関係もしくは部分関係に基づくルールであり、対象領域が同一でない静止画像対については同種性に基づくルールである。なお、大小関係、部分関係、同種性に基づくルールについては、第3実施形態以降に詳しく説明する。
 まず、画像入力部210は、静止画像501の画像情報を関連性決定部204に入力してステップS403に進む。入力した静止画像が開始画像であれば、ステップS403からステップS401に戻り、画像入力部210は、関連性決定部204に2枚目の静止画像502の画像情報を入力する(ステップS403)。
 次にステップS405において、関連性決定部204は、静止画像501、502から対象領域を検出する。関連性決定部204には、対象としてあらかじめ建造物・草花・人が登録されており、それぞれのモデルが学習されているものとする。そして、関連性決定部204は、静止画像501、502からそれぞれ、建造物の対象領域として実線矩形で囲まれた箇所を検出する。
 関連性決定部204は、対象領域0および対象領域1の画素情報から画像特徴量を抽出し、領域間の類似性をもとに同一性・大小関係・部分関係・同種性を判定する。対象領域0と1は、建造物の種別として検出されているため、同種性ありと判定される。また、静止画像501上の破線矩形の領域が、対象領域1と対象領域0の共通領域として検出される。対象領域1と0は大小関係にあることが判定される。また共通領域以外の領域が対象領域0上には存在しないため、部分関係の関係性はないと判定される。よって、静止画像501における静止画像502との間の関連性フラグは、同一性・大小関係・部分関係・同種性の順に1、−1、0、1となる(ステップS407)。
 提示方法決定部202は、画像関連性情報として、画像IDと関連性フラグをもとに提示方法を決定する。静止画像501と静止画像502の対象領域は同一のため、大小関係もしくは部分関係に基づくルールを適用する。提示方法決定部202は、開始画像である静止画像501の提示時間を初期値Tsにする。提示方法決定部202は、静止画像501、502の大小関係が小・大の関係であるため、静止画像502の提示時間を、a*Tsにする。また、静止画像501、502に大小関係があることから、提示方法決定部202は、静止画像501、502の切り替わりのエフェクトとして視覚的変化の少ないディゾルブを挿入する(ステップS409)。
 スライドショー生成部203は、決定した提示時間・エフェクトで、静止画像501、502を用いてスライドショーを生成する(ステップS411)。上記のステップはすべての静止画像について行われる(ステップS413)。
 図6において、静止画像から検出される対象領域の種別が対象領域601である。各関連性種別に対する関連性フラグが602である。提示方法決定部202によって決定される提示時間長およびエフェクトが、提示時間長603およびエフェクト604である。
 以上、本実施形態は、連続する静止画像が異なる個数の顔画像を含む場合であっても、互いに内容的な意味に関連性がある場合には、関連性があることを視聴者に提示できる。その理由は、提示方法決定部202が、連続する静止画像の提示方法を、互いの関連性に応じて制御するためである。
 また、本実施形態は、連続する静止画像が同じ個数の顔画像を含む場合であっても、互いに内容的な意味に関連性がない場合には、関連性がないことを視聴者に提示できる。
 さらに、本実施形態は、様々な時間間隔で撮影された静止画像群についても、静止画像の間に関連性がある場合には、関連性があることを視聴者が理解しやすい方法で提示できる。
 本実施形態は、一定の時間間隔で撮影された静止画像群についても、互いに内容的な意味に関連性がない場合には、関連性がないことを視聴者が理解しやすい方法で提示できる。
 [第3実施形態]
 第2実施形態に開示した関連性に代えて、あるいは加えて以下の関連性のいずれか1つの変化に応じて提示方法を変更し、以下の提示ルールのいずれか1つを採用してもよい。
 (関連性2. 対象の大小関係)
 関連性決定部204は、関連性を、静止画像対に含まれる2つの静止画像に表わされた対象の大小関係によって決定してもよい。つまり、関連性決定部204は、静止画像対に含まれる2つの静止画像において、あらかじめ登録された特定対象が含まれている領域の大きさの変化によって関連性を決定してもよい。関連性2は、このように決定された関連性である。
 「対象の大小関係にある」とは、スライドショー内で連続する静止画像対に含まれる対象が同一であり、かつ対象領域の面積に規定値以上の差があることである。例えば、対象の周囲を含んだ画像と、対象のみを撮影した画像を組み合わせてスライドショーを生成することで、対象を紹介するケースがある。
 関連性決定部204は、対象の大小関係を、同一と判定された対象領域に共通する部分領域の面積、または共通する部分領域に含まれる特徴点間の距離によって判定できる。関連性決定部204は、例えば、特徴点間の距離が大であるほど対象が大きく撮影されていると判定できる。関連性決定部204は、スライドショー内で連続する静止画像対の間で同一と判定された対象領域の間で判定してもよい。この場合、関連性決定部204は、関連性2についての関連性フラグには、ある静止画像中の対象領域の面積よりもその次の静止画像中の対象領域の面積の方が大きい場合には1を、小さい場合には−1を、大小関係が存在しない場合には0を設定する。あるいは、関連性決定部204は、スライドショーに含まれる全静止画像から検出された対象領域のうち、同一と判定された対象領域に共通する部分領域の面積もしくは特徴点間距離を比較して、対象の大小関係を決定してもよい。関連性決定部204は、例えば、同一と判定された対象領域に共通する部分領域の最大面積Smaxと最小面積Sminをもとに、(Smax+2Smin)/3よりも小さい同一の対象領域を小とすればよい。また、関連性決定部204は、(Smax+2Smin)/3よりも大きく(2Smax+Smin)/3より小さい同一の対象領域を中とすればよい。また、関連性決定部204は、(2Smax+Smin)/3よりも大きい同一の対象領域を大とすればよい。この場合、関連性決定部204は、ある静止画像中の対象領域とその次の静止画像中の対象領域とが、小と中もしくは中と大の関係であれば関連性フラグに1を設定する。関連性決定部204は、ある静止画像中の対象領域とその次の静止画像中の対象領域とが、小と大の関係であれば2を設定する。関連性決定部204は、ある静止画像中の対象領域とその次の静止画像中の対象領域とが、大と中もしくは中と小の関係であれば関連性フラグに−1を設定する。関連性決定部204は、ある静止画像中の対象領域とその次の静止画像中の対象領域とが、大と小の関係であれば−2を設定する。関連性決定部204は、ある静止画像中の対象領域とその次の静止画像中の対象領域とに、大小関係が存在しない場合は、関連性フラグに0を設定する。
 提示方法決定部202は、大から小への対象の変化が連続した場合には、関連性の変化は無いものとして、同様の提示方法の変更を行なう。例えば、提示方法決定部202は、同じ時間間隔で徐々に提示時間を短くしていく。提示方法決定部202は、詳しくは以下のルールに基づいて提示方法を制御する。
 〔対象の大小関係に応じたルール〕
 (2−1)提示時間に関するルール
 提示方法決定部202は、連続する静止画像対に含まれる対象の大小関係をもとに、静止画像対の提示時間を決定する。例えば、提示方法決定部202は、対象の大小関係にある静止画像群のうち、はじめに提示される静止画像の提示時間を初期値Tsにする。そして、提示方法決定部202は、Tsを基準として後続の静止画像の提示時間を決定する。また、提示方法決定部202は、対象の大小関係にある静止画像群のうち、視認性の高い静止画像の提示時間をTpにしてもよい。そして、提示方法決定部202は、Tpを基準に後続の静止画像の提示時間を決定してもよい。また、提示方法決定部202は、大小関係にある静止画像群のうち、静止画像の提示時間がTq以下になった次の静止画像の提示時間を初期値Tsとしてもよい。そして、提示方法決定部202は、Tsを基準に後続の静止画像の提示時間を決定してもよい。また、提示方法決定部202は、大小関係にある静止画像群のうち最後に提示される静止画像の提示時間をTsに設定してもよい。提示方法決定部202は、あらかじめ設定しておいたスライドショー全体の提示時間から、提示する画像数に応じてTs、Tpの値を算出してもよい。また、連続する静止画像対に含まれる対象に大小関係がない場合には、提示方法決定部202は、前の静止画像の提示時間と独立に後続の静止画像の提示時間を決定する。例えば、提示方法決定部202は、後続の静止画像の提示時間を初期値Tsに設定してもよい。提示方法決定部202は、後続の静止画像の提示時間を規定範囲内のランダムな値に設定してもよい。
 図7を用いて様々な大きさの対象Bを撮影した静止画像を再生させる場合について説明する。関連性決定部204は、連続する静止画像間の大小関係を、スライドショーに含まれる全静止画像から検出された対象領域のうち同一と判定された対象領域の間の面積を比較して決定するとする。また、提示方法決定部202は、ある静止画像の提示時間に対し、関連性フラグ分パラメータaを乗算させることで、次の静止画像の提示時間を算出するとする。このとき、始めの静止画像701の提示時間は初期値Tsであり、静止画像701、702は小と中の関係であり、静止画像702、703は中と大の関係であり、静止画像703、704は大と小の関係であるとする。このとき、静止画像701、702の関連性フラグが1のため、静止画像702の提示時間はa×Tsとなる(aの乗算)。さらに静止画像703の関連性フラグが1のため、静止画像703の提示時間はa×a×Tsとなる(aの乗算)。静止画像703、704の関連性フラグは−2のため、704の提示時間はTsとなる(a×aの除算)。パラメータaを0から1の間に設定されていると、対象Bが小さく撮影された静止画像(ロングショット)が長く、対象Bがより大きく撮影された静止画像(ミドルショット、タイトショット)は短く提示される。
 これにより、利用者は、対象B以外の情景が映りこんだ情報量の多い静止画像について内容を理解することができる。そして、利用者は、以降の内容が前の静止画像の一部分であることを直感的に理解することができる。また、本実施形態の情報処理装置200は、連続する画像が同じ対象を含む画像であっても、連続する画像の提示時間が変化する映像を生成することができる。そのため、本実施形態には、視聴者を飽きさせないテンポ感あるスライドショーを生成できるという効果がある。
 (2−2)エフェクト・BGM・ジングルに関するルール
 提示方法決定部202は、連続する静止画像対に含まれる対象の大小関係をもとに、静止画像対の間に挿入するエフェクト・BGM・ジングルを決定する。例えば、連続する静止画像対に含まれる対象が大小関係にある場合には、提示方法決定部202は、静止画像の切り替え時に視覚的な変化の少ないエフェクトとしてあらかじめ登録された特殊効果(ディゾルブやフェード等)を挿入する。連続する静止画像対に含まれる対象が大小関係にない場合には、提示方法決定部202は、静止画像の切り替え時に視覚的な変化の大きいエフェクトとしてあらかじめ登録された特殊効果(ページめくり、ワイプ等のDVE)を挿入する。また例えば、連続する静止画像対に含まれる対象が大小関係にある場合には、提示方法決定部202は、静止画像対の提示中同じBGMを流す。連続する静止画像対に含まれる対象が同一でない場合には、提示方法決定部202は、静止画像の切り替え時にBGMを止めるもしくは異なるBGMに切り替える。
 また、提示方法決定部202は、大小関係が存在しない画像間に、ジングルを挿入してもよい。これにより、大小関係の対象を撮影した静止画像群は、画像や音響の変化がなく滑らかに接続される。そのため、視聴者は、静止画像が互いに変化がなくほぼ同じ内容であることを容易に理解できる。また、連続する静止画像対に含まれる対象が大小関係に無い場合には、画像や音響が大きく変化するために、視聴者は内容に変化があったことに気づきスライドショーの内容理解に集中することができる。
 (関連性3. 対象の部分関係)
 関連性決定部204は、関連性を、静止画像対に含まれる2つの静止画像に表わされた対象の部分関係によって決定してもよい。つまり、関連性決定部204は、静止画像対に含まれる2つの静止画像に表わされた対象が全体と部分との関係にあるかによって決定してもよい。関連性3は、このように決定された関連性である。
 「対象の部分関係にある」とは、対象スライドショー内で連続する静止画像対に映っている対象が同一であり、かつ互いに異なる対象の部分を撮影した画像となっている関係を表す。例えば、広い景色や、大きな対象や、長い対象を撮影したい場合に、対象の一部を撮影した静止画像を組み合わせてスライドショー再生することで全体を表現するケースがこれにあたる。
 関連性決定部204は、ある静止画像中の対象領域とその次の静止画像中の対象領域とが対象の部分関係にある場合には、関連性3についての関連性フラグに1を設定する。関連性決定部204は、対象の部分関係でない場合には、関連性3についての関連性フラグに0を設定する。関連性決定部204は、対象の部分関係を、スライドショー内で連続する静止画像中の同一と判定された対象領域に共通する部分領域(共通領域)をもとに判定できる。例えば、関連性決定部204は、対象領域の一方をテンプレートとする。そして、関連性決定部204は、もう一方の対象領域を走査して差分の少ない位置を検出し重なった領域を共通領域とする。関連性決定部204は、各対象領域の共通領域以外の領域がどちらも規定面積以上の場合に、それらが対象の部分関係にあると判定する。あるいは、関連性決定部204は、スライドショーに含まれる全静止画像から同一と判定された対象領域の相対的な位置をもとに、判定を行ってもよい。
 提示方法決定部202は、全体から部分への対象の変化が連続した場合には、関連性の変化は無いものとして、例えば、同じ時間間隔で徐々に提示時間を短くしていく。提示方法決定部202は、詳しくは以下のルールに基づいて提示方法を制御する。
 〔対象の部分関係に応じたルール〕
 (3−1)提示時間に関するルール
 提示方法決定部202は、連続する静止画像対に含まれる対象の部分関係をもとに、静止画像対の提示時間を決定する。提示方法決定部202は、例えば、対象の部分関係にある静止画像群のうち、はじめに提示される静止画像の提示時間を初期値Tsにする。そして、提示方法決定部202は、Tsを基準として後続の静止画像の提示時間を決定する。また、提示方法決定部202は、対象の部分関係にある静止画像群のうち、視認性の高い静止画像の提示時間をTpにしてもよい。そして、提示方法決定部202は、Tpを基準に後続の静止画像の提示時間を決定してもよい。また、提示方法決定部202は、部分関係にある静止画像群のうち、静止画像の提示時間がTq以下になった次の静止画像の提示時間を初期値Tsとしてもよい。そして、提示方法決定部202は、Tsを基準に後続の静止画像の提示時間を決定してもよい。また、提示方法決定部202は、部分関係にある静止画像群のうち、最後に提示される画像の提示時間をTsに設定してもよい。提示方法決定部202は、あらかじめ設定しておいたスライドショー全体の提示時間から、提示する画像数に応じてTs、Tpの値を算出してもよい。連続する静止画像対に含まれる対象に部分関係がない場合には、提示方法決定部202は、前の静止画像の提示時間と独立に後続の静止画像の提示時間を決定する。例えば、提示方法決定部202は、後続の静止画像の提示時間を初期値Tsに設定してもよい。提示方法決定部202は、後続の静止画像の提示時間を規定範囲内のランダムな値に設定してもよい。
 図8を用いて、風景を撮影した静止画像を再生させる場合について説明する。関連性決定部204は、連続する静止画像間の部分関係を、スライドショーに含まれる全静止画像から検出された対象領域のうち同一と判定された対象領域の間で共通する部分領域と、対象領域との位置関係をもとに判定したとする。また、提示方法決定部202は、ある静止画像の提示時間に対して規定パラメータを乗算させることで、次の静止画像の提示時間を算出するとする。
 関連性決定部204は、始めの静止画像801の提示時間を初期値Tsにする。静止画像801と802、802と803は部分関係があり、静止画像803と804とは部分関係がない。このとき、始めの静止画像801の提示時間は初期値Tsであり、静止画像801、802の関連性フラグが1であるため、静止画像802の提示時間はa×Tsとなる。さらに静止画像802、803の関連性フラグも1であるため、静止画像803の提示時間はa Tsとなる。静止画像803と804の関連性フラグは0であるため、提示方法決定部202は、静止画像804の提示時間を初期値にもどしてTsにする。
 パラメータaに、0から1の間で、かつ、静止画像間で一致する部分領域の面積が大であるほど小さい値が設定されていると、風景について初めて提示される静止画像801が長く提示される。そして、その他の部分は前に提示された画像との重複する情報量に応じた提示時間で提示される。これにより、利用者は、風景について始めに提示された静止画像について内容を理解することができる。そして、利用者は、以降の内容が始めの静止画像とほぼ同等の内容であることを理解することができる。また、本実施形態の情報処理装置200は、連続する画像が同じ対象を含む画像であっても、連続する画像の提示時間が変化する映像を生成することができる。そのため、本実施形態には、視聴者を飽きさせないテンポ感あるスライドショーを生成できるという効果がある。
 (3−2)エフェクト・BGM・ジングルに関するルール
 提示方法決定部202は、連続する静止画像対に含まれる対象の部分関係をもとに、静止画像対の間に挿入するエフェクト・BGM・ジングルを決定する。例えば、連続する静止画像対に含まれる対象が部分関係にある場合には、提示方法決定部202は、静止画像の切り替え時に視覚的な変化の少ないエフェクトとしてあらかじめ登録された特殊効果(ディゾルブやフェード等)を挿入する。連続する静止画像対に含まれる対象が部分関係にない場合には、提示方法決定部202は、静止画像の切り替え時に視覚的な変化の大きいエフェクトとしてあらかじめ登録された特殊効果(ページめくり、ワイプ等のDVE)を挿入する。また例えば、連続する静止画像対が対象の部分関係にある場合には、提示方法決定部202は、静止画像対の提示中同じBGMを流す。連続する静止画像対に含まれる対象が同一でない場合には、提示方法決定部202は、静止画像の切り替え時にBGMを止めるもしくは異なるBGMに切り替える。また、提示方法決定部202は、大小関係が存在しない画像間に、ジングルを挿入してもよい。これにより、連続する静止画像対が対象の部分関係にある場合、その静止画像対は画像や音響の変化がなく滑らかに接続される。そのため、視聴者は、静止画像が互いに変化がなくほぼ同じ内容であることを容易に理解できる。また、連続する静止画像対が部分関係にない場合には、画像や音響が大きく変化するために、視聴者は内容に変化があったことに気づきスライドショーの内容理解に集中することができる。
 (関連性4. 対象の同種性)
 関連性決定部204は、関連性を、静止画像対に含まれる2つの静止画像に表わされた対象が同種か否かによって決定してもよい。関連性4は、このように決定された関連性である。
 「対象が同種である」とは、スライドショー内で連続する静止画像対に映っている主要な対象が、互いに同じ種別の対象であることとする。関連性決定部204は、関連性4についての関連性フラグには、ある静止画像中の対象領域とその次の静止画像中の対象領域とが同種の関係にある場合には、関連性4についての関連性フラグに1を設定する。関連性決定部204は、ある静止画像中の対象領域とその次の静止画像中の対象領域とが異種である場合には、関連性4についての関連性フラグに0を設定する。対象の同種性の判別は、同種性を判別したい各種別に属する対象の画像データ(登録データ)をもとに、機械学習に基づく方法で実現できる。まず、関連性決定部204は、登録データから各種別に属する対象の画像特徴量を抽出する。関連性決定部204は、画像特徴量として、色ヒストグラムやエッジヒストグラム等の大域特徴を用いてもよい。関連性決定部204は、画像特徴量として、HoG(Histograms of Oriented Gradients)やSIFT等の局所特徴量を用いてもよい。関連性決定部204は、大域特徴を用いてSVM(Sigular Value Decomposition)やニューラルネットワークやGMM(Gaussian Mixture Model)等で学習を行ってもよい。あるいは、関連性決定部204は、局所特徴量からBoW(Bag of Words)のように特徴量空間の変換を行った上で学習を行ってもよい。関連性決定部204は、スライドショーに含まれる各静止画像中の対象領域について同種性を判別する際は、各対象領域の画像特徴量と、学習の結果得られた各種別のモデルとの間でそれぞれ類似性を求める。そして、関連性決定部204は、対象領域を規定値以上の類似度を得た最も近いモデルの種別と判定する。関連性決定部204は、同じ種別と判定された対象領域を同種と判定する。関連性決定部204は、同種性を、上記以外の方法で判定しても構わない。
 同種の対象を含む画像が3つ連続した場合には、提示方法決定部202は、関連性の変化は無いものとして、例えば、同じ時間間隔で徐々に提示時間を短くしていく。提示方法決定部202は、詳しくは以下のルールに基づいて提示方法を制御する。
 〔対象の同種性に応じたルール〕
 (4−1)提示時間に関するルール
 提示方法決定部202は、連続する静止画像対に含まれる対象の同種性をもとに、静止画像対の提示時間を決定する。例えば、提示方法決定部202は、同種の対象を含む静止画像群のうち、はじめに提示される静止画像の提示時間を初期値Tsにする。そして、提示方法決定部202は、Tsを基準として後続の静止画像の提示時間を決定する。また、提示方法決定部202は、同種の対象を含む静止画像群のうち、視認性の高い静止画像の提示時間をTpにしてもよい。そして、提示方法決定部202は、Tpを基準に後続の静止画像の提示時間を決定してもよい。また、提示方法決定部202は、同種の対象を含む静止画像群のうち、静止画像の提示時間がTq以下になった次の静止画像の提示時間を初期値Tsにしてもよい。そして、提示方法決定部202は、Tsを基準に後続の静止画像の提示時間を決定してもよい。また、提示方法決定部202は、同種の対象を含む静止画像群のうち、最後に提示される画像の提示時間をTsに設定してもよい。提示方法決定部202は、あらかじめ設定しておいたスライドショー全体の提示時間から、提示する画像数に応じて、Ts、Tpの値を算出してもよい。連続する静止画像対に含まれる対象が同種でない場合には、提示方法決定部202は、前の静止画像の提示時間と独立に後続の静止画像の提示時間を決定する。提示方法決定部202は、後続の静止画像の提示時間を例えば、初期値Tsに設定してもよい。提示方法決定部202は、後続の静止画像の提示時間を、規定範囲内のランダムな値に設定してもよい。
 図9を用いて、花を撮影した静止画像を再生させる場合について説明する。関連性決定部204は、連続する静止画像間の同種性を、機械学習に基づく方法で判定したとする。また、提示方法決定部202は、ある静止画像の提示時間に対して、関連性フラグ分のパラメータを乗算させることで、次の静止画像の提示時間を算出する。提示方法決定部202は、始めの静止画像901の提示時間を初期値Tsにする。静止画像901、902、静止画像902、903は同種の関係であり、静止画像903、904は異種の関係である。このとき、始めの静止画像901、902の関連性フラグが1であるため、静止画像902の提示時間はa×Tsになる。さらに静止画像902、903の関連性フラグが1であるため、静止画像903の提示時間はa Tsになる。静止画像903、904の関連性フラグは0であるため、提示方法決定部202は、静止画像904の提示時間を初期値にもどしてTsにする。パラメータaを0から1の間に設定されていると、植物を含む静止画像のうち初めてに提示された静止画像901が長く提示される。そして、後続の静止画像は、901から離れるほど短い提示時間で提示される。これにより、利用者は、始めに提示された静止画像から画像内容が植物であることを理解することができる。そして、利用者は以降の静止画像の内容がほぼ同等であることを理解することができる。
 また、本実施形態の情報処理装置200は、連続する画像が同じ対象を含む画像であっても、連続する画像の提示時間が変化する映像を生成することができる。そのため、本実施形態には、視聴者を飽きさせないテンポ感あるスライドショーを生成できるという効果がある(本実施形態は、花畑で撮影した複数の花の画像を、同種の被写体を順に再生することで、この種の被写体が沢山存在したことを表現できる)。
 (4−2)エフェクト・BGM・ジングルに関するルール
 提示方法決定部202は、連続する静止画像対に含まれる対象の同種性をもとに、静止画像対の間に挿入するエフェクト・BGM・ジングルを決定する。提示方法決定部202は、例えば、連続する静止画像対に含まれる対象が同種である場合には、静止画像の切り替え時に視覚的な変化の少ないエフェクトとしてあらかじめ登録された特殊効果(ディゾルブやフェード等)を挿入する。提示方法決定部202は、連続する静止画像対に含まれる対象が異種である場合には、静止画像の切り替え時に視覚的な変化の大きいエフェクトとしてあらかじめ登録された特殊効果(ページめくり、ワイプ等のDVE)を挿入する。また例えば、連続する静止画像対が同種である場合には、提示方法決定部202は、静止画像対の提示中同じBGMを流す。連続する静止画像対が異種である場合には、提示方法決定部202は、静止画像の切り替え時にBGMを止めるもしくは異なるBGMに切り替える。また、提示方法決定部202は、異種の静止画像間に、ジングルを挿入してもよい。これにより、連続する静止画像対に含まれる対象が同種である場合、その静止画像対は、画像や音響の変化がなく滑らかに接続される。そのため、視聴者は、静止画像が互いに変化がなくほぼ同じ内容であることを容易に理解できる。また、連続する静止画像対が異種である場合には、画像や音響が大きく変化するために、視聴者は内容に変化があったことに気づきスライドショーの内容理解に集中することができる。
 (関連性5. 撮影場所の同一性)
 関連性決定部204は、関連性を、静止画像対に含まれる2つの静止画像の撮影場所の共通性によって決定してもよい。関連性5は、このように決定された関連性である。
 「撮影場所が同一である」とは、スライドショー内で連続する静止画像対を撮影した場所が同一であることとする。関連性決定部204は、ある静止画像の撮影場所とその次の静止画像の撮影場所が同じである場合には、関連性5についての関連性フラグに1を設定する。ある静止画像の撮影場所とその次の静止画像の撮影場所が異なる場合には、関連性決定部204は、関連性5についての関連性フラグに0を設定する。関連性決定部204は、撮影場所の同一性を、静止画像中の対象領域以外の領域(背景領域)の類似度をもとに判定できる。例えば、関連性決定部204は、静止画像から対象領域と背景領域を分離してもよい。そして、関連性決定部204は、背景領域から抽出した画像特徴量が類似する場合に、同一の撮影場所と判定してもよい。関連性決定部204は、撮影場所の同一性を、上記以外の方法で判定しても構わない。関連性決定部204は、撮影場所の同一性を、スライドショー内で連続する静止画像の間で背景の類似性を判定してもよい。あるいは、関連性決定部204は、スライドショーに含まれる全静止画像中の背景領域の同一性をもとに、撮影場所の同一性を判定してもよい。関連性決定部204は、画像情報に加えて、メタ情報である撮影場所やセンサ情報であるGPSを組み合わせて、撮影場所の同一性を判定してもよい。
 提示方法決定部202は、同じ撮影場所で撮影された画像が3つ連続した場合には、関連性の変化は無いものとして、例えば、同じ時間間隔で徐々に提示時間を短くしていく。提示方法決定部202は、詳しくは以下のルールに基づいて提示方法を制御する。
 〔撮影場所の同一性に応じたルール〕
 (5−1)提示時間に関するルール
 提示方法決定部202は、連続する静止画像対の撮影場所の同一性をもとに、静止画像対の提示時間を決定する。提示方法決定部202は、例えば、同一の場所で撮影された静止画像群のうち、はじめに提示される静止画像の提示時間を初期値Tsにする。そして、提示方法決定部202は、Tsを基準として後続の静止画像の提示時間を決定する。また、提示方法決定部202は、同一の場所で撮影された静止画像群のうち、視認性の高い静止画像の提示時間をTpにしてもよい。そして、提示方法決定部202は、Tpを基準に後続の静止画像の提示時間を決定してもよい。また、提示方法決定部202は、同一の場所で撮影された静止画像群のうち、静止画像の提示時間がTq以下になった次の静止画像の提示時間を初期値Tsにしてもよい。そして、提示方法決定部202は、Tsを基準に後続の静止画像の提示時間を決定してもよい。また、提示方法決定部202は、同一の場所で撮影された静止画像群のうち、最後に提示される画像の提示時間をTsに設定してもよい。提示方法決定部202は、あらかじめ設定してスライドショー全体の提示時間から、提示する画像数に応じて、Ts、Tpの値を算出してもよい。連続する静止画像対が異なる場所で撮影されている場合には、提示方法決定部202は、前の静止画像の提示時間と独立に後続の静止画像の提示時間を決定する。提示方法決定部202は、例えば、後続の静止画像の提示時間を、初期値Tsに設定してもよい。提示方法決定部202は、後続の静止画像の提示時間を、規定範囲内のランダムな値に設定してもよい。
 (5−2)エフェクト・BGM・ジングルに関するルール
 提示方法決定部202は、連続する静止画像対の撮影場所の同一性をもとに、静止画像対の間に挿入するエフェクト・BGM・ジングルを決定する。例えば、連続する静止画像対が同じ場所で撮影された場合には、提示方法決定部202は、静止画像の切り替え時に視覚的な変化の少ないエフェクトとしてあらかじめ登録された特殊効果(ディゾルブやフェード等)を挿入する。連続する静止画像対が異なる場所で撮影された場合には、提示方法決定部202は、静止画像の切り替え時に視覚的な変化の大きいエフェクトとしてあらかじめ登録された特殊効果(ページめくり、ワイプ等のDVE)を挿入する。また例えば、連続する静止画像対が同じ場所で撮影された場合には、提示方法決定部202は、静止画像対の提示中同じBGMを流す。連続する静止画像対が異なる場所で撮影された場合には、提示方法決定部202は、静止画像の切り替え時にBGMを止めるもしくは異なるBGMに切り替える。また、提示方法決定部202は、異なる場所で撮影された静止画像間に、ジングルを挿入してもよい。これにより、連続する静止画像対が同じ場所で撮影された場合、その静止画像対は、画像や音響の変化がなく滑らかに接続される。そのため、視聴者は、静止画像が互いに変化がなくほぼ同じ内容であることを容易に理解できる。また、異なる場所で撮影された場合には、画像や音響的が大きく変化するため、視聴者は内容に変化があることに気づきスライドショーの内容理解に集中することができる。
 (関連性6. 撮影時間帯の同一性)
 関連性決定部204は、関連性を、静止画像対に含まれる2つの静止画像の撮影時間帯の共通性によって決定してもよい。関連性6は、このように決定された関連性である。
 「撮影時間帯が同一である」とは、スライドショー内で連続する静止画像対を撮影した時間帯が同一であることとする。関連性決定部204は、ある静止画像の撮影時間帯とその次の静止画像の撮影時間帯が同じである場合には、関連性6についての関連性フラグに1を設定する。関連性決定部204は、ある静止画像の撮影時間帯とその次の静止画像の撮影時間帯が異なる場合には関連性6についての関連性フラグに0を設定する。関連性決定部204は、撮影時間帯の同一性を、静止画像中の背景領域の色情報をもとに判定できる。関連性決定部204は、例えば、1日を複数の時間帯に分割し、各時間帯における太陽光の色ヒストグラムの統計量を保持する。そして、関連性決定部204は、静止画像の背景領域中にいずれかの時間帯の統計量と近い部分領域が含まれるときに、その時間帯に撮影された静止画像と判定する。関連性決定部204は、各静止画像の撮影時間帯を推定する。そして、関連性決定部204は、推定時間が同じ場合に撮影時間帯が同一と判定する。関連性決定部204は、撮影時間帯の同一性を、上記以外の方法で判定しても構わない。関連性決定部204は、撮影時間帯の同一性を、スライドショー内で連続する静止画像の間で撮影時間帯の類似性によって判定してもよい。あるいは、関連性決定部204は、スライドショーに含まれる全静止画像中の撮影時間帯の同一性をもとに、撮影時間帯の同一性を判定してもよい。関連性決定部204は、画像情報に加えて、メタ情報である撮影時刻と組み合わせて、撮影時間帯の同一性を判定してもよい。
 提示方法決定部202は、同じ撮影時間帯に撮影された画像が3つ連続した場合には、関連性の変化は無いものとして、例えば、同じ時間間隔で徐々に提示時間を短くしていく。提示方法決定部202は、詳しくは以下のルールに基づいて提示方法を制御する。
 〔撮影時間帯の同一性に応じたルール〕
 (6−1)提示時間に関するルール
 提示方法決定部202は、連続する静止画像対の撮影時間帯の同一性をもとに、静止画像対の提示時間を決定する。提示方法決定部202は、例えば、同一の時間帯に撮影された静止画像群のうち、はじめに提示される静止画像の提示時間を初期値Tsにする。そして、提示方法決定部202は、Tsを基準として後続の静止画像の提示時間を決定する。また、提示方法決定部202は、同一の時間帯に撮影された静止画像群のうち、視認性の高い静止画像の提示時間をTpにしてもよい。そして、提示方法決定部202は、Tpを基準に後続の静止画像の提示時間を決定してもよい。また、提示方法決定部202は、同一の時間帯に撮影された静止画像群のうち、静止画像の提示時間がTq以下になった次の静止画像の提示時間を初期値Tsにしてもよい。そして、提示方法決定部202は、Tsを基準に後続の静止画像の提示時間を決定してもよい。また、提示方法決定部202は、同一の時間帯に撮影された静止画像群のうち、最後に提示される画像の提示時間をTsに設定してもよい。提示方法決定部202は、あらかじめ設定しておいたスライドショー全体の提示時間から、提示する画像数に応じて、Ts、Tpの値を算出してもよい。連続する静止画像対が異なる時間帯に撮影された場合には、提示方法決定部202は、前の静止画像の提示時間と独立に後続の静止画像の提示時間を決定する。提示方法決定部202は、後続の静止画像の提示時間を、例えば、初期値Tsに設定してもよい。提示方法決定部202は、後続の静止画像の提示時間を、規定範囲内のランダムな値に設定してもよい。
 (6−2)エフェクト・BGM・ジングルに関するルール
 提示方法決定部202は、連続する静止画像対の撮影時間帯の同一性をもとに、静止画像対の間に挿入するエフェクト・BGM・ジングルを決定する。提示方法決定部202は、例えば、連続する静止画像対が同一の時間帯に撮影された場合には、静止画像の切り替え時に視覚的な変化の少ないエフェクトとしてあらかじめ登録された特殊効果(ディゾルブやフェード等)を挿入する。提示方法決定部202は、連続する静止画像対が異なる時間帯に撮影された場合には、静止画像の切り替え時に視覚的な変化の大きいエフェクトとしてあらかじめ登録された特殊効果(ページめくり、ワイプ等のDVE)を挿入する。また例えば、連続する静止画像対が同一の時間帯に撮影された場合には、提示方法決定部202は、静止画像対の提示中同じBGMを流す。連続する静止画像対が異なる時間帯に撮影された場合には、提示方法決定部202は、静止画像の切り替え時にBGMを止めるもしくは異なるBGMに切り替える。また、提示方法決定部202は、異なる時間帯の静止画像間に、ジングルを挿入してもよい。これにより、連続する静止画像対が同一の時間帯に撮影された場合、その静止画像対は画像や音響の変化がなく滑らかに接続される。そのため、視聴者は、静止画像が互いに変化がなくほぼ同じ内容であることを容易に理解できる。また、連続する静止画像対が異なる時間帯に撮影された場合には、画像や音響が大きく変化するために、視聴者は内容に変化があったことに気づきスライドショーの内容理解に集中することができる。連続する静止画像対は、提示ルールとして、上記のいずれか1つのルールを適用してもよい。連続する静止画像対は、提示ルールとして、複数のルールを組み合わせて用いてもよい。スライドショー生成部203は、提示方法決定部202から入力された提示方法情報と、画像入力部210から入力された画像情報をもとに、スライドショーを生成する。
 [第4実施形態]
 第2実施形態に示したように、関連性決定部204は、スライドショーに含まれる全静止画像から検出した全対象領域を、類似性に基づいてグループ化してもよい。そして、関連性決定部204は、隣接する静止画像対から検出された対象領域が同じグループに属する場合には同一の対象と判定してもよい。
 本実施形態では、上述のように関連性決定部204が静止画像をグループ化する場合の処理の流れについて説明する。
 対象領域の検出手法は、あらかじめ登録された特定対象を検出する場合の検出手法と、登録されていない一般対象を検出する場合の検出手法とに分けられる。関連性決定部204は、特定対象を検出する場合は、登録された各対象の画像データをテンプレートとしてもよい。そして、関連性決定部204は、様々な解像度に変換したテンプレートで入力画像を走査してもよい。そして、関連性決定部204は、テンプレートと同じ位置の画素値の差分が小さい領域を対応する対象領域として検出してもよい。
 図10のフローチャートを参照して、本実施の形態の動作について詳細に説明する。本実施の形態でも、図5に示す静止画像をスライドショー再生させる場合を例に説明する。関連性決定部204は、関連性種別として、同一性・大小関係・部分関係を判定する。提示ルールの決定方法は、第1実施の形態と同様とする。
 まず、画像入力部210は、関連性決定部204に、すべての静止画像の画像情報を入力する(ステップS1001)。
 関連性決定部204は、全静止画像から画像特徴量を抽出する。そして、関連性決定部204は、画像特徴量の類似性をもとに同一対象を撮影した静止画像をまとめるようにグループ化する(ステップS1003)。図5の例では、静止画像501、502、503、504がグループA、静止画像505、506、507、508、509、510がグループB、511、512、513がグループCに分類される。
 関連性決定部204は、同一のグループに属する静止画像間で、大小関係と部分関係を判定する(S1005)。関連性決定部204は、全静止画像からSIFT等の局所特徴点を抽出して互いに対応点を取る。すると、グループAについては、静止画像502、503、504が静止画像501に包含されていることが分かる。そして、静止画像502、503、504は互いに共通領域が小さいことが分かる。よって、関連性決定部204は、静止画像501と静止画像502の間に大小関係があることを判定する。そして、関連性比較部201は、静止画像502と静止画像503、静止画像503と静止画像504に部分関係があることを判定する。
 関連性決定部204は同様に局所特徴量の比較を行うことで、グループBについては、静止画像506、507、508、509、510は505に内包されることがわかる。さらに、静止画像506、507は静止画像513に内包されることがわかる。よって、静止画像505が大の関係であり、静止画像511、512、513は中の関係であり、静止画像506、507、は小の関係であると判定される。また、静止画像506、507は、同一対象の一部分を撮影した静止画像であると判定される。静止画像511、512、513も同一対象の一部を撮影した静止画像であると判定される。グループCについては、いずれも共有領域が画面全体を占めるため、大小関係・部分関係はいずれもなしと判定される。上記の結果、提示方法決定部202は、提示時間長1101およびエフェクト1102を図11のように決定する(S1007)。最後に、スライドショー生成部203は、決定した提示方法を用いてスライドショーを生成する(S1009)
 本実施形態には、第2実施形態の効果に加えて、グループ同士の関係性を明瞭にスライドショーに表わすことが可能となると言う効果がある。
 [他の実施形態]
 以上、本発明の実施形態について詳述したが、それぞれの実施形態に含まれる別々の特徴を如何様に組み合わせたシステムまたは装置も、本発明の範疇に含まれる。
 また、本発明は、複数の機器から構成されるシステムに適用されてもよいし、単体の装置に適用されてもよい。さらに、本発明は、実施形態の機能を実現する情報処理プログラムが、システムあるいは装置に直接あるいは遠隔から供給される場合にも適用可能である。したがって、本発明の機能をコンピュータで実現するために、コンピュータにインストールされるプログラム、あるいはそのプログラムを格納した媒体、そのプログラムをダウンロードさせるWWW(World Wide Web)サーバも、本発明の範疇に含まれる。
 例えば、情報処理装置100、情報処理装置200は、それぞれ、コンピュータ及びコンピュータを制御するプログラム、専用のハードウェア、又は、コンピュータ及びコンピュータを制御するプログラムと専用のハードウェアの組合せにより実現することができる。
 関連性比較部101、提示方法決定部102、スライドショー生成部103、関連性比較部201、提示方法決定部202、スライドショー生成部203、関連性決定部204は、例えば、プログラムを記憶する記録媒体からメモリに読み込まれた、各部の機能を実現するための専用のプログラムと、そのプログラムを実行するプロセッサにより実現することができる。あるいは、関連性比較部101、提示方法決定部102、スライドショー生成部103、関連性比較部201、提示方法決定部202、スライドショー生成部203、関連性決定部204の一部又は全部を、各部の機能を実現する専用の回路によって実現することもできる。
 以上、実施形態を参照して本発明を説明したが、本発明は上記実施形態に限定されるものではない。本発明の構成や詳細には、本発明のスコープ内で当業者が理解し得る様々な変更をすることができる。
 以上、実施形態を参照して本願発明を説明したが、本願発明は上記実施形態に限定されものではない。本願発明の構成や詳細には、本願発明のスコープ内で当業者が理解し得る様々な変更をすることができる。
 この出願は、2011年5月12日に出願された日本出願特願2011−107103を基礎とする優先権を主張し、その開示の全てをここに取り込む。
Hereinafter, exemplary embodiments of the present invention will be described in detail with reference to the drawings. However, the components described in the following embodiments are merely examples, and are not intended to limit the technical scope of the present invention only to them.
[First Embodiment]
An information processing apparatus 100 as a first embodiment of the present invention will be described with reference to FIG. The information processing apparatus 100 is an apparatus that generates a slide show including at least three still images.
The information processing apparatus 100 includes an association comparison unit (comparison unit) 101, a presentation method determination unit 102, and a slide show generation unit (generation unit) 103.
The relevance comparison unit 101 includes a first relevance between still images of the first still image pair included in the still image group and a still image of the second still image pair in the still image group including at least three still images. The second relevance between each other is compared.
The presentation method determination unit 102 identifies the presentation method of the second still image pair from the first information that identifies the presentation method of the first still image pair based on the change from the first relevance to the second relevance. The second information is determined.
The slide show generation unit 103 generates a slide show including the first still image pair and the second still image pair based on the presentation method determined by the presentation method determination unit 102.
According to the above configuration, it is possible to generate a slide show that presents still images according to a change in relevance between still images.
[Second Embodiment]
Next, an information processing apparatus 200 according to the second embodiment of the present invention will be described with reference to FIG. FIG. 2 is a block diagram for explaining the configuration of the information processing apparatus 200 according to this embodiment.
(Constitution)
The information processing apparatus 200 includes an image input unit 210 that inputs image information from an imaging device 250 such as a digital camera or a digital video camera, a relevance comparison unit (comparison unit) 201 that compares relevance of input images, A relevance determination unit 204 that determines gender. The information processing apparatus 200 also includes a presentation method determination unit 202 that determines the display length of each image included in the slide show, the effect at the time of switching images, the BGM (Background Music) being displayed, the jingle at the time of switching, and the like. Further, the information processing apparatus 200 includes a slide show generation unit (generation unit) 203 that generates a slide show by combining still images input by the presentation method determined by the presentation method determination unit 202. The information processing apparatus 200 is also connected to a display 260 for displaying a stationary slide show.
The image information input by the image input unit 210 includes an image ID (identifier) for identifying a still image, a presentation order in a slide show, and pixel information of the still image. The image information may include meta information describing a subject, a shooting location, and a shooting time shown in a still image, and sensor information such as GPS (Global Positioning System).
The relevancy determination unit 204 determines the first relevance based on the commonality of objects represented in two still images included in the first still image pair. The relevance determination unit 204 determines the second relevance based on the commonality of objects represented in two still images included in the second still image pair. The first still image pair and the second still image pair are pairs of continuous still images, respectively.
The relevancy determination unit 204 inputs the image ID of the still image and the relevance flag to the relevance comparison unit 201 as the image relevance information. The relevancy determination unit 204 may input pixel information in addition to the above as image relevance information. The relevance flag is data representing a relevance type existing between a current still image and a still image presented thereafter among relevance types defined in advance. Alternatively, the relevance flag is data indicating that no relevance type exists between these images (no relevance). A flag 1 is set to the relevance flags of all relevance types existing between a certain still image and a subsequent still image. A flag 0 is set for the relationship flag of the relationship type that does not exist. Alternatively, the relevance flag may be set with any numerical value that has meaning depending on the relevance type.
The relationship comparison unit 201 compares the first relationship between the still images of the first still image pair included in the still image group and the second relationship between the still images of the second still image pair. Specifically, the relevance comparison unit 201 determines the relevance between pairs of still images that are consecutive in the slide show. More specifically, the still image group includes continuous first, second, and third still images. The first still image pair is a pair of a first still image and a second still image. The second still image pair is a pair of the second still image and the third still image.
The presentation method determination unit 202 presents each still image based on the image relevance information input from the relevance comparison unit 201, the image information input from the image input unit 210, and a pre-registered presentation rule. Decide how. Then, the presentation method determination unit 202 inputs the presentation method information to the slide show generation unit 203.
The slide show generation unit 203 generates a slide show by combining still images based on the determined presentation method.
(Relevance determination method)
Hereinafter, a method for determining relevance in the relevance determination unit 204 will be described. Although various types of relevance can be considered, in the present embodiment, the identity of the following objects will be described as an example.
Relevance Object identity
The relevancy determination unit 204 determines relevance based on the commonality (identity) of objects represented in two still images included in the still image pair. The relationship 1 is the relationship determined in this way. The identity can be determined by the commonality of the feature quantities derived from a plurality of still images included in the still image group.
In other words, the relevance 1 is a relevance indicating that the objects shown in the still image pairs continuous in the slide show are the same. The relevance determination unit 204 sets 1 in the relevance flag for relevance 1 if it is the same, and 0 if it is not the same. The relevancy determination unit 204 can determine the identity of the target based on the similarity of the target area images detected from the still images. The target region is a region on a still image of a target having a certain image-like pattern such as a stationary object such as a tree or a building, or a moving body such as a human or an animal. The target area may be a partial area in the still image. The target area may be the entire still image.
The relevancy determination unit 204 detects a target region from a pair of still images that are continuous in the slide show. Then, the relevancy determination unit 204 determines the identity of the target based on the similarity of the target area. Alternatively, the relevancy determination unit 204 may determine the identity based on the similarity between the target areas detected from all the still images included in the slide show. In this case, the relevance determination unit 204 groups all detected target areas based on similarity. Then, the relevance determination unit 204 determines that the target areas are the same when the target areas detected from adjacent still image pairs belong to the same group.
The detection method of the target area is divided into a detection method for detecting a specific target registered in advance and a detection method for detecting a general target that is not registered. When the relevancy determination unit 204 detects a specific target, the image data of each registered target may be used as a template. The relevancy determination unit 204 may scan the input image with templates converted into various resolutions. The relevancy determination unit 204 may detect a region having a small difference in pixel values at the same position as the template as a corresponding target region.
Alternatively, the relevancy determination unit 204 may extract an image feature amount expressing color, texture, and shape from each partial region of the input image. The relevancy determination unit 204 may set a partial region having an image feature amount similar to the registered image feature amount of each target as a corresponding target region. In addition, when the specific target is a person, there is a method of using information obtained from the entire face. For example, there is a method of storing an image showing various faces as a template and determining that a face exists in the input image when the difference between the input image and the template is equal to or smaller than a threshold value. In addition, it is conceivable to store a model combining color information such as skin color, edge direction and density in advance, and determine that a face exists when an area similar to the model is detected from an input frame. In addition, there is a method of performing face detection using a template created by utilizing the fact that the contour of the face (head) is an ellipse and the eyes and mouth are elongated. Furthermore, the face detection method using the characteristics of the luminance distribution that the cheeks and forehead are bright and the eyes and mouth are low, and the face is detected using the face symmetry and skin color area and position. The method of performing etc. is mentioned.
In addition, there is a method for statistically learning the feature value distribution obtained from a large number of human face and non-face learning samples and determining whether the feature value obtained from the input image belongs to the face or non-face distribution. is there. Examples of such a method include a method using a neural network, a support vector machine, an AdaBoost method, and the like. Further, when detecting a general target, for example, Normalized Cut, Saliency Map, Depth of Field (DoF), or the like may be used. Normalized Cut is a method of dividing an image into a plurality of regions. For details, see Jianbo Shi and Jitendra Malik, “Normalized Cuts and Image Segmentation”, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 22, no. 8. August 2000 has a disclosure. The relevancy determination unit 204 may detect a region located in the center of the screen among regions divided by Normalized Cut as a target region.
Further, Saliency Map is a method of calculating an object region in an image from visual attention. For SaliencyMap, see L. Itti, C.I. Koch and E.M. Niebur, “A Model of Saliency-based Visual Attention for Rapid Scene Analysis,” IEEE Trans. Pattern Analysis and Machine Intelligence, Vol. 20, no. 11, pp. 1254-1259, 1998. The relevancy determination unit 204 may detect, as a target region, a region for which a high importance level has been calculated by the Saliency Map. Dof is a method based on the characteristic that the target edge existing within the depth of field is not blurred and the edge outside the depth of field is blurred. For details, see 3Du-Ming Tsai, Hu-Jong Wang, “Segmenting focused objects in complex visual images”, Pattern Recognition Letters, Vol. 19, pp. 929 940, 1998. There is a disclosure. The relevancy determination unit 204 may calculate the amount of blur based on the thickness of the edge, combine edges with less blur, and detect a focused area as a target area.
The relevancy determination unit 204 is a position in a still image or high visibility (lighting conditions, orientation, angle, position on the screen, hiding by another object, blur, facial expression (in the case of a person), etc. One evaluation area may be detected for each still image based on an evaluation value indicating the quality of the reflection based on the above or the appearance frequency of a plurality of images. The relevancy determination unit 204 may combine a plurality of detected target areas into one target area. The relevancy determination unit 204 may extract the image feature amount, and calculate the similarity between the target regions based on a scale that calculates a higher value as the difference in the image feature amount is smaller. The relevancy determination unit 204 can calculate an image feature amount based on image information such as a color, an edge, and a texture detected from the target region. Alternatively, the relevancy determination unit 204 may detect local feature points such as SIFT (Scale-Invariant Feature Transform) from each target image region. Then, the relevance determination unit 204 may associate feature points between image regions. The relevancy determination unit 204 may use a scale that calculates a higher value as the number of associated feature points is larger or the positional relationship between the associated feature points is similar between images.
(Presentation method)
When there is no change in the relevance, the presentation method determination unit 202 selects the second presentation method information from the first presentation method information so that the presentation method in the second still image pair changes, similarly to the change in the presentation method in the first still image pair. Determine presentation method information. For example, when there is no relevance change, the presentation method determination unit 202 determines the presentation method information so that the presentation time changes in the second still image pair, similarly to the change in the presentation time in the first still image pair. .
The presentation method information is data indicating the presentation method of each still image. The presentation method information includes an image ID and a presentation time. In addition to the above, the presentation method information may include an effect, BGM, audio jingle, and video jingle. The presentation rule is a rule that defines a method of presenting a still image according to the relevance type. It is assumed that the presentation method determination unit 202 holds a parameter that defines each presentation time of continuous still image pairs as a presentation rule. In addition to the presentation time, the presentation method determination unit 202 may hold control parameters related to effects, BGM, and jingles (short video, music, and sound effects) inserted between still images. In addition, the presentation rule may define a presentation method in a case where no relevance type exists in a continuous still image pair.
In the present embodiment, the presentation method determination unit 202 determines the presentation time of a still image pair based on the identity of objects included in successive still image pairs.
For example, when the targets included in consecutive still image pairs are the same, the presentation method determination unit 202 sets the presentation time of the still image to be presented first to the initial value Ts. The presentation method determination unit 202 determines the presentation time of the subsequent still image based on Ts. If the objects included in the continuous still image pairs are not the same, the presentation method determination unit 202 determines the subsequent presentation time independently of the previous still image presentation time. The presentation method determination unit 202 may set the subsequent presentation time to, for example, the initial value Ts. The presentation method determination unit 202 may set the subsequent presentation time to a random value within a specified range.
Note that the presentation method determination unit 202 may set Tp as the presentation time of a still image with high visibility among the group of still images obtained by photographing the same object. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Tp. In addition, the presentation method determination unit 202 may set the presentation time of the next still image in which the presentation time of the still image is equal to or less than Tq among the still image groups obtained by photographing the same target as the initial value Ts. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Ts. In addition, the presentation method determination unit 202 may set the presentation time of the still image presented last among the still image groups obtained by photographing the same target as the initial value Ts. The presentation method determination unit 202 may calculate the values of Ts and Tp according to the number of images to be presented in consideration of the preset presentation time of the entire slide show.
A case where a group of still images obtained by photographing persons A in various directions is reproduced will be described with reference to FIG. The presentation method determination unit 202 calculates the presentation time of the subsequent still image by multiplying the parameter a by the presentation time of a certain still image. At this time, assuming that the presentation time of the first still image 301 is the initial value Ts, the presentation time of the subsequent still images 302 to 305 is expressed by the following equation (1).
Figure JPOXMLDOC01-appb-M000001
Further, when the visibility evaluation value of the still image 303 facing the front is equal to or greater than the threshold, if the presentation time of the still image 303 is Tp, the presentation time of the subsequent still image is expressed by the following equation (2). .
Figure JPOXMLDOC01-appb-M000002
When the parameter a is set between 0 and 1, among the still images including the person A, the initially presented still image 301 and the still image 303 with the good reflection of the person A are presented for a long time. Other still images including A are presented gradually and gradually as they move away from the still image 303. As a result, in this embodiment, the user can understand the content of the still image with the first moment when the target first appears or a good reflection, and the other images have almost the same content as the understood content. There is an effect that can be understood. Further, the information processing apparatus 200 according to the present embodiment can generate a video in which the presentation time of consecutive images changes even if the consecutive images include the same target. For this reason, this embodiment has an effect that a slide show with a tempo that does not bore viewers can be generated.
The presentation method determination unit 202 also determines an effect, a BGM, and a jingle to be inserted between still image pairs based on the identity of objects included in successive still image pairs. For example, when the target included in the pair of still images is the same, the presentation method determination unit 202 has a special effect (such as dissolve or fade) registered in advance as an effect with little visual change when switching still images. Insert. If they are not the same, the presentation method determination unit 202 inserts special effects (DVE (Digital Video Effects) such as page turning and wipe) registered in advance as effects having a large visual change when switching still images. The presentation method determination unit 202 gradually shortens (lengthens) the length of the effect jingle. The presentation method determination unit 202 gradually decreases (increases) the volume of the BGM.
Further, for example, when the targets included in the continuous still image pairs are the same, the presentation method determination unit 202 plays the same BGM during the presentation of the still image pairs. When the targets included in the continuous still image pairs are not the same, the presentation method determination unit 202 stops the BGM or switches to a different BGM when switching the still images. In addition, the presentation method determination unit 202 may insert jingles between images that do not have identity. Thereby, the still image group which image | photographed the same object is connected smoothly, without a change of an image or a sound. Therefore, the viewer can easily understand that the still images have almost the same contents without any change. In addition, since the images and sound greatly change when they are not the same target, the viewer can notice that the content has changed and can concentrate on understanding the content of the slide show.
(Operation)
Next, the operation of the present embodiment will be described in detail with reference to the flowchart of FIG. Here, as an example, consider a case where a slide show for reproducing still images 501 to 513 shown in FIG. 5 in ascending order is generated. This slide show tells about the events of shooting flowers and people in a greenhouse in the building. A rectangle in FIG. 5 is an example of a target region detected from each still image by the relevancy determination unit 204.
The presentation rules for controlling the presentation method are rules based on the magnitude relationship or partial relationship for still image pairs with the same target area, and rules based on homogeneity for still image pairs with the same target area. . The rules based on the magnitude relationship, the partial relationship, and the homogeneity will be described in detail in the third and subsequent embodiments.
First, the image input unit 210 inputs the image information of the still image 501 to the relevancy determination unit 204, and proceeds to step S403. If the input still image is a start image, the process returns from step S403 to step S401, and the image input unit 210 inputs the image information of the second still image 502 to the relevancy determination unit 204 (step S403).
Next, in step S <b> 405, the relevance determination unit 204 detects a target area from the still images 501 and 502. Assume that in the relevancy determination unit 204, buildings, flowers, and people are registered in advance as targets, and each model is learned. Then, the relevancy determination unit 204 detects a part surrounded by a solid line rectangle as a target area of the building from the still images 501 and 502.
The relevancy determination unit 204 extracts image feature amounts from the pixel information of the target region 0 and the target region 1, and determines identity, magnitude relationship, partial relationship, and homogeneity based on the similarity between the regions. Since the target areas 0 and 1 are detected as the types of buildings, it is determined that there is a homogeneity. Also, a broken-line rectangular area on the still image 501 is detected as a common area of the target area 1 and the target area 0. It is determined that the target areas 1 and 0 have a magnitude relationship. Further, since there is no area other than the common area on the target area 0, it is determined that there is no partial relationship. Therefore, the relevance flags between the still image 501 and the still image 502 are 1, −1, 0, 1 in the order of identity, magnitude relationship, partial relationship, and homogeneity (step S407).
The presentation method determination unit 202 determines the presentation method based on the image ID and the relevance flag as the image relevance information. Since the target areas of the still image 501 and the still image 502 are the same, a rule based on a magnitude relationship or a partial relationship is applied. The presentation method determination unit 202 sets the presentation time of the still image 501 that is the start image to the initial value Ts. The presentation method determination unit 202 sets the presentation time of the still image 502 to a * Ts because the magnitude relationship between the still images 501 and 502 is a small / large relationship. Since the still images 501 and 502 have a magnitude relationship, the presentation method determination unit 202 inserts a dissolve with little visual change as an effect of switching between the still images 501 and 502 (step S409).
The slide show generation unit 203 generates a slide show using the still images 501 and 502 with the determined presentation time / effect (step S411). The above steps are performed for all still images (step S413).
In FIG. 6, the type of the target area detected from the still image is the target area 601. The relevance flag for each relevance type is 602. The presentation time length and the effect determined by the presentation method determination unit 202 are the presentation time length 603 and the effect 604.
As described above, in the present embodiment, even when consecutive still images include different numbers of face images, if the content meanings are related to each other, it can be shown to the viewer that they are related. . The reason is that the presentation method determination unit 202 controls the presentation method of continuous still images according to the relevance of each other.
In addition, in the present embodiment, even when consecutive still images include the same number of face images, if there is no relevance in terms of content, it can be shown to the viewer that there is no relevance. .
Further, in the present embodiment, even when still images captured at various time intervals are related to each other, still images can be presented in a manner that is easy for the viewer to understand. .
In the present embodiment, even in the case of still image groups captured at regular time intervals, when the content meanings are not related to each other, it is possible to present that there is no relationship in a way that the viewer can easily understand.
[Third Embodiment]
Instead of or in addition to the relevance disclosed in the second embodiment, the presentation method may be changed according to any one of the following relevance changes, and any one of the following presentation rules may be adopted: .
(Relevance 2. Target size relationship)
The relevancy determination unit 204 may determine the relevance based on the magnitude relationship between the objects represented in the two still images included in the still image pair. That is, the relevancy determination unit 204 may determine relevance by changing the size of a region including a specific object registered in advance in two still images included in a still image pair. The relevance 2 is the relevance determined in this way.
The “target size relationship” means that the targets included in the still image pairs that are continuous in the slide show are the same, and the area of the target region has a difference greater than a specified value. For example, there is a case where a target is introduced by generating a slide show by combining an image including the periphery of the target and an image obtained by photographing only the target.
The relevancy determination unit 204 can determine the magnitude relationship between objects based on the areas of partial areas common to the target areas determined to be the same or the distances between feature points included in the common partial areas. For example, the relevancy determination unit 204 can determine that the larger the distance between feature points is, the larger the object is photographed. The relevancy determination unit 204 may determine between target areas determined to be the same between a pair of still images that are consecutive in a slide show. In this case, the relevancy determination unit 204 determines that the relevance flag for relevance 2 has a larger area of the target area in the next still image than the area of the target area in the still image. 1 is set when it is small, and -1 is set when there is no magnitude relationship. Alternatively, the relevancy determination unit 204 compares the areas of the partial areas common to the target areas determined to be the same among the target areas detected from all the still images included in the slide show or the distance between the feature points, The magnitude relationship may be determined. For example, the relevancy determination unit 204 reduces the same target area smaller than (Smax + 2Smin) / 3 based on the maximum area Smax and the minimum area Smin of the partial areas common to the target areas determined to be the same. That's fine. The relevancy determination unit 204 may set the same target region larger than (Smax + 2Smin) / 3 and smaller than (2Smax + Smin) / 3. The relevancy determination unit 204 may increase the same target area larger than (2Smax + Smin) / 3. In this case, the relevancy determination unit 204 sets the relevance flag to 1 if the target area in a still image and the target area in the next still image have a small-medium or medium-large relationship. . The relevancy determination unit 204 sets 2 if the target area in a certain still image and the target area in the next still image have a small and large relationship. The relevance determination unit 204 sets −1 in the relevance flag if the target area in a certain still image and the target area in the next still image have a large-medium or medium-small relationship. The relevancy determination unit 204 sets −2 if the target area in a certain still image and the target area in the next still image have a large and small relationship. When there is no magnitude relationship between the target area in a still image and the target area in the next still image, the relevancy determination unit 204 sets 0 in the relevance flag.
The presentation method determination unit 202 changes the presentation method in the same manner assuming that there is no change in relevance when the change of the object from large to small continues. For example, the presentation method determination unit 202 gradually shortens the presentation time at the same time interval. Specifically, the presentation method determination unit 202 controls the presentation method based on the following rules.
[Rules according to the target size]
(2-1) Rules regarding presentation time
The presentation method determination unit 202 determines the presentation time of a still image pair based on the size relationship of objects included in successive still image pairs. For example, the presentation method determination unit 202 sets the presentation time of the first still image to be presented as the initial value Ts among the still image groups having a target size relationship. Then, the presentation method determination unit 202 determines the presentation time of the subsequent still image based on Ts. In addition, the presentation method determination unit 202 may set Tp as the presentation time of a still image with high visibility among a group of still images having a target size relationship. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Tp. In addition, the presentation method determination unit 202 may set, as the initial value Ts, the presentation time of the next still image in which the presentation time of the still image is equal to or less than Tq among the still image groups having a magnitude relationship. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Ts. In addition, the presentation method determination unit 202 may set the presentation time of the still image presented last among the still image groups having a magnitude relationship to Ts. The presentation method determination unit 202 may calculate the values of Ts and Tp according to the number of images to be presented from the preset presentation time of the entire slide show. When there is no magnitude relationship between the objects included in the continuous still image pair, the presentation method determination unit 202 determines the presentation time of the subsequent still image independently of the presentation time of the previous still image. For example, the presentation method determination unit 202 may set the presentation time of the subsequent still image to the initial value Ts. The presentation method determination unit 202 may set the presentation time of the subsequent still image to a random value within a specified range.
A case where a still image obtained by photographing the target B of various sizes is reproduced will be described with reference to FIG. Assume that the relevancy determination unit 204 determines the magnitude relationship between successive still images by comparing the areas between target regions determined to be the same among target regions detected from all the still images included in the slide show. . The presentation method determination unit 202 calculates the presentation time of the next still image by multiplying the presentation time of a certain still image by the relevance flag parameter a. At this time, the presentation time of the first still image 701 is the initial value Ts, the still images 701 and 702 have a small and medium relationship, the still images 702 and 703 have a medium and large relationship, and the still images 703, Assume that 704 has a relationship between large and small. At this time, since the relevance flag of the still images 701 and 702 is 1, the presentation time of the still image 702 is a × Ts (multiplication of a). Further, since the relevance flag of the still image 703 is 1, the presentation time of the still image 703 is a × a × Ts (multiplication of a). Since the relevance flag of the still images 703 and 704 is −2, the presentation time of 704 is Ts (a × a division). When the parameter a is set between 0 and 1, a still image (long shot) in which the subject B is photographed small is long, and a still image (middle shot and tight shot) in which the subject B is photographed larger is short. Presented.
Thereby, the user can understand the content of a still image with a large amount of information in which a scene other than the target B is reflected. The user can intuitively understand that the subsequent contents are a part of the previous still image. Further, the information processing apparatus 200 according to the present embodiment can generate a video in which the presentation time of consecutive images changes even if the consecutive images include the same target. For this reason, this embodiment has an effect that a slide show with a tempo that does not bore viewers can be generated.
(2-2) Rules regarding effects, BGM, and jingles
The presentation method determination unit 202 determines an effect, a BGM, and a jingle to be inserted between a pair of still images based on the size relationship of objects included in successive still image pairs. For example, when the objects included in successive still image pairs have a magnitude relationship, the presentation method determination unit 202 performs special effects (dissolve or fade) registered in advance as effects with little visual change when switching still images. Etc.). When the objects included in the continuous still image pairs are not in a size relationship, the presentation method determination unit 202 displays special effects (page turning, wipe, etc.) registered in advance as effects having a large visual change when switching still images. DVE) is inserted. Further, for example, when the objects included in the continuous still image pairs are in a size relationship, the presentation method determination unit 202 plays the same BGM during the presentation of the still image pairs. When the targets included in the continuous still image pairs are not the same, the presentation method determination unit 202 stops the BGM or switches to a different BGM when switching the still images.
In addition, the presentation method determination unit 202 may insert jingles between images that do not have a magnitude relationship. Thereby, the still image group which image | photographed the object of magnitude relation is connected smoothly, without a change of an image or a sound. Therefore, the viewer can easily understand that the still images have almost the same contents without any change. In addition, when the target included in a pair of still images does not have a large or small relationship, the image and sound change greatly, so that the viewer notices that the content has changed and concentrates on understanding the content of the slide show. Can do.
(Relevance 3. Target partial relationship)
The relevancy determination unit 204 may determine the relevance based on a partial relationship of objects represented in two still images included in the still image pair. That is, the relevancy determination unit 204 may determine whether the object represented in the two still images included in the still image pair is in a relationship between the whole and the part. The relationship 3 is the relationship determined in this way.
“In a partial relationship of objects” represents a relationship in which the objects shown in consecutive still image pairs in the target slide show are the same and are images obtained by capturing different target parts. For example, when it is desired to shoot a wide landscape, a large object, or a long object, a case where a whole image is expressed by combining a still image obtained by capturing a part of the object and performing a slide show.
When the target area in a still image and the target area in the next still image have a target partial relationship, the relevancy determination unit 204 sets 1 in the relevance flag for the relevance 3. The relevance determination unit 204 sets 0 in the relevance flag for relevance 3 when the partial relationship is not the target. The relevancy determination unit 204 can determine the target partial relationship based on a partial area (common area) that is common to the target areas determined to be the same in consecutive still images in the slide show. For example, the relevancy determination unit 204 uses one of the target areas as a template. Then, the relevancy determination unit 204 scans the other target area, detects a position with a small difference, and sets the overlapping area as a common area. The relevance determining unit 204 determines that the regions other than the common region of each target region are in a target partial relationship when both regions are equal to or larger than the specified area. Alternatively, the relevancy determination unit 204 may perform the determination based on the relative position of the target area determined to be the same from all the still images included in the slide show.
When the change of the object from the whole to the part continues, the presentation method determination unit 202 assumes that there is no change in relevance, and gradually shortens the presentation time, for example, at the same time interval. Specifically, the presentation method determination unit 202 controls the presentation method based on the following rules.
[Rules according to the target partial relationship]
(3-1) Rules regarding presentation time
The presentation method determination unit 202 determines the presentation time of a still image pair based on the target partial relationship included in the continuous still image pair. For example, the presentation method determination unit 202 sets the presentation time of the first still image to be presented as the initial value Ts among the still image groups in the target partial relationship. Then, the presentation method determination unit 202 determines the presentation time of the subsequent still image based on Ts. In addition, the presentation method determination unit 202 may set Tp as the presentation time of a still image with high visibility among the still image groups in the target partial relationship. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Tp. In addition, the presentation method determination unit 202 may set, as an initial value Ts, the presentation time of the next still image in which the presentation time of the still image is equal to or less than Tq among the still image groups having a partial relationship. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Ts. In addition, the presentation method determination unit 202 may set the presentation time of the last presented image among the still image groups having a partial relationship to Ts. The presentation method determination unit 202 may calculate the values of Ts and Tp according to the number of images to be presented from the preset presentation time of the entire slide show. When there is no partial relationship between the objects included in the continuous still image pairs, the presentation method determination unit 202 determines the presentation time of the subsequent still image independently of the presentation time of the previous still image. For example, the presentation method determination unit 202 may set the presentation time of the subsequent still image to the initial value Ts. The presentation method determination unit 202 may set the presentation time of the subsequent still image to a random value within a specified range.
A case where a still image obtained by photographing a landscape is reproduced will be described with reference to FIG. The relevancy determination unit 204 includes a partial area that is common among target areas determined to be the same among target areas detected from all still images included in a slide show, and a target area between partial still images. Is determined based on the positional relationship between and. Further, it is assumed that the presentation method determination unit 202 calculates the presentation time of the next still image by multiplying the presentation time of a certain still image by a specified parameter.
The relevancy determination unit 204 sets the presentation time of the first still image 801 to the initial value Ts. Still images 801 and 802 and 802 and 803 have a partial relationship, and still images 803 and 804 have no partial relationship. At this time, the presentation time of the first still image 801 is the initial value Ts, and the relevance flag of the still images 801 and 802 is 1, so the presentation time of the still image 802 is a × Ts. Furthermore, since the relevance flag of the still images 802 and 803 is also 1, the presentation time of the still image 803 is a 2 Ts. Since the association flag between the still images 803 and 804 is 0, the presentation method determination unit 202 returns the presentation time of the still image 804 to the initial value and sets it to Ts.
When the parameter a is set to a value between 0 and 1 and the smaller the area of the matching partial region is, the still image 801 presented for the first time for the landscape is presented longer. The The other parts are presented at a presentation time corresponding to the amount of information overlapping with the previously presented image. Thereby, the user can understand the content about the still image initially shown about the scenery. Then, the user can understand that the subsequent contents are substantially the same as the first still image. Further, the information processing apparatus 200 according to the present embodiment can generate a video in which the presentation time of consecutive images changes even if the consecutive images include the same target. For this reason, this embodiment has an effect that a slide show with a tempo that does not bore viewers can be generated.
(3-2) Rules regarding effects, BGM, and jingles
The presentation method determination unit 202 determines an effect, a BGM, and a jingle to be inserted between a pair of still images based on a target partial relationship included in a pair of still images. For example, when the targets included in the continuous still image pairs are in a partial relationship, the presentation method determination unit 202 uses special effects (dissolve and fade) that are registered in advance as effects with little visual change when switching still images. Etc.). When the target included in the continuous still image pair is not in a partial relationship, the presentation method determination unit 202 displays a special effect (page turning, wipe, etc.) registered in advance as an effect having a large visual change when switching still images. DVE) is inserted. In addition, for example, when consecutive still image pairs are in a target partial relationship, the presentation method determination unit 202 plays the same BGM during presentation of still image pairs. When the targets included in the continuous still image pairs are not the same, the presentation method determination unit 202 stops the BGM or switches to a different BGM when switching the still images. In addition, the presentation method determination unit 202 may insert jingles between images that do not have a magnitude relationship. Thereby, when a continuous still image pair has a target partial relationship, the still image pair is smoothly connected without any change in image or sound. Therefore, the viewer can easily understand that the still images have almost the same contents without any change. In addition, when a pair of still images is not in a partial relationship, the image and sound change greatly, so that the viewer can notice that the content has changed and can concentrate on understanding the content of the slide show.
(Relevance 4. Target homogeneity)
The relevancy determination unit 204 may determine relevance depending on whether or not the objects represented in the two still images included in the still image pair are of the same type. The relationship 4 is the relationship determined in this way.
“The objects are of the same type” means that main objects appearing in a pair of still images that are consecutive in the slide show are objects of the same type. The relevancy determination unit 204 determines that the relevance flag for relevance 4 is related to relevance 4 when the target area in a certain still image and the target area in the next still image have the same kind of relationship. 1 is set in the relevance flag. When the target area in a certain still image and the target area in the next still image are different from each other, the relevancy determination unit 204 sets 0 in the relevance flag for the relevance 4. Discrimination of the homogeneity of the object can be realized by a method based on machine learning based on the image data (registered data) of the object belonging to each type for which homogeneity is desired. First, the relevancy determination unit 204 extracts target image feature amounts belonging to various types from registered data. The relevancy determination unit 204 may use a global feature such as a color histogram or an edge histogram as the image feature amount. The relevancy determination unit 204 may use local feature amounts such as HoG (Histograms of Oriented Gradients) and SIFT as the image feature amounts. The relevancy determination unit 204 may perform learning by using a global feature, such as SVM (Signal Value Decomposition), a neural network, or a GMM (Gaussian Mixture Model). Alternatively, the relevancy determination unit 204 may perform learning after converting the feature amount space from the local feature amount like BoW (Bag of Words). When determining the homogeneity of the target areas in each still image included in the slide show, the relevance determining unit 204 determines between the image feature amount of each target area and various types of models obtained as a result of learning. Each seeks similarity. Then, the relevancy determination unit 204 determines that the target region is the closest model type that has obtained a similarity equal to or greater than a specified value. The relevancy determination unit 204 determines that the target areas determined to be the same type are the same type. The relevancy determination unit 204 may determine the homogeneity by a method other than the above.
When three images including the same type of target are consecutive, the presentation method determination unit 202 assumes that there is no change in relevance and, for example, gradually shortens the presentation time at the same time interval. Specifically, the presentation method determination unit 202 controls the presentation method based on the following rules.
[Rules according to target homogeneity]
(4-1) Rules regarding presentation time
The presentation method determination unit 202 determines the presentation time of a still image pair based on the homogeneity of objects included in successive still image pairs. For example, the presentation method determination unit 202 sets the presentation time of the still image presented first among the still image groups including the same type of target to the initial value Ts. Then, the presentation method determination unit 202 determines the presentation time of the subsequent still image based on Ts. In addition, the presentation method determination unit 202 may set Tp as the presentation time of a still image with high visibility among a group of still images including the same type of target. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Tp. In addition, the presentation method determination unit 202 may set the presentation time of the next still image in which the presentation time of the still image is equal to or less than Tq among the still image group including the same type of target as the initial value Ts. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Ts. In addition, the presentation method determination unit 202 may set the presentation time of the last presented image among the still image groups including the same type of target to Ts. The presentation method determination unit 202 may calculate the values of Ts and Tp according to the number of images to be presented from the preset presentation time of the entire slide show. When the objects included in the continuous still image pairs are not of the same type, the presentation method determination unit 202 determines the presentation time of the subsequent still image independently of the presentation time of the previous still image. The presentation method determination unit 202 may set the presentation time of subsequent still images to, for example, the initial value Ts. The presentation method determination unit 202 may set the presentation time of the subsequent still image to a random value within a specified range.
A case where a still image obtained by photographing a flower is reproduced will be described with reference to FIG. Assume that the relevancy determination unit 204 determines the homogeneity between consecutive still images by a method based on machine learning. Also, the presentation method determination unit 202 calculates the presentation time of the next still image by multiplying the presentation time of a certain still image by the parameter for the relevance flag. The presentation method determination unit 202 sets the presentation time of the first still image 901 to the initial value Ts. The still images 901 and 902 and the still images 902 and 903 have the same type of relationship, and the still images 903 and 904 have a different type of relationship. At this time, since the relevance flag of the first still images 901 and 902 is 1, the presentation time of the still image 902 is a × Ts. Furthermore, since the relevance flag of the still images 902 and 903 is 1, the presentation time of the still image 903 is a 2 Ts. Since the relevance flag of the still images 903 and 904 is 0, the presentation method determination unit 202 returns the presentation time of the still image 904 to the initial value and sets it to Ts. When the parameter a is set between 0 and 1, a still image 901 presented for the first time among the still images including plants is presented for a long time. Then, the subsequent still image is presented in a presentation time that is shorter as the distance from 901 increases. Thereby, the user can understand that the image content is a plant from the initially presented still image. Then, the user can understand that the contents of the subsequent still images are almost the same.
Further, the information processing apparatus 200 according to the present embodiment can generate a video in which the presentation time of consecutive images changes even if the consecutive images include the same target. Therefore, this embodiment has an effect that it is possible to generate a slide show with a tempo that does not bore the viewer (this embodiment reproduces images of a plurality of flowers taken in a flower field in order of the same kind of subject in order. Can express that there were many subjects of this type).
(4-2) Rules regarding effects, BGM, and jingles
The presentation method determination unit 202 determines an effect, a BGM, and a jingle to be inserted between a pair of still images based on the homogeneity of objects included in successive still image pairs. For example, when the target included in the pair of still images is the same type, the presentation method determination unit 202 uses special effects (such as dissolves and fades) registered in advance as effects with little visual change when switching still images. ) Is inserted. The presentation method determination unit 202, when the target included in the continuous still image pair is different, special effects (such as page turning and wipe) registered in advance as effects having a large visual change when switching still images. DVE) is inserted. In addition, for example, when continuous still image pairs are of the same type, the presentation method determination unit 202 plays the same BGM during presentation of the still image pairs. When continuous still image pairs are heterogeneous, the presentation method determination unit 202 stops BGM or switches to a different BGM when switching still images. In addition, the presentation method determination unit 202 may insert jingles between different types of still images. Thereby, when the object contained in the continuous still image pair is the same type, the still image pair is smoothly connected without any change in image or sound. Therefore, the viewer can easily understand that the still images have almost the same contents without any change. In addition, when the pairs of still images are different, the image and sound change greatly, so that the viewer can notice that the content has changed and can concentrate on understanding the content of the slide show.
(Relevance 5. Sameness of shooting location)
The relevancy determination unit 204 may determine the relevance based on the commonality of the shooting locations of two still images included in the still image pair. The relationship 5 is the relationship determined in this way.
“The shooting locations are the same” means that the locations where the still image pairs that were consecutive in the slide show were shot are the same. The relevancy determination unit 204 sets 1 in the relevance flag for relevance 5 when the shooting location of a certain still image is the same as the shooting location of the next still image. When the shooting location of a certain still image is different from the shooting location of the next still image, the relevancy determination unit 204 sets 0 for the relevance flag for relevance 5. The relevance determination unit 204 can determine the identity of the shooting location based on the similarity of an area (background area) other than the target area in the still image. For example, the relationship determination unit 204 may separate the target area and the background area from the still image. Then, the relevancy determination unit 204 may determine the same shooting location when the image feature values extracted from the background region are similar. The relevancy determination unit 204 may determine the identity of the shooting location by a method other than the above. The relevancy determination unit 204 may determine the similarity of the shooting location and the similarity of the background between still images consecutive in the slide show. Alternatively, the relevancy determination unit 204 may determine the identity of the shooting location based on the identity of the background areas in all the still images included in the slide show. The relevancy determination unit 204 may determine the identity of the shooting location by combining the shooting location as meta information and the GPS as sensor information in addition to the image information.
For example, when three images taken at the same shooting location are consecutive, the presentation method determination unit 202 assumes that there is no change in relevance and, for example, gradually reduces the presentation time at the same time interval. Specifically, the presentation method determination unit 202 controls the presentation method based on the following rules.
[Rules according to the identity of the shooting location]
(5-1) Rules regarding presentation time
The presentation method determination unit 202 determines the presentation time of still image pairs based on the identity of the shooting locations of consecutive still image pairs. For example, the presentation method determination unit 202 sets the presentation time of a still image presented first among the still images captured at the same place as the initial value Ts. Then, the presentation method determination unit 202 determines the presentation time of the subsequent still image based on Ts. In addition, the presentation method determination unit 202 may set Tp as the presentation time of a still image with high visibility among the still images captured at the same place. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Tp. In addition, the presentation method determination unit 202 may set the presentation time of the next still image in which the presentation time of the still image is equal to or less than Tq in the still image group captured at the same place as the initial value Ts. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Ts. In addition, the presentation method determination unit 202 may set the presentation time of the last presented image in the group of still images captured at the same place as Ts. The presentation method determination unit 202 may calculate values of Ts and Tp according to the number of images to be presented from preset presentation time of the entire slide show. When successive still image pairs are taken at different locations, the presentation method determination unit 202 determines the presentation time of the subsequent still image independently of the presentation time of the previous still image. For example, the presentation method determination unit 202 may set the presentation time of the subsequent still image to the initial value Ts. The presentation method determination unit 202 may set the presentation time of the subsequent still image to a random value within a specified range.
(5-2) Rules regarding effects, BGM, and jingles
The presentation method determination unit 202 determines an effect, a BGM, and a jingle to be inserted between a pair of still images based on the identity of the shooting locations of consecutive still image pairs. For example, when consecutive still image pairs are shot at the same place, the presentation method determination unit 202 uses special effects (such as dissolves and fades) registered in advance as effects with little visual change when switching still images. Insert. When consecutive pairs of still images are taken at different locations, the presentation method determination unit 202 displays special effects (DVE such as page turning and wipe) that are registered in advance as effects having a large visual change when switching still images. ) Is inserted. Further, for example, when consecutive still image pairs are photographed at the same place, the presentation method determination unit 202 plays the same BGM during presentation of the still image pairs. When consecutive still image pairs are captured at different locations, the presentation method determination unit 202 stops BGM or switches to a different BGM when switching still images. In addition, the presentation method determination unit 202 may insert a jingle between still images taken at different places. Thereby, when continuous still image pairs are taken at the same place, the still image pairs are smoothly connected without any change in image or sound. Therefore, the viewer can easily understand that the still images have almost the same contents without any change. Also, when the image is taken at a different place, the image and sound change greatly, so that the viewer can notice that there is a change in the content and can concentrate on understanding the content of the slide show.
(Relevance 6. Identity of shooting times)
The relevancy determination unit 204 may determine the relevance based on the commonality of the shooting time zones of two still images included in the still image pair. The relationship 6 is the relationship determined in this way.
“The shooting time zone is the same” means that the time zone in which a pair of still images is taken in the slide show is the same. The relevancy determination unit 204 sets 1 in the relevance flag for relevance 6 when the shooting time zone of a certain still image is the same as the shooting time zone of the next still image. The relevancy determination unit 204 sets a relevance flag for relevance 6 to 0 when the shooting time zone of a certain still image is different from the shooting time zone of the next still image. The relevancy determination unit 204 can determine the identity of the shooting time period based on the color information of the background area in the still image. For example, the relevance determination unit 204 divides a day into a plurality of time zones, and holds the statistics of the color histogram of sunlight in each time zone. Then, when the background area of the still image includes a partial area close to a statistic of any time zone, the relevancy determination unit 204 determines that the still image is captured during that time zone. The relevancy determination unit 204 estimates the shooting time period of each still image. Then, the relevance determination unit 204 determines that the shooting time zones are the same when the estimated time is the same. The relevancy determination unit 204 may determine the identity of the shooting time period by a method other than the above. The relevance determination unit 204 may determine the identity of the shooting time zones based on the similarity of the shooting time zones between consecutive still images in the slide show. Alternatively, the relevancy determination unit 204 may determine the identity of the shooting time zone based on the identity of the shooting time zones in all the still images included in the slide show. The relevancy determination unit 204 may determine the identity of the shooting time period in combination with the shooting time that is meta information in addition to the image information.
The presentation method determination unit 202 assumes that there is no change in relevance when three images taken in the same shooting time period are consecutive, and for example, gradually reduces the presentation time at the same time interval. Specifically, the presentation method determination unit 202 controls the presentation method based on the following rules.
[Rules according to the identity of the shooting period]
(6-1) Rules regarding presentation time
The presentation method determination unit 202 determines the presentation time of still image pairs based on the identity of the shooting time zones of consecutive still image pairs. For example, the presentation method determination unit 202 sets the presentation time of the first still image to be presented as the initial value Ts in the group of still images taken in the same time zone. Then, the presentation method determination unit 202 determines the presentation time of the subsequent still image based on Ts. In addition, the presentation method determination unit 202 may set the presentation time of a still image with high visibility among the still image groups taken in the same time zone as Tp. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Tp. In addition, the presentation method determination unit 202 may set the presentation time of the next still image in which the presentation time of the still image is equal to or less than Tq among the still image groups photographed in the same time zone as the initial value Ts. And the presentation method determination part 202 may determine the presentation time of a subsequent still image on the basis of Ts. In addition, the presentation method determination unit 202 may set the presentation time of the image to be presented last among the still image groups captured in the same time zone as Ts. The presentation method determination unit 202 may calculate the values of Ts and Tp according to the number of images to be presented from the preset presentation time of the entire slide show. When successive still image pairs are photographed at different time periods, the presentation method determination unit 202 determines the presentation time of the subsequent still image independently of the presentation time of the previous still image. The presentation method determination unit 202 may set the presentation time of subsequent still images to, for example, the initial value Ts. The presentation method determination unit 202 may set the presentation time of the subsequent still image to a random value within a specified range.
(6-2) Rules regarding effects, BGM, and jingles
The presentation method determination unit 202 determines an effect, a BGM, and a jingle to be inserted between a pair of still images based on the identity of the shooting time zones of consecutive still image pairs. For example, when a pair of still images is shot in the same time zone, the presentation method determination unit 202 can use special effects (dissolve or fade) that are registered in advance as effects with little visual change when switching still images. Etc.). The presentation method determination unit 202, when a pair of still images is shot at different time zones, special effects (such as page turning and wipe) registered in advance as effects having a large visual change when switching still images. DVE) is inserted. Also, for example, when consecutive still image pairs are captured in the same time zone, the presentation method determination unit 202 plays the same BGM during presentation of still image pairs. When consecutive pairs of still images are captured in different time zones, the presentation method determination unit 202 stops BGM or switches to a different BGM when switching still images. In addition, the presentation method determination unit 202 may insert jingles between still images in different time zones. Thus, when consecutive still image pairs are taken in the same time zone, the still image pairs are smoothly connected without any change in image or sound. Therefore, the viewer can easily understand that the still images have almost the same contents without any change. In addition, when consecutive pairs of still images are taken at different times, the image and sound change greatly, so that the viewer can notice that the content has changed and concentrate on understanding the content of the slide show. it can. Any one of the above-described rules may be applied as a presentation rule to successive still image pairs. Continuous still image pairs may be used in combination with a plurality of rules as presentation rules. The slide show generation unit 203 generates a slide show based on the presentation method information input from the presentation method determination unit 202 and the image information input from the image input unit 210.
[Fourth Embodiment]
As shown in the second embodiment, the relevancy determination unit 204 may group all target areas detected from all still images included in the slide show based on similarity. Then, the relevancy determination unit 204 may determine that the target areas are the same when the target areas detected from the adjacent still image pairs belong to the same group.
In the present embodiment, a processing flow when the relevancy determination unit 204 groups still images as described above will be described.
The detection method of the target area is divided into a detection method for detecting a specific target registered in advance and a detection method for detecting a general target that is not registered. When detecting the specific target, the relevance determination unit 204 may use the registered image data of each target as a template. Then, the relevancy determination unit 204 may scan the input image with templates converted into various resolutions. Then, the relevancy determination unit 204 may detect a region having a small difference in pixel values at the same position as the template as a corresponding target region.
The operation of this embodiment will be described in detail with reference to the flowchart of FIG. Also in this embodiment, a case where the still image shown in FIG. 5 is reproduced as a slide show will be described as an example. The relevancy determination unit 204 determines the identity, the magnitude relationship, and the partial relationship as the relevance type. The presentation rule determination method is the same as in the first embodiment.
First, the image input unit 210 inputs image information of all still images to the relevancy determination unit 204 (step S1001).
The relevancy determination unit 204 extracts image feature amounts from all still images. Then, the relevancy determination unit 204 groups the still images obtained by photographing the same object based on the similarity of the image feature amounts (step S1003). In the example of FIG. 5, still images 501, 502, 503, and 504 are classified into group A, and still images 505, 506, 507, 508, 509, and 510 are classified into group B, 511, 512, and 513.
The relevancy determination unit 204 determines the magnitude relationship and the partial relationship between still images belonging to the same group (S1005). The relevancy determination unit 204 extracts local feature points such as SIFT from all still images and takes corresponding points with each other. Then, for group A, it can be seen that still images 502, 503, and 504 are included in the still image 501. It can be seen that the common areas of the still images 502, 503, and 504 are small. Therefore, the relevancy determination unit 204 determines that there is a magnitude relationship between the still image 501 and the still image 502. Then, the relationship comparison unit 201 determines that there is a partial relationship between the still image 502 and the still image 503, and the still image 503 and the still image 504.
The relevancy determination unit 204 compares local feature amounts in the same manner, so that it can be understood that the still images 506, 507, 508, 509, and 510 are included in 505 for the group B. Further, it can be seen that the still images 506 and 507 are included in the still image 513. Therefore, it is determined that the still image 505 has a large relationship, the still images 511, 512, and 513 have a medium relationship, and the still images 506 and 507 have a small relationship. The still images 506 and 507 are determined to be still images obtained by photographing a part of the same target. The still images 511, 512, and 513 are also determined to be still images obtained by capturing a part of the same object. For group C, since the shared area occupies the entire screen, it is determined that there is no magnitude relationship / partial relationship. As a result, the presentation method determination unit 202 determines the presentation time length 1101 and the effect 1102 as shown in FIG. 11 (S1007). Finally, the slide show generating unit 203 generates a slide show using the determined presentation method (S1009).
In the present embodiment, in addition to the effect of the second embodiment, there is an effect that the relationship between groups can be clearly expressed in a slide show.
[Other Embodiments]
As mentioned above, although embodiment of this invention was explained in full detail, the system or apparatus which combined the separate characteristic contained in each embodiment how was included in the category of this invention.
In addition, the present invention may be applied to a system composed of a plurality of devices, or may be applied to a single device. Furthermore, the present invention can also be applied to a case where an information processing program that implements the functions of the embodiments is supplied directly or remotely to a system or apparatus. Therefore, in order to realize the functions of the present invention on a computer, a program installed in the computer, a medium storing the program, and a WWW (World Wide Web) server that downloads the program are also included in the scope of the present invention. .
For example, each of the information processing apparatus 100 and the information processing apparatus 200 can be realized by a computer and a program for controlling the computer, dedicated hardware, or a combination of the computer and the program for controlling the computer and dedicated hardware. .
The relevance comparison unit 101, the presentation method determination unit 102, the slide show generation unit 103, the relevance comparison unit 201, the presentation method determination unit 202, the slide show generation unit 203, and the relevance determination unit 204 are, for example, from a recording medium that stores a program. It can be realized by a dedicated program for realizing the function of each unit read into the memory and a processor for executing the program. Alternatively, some or all of the relevance comparison unit 101, the presentation method determination unit 102, the slide show generation unit 103, the relevance comparison unit 201, the presentation method determination unit 202, the slide show generation unit 203, and the relevance determination unit 204 may be replaced with each part. It can also be realized by a dedicated circuit for realizing the function.
The present invention has been described above with reference to the embodiments, but the present invention is not limited to the above embodiments. Various changes that can be understood by those skilled in the art can be made to the configuration and details of the present invention within the scope of the present invention.
While the present invention has been described with reference to the embodiments, the present invention is not limited to the above embodiments. Various changes that can be understood by those skilled in the art can be made to the configuration and details of the present invention within the scope of the present invention.
This application claims the priority on the basis of Japanese application Japanese Patent Application No. 2011-107103 for which it applied on May 12, 2011, and takes in those the indications of all here.

Claims (17)

  1.  少なくとも3つの静止画像を含む静止画像群において、前記静止画像群に含まれる第1静止画像対の静止画像同士の第1関連性と第2静止画像対の静止画像同士の第2関連性とを比較する比較手段と、
     前記第1関連性から第2関連性への変化に基づいて、前記第1静止画像対の提示方法を特定する第1情報から、前記第2静止画像対の提示方法を特定する第2情報を決定する提示方法決定手段と、
     前記第1情報により特定された提示方法及び前記第2情報により特定された提示方法に基づいて、前記第1静止画像対および前記第2静止画像対を含むスライドショーを生成する生成手段と、
     を備えたことを特徴とする情報処理装置。
    In a still image group including at least three still images, the first relationship between the still images of the first still image pair and the second relationship between the still images of the second still image pair included in the still image group. A comparison means for comparing;
    Based on the change from the first relevance to the second relevance, the second information specifying the presentation method of the second still image pair from the first information specifying the presentation method of the first still image pair. A presentation method determining means for determining;
    Generating means for generating a slide show including the first still image pair and the second still image pair based on the presentation method specified by the first information and the presentation method specified by the second information;
    An information processing apparatus comprising:
  2.  前記第1静止画像対および前記第2静止画像対は、それぞれ連続する静止画像の対であることを特徴とする請求項1に記載の情報処理装置。 The information processing apparatus according to claim 1, wherein the first still image pair and the second still image pair are pairs of continuous still images.
  3.  前記静止画像群は連続する第1、第2、第3静止画像を含み、前記第1静止画像対は、前記第1静止画像と前記第2静止画像の対であり、前記第2静止画像対は、前記第2静止画像と前記第3静止画像の対であることを特徴とする請求項2に記載の情報処理装置。 The still image group includes continuous first, second, and third still images, and the first still image pair is a pair of the first still image and the second still image, and the second still image pair. The information processing apparatus according to claim 2, wherein the information processing device is a pair of the second still image and the third still image.
  4.  前記第1関連性を、
     前記第1静止画像対に含まれる2つの静止画像に表わされた対象の共通性によって決定し、
     前記第2関連性を、
     前記第2静止画像対に含まれる2つの静止画像に表わされた対象の共通性によって決定する関連性決定手段をさらに備えたことを特徴とする請求項1、2または3に記載の情報処理装置。
    The first relationship is
    Determined by the commonality of objects represented in two still images included in the first still image pair,
    The second relationship is
    The information processing according to claim 1, 2, or 3, further comprising a relevance determining unit that determines the commonality of objects represented in two still images included in the second still image pair. apparatus.
  5.  前記第1関連性を、
     前記第1静止画像対に含まれる2つの静止画像に表わされた対象が同一か否かによって決定し、
     前記第2関連性を、
     前記第2静止画像対に含まれる2つの静止画像に表わされた対象が同一か否かによって決定する関連性決定手段をさらに備えたことを特徴とする請求項4に記載の情報処理装置。
    The first relationship is
    Determining whether the objects represented in the two still images included in the first still image pair are the same;
    The second relationship is
    5. The information processing apparatus according to claim 4, further comprising a relevance determining unit that determines whether or not the objects represented in two still images included in the second still image pair are the same.
  6.  前記第1関連性を、
     前記第1静止画像対に含まれる2つの静止画像に表わされた対象が同種か否かによって決定し、
     前記第2関連性を、
     前記第2静止画像対に含まれる2つの静止画像に表わされた対象が同種か否かによって決定する関連性決定手段をさらに備えたことを特徴とする請求項4に記載の情報処理装置。
    The first relationship is
    Determining whether the objects represented in the two still images included in the first still image pair are of the same type,
    The second relationship is
    5. The information processing apparatus according to claim 4, further comprising a relevancy determination unit that determines whether or not the objects represented in two still images included in the second still image pair are of the same type.
  7.  前記第1関連性を、
     前記第1静止画像対に含まれる2つの静止画像に表わされた対象が全体と部分との関係にあるかによって決定し、
     前記第2関連性を、
     前記第2静止画像対に含まれる2つの静止画像に表わされた対象が全体と部分との関係にあるかによって決定する関連性決定手段をさらに備えたことを特徴とする請求項4に記載の情報処理装置。
    The first relationship is
    Determining whether the object represented by the two still images included in the first still image pair has a relationship between the whole and a part;
    The second relationship is
    5. The relevance determining means for determining whether a target represented by two still images included in the second still image pair is in a relationship between a whole and a part. Information processing device.
  8.  前記第1関連性を、
     前記第1静止画像対に含まれる2つの静止画像の撮影場所または撮影時間の共通性によって決定し、
     前記第2関連性を、
     前記第2静止画像対に含まれる2つの静止画像の撮影場所または撮影時間の共通性によって決定する関連性決定手段をさらに備えたことを特徴とする請求項4乃至7のいずれか1項に記載の情報処理装置。
    The first relationship is
    It is determined by the common place of shooting location or shooting time of two still images included in the first still image pair,
    The second relationship is
    8. The apparatus according to claim 4, further comprising a relevance determining unit that determines a shooting location or a shooting time of two still images included in the second still image pair. Information processing device.
  9.  前記第1関連性および前記第2関連性を、
     前記静止画像群に含まれる複数の静止画像から導き出される特徴量の共通性によって決定する関連性決定手段をさらに備えたことを特徴とする請求項4乃至7のいずれか1項に記載の情報処理装置。
    The first relationship and the second relationship,
    The information processing according to any one of claims 4 to 7, further comprising a relevance determining unit that determines the commonality of feature quantities derived from a plurality of still images included in the still image group. apparatus.
  10.  前記関連性決定手段は、
     前記第1関連性を、
     前記第1静止画像対に含まれる2つの静止画像の画像特徴量の共通性によって決定し、
     前記第2関連性を、
     前記第2静止画像対に含まれる2つの静止画像の画像特徴量の共通性によって決定することを特徴とする請求項9に記載の情報処理装置。
    The relevance determining means includes
    The first relationship is
    Determined by the commonality of image feature quantities of two still images included in the first still image pair,
    The second relationship is
    The information processing apparatus according to claim 9, wherein the information processing apparatus is determined based on a commonality of image feature amounts of two still images included in the second still image pair.
  11.  前記関連性決定手段は、
     前記第1関連性を、
     前記第1静止画像対に含まれる2つの静止画像において、あらかじめ登録された特定対象が含まれている領域の共通性によって決定し、
     前記第2関連性を、
     前記第2静止画像対に含まれる2つの静止画像において、あらかじめ登録された特定対象が含まれている領域の共通性によって決定することを特徴とする請求項9に記載の情報処理装置。
    The relevance determining means includes
    The first relationship is
    In the two still images included in the first still image pair, it is determined by the commonality of the regions including the specific target registered in advance,
    The second relationship is
    10. The information processing apparatus according to claim 9, wherein the information processing apparatus determines the two still images included in the second still image pair based on a commonality of regions including a specific target registered in advance.
  12.  前記関連性決定手段は、
     前記第1関連性を、
     前記第1静止画像対に含まれる2つの静止画像において、あらかじめ登録された特定対象が含まれている領域の大きさの変化によって決定し、
     前記第2関連性を、
     前記第2静止画像対に含まれる2つの静止画像において、あらかじめ登録された特定対象が含まれている領域の大きさの変化によって決定することを特徴とする請求項9に記載の情報処理装置。
    The relevance determining means includes
    The first relationship is
    In the two still images included in the first still image pair, it is determined by a change in the size of a region including a specific object registered in advance.
    The second relationship is
    The information processing apparatus according to claim 9, wherein the two still images included in the second still image pair are determined by a change in size of a region including a specific target registered in advance.
  13.  前記提示方法決定手段は、前記第2情報として、
     前記第2静止画像対に含まれる2つの静止画像の提示時間の差分、それら静止画像の切り替え時のエフェクトまたはジングル、それらの静止画像を提示する際の背景音楽を決定することを特徴とする請求項1乃至12のいずれか1項に記載の情報処理装置。
    The presenting method determining means, as the second information,
    The difference between presentation times of two still images included in the second still image pair, an effect or jingle when switching the still images, and background music when presenting the still images are determined. Item 13. The information processing apparatus according to any one of Items 1 to 12.
  14.  前記提示方法決定手段は、
     前記第1関連性と第2関連性との間に変化が無い場合、前記第1静止画像対における提示方法の変化と同様に、前記第2静止画像対における提示方法が変化するように前記第1情報から前記第2情報を決定することを特徴とする請求項1乃至13のいずれか1項に記載の情報処理装置。
    The presenting method determining means includes
    If there is no change between the first association and the second association, the presentation method in the second still image pair is changed in the same manner as the presentation method in the first still image pair. The information processing apparatus according to claim 1, wherein the second information is determined from one information.
  15.  前記提示方法決定手段は、
     前記第1関連性と第2関連性との間に変化が無い場合、前記第1静止画像対における提示時間の変化と同様に、前記第2静止画像対において提示時間が変化するように、前記第1情報から前記第2情報を決定することを特徴とする請求項1乃至14のいずれか1項に記載の情報処理装置。
    The presenting method determining means includes
    When there is no change between the first relevance and the second relevance, the presentation time is changed in the second still image pair as in the change in the presentation time in the first still image pair. The information processing apparatus according to claim 1, wherein the second information is determined from the first information.
  16.  少なくとも3つの静止画像を含む静止画像群において、前記静止画像群に含まれる第1静止画像対の静止画像同士の第1関連性と第2静止画像対の静止画像同士の第2関連性とを比較し、
     前記第1関連性から第2関連性への変化に基づいて、前記第1静止画像対の提示方法を特定する第1情報から、前記第2静止画像対の提示方法を特定する第2情報を決定し、
     前記第1情報及び前記第2情報により特定された提示方法に基づいて、前記第1静止画像対および前記第2静止画像対を含むスライドショーを生成する
     ことを特徴とする情報処理方法。
    In a still image group including at least three still images, the first relationship between the still images of the first still image pair and the second relationship between the still images of the second still image pair included in the still image group. Compare and
    Based on the change from the first relevance to the second relevance, the second information specifying the presentation method of the second still image pair from the first information specifying the presentation method of the first still image pair. Decide
    An information processing method, comprising: generating a slide show including the first still image pair and the second still image pair based on the presentation method specified by the first information and the second information.
  17.  コンピュータを、
     少なくとも3つの静止画像を含む静止画像群において、前記静止画像群に含まれる第1静止画像対の静止画像同士の第1関連性と第2静止画像対の静止画像同士の第2関連性とを比較する比較手段と、
     前記第1関連性から第2関連性への変化に基づいて、前記第1静止画像対の提示方法を特定する第1情報から、前記第2静止画像対の提示方法を特定する第2情報を決定する提示方法決定手段と、
     前記第1情報及び前記第2情報により特定された提示方法に基づいて、前記第1静止画像対および前記第2静止画像対を含むスライドショーを生成する生成手段と
     して動作させることを特徴とする情報処理プログラムを記憶する記録媒体。
    Computer
    In a still image group including at least three still images, the first relationship between the still images of the first still image pair and the second relationship between the still images of the second still image pair included in the still image group. A comparison means for comparing;
    Based on the change from the first relevance to the second relevance, the second information specifying the presentation method of the second still image pair from the first information specifying the presentation method of the first still image pair. A presentation method determining means for determining;
    Based on the presentation method specified by the first information and the second information, the display unit is operated as a generation unit that generates a slide show including the first still image pair and the second still image pair. A recording medium for storing an information processing program.
PCT/JP2012/061788 2011-05-12 2012-04-27 Information processing device, information processing method, and information processing program WO2012153744A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2011-107103 2011-05-12
JP2011107103A JP2014170979A (en) 2011-05-12 2011-05-12 Information processing apparatus, information processing method, and information processing program

Publications (1)

Publication Number Publication Date
WO2012153744A1 true WO2012153744A1 (en) 2012-11-15

Family

ID=47139222

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2012/061788 WO2012153744A1 (en) 2011-05-12 2012-04-27 Information processing device, information processing method, and information processing program

Country Status (2)

Country Link
JP (1) JP2014170979A (en)
WO (1) WO2012153744A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2015035727A (en) * 2013-08-09 2015-02-19 株式会社リコー Display system, information terminal, display device, reproduction control program, reproduction program, and reproduction control method
CN111083361A (en) * 2019-12-11 2020-04-28 维沃移动通信有限公司 Image acquisition method and electronic equipment
CN114886417A (en) * 2022-05-10 2022-08-12 南京布尔特医疗技术发展有限公司 Intelligent safety nursing monitoring system and method

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6532283B2 (en) * 2015-05-12 2019-06-19 キヤノン株式会社 IMAGE PROCESSING APPARATUS, IMAGE PROCESSING METHOD, AND PROGRAM
JP6729678B2 (en) 2016-02-26 2020-07-22 日本電気株式会社 Information processing apparatus, suspect information generation method and program

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005348371A (en) * 2004-06-07 2005-12-15 Fuji Photo Film Co Ltd Electronic album display system, electronic album display method, and electronic album display program
JP2006129453A (en) * 2004-09-29 2006-05-18 Nikon Corp Image reproduction apparatus and image reproduction program
JP2006261877A (en) * 2005-03-16 2006-09-28 Casio Comput Co Ltd Image reproducing apparatus and program
JP2008061032A (en) * 2006-08-31 2008-03-13 Sony Corp Image reproducing apparatus and image reproducing method, and computer program
WO2008133046A1 (en) * 2007-04-13 2008-11-06 Nec Corporation Photograph grouping device, photograph grouping method and photograph grouping program
JP2010021819A (en) * 2008-07-11 2010-01-28 Casio Comput Co Ltd Image display apparatus, image display method, and program
JP2010206508A (en) * 2009-03-03 2010-09-16 Olympus Imaging Corp Display device, imaging device, and program for display device

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005348371A (en) * 2004-06-07 2005-12-15 Fuji Photo Film Co Ltd Electronic album display system, electronic album display method, and electronic album display program
JP2006129453A (en) * 2004-09-29 2006-05-18 Nikon Corp Image reproduction apparatus and image reproduction program
JP2006261877A (en) * 2005-03-16 2006-09-28 Casio Comput Co Ltd Image reproducing apparatus and program
JP2008061032A (en) * 2006-08-31 2008-03-13 Sony Corp Image reproducing apparatus and image reproducing method, and computer program
WO2008133046A1 (en) * 2007-04-13 2008-11-06 Nec Corporation Photograph grouping device, photograph grouping method and photograph grouping program
JP2010021819A (en) * 2008-07-11 2010-01-28 Casio Comput Co Ltd Image display apparatus, image display method, and program
JP2010206508A (en) * 2009-03-03 2010-09-16 Olympus Imaging Corp Display device, imaging device, and program for display device

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2015035727A (en) * 2013-08-09 2015-02-19 株式会社リコー Display system, information terminal, display device, reproduction control program, reproduction program, and reproduction control method
CN111083361A (en) * 2019-12-11 2020-04-28 维沃移动通信有限公司 Image acquisition method and electronic equipment
CN114886417A (en) * 2022-05-10 2022-08-12 南京布尔特医疗技术发展有限公司 Intelligent safety nursing monitoring system and method
CN114886417B (en) * 2022-05-10 2023-09-22 南京布尔特医疗技术发展有限公司 Intelligent safety nursing monitoring system and method

Also Published As

Publication number Publication date
JP2014170979A (en) 2014-09-18

Similar Documents

Publication Publication Date Title
US11321385B2 (en) Visualization of image themes based on image content
US11132578B2 (en) System and method for creating navigable views
US11094131B2 (en) Augmented reality apparatus and method
KR101605983B1 (en) Image recomposition using face detection
US8548249B2 (en) Information processing apparatus, information processing method, and program
Goferman et al. Context-aware saliency detection
US8908976B2 (en) Image information processing apparatus
JP2017508192A (en) Object selection and tracking for display segmentation and video frame clustering
WO2012153744A1 (en) Information processing device, information processing method, and information processing program
CN111638784B (en) Facial expression interaction method, interaction device and computer storage medium
CN111881755B (en) Method and device for cutting video frame sequence
CN111491187A (en) Video recommendation method, device, equipment and storage medium
CN110418148B (en) Video generation method, video generation device and readable storage medium
JP5776471B2 (en) Image display system
US20160140748A1 (en) Automated animation for presentation of images
JP5850188B2 (en) Image display system
WO2012153747A1 (en) Information processing device, information processing method, and information processing program
JP2009289210A (en) Device and method for recognizing important object and program thereof
CN109791556B (en) Method for automatically creating collage from mobile video
Huang et al. Automatic detection of object of interest and tracking in active video
KR101573482B1 (en) Apparatus for inserting advertisement using frame clustering and method thereof
WO2023039865A1 (en) Image processing method, video processing method, training method, device, program product, and storage medium
Shankar et al. A novel semantics and feature preserving perspective for content aware image retargeting
CN116980695A (en) Video processing method, device, equipment and storage medium
Chapdelaine et al. Designing caption production rules based on face, text, and motion detection

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12782029

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12782029

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP