US9106872B2 - Image processing apparatus, image processing method, and program - Google Patents
Image processing apparatus, image processing method, and program Download PDFInfo
- Publication number
- US9106872B2 US9106872B2 US12/565,499 US56549909A US9106872B2 US 9106872 B2 US9106872 B2 US 9106872B2 US 56549909 A US56549909 A US 56549909A US 9106872 B2 US9106872 B2 US 9106872B2
- Authority
- US
- United States
- Prior art keywords
- image
- frame
- face
- unit
- moving picture
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
- 238000012545 processing Methods 0.000 title claims abstract description 98
- 238000003672 processing method Methods 0.000 title claims description 13
- 230000009466 transformation Effects 0.000 claims abstract description 264
- 238000003860 storage Methods 0.000 claims abstract description 137
- 239000002131 composite material Substances 0.000 claims abstract description 42
- 238000000034 method Methods 0.000 claims description 382
- PXFBZOLANLWPMH-UHFFFAOYSA-N 16-Epiaffinine Natural products C1C(C2=CC=CC=C2N2)=C2C(=O)CC2C(=CC)CN(C)C1C2CO PXFBZOLANLWPMH-UHFFFAOYSA-N 0.000 claims description 254
- 230000008921 facial expression Effects 0.000 claims description 119
- 230000001131 transforming effect Effects 0.000 abstract description 38
- 230000002123 temporal effect Effects 0.000 abstract description 16
- 230000008569 process Effects 0.000 description 224
- 238000005034 decoration Methods 0.000 description 160
- 230000000875 corresponding effect Effects 0.000 description 152
- 238000010586 diagram Methods 0.000 description 133
- 239000013598 vector Substances 0.000 description 133
- 230000003287 optical effect Effects 0.000 description 92
- 230000033001 locomotion Effects 0.000 description 85
- 238000004364 calculation method Methods 0.000 description 71
- 239000000872 buffer Substances 0.000 description 69
- 238000001514 detection method Methods 0.000 description 60
- 239000011159 matrix material Substances 0.000 description 41
- 230000002596 correlated effect Effects 0.000 description 40
- 238000003384 imaging method Methods 0.000 description 37
- 230000014509 gene expression Effects 0.000 description 34
- 210000000887 face Anatomy 0.000 description 28
- 230000007704 transition Effects 0.000 description 15
- 230000000694 effects Effects 0.000 description 11
- 239000000284 extract Substances 0.000 description 8
- 238000001914 filtration Methods 0.000 description 8
- 230000004048 modification Effects 0.000 description 8
- 238000012986 modification Methods 0.000 description 8
- 230000001174 ascending effect Effects 0.000 description 6
- 230000001815 facial effect Effects 0.000 description 6
- 238000012546 transfer Methods 0.000 description 6
- 230000008859 change Effects 0.000 description 5
- 238000011156 evaluation Methods 0.000 description 5
- 230000004044 response Effects 0.000 description 5
- 241000270295 Serpentes Species 0.000 description 4
- 238000004458 analytical method Methods 0.000 description 4
- 230000000873 masking effect Effects 0.000 description 4
- 238000006243 chemical reaction Methods 0.000 description 3
- 230000007423 decrease Effects 0.000 description 3
- 238000009826 distribution Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 230000010354 integration Effects 0.000 description 3
- 101100365736 Candida albicans (strain SC5314 / ATCC MYA-2876) SEP7 gene Proteins 0.000 description 2
- 101150105385 SHS1 gene Proteins 0.000 description 2
- 230000001133 acceleration Effects 0.000 description 2
- 230000005540 biological transmission Effects 0.000 description 2
- 230000000295 complement effect Effects 0.000 description 2
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 238000000354 decomposition reaction Methods 0.000 description 2
- 230000010339 dilation Effects 0.000 description 2
- 239000006185 dispersion Substances 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 238000010606 normalization Methods 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 238000000513 principal component analysis Methods 0.000 description 2
- 230000002035 prolonged effect Effects 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 238000013519 translation Methods 0.000 description 2
- 238000012800 visualization Methods 0.000 description 2
- 244000025254 Cannabis sativa Species 0.000 description 1
- 241000282326 Felis catus Species 0.000 description 1
- 241001465754 Metazoa Species 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000000593 degrading effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 210000005069 ears Anatomy 0.000 description 1
- 238000003708 edge detection Methods 0.000 description 1
- 230000008030 elimination Effects 0.000 description 1
- 238000003379 elimination reaction Methods 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 238000004091 panning Methods 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
- 230000000644 propagated effect Effects 0.000 description 1
- 239000011435 rock Substances 0.000 description 1
- 238000010187 selection method Methods 0.000 description 1
- 230000007480 spreading Effects 0.000 description 1
- 238000003892 spreading Methods 0.000 description 1
- 239000004575 stone Substances 0.000 description 1
Images
Classifications
-
- H04N5/44543—
-
- G06K9/00308—
-
- G06K9/00335—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
- G06V40/175—Static expression
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/44008—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/4402—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/441—Acquiring end-user identification, e.g. using personal code sent by the remote control or by inserting a card
- H04N21/4415—Acquiring end-user identification, e.g. using personal code sent by the remote control or by inserting a card using biometric characteristics of the user, e.g. by voice recognition or fingerprint scanning
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/485—End-user interface for client configuration
Definitions
- the present invention relates to an image processing apparatus, and specifically, relates to an image processing apparatus and an image processing method which analyze an image to execute image processing, and a program that causes a computer to execute the method thereof.
- imaging apparatuses such as digital video cameras (e.g., integrated camera/recorders) have come into widespread use. For example, a tourist traveling various tourist locations will often take moving pictures of a person or persons using an imaging apparatus, with a famous building or scenery or the like in the background.
- digital video cameras e.g., integrated camera/recorders
- the moving picture thus taken can be played, for example, using a player within the home.
- playing the whole moving picture from the beginning can mean that the play time is prolonged, and there is a possibility that interest in the moving picture may decrease.
- playing multiple moving pictures taken at multiple tourist sites can mean that the play time is further prolonged, and interest in viewing all the moving pictures may decrease.
- the time used for viewing of a moving picture is relatively long, and accordingly, a viewer is often satisfied with viewing a moving picture that has been played and viewed once, and consequently, there is little call for viewing such a moving picture repeatedly.
- This film roll image is, for example, an image wherein the thumbnail images of still images that represent scenes making up a moving picture are disposed in the playback point-in-time order of scenes corresponding to each thereof.
- the imaging range corresponding to the thumbnail image thereof is displayed, but other imaging spaces at the scene corresponding to the thumbnail image thereof are not displayed.
- interest in the scene thereof can be furthered.
- an image processing apparatus and an image processing method including: a front-most image information storage unit configured to store information relating to a front-most image that is one image of a plurality of compositing target images serving as compositing targets of an imaged image making up an imaged moving picture; a transformation information storage unit configured to store transformation information used for transforming another compositing target image of the plurality of compositing target images with the front-most image as a reference; an image transforming unit configured to transform the plurality of compositing target images sequentially toward the front-most image from the compositing target image positioned at the forefront or last of the temporal axis based on the transformation information; an image compositing unit configured to overwrite-composite the transformed compositing target images sequentially to generate a composited image; and a display control unit configured to display the composited image sequentially.
- the plurality of compositing target images are transformed sequentially toward the front-most image from the compositing target image positioned at the forefront or last of the temporal axis based on the transformation information, and the transformed compositing target images are overwritten and composited sequentially to display a composited image sequentially.
- the front-most image may be the compositing target image other than the compositing target image positioned at the forefront or last on the temporal axis, with the image transforming unit transforming the compositing target images adjacent to the front-most image, following which sequentially transforming the plurality of compositing target images toward the front-most image from the compositing target image positioned at the edge portion of the opposite side of the transformed compositing target images on the temporal axis based on the transformation information.
- the plurality compositing target images are transformed sequentially toward the front-most image from the compositing target image positioned at the edge portion of the opposite side of the transformed compositing target images on the temporal axis based on the transformation information.
- the image compositing unit may add a decoration image to the peripheries of the transformed compositing target images, and sequentially overwrite-composite the compositing target images to which the decoration image thereof has been added to generate a composited image.
- the image processing apparatus may further include an operation accepting unit configured to accept a specification operation used to specifying whether to execute the transformation from which edge portion on the temporal axis, with the image transforming unit executing the transformation toward the front-most image from the composited target image positioned at the specified edge portion.
- an operation accepting unit configured to accept a specification operation used to specifying whether to execute the transformation from which edge portion on the temporal axis, with the image transforming unit executing the transformation toward the front-most image from the composited target image positioned at the specified edge portion.
- the image processing apparatus may further include an object detecting unit configured to detect an object included in the compositing target images, with the case that the object is included in the transformed compositing target image, in order to prevent the next compositing target image from being overwritten on an object region including the object thereof, the image compositing unit moving and overwrite-compositing the next compositing target image thereof.
- an object detecting unit configured to detect an object included in the compositing target images, with the case that the object is included in the transformed compositing target image, in order to prevent the next compositing target image from being overwritten on an object region including the object thereof, the image compositing unit moving and overwrite-compositing the next compositing target image thereof.
- the plurality of compositing target images may be images that are transformed and composited based on the transformation information, thereby creating a panorama image corresponding to a portion of imaging space included in the imaged moving picture.
- FIG. 1 is a block diagram illustrating a functional configuration example of an image processing apparatus according to a first embodiment of the present invention
- FIG. 2 is a block diagram illustrating a functional configuration example of a camera work detecting unit according to the first embodiment of the present invention
- FIGS. 3A and 3B are diagrams schematically illustrating files stored in a moving picture storage unit and a metadata storage unit, according to the first embodiment of the present invention
- FIG. 4 is a diagram schematically illustrating the storage content of a compositing target image metadata storage unit according to the first embodiment of the present invention
- FIGS. 5A through 5C are diagrams illustrating an example of an image making up a moving picture
- FIGS. 6A through 6C are diagrams illustrating an image simplified by omitting the background or the like is omitted regarding an image making up a moving picture
- FIG. 7 is a diagram illustrating a case where a rectangle is transformed using affine transformation parameters
- FIG. 8 is a diagram schematically illustrating a face detecting method by a face detecting unit according to the first embodiment of the present invention.
- FIG. 9 is a diagram conceptually illustrating discriminant axis information used for calculation of a facial expression evaluated value, and a flow at the time of generating this discriminant axis information
- FIG. 10 is a diagram schematically illustrating the masking process for a sample image to be input at the time of the PCA process according to the first embodiment of the present invention
- FIG. 11 is a diagram schematically illustrating relationship between a determinant axis in the pixel space and the PCA space, and a face image to be evaluated;
- FIG. 12 is a diagram illustrating a calculation example in the case that the facial expression evaluated value according to the first embodiment of the present invention is output as a numeric value
- FIG. 13 is a flowchart illustrating the procedure of a metadata recording process by the image processing apparatus according to the first embodiment of the present invention
- FIG. 14 is a flowchart illustrating an affine transformation parameter calculating procedure included in the procedure of the metadata recording process by the image processing apparatus according to the first embodiment of the present invention
- FIG. 15 is a diagram illustrating an example of the transition of a moving picture imaged by a camera
- FIG. 16 is a diagram schematically illustrating the locus of a subject included in images selected by a compositing target image selecting unit according to an embodiment of the present invention
- FIG. 17 is a diagram illustrating an image compositing example in the case that images are composited by an image compositing unit according to the first embodiment of the present invention.
- FIG. 18 is a diagram illustrating an example of the transition of a moving picture imaged by the camera.
- FIG. 19 is a diagram schematically illustrating the locus of a subject included in images selected by the compositing target image selecting unit according to an embodiment of the present invention.
- FIG. 20 is a diagram illustrating an image compositing example in the case that images are composited by the image compositing unit according to the first embodiment of the present invention
- FIG. 21 is a diagram illustrating an example of the transition of a moving picture imaged by the camera.
- FIG. 22 is a diagram schematically illustrating the locus of a subject included in images selected by the compositing target image selecting unit according to an embodiment of the present invention.
- FIG. 23 is a diagram illustrating an image compositing example in the case that images are composited by the image compositing unit according to the first embodiment of the present invention.
- FIG. 24 is a diagram illustrating a reliability determination score calculated by the camera work detecting unit according to the first embodiment of the present invention, in a time-series manner;
- FIG. 25 is a diagram illustrating an integral zoom component calculated with an affine transformation parameter calculated by the camera work detecting unit according to the first embodiment of the present invention, in a time-series manner;
- FIG. 26 is a diagram illustrating an integral translational component calculated with an affine transformation parameter calculated by the camera work detecting unit according to the first embodiment of the present invention, in a time-series manner;
- FIGS. 27A and 27B are diagrams schematically illustrating an image included in an image compositing target section selected by a section selecting unit according to the first embodiment of the present invention, in a time-series manner;
- FIG. 28 is a diagram schematically illustrating a selecting method in the case that a compositing target image is selected by the compositing target image selecting unit according to the first embodiment of the present invention
- FIG. 29 is a diagram schematically illustrating a moving picture to be determined by the section selecting unit and the compositing target image selecting unit, according to the first embodiment of the present invention, in a time-series manner;
- FIG. 30 is a diagram illustrating an example of a compositing target image selected regarding a moving picture imaged by the camera
- FIG. 31 is a diagram illustrating the transition of a composited image created by the image compositing unit according to the first embodiment of the present invention.
- FIG. 32 is a diagram illustrating the transition of a composited image created by the image compositing unit according to the first embodiment of the present invention.
- FIG. 33 is another compositing example of a composited image created by the image compositing unit according to the first embodiment of the present invention.
- FIG. 34 is a flowchart illustrating the procedure of a compositing target image metadata recording process by the image processing apparatus according to the first embodiment of the present invention
- FIG. 35 is a flowchart illustrating a section extracting procedure included in the procedure of the compositing target image metadata recording process by the image processing apparatus according to the first embodiment of the present invention
- FIG. 36 is a flowchart illustrating a frame selecting procedure included in the procedure of the compositing target image metadata recording process by the image processing apparatus according to the first embodiment of the present invention
- FIG. 37 is a flowchart illustrating the procedure of a composited image display process by the image processing apparatus according to the first embodiment of the present invention.
- FIG. 38 is a flowchart illustrating an image compositing procedure included in the procedure of the composited image display process by the image processing apparatus according to the first embodiment of the present invention.
- FIG. 39 is a block diagram illustrating a functional configuration example of a camera work detecting unit according to a second embodiment of the present invention.
- FIG. 40 is a diagram illustrating an overview of a calculating method at the time of calculating a reliability determination score by a reliability determination score calculating unit according to the second embodiment of the present invention.
- FIG. 41 is a block diagram illustrating a functional configuration example of a camera work detecting unit according to the second embodiment of the present invention.
- FIG. 42 is a diagram illustrating an overview of a calculating method at the time of calculating a reliability determination score by a reliability determination score calculating unit according to the second embodiment of the present invention.
- FIG. 43 is a block diagram illustrating a functional configuration example of an image processing apparatus according to a third embodiment of the present invention.
- FIGS. 44A and 44B are diagrams schematically illustrating content stored in a reference information storage unit and a metadata storage unit, according to the third embodiment of the present invention.
- FIG. 45 is a diagram illustrating an example of a selection screen displayed on a display unit according to the third embodiment of the present invention.
- FIG. 46 is a diagram illustrating an example of scores used for calculation of a selection determination score by a compositing target image selecting unit according to the third embodiment of the present invention.
- FIGS. 47A and 47B are diagrams illustrating an overview of a selecting method in the case that frame selection is executed by the compositing target image selecting unit according to the third embodiment of the present invention.
- FIGS. 48A and 48B are diagrams illustrating an overview of the selecting method in the case that frame selection is executed by the compositing target image selecting unit according to the third embodiment of the present invention.
- FIG. 49 is a flowchart illustrating the procedure of a frame selecting process by the image processing apparatus according to the third embodiment of the present invention.
- FIG. 50 is a block diagram illustrating a functional configuration example of an image processing apparatus according to a fourth embodiment of the present invention.
- FIG. 51 is a diagram schematically illustrating the storage content of a compositing target image metadata storage unit according to a fourth embodiment of the present invention.
- FIGS. 52A through 52C are diagrams illustrating an overview of a determining method arranged to determine the heaviness of a white frame by a decoration image determining unit according to the fourth embodiment of the present invention.
- FIGS. 53A through 53C are diagrams illustrating an example of a decoration image determined by the decoration image determining unit according to the fourth embodiment of the present invention.
- FIGS. 54A through 54C are diagrams illustrating an example of a determining method for a decoration image by the decoration image determining unit according to the fourth embodiment of the present invention.
- FIGS. 55A and 55B are diagrams illustrating an example of a determining method for a decoration image by the decoration image determining unit according to the fourth embodiment of the present invention.
- FIGS. 56A through 56D are diagrams illustrating a compositing example in which decoration images determined by the decoration image determining unit according to the fourth embodiment are composited;
- FIGS. 57A through 57D are diagrams illustrating a compositing example in which decoration images determined by the decoration image determining unit according to the fourth embodiment are composited;
- FIGS. 58A through 58D are diagrams illustrating a compositing example in which decoration images determined by the decoration image determining unit according to the fourth embodiment are composited;
- FIG. 59 is a diagram illustrating an example of a decoration image determining method by the decoration image determining unit according to the fourth embodiment, which takes faces into consideration;
- FIGS. 60A through 60C are diagrams illustrating an example of a shadow determined to be a decoration image to be added to a compositing target image by the decoration image determining unit according to the fourth embodiment
- FIGS. 61A and 61B are diagrams illustrating an overview of a determining method arranged to determine the direction of a shadow by the decoration image determining unit according to the fourth embodiment
- FIG. 62 is a diagram illustrating an example of a shadow determining method by the decoration image determining unit according to the fourth embodiment, which takes faces into consideration;
- FIG. 63 is a flowchart illustrating the procedure of a compositing target image metadata recording process by the image processing apparatus according to the fourth embodiment of the present invention.
- FIG. 64 is a flowchart illustrating a decoration image determining procedure included in the procedure of the compositing target image metadata recording process by the image processing apparatus according to the fourth embodiment of the present invention.
- FIG. 65 is a flowchart illustrating a decoration image determining procedure included in the procedure of the compositing target image metadata recording process by the image processing apparatus according to the fourth embodiment of the present invention.
- FIG. 66 is a flowchart illustrating the procedure of a composited image display process by the image processing apparatus according to the fourth embodiment of the present invention.
- FIG. 67 is a flowchart illustrating an image compositing procedure included in the procedure of the composited image display process by the image processing apparatus according to the fourth embodiment of the present invention.
- FIG. 68 is a diagram illustrating a configuration example of a multi-core processor according to a fifth embodiment of the present invention.
- FIG. 69 is a diagram illustrating a configuration example of a control processor core according to the fifth embodiment of the present invention.
- FIG. 70 is a diagram illustrating a configuration example of an arithmetic processor core according to the fifth embodiment of the present invention.
- FIG. 71 is a diagram schematically illustrating a computation method of the multi-core processor according to the fifth embodiment of the present invention.
- FIG. 72 is a diagram schematically illustrating the flows of a program and data in the case that a computation is executed by the multi-core processor according to the fifth embodiment of the present invention.
- FIG. 73 is a diagram schematically illustrating an overview of a computation method arranged to execute a process as to a plurality of data by each corresponding command, and an overview of an SIMD computation that is a computation method arranged to execute a process as to a plurality of data by a single command;
- FIG. 74 is a diagram illustrating a configuration example of a program executed by the control processor core or arithmetic processor core, according to the fifth embodiment of the present invention.
- FIG. 75 is a diagram schematically illustrating a data structure and a processing flow in the case that image data stored in main memory according to the fifth embodiment of the present invention is subjected to a filtering process using a Sobel filter;
- FIG. 76 is a diagram schematically illustrating a data flow in the case that the Sobel filter is used to execute SIMD computation regarding image data stored in the main memory according to the fifth embodiment of the present invention.
- FIG. 77 is a diagram schematically illustrating a vector creating method arranged to create nine vectors from the image data stored in a first buffer in the case that the filtering process is executed using the Sobel filter according to the fifth embodiment of the present invention
- FIG. 78 is a diagram schematically illustrating a vector computing method arranged to use SIMD computation to execute a vector computation regarding vector data in the case that the filtering process is executed using the Sobel filter according to the fifth embodiment of the present invention
- FIG. 79 is a diagram schematically illustrating the flow of a camera work parameter calculating process according to the fifth embodiment of the present invention in a time-series manner.
- FIG. 80 is a diagram schematically illustrating a Blu-ray Disc which is an example of a recording medium, data recorded in the Blu-ray Disc, and the internal configuration of a Blu-ray player capable of playing the Blu-ray Disc.
- First Embodiment (composited image creating control: an example in which compositing target images are selected from a moving picture, and these are used to create a composited image)
- Second Embodiment image compositing target section selecting control: a modification wherein an image compositing target section to select compositing target images is selected from a moving picture
- FIG. 1 is a block diagram illustrating a functional configuration example of an image processing apparatus 100 according to a first embodiment of the present invention.
- the image processing apparatus 100 includes a moving picture input unit 110 , a camera work detecting unit 120 , a face detecting unit 130 , a facial expression evaluated value calculating unit 140 , a discriminant axis information storage unit 141 , a recording control unit 150 , a section selecting unit 160 , and a compositing target image selecting unit 170 .
- the image processing apparatus 100 includes a file obtaining unit 180 , an image transforming unit 190 , a moving picture storage unit 200 , a metadata storage unit 210 , an image compositing unit 220 , and image memory 230 .
- the image processing apparatus 100 includes a display control unit 240 , a display unit 250 , an operation accepting unit 260 , and a compositing target image metadata storage unit 270 .
- the image processing apparatus 100 can be realized by a personal computer which can subject a moving picture imaged by an imaging apparatus, for example, such as a digital video camera or the like, to video analysis for extracting feature quantity, and to various types of image process using this extracted feature quantity.
- the moving picture input unit 110 is a moving picture input unit configured to input a moving picture imaged by an imaging apparatus such as a digital video camera or the like (hereafter, simply referred to as “camera”).
- the moving picture input unit 110 outputs the input moving picture to the camera work detecting unit 120 , face detecting unit 130 , facial expression evaluated value calculating unit 140 , and recording control unit 150 .
- the camera work detecting unit 120 analyzes the moving picture output from the moving picture input unit 110 to detect information of motion at the time of imaging (also referred to as “camera work”), and calculates camera work parameters based on the motion information of the camera. Also, when calculating camera work parameters, the camera work detecting unit 120 calculates a reliability determination score that indicates the likelihood of the camera work parameters, and outputs the calculated camera work parameters and the reliability determination score to the recording control unit 150 . Specifically, the camera work detecting unit 120 extracts a feature point from each image making up the moving picture, and calculates an optical flow (movement vector) as to this feature point.
- an optical flow movement vector
- the camera work detecting unit 120 analyzes the calculated optical flow as to the feature point to select a feature point that exhibits dominant motion, and estimates the motion of the camera based on the optical flow as to the feature point that exhibits dominant motion.
- “dominant motion” means regular motion that a relatively great number of optical flows exhibit out of multiple optical flows regarding multiple feature points.
- a reliability determination score is calculated based on a rate between the feature point that exhibits dominant motion, and feature points other than that feature point. Note that the internal configuration of the camera work detecting unit 120 will be described in detail with reference to FIG. 2 . Note that the camera work detecting unit 120 is an example of the calculating unit referred to in the Summary of the Invention.
- the face detecting unit 130 detects the face of a person included in an image making up a moving picture output from the moving picture input unit 110 , and outputs face detection information relating to the detected face to the facial expression evaluated value calculating unit 140 , and recording control unit 150 .
- this face detecting method include a face detecting method by matching between a template in which facial luminance distribution information is recorded, and an actual image (e.g., see Japanese Unexamined Patent Application Publication No. 2004-133637), and a face detecting method based on a skin-colored portion, and the feature quantity of the face of a person, included in an image.
- a face detecting method may be employed wherein, with regard to a face image that is an image including a face, luminance difference between two pixels on the face image is learned, the feature quantity obtained from this learning is stored as face detection reference data beforehand, and face detection is executed using this face detection reference data.
- the face detection method using this face detection reference data will be described in detail with reference to FIG. 8 .
- the face detection information includes the position and size of the detected face image. Examples of the face position included in the face detection information include the position of a rectangular face detection window surrounding the detected face (e.g., the coordinates of the upper left corner of the face detection window, hereafter, referred to as “face position information”).
- examples of the size of a face included in the face detection information include the size of the face detection window thereof (e.g., the number of pixels in the horizontal direction and vertical direction of the face detection window).
- the face detection information may include roll, pitch, and yaw that indicate the orientation of the detected face.
- the facial expression evaluated value calculating unit 140 uses the discriminant axis information stored in the discriminant axis information storage unit 141 to calculate a facial expression evaluated value relating to the facial expression of the face detected by the face detecting unit 130 , and outputs the calculated facial expression evaluated value to the recording control unit 150 .
- a face image that is a rectangular region including at least a portion of the face detected by the face detecting unit 130 is extracted from an image making up the moving picture output from the moving picture input unit 110 . This extracted face image is normalized, and a facial expression evaluated value regarding the normalized face image is calculated.
- the facial expression evaluated value Based on this calculated facial expression evaluated value, determination can be made whether or not the facial expression of the face to be calculated is a specific facial expression. Note that, with the first embodiment of the present invention, an example will be described wherein a smiling face is determined to be a specific facial expression. Also, the facial expression evaluated value calculation will be described in detail with reference to FIGS. 9 through 12 . Also, the facial expression evaluated value calculating unit 140 is an example of the determining unit referred to in the Summary of the Invention.
- the discriminant axis information storage unit 141 stores discriminant axis information used for calculation of a facial expression evaluated value by the facial expression evaluated value calculating unit 140 , and supplies the stored discriminant axis information to the facial expression evaluated value calculating unit 140 . Note that the discriminant axis information will be described in detail with reference to FIG. 9 and so forth.
- the recording control unit 150 executes recording control as to the moving picture storage unit 200 , metadata storage unit 210 , and the compositing target image metadata storage unit 270 . That is to say, the recording control unit 150 records the moving picture output from the moving picture input unit 110 in the moving picture storage unit 200 as a moving picture file. Also, the recording control unit 150 records the affine transformation parameter and the reliability determination score output from the camera work detecting unit 120 , the face detection information output from the face detecting unit 130 , and the facial expression evaluated value output from the facial expression evaluated value calculating unit 140 in the metadata storage unit 210 . Such each piece of information is recorded in the metadata storage unit 210 in a manner correlated with a moving picture and a frame (image) as a metadata file. Also, the recording control unit 150 records the compositing target image information (frame selection information) output from the compositing target image selecting unit 170 in the compositing target image metadata storage unit 270 as compositing target image metadata.
- frame selection information output from the compositing target
- the section selecting unit 160 uses the metadata stored in the metadata storage unit 210 to select an image compositing target section that is a section serving as a target for creating a composited image regarding the moving picture stored in the moving picture storage unit 200 , and outputs information relating to the selected image compositing target section to the compositing target image selecting unit 170 . Note that selection of an image compositing target section will be described in detail with reference to FIGS. 24 through 26 .
- the compositing target image selecting unit 170 uses the metadata stored in the metadata storage unit 210 to select a frame for creating a composited image from frames included in the image compositing target section output from the section selecting unit 160 , and outputs compositing target image information (frame selection information) relating to the selected frame to the recording control unit 150 . Specifically, the compositing target image selecting unit 170 first selects a front-most image (front-most frame) to be overlaid and composited on the top side of the composited image, and with the selected front-most image as a reference image, selects another compositing target image in order.
- Affine transformation parameters used for subjecting the selected compositing target image to affine transformation are calculated at the time of this selection with the front-most image as a reference image. Note that compositing target image selection will be described in detail with reference to FIGS. 27A through 29 . Also, the section selecting unit 160 and the compositing target image selecting unit 170 are examples of the selecting unit referred to in the Summary of the Invention. Also, the compositing target image selecting unit 170 is an example of the image selecting unit referred to in the Summary of the Invention.
- the moving picture storage unit 200 stores the moving picture output from the moving picture input unit 110 as a moving picture file based on the control of the recording control unit 150 . Also, the moving picture storage unit 200 supplies the moving picture file to the file obtaining unit 180 in response to a request from the file obtaining unit 180 , and supplies the moving picture file to the display control unit 240 in response to a request from the display control unit 240 . Note that the moving picture file to be stored in the moving picture storage unit 200 will be described in detail with reference to FIGS. 3A and 3B .
- the metadata storage unit 210 stores each piece of information output from the camera work detecting unit 120 , face detecting unit 130 , and facial expression evaluated value calculating unit 140 in a manner correlated with a moving picture and frame as a metadata file based on the control of the recording control unit 150 . Also, the metadata storage unit 210 supplies the metadata file to the section selecting unit 160 and the compositing target image selecting unit 170 . Note that the metadata file to be stored in the metadata storage unit 210 will be described in detail with reference to FIGS. 3A and 3B .
- the compositing target image metadata storage unit 270 stores the compositing target image information output from the compositing target image selecting unit 170 as a compositing target image metadata file based on the control of the recording control unit 150 . Also, the compositing target image metadata storage unit 270 supplies the compositing target image metadata file to the file obtaining unit 180 in response to a request from the file obtaining unit 180 . Note that the metadata file to be stored in the compositing target image metadata storage unit 270 will be described in detail with reference to FIG. 4 . Note that the compositing target image metadata storage unit 270 is an example of the -most image information storage unit and transformation information storage unit referred to in the Summary of the Invention.
- the file obtaining unit 180 obtains each file stored in the moving picture storage unit 200 and the metadata storage unit 210 in response to operation input accepted by the operation accepting unit 260 , and supplies the information of the obtained each file to the image transforming unit 190 . Specifically, in the case that an instruction operation used for displaying a composited image has been accepted by the operation accepting unit 260 , the file obtaining unit 180 obtains the moving picture file from the moving picture storage unit 200 . Also, the file obtaining unit 180 obtains the compositing target image metadata file corresponding to this moving picture file from the compositing target image metadata storage unit 270 .
- the file obtaining unit 180 uses the compositing target image metadata file to extract an image to be composited from each image making up the obtained moving picture file, and outputs this extracted image, and the compositing target image metadata corresponding to this image to the image transforming unit 190 .
- the image transforming unit 190 subjects the image output from the file obtaining unit 180 to affine transformation using the compositing target image metadata corresponding to this image, and outputs the image subjected to affine transformation to the image compositing unit 220 .
- the image transforming unit 190 subjects another image to affine transformation in order with the position of a single image (e.g., front-most image) as a reference.
- the image transforming unit 190 subjects an image to affine transformation in order in accordance with the display order according to the instruction operation thereof. Note that such image transformation will be described in detail with reference to FIGS. 15 through 23 , 31 , 32 , and so forth.
- the image compositing unit 220 composites the image subjected to affine transformation by the image transforming unit 190 to create a composited image using the image memory 230 , and outputs the created composited image to the display control unit 240 sequentially. Also, when compositing the image subjected to affine transformation by the image transforming unit 190 , the image compositing unit 220 composites the image by adding a white frame to the periphery of the image. Such image compositing will be described in detail with reference to FIGS. 31 and 32 , and other drawings.
- the image memory 230 is a work buffer used at the creating a composited image by the image compositing unit 220 , holds history images including the image subjected to affine transformation by the image transforming unit 190 , and supplies the held history images to the image compositing unit 220 .
- the image memory 230 is also called a “canvas”, equivalent to one image worth of memory, where image data is stored.
- the display control unit 240 displays the composited image output from the image compositing unit 220 on the display unit 250 in accordance with the content of the operation input accepted by the operation accepting unit 260 . Also, in the case that an instruction operation used for playing a moving picture has been accepted by the operation accepting unit 260 , the display control unit 240 obtains the moving picture file according to this instruction operation from the moving picture storage unit 200 , and controls the display unit 250 to display an image making up this moving picture file.
- the display unit 250 displays each image based on the control of the display control unit 240 .
- the display unit 250 can be realized with, for example, the display of a personal computer, or a television set. Note that a display example of a composited image will be described in detail with reference to FIGS. 31 through 33 and so forth.
- the operation accepting unit 260 accepts operation input from a keyboard made up of various types of keys, a mouse (pointing device), or the like, and outputs the content of accepted operation input to the file obtaining unit 180 , image transforming unit 190 , or display control unit 240 .
- FIG. 2 is a block diagram illustrating a functional configuration example of the camera work detecting unit 120 according to the first embodiment of the present invention.
- the camera work detecting unit 120 includes a feature point extracting unit 121 , an optical flow calculating unit 122 , a camera work parameter calculating unit 123 , and a reliability determination score calculating unit 124 .
- the feature point extracting unit 121 extracts feature points from the whole image corresponding to a frame making up the moving picture output from the moving picture input unit 110 , and outputs the extracted feature points to the optical flow calculating unit 122 and the reliability determination score calculating unit 124 .
- the feature point extracting unit 121 extracts feature points from the whole image regarding the top frame of frames making up the moving picture output from the moving picture input unit 110 .
- the feature point extracting unit 121 compares to the image corresponding to the previous frame to extract feature points from a new imaged region portion regarding a frame other than the top frame.
- a point can be extracted as a feature point where the gradient of the edge is strong in the vertical direction and in the horizontal direction (in general, referred to as “corner point”, hereinafter, referred to as “corner point”).
- This corner point is a feature point that is strong against optical flow calculation, and can be obtained with edge detection. Note that extraction of a corner point will be described in detail with reference to FIGS. 5A through 6C .
- the feature point extracting unit 121 extracts feature points from the whole image regarding the top frame, and compares to the image corresponding to the previous frame to extract feature points from a new imaged region portion regarding a frame other than the top frame.
- feature points may also be extracted from the whole image regarding each frame other than the top frame in accordance with processing capability.
- the optical flow calculating unit 122 calculates an optical flow as to each feature point output from the feature point extracting unit 121 , and outputs the calculated optical flow to the camera work parameter calculating unit 123 . Specifically, the optical flow calculating unit 122 compares each image corresponding to consecutive two frames (the current frame and the immediately previous frame) making up the moving picture output from the moving picture input unit 110 , thereby calculating the optical flow of the current frame. That is to say, the optical flow calculating unit 122 calculates an optical flow as to the image corresponding to the current frame regarding each feature point of the image corresponding to the immediately previous frame. This optical flow is calculated for each frame making up the moving picture.
- optical flow calculating unit 122 is an example of the motion amount calculating unit referred to in the Summary of the Invention.
- the camera work parameter calculating unit 123 executes a camera work parameter calculating process wherein the optical flow corresponding to each feature point output from the optical flow calculating unit 122 is used to calculate camera work parameters. Subsequently, the camera work parameter calculating unit 123 outputs the calculated camera work parameters to the recording control unit 150 , and outputs information relating to the feature point used for calculation of the camera work parameters to the reliability determination score calculating unit 124 .
- the camera work parameters are transformation information (the motion information of the camera) relating to the consecutive two frames (the current frame and the immediately previous frame).
- each compositing target image selected regarding the moving picture to be played is transformed and composited in accordance with the motion of the camera.
- the motion of the camera is extracted using the optical flow calculated by the optical flow calculating unit 122 , and camera work parameters (transformation parameters) are calculated based on the extracted motion.
- affine transformation as an image transforming method used for transforming a compositing target image.
- description will be made regarding an example using affine transformation parameters calculated based on an optical flow as camera work parameters.
- another image transforming method may be employed using projective-transformation parameters or the like as camera work parameters.
- an affine transformation parameter may be obtained by a computation using a vector between three points.
- projective-transformation parameters may be obtained by a computation using a vector between four points.
- camera work parameters are transformation information to transform another imaged image with at least one imaged image of imaged images making up an imaged moving picture as a reference, and include at least position information and attitude information described in the coordinates system of the camera. That is to say, camera work parameters include information relating to the position and attitude of the camera in the case of being imaged by a user. Also, the motion of the camera according to the operation by a user, for example, such as zoom in, zoom out, pan, tilt, rotation, or the like can be estimated based on the affine transformation parameters calculated by the camera work parameter calculating unit 123 . Note that calculation of affine transformation parameters will be described in detail with reference to FIGS. 5A through 6C .
- the camera work parameter calculating unit 123 is an example of the motion information calculating unit referred to in the Summary of the Invention.
- the reliability determination score calculating unit 124 calculates a reliability determination score based on information relating to the feature point output from the feature point extracting unit 121 , and the feature point output from the camera work parameter calculating unit 123 , and outputs the calculated reliability determination score to the recording control unit 150 .
- This reliability determination score is calculated based on a rate between the number of feature points in the whole screen of an image to be subjected to calculation of affine transformation parameters, and the number of feature points that exhibit dominant motion in the whole screen. Calculation of this reliability determination score will be described in detail with reference to FIGS. 5A through 6C .
- FIGS. 3A and 3B are diagrams schematically illustrating files stored in the moving picture storage unit 200 and the metadata storage unit 210 , according to the first embodiment of the present invention.
- FIG. 3A illustrates moving picture files 201 through 203 stored in the moving picture storage unit 200 , metadata files 211 through 213 stored in the metadata storage unit 210 in a manner correlated with the moving picture files 201 through 203 .
- a moving picture ID that is identification information used for identifying each moving picture file stored in the moving picture storage unit 200 is appended to each moving picture file. For example, “# 1 ” is appended to the moving picture file 201 , “# 2 ” is appended to the moving picture file 202 , and “#n” is appended to the moving picture file 203 .
- FIG. 3B schematically illustrates the moving picture file 201 stored in the moving picture storage unit 200 , the metadata file 211 stored in the metadata storage unit 210 in a manner correlated with the moving picture file 201 .
- the moving picture file 201 is a moving picture file made up of m frames, and these m frames are shown as frames “ 1 ” 204 through “m” 207 .
- a moving picture ID 214 a moving picture ID 214
- a frame number 215 a frame number 215
- affine transformation parameters 216 a reliability determination score 217
- face detection information 218 face detection information
- a facial expression evaluated value 219 are stored in a correlated manner.
- the moving picture ID 214 is a moving picture ID appended to the corresponding moving picture file, and for example, stores “# 1 ” appended to the moving picture file 201 .
- the frame number 215 is a serial number of each frame making up the corresponding moving picture file, and for example, stores “ 1 ” through “m” corresponding to the frames “ 1 ” 204 through “m” 207 making up the moving picture of the moving picture file 201 .
- the affine transformation parameters 216 are affine transformation parameters calculated regarding each frame of the moving picture corresponding to the frame number 215 .
- the affine transformation parameters 216 “a 1 , b 1 , c 1 , d 1 , e 1 , f 1 ” corresponding to the “ 1 ” of the frame number 215 are the affine transformation parameters of a unit matrix.
- the affine transformation parameters 216 of “ai, bi, ci, di, ei, fi” corresponding to the “i (i is an integer of 2 or more)” of the frame number 215 are the affine transformation parameters of the frame “i” as to the immediately previous frame “i ⁇ 1”.
- the reliability determination score 217 is a reliability determination score calculated regarding each frame of the moving picture corresponding to the frame number 215 , and for example, stores a value of 0 through 1.
- the face detection information 218 is face detection information detected regarding each frame of the moving picture corresponding to the frame number 215 , and stores the face detection information output from the face detecting unit 130 .
- FIG. 3B illustrates the face detection information of which the specific content is omitted. In the case that multiple faces have been detected from one frame, such a plurality of face detection information is stored in a manner correlated with the frame thereof.
- the facial expression evaluated value 219 is a facial expression evaluated value calculated regarding a face included in each frame of the moving picture corresponding to the frame number 215 , and stores the facial expression evaluated value output from the facial expression evaluated value calculating unit 140 . Note that, in the case that no face is included in the corresponding frame, the facial expression evaluated value 219 stores “ 0 ”. For example, in the case that no face has been detected from the frames corresponding to the “ 1 ” and “m” of the frame number 215 , no facial expression evaluated value is calculated by the facial expression evaluated value calculating unit 140 , and accordingly, the facial expression evaluated value 219 stores “ 0 ”.
- the facial expression evaluated value 219 stores the facial expression evaluated value calculated by the facial expression evaluated value calculating unit 140 .
- the facial expression evaluated value 219 corresponding to the “i” of the frame number 215 stores “ 124 ”
- the facial expression evaluated value 219 corresponding to the “j” of the frame number 215 stores “ 405 ”.
- determination can be made that there is a high possibility that the face detected regarding the frame “j” resembles a smiling face more than the face detected regarding the frame “i”.
- the i and j are integers that satisfy 1 ⁇ i ⁇ j ⁇ m.
- the facial expression evaluated values calculated regarding these multiple faces are stored in a manner correlated with the frame thereof.
- FIG. 4 is a diagram schematically illustrating the storage content of the compositing target image metadata storage unit 270 according to the first embodiment of the present invention.
- a moving picture ID 271 , a section number 272 , a frame number 273 , affine transformation parameters 274 , a front-most frame 275 , and face data 276 are stored in the compositing target image metadata storage unit 270 in a correlated manner.
- the moving picture ID 271 is a moving picture ID appended to the corresponding moving picture file, and for example, stores “# 1 ”, “# 2 ”, and so forth appended to each moving picture file stored in the moving picture storage unit 200 .
- the section number 272 is a serial number used for identifying an image compositing target section selected by the section selecting unit 160 , and for example, stores “# 101 ”, “# 102 ”, “# 103 ”, and so forth in the selected order regarding the moving picture file 201 .
- a composited image is created regarding the moving picture file 201 .
- a composited image is created for each section of the section number 272 (e.g., “# 101 ”, “# 102 ”, “# 103 ”).
- the frame number 273 is the frame number of the frame selected by the compositing target image selecting unit 170 with the image compositing target section selected by the section selecting unit 160 .
- This frame number corresponds to the frame number 215 shown in FIG. 3B .
- the frame numbers “ 1 ”, “ 6 ”, “ 9 ”, and “ 17 ” of these frames are stored.
- the affine transformation parameters 274 are affine transformation parameters used for subjecting each frame selected by the compositing target image selecting unit 170 to affine transformation.
- affine transformation parameters to be stored in the affine transformation parameters 274 are affine transformation parameters used for subjecting another frame to affine transformation with one frame within the same image compositing target section as a reference frame. This reference frame may be set to, for example, the front-most frame.
- affine transformation parameters to be stored in the affine transformation parameters 274 in a manner correlated with this reference frame are the affine transformation parameters of a unit matrix.
- the front-most frame 275 is information that indicates the front-most frame selected by the compositing target image selecting unit 170 . For example, in the case that the corresponding frame has been selected as the front-most frame, the front-most frame 275 stores “ 1 ”, and in the case that the corresponding frame has not been selected as the front-most frame, the front-most frame 275 stores “ 0 ”. For example, of each frame of section number 272 “# 101 ” of the moving picture ID 271 “# 1 ”, in the case that the frame “ 9 ” has been selected as the front-most frame, the front-most frame 275 of the frame “ 9 ” stores “ 1 ”. In this case, of each frame of the section number 272 “# 101 ”, with regard to a frame other than the frame “ 9 ”, the front-most frame 275 stores “ 0 ”.
- the face data 276 is information relating to a face included in each frame selected by the compositing target image selecting unit 170 .
- the position information and size information of a face are stored.
- the position information of a face is shown with (xn, yn)
- the size information of the face is shown with (Hn, Wn).
- n is a numeric value corresponding to the frame number 273 . Note that in the case that multiple faces have been detected from one frame, such a plurality of face data is stored in a manner correlated with the frame thereof.
- FIGS. 5A through 5C are diagrams illustrating an example of an image making up a moving picture.
- FIGS. 6A through 6C are diagrams illustrating an image simplified by omitting the background or the like regarding an image making up a moving picture.
- FIGS. 5A through 5C illustrate an image 300 as an example of an image making up a moving picture.
- FIG. 6A illustrates an image 320 simplified by omitting the background or the like regarding the image corresponding to the immediately previous frame of the image 300 .
- FIGS. 6B and 6C illustrate an image 330 simplified by omitting the background or the like regarding the image 300 .
- the images 300 , 320 , and 330 shown in FIG. 5A through 6C include images 301 , 321 , and 331 of a horse on which a person is straddling, and images 302 , 322 , and 332 of a snake disposed in front of the horse images 301 , 321 , and 331 . Also, as shown in FIGS. 5A through 5C , there are provided a flag, a chair, and so forth on the background of these images, and this flag is fluttering in the wind.
- the image 320 shown in FIG. 6A is an image simplified regarding the image corresponding to the immediately previous frame of the frame corresponding to the images 300 and 330 shown in FIGS. 5A through 5C , 6 B, and 6 C.
- the images 320 and 330 corresponding to two consecutive frames are images that indicate transition in the case that the size of a subject within the screen increases gradually. That is to say, at the time of this imaging a zoom-in operation is performed wherein an operation used for increasing the size of a subject within the screen is increased gradually.
- FIGS. 6A through 6C description will be made regarding a method as an example wherein affine transformation parameters are calculated using the optical flows corresponding to three corner points detected from the images 320 and 330 .
- corner points 333 , 334 , and 335 corresponding to the corner points 323 , 324 , and 325 of the image 320 are detected.
- the horse images 321 and 331 and snake images 322 and 332 include in the images 320 and 330 shown in FIGS. 6A and 6B are installed on the ground, and accordingly, do not move regardless of the motion of the camera. Therefore, the motion of the camera can be estimated accurately based on the optical flows obtained as to the corner points detected regarding the horse images 321 and 331 , and the snake images 322 and 332 . For example, as shown in FIG. 6C , based on the three optical flows 337 through 339 detected in the image 330 , it can be estimated that the image 330 is an image enlarged from the image 320 with the point 336 as the center.
- an object that moves regardless of the motion of the camera is included in an image, such as a flag fluttering in the wind or the like.
- the image 300 shown in FIGS. 5A through 5C includes a flag fluttering in the wind.
- a corner point is detected regarding an object that moves regardless of the motion of the camera, and the motion of the camera is estimated using the optical flow obtained as to this corner point, it is difficult to estimate the motion of the camera accurately.
- corner points 303 through 305 are corner points corresponding to the corner points 333 through 335 shown in FIGS. 6B and 6C .
- corner points 306 through 311 are corner points detected regarding the flags disposed on the background of the horse image 301 . These flags are fluttering in the wind, and accordingly, the motion of each of the flags due to influence of the wind has been detected as an optical flow. That is to say, optical flows corresponding to the corner points 306 through 311 are optical flows detected regarding the flags that move regardless of the motion of the camera.
- the three optical flows used in the case of calculating affine transformation parameters include an optical flow corresponding to at least one corner point of the corner points 306 through 311 , it is difficult to detect the motion of the camera accurately. In this case, it is difficult to calculate accurate affine transformation parameters.
- an optical flow as to an object that moves regardless of the motion of the camera (optical flows corresponding to the corner points 306 through 311 shown in FIG. 5B ) is detected from an imaged image in some cases.
- an optical flow having a particular regularity in relationship with the motion of the camera (an optical flow other than the optical flows corresponding to the corner points 306 through 311 shown in FIG. 5B ) is detected from an imaged image in some cases.
- an affine transformation parameter calculating process is executed multiple times wherein affine transformation parameters are calculated based on the three optical flows, thereby obtaining multiple affine transformation parameters.
- the optimal affine transformation parameters are selected from these multiple affine transformation parameters. Note that, with this example, we will say that the size of a moving object included in each image making up a moving picture is relatively small as to the area of the image.
- affine transformation In the case that the position of a moving source is set to (x, y), and the position of a moving destination after affine transformation is set to (x′, y′) on the two-dimensional coordinates, the matrix expression of affine transformation can be represented with Expression 1. Note that the matrix of 3 ⁇ 3 of the right side of Expression 1 is an affine matrix.
- a through f are affine transformation parameters.
- the affine transformation parameters can be represented with the following expression with the affine transformation parameters as a matrix AM.
- a zoom component XZ in the X direction, a zoom component YZ in the Y direction, a translational component XT in the X direction, a translational component YT in the Y direction, a rotational component ⁇ x in the X direction, and a rotational component ⁇ y in the Y direction can be obtained, respectively.
- AM ( a b c d e f )
- y tan - 1 ( b e ⁇ )
- each component of camera work (zoom components in the X and Y directions, translational components, rotational components) can be obtained from the affine transformation parameters. Note that the rotational components will be described with reference to FIG. 7 .
- FIG. 7 is a diagram illustrating a case where a rectangle 340 is transformed using affine transformation parameters.
- one peak of the rectangle 340 is the origin, and with the x-y coordinates wherein the two sides of the rectangle 340 adjacent to the peak corresponding to the origin are taken as the x axis and the y axis, a rectangle after the rectangle 340 is transformed using predetermined affine transformation parameters is taken as a rectangle 341 .
- an angle formed by the sides on the x axis side of the rectangles 340 and 341 is a rotational component ⁇ x
- an angle formed by the sides on the y axis side of the rectangles 340 and 341 is a rotational component ⁇ y.
- the addition value of camera work components is a value obtained by adding each component of the camera work corresponding to each frame from the top frame of the image compositing target section serving as a calculation target to the current frame serving as a calculation target.
- the frames from the top frame to the current frame are taken as frames 1 through 5
- the translational components in the X direction of the camera work components of the frames 1 through 5 are taken as XT 1 through XT 5 .
- the addition value of the translational components in the X direction of the camera work components can be calculated by “XT 1 +XT 2 +XT 3 +XT 4 +XT 5 ”.
- the addition value of each component of another camera work can be calculated in the same way.
- affine transformation parameters are calculated using the three optical flows corresponding to the selected three feature points. For example, affine transformation parameters are calculated using the optical flows (indicated with an arrow connected to a white circle) corresponding to the three corner points selected from the corner points (indicated with a white circle) in the image 300 shown in FIG. 5B .
- the affine transformation parameters can be obtained using Expression 1.
- the scores of the affine transformation parameters are calculated. Specifically, the positions of the movement destinations of all the feature points in the image corresponding to the immediately previous frame of the current frame are obtained using the obtained affine transformation parameters. Subsequently, the position of the feature point obtained using the affine transformation parameters, and the position of the feature point detected at the current frame are compared, and the difference value of the positions of the two feature points corresponding to each other is calculated for each feature point. For example, absolute difference between the positions of the two feature points corresponding to each other is calculated as a difference value.
- the calculated difference value, and a predetermined threshold are compared for each feature point, and the number of feature points of which the difference value is smaller than the threshold is obtained as the scores of the affine transformation parameters.
- three feature points are selected at random from the feature points of which the optical flows have been detected.
- a process is repeated a predetermined number of times wherein the scores of the affine transformation parameters are calculated based on the optical flows corresponding to these feature points, thereby calculating the multiple scores of the affine transformation parameters.
- This predetermined number of times may be set as appropriate according to the types of image to be compared, the processing capability of the image processing apparatus 100 , or the like, or a fixed value may be used. For example, twenty times or so may be set as this predetermined number of times, taking the processing capability of the image processing apparatus 100 into consideration.
- affine transformation parameters are calculated using the three optical flows corresponding to the selected three corner points, as described above, these three optical flows have a particular regularity. Therefore, affine transformation parameters used for transforming the image corresponding to the immediately previous frame in accordance with a certain rule are obtained. Also, a relatively small value is calculated as the difference value between the position of the corner point obtained using the affine transformation parameters, and the position of the corner point detected at the current frame, obtained regarding a corner point other than the corner points 306 through 311 . Therefore, the scores of the affine transformation parameters become great in value.
- the affine transformation parameter having the highest score value is selected as a representative affine transformation parameter.
- the selected representative affine transformation parameter is used to calculate affine transformation parameters used for subjecting the current frame to affine transformation with the immediately previous frame as a reference frame, and the affine transformation parameters are recorded in the metadata storage unit 210 in a manner correlated with the current frame.
- the image can be subjected to affine transformation using the optimal affine transformation parameters.
- the reliability determination score thus calculated is recorded in the metadata storage unit 210 in a manner correlated with the current frame.
- An image compositing target section is selected from a moving picture using this reliability determination score.
- FIG. 8 is a diagram schematically illustrating a face detecting method by the face detecting unit 130 according to the first embodiment of the present invention.
- (a) and (b) illustrate an image 450 making up a moving picture output from the moving picture input unit 110 .
- the face 452 of a person 451 is included in the image 450 .
- the face 452 is detected from the image 450 by the face detecting unit 130 .
- description will be made as an example regarding a case where face detection is executed using the face detection reference data.
- a face detection window 453 having a certain size is disposed at the left upper corner of the image 450 , and an estimate used for determining whether or not a face is included in the image within the face detection window 453 is calculated based on the face detection reference data.
- the face detection window is shifted in the right-sided direction (direction indicated with an arrow 454 ) by one pixel, and similarly, an estimate is calculated.
- the face detection window is shifted sequentially in the right-sided direction by one pixel at a time, thereby calculating an estimate sequentially.
- the face detection window is shifted to one-pixel lower side, and is moved to the left edge of the image 450 .
- the face detection window is shifted sequentially in the right-sided direction by one pixel at a time, thereby calculating an estimate sequentially.
- an estimate is calculated sequentially.
- the face detection window is reduced by a predetermined scale factor, and the same process is executed sequentially.
- the face detection window having a certain size is used to execute calculation of an estimate sequentially regarding the image 450 to be reduced sequentially. Subsequently, a face is detected based on calculated each estimate, and a face region that is a rectangular region including at least a portion of this face is obtained. For example, as shown in (b) in FIG. 8 , the face 452 of the person 451 included in the image 450 is detected, and a face region 455 including the face 452 is obtained. Also, face detection information relating to the detected face (the position, size, and so forth of the face) is obtained. For example, as shown in (c) in FIG.
- the face detecting unit 130 outputs such face detection information to the facial expression evaluated value calculating unit 140 and the recording control unit 150 .
- the facial expression evaluated value calculating unit 140 extracts the image (image data) of the face region 455 from the image 450 based on the face detection information output from the face detecting unit 130 . That is to say, the facial expression evaluated value calculating unit 140 accesses the memory (RAM) in which the image data serving as a face detection target is temporally stored, and reads out only the image data within the face region corresponding to the face detection information output from the face detecting unit 130 .
- RAM memory
- the facial expression evaluated value calculating unit 140 transforms the extracted image into an image of a certain size, and normalizes this to generate a normalized face image. That is to say, the extracted image is subjected to resolution conversion as an image of a certain size (certain resolution), thereby generating a normalized face image.
- the image size after normalization is a size serving as a processing unit when the facial expression evaluated value calculating unit 140 calculates a facial expression evaluated value relating to a face.
- a certain size for normalization is a size of 48 pixels ⁇ 48 pixels.
- the facial expression evaluated value calculating unit 140 calculates a facial expression evaluated value that indicates the degree of facial expression included in a normalized face image based on the normalized face image, and the discriminant axis information stored in the discriminant axis information storage unit 141 .
- This calculated facial expression evaluated value is output to the recording control unit 150 .
- This facial expression evaluated value is a value that indicates a degree whether or not the facial expression to be determined is similar to which of two different facial expressions.
- description will be made regarding an example wherein the two different facial expressions are “smiling face” and “normal facial expression”, and evaluation is made that a degree that the facial expression to be determined is “smiling face” is strong as the calculated facial expression evaluated value increases.
- evaluation is made that a degree that the facial expression to be determined is “normal facial expression” is strong as the calculated facial expression evaluated value decreases. Note that this facial expression evaluated value calculating method will be described in detail with reference to FIGS. 9 through 12 .
- FIG. 9 is a diagram conceptually illustrating discriminant axis information used for calculation of a facial expression evaluated value, and a flow at the time of generating this discriminant axis information.
- a facial expression evaluated value calculating method using Fisher's linear discriminant analysis With this facial expression evaluated value calculating method, first, a great number of sample images including one of the two different facial expressions are prepared beforehand. Subsequently, let us consider as a two-class problem between two different facial expressions, linear discriminant analysis (LDA) is executed based on the data of these sample images. Thus, a discriminant axis 466 (shown in (c) in FIG. 9 ) is formed wherein these two different facial expressions are discriminated well. Subsequently, this discriminant axis 466 is stored in the discriminant axis information storage unit 141 . Also, at the time of facial expression evaluation, a facial expression evaluated value is calculated by obtaining an inner product between the data of a face image to be evaluated and the discriminant axis 466 . Description will be made below specifically with reference to the drawings.
- FIG. 9 illustrates a group of smiling face sample images as a smiling face sample image group 461 , and illustrates a group of normal facial expression sample images as a normal facial expression sample image group 462 .
- each sample image included in the sample image groups 461 and 462 is an image normalized such that a face becomes a certain size within a rectangular image of a certain size, for example, of 48 pixels ⁇ 48 pixels.
- the data of each sample image included in the sample image groups 461 and 462 is handled as 48 ⁇ 48 dimensional vector data, thereby executing the LDA process.
- vector space in which the 48 ⁇ 48 dimensional vector data is handled is an extremely high dimensional space including 48 ⁇ 48 coordinate axes. Therefore, before the LDA process is executed, principal component analysis (PCA is executed regarding the 48 ⁇ 48 dimensional vector data. Subsequently, the 48 ⁇ 48 dimensional vector data is converted (dimensionally compressed) into low dimensional space data that represents only the features of a face effectively.
- PCA principal component analysis
- FIG. 10 is a diagram schematically illustrating the masking process for a sample image to be input at the time of the PCA process according to the first embodiment of the present invention.
- (a) illustrates a sample image 470 that belongs to the smiling face sample image group 461 .
- a smiling face 475 is included in the sample image 470 .
- the sample image 470 is an image normalized in a rectangular shape such that the face 475 becomes a certain size.
- the PCA space 465 shown in (b) and (c) in FIG. 9 is partial space including only a coordinate axis that effectively represents the features of a face included in a sample image.
- Each sample image that belongs to the smiling face sample image group 461 or the normal facial expression sample image group 462 that have been dimensionally compressed by the above PCA process is converted into vector data on the PCA space 465 , such as shown in (b) in FIG. 9 , which schematically illustrates a state in which each sample image that belongs to the smiling sample image group 461 or the normal facial expression sample image group 462 is projected on the PCA space 465 .
- FIG. 9 which schematically illustrates a state in which each sample image that belongs to the smiling sample image group 461 or the normal facial expression sample image group 462 is projected on the PCA space 465 .
- the distance between the sample images is relatively near distance.
- a cluster formed by the sample images thereof is taken as a cluster 463 .
- a cluster formed by the sample images thereof is taken as a cluster 464 .
- a projective axis (discriminant axis 466 ) which separates the clusters 463 and 464 appropriately is formed by the LDA process. Such a discriminant axis 466 is referred to as “Fisher's projective axis”.
- a discriminant axis is obtained such that dispersion within a cluster and between clusters projected on an N′-dimensional eigenvector becomes the maximum. That is to say, an eigenvector corresponding to the maximum eigenvalue of each covariance matrix within a cluster and between clusters is obtained, and this is taken as a vector (Fisher vector) on the discriminant axis 466 . Relationship between each covariance matrix and an eigenvalue/eigenvector is shown in Expressions 11 and 12.
- R W represents an intra-cluster covariance matrix
- R B represents an inter-cluster covariance matrix
- ⁇ represents an eigenvalue.
- lower-upper (LU) decomposition method QR decomposition method (Q: orthogonal matrix
- R upper triangular matrix
- Gaussian elimination may be employed, respectively.
- Information such as the coefficient of each component of Fisher vector, and so forth is stored in the discriminant axis information storage unit 141 as the information of the discriminant axis 466 thus calculated.
- the information of the discriminant axis 466 is information including the coefficient of a vector that indicates the discriminant axis of a facial expression obtained by executing linear discriminant analysis based on a signal component obtained by principal component analysis from the image data of a great number of face sample images relating to two different facial expressions, and so forth.
- FIG. 11 is a diagram schematically illustrating relationship between a determinant axis in the pixel space and the PCA space, and a face image to be evaluated.
- the image data of a face (face image) detected from an imaged image is subjected to the PCA process to extract a principal component.
- the facial expression of the face image corresponding to the image data thereof is evaluated as a projective component as to the discriminant axis 466 (vector A d ) of the face image vector subjected to the PCA process (input face image vector), such as shown in the PCA space 465 in FIG. 11 . That is to say, as shown in Expression 15, a facial expression evaluated value E exp can be calculated by inner product between the input face image vector and the Fisher vector.
- the information of the Fisher vector may be converted into information on pixel space S pxl (the dimensional space that the original image data before the PCA process has).
- Expressions 13 and 14 are expressions that represent the input face image vector and Fisher vector as a vector on the pixel space S pxl .
- FIG. 11 conceptually illustrates relationship represented with Expressions 13 and 14.
- a subtraction result between a Fisher vector component and the constant C on the pixel space S pxl can be calculated beforehand. Therefore, this subtraction result and the constant C are stored in the discriminant axis information storage unit 141 as discriminant axis information.
- the facial expression evaluated value calculating unit 140 calculates the vector of the face image detected from the imaged image, following which executes the inner product computation of Expression 16 without subjecting this vector to the PCA process.
- subtraction, multiplication, and addition are executed 48 ⁇ 48 times at the maximum, and actually, only computations of coefficients corresponding to around 40 principal components ⁇ 1 through ⁇ N , are executed.
- Such a facial expression evaluated value calculating method is employed, whereby facial expression evaluation can be executed with high precision while reducing processing load, for example, as compared to a facial expression evaluating method wherein a facial expression is evaluated by matching between a great number of face image templates and the detected face image.
- a facial expression evaluated value calculating method used for the first embodiment of the present invention, after the detected face is normalized to a certain size, the face image thereof is replaced with vector information, whereby this can be applied to an inner product computation as is (or partially masked).
- the inner product computation thereof can be simplified as a simple computation made up of around 40-dimensional subtraction, multiplication, and addition.
- FIG. 12 is a diagram illustrating a calculation example in the case that the facial expression evaluated value according to the first embodiment of the present invention is output as a numeric value.
- the average 482 of the distribution 481 of a smiling face image on the PCA space, and the average 484 of the distribution 483 of a normal facial expression face image on the PCA space are obtained.
- a projective point as to the discriminant axis 466 of the averages 482 and 484 is determined.
- the facial expression evaluated value E exp is converted into a numeric value with a middle point 485 of the projective points of the averages 482 and 484 as a reference. That is to say, as shown in FIG.
- distance between a projective point 486 as to the discriminant axis 466 of the face image, and the middle point 485 of the projective points of the averages 482 and 484 is taken as the facial expression evaluated value E exp
- a side where the smiling face sample image is distributed is taken as a positive numeric value.
- calculation of a facial expression evaluated value is executed regarding each face of these by the facial expression evaluated value calculating unit 140 . Subsequently, the facial expression evaluated value calculated regarding each face of these is stored in the metadata storage unit 210 .
- FIG. 13 is a flowchart illustrating the procedure of a metadata recording process by the image processing apparatus 100 according to the first embodiment of the present invention.
- a moving picture file is input to the moving picture input unit 110 (step S 901 ). Subsequently, the moving picture file input to the moving picture input unit 110 is decoded, the image of one frame in the time-series sequence is obtained (step S 902 ).
- the face detecting unit 130 executes a face detecting process regarding the obtained one frame (step S 903 ).
- the facial expression evaluated value calculating unit 140 executes a facial expression evaluated value calculating process regarding the face detected by the face detecting process (step S 904 ).
- step S 905 determination is made whether or not the obtained one frame is the top frame of the moving picture file input to the moving picture input unit 110 (step S 905 ).
- step S 905 feature points are extracted from the whole of the image corresponding to this top frame (step S 906 ).
- step S 907 the affine transformation parameters of a unit matrix are selected as affine transformation parameters (step S 907 ), “1.0” is calculated as a reliability determination score, and the flow proceeds to step S 913 .
- step S 905 feature points are extracted from a new imaged region with the image corresponding to the immediately previous frame as a reference (step S 905 ). That is to say, a feature point that has already been extracted with the image corresponding to the immediately previous frame can be obtained by the optical flow corresponding to this feature point, and accordingly, this feature point is not extracted with the image corresponding to the current frame. Subsequently, the number of feature points with the whole image corresponding to the immediately previous frame is held (step S 910 ).
- step S 911 the optical flow as to each feature point extracted from the image corresponding to the immediately previous frame is calculated.
- the camera work parameter calculating unit 123 executes an affine transformation parameter calculating process (step S 920 ). This affine transformation parameter calculating process will be described in detail with reference to FIG. 14 . Note that step S 920 is an example of the calculating procedure referred to in the Summary of the Invention.
- a reliability determination score is calculated based on the number of feature points in the held whole image, and the score of the selected representative affine transformation parameter (step S 912 ).
- Each piece of metadata obtained regarding the current frame is recorded in the metadata storage unit 210 in a manner correlated with the current frame (step S 913 ). Note that in the case that the current frame is the top frame, the affine transformation parameters of the selected unit matrix are recorded in the metadata storage unit 210 in a manner correlated with the top frame.
- the image corresponding to the current frame, and the feature points of this image are stored in an overwrite manner (step S 914 ).
- step S 915 determination is made whether or not the current frame is the last frame of the moving picture file input to the moving picture input unit 110 (step S 915 ).
- the flow returns to step S 902 , where the metadata recording process is repeated (steps S 902 through S 914 , and S 920 ).
- the metadata recording process is ended.
- FIG. 14 is a flowchart illustrating an affine transformation parameter calculating procedure (the procedure in step S 920 shown in FIG. 13 ) included in the procedure of the metadata recording process by the image processing apparatus 100 according to the first embodiment of the present invention.
- a variable i is initialized to “ 1 ” (step S 921 ).
- M feature points are selected from the feature points of which the optical flows have been detected (step S 922 ).
- affine transformation parameters are employed as camera work parameters
- three feature points are selected randomly.
- projective-transformation parameters are employed as camera work parameters
- four feature points are selected randomly.
- affine transformation parameters are calculated based on M optical flows calculated corresponding to the selected M feature points (step S 923 ).
- the scores of the affine transformation parameters are calculated (step S 924 ). Specifically, the calculated affine transformation parameters are used to obtain the positions of the movement destinations of all the feature points of the image corresponding to the immediately previous frame. Subsequently, the positions of the feature points obtained with the affine transformation parameters thereof, and the positions of the feature points of the image corresponding to the current frame obtained at the time of calculating an optical flow in step S 911 are compared. A difference value between the positions of two feature points corresponding to each other is calculated for each feature point. For example, absolute distance between two positions corresponding to each other is calculated as a difference value. Subsequently, the calculated difference value and a predetermined threshold are compared for each feature point, and the number of feature points of which the difference value is smaller than the threshold is obtained as the scores of the affine transformation parameters.
- step S 926 “ 1 ” is added to the variable i (step S 925 ), and determination is made whether or not the variable i is greater than a constant N (step S 926 ). In the case that the variable i is equal to or smaller than the constant N (step S 926 ), the flow returns to step S 922 , where the score calculating process for affine transformation parameters is repeated (steps S 922 through S 925 ). For example, 20 may be employed as the constant N.
- the affine transformation parameter having the highest score value is selected as the representative affine transformation parameter (step S 927 ).
- affine transformation parameters are detected as camera work parameters based on the optical flow detected from an image making up a moving picture.
- a sensor such as an acceleration sensor or gyro sensor or the like
- a zoom button used at the time of performing a zoom operation are provided to the camera, and the movement amount of the camera at the time of imaging is detected by this sensor and zoom button.
- camera work parameters may be obtained based on the movement amount of the camera.
- an arrangement may be made wherein multiple camera work parameters are detected by the camera work parameter calculating unit 123 beforehand, and based on the movement amount of the camera detected at the time of imaging, one camera work parameter is selected from the multiple camera work parameters.
- each image shown in FIGS. 15 through 23 is simplified for the sake of explanatory convenience, and of images making up a moving picture, images of which the movement amount between images is comparatively small are illustrated. Also, let us say that each image shown in FIGS. 15 through 23 is an image selected by the compositing target image selecting unit 170 . Also, a case where a subsequent frame is overwritten on a previous frame on the temporal axis is shown as an example.
- FIG. 15 is a diagram illustrating an example of the transition of a moving picture imaged by the camera.
- FIG. 15 illustrates images 401 through 403 making up a moving picture in the case of imaging a person 400 with a mountain as the background. Also, this example illustrates a case where a user is shooting images while moving the direction of the lens of the camera to the right and upper side user.
- the person 400 included in the moving picture to be imaged by the camera moves from the right side to the left side and moves to the lower side in the image making up the moving picture thereof.
- FIG. 16 is a diagram schematically illustrating the locus of a subject included in the images 401 through 403 selected by the compositing target image selecting unit 170 according to an embodiment of the present invention. Note that, with the images 402 and 403 shown in (b) and (c) in FIG. 16 , the image corresponding to the previous frame is shown with a dashed line.
- the image 401 shown in (a) in FIG. 16 is the same as the image 401 shown in (a) in FIG. 15 .
- a solid-line portion of the image 402 shown in (b) in FIG. 16 is the same as the image 402 shown in (b) in FIG. 15
- arrows 404 through 406 of the image 402 shown in (b) in FIG. 16 are arrows that indicate the locus of a subject included in the images 401 and 402 .
- a solid-line portion of the image 403 shown in (c) in FIG. 16 is the same as the image 403 shown in (c) in FIG. 15
- a dashed-line portion of the image 403 shown in (c) in FIG. 16 is the same as the solid-line portion of the image 402 shown in (b) in FIG. 16 .
- arrows 407 through 409 of the image 403 shown in (c) in FIG. 16 are arrows that indicate the locus of a subject included in the images 402 and 403 .
- the person 400 and the mountain of the background included in the image are moved along with movement of the camera. Based on each optical flow detected by this movement, affine transformation parameters between images can be obtained. For example, in the case that the image 401 is taken as a reference image, the affine matrix of the affine transformation parameters calculated regarding each image of the images 401 and 402 is multiplied, whereby affine transformation parameters used for subjecting the image 402 to affine transformation can be obtained. Also, the affine matrix of the affine transformation parameters calculated regarding each image of the images 401 through 403 is multiplied, whereby affine transformation parameters used for subjecting the image 403 to affine transformation can be obtained. Now, let us say that the affine transformation parameters of the reference image are the affine transformation parameters of a unit matrix.
- FIG. 17 is a diagram illustrating an image compositing example in the case that images are composited by the image compositing unit 220 according to the first embodiment of the present invention.
- description will be made regarding a case where the images 402 and 403 are subjected to affine transformation using the affine transformation parameters stored in the compositing target image metadata storage unit 270 with the image 401 as a reference image.
- the image transforming unit 190 subjects the image 401 to affine transformation by the matrix of A 1 with the position and size of the image 401 of the top frame as a reference.
- the image 401 is a reference image, and accordingly, A 1 is a unit matrix. Therefore, the position and size of the image 401 are not transformed.
- the image transforming unit 190 subjects the image 402 to affine transformation using the affine transformation parameters correlated with the image 402 .
- the matrix of the affine transformation parameters correlated with the image 402 is A 2 .
- the image transforming unit 190 subjects the image 402 to affine transformation by the matrix of A 2 with the position and size of the image 401 of the top frame as a reference. Note that, with the example shown in (b) in FIG. 17 , as shown in arrows 415 and 416 , the position alone of the image 402 is transformed.
- the image compositing unit 220 overwrites the image 402 subjected to affine transformation by the image transforming unit 190 on the image 401 corresponding to the previous frame so as to be overlaid thereupon and both be composited. Specifically, of the region of the image 401 , a region 410 overlaid with the image 402 is overwritten by the image of the image 402 . Also, of the region of the image 401 , a region 411 not overlaid with the image 402 is not overwritten by the image of the image 402 . That is to say, in the case that the image 402 corresponding to the second frame is composited, as shown in (b) in FIG. 17 , an image is created wherein the whole portion of the image 402 , and the portion corresponding to the region 411 of the image 401 are composited.
- the image transforming unit 190 subjects the image 403 to affine transformation using the affine transformation parameters correlated with the image 403 .
- the matrix of the affine transformation parameters corresponding to the image 403 is A 3 .
- the image transforming unit 190 subjects the image 403 to affine transformation by the matrix A 3 with the position and size of the image 401 as a reference. Note that, with the image shown in (c) in FIG. 17 , as shown in arrows 417 and 418 , the position alone of the image 403 is transformed.
- the image compositing unit 220 overwrites the image 403 subjected to affine transformation by the image transforming unit 190 on the composited image of the images 401 and 402 corresponding to the previous frame so as to be overlaid thereupon and both be composited. Specifically, of the region of the composited image of the images 401 and 402 , regions 413 and 414 overlaid with the image 403 are overwritten by the image of the image 403 . Also, of the region of the composited image of the images 401 and 402 , regions 411 and 412 not overlaid with the image 403 are composited with the composited image of the images 401 and 402 .
- FIG. 18 is a diagram illustrating an example of the transition of a moving picture imaged by the camera.
- FIG. 18 illustrates images 421 through 423 corresponding to consecutive frames included in a moving picture in the case of imaging a person 420 with a mountain as the background.
- This example illustrates a case where a user is shooting images while zooming in.
- the size of the person 420 included in the moving picture imaged by the camera is increasing gradually on an image making up the moving picture thereof. Note that, though the position of the camera is somewhat moved at the time of zooming in some cases, with this example, description will be made without taking movement of the position of the camera into consideration.
- FIG. 19 is a diagram schematically illustrating the locus of a subject included in the images 421 through 423 selected by the compositing target image selecting unit 170 according to an embodiment of the present invention. Note that an affine transformation parameter calculation example shown in FIG. 19 is the same as the affine transformation parameter calculation example shown in FIG. 16 , and accordingly, description thereof will be omitted here.
- FIG. 20 is a diagram illustrating an image compositing example in the case that images are composited by the image compositing unit 220 according to the first embodiment of the present invention. Note that an image compositing example shown in FIG. 20 is the same as the image compositing example shown in FIG. 17 , and accordingly, description thereof will be omitted here.
- FIG. 21 is a diagram illustrating an example of the transition of a moving picture imaged by the camera.
- FIG. 21 illustrates images 441 through 443 corresponding to consecutive frames included in a moving picture in the case of imaging a person 440 with a mountain as the background.
- This example illustrates a case where a user is performing imaging while rotating the camera with the optical axis direction as the rotation center.
- the person 440 included in the moving picture imaged by the camera is rotating gradually on an image making up the moving picture thereof. Note that, though the position of the camera is somewhat moved due to the rotation of the camera in some cases, with this example, description will be made without taking movement of the position of the camera into consideration.
- FIG. 22 is a diagram schematically illustrating the locus of a subject included in the images 441 through 443 selected by the compositing target image selecting unit 170 according to an embodiment of the present invention. Note that an affine transformation parameter calculation example shown in FIG. 22 is the same as the affine transformation parameter calculation example shown in FIG. 16 , and accordingly, description thereof will be omitted here.
- FIG. 23 is a diagram illustrating an image compositing example in the case that images are composited by the image compositing unit 220 according to the first embodiment of the present invention. Note that an image compositing example shown in FIG. 23 is the same as the image compositing example shown in FIG. 17 , and accordingly, description thereof will be omitted here.
- the zoom rate is small, i.e., change in zoom is small.
- Movement amount in the horizontal direction or vertical direction is equal to or greater than a certain number of pixels.
- the length of an image compositing target section is length within a certain range (e.g., within a range of 3 seconds to 20 seconds).
- a change rate between frames is within a certain range.
- the change rate between frames is a value that indicates each component of camera work calculated by affine transformation parameters relating to between frames.
- FIG. 24 is a diagram illustrating a reliability determination score calculated by the camera work detecting unit 120 according to the first embodiment of the present invention, in a time-series manner.
- (a) schematically illustrates a frame serving as a reliability determination score calculation target using a rectangle. Note that, in (a) in FIG. 24 , a frame serial number is appended to the inside of a rectangle that indicates each frame.
- (b) illustrates a graph that represents a reliability determination score calculated regarding frames 1 through 15 shown in (a) in FIG. 24 , for each frame.
- the horizontal axis is an axis that indicates imaging time
- the vertical axis is an axis that indicates the value of a reliability determination score calculated regarding each frame.
- the threshold of the above condition (1) is a threshold TH 1 .
- a value of 0.5 through 0.8 can be set as the threshold TH 1 .
- a reliability determination score SHS is a value calculated by the reliability determination score calculating unit 124 , and is stored in the metadata storage unit 210 in a manner correlated with each frame making up the moving picture file stored in the moving picture storage unit 200 .
- the certain range in the above condition (4) is a range between K 1 and K 2 .
- a reliability determination score calculated regarding each frame is represented with a white circle. Of white circles that represent a reliability determination score, a score exceeding the threshold TH 1 is represented with a heavy-line circle.
- the reliability determination score of frames 1 and 3 are below the threshold TH 1 , and accordingly, the frames 1 and 3 are not determined to be an image compositing target section candidate.
- the reliability determination score of a frame 2 exceeds the threshold TH 1 , but the reliability scores of the adjacent frames 1 and 3 are below the threshold TH 1 , and accordingly, the above condition (4) is unsatisfied. Therefore, the frame 2 is not determined to be an image compositing target section candidate.
- the reliability determination scores of frames 4 through 14 exceed the threshold TH 1 .
- the length of the frame 14 exceeds the certain range (range between K 1 and K 2 ) of the above condition (4).
- the sections selecting unit 160 extracts consecutive frames, included within the certain range of the above condition (4), of which the reliability determination score SHS exceeds the threshold TH 1 , as image compositing target section candidates.
- the section selecting unit 160 determines whether or not the change rate of each frame (each component of camera work) included in the image compositing target section candidates satisfies the following conditions (11) through (17).
- XZth 1 ⁇ XZ ⁇ XZth 2 (11)
- YZth 1 ⁇ YZ ⁇ YZth 2 (12)
- XT ⁇ XTth (13)
- YT ⁇ YTth (14)
- XZ represents a zoom component in the X direction of camera work
- YZ represents a zoom component in the Y direction of camera work
- XT represents a translational component in the X direction of camera work
- YT represents a translational component in the Y direction of camera work
- XR represents a rotational component in the X direction of camera work
- YR represents a rotational component in the Y direction of camera work
- XZth 1 , XZth 2 , YZth 1 , YZth 2 , XTth, YTth, XRth, YRth, and ADRth represent a threshold.
- a value of 0.7 through 0.9 may be set as the thresholds XZth 1 and YZth 1 .
- a value of 1.1 through 1.3 may be set as the thresholds XZth 2 and YZth 2 .
- a value of W/20 through W/5 may be set as the threshold XTth.
- W is a value that indicates the number of pixels in the horizontal direction of an image to be processed.
- a value of H/20 through H/5 may be set as the threshold YTth.
- H is a value that indicates the number of pixels in the vertical direction of an image to be processed.
- 3 deg through 5 deg may be set as the thresholds XRth and YRth.
- 0.5 deg through 1.0 deg may be set as the threshold ADRth.
- FIG. 25 is a diagram illustrating an integral zoom component calculated with affine transformation parameters calculated by the camera work detecting unit 120 according to the first embodiment of the present invention, in a time-series manner.
- (a) schematically illustrates a frame serving as an integral zoom component calculation target using a rectangle. Note that frames 1 through 15 shown in (a) in FIG. 25 are the same as those shown in (a) in FIG. 24 .
- FIG. 25 illustrates a graph that represents an integral zoom component in the horizontal direction (X direction) calculated regarding the frames 1 through 15 shown in (a) in FIG. 25 for each frame.
- the horizontal axis is an axis that indicates imaging point-in-time
- the vertical axis is an axis that indicates the value of an integral zoom component calculated regarding each frame.
- This integral zoom component is the value of a zoom component of an affine matrix calculated by multiplication of the affine matrix of the affine transformation parameters correlated with each frame from the reference frame to the target frame.
- the affine transformation parameters correlated with the reference frame are the values of a unit matrix.
- the value of the integral zoom component of the reference frame is 1.0
- the value of the integral zoom component of the target frame is a relative value as to the reference frame.
- this integral zoom component is calculated regarding a zoom component in the X direction and a zoom component in the Y direction of camera work. Note that, with the example shown in FIG. 25 , the integral zoom component calculated regarding the zoom components in the X direction is illustrated as an example.
- the integral zoom component calculated regarding each frame is illustrated with a white circle.
- the white circle corresponding to the top frame is illustrated as an integral zoom component 501 .
- the white circle that indicates the maximum value is illustrated as an integral zoom component 502
- the white circle that indicates the minimum value is illustrated as an integral zoom component 503 .
- the horizontal axis of the graph shown in (b) in FIG. 25 indicates each frame number according to imaging point-in-time.
- the value of the integral zoom component 502 that indicates the maximum value is illustrated as M 2
- the value of the integral zoom component 503 that indicates the minimum value is illustrated as M 1 .
- the zoom rate used for the condition (2) is a value that indicates a ratio between the maximum value and the minimum value of the integral zoom components.
- This condition (2) is determined based on whether or not the zoom rate as to each frame included in the image compositing target section candidates is included in a certain range.
- This certain range may be, for example, a range below the threshold XZth 2 such as shown in the following. XZm 1 /XZm 2 ⁇ XZth 2
- XZm 1 is the maximum value of the integral zoom components in the X direction calculated regarding each frame included in the image compositing target section candidates
- XZm 2 is the minimum value of the integral zoom components thereof. Note that, for example, a value of 1.5 through 3.0 may be set as the threshold XZth 2 .
- the zoom rate calculated regarding the zoom components in the Y direction can be determined based on whether or not the following expression is satisfied. YZm 1 /YZm 2 ⁇ YZth 2
- YZm 1 is the maximum value of the integral zoom components in the Y direction calculated regarding each frame included in the image compositing target section candidates
- YZm 2 is the minimum value of the integral zoom components thereof. Note that, for example, a value of 1.5 through 3.0 may be set as the threshold YZth 2 .
- the zoom rates in the X direction and in the Y direction calculated regarding each frame included in the image compositing target section candidates that satisfy the conditions (11) through (17) satisfy the condition (2) determination is made whether or not the condition (3) is satisfied.
- the zoom rates in the X direction and in the Y direction calculated regarding each frame included in the image compositing target section candidates do not satisfy the condition (2), the image compositing target section candidate thereof is not determined to be an image compositing target section.
- FIG. 26 is a diagram illustrating an integral translational component calculated with affine transformation parameters calculated by the camera work detecting unit 120 according to the first embodiment of the present invention, in a time-series manner.
- (a) schematically illustrates a frame serving as an integral translational component calculation target using a rectangle. Note that the frames 1 through 15 shown in (a) in FIG. 26 are the same as those shown in (a) in FIG. 24 .
- FIG. 26 illustrates a graph that represents an integral translational component in the horizontal direction (X direction) calculated regarding the frames 1 through 15 shown in (a) in FIG. 26 for each frame.
- the horizontal axis is an axis that indicates imaging point-in-time
- the vertical axis is an axis that indicates the value of an integral translational component calculated regarding each frame.
- This integral translational component is the value of a translational component of an affine matrix calculated by multiplication of the affine matrix of the affine transformation parameters correlated with each frame from the reference frame to the target frame. Now, let us say that the affine transformation parameters correlated with the reference frame is the value of a unit matrix.
- the value of the integral translational component of the reference frame is 0.0
- the value of the integral translational component of the target frame is a relative value as to the reference frame.
- this integral translational component is calculated regarding a translational component in the X direction and a translational component in the Y direction of camera work. Note that, with the example shown in FIG. 26 , the integral translational component calculated regarding the translational components in the X direction is illustrated as an example.
- the integral translational component calculated regarding each frame is illustrated with a white circle.
- a white circle that indicates an integral translational component calculated regarding each frame included in the image compositing target section candidates 500 is illustrated with a heavy line.
- the horizontal axis of the graph shown in (b) in FIG. 26 indicates each frame number according to imaging point-in-time.
- the thresholds XTth 2 and ⁇ XTth 2 used for the condition (2) are indicated.
- the condition (3) is determined based on whether or not movement amount in the horizontal direction or vertical direction is equal to or greater than a certain number of pixels. Specifically, determination is made based on whether or not at least one of panning and tilting is relatively great. For example, such as shown in the following, determination is made based on whether or not the absolute value of the integral translational component XT 1 exceeds the threshold XTth 2 .
- a value of 3W through 10W may be set as the threshold XTth 2 .
- W is a value that indicates the number of pixels in the horizontal direction of an image.
- determination may be made based on whether or not the following expression is satisfied.
- a value of 3H through 10H may be set as the threshold YTth 2 .
- H is a value that indicates the number of pixels in the vertical direction of an image.
- the values of the integral translational components in the X direction calculated regarding the frames 12 and 13 included in the image compositing target section candidates 500 exceed the threshold XTth 2 . That is to say, with the example shown in FIG. 26 , the absolute value of an integral translational component in the X direction calculated regarding each frame included in image compositing target section candidates 500 exceeds the threshold XTth 2 , and accordingly, determination is made that the condition (3) is satisfied. Also, even in the case that the absolute value of an integral translational component in the Y direction calculated regarding each frame included in image compositing target section candidates 500 exceeds the threshold XTth 2 , determination is made that the condition (3) is satisfied.
- the image compositing target section candidates determined to satisfy the condition (3) are selected by the section selecting unit 160 as image compositing target sections.
- the image compositing target section candidates thereof are not determined to be image compositing target sections.
- determination is made whether or not the conditions (11) through (17) are satisfied regarding each frame included in the image compositing target section candidates that satisfy the conditions (1) and (4). Subsequently, determination is made whether or not the conditions (2) and (3) are satisfied regarding each frame included in the image compositing target section candidates that satisfy the conditions (11) and (17). However, these determinations may be made in a different sequence. Also, of the conditions (1) through (4), and the conditions (11) through (17), one of the determination conditions may be omitted. Also, for example, an image compositing target section may be selected using another determination condition regarding each frame included in the image compositing target section candidates. For example, it may be set as a determination condition that a rotational rate is small, using an integral rotational component.
- All of the frames included in the image compositing target section thus selected are used to create a composited image.
- an experience which is not experienced by simple playback of a moving picture, may be provided to a user by creating a composited image, wherein multiple still images imaged over space of a relatively wide range at the same imaging place are joined together. That is to say, a panorama-style composited image made up of still images which is not experienced by from playback of a moving picture (e.g., such as an image in which photos are joined together) can be provided to the user. Therefore, with the first embodiment of the present invention, frames used for creation of a composited image are further selected from each frame included in the image compositing target section.
- an image including a face is selected preferentially at the image compositing target section, this image is selected a front-most frame, another frame is selected based on the overlaying rate with this front-most frame.
- an image having the highest facial expression evaluated value that indicates the degree of a smiling face may be selected as a front-most frame.
- a blurring condition evaluated value or the like other than a facial expression evaluated value may be employed for determination.
- FIGS. 27A and 27B are diagrams schematically illustrating an image included in the image compositing target section selected by the section selecting unit 160 according to the first embodiment of the present invention, in a time-series manner.
- FIG. 27A illustrates the transition of an image in the case that a user has been shooting images while moving the direction of the lens of the camera in the horizontal direction with a stationary dog, person, and car as subjects.
- FIG. 27B illustrates the transition of an image in the case that a user has been shooting images while moving the direction of the lens of the camera in the horizontal direction with a tree and a mountain as subjects. Note that the images shown in FIGS. 27A through 27B are simplified for the sake of explanatory convenience, and the movement amount between consecutive two frames is illustrated greatly.
- the images 512 through 515 include a person's face. Also, let us say that of the faces included in the images 512 through 515 , the facial expression evaluated value calculated regarding the face included in the image 513 is the highest value.
- the compositing target image selecting unit 170 selects the image including a face as the front-most frame. Also, as shown in FIG. 27A , in the case that multiple images include a face, the image including a face having the highest value of facial expression evaluated value calculated regarding these faces is selected as the front-most frame. With the example shown in FIG. 27A , the image 513 is selected.
- images 521 through 528 shown in FIG. 27B do not include a person's face.
- the compositing target image selecting unit 170 selects the top image on the time axis of images included in the image compositing target section as the front-most frame.
- the image 521 corresponding to the top frame included in the image compositing target section is selected.
- the front-most frame thus selected from the image compositing target section is taken as a reference frame, and another frame is selected base on this reference frame.
- description will be made regarding an example wherein a frame is selected based on overlaid regions between images include in the image compositing target section.
- FIG. 28 is a diagram schematically illustrating a selecting method in the case that a compositing target image is selected by the compositing target image selecting unit 170 according to the first embodiment of the present invention.
- description will be made regarding a case where two work buffers are used to select another frame as to the front-most frame.
- a first work buffer 531 and a second work buffer 532 are work buffers that represent each pixel making up an image by one bit. That it to say, in the case that an image is pasted to the first work buffer 531 and the second work buffer 532 , a bit at the pasting position is set to on. Also, a selected image is pasted to the first work buffer 531 as a composited image, and one image serving as a selection determination target is subjected to affine transformation and pasted to the second work buffer 532 . Subsequently, an overlaid portion between the composited image pasted to the first work buffer 531 , and the image pasted to the second work buffer 532 is calculated.
- the area of the region of the calculated overlaid portion is an area C
- the area of the composited image pasted to the first work buffer 531 is an area A
- the area of the image pasted to the second work buffer 532 is an area B.
- a value of the area C/the area A, and a value of the area C/the area B are calculated, and of these two values, the greater value is calculated as an overlaying rate.
- the image thereof is selected. For example, 60% may be set as this threshold.
- the image wherein the calculated overlaying rate is below the threshold is selected as a compositing target image, and is also pasted to the first work buffer 531 .
- the same determining process is executed regarding the next frame.
- This affine transformation is executed using the affine transformation parameters stored in the metadata storage unit 210 . Note that the example shown in FIG. 28 is an example wherein of the image 536 , only the position has been transformed, and the angle and size have not been transformed.
- an overlaid portion 538 between the composited image (composited image of the images 534 and 535 ) pasted to the first work buffer 531 , and the image 536 pasted to the second work buffer 532 is calculated.
- a work buffer range 533 shown in (b) in FIG. 28 is a range that indicates the size corresponding to the first work buffer 531 and the second work buffer 532 .
- the calculated overlaid portion 538 is represented with a rectangle shaded internally, and other image portions are represented with a dotted line.
- the area of the calculated overlaid portion 538 is an area C 1
- the area of the composited image (composited image of the images 534 and 535 ) pasted to the first work buffer 531 is an area A 1
- the area of the image 536 pasted to the second work buffer 532 is an area B 1 .
- the value of the area A 1 is greater than the value of the area B 1
- a value of the area C 1 /the area B 1 is greater than a value of the area C 1 /the area A 1 . Therefore, the value of the area C 1 /the area B 1 is calculated as an overlaid rate.
- the image 536 is selected as a compositing target image.
- the values of a matrix are calculated wherein the affine matrixes of the affine transformation parameters from the front-most frame to the current frame are multiplied.
- the values of a unit matrix are used.
- the affine transformation parameters of the calculated matrix, and the frame number and face data of the selected compositing target image are recorded in the compositing target image metadata storage unit 270 .
- the compositing target image selected as the front-most image information that indicates the front-most frame is recorded in the compositing target image metadata storage unit 270 along with such information.
- the image 536 is pasted to the first work buffer 531 .
- the composited image of the images 534 through 536 is pasted to the first work buffer 531 .
- a determining process is executed regarding the next image 537 in the same way.
- the image 536 is not pasted to the first work buffer 531 . In this case, the determining process is executed between the next image, the images 534 and 535 .
- a compositing target image is selected by calculating an overlaying rate sequentially using the two work buffers, but a compositing target image may be selected by calculating an overlaying rate using the affine transformation parameters alone stored in the metadata storage unit 210 .
- FIG. 29 is a diagram schematically illustrating a moving picture to be determined by the section selecting unit 160 and the compositing target image selecting unit 170 , according to the first embodiment of the present invention, in a time-series manner.
- (a) schematically illustrates a frame making up a moving picture file 540 that is a moving picture file stored in the moving picture storage unit 200 .
- (b) illustrates an image compositing target section group 541 that is a group of the image compositing target sections selected regarding the moving picture file 540 .
- FIG. 29 (a) schematically illustrates a frame making up a moving picture file 540 that is a moving picture file stored in the moving picture storage unit 200 .
- (b) illustrates an image compositing target section group 541 that is a group of the image compositing target sections selected regarding the moving picture file 540 .
- FIG. 29 (c) illustrates a face detection section group 544 that is a group of face detection sections that indicate a section where a face has been detected regarding each frame making up the moving picture file 540 .
- FIG. 29 (d) illustrates a smiling face detection section group 546 that is a group of smiling face detection sections that indicate a section where a face detected regarding each frame making up the moving picture file 540 has been determined to be a smiling face.
- FIG. 29 illustrates the front-most frame selected at the image compositing target section, and a search direction 548 that is a direction where another frame is searched with the front-most frame as a reference frame.
- a search direction 548 that is a direction where another frame is searched with the front-most frame as a reference frame.
- FIG. 29 (f) illustrates a selection frame group 554 that is a group of frames (compositing target images) selected regarding the moving picture file 540 .
- each frame making up the moving picture file 540 is illustrated with a rectangle to which a frame number is appended internally.
- each section is illustrated with a rectangle shaded internally.
- the front-most frame is represented with a white circle, and the search direction from the front-most frame is represented with an arrow.
- the image compositing target sections 542 and 543 have been selected by the section selecting unit 160 regarding the moving picture file 540 .
- a face has been detected at the face detection section 545 , of frames included in the image compositing target section 542 .
- a facial expression evaluated value has been a certain value or more at the smiling face detection section 547 , and has been determined to be a smiling face.
- a frame having the highest facial expression evaluated value is a frame h.
- the frame h is selected as the front-most frame 549 at the image compositing target section 542 .
- a frame selecting process is executed toward a search direction 550 that is a front-sided search direction on the temporal axis, and a search direction 551 that is a rear-sided search direction on the temporal axis.
- This selecting method is the same as the selecting method shown in FIG. 28 . In this case, for example, let us say that frames g+2, . . . , h, . . . , h+m, . . . , i have been selected.
- the frame j+1 on the most front side on the temporal axis is selected as a front-most frame 552 .
- a frame is selected toward a search direction 553 that is a search direction on the rear side on the temporal axis.
- This selecting method is the same as the selecting method shown in FIG. 28 . In this case, for example, let us say that frames j+1, . . . , j+n, . . . , k ⁇ 3 have been selected.
- a single or multiple image compositing target sections are selected regarding a single moving picture file, and multiple frames are selected from the image compositing target sections thereof.
- a composited image is created using the multiple frames thus selected.
- the identification number of the selected image compositing target section, and the selected each frame number are recorded in the compositing target image metadata storage unit 270 in a correlated manner.
- the calculated affine transformation parameters, front-most frame information that indicates the front-most frame, and face data that is data relating to a detected face regarding a frame are recorded in the compositing target image metadata storage unit 270 in a correlated manner.
- a composited image can be displayed by the user operations using each piece of information stored in the compositing target image metadata storage unit 270 . For example, a composited image of the frames g+2, . . . , h, . . . , h+m, . . .
- i selected at the image compositing target section 542 can be displayed. Also, a composited image of frames j+1, . . . , j+n, . . . , k ⁇ 3 selected at the image compositing target section 543 can be displayed.
- the front-most frame may be selected using the position and size included in face detection information. For example, an image may be selected as the front-most frame wherein the position of the detected face is shifted to the center, and the size thereof is relatively great.
- FIG. 30 is a diagram illustrating an example of a compositing target image selected regarding a moving picture imaged by the camera.
- FIG. 30 illustrates images 561 through 568 selected regarding a moving picture in the case that a lady standing with the slope of a mountain including rocks as the background is imaged while moving the camera. Let us say that this moving picture has been imaged in time-series along an arrow direction with the image corresponding to the frame on the front side on the temporal axis as an image 561 . Also, let us say that the position of the lady serving as a subject is generally the center of the image, and an image 566 serving as a smiling face has been selected as the front-most frame.
- FIGS. 31 and 32 are diagrams illustrating the transition of a composited image created by the image compositing unit 220 according to the first embodiment of the present invention.
- Images 571 through 578 shown in FIGS. 31 and 32 are images wherein a white frame is added to the peripheries of the images 561 through 568 shown in FIG. 30 , and are the same as the images 561 through 568 except that a white frame is appended thereto.
- the size of a composited image to be created regarding the images 571 through 578 is calculated using the affine transformation parameters correlated with other images 561 through 575 , 577 , and 568 .
- the position and size of the front-most frame is calculated.
- other images are subjected to affine transformation sequentially, and are composited in the image memory 230 . For example, as shown in (a) through (d) in FIG. 31 , and (a) in FIG.
- the top framer through the immediately previous frame of the front-most image are composited sequentially in time series. That is to say, the images 571 through 575 are composited sequentially.
- the last frame through the immediately previous frame of the front-most image are composited sequentially in time series. That is to say, the images 577 and 578 are composited sequentially.
- the front-most image 576 selected at the same image compositing target section is composited.
- the images are overlaid from each of the top frame and the last frame regarding each frame selected at the same image compositing target section, and finally, the image selected as the front-most frame is overwritten thereupon.
- the image selected as the front-most frame is displayed on the top side, whereby a clear composited image can be created.
- a metadata file stored in the compositing target image metadata storage unit 270 is used to composite partial images making up the moving picture, whereby a panorama-style composited image that represents one scene included in the moving picture thereof can be created.
- a white frame is added to frames that satisfy a certain condition to composite the frames without using all the consecutive frames making up the moving picture, whereby a composited image can be displayed such that photos are adhered together.
- an interesting image that completely differs from an image viewed at the time of common moving picture playback can be provided.
- a panorama-style composited image can be displayed automatically regarding each scene included in the moving picture, whereby the user can readily understand whether or not the moving picture thereof is made up of what kinds of scenes.
- an imaged scene can be displayed selectively, whereby a significant image can be provided to the user.
- a person's smiling face image is taken as the front-most image, and based on this image another image serving as a compositing target is selected, whereby a panorama composited image can be provided such that accent is given to the person of a smiling face thereof.
- FIG. 33 is another compositing example of a composited image created by the image compositing unit 220 according to the first embodiment of the present invention.
- (a) illustrates a display method selection screen 580 to select a display method for a composited image created by the image compositing unit 220 and displayed on the display unit 250 .
- Check fields 581 and 582 , a “panorama-style composited image display” button 583 , and a “cancel” button 584 are provided to the display method selection screen 580 .
- the check field 581 is, as shown in FIGS. 31 and 32 , a check field used for selecting a display method wherein the front-most frame is overlaid and composited on the top side and displayed.
- the check field 582 is, as shown in (b) in FIG. 33 , a check field used for selecting a display method wherein each frame is composited and displayed while being overlaid in the time-series sequence. Operation input to add a check mark to these check fields is performed using a cursor 585 .
- the “panorama-style composited image display” button 583 is a button to be pressed at the time of displaying a composited image on the display unit 250 . That is to say, after a check mark is added to one of the check fields 581 and 582 , the “panorama-style composited image display” button 583 is pressed using the cursor 585 , whereby images composited by the compositing method to which a check mark has been added are displayed sequentially.
- the “cancel” button 584 is a button to be pressed in the case of eliminating the check mark added to one of the check fields 581 and 582 .
- the “panorama-style composited image display” button 583 has been pressed in a state in which a check mark is added to the check field 581 , as shown in FIGS. 32 and 33 , the frames on both sides of the front-most frame are composited sequentially, and the front-most frame is overlaid and composited on the top side.
- the “panorama-style composited image display” button 583 has been pressed in a state in which a check mark is added to the check field 582 , as shown in (b) in FIG. 33 , each frame is composited and displayed by being overlaid in the time-series sequence.
- FIG. 33 illustrates the images 571 through 578 have been composited in the time-series sequence.
- the images 571 through 578 are the same as the images 571 through 578 shown in FIGS. 31 and 32 .
- imaging time can be clearly understood with reference to the displayed composited image.
- the next image is overwritten up to a region near a person included in the image 576 selected as the front-most frame, there is a possibility that the person thereof may be hidden. Therefore, for example, face data stored in the compositing target image metadata storage unit 270 in correlated with the image 576 is used, whereby other images can be prevented from overlaying on a certain range from the face region.
- FIG. 33 schematically illustrates a case where an image overwritten within a certain range (e.g., distance L 1 ) from a face region 586 included in the image 576 corresponding to the front-most frame.
- a certain range e.g., distance L 1
- the image 577 is overwritten within the distance L 1 from the face region 586 included in the image 576 . Therefore, as shown in (c) in FIG. 33 , the image 577 is moved in an arrow direction 587 so as to be apart the distance L 1 or more from the face region 586 included in the image 576 .
- the images after the moved image are also moved by the same distance. That is to say, the image 578 is moved by the same distance as the image 577 .
- the surroundings of a person's face included in the image 576 are seen, whereby a composited image can be further enjoyed.
- an arrangement may be made wherein composited images relating to a single or multiple moving pictures are created and stored beforehand, each composited image thus stored is displayed based on the user's operation.
- an arrangement may be made wherein the compositing transition of a composited image is stored sequentially for each image beforehand, such transition is displayed sequentially.
- FIG. 34 is a flowchart illustrating the procedure of a compositing target image metadata recording process by the image processing apparatus 100 according to the first embodiment of the present invention.
- the section selecting unit 160 executes a section extracting process (step S 930 ). This section selecting process will be described in detail with reference to FIG. 35 . Subsequently, the compositing target image selecting unit 170 executes a frame selecting process (step S 950 ). This frame selecting process will be described in detail with reference to FIG. 36 . Note that steps S 930 and S 950 are examples of the selecting procedure referred to in the Summary of the Invention.
- FIG. 35 is a flowchart illustrating a section extracting procedure (the procedure in step S 930 shown in FIG. 34 ) included in the procedure of the compositing target image metadata recording process by the image processing apparatus 100 according to the first embodiment of the present invention.
- a variable t is initialized to “ 1 ” (step S 931 ), and a frame t making up a moving picture file is set as a reference frame (step S 932 ). Subsequently, the affine translation parameter and reliability determination score corresponding to the frame t are obtained from a metadata file stored in the metadata storage unit 210 (step S 933 ).
- elapsed time T from the reference frame to the frame t is calculated (step S 934 ), and determination is made whether or not the elapsed time T exceeds a threshold TM 1 (step S 935 ).
- the flow proceeds to step S 946 .
- determination is made whether or not the reliability determination score SHS corresponding to the frame t exceeds the threshold TH 1 (step S 936 ).
- each component of the camera work is calculated from the affine transformation parameters corresponding to the frame t (step S 937 ). Subsequently, determination is made whether or not all the values of the components of the camera work are included in the range of a threshold (step S 938 ). In the case that all the values of the components of the camera work are included in the range of a threshold (step S 938 ), the zoom rate corresponding to each frame from the reference frame to the frame t is calculated (step S 939 ).
- step S 940 determination is made whether or not the calculated zoom rate is included in the range of a threshold.
- step S 940 the movement amount corresponding to each frame from the reference frame to the frame t (integral translational component) is calculated (step S 941 ).
- step S 942 determination is made whether or not the calculated movement amount is included in the range of a threshold.
- step S 943 determination is made whether or not the frame t is the last frame, of frames making up a moving picture to be processed.
- step S 943 In the case that the frame t is not the last frame (step S 943 ), “ 1 ” is added to the variable t (step S 944 ), and the flow returns to step S 933 , where the section selecting process is repeated. On the other hand, in the case that the frame t is the last frame (step S 943 ), the section selecting process is ended.
- step S 936 the reliability determination score SHS corresponding to the frame t does not exceed the threshold TH 1 (step S 936 ), or in the case that all the values of the components of the camera work are not included in the range of a threshold (step S 938 )
- step S 945 the flow proceeds to step S 945 .
- step S 945 determination is made whether or not the elapsed time T is below an image compositing target section range TM 2 (step S 945 ).
- the section from the reference frame to the frame t is selected as an image compositing target section (step S 946 ).
- step S 945 “ 1 ” is added to the variable t (step S 947 ), and the flow returns to step S 932 , where the section selecting process is repeated.
- FIG. 36 is a flowchart illustrating a frame selecting procedure (the procedure in step S 950 shown in FIG. 34 ) included in the procedure of the compositing target image metadata recording process by the image processing apparatus 100 according to the first embodiment of the present invention.
- the affine transformation parameters, face detection information, and facial expression evaluated value corresponding to each frame included in one of the image compositing target section selected by the section selecting unit 160 are obtained from a metadata file store in the metadata storage unit 210 (step S 951 ). Subsequently, the size for compositing the image corresponding to each frame is calculated using the affine transformation parameters corresponding to each frame included in the obtained image compositing target section, and two work buffers having this size are secured (step S 952 ).
- step S 953 determination is made based on the obtained face detection information whether or not a face is included in the image corresponding to each frame included in the image compositing target section serving as a selection target.
- the image having the highest value of the facial expression evaluated value corresponding to each image including a face is selected as the front-most frame (step S 954 ).
- the top frame of the frames included in the image compositing target section is selected as the front-most frame (step S 955 ).
- step S 956 the selected front-most frame is pasted to the first work buffer (step S 956 ), and each piece of metadata relating to the selected front-most frame is recorded in the compositing target image metadata storage unit 270 (step S 957 ).
- step S 958 an overlaying rate between the selected front-most frame, and the frame adjacent to one of the directions of the front-most frame is calculated (step S 958 ). That is to say, the frame serving as an overlaying rate calculation target is pasted to the second work buffer, and an overlaying rate as to the front-most frame pasted to the first work buffer is calculated.
- step S 959 determination is made whether or not the calculated overlaying rate is below a threshold.
- the frame serving as the overlaying rate calculation target is selected, and the selected frame is pasted to the first work buffer (step S 960 ).
- each piece of metadata relating to the selected frame is recorded in the compositing target image metadata storage unit 270 (step S 961 ), and the flow proceeds to step S 962 .
- step S 959 determination is made whether or not calculation of an overlaying rate regarding each frame included in the image compositing target section serving as a selection target has been completed (step S 962 ).
- step S 962 the overlaying rate of a frame adjacent to the frame of which the overlaying rate has been calculated immediately before is calculated (step S 964 ). That is to say, the frame serving as an overlaying rate calculation target is pasted to the second work buffer, and an overlaying rate as to the composited image pasted to the first work buffer is calculated. Note that, in the case that the frame of which the overlaying rate has been calculated immediately before is the top frame or the last frame of the image compositing target section, the overlaying rate of the frame adjacent to the other direction of the front-most frame is calculated.
- step S 962 determination is made whether or not the frame selecting process has been completed regarding all of the image compositing target sections selected by the section selecting unit 160 (step S 963 ). In the case that the frame selecting process has not been completed regarding all the image compositing target sections (step S 963 ), the flow returns to step S 951 , where the frame selecting process is repeated (steps S 951 through S 964 ). On the other hand, in the case that the frame selecting process has been completed regarding all the image compositing target sections (step S 963 ), the operation of the frame selecting process is ended.
- FIG. 37 is a flowchart illustrating the procedure of a composited image display process by the image processing apparatus 100 according to the first embodiment of the present invention.
- the front-most frame is overlaid and composited on the top side.
- description will be made regarding an example wherein composited images relating to a moving picture file stored in the moving picture storage unit 200 are displayed sequentially.
- the file obtaining unit 180 obtains each piece of metadata stored in the compositing target image metadata storage unit 270 (step S 970 ). Subsequently, the file obtaining unit 180 decodes the moving picture file corresponding to each piece of metadata obtained, and obtains each frame making up the moving picture file (step S 971 ). Subsequently, the size for compositing the image corresponding to each frame is calculated using each of affine transformation parameters included in the metadata of one of the obtained image compositing target sections serving as a compositing target, and a work buffer having this size is secured in the image memory 230 (step S 972 ). Also, based on the size of the composited image thus calculated, the position and size of the front-most frame are calculated.
- step S 973 determination is made based on the front-most frame information included in the obtained metadata whether or not the front-most frame is the top frame at the image compositing target section.
- the front-most frame is the top frame at the image compositing target section (step S 973 )
- images are extracted in the ascending order of frame numbers (step S 974 ).
- the image transforming unit 190 uses the affine transformation parameters correlated with the extracted image to subject the image thereof to affine transformation (step S 975 ).
- the image compositing unit 220 adds a white frame to the periphery of the image subjected to affine transformation, and holds the image to which the white frame has been added in the image memory 230 (step S 976 ). That is to say, the image subjected to affine transformation to which the white frame has been added is composited as to a history image stored in the image memory 230 . Note that in the case of the first image, only the image thereof is held in the image memory 230 .
- the display control unit 240 controls the display unit 250 to display the composited image held at the image memory 230 (step S 977 ).
- step S 979 the flow returns to step S 972 , where the composited image display process is repeated (steps S 972 through S 978 and S 980 ).
- step S 979 the operation of the composited image display process is ended.
- step S 973 an image compositing process is executed (step S 980 ). This image compositing process will be described in detail with reference to FIG. 38 .
- FIG. 38 is a flowchart illustrating an image compositing procedure (the procedure in step S 980 shown in FIG. 37 ) included in the procedure of the composited image display process by the image processing apparatus 100 according to the first embodiment of the present invention.
- image compositing is executed toward the front-most frame from the top frame at the image compositing target section, image compositing is executed toward the front-most frame from the last frame, and finally the front-most frame is composited.
- images are extracted in the ascending order of frame numbers (step S 981 ).
- the image transforming unit 190 uses the affine transformation parameters correlated with the extracted image to subject the image thereof to affine transformation (step S 982 ).
- the image compositing unit 220 adds a white frame to the periphery of the image subjected to affine transformation, and holds the image to which the white frame has been added in the image memory 230 (step S 983 ). That is to say, the image subjected to affine transformation to which the white frame has been added is composited as to a history image stored in the image memory 230 .
- the display control unit 240 controls the display unit 250 to display the composited image held at the image memory 230 (step S 984 ).
- step S 985 the flow returns to step S 981 , where the image compositing process is repeated from the top frame to the frame immediately before the front-most frame (steps S 981 through S 984 ).
- image held at the image memory 230 is the frame immediately before the front-most frame (step S 985 )
- images are extracted sequentially in the descending order of frame numbers (step S 986 ).
- the image transforming unit 190 uses the affine transformation parameters correlated with the extracted image to subject the image thereof to affine transformation (step S 987 ).
- the image compositing unit 220 adds a white frame to the periphery of the image subjected to affine transformation, and holds this image to which the white frame has been added in the image memory 230 (step S 988 ).
- the image subjected to affine transformation to which the white frame has been added is composited as to the history image held at the image memory 230 .
- the display control unit 240 controls the display unit 250 to display the composited image held at the image memory 230 (step S 989 ).
- step S 990 the flow returns to step S 986 , where the image compositing process is repeated from the last frame to the frame immediately after the front-most frame (steps S 986 through S 989 ).
- the image holding at the image memory 230 is the frame immediately after the front-most frame (step S 990 ).
- the image of the front-most frame is extracted (step S 991 ).
- the image transforming unit 190 uses the affine transformation parameters correlated with the extracted image of the front-most frame to subject the image of the front-most frame to affine transformation (step S 992 ).
- the affine transformation parameters correlated with the front-most frame is the values of a unit matrix, and accordingly, the image of the front-most frame is not transformed actually.
- the image compositing unit 220 adds a white frame to the periphery of the image of the front-most frame subjected to affine transformation, and holds the image to which the white frame has been added in the image memory 230 (step S 993 ). That is to say, the image subjected to affine transformation to which the white frame has been added is composited as to the history image held at the image memory 230 .
- the display control unit 240 controls the display unit 250 to display the composited image held at the image memory 230 (step S 994 ).
- steps S 975 , S 982 , S 987 , and S 992 are an example of the image transforming procedure referred to in the Summary of the Invention.
- steps S 976 , S 983 , S 988 , and S 993 are an example of the image compositing procedure referred to in the Summary of the Invention.
- steps S 977 , S 984 , S 989 , and S 994 are an example of the display control procedure referred to in the Summary of the Invention.
- the functional configuration according to the second embodiment of the present invention is generally the same as that of the image processing apparatus 100 except that the camera work detecting unit 120 included in the image processing apparatus 100 shown in FIG. 1 differs. Therefore, the configuration other than the camera work detecting unit 120 will be denoted with the same reference numerals as those in the first embodiment of the present invention, and only the configuration having a different function will be described, and other description will be omitted.
- FIG. 39 is a block diagram illustrating a functional configuration example of a camera work detecting unit 125 according to the second embodiment of the present invention.
- the camera work detecting unit 125 includes a feature point extracting unit 121 , an optical flow calculating unit 122 , a camera work parameter calculating unit 123 , and a reliability determination score calculating unit 126 .
- the feature point extracting unit 121 , optical flow calculating unit 122 , and camera work parameter calculating unit 123 are the same as those in the camera work detecting unit 120 shown in FIG. 2 . Therefore, description will be made below with the reliability determination score calculating unit 126 as the center. Also, with this example, description will be made regarding an example wherein a reliability determination score is calculated using the luminance difference value between pixels included in an overlaid region of consecutive two frames.
- the reliability determination score calculating unit 126 uses the camera work parameters output from the camera work parameter calculating unit 123 to calculate a reliability determination score relating to consecutive two frames making up the moving picture output from the moving picture input unit 110 , and outputs the calculated reliability determination score to the recording control unit 150 . Note that reliability determination score calculation will be described in detail with reference to FIG. 40 .
- FIG. 40 is a diagram illustrating an overview of a calculating method at the time of calculating a reliability determination score by the reliability determination score calculating unit 126 according to the second embodiment of the present invention.
- FIG. 40 illustrates a previous frame 590 and a subsequent frame on the temporal axis as consecutive two frames making up the moving picture.
- (a) in FIG. 40 illustrates a case where the frame 591 is subjected to affine transformation with the frame 590 as a reference, and the frame 591 after affine transformation is overlaid as to the frame 590 .
- the range of an overlaid region 592 between the frames 590 and 591 is indicated with a heavy line.
- (b) illustrates the frame 590 shown in (a) in FIG.
- FIG. 40 illustrates the frame 591 shown in (a) in FIG. 40 .
- a square included in the frames 590 and 591 indicates a pixel making up each frame.
- (b) and (c) in FIG. 40 illustrate only the pixels included in the overlaid region 592 using a square.
- the reliability determination score calculating unit 126 calculates a reliability determination score, first, with regard to the consecutive two frames making up the moving picture output from the moving picture input unit 110 , the subsequent frame is subjected to affine transformation with the previous frame on the temporal axis as a reference frame. This affine transformation is executed using the camera work parameters output from the camera work parameter calculating unit 123 (affine transformation parameters relating to the previous frame (reference frame) and the subsequent frame). Subsequently, the reliability determination score calculating unit 126 calculates the absolute value of the luminance difference value between pixels included in the overlaid region after affine transformation.
- the reliability determination score calculating unit 126 calculates the total value of the absolute value of the difference value calculated for each pixel, and calculates as a reliability determination score a value obtained by dividing the calculated total value by the number of pixels serving as calculation targets.
- R represents the overlaid region between the frames 590 and 591 . That is to say, such as arrows 593 through 595 shown in (b) and (c) in FIG. 40 , of pixels included in the overlaid region between the frames 590 and 591 , the absolute value of the luminance difference value of two pixels disposed in the corresponding positions is calculated sequentially. Subsequently, the reliability determination score SHS 1 is calculated using the absolute values of these difference values.
- the reliability determination score calculated by the reliability determination score calculating unit 126 is output to the recording control unit 150 , and is recorded in the metadata storage unit 210 .
- the reliability determination score is recorded in the reliability determination score 217 of the metadata file 211 shown in FIG. 3B .
- the recorded content of the reliability determination score 217 has a different numeric value from the case shown in FIG. 3B .
- the section selecting unit 160 uses this reliability determination score to execute selection of an image compositing target section based on whether or not the reliability determination score exceeds a threshold. For example, in the case that the reliability determination score is below a certain threshold, the reliability can be determined to be high.
- a reliability determination score is calculated using all the pixels included in the overlaid region of the consecutive two frames.
- a reliability determination score may be calculated using information relating to a feature point serving as an affine transformation parameter calculation target by the camera work parameter calculating unit 123 .
- FIG. 41 is a block diagram illustrating a functional configuration example of the camera work detecting unit 127 according to the second embodiment of the present invention.
- the camera work detecting unit 127 includes a feature point extracting unit 121 , an optical flow calculating unit 122 , a camera work parameter calculating unit 123 , and a reliability determination score calculating unit 128 .
- the feature point extracting unit 121 , optical flow calculating unit 122 , and camera work parameter calculating unit 123 are the same as those of the camera work detecting unit 120 shown in FIG. 2 . Therefore, description will be made below with the reliability determination score calculating unit 128 as the center. Also, with this example, description will be made regarding an example wherein, of pixels included in the overlaid region of the consecutive two frames, a reliability determination score is calculated using the luminance difference value of pixels disposed around a feature point that exhibits dominant motion.
- the reliability determination score calculating unit 128 calculates a reliability determination score relating to the consecutive two frames making up the moving picture output from the moving picture input unit 110 , and outputs the calculated reliability determination score to the recording control unit 150 .
- the camera work parameters calculated by the camera work parameter calculating unit 123 are used. Note that the reliability determination score calculation will be described in detail with reference to FIG. 42 .
- FIG. 42 is a diagram illustrating an overview of a calculating method at the time of calculating a reliability determination score by the reliability determination score calculating unit 128 according to the second embodiment of the present invention.
- the relationship between the frames 590 and 591 shown in FIG. 42 , and the overlaid region 592 between the frames 590 and 591 are the same as those in FIG. 40 , and accordingly, description thereof will be omitted here.
- (b) and (c) in FIG. 42 illustrate the range of the overlaid region 592 using a dashed line.
- each pixel included in the overlaid region 592 shown in (a) through (c) in FIG.
- a feature point that exhibits dominant motion is represented with a single circle within a square, and the other feature points are represented with a double circle within a square.
- the overlaid region 592 shown in (a) through (c) in FIG. 40 includes four feature points that exhibit dominant motion, and the other single feature point.
- description will be made regarding an example wherein, of pixels included in the overlaid region 592 , pixels (e.g., eight neighborhood pixels) around four feature points that exhibit dominant motion are used to calculate a luminance difference value.
- pixels around the other single feature point are not used.
- each pixel included in regions 601 through 608 is used as a reliability determination score calculation target.
- the reliability determination score calculating unit 128 calculates a reliability determination score, first, of the consecutive two frames making up the moving picture output from the moving picture input unit 110 , the previous frame is taken as a reference frame on the temporal axis, and the subsequent frame is subjected to affine transformation. This affine transformation is the same as the example shown in FIG. 40 . Subsequently, the reliability determination score calculating unit 128 calculates, of the pixels included in the consecutive two frames, the absolute value of the luminance difference value of pixels around a feature point that exhibits dominant motion included in the overlaid region after affine transformation.
- the reliability determination score calculating unit 128 calculates the total value of the absolute value of the difference value calculated for each pixel around a feature point that exhibits dominant motion, and calculates as a reliability determination score a value obtained by dividing the calculated total value by the number of pixels serving as calculation targets.
- x-y coordinates wherein the left upper corner of the frame 590 is taken as the origin, the side of the upper side of the frame 590 is taken as the x axis, and the side of the left side of the frame 590 is taken as the y axis.
- the luminance value at the coordinates (i, j) of the overlaid region of the frame 590 is I11 (i, j)
- the luminance value at the coordinates (i, j) of the overlaid region of the frame 591 is I12(i, j).
- P1num the number of pixels used for calculation of a difference value
- R 1 represents the regions (e.g., eight neighborhood regions) around a feature point that exhibits dominant motion, of the overlaid region between the frames 590 and 591 .
- R 1 is taken as a region in which such pixels are not included. That is to say, the absolute value of the luminance difference value of two pixels disposed in the corresponding positions is calculated regarding the regions 601 through 608 shown in (b) and (c) in FIG. 42 . Subsequently, the absolute values of these difference values are used to calculate the reliability determination score SHS 2 .
- the reliability determination score calculated by the reliability determination score calculating unit 128 is output to the recording control unit 150 , and is recorded in the metadata storage unit 210 .
- the reliability determination score is recorded in the reliability determination score 217 of the metadata file 211 shown in FIG. 3B .
- the recorded content of the reliability determination score 217 has a different numeric value from the case shown in FIG. 3B .
- the section selecting unit 160 uses this reliability determination score to execute selection of an image compositing target section based on whether or not the reliability determination score exceeds a threshold. For example, in the case that the reliability determination score is below a certain threshold, the reliability can be determined to be high.
- a luminance difference value is used to calculate a reliability determination score.
- a difference value is calculated regarding each of G (Green), R (Red), and B (Blue), and based on this difference value a reliability determination score is calculated.
- an arrangement may be made wherein a Gaussian filter is applied to the images corresponding to consecutive two frames to smooth the images, the smoothed images are reduced, and the pixels of the reduced images are used to calculate a reliability determination score in the same way as above.
- camera sensor information may be used to calculate a reliability determination score.
- a moving picture is recorded by a camera including a triaxial acceleration sensor and a triaxial gyro sensor
- the sensor information output from these sensors is recorded in a manner correlated with the moving picture as camera motion information.
- the moving picture correlated with the sensor information is input to the moving picture input unit 110 , and the difference value between the motion information calculated regarding each frame making up this moving picture, and the motion information specified with the sensor information is calculated. Subsequently, this difference value may be used as a reliability determination score.
- reliability determination score in the case that the reliability determination score does not exceed a threshold, reliability can be determined to be high. That is to say, a method may be employed wherein, in the case that the degree of matching between image motion estimated using the sensor information output from the camera, and image motion calculated from an image making up a moving picture is great, reliability is determined to be high.
- FIG. 43 is a block diagram illustrating a functional configuration example of an image processing apparatus 101 according to the third embodiment of the present invention.
- the image processing apparatus 101 includes a compositing target image selecting unit 171 , a face attribute state determining unit 280 , a reference information storage unit 281 , and a metadata storage unit 290 .
- the other functional configuration is generally the same as that in the image processing apparatus 100 shown in FIG. 1 , and accordingly, the same reference numerals as the reference numerals of the image processing apparatus 100 are denoted thereto, description will be made with components having a different function as the center, and other description will be omitted.
- the compositing target image selecting unit 171 uses metadata stored in the metadata recording unit 290 to select a frame used to create a composited image from frames included in the image compositing target section output from the section selecting unit 160 , and outputs compositing target image information relating to the selected frame (frame selection information) to the recording control unit 150 . Note that the compositing target image selection will be described in detail with reference to FIGS. 45 through 48 .
- the face attribute state determining unit 280 uses the reference information stored in the reference information storage unit 281 to determine the attributes and state of the face detected by the face detecting unit 130 , and outputs determination results to the recording control unit 150 . Specifically, the face attribute state determining unit 280 calculates a correlation value that indicates correlativity between a face image extracted from an image making up the moving picture output from the moving picture input unit 110 and normalized, and various types of data stored in the reference information storage unit 281 , determines the height of correlativity based on the calculated correlation value, and extracts various types of data relating to the normalized face image. Such an extracting process is executed repeatedly, and based on extracted various types of data, the facial expression and so forth of the normalized face image are determined.
- a technique may be employed to execute template matching with a database in which a great number of face data is integrated (e.g., see Japanese Unexamined Patent Application Publication No. 2003-271933, Japanese Unexamined Patent Application Publication No. 2004-30629, etc.).
- a smiling face database is established from a great number of smiling face sample images beforehand, this and a face image extracted partially from the image are subjected to a template matching process, whereby determination can be made whether the face is a smiling face or a non-smiling face.
- determination can also be made regarding the other attributes or states. Note that determination may be made using the evaluated value shown in the first embodiment of the present invention.
- the face attribute state determining unit 280 is an example of the determining unit referred to in the Summary of the Invention.
- the reference information storage unit 281 stores reference information such as various portions of a face used for gender determination, and supplies the stored reference information to the face attribute state determining unit 280 . Note that the reference information stored in the reference information storage unit 281 will be described in detail with reference to FIGS. 44A and 44B .
- the metadata storage unit 290 stores various types of information output from the camera work detecting unit 120 , face detecting unit 130 , and face attribute state determining unit 280 in a manner correlated with a moving picture and a frame as a metadata file based on the control of the recording control unit 150 . Also, the metadata storage unit 290 supplies a metadata file to the section selecting unit 160 and the compositing target image selecting unit 171 . Note that the metadata file stored in the metadata storage unit 290 will be described in detail with reference to FIGS. 44A and 44B .
- FIGS. 44A and 44B are diagrams schematically illustrating content stored in the reference information storage unit 281 and the metadata storage unit 290 , according to the third embodiment of the present invention.
- FIG. 44A illustrates reference information stored in the reference information storage unit 281 .
- This reference information is, for example, an image itself such as the eyes, mouth, or the like, i.e., a feature database serving as a person's face, a feature database relating to a specific person's face, or the like.
- a gender database 282 , an age database 283 , a smiling face database 284 , an eye closing database 285 , and a characters database 286 are stored in the reference information storage unit 281 .
- the gender database 282 is a database used for determining the gender of a face image
- the age database 283 is a database used for determining the age of a face image.
- the smiling face database 284 is a database used for determining the face smiling face of a face image
- the eye closing database 285 is a database used for determining eye closing of a face image.
- the characters database 286 is a database used for determining a person determined by the user as characters. In general, according to feature data, a great number of various types of databases can be held with relatively little storage capacity as compared to an image itself.
- FIG. 44B schematically illustrates a metadata file 291 stored in the metadata storage unit 290 correlated with the moving picture file 201 stored in the moving picture storage unit 200 .
- the correspondence relation between the moving picture file 201 and the metadata file 291 is the same as the example shown in FIG. 3B , and accordingly, description thereof will be omitted here.
- a moving picture ID 292 , a frame number 293 , an affine transformation parameter 294 , a reliability determination score 295 , face detection information 296 , gender 297 , age 298 , smiling face 299 , eye closing 287 , and characters 288 are stored in the metadata file 291 in a correlated manner.
- the moving picture ID 292 , frame number 293 , affine transformation parameter 294 , reliability determination score 295 , and face detection information 296 are the same as those shown in FIG. 3B , and accordingly, description thereof will be omitted here.
- the gender 297 is gender determined regarding face included in each frame of the image corresponding to the frame number 293 , where the determination result output from the face attribute state determining unit 280 is stored. For example, in the case that no face is included in the corresponding frame, nothing is stored. For example, in the case that the face included in the corresponding frame has been determined to be a male, “ 1 ” is stored in the gender 297 . On the other hand, in the case that the face included in the corresponding frame has been determined to be a female, “ 0 ” is stored in the gender 297 .
- the age 298 is an age determined regarding the face included in each frame of the moving picture corresponding to the frame number 293 , where the determination result output from the face attribute state determining unit 280 is stored. For example, in the case that no face is included in the corresponding frame, nothing is stored. For example, in the case that the face included in the corresponding frame has been determined to be an adult, “ 1 ” is stored in the age 298 . On the other hand, in the case that the face included in the corresponding frame has been determined to be a child, “ 0 ” is stored in the age 298 .
- the smiling face 299 is whether there is a smiling face determined regarding a face included in each frame of the moving picture corresponding to the frame number 293 , where the determination result output from the face attribute state determining unit 280 is stored. For example, in the case that no face is included in the corresponding frame, nothing is stored. For example, in the case that the face included in the corresponding frame has been determined to be a smiling face, “ 1 ” is stored in the smiling face 299 . On the other hand, in the case that the face included in the corresponding frame has been determined to be a non-smiling face, “ 0 ” is stored in the smiling face 299 .
- the eye closing 287 is whether there is an eye closed face determined regarding a face included in each frame of the moving picture corresponding to the frame number 293 , where the determination result output from the face attribute state determining unit 280 is stored. For example, in the case that no face is included in the corresponding frame, nothing is stored. For example, in the case that the face included in the corresponding frame has been determined to be an eye closed face, “ 1 ” is stored in the eye closing 287 . On the other hand, in the case that the face included in the corresponding frame has been determined to be a non-eye-closed face, “ 0 ” is stored in the eye closing 287 .
- the characters 288 is whether there is a character determined regarding a face included in each frame of the moving picture corresponding to the frame number 293 , where the determination result output from the face attribute state determining unit 280 is stored. For example, in the case that no face is included in the corresponding frame, nothing is stored. For example, in the case that the face included in the corresponding frame has been determined to be Mr. A, “A” is stored in the characters 288 . On the other hand, in the case that the face included in the corresponding frame has been determined to be Mr. B, “B” is stored in the characters 288 .
- FIG. 45 is a diagram illustrating an example of a selection screen displayed on the display unit 250 according to the third embodiment of the present invention.
- a selection screen 610 is a selection screen wherein the user performs a selection operation to display a desired composited image.
- the selection screen 610 includes, for example, check fields used for selecting a gender 611 , age 612 , characters 613 , and facial expression 614 as facial attributes.
- the selection screen 610 includes a compositing sequence 615 used for selecting a compositing sequence, and a pull-down button 617 used for selecting an overlaying rate 616 .
- the selection screen 610 includes a “panorama-style composited image display” button 619 , and a “cancel” button 620 .
- a cursor 618 is used to add a check to the female field of the gender 611 , to add a check to the adult field of the age 612 , to add a check to the characters field of the characters 613 , and to add a check to the smiling face field of the facial expression 614 .
- the names of the characters registered on the characters database 286 e.g., Mr. A, Mr. B, and Mr. C
- a check is added to the field of a desired name from these characters. For example, a check is added to Mr. B.
- the cursor 618 is used to add a check to the ascending field of the compositing sequence 615 . Also, the cursor 618 is used to press the pull-down button 617 , whereby an overlaying rate list of “high”, “middle”, and “low” are displayed as shown in FIG. 45 .
- a desired overlaying rate can be selected from this overlaying rate list.
- an overlaying rate of 60% through 70% is selected, and in the case that “middle” has been selected, an overlaying rate of 60% through 40% is selected, and in the case that “low” has been selected, an overlaying rate of 40% through 35% is selected.
- the user selects a desired overlaying rate from the overlaying rate list, but the user may input a desired overlaying rate directly. Note that an item that the user does not desire specifically may not be selected.
- the user uses the cursor 618 to press the “panorama-style composited image display” button 619 .
- the selection information of each item selected at the time of pressing the “panorama-style composited image display” button 619 is output from the operation accepting unit 260 to the compositing target image selecting unit 171 .
- the compositing target image selecting unit 171 executes a frame selecting process. Note that, upon the “cancel” button 620 being pressed, the selection state at each item is released.
- FIG. 46 is a diagram illustrating an example of scores used for calculation of a selection determination score by the compositing target image selecting unit 171 according to the third embodiment of the present invention.
- the selection determination score is a value used for selecting a frame, and is calculated based on each piece of information relating to a face included in a frame. Also, in the case that multiple faces are included in one frame, scores relating to these faces are calculated for each face, and the value to which the score for each face is added is calculated as the selection determination score. Also, the selection determination score is calculated, for example, based on the item selected at the selection screen 610 shown in FIG. 45 , the score shown in FIG. 46 , and the attributes and state relating a face. As the attributes and state relating a face the metadata stored in the metadata storage unit 290 is used.
- an attribute value Z 1 that indicates the determined face attribute, a status value J 1 that indicates the determined face state are calculated, and the attribute value Z 1 and the status value J 1 are added, whereby the selection determination score is calculated.
- the attribute value Z 1 is, for example, a value determined with a gender 621 , age 622 , and whether or not there is a specific person 623 , as face attributes.
- the status value J 1 is, for example, a value determined with a smiling face 624 , eye closing 625 , a face position within the image (distance 626 from the image center), a face size 627 within the image, as face states.
- the score shown in FIG. 46 relating to the matched item is added to the attribute value Z 1 .
- the score shown in FIG. 46 is added to the status value J 1 .
- each score corresponding to the metadata stored in the metadata storage unit 290 is added to the status value J 1 . Subsequently, the calculated attribute value Z 1 and status value J 1 are added, whereby the selection determination score is calculated.
- the attribute value Z 1 and the status value J 1 are calculated for each face, and a value obtained by adding the attribute value Z 1 and the status value J 1 is calculated as the selection determination score of the image thereof.
- each value is added at the time of calculating each score, but for example, each score may be calculated by multiplying each value.
- an arrangement may be made wherein the weighting coefficient corresponding to a facial attribute and state is used instead of each score shown in FIG. 46 , whereby each score is calculated.
- a front-most image (front-most frame) is selected using the selection determination score thus calculated. That is to say, the selection determination score is calculated regarding each image included in the image compositing target section selected by the section selecting unit 160 , and of the calculated selection determination scores, an image having the highest calculated value is selected as a front-most image. Subsequently, the other frames other than the front-most frame are selected. This selection method will be described in detail with reference to FIGS. 47A through 48B .
- FIGS. 47A and 47B are diagrams illustrating an overview of a selecting method in the case that frame selection is executed by the compositing target image selecting unit 171 according to the third embodiment of the present invention.
- FIG. 47A illustrates an example in the case of calculating the selection determination score of a frame to be overlaid with the lower side of the already selected frame. For example, let us say that a frame 640 is a frame selected as the front-most frame. In the case that another frame is selected with the frame 640 as a reference frame, the scores relating to a face included in the region to be overwritten with the frame 640 are excluded.
- the frame 641 For example, let us say that in the case that faces 643 through 646 are included in a frame 641 , at least a part of the image 643 and 644 is included in an overlaid region 642 as to the front-most frame 640 .
- determination is made whether or not the overlaid region 642 has an overlaying rate within a predetermined range.
- the setting rang of this overlaying rate is the content selected at the selection screen 610 shown in FIG. 45 .
- the frame 641 is not selected.
- the selection determination score is calculated regarding the frame 641 .
- the selection determination score of the frame 641 is calculated using a value relating to the faces 645 and 646 not included in the overlaid region 642 of the faces 643 through 646 included in the frame 641 .
- FIG. 47B illustrates an example in the case of calculating the selection determination score of a frame to be overlaid with the lower side of the already selected frame.
- a frame 660 is a frame serving as an overwrite compositing target as to a frame 650 .
- the score relating to a face included in a region other than the region to be overwritten as to the frame 650 is added. For example, let us say that faces 651 through 653 are included in the frame 650 , and at least a part of the face 653 is included in an overlaid region 655 as to the frame 660 .
- faces 661 through 664 are included in the frame 660 .
- determination is made whether or not the overlaid region 655 has an overlaying rate within a predetermined range.
- the frame 660 is not selected.
- the selection determination score is calculated regarding the frame 660 .
- the selection determination score of the frame 660 is calculated using values relating to the faces 651 , 652 , and 661 through 664 not included in the overlaid region 655 of the frame 650 , of the faces 651 through 653 , and 663 and 664 included in the frames 650 and 660 . That is to say, in the case that two images are composited, a value relating to a face included in a region serving as a display target is calculated as the selection determination score.
- FIGS. 48A and 48B are diagrams illustrating an overview of the selecting method in the case that frame selection is executed by the compositing target image selecting unit 171 according to the third embodiment of the present invention.
- FIGS. 48A and 48B description will be made regarding a case where the overlaying sequence is specified by the user's operation beforehand.
- a selecting method in the case of overwrite-compositing toward a new image from an old image at imaging point-in-time in the case of overwrite-compositing in the ascending order
- the image selected as the front-most frame is overwrite-composited on the top side.
- FIG. 48A illustrates an example wherein an image is selected toward a search direction 671 with a front-most frame 670 as a reference frame.
- search is executed in a direction that traces back the time series from a front-most image 676 , and accordingly, the already selected images are overlaid above the image serving as a selection target. Therefore, the score is calculated while taking an overlaid region into consideration, and of images included in a range of the specified overlaying rate, an image having the highest score is selected. For example, let us say that images 675 and 674 have been selected with an image 676 as a reference image corresponding to the reference frame 670 .
- the selection determination score calculated regarding a face included in the image 673 is calculated.
- a face 678 of four faces included in the image 673 a portion thereof is included in the overlaid region 677 . Therefore, in the case that the selection determination score regarding a face included in the image 673 is calculated, the selection determination score is calculated using three faces other than the face 678 included in the image 673 .
- FIG. 48B illustrates an example wherein an image is selected toward a search direction 682 with a front-most frame 680 as a reference frame.
- search is executed in the time-series direction from a front-most image 683 , and accordingly, an image serving as a selection target is overlaid above already selected images.
- the front-most image 683 alone is overlaid above all the images. Therefore, with regard to the image 684 , in the same way as with the case shown in FIG. 48A , the score is calculated while taking an overlaid region as to the image 683 into consideration, and of images included in a range of the specified overlaying rate, an image having the highest score is selected.
- an image is selected wherein a total score between an image serving as a selection target and an image overlaid with the above image is not reduced as much as possible within a range of the specified overlaying rate, and a score on the current image is high. That is to say, an image is selected wherein a total score on the already selected composited image, and a score on the current image are the highest.
- the images 684 and 685 have been selected with the image 683 corresponding to the reference frame 680 as a reference image.
- determination of the overlaying rate of an overlaid region 687 between the images 685 and 686 is executed.
- the selection determination score calculated regarding each face included in a composited image between the images 683 through 685 , and the image 686 is calculated.
- three faces included in the image 686 become calculation targets regardless whether or not the three faces are included in the overlaid region 687 .
- a selecting process is executed in the direction opposite of the case of ascending order.
- search is executed in a direction that traces back the time series from the front-most image, and in a direction toward the time series from the front-most image.
- the selecting process in the same way as with FIG. 48A is executed.
- Such compositing order is used as compositing order at the time of image compositing, and accordingly, such compositing order is stored in correlated with the image compositing target section of the compositing target image metadata storage unit 270 .
- a unique attribute relating to a face included in an image is taken as a facial attribute
- an attribute relating to a facial state is taken as a state
- a selection determination score is calculated using each piece of information relating to such a face.
- a selection determination score may be calculated using each piece of information relating to another object.
- a selection determination score may be calculated using each piece of information relating to various types of object, such as a pet such as a cat, or a dog, an animal, a house, a vehicle, or the like.
- a selection determination score may be calculated using a saliency map.
- This saliency map represents a region that a people will observe by a score for each pixel.
- a value obtained by integrating the score for each pixel of a region other than an overlaid region may be calculated as the selection determination score of the image.
- a value obtained by adding the score calculated using the saliency map, and a score relating to an object may be calculated as a selection determination score.
- FIG. 49 is a flowchart illustrating the procedure of a frame selecting process by the image processing apparatus 101 according to the third embodiment of the present invention. This procedure is a modification of the procedure in step S 950 shown in FIG. 34 . Also, this example illustrates an example wherein the affine transformation parameters alone stored in the metadata storage unit 210 are used to calculate an overlaying rate. Also, this example illustrates a case where the compositing order has not specified.
- the metadata corresponding to each frame included in one image compositing target section selected by the section selecting unit 160 is obtained from the metadata file stored in the metadata storage unit 210 (step S 1001 ). Subsequently, the selection information of each item accepted by the operation accepting unit 260 is obtained (step S 1002 ).
- a selection determination score is calculated regarding each frame included in the image compositing target section serving as a selection target, and of the calculated selection determination scores, a frame having the highest value is selected as a front-most frame (step S 1003 ). Also, the selected front-most frame is selected as a representative frame. Subsequently, each piece of metadata relating to the selected front-most frame is recorded in the compositing target image metadata storage unit 270 (step S 1004 ).
- step S 1005 search is executed in the descending order from the selected representative frame, frames within a range of the specified overlaying rate are extracted, and a selection determination score is calculated regarding these frames (step S 1005 ).
- a frame having the highest value is selected as a new representative frame (step S 1006 ).
- Each piece of metadata relating to the new selected representative frame is recorded in the compositing target image metadata storage unit 270 (step S 1007 ). Determination is made whether or not calculation of a selection determination score has been executed up to the top frame of the image compositing target section serving as a selection target (step S 1008 ).
- step S 1008 the flow returns to step S 1005 , where a representative frame selecting process is repeated (steps S 1005 through S 1007 ).
- step S 1008 the already selected front-most frame is selected as a representative frame (step S 1009 ).
- search is executed in the ascending order from the selected representative frame, frames within a range of the specified overlaying rate are extracted, and a selection determination score is calculated regarding these frames (step S 1010 ).
- a frame having the highest value is selected as a new representative (step S 1011 ).
- step S 1012 Each piece of metadata relating to the new selected representative frame is recorded in the compositing target image metadata storage unit 270 (step S 1012 ). Determination is made whether or not calculation of a selection determination score has been executed up to the last frame of the image compositing target section serving as a selection target (step S 1013 ). In the case that calculation of a selection determination score has not been executed up to the last frame of the image compositing target section serving as a selection target (step S 1013 ), the flow returns to step S 1010 , where the representative frame selecting process is repeated (steps S 1010 through S 1012 ).
- step S 1013 determination is made whether or not the frame selecting process has been completed regarding all the image compositing target sections selected by the section selecting unit 160 (step S 1014 ).
- step S 1014 determination is made whether or not the frame selecting process has been completed regarding all the image compositing target sections selected by the section selecting unit 160 (step S 1014 ).
- step S 1014 the flow returns to step S 1001 , where the frame selecting process is repeated (steps S 1001 through S 1013 ).
- step S 1014 the operation of the frame selecting process is ended.
- the attribute and state and the like relating a face included in an image are used to select the front-most image and the compositing target image, whereby an interesting image relating to a person included in a moving picture can be provided.
- each item such as the attribute and state and the like relating to a face can be set and displayed according to the user's preference, whereby a composited image according to the user's preference can be provided.
- FIG. 50 is a block diagram illustrating a functional configuration example of an image processing apparatus 102 according to the fourth embodiment of the present invention.
- the image processing apparatus 102 includes an image compositing unit 221 , a compositing target image metadata storage unit 277 , and a decoration image determining unit 700 .
- the other functional configuration is generally the same as that in the image processing apparatus 100 shown in FIG. 1 , and accordingly, the same reference numerals as the reference numerals of the image processing apparatus 100 are denoted thereto, and description will be made with components having a different function as the center, and the other description will be omitted.
- the recording control unit 150 records information relating to the decoration image output from the decoration image determining unit 700 in the compositing target image metadata storage unit 277 as compositing target image metadata.
- the compositing target image selecting unit 170 selects a composited image from frames included in an image compositing target section, and outputs compositing target image information relating to the selected compositing target image to the recording control unit 150 and the decoration image determining unit 700 .
- the decoration image determining unit 700 determines a decoration image to be added to each frame corresponding to the compositing target image information output from the compositing target image selecting unit 170 , and outputs the decoration image information relating to the determined decoration image to the recording control unit 150 .
- the decoration image determining unit 700 determines a decoration image based on the metadata stored in the metadata storage unit 210 .
- the decoration image determining unit 700 determines a decoration image based on the metadata stored in the metadata storage unit 210 , and each frame corresponding to the compositing target image information. Note that a decoration image determining method will be described in detail with reference to FIGS. 52A through 62 .
- the compositing target image metadata storage unit 277 stores the compositing target image information output from the compositing target image selecting unit 170 as a compositing target image metadata file based on the control of the signal 150 . Also, the compositing target image metadata storage unit 277 stores the decoration image information output from the decoration image determining unit 700 as a compositing target image metadata file. Also, the compositing target image metadata storage unit 277 supplies the compositing target image metadata file to the file obtaining unit 180 in response to a request from the file obtaining unit 180 . Note that the metadata file stored in the compositing target image metadata storage unit 277 will be described in detail with reference to FIG. 51 .
- the file obtaining unit 180 outputs the compositing target image metadata obtained from the compositing target image metadata storage unit 277 to the image compositing unit 221 .
- the image compositing unit 221 uses the image memory 230 to add a decoration image according to the content of compositing target image metadata to the images subjected to affine transformation by the image transforming unit 190 , composites the images to which the decoration image has been added to create a composited image.
- FIG. 51 is a diagram schematically illustrating the storage content of the compositing target image metadata storage unit 277 according to the fourth embodiment of the present invention.
- a moving picture ID 271 , a section number 272 , a frame number 273 , affine transformation parameters 274 , a front-most frame 275 , face data 276 , and decoration image data 278 are stored in the compositing target image metadata storage unit 277 in a correlated manner.
- the moving picture ID 271 , section number 272 , frame number 273 , affine transformation parameters 274 , front-most frame 275 , and face data 276 are the same as those of an example shown in FIG. 4 , description thereof will be omitted here.
- the decoration image data 278 is information relating to a decoration image to be added to the compositing target image determined by the decoration image determining unit 700 .
- this information for example, the position, size, pattern, color, type, and the like of a decoration image to be added to a compositing target image are stored.
- a decoration image is added to the periphery of the corresponding frame. Note that the example shown in FIG. 51 illustrates information relating to a decoration image in an omitted manner.
- FIGS. 52A through 52C are diagrams illustrating an overview of a determining method arranged to determine the heaviness of a white frame by the decoration image determining unit 700 according to the fourth embodiment of the present invention.
- FIG. 52A illustrates consecutive three frames 711 through 713 in time series, of the frames selected by the compositing target image selecting unit 170 .
- a composited image 714 indicates an image wherein the consecutive three frames 711 through 713 in time series are subjected to affine transformation, and are then composited.
- the frame 713 is a frame adjacent to the frame 712 in time series, and indicates a state after affine transformation.
- the decoration image determining unit 700 obtains each of these frames from the moving picture file stored in the moving picture storage unit 200 based on information relating to the frames selected by the compositing target image selecting unit 170 . Also, the decoration image determining unit 700 includes a work buffer used for holding a composited image.
- a frame 714 that is the next frame after affine transformation is overlaid above the composited image 714 wherein up to the immediately previous frame of the selected frames are composited.
- a feature point is extracted from the composited image 714 , and an optical flow is calculated regarding this feature point. Subsequently, camera work parameters and reliability are calculated.
- the calculation results differ from the camera work parameters and reliability calculated regarding the whole image in some cases.
- the frames selected by the compositing target image selecting unit 170 are not consecutive frames making up a moving picture, and accordingly, there is a possibility that inter-frame shift may be caused due to multiplication of multiple affine transformation parameters.
- a region 716 is an overlaid region between the frames 711 and 713
- the region other than the region 716 is an overlaid region between the frames 712 and 713 .
- 52A schematically illustrates the feature points extracted from the overlaid region 715 using black circles and white circles. Also, optical flows calculated regarding these feature points are represented with arrows with feature points as the origins.
- the feature points indicated with black circles of the overlaid region 715 are feature points that exhibit dominant motion, and the white circles are the other feature points.
- FIG. 52B illustrates a case where the affine transformation parameters calculated regarding the overlaid region 715 shown in FIG. 52A are used to subject the overlaid region 715 to affine transformation.
- FIG. 52B uses four arrows to illustrate the transition of movement with the affine transformation parameters, and illustrates the region after the overlaid region 715 is subjected to affine transformation as an overlaid region 717 after affine transformation.
- FIG. 52C illustrates only the overlaid region 715 , and the overlaid region 717 after affine transformation.
- the heaviness of a white frame is determined based on the size of shift.
- Over-extension rate DR 1 (1 ⁇ ( C/A ))
- Over-extension rate DR 2 (1 ⁇ ( C/B ))
- C represents the area of a region wherein an overlaid region of target images, and a region after affine transformation of this overlaid region are composited.
- A represents the area of the overlaid region of target images
- B represents the area of the overlaid region of target images after affine transformation.
- the heaviness of a white frame serving as a decoration image is calculated using the following expressions.
- the heaviness of a white frame SH 1 to be added to the sides of both ends in the horizontal direction W ⁇ DR
- the heaviness of a white frame SH 2 to be added to the sides of both ends in the vertical direction H ⁇ DR
- An upper limit and a lower limit may be provided to a over-extension rate so as to prevent a white frame from becoming greatly heavy as to an image serving as a providing target.
- a range of an upper limit and a lower limit of a over-extension rate for example, 0.1 through 0.03 may be set.
- the heavinesses SH 1 and SH 2 to be added to the sides of both ends in the horizontal direction and in the vertical direction thus calculated are output to the recording control unit 150 , and are recorded in the decoration image data 278 of the compositing target image metadata storage unit 277 .
- the over-extension rate is small, shift between a target image serving as an object to which a white frame is added, a composited image to be overlaid below that image is small. Therefore, even in the case that the heaviness of a white frame to be added to the target image thereof is thinned, it can be conceived to form a natural image between the target image and the composited image.
- the over-extension rate is great, shift between the target image, and a composited image to be overlaid below that image is great. Therefore, the heaviness of a white frame to be added to the target image thereof is increased, whereby the shift thereof can be obscured.
- a white frame to be added to the image unnaturalness due to shift can be eliminated.
- the position to which a white frame is added may be one of the outer side, inner side, and above the edge of the target image. Examples to add such a white frame are shown in FIGS. 56A through 58D .
- FIGS. 53A through 53C are diagrams illustrating an example of a decoration image determined by the decoration image determining unit 700 according to the fourth embodiment of the present invention.
- FIG. 53A illustrates an example wherein a white frame image 721 is added to the outer side of the periphery of a frame 720
- FIG. 53B illustrates an example wherein a white frame image 722 is added so as to include the periphery of the frame 720
- FIG. 53C illustrates an example wherein a white frame image 723 is added to the inner side of the periphery of the frame 720 .
- FIGS. 53A through 53C enlarge white frames as to images to facilitate visualization. Also, FIGS. 53A through 53C illustrate the periphery of an image using a heavy line so as to facilitate visualization for the size of the frame 720 .
- the heaviness of a white frame thus calculated may be determined for each image serving as a calculation target.
- a over-extension rate having the highest value may be used to unify the white frame ratio of each frame included in the section thereof.
- the average value of over-extension rates may be employed.
- the position to which a white frame is added may be changed according to relative positional relationship between images.
- FIGS. 54A through 54C are diagrams illustrating an example of a determining method for a decoration image by the decoration image determining unit 700 according to the fourth embodiment of the present invention.
- FIG. 54A illustrates a case where a frame 740 is subjected to affine transformation with a frame 730 as a reference frame.
- description will be made regarding an example wherein the position of an image to be added to the periphery on the upper side of the frame 740 is determined.
- movement vectors in the case that the frame 740 is subjected to affine transformation, and points at the two corners on the upper side of the frame 370 are moved, with the frame 730 as a reference frame, are illustrated with movement vectors 731 and 732 .
- normal vectors on the upper side of the frame 730 are illustrated with normal vectors 733 and 734 .
- FIG. 54B illustrates relationship between the movement vectors 731 and 732 , and the normal vectors 733 and 734 .
- an angle 736 formed of a vector 735 obtained by adding the movement vectors 731 and 732 , and the vector 733 is calculated.
- an angle 737 formed of the vectors 735 and 734 is calculated.
- the value of the angle 736 , and the value of the angle 737 are compared. As a result of this comparison, in the case that the value of the angle 736 is smaller, a white frame is disposed in the inner side of the periphery of the frame 740 .
- a white frame is disposed in the outer side of the periphery of the frame 740 .
- the value of the angle 737 is smaller, and accordingly, a white frame to be added to the periphery on the upper side of the frame 740 is determined to be disposed on the outer side thereof.
- the position of a white frame can be determined in the same way.
- FIG. 54C illustrates a white frame image 741 serving as an example of a white frame to be added to the frame 740 .
- the position of the white frame image can be determined according to the relative position as to the reference frame.
- a white frame image may be determined to be disposed in either the outer side or the inner side.
- a white frame image may be determined not only to be disposed in either the outer side or the inner side of the periphery, but also to be disposed so as to straddle the periphery thereof.
- the center position of a white frame image may be changed according to an angle formed of an addition vector and a normal vector.
- a white frame image to be disposed regarding the same side may be disposed differently at both ends of the side thereof.
- FIGS. 55A and 55B are diagrams illustrating an example of a determining method for a decoration image by the decoration image determining unit 700 according to the fourth embodiment of the present invention.
- FIG. 55A illustrates a case where a frame 760 is subjected to affine transformation with a frame 750 as a reference frame.
- a frame 760 is subjected to affine transformation with a frame 750 as a reference frame.
- an angle formed of a movement vector 751 , and a normal vector 753 is small
- an angle formed of a movement vector 752 and a normal vector 754 is small
- around the termination of the movement vector 751 is the inner side.
- around the termination of the movement vector 752 may be determined to be disposed in the other side.
- the other sides may be determined in the same way.
- a white frame image may be disposed such as shown in FIG. 55B .
- a white frame image is disposed in such a way, there is a possibility that the outer shape of the image may be changed.
- FIG. 55B in the case that a frame 760 is a rectangle, an image to which a white frame image has been added becomes a parallelogram in some cases.
- the target frame is a rectangle
- a case is assumed wherein an image to which a white frame image has been added becomes a trapezoid
- a case is assumed wherein an image to which a white frame image has been added becomes a trapezoid
- FIGS. 56A through 56D are diagrams illustrating a compositing example in which decoration images determined by the decoration image determining unit 700 according to the fourth embodiment are composited.
- the image compositing unit 220 executes this decoration image compositing based on the content of decoration image data 278 of the compositing target image metadata storage unit 277 . That is to say, based on the content of affine transformation parameters 274 of the compositing target image metadata storage unit 277 , the image is subjected to affine transformation by the image transforming unit 190 .
- the image compositing unit 220 executes image compositing by adding a decoration image to an image after affine transformation.
- FIGS. 56A through 56D influence of shift in the translational direction will be described. Note that FIGS. 56A through 57D illustrate an example wherein a white frame is added as a decoration image.
- FIG. 56A illustrates compositing target images 738 and 739 .
- FIG. 56B illustrates an example wherein the compositing target image 739 is subjected to affine transformation with the compositing target image 738 as a reference image. With the example shown in FIG. 56B , a case is illustrated wherein the compositing target image 739 is shifted to lower right. For example, with an image region surrounded with a circle 742 , the portion of a roof included in the compositing target image 738 is seen excessively. Also, for example, with an image region surrounded with a circle 743 , the portion of grass included in the compositing target image 738 is hidden with the compositing target image 739 .
- a decoration image 745 determined by the decoration image determining unit 700 is composited as to the compositing target image 739 , for example, whereby an unnatural image included in the region portions surrounded with circles 742 through 744 can be hidden.
- FIG. 56D in the case that a decoration image 746 is composited as to the compositing target image 739 in the opposite direction of the region determined by the decoration image determining unit 700 , for example, an unnatural image included in the region portions surrounded with circles 742 through 744 may not be hidden.
- FIGS. 57A through 57D are diagrams illustrating a compositing example in which decoration images determined by the decoration image determining unit 700 according to the fourth embodiment are composited. Note that, with the example shown in FIGS. 57A through 57D , influence of shift in the enlargement/reduction direction will be described.
- FIG. 57A illustrates compositing target images 747 and 748 .
- FIG. 57B illustrates an example of a case where the composited image 748 is subjected to affine transformation and composited with the compositing target image 747 as a reference image.
- FIG. 57B a case is shown wherein the compositing target image 747 is shifted in the reduction direction.
- the portion of a roof included in the compositing target image 747 is seen excessively.
- the portion of an electric light included in the compositing target image 747 is seen excessively.
- a decoration image 757 determined by the decoration image determining unit 700 is composited as to the compositing target image 748 , for example, whereby an unnatural image included in the region portions surrounded with circles 749 , 755 , and 756 can be hidden.
- FIG. 57C a decoration image 757 determined by the decoration image determining unit 700 is composited as to the compositing target image 748 , for example, whereby an unnatural image included in the region portions surrounded with circles 749 , 755 , and 756 can be hidden.
- a decoration image 758 is composited as to the compositing target image 748 in the opposite direction of the region determined by the decoration image determining unit 700 , for example, an unnatural image included in the region portions surrounded with the circles 749 , 755 , and 756 may not be hidden.
- FIGS. 58A through 58D are diagrams illustrating a compositing example in which the decoration image determined by the decoration image determining unit 700 according to the fourth embodiment is composited. Note that, with the example shown in FIGS. 58A through 58D , influence of shift in the rotational direction will be described.
- FIG. 58A illustrates compositing target images 759 and 762 .
- FIG. 58B illustrates an example of a case where the composited image 762 is subjected to affine transformation and composited with the compositing target image 759 as a reference image. With the example shown in FIG. 58B , a case is shown wherein the compositing target image 762 is shifted in the clockwise direction. For example, with an image region surrounded with a circle 763 , the portion of a stone pavement included in the compositing target image 759 is seen unnaturally. Therefore, as shown in FIG.
- a decoration image 764 determined by the decoration image determining unit 700 is composited as to the compositing target image 762 , for example, whereby an unnatural image included in the region portion surrounded with a circle 763 can be hidden.
- a decoration image 765 is composited as to the compositing target image 762 in the opposite direction of the region determined by the decoration image determining unit 700 , for example, an unnatural image included in the region portion surrounded with the circle 763 may not be hidden.
- the compositing target is shifted simply in the rotational direction, as described in FIGS. 54A through 54C , the sum of two movement vectors becomes 0. Therefore, a white frame image may be determined not only to be disposed in either the outer side or the inner side of the periphery, but also to be disposed above the periphery thereof.
- the decoration image determined by the decoration image determining unit 700 is composited as to a compositing target image, whereby a panorama-style natural composited image can be provided to the user.
- a face included in the composited image overlaid on the lower side may be hidden.
- the position of a decoration image can be changed in a direction where the face is not hidden.
- the heaviness of a decoration image may be adjusted.
- FIG. 59 is a diagram illustrating an example of a decoration image determining method by the decoration image determining unit 700 according to the fourth embodiment, which takes faces into consideration. Now, an example is shown wherein the decoration images of compositing target images 766 through 768 are determined. Also, let us say that the compositing target images 767 and 768 include a face, and these regions including a face are represented with face regions 769 and 770 .
- the compositing target images 767 and 768 are subjected to affine transformation and composited with the compositing target image 766 as a reference image.
- the compositing target image 768 is overwritten and composited as to the compositing target image 767 , and accordingly, for example, in the case that a decoration image to be added to the compositing target image 768 is great (or in the case of being added to the outer side), there is a possibility that a face included in the face region 769 may be hidden.
- a face region included in a compositing target image can be determined using the face data stored in face data 276 of the compositing target image metadata storage unit 277 . Therefore, for example, the decoration image determining unit 700 calculates whether or not the region of a decoration image determined before a face is taken into consideration, and a face region are overlaid, and in the case that the region of the decoration image and the face region are overlaid, the position of the decoration image is changed to the opposite side of the determined position.
- the decoration image 771 is changed to be added to the inner side, i.e., changed to be a decoration image 772 .
- the decoration image 773 is changed to be added to the outer side, i.e., changed to be a decoration image 774 .
- the heaviness of the decoration image may be changed by calculating the distance between the decoration image and the face region.
- the position of the decoration image is changed to a position where an overlaid area is small.
- the decoration may be determined to be added to the outer side of the compositing target image.
- an arrangement may be made wherein the distance between the face region and the decoration image is calculated for each side beforehand, the maximum heaviness of the decoration image not to be overlaid above a face is stored in the compositing target image metadata storage unit 277 beforehand, and the decoration image is added using these values.
- FIGS. 60A through 60C are diagrams illustrating an example of a shadow determined to be a decoration image to be added to a compositing target image by the decoration image determining unit 700 according to the fourth embodiment of the present invention.
- a shadow to be added to a compositing target image is indicated by adding a diagonal stroke internally.
- FIG. 60A illustrates the basic shape of a shadow to be added to a compositing target image.
- a shadow such as a shadow 351
- an affect can also be produced wherein light is applied from obliquely upward or obliquely downward of a compositing target image.
- FIG. 60B illustrates a case where an effect is produced such that light is applied from obliquely upward of a compositing target image to create a composited image. That is to say, a composited image to which the shadow 352 has been added is composited sequentially, whereby the composited image shown in FIG. 60B can be created.
- FIG. 60C illustrates a case where an effect is produced wherein light is applied from obliquely downward of a compositing target image to create a compositing target. That is to say, a composited image to which the shadow 353 has been added is composited sequentially, whereby the composited image shown in FIG. 60C can be created.
- the heaviness of a shadow may be set to a certain value, for example. Also, for example, in the same way as the case of a white frame described above, the heaviness of a shadow may be determined according to a over-extension rate.
- FIGS. 61A and 61B are diagrams illustrating an overview of a determining method arranged to determine the direction of a shadow by the decoration image determining unit 700 according to the fourth embodiment.
- FIG. 61A illustrates compositing target images 354 through 357 held at a work buffer 358 .
- the lengths of the sides of a compositing target image overwritten above another compositing target image are calculated, the outward normal vectors of the sides thereof are calculated, and the sizes of the normal vectors are taken as the lengths of overwritten portions of the sides.
- a side overwritten above another compositing target image is indicated with a heavy line.
- FIG. 61B illustrates composited images formed the compositing target images 354 through 357 to which a shadow thus determined is added. The direction of a shadow is thus determined, whereby the shadow is projected above a compositing target image, and accordingly, the effect of the shadow can be enhanced.
- the direction of a shadow may be selected and determined from several directions determined beforehand with reference to the direction of a normal vector.
- an arrangement may be made wherein an affine matrix obtained by multiplying the affine transformation parameters correlated with each compositing target image is used to obtain the integral translational component of camera work, and the direction of this integral translational component is taken as the direction of a shadow.
- a face included in a composited image overlaid on the lower side may be hidden.
- the position of a shadow can be changed to a direction where the face is not hidden.
- the heaviness of a shadow may be adjusted.
- the color of the shadow may be diluted. An example will be shown below wherein in the case that a face is covered with a shadow, the direction of a normal vector at the side thereof is inverted, thereby changing the direction of the shadow.
- FIG. 62 is a diagram illustrating an example of a shadow determining method by the decoration image determining unit 700 according to the fourth embodiment, which takes faces into consideration.
- the example shown in FIG. 62 is the same as the case shown in FIG. 61A except that of compositing target images 354 through 357 held at the work buffer 358 , a face region 359 is included in the compositing target image 356 .
- a normal vector is obtained in the same way as the case shown in FIG. 61A , but in the case that the face region is covered with a shadow, the direction of the normal vector at the side thereof is inverted.
- the direction of the normal vector at the side thereof is inverted. That is to say, the direction of the normal vector thereof is inverted such as a normal vector 360 .
- a method used for determining the direction of a shadow is the same as with the case shown in FIG. 61A except that the direction of a normal vector is inverted, and accordingly, description thereof will be omitted here.
- the direction of the integral vector of a normal vector calculated after the direction of a normal vector is inverted is almost the same as with the case of not inverting the direction thereof, and accordingly, the face is hidden with a shadow.
- the directions of the shadows are upper-right directions, whereby the face can be prevented from being hidden with the shadows.
- the density of the whole shadow may be diluted according to the direction and length of an inverted normal vector.
- the heaviness of a shadow may be thinned according to the direction and length of an inverted normal vector.
- a shadow changing method may be used wherein the density of the whole shadow is diluted by a certain rate, or the heaviness thereof is thinned by a certain rate.
- a white frame or a shadow is determined to be a decoration image
- the fourth embodiment of the present invention may also be applied to a case where a white frame is added to the periphery of a compositing target image, and also a shadow is added to the periphery of this white frame.
- a decoration image to be added to a compositing target image is determined at the time of frame selection, but for example, a decoration image to be added to a compositing target image may be determined at the time of display of a composited image.
- a decoration image may be determined according to the user's operation. For example, whether or not there is a white frame, the position of a white frame, the heaviness of a white frame, whether or not there is a shadow, the direction of a shadow, or the like, of a compositing target image may be specified by the user's operation.
- a white frame or the heaviness and position of a shadow are determined to be a decoration image, but for example, the color or mode of a decoration image may be determined based on a over-extension rate or normal vector.
- a decoration image is determined based on the relative positional relationship between compositing target images, or the like, but for example, a decoration image may be determined based on the attributes, color, or the like of a subject included in a compositing target image.
- the color histogram of the peripheral portion of a compositing target image is calculated, and the complementary color of a most frequent color may be determined to be the color of a frame.
- a frame to which a color other than white has been added is added to a compositing target image, whereby the edge of the compositing target image can be prevented from falling into obscurity.
- a reddish color is employed as the complementary color as to green.
- the color or mode of a frame may be determined using each piece of metadata relating to a face shown in the third embodiment of the present invention.
- the color of a frame can be changed according to the gender or age of a person included in a compositing target image.
- the color of a frame can be changed according to a point in time wherein a compositing target image has been imaged.
- compositing may be executed while correcting shift at the time of displaying a composited image using the camera work parameters and reliability calculated by the method shown in FIGS. 52A through 52C .
- the camera work parameters and reliability calculated by the method shown in FIGS. 52A through 52C are stored in the compositing target image metadata storage unit 277 beforehand. Subsequently, when displaying a compositing target image, with regard to a compositing target image having high reliability, the affine transformation parameters used for transforming a compositing target image can be corrected using the corresponding camera work parameters.
- FIG. 63 is a flowchart illustrating the procedure of a compositing target image metadata recording process by the image processing apparatus 102 according to the fourth embodiment of the present invention.
- This procedure is a modification of the procedure shown in FIG. 34 , the same procedure as the procedure shown in FIG. 34 will be denoted with the same reference numerals, and description thereof will be omitted here.
- the decoration image determining unit 700 executes a decoration image determining process (step S 1020 ). This decoration image determining process will be described in detail with reference to FIG. 64 .
- FIG. 64 is a flowchart illustrating a decoration image determining procedure (the procedure in step S 1020 shown in FIG. 63 ) included in the procedure of the compositing target image metadata recording process by the image processing apparatus 102 according to the fourth embodiment of the present invention.
- a white frame is determined to be a decoration image.
- a white frame to be added to each frame included in the same image compositing target section has the same heaviness.
- the metadata corresponding to each compositing target image included in a single image compositing target section selected by the compositing target image selecting unit 170 is obtained from a metadata file stored in the metadata storage unit 210 (step S 1021 ).
- the size for compositing each compositing target image using the affine transformation parameters corresponding to each compositing target image included in the obtained image compositing target section is calculated, and two work buffers having this size are secured (step S 1022 ).
- the two work buffers are work buffers that represent each pixel making up an image by 8 bits, and with this example, these work buffers are indicated as a first work buffer and a second work buffer.
- the front-most frame is selected as a reference frame (step S 1023 ).
- the reference frame is pasted to the first work buffer (step S 1024 ).
- the frame adjacent to the reference frame is pasted to the second work buffer (step S 1025 ).
- the frame pasted immediately before is eliminated, and the frame alone adjacent to the reference frame is pasted to the second work buffer.
- the over-extension rate of an overlaid region between the history image pasted to the first work buffer, and the frame to be pasted to the second work buffer is calculated (step S 1026 ). This over-extension rate is calculated by the calculation method shown in FIGS. 52A through 52C .
- step S 1027 determination is made whether or not the calculation of a over-extension rate has been completed.
- the frame of which the over-extension rate has been calculated immediately before is selected as a reference frame (step S 1028 ), and the flow returns to step S 1024 .
- the new selected reference frame is pasted above the history image of the first work buffer (step S 1024 ). Note that in the case that the frame of which the over-extension rate has been calculated immediately before is the top frame or last frame of the image compositing target section, a frame adjacent to the other direction of the front-most frame is selected as a reference frame.
- step S 1027 the heaviness of a white frame to be added to each compositing target image included in the image compositing target section serving as a selection target is calculated based on the over-extension rate calculated regarding each compositing target section (step S 1029 ). Subsequently, the calculated heaviness of the white frame is recorded in the compositing target image metadata storage unit 277 (step S 1030 ). Determination is made whether or not the decoration image determining process has been completed regarding all the image compositing target sections selected by the section selecting unit 160 (step S 1031 ).
- step S 1031 the flow returns to step S 1021 , where the decoration image determining process is repeated (steps S 1021 through S 1030 ).
- step S 1031 the operation of the decoration image determining process is ended.
- FIG. 65 is a flowchart illustrating a decoration image determining procedure (the procedure in step S 1020 shown in FIG. 63 ) included in the procedure of the compositing target image metadata recording process by the image processing apparatus 102 according to the fourth embodiment of the present invention.
- This example is an example wherein a shadow is determined to be a decoration image, and is a modification of the procedure shown in FIG. 64 . Therefore, only the procedure different from the procedure shown in FIG. 64 will be described, ant other description will be omitted.
- a normal vector between the history image pasted to the first work buffer, and the frame pasted to the second work buffer is calculated (step S 1041 ). This normal vector is calculated by the calculation method shown in FIGS. 61A and 61B .
- step S 1027 the integral value of the calculated normal vectors is calculated regarding each compositing target image, the direction of a normal vector obtained by integration is calculated as the direction where a shadow is added (step S 1042 ). Subsequently, the calculated direction where a shadow is added is stored in the compositing target image metadata storage unit 277 (step S 1043 ).
- FIG. 66 is a flowchart illustrating the procedure of a composited image display process by the image processing apparatus 102 according to the fourth embodiment of the present invention.
- This procedure is a modification of the procedure shown in FIG. 37 , and accordingly, the same procedure as the procedure shown in FIG. 37 will be denoted with the same reference numerals, and description thereof will be omitted here.
- the image compositing unit 221 adds a decoration image to the periphery of the image subjected to affine transformation based on decoration image data included in each piece of metadata obtained (step S 1051 ), and holds the image to which the decoration image has been added at the image memory 230 (step S 1052 ). That is to say, the image subjected to affine transformation to which the decoration image has been added is composited as to the history image held at the image memory 230 .
- step S 1060 an image compositing process is executed (step S 1060 ). This image compositing process will be described in detail with reference to FIG. 67 .
- FIG. 67 is a flowchart illustrating an image compositing procedure (the procedure in step S 1060 shown in FIG. 66 ) included in the procedure of the composited image display process by the image processing apparatus 102 according to the fourth embodiment of the present invention.
- This procedure is a modification of the procedure shown in FIG. 38 , and accordingly, the same procedure as the procedure shown in FIG. 38 will be denoted with the same reference numerals, and description thereof will be omitted here.
- the image compositing unit 221 adds a decoration image to the periphery of the image subjected to affine transformation based on decoration image data included in each piece of metadata obtained (step S 1061 ), and holds the image to which the decoration image has been added at the image memory 230 (step S 1062 ).
- the image compositing unit 221 adds a decoration image to the periphery of the image subjected to affine transformation based on decoration image data included in each piece of metadata obtained (step S 1063 ), and holds the image to which the decoration image has been added at the image memory 230 (step S 1064 ).
- the image compositing unit 221 adds a decoration image to the periphery of the image subjected to affine transformation based on decoration image data included in each piece of metadata obtained (step S 1065 ), and holds the image to which the decoration image has been added at the image memory 230 (step S 1066 ).
- FIG. 68 is a diagram illustrating a configuration example of a multi-core processor 800 according to the fifth embodiment of the present invention.
- the multi-core processor 800 is a processor wherein different types of processor cores are implemented in a single CPU (Central Processing Unit) package. Specifically, two types of processor cores which can handle all applications are implemented in the multi-core processor 800 to maintain the processing capability of each processor core and to realize a simple configuration.
- the two types of processor cores are one type of cores, and the other type of cores optimized partly for a predetermined application.
- the multi-core processor 800 includes a control processor core 801 , arithmetic processor cores (# 1 ) 811 through (# 8 ) 818 , and a bus 802 , and is connected to main memory 781 . Also, the multi-core processor 800 is connected to other devices, for example, a graphics device 782 , an I/O device 783 , and so forth. As the multi-core processor 800 , for example, a Cell (Cell Broadband Engine) that is a microprocessor developed by the present applicant and others may be employed.
- Cell Cell Broadband Engine
- the control processor 801 is a control processor core configured to principally execute frequent thread switching such as an operating system, or the like. Note that the control processor core 801 will be described in detail with reference to FIG. 69 .
- the arithmetic processor cores (# 1 ) 811 through (# 8 ) 818 are simple small arithmetic processor cores which excel at the process of a multimedia system. Note that the arithmetic processor cores (# 1 ) 811 through (# 8 ) 818 will be described in detail with reference to FIG. 70 .
- the bus 802 is a high-speed bus called as an element interconnect bus (EIB). Also, the bus 802 is connected to the control processor 801 , and each of the arithmetic processor cores (# 1 ) 811 through (# 8 ) 818 , and data access by each processor core is executed via the bus 802 .
- EIB element interconnect bus
- the main memory 781 is connected to the bus 802 , and is main memory which stores various types of program to be loaded to each processor core, data for the process by each processor core, and data processed by each processor core.
- the graphics device 782 is a graphics device connected to the bus 802
- the I/O device 783 is an external input/output device connected to the bus 802 .
- FIG. 69 is a diagram illustrating a configuration example of the control processor core 801 according to the fifth embodiment of the present invention.
- the control processor core 801 includes a control processor unit 803 , and a control processor storage system 806 .
- the control processor unit 803 is a unit making up a core configured to execute the arithmetic process of the control processor core 801 . Also, the control processor unit 803 includes a command set with the architecture of a microprocessor as basis. A command cache 804 and a data cache 805 are implemented in the control processor unit 803 as primary caches.
- the command cache 804 is, for example, a command cache of 32 KB
- the data cache 805 is, for example, a data cache of 32 KB.
- the control processor storage system 806 is a unit which controls data access from the control processor unit 803 to the main memory 781 . Also, with the control processor storage system 806 , a secondary cache 807 of 512 KB is mounted to increase memory access speed from the control processor unit 803 .
- FIG. 70 is a diagram illustrating a configuration example of the arithmetic processor core (# 1 ) 811 according to the fifth embodiment of the present invention.
- the arithmetic processor (# 1 ) 811 includes an arithmetic processor unit 820 and a memory flow controller 822 .
- the arithmetic processor cores (# 2 ) 812 through (# 8 ) 818 have the same configuration as the arithmetic processor core (# 1 ) 811 , and accordingly, description thereof will be omitted here.
- the arithmetic processor unit 820 is a unit serving as a core which executes the arithmetic process of the arithmetic processor core (# 1 ) 811 , and includes an original command set different from the control processor unit 803 of the control processor core 801 . Also, a local store (LS) 821 is implemented in the arithmetic processor unit 820 .
- LS local store
- the local store 821 is the dedicated memory of the arithmetic processor unit 820 , and is the unique memory that can refer to the arithmetic processor unit 820 directly. For example, memory of which the capacity is 256 KB can be employed as the local store 821 . Note that the arithmetic processor unit 820 has to take advantage of the memory flow controller 822 to access the local store above the main memory 781 and the other arithmetic processor cores (arithmetic processor cores (# 2 ) 812 through (# 8 ) 818 ).
- the memory flow controller 822 is a unit used to exchange data with the main memory 781 or other arithmetic processor cores or the like, and is a unit called as a memory flow controller (MFC).
- MFC memory flow controller
- the arithmetic processor unit 820 requests data transfer or the like as to the memory flow controller 822 via an interface called as a channel.
- model has been familiar as the most fundamental model of this programming model wherein a main program is executed over the control processor core 801 , and a sub program is executed over the arithmetic processor cores (# 1 ) 811 through (# 8 ) 818 .
- the computation method of the multi-core processor 800 using this model will be described in detail with reference to the drawings.
- FIG. 71 is a diagram schematically illustrating the computation method of the multi-core processor 800 according to the fifth embodiment of the present invention.
- the control processor core 801 controls each arithmetic processor core to use data 787 (a portion of the data 785 ) for the process of a task 786 that is a portion of the task 784 to execute the task 786 .
- the control processor core 801 controls each arithmetic processor core to use the data 787 (a portion of the data 785 ) for the process of a task 786 that is a portion of the task 784 to execute the task 786 .
- a computation process is executed by each arithmetic processor core for each frame making up a moving picture.
- the multi-core processor 800 executes a computation, whereby the arithmetic processor cores (# 1 ) 811 through (# 8 ) 818 are used in parallel, and many computations can be executed in comparatively little time. Also, a single instruction/multiple data (SIMD) computation is executed over the arithmetic processor cores (# 1 ) 811 through (# 8 ) 818 , whereby relatively many computations can be executed with a still smaller number of commands. Note that SIMD computation will be described in detail with reference to FIGS. 75 through 78 .
- SIMD single instruction/multiple data
- FIG. 72 is a diagram schematically illustrating the flows of a program and data in the case that a computation is executed by the multi-core processor 800 according to the fifth embodiment of the present invention.
- description will be made regarding the arithmetic processor core (# 1 ) 811 of the arithmetic processor cores (# 1 ) 811 through (# 8 ) 818 as an example, but a computation can also be executed similarly regarding the arithmetic processor cores (# 2 ) 812 through (# 8 ) 818 .
- control processor core 801 transmits to the arithmetic processor core (# 1 ) 811 an instruction to load an arithmetic processor core program 823 stored in the main memory 781 to the local store 821 of the arithmetic processor core (# 1 ) 811 .
- the arithmetic processor core (# 1 ) 811 loads the arithmetic processor core program 823 stored in the main memory 781 to the local store 821 .
- control processor core 801 instructs the arithmetic processor core (# 1 ) 811 to execute the arithmetic processor core program 825 stored in the local store 821 .
- the arithmetic processor core (# 1 ) 811 transfers data 824 for the execution process of the arithmetic processor core program 825 stored in the local store 821 from the main memory 781 to the local store 821 .
- the arithmetic processor core (# 1 ) 811 processes the data 826 transferred from the main memory 781 based on the arithmetic processor core program 825 stored in the local store 821 , and executes a process according to a condition to store the processing results in the local store 821 .
- the arithmetic processor core (# 1 ) 811 transfers the processing results executed based on the arithmetic processor core program 825 stored in the local store 821 from the local store 821 to the main memory 781 .
- the arithmetic processor core (# 1 ) 811 informs the control processor core 801 of end of the computation.
- SIMD computation is a computation method wherein a process as to a plurality of data is executed with a single command.
- FIG. 73 is a diagram schematically illustrating an overview of a computation method arranged to execute a process as to a plurality of data by each corresponding command.
- the computation method shown in (a) in FIG. 73 is a common computation method, and is referred to as, for example, a scalar computation.
- the processing results of data “C 1 ” are obtained with a command used for adding data “A 1 ” and data “B 1 ”.
- a command to add data “A 2 ”, “A 3 ”, and “A 4 ”, and data “B 2 ”, “B 3 ”, and “B 4 ” on the same row respectively is executed regarding each of the computations.
- FIG. 73 is a diagram schematically illustrating an overview of SIMD computation that is a computation method arranged to execute a process as to a plurality of data by a single command.
- data integrated for SIMD computation (each piece of data surrounded with dotted lines 827 and 828 ) is referred to as vector data in some cases.
- SIMD computation executed using such vector data is referred to as a vector computation in some cases.
- vector data surrounded with a dotted line 827 is “A 1 ”, “A 2 ”, “A 3 ”, and “A 4 ”
- vector data surrounded with the dotted line 828 is “B 1 ”, “B 2 ”, “B 3 ”, and “B 4 ”.
- control processor core 801 of the multi-core processor 800 executes a command relating to these SIMD computations, and the arithmetic processor cores (# 1 ) 811 through (# 8 ) 818 execute a parallel process regarding the computations of a plurality of data as to such a command.
- a process to add between data “A 1 ” and “B 1 ”, a process to subtract between data “A 2 ” and “B 2 ”, a process to multiply between data “A 3 ” and “B 3 ”, and a process to divide between data “A 4 ” and “B 4 ” are not realized by SIMD computation. That is to say, in the case that a different process is executed as to each of a plurality of data, a process by SIMD computation is not realized.
- FIG. 74 is a diagram illustrating a configuration example of a program executed by the control processor core 801 or the arithmetic processor core (# 1 ) 811 , according to the fifth embodiment of the present invention.
- the arithmetic processor core (# 1 ) 811 alone will be illustrated, the same process will also be executed regarding the arithmetic processor core (# 2 ) 812 through (# 8 ) 818 .
- the control processor core 801 executes decode 852 , interlace 853 , and resize 854 as decode 851 .
- the decode 852 is a process to decode a moving picture file.
- the interlace 853 is a process to remove the interlace of each frame decoded.
- the resize 854 is a process to reduce each frame of which the interlace has been removed.
- control processor core 801 executes command transmissions 857 and 859 , and termination notice receptions 858 and 860 as arithmetic processor core management 856 .
- the command transmissions 857 and 859 are processes to transmit an execution command of SIMD computation as to the arithmetic processor cores (# 1 ) 811 through (# 8 ) 818 .
- the termination notice receptions 858 and 860 are processes to receive the termination notice of SIMD computation from the arithmetic processor cores (# 1 ) 811 through (# 8 ) 818 as to the above command.
- the control processor core 801 executes a camera work parameter calculating process 862 as camera work detection 861 .
- the camera work parameter calculating process 862 is a process to calculate affine transformation parameters for each frame based on the optical flow calculated with SIMD computation by the arithmetic processor cores (# 1 ) 811 through (# 8 ) 818 .
- the arithmetic processor core (# 1 ) 811 executes a Sobel filter process 864 , a second moment matrix process 865 as a feature point extracting process 863 . Also, the arithmetic processor core (# 1 ) 811 executes a separable filter process 866 , a Harris corner-points extracting (Calc Harris) process 867 , a dilation process 868 , and a sorting process 869 .
- the Sobel filter process 864 is a process to calculate a value dx in the X direction obtained by using a P2 filter (x direction), and a value dy in the y direction obtained by using a filter in the Y direction. Note that the calculation of the value dx in the X direction will be described in detail with reference to FIGS. 75 through 78 .
- the second moment matrix process 865 is a process to calculate each value of dx 2 , dy 2 , dx ⁇ dy using the dx and dy calculated by the Sobel filter process 864 .
- the separable filter process 866 is a process to apply a Gaussian filter (blurring process) to the images of the value of dx 2 , dy 2 , dx ⁇ dy calculated by the second moment filter process 865 .
- the Harris corner-points extracting process 867 is a process to use each value of the dx 2 , dy 2 , dx ⁇ dy subjected to a blurring process by the separable filter process 866 to calculate the score of a Harris corner.
- the dilation process 868 is a process to execute a blurring process as to an image made up of the score of a Harris corner calculated by the Harris corner-points extracting process 867 .
- the sorting process 869 is a process to sort pixels in the descending order of the score of a Harris corner calculated by the Harris corner-points extracting process 867 , pick up pixels by a predetermined number of pixels from a pixel having the highest score, and extract this picked-up pixels as feature points.
- the arithmetic processor core (# 1 ) 811 executes a pyramid image creating process 871 as an optical flow computing process 870 , and executes an optical flow calculating process 872 .
- the pyramid image creating process 871 is a process to sequentially create an image reduced through a predetermined number of stages from the image size at the time of imaging by the camera, and the created image is referred to as a multi-resolution image.
- the optical flow calculating process 872 is a process to calculate an optical flow regarding the smallest image of the multi-resolution images created by the pyramid image creating process 871 , and use this calculation result to calculate an optical flow again regarding a image having one-rank up resolution of the smallest image. Such a series of processes are executed repeatedly until the image to be processed reaches the largest image.
- the feature point extracting process, and the optical flow calculating process are executed in parallel with SIMD computation using the multi-core processor 800 , whereby processing results can be obtained.
- the feature point extracting process is a feature point extracting process executed by the feature point extracting unit 121 shown in FIG. 2 and others
- the optical flow calculating process is an optical flow calculating process executed by the optical flow calculating unit 122 .
- the feature point extracting process, and the optical flow calculating process shown in FIG. 74 and others are an example, another process made up of various types of filter process and a threshold process and the like as to an image making up a moving picture may be used to execute SIMD computation by the multi-core processor 800 .
- FIG. 75 is a diagram schematically illustrating a data structure and a processing flow in the case that image data stored in the main memory 781 according to the fifth embodiment of the present invention is subjected to a filtering process using the Sobel filter 830 .
- This image data is image data corresponding to a single frame making up a moving picture imaged by the camera.
- the image data stored in the main memory 781 shown in FIG. 75 is illustrated in a simplified manner with the number of horizontal pixels as 32 pixels.
- the Sobel filter 830 is a 3 ⁇ 3 edge extracting filter.
- the image data stored in the main memory 781 is subjected to a filtering process using the Sobel filter 830 , and the results of this filtering process are output.
- SIMD computation is used to obtain the results of four filters at once.
- FIG. 76 is a diagram schematically illustrating a data flow in the case that the Sobel filter 830 is used to execute SIMD computation regarding image data stored in the main memory 781 according to the fifth embodiment of the present invention.
- a predetermined number of rows including the first row of the image data stored in the main memory 781 are DMA (Direct Memory Access)-transferred to a first buffer 831 provided to the local store 821 of an arithmetic processor core.
- a predetermined number of rows wherein each row DMA-transferred to the first buffer 831 is shifted downward by one are DMA-transferred to a second buffer 832 .
- the predetermined number of rows is, for example, three rows.
- the double buffers are used, whereby delay due to DMA transfer can be covered.
- FIG. 77 is a diagram schematically illustrating a vector creating method arranged to create nine vectors from the image data stored in the first buffer 831 in the case that the filtering process is executed using the Sobel filter 830 according to the fifth embodiment of the present invention.
- FIG. 76 after DMA transfer is executed, nine vectors are created from the image data stored in the first buffer 831 .
- vector data 841 is created with four pieces of data from the left corner, and according to four pieces of data obtained by shifting the above four pieces to the right side by one, vector data 842 is created.
- vector data 843 is created.
- vector data 844 through 849 are created with four pieces of data.
- FIG. 78 is a diagram schematically illustrating a vector computing method arranged to use SIMD computation to execute a vector computation regarding vector data 841 through 849 in the case that the filtering process is executed using the Sobel filter 830 according to the fifth embodiment of the present invention.
- SIMD computation is executed sequentially regarding the vector data 841 through 843 , whereby a vector A is obtained.
- SIMD computation first, SIMD computation of ⁇ 1 ⁇ vector data 841 is executed.
- SIMD computation of 0 ⁇ vector data 841 is executed and SIMD computation of 1 ⁇ vector data 843 is executed.
- the computation result has been determined to be 0, and accordingly, this computation may be omitted.
- 1 ⁇ vector data 843 the computation result has been determined to be the same value as the vector data 843 , and accordingly this computation may be omitted.
- an addition process between the computation result of ⁇ 1 ⁇ vector data 841 , and the computation result of 0 ⁇ vector data 842 is executed with SIMD computation.
- an addition process between the result of the above addition process and the computation result of 1 ⁇ vector data 843 is executed with SIMD computation.
- the computation of a data structure serving as vector data 1 ⁇ vector data 2 +vector data 3 can be realized with SIMD computation. Therefore, with regard to the computation of the vector A, SIMD computation regarding, for example, 0 ⁇ vector data 842 , and 1 ⁇ vector data 843 will be omitted.
- ⁇ 1 ⁇ vector data 841 +vector data 843 may be executed with a one-time SIMD computation.
- SIMD computation is executed regarding the vector data 844 through 846 to obtain a vector B
- SIMD computation is executed regarding the vector data 847 through 849 to obtain a vector C.
- SIMD computation is executed regarding the obtained vectors A through C to obtain a vector D.
- SIMD computation is executed, whereby the results equivalent to the number of vector elements (four pieces of data in this example) can be obtained at the same time.
- the same process is executed repeatedly while shifting the position of data to be extracted to the right side by one, whereby calculation of the vector D regarding each piece of the image data is executed sequentially. Subsequently, in the case that the process up to the right edge of the image data stored in the first buffer 831 shown in FIG. 76 has been completed, the processing results are DMA-transferred to the main memory 781 .
- FIG. 79 is a diagram schematically illustrating the flow of a camera work parameter calculating process according to the fifth embodiment of the present invention in a time-series manner.
- SIMD computation is executed using the multi-core processor 800 , whereby decoding and analyzing processes regarding a moving picture can be executed in parallel. Therefore, the analyzing time of one frame making up a moving picture can be reduced as compared to the decode time.
- t 1 represents time for the decoding process of one frame making up a moving picture by the control processor core 801 .
- t 2 represents time for the feature point extracting process of one frame making up a moving picture by the arithmetic processor core (# 1 ) 811 through (# 8 ) 818 .
- t 3 represents time for the optical flow calculating process of one frame making up a moving picture by the arithmetic processor core (# 1 ) 811 through (# 8 ) 818 .
- t 4 represents time for the camera work detecting process of one frame making up a moving picture by the control processor core 801 .
- t 5 represents time for the camera work detecting process regarding one frame making up a moving picture by the control processor core 801 and the arithmetic processor core (# 1 ) 811 through (# 8 ) 818 .
- t 6 represents time for a process to manage the arithmetic processor core (# 1 ) 811 through (# 8 ) 818 by the control processor 801 .
- t 1 is set to 25.0 ms
- t 2 is set to 7.9 ms
- t 3 is set to 6.7 ms
- t 4 is set to 1.2 ms
- t 5 is set to 15.8 ms.
- FIG. 80 (a) is a top view schematically illustrating a Blu-ray Disc (registered trademark) 880 which is an example of a recording medium, and (b) in FIG. 80 a diagram schematically illustrating data 881 through data 884 recorded in the Blu-ray Disc 880 .
- a moving picture content 882 , a caption 883 , metadata 884 , and a Java (registered trademark) program 881 are recorded in the Blu-ray Disc 880 .
- the moving picture content 882 is a moving picture imaged by a camera or the like
- the caption 883 is the caption of the moving picture content 882
- the metadata 884 is metadata obtained by analyzing the moving picture content 882 (e.g., each piece of information shown in FIGS. 3 , 4 , etc.).
- the Java program 881 is a Java program according to moving picture playback with each embodiment of the present invention.
- FIG. 80 is a diagram schematically illustrating the internal configuration of a Blu-ray player capable of playing the Blu-ray Disc 880 .
- a CPU 891 a CPU 891 , an OS 892 , a Java VM (Java virtual machine), and a library 893 are implemented as default, and accordingly, a Java program can be executed. Therefore, the Blu-ray Disc 880 is mounted on the Blu-ray player 890 , whereby the Blu-ray player 890 can load and execute the Java program 881 .
- the metadata 884 is used to execute display of a panorama image corresponding to a moving picture according to each embodiment of the present invention, search of a moving picture from multiple moving pictures, or the like. That is to say, moving picture playback according to each embodiment of the present invention can be realized with all the Blu-ray players without using a dedicated PC software or the like.
- an atmosphere at an imaging place, a situation of the place thereof, and so forth can be propagated accurately, and the imaging space at the imaging place thereof can be recognized.
- the moving picture thereof can be made to seem more fun, and interest relating the moving picture can be raised.
- a prominent image is selected from a moving picture, and a composited image is created from this image, whereby a chance to view a scene imaged by the moving picture again can be readily provided.
- multiple images may be composited in a panorama manner to create one composited image such that multiple paper photos are overlaid.
- the embodiments of the present invention may be applied to an image processing apparatus in which an image output unit which outputs image information used for displaying a composited image on another image display apparatus. Further, the embodiments of the present invention may be applied to a moving picture player capable of playing a moving picture, an imaging apparatus such as a digital video camera or the like capable of playing an imaged moving picture, a photo frame capable of displaying an image based on image data, or the like.
- the embodiments of the present invention description has been made with the image processing apparatus as an example, but the embodiments of the present invention may be applied to a moving picture player or the like capable of playing a moving picture. Also, with the embodiments of the present invention, description has been made regarding a moving picture imaged by the camera, but the embodiments of the present invention may also be applied to, for example, a moving picture or the like wherein, in the case that a moving picture imaged by the camera is edited, a moving picture after editing or animation or the like are partially composited.
- the procedures described in the embodiments of the present invention may be regarded as a method including such a series of procedures, or may be regarded as a program used for causing a computer to execute such a series of procedures, or may be regarded a recording medium configured to store the program thereof.
- a CD Compact Disc
- MD MiniDisc
- DVD Digital Versatile Disk
- a memory card a Blu-ray Disc (registered trademark), or the like may be employed as the recording medium thereof.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Theoretical Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Biomedical Technology (AREA)
- Social Psychology (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Psychiatry (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Television Signal Processing For Recording (AREA)
- Processing Or Creating Images (AREA)
- Image Processing (AREA)
- Studio Devices (AREA)
- Image Analysis (AREA)
- Studio Circuits (AREA)
Abstract
Description
SHS=Nm1/N
R B {right arrow over (μ)}=λR W{right arrow over (μ)}
R W −1 R B{right arrow over (μ)}=λ{right arrow over (μ)}
Fisher vector {right arrow over (A)}d eigenvector {right arrow over (μ)}max as to the maximum eigenvalue λmax
XZth1<XZ<XZth2 (11)
YZth1<YZ<YZth2 (12)
XT<XTth (13)
YT<YTth (14)
XR<XRth (15)
YR<YRth (16)
|XR−YR|<ADRth (17)
XZm1/XZm2<XZth2
YZm1/YZm2<YZth2
|XT1|>XTth2
|YT1|>YTth2
Over-extension rate DR1=(1−(C/A))
Over-extension rate DR2=(1−(C/B))
The heaviness of a white frame SH1 to be added to the sides of both ends in the horizontal direction=W×DR
The heaviness of a white frame SH2 to be added to the sides of both ends in the vertical direction=H×DR
S=(dx 2 ×dy 2 −dx·dy×dx·dy)/(dx 2 +dy 2+ε)
Claims (21)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2008-275143 | 2008-10-27 | ||
JP2008275143A JP4623201B2 (en) | 2008-10-27 | 2008-10-27 | Image processing apparatus, image processing method, and program |
Publications (2)
Publication Number | Publication Date |
---|---|
US20100104217A1 US20100104217A1 (en) | 2010-04-29 |
US9106872B2 true US9106872B2 (en) | 2015-08-11 |
Family
ID=41718785
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/565,499 Active 2033-03-23 US9106872B2 (en) | 2008-10-27 | 2009-09-23 | Image processing apparatus, image processing method, and program |
Country Status (6)
Country | Link |
---|---|
US (1) | US9106872B2 (en) |
EP (1) | EP2180698B1 (en) |
JP (1) | JP4623201B2 (en) |
CN (1) | CN101729784A (en) |
BR (1) | BRPI0904357A2 (en) |
RU (1) | RU2009139617A (en) |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150302633A1 (en) * | 2014-04-22 | 2015-10-22 | Google Inc. | Selecting time-distributed panoramic images for display |
US9667859B1 (en) * | 2015-12-28 | 2017-05-30 | Gopro, Inc. | Systems and methods for determining preferences for capture settings of an image capturing device |
US9836054B1 (en) | 2016-02-16 | 2017-12-05 | Gopro, Inc. | Systems and methods for determining preferences for flight control settings of an unmanned aerial vehicle |
US9892760B1 (en) | 2015-10-22 | 2018-02-13 | Gopro, Inc. | Apparatus and methods for embedding metadata into video stream |
US9922387B1 (en) | 2016-01-19 | 2018-03-20 | Gopro, Inc. | Storage of metadata and images |
US9934222B2 (en) | 2014-04-22 | 2018-04-03 | Google Llc | Providing a thumbnail image that follows a main image |
US9967457B1 (en) | 2016-01-22 | 2018-05-08 | Gopro, Inc. | Systems and methods for determining preferences for capture settings of an image capturing device |
US9973792B1 (en) | 2016-10-27 | 2018-05-15 | Gopro, Inc. | Systems and methods for presenting visual information during presentation of a video segment |
USD830407S1 (en) | 2014-04-22 | 2018-10-09 | Google Llc | Display screen with graphical user interface or portion thereof |
USD830399S1 (en) | 2014-04-22 | 2018-10-09 | Google Llc | Display screen with graphical user interface or portion thereof |
US10187607B1 (en) | 2017-04-04 | 2019-01-22 | Gopro, Inc. | Systems and methods for using a variable capture frame rate for video capture |
USD868092S1 (en) | 2014-04-22 | 2019-11-26 | Google Llc | Display screen with graphical user interface or portion thereof |
US20210134049A1 (en) * | 2017-08-08 | 2021-05-06 | Sony Corporation | Image processing apparatus and method |
Families Citing this family (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2009151896A (en) | 2007-12-21 | 2009-07-09 | Sony Corp | Image processing system, motion picture reproducing system, and processing method and program for them |
JP5226600B2 (en) * | 2009-04-28 | 2013-07-03 | 富士フイルム株式会社 | Image deformation apparatus and operation control method thereof |
EP2524324B1 (en) * | 2010-01-11 | 2019-04-24 | InterDigital Madison Patent Holdings | Method of navigation among identifiers placed in zones and device to implement such method |
JP5521778B2 (en) * | 2010-05-26 | 2014-06-18 | 富士通株式会社 | Measuring device and control device |
JP5665380B2 (en) * | 2010-06-15 | 2015-02-04 | キヤノン株式会社 | Image processing apparatus, image processing apparatus control method, program, and recording medium |
JP2012191273A (en) * | 2011-03-08 | 2012-10-04 | Toshiba Corp | Electronic apparatus and display method |
JP2012198781A (en) * | 2011-03-22 | 2012-10-18 | Sony Corp | Information processing apparatus, information processing method, and program |
JP5821625B2 (en) | 2011-08-29 | 2015-11-24 | カシオ計算機株式会社 | Image editing apparatus and program |
US9560334B2 (en) * | 2011-09-08 | 2017-01-31 | Qualcomm Incorporated | Methods and apparatus for improved cropping of a stereoscopic image pair |
JP2013074572A (en) | 2011-09-29 | 2013-04-22 | Casio Comput Co Ltd | Image processing apparatus, image processing method, and program |
KR101826704B1 (en) | 2013-09-13 | 2018-02-08 | 인텔 코포레이션 | Video production sharing apparatus and method |
WO2015137788A1 (en) * | 2014-03-14 | 2015-09-17 | Samsung Electronics Co., Ltd. | Electronic apparatus for providing health status information, method of controlling the same, and computer-readable storage medium |
US10664687B2 (en) * | 2014-06-12 | 2020-05-26 | Microsoft Technology Licensing, Llc | Rule-based video importance analysis |
CN107111468B (en) * | 2014-10-14 | 2021-06-11 | 巴尔科股份有限公司 | Display system with virtual display |
JP2016163089A (en) * | 2015-02-27 | 2016-09-05 | 株式会社Jvcケンウッド | Image connection device, image connection method, and image connection program |
CN107015788B (en) * | 2016-10-19 | 2020-10-30 | 创新先进技术有限公司 | Method and device for displaying images on mobile device in animation mode |
US10346461B1 (en) * | 2018-04-02 | 2019-07-09 | Pond5 Inc. | Method and system for image searching by color |
US10242035B1 (en) * | 2018-04-02 | 2019-03-26 | Pond5, Inc. | Method and system for image searching |
CN110956603B (en) | 2018-09-25 | 2023-04-21 | Oppo广东移动通信有限公司 | Detection method and device for edge flying spot of depth image and electronic equipment |
CN113168368B (en) * | 2018-11-28 | 2023-09-29 | 株式会社特拉斯特技术 | Programming device and recording medium |
US12067051B1 (en) * | 2020-03-19 | 2024-08-20 | Kipling Conrad Singh Warner | Music and content recommendation, identification, similarity evaluation, and matching |
CN111563922B (en) * | 2020-03-26 | 2023-09-26 | 北京迈格威科技有限公司 | Visual positioning method, visual positioning device, electronic equipment and storage medium |
CN113033552B (en) * | 2021-03-19 | 2024-02-02 | 北京字跳网络技术有限公司 | Text recognition method and device and electronic equipment |
Citations (50)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5818975A (en) * | 1996-10-28 | 1998-10-06 | Eastman Kodak Company | Method and apparatus for area selective exposure adjustment |
US6009190A (en) * | 1997-08-01 | 1999-12-28 | Microsoft Corporation | Texture map construction method and apparatus for displaying panoramic image mosaics |
US6078701A (en) * | 1997-08-01 | 2000-06-20 | Sarnoff Corporation | Method and apparatus for performing local to global multiframe alignment to construct mosaic images |
US6246413B1 (en) * | 1998-08-17 | 2001-06-12 | Mgi Software Corporation | Method and system for creating panoramas |
US20010005208A1 (en) * | 1997-06-20 | 2001-06-28 | Nippon Telegraph And Telephone Corporation | Scheme for interactive video manipulation and display of moving object on background image |
US6393162B1 (en) * | 1998-01-09 | 2002-05-21 | Olympus Optical Co., Ltd. | Image synthesizing apparatus |
US20020122113A1 (en) * | 1999-08-09 | 2002-09-05 | Foote Jonathan T. | Method and system for compensating for parallax in multiple camera systems |
US20020136590A1 (en) * | 1997-07-21 | 2002-09-26 | Barry Himmel | Personalized event/theme photograph album |
US20020135672A1 (en) * | 1999-01-08 | 2002-09-26 | Sezan M. Ibrahim | System for constructing mosaic images |
US20020196340A1 (en) * | 2001-04-24 | 2002-12-26 | Matsushita Electric Industrial Co., Ltd. | Image synthesis display method and apparatus for vehicle camera |
JP2003304444A (en) | 2002-04-08 | 2003-10-24 | Inc Corp | Apparatus and method for compositing image |
US20030234866A1 (en) * | 2002-06-21 | 2003-12-25 | Ross Cutler | System and method for camera color calibration and image stitching |
US6677981B1 (en) * | 1999-12-31 | 2004-01-13 | Stmicroelectronics, Inc. | Motion play-back of still pictures comprising a panoramic view for simulating perspective |
US20040189691A1 (en) * | 2003-03-28 | 2004-09-30 | Nebojsa Jojic | User interface for adaptive video fast forward |
WO2005024723A1 (en) | 2003-09-08 | 2005-03-17 | Nec Corporation | Image combining system, image combining method, and program |
US20050058360A1 (en) * | 2003-09-12 | 2005-03-17 | Thomas Berkey | Imaging system and method for displaying and/or recording undistorted wide-angle image data |
US6891561B1 (en) * | 1999-03-31 | 2005-05-10 | Vulcan Patents Llc | Providing visual context for a mobile active visual display of a panoramic region |
US20050099494A1 (en) * | 2003-11-10 | 2005-05-12 | Yining Deng | Digital camera with panoramic image capture |
US20050129272A1 (en) * | 2001-11-30 | 2005-06-16 | Frank Rottman | Video monitoring system with object masking |
WO2006005231A1 (en) * | 2004-07-14 | 2006-01-19 | Kun Ma | A portable panoramic photographic apparatus |
US7027054B1 (en) * | 2002-08-14 | 2006-04-11 | Avaworks, Incorporated | Do-it-yourself photo realistic talking head creation system and method |
US20070064974A1 (en) * | 2005-09-19 | 2007-03-22 | International Business Machines Corporation | Video data management |
US20070104383A1 (en) * | 2002-11-14 | 2007-05-10 | Microsoft Corporation | Stabilization of objects within a video sequence |
US7227893B1 (en) * | 2002-08-22 | 2007-06-05 | Xlabs Holdings, Llc | Application-specific object-based segmentation and recognition system |
US20070162873A1 (en) * | 2006-01-10 | 2007-07-12 | Nokia Corporation | Apparatus, method and computer program product for generating a thumbnail representation of a video sequence |
US20070182818A1 (en) * | 2005-09-02 | 2007-08-09 | Buehler Christopher J | Object tracking and alerts |
US20070206945A1 (en) * | 2006-03-01 | 2007-09-06 | Delorme David M | Method and apparatus for panoramic imaging |
JP2007259477A (en) | 2002-09-30 | 2007-10-04 | Toshiba Corp | Image composition method, apparatus, and program |
US20080034306A1 (en) | 2006-08-04 | 2008-02-07 | Bas Ording | Motion picture preview icons |
US20080050035A1 (en) * | 2006-08-28 | 2008-02-28 | Shingo Tsurumi | Information Processing Apparatus, Imaging Apparatus, Information Processing System, Device Control Method and Program |
WO2008035146A1 (en) | 2006-09-18 | 2008-03-27 | Sony Ericsson Mobile Communications Ab | Video pattern thumbnails and method |
US20080159708A1 (en) | 2006-12-27 | 2008-07-03 | Kabushiki Kaisha Toshiba | Video Contents Display Apparatus, Video Contents Display Method, and Program Therefor |
JP2008167155A (en) | 2006-12-28 | 2008-07-17 | Sony Corp | Reproducing device and reproduction method, and program |
US20080192125A1 (en) * | 2007-02-14 | 2008-08-14 | Samsung Electronics Co., Ltd. | Panoramic photography method and apparatus |
EP1970908A2 (en) | 2007-03-15 | 2008-09-17 | Sony Corporation | Information processing apparatus, imaging apparatus, image display control method and computer program |
JP2008245233A (en) | 2007-03-01 | 2008-10-09 | Nippon Hoso Kyokai <Nhk> | Shielding control apparatus, object shielding control system, and shielding control program |
US20090208062A1 (en) * | 2008-02-20 | 2009-08-20 | Samsung Electronics Co., Ltd. | Method and a handheld device for capturing motion |
US7602417B2 (en) * | 2005-08-05 | 2009-10-13 | Canon Kabushiki Kaisha | Image processing method, imaging apparatus, and storage medium storing control program of image processing method executable by computer |
US20090290013A1 (en) * | 2008-05-20 | 2009-11-26 | Sony Corporation | Image pickup apparatus, image pickup method, playback control apparatus, playback control method, and program |
US20100272369A1 (en) * | 2006-02-24 | 2010-10-28 | Morpho, Inc. | Image processing apparatus |
US20100302280A1 (en) * | 2009-06-02 | 2010-12-02 | Microsoft Corporation | Rendering aligned perspective images |
US20110057948A1 (en) * | 2009-09-04 | 2011-03-10 | Sony Corporation | Method and apparatus for image alignment |
US20110058014A1 (en) * | 2009-09-10 | 2011-03-10 | Noriyuki Yamashita | Image processing device, image processing method, and program |
US20110058749A1 (en) * | 2009-09-04 | 2011-03-10 | Sony Corporation | Method and apparatus for determining the mis-alignment in images |
US20110149016A1 (en) * | 2009-12-21 | 2011-06-23 | Sony Corporation | Image processing apparatus and method, and program |
US20120133639A1 (en) * | 2010-11-30 | 2012-05-31 | Microsoft Corporation | Strip panorama |
US8345963B2 (en) * | 2003-02-26 | 2013-01-01 | Facebook, Inc. | System for image analysis in a network that is structured with multiple layers and differentially weighted neurons |
US8345921B1 (en) * | 2009-03-10 | 2013-01-01 | Google Inc. | Object detection with false positive filtering |
US20130106990A1 (en) * | 2011-11-01 | 2013-05-02 | Microsoft Corporation | Planar panorama imagery generation |
US20130108105A1 (en) * | 2011-10-31 | 2013-05-02 | Electronics And Telecommunications Research Institute | Apparatus and method for masking privacy region based on monitored video image |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2003271933A (en) | 2002-03-18 | 2003-09-26 | Sony Corp | Face detector, face detecting method, and robot device |
JP4329398B2 (en) | 2002-05-10 | 2009-09-09 | ソニー株式会社 | Face detection apparatus and method, program, and recording medium |
JP2004133637A (en) | 2002-10-09 | 2004-04-30 | Sony Corp | Face detector, face detection method and program, and robot apparatus |
JP5161523B2 (en) | 2007-03-30 | 2013-03-13 | ヤマハ発動機株式会社 | Saddle type vehicle, power unit and continuously variable transmission |
-
2008
- 2008-10-27 JP JP2008275143A patent/JP4623201B2/en not_active Expired - Fee Related
-
2009
- 2009-09-23 US US12/565,499 patent/US9106872B2/en active Active
- 2009-09-30 EP EP20090171855 patent/EP2180698B1/en not_active Not-in-force
- 2009-10-23 BR BRPI0904357-8A patent/BRPI0904357A2/en not_active IP Right Cessation
- 2009-10-26 RU RU2009139617/08A patent/RU2009139617A/en not_active Application Discontinuation
- 2009-10-27 CN CN200910206897A patent/CN101729784A/en active Pending
Patent Citations (52)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5818975A (en) * | 1996-10-28 | 1998-10-06 | Eastman Kodak Company | Method and apparatus for area selective exposure adjustment |
US20010005208A1 (en) * | 1997-06-20 | 2001-06-28 | Nippon Telegraph And Telephone Corporation | Scheme for interactive video manipulation and display of moving object on background image |
US20020136590A1 (en) * | 1997-07-21 | 2002-09-26 | Barry Himmel | Personalized event/theme photograph album |
US6009190A (en) * | 1997-08-01 | 1999-12-28 | Microsoft Corporation | Texture map construction method and apparatus for displaying panoramic image mosaics |
US6078701A (en) * | 1997-08-01 | 2000-06-20 | Sarnoff Corporation | Method and apparatus for performing local to global multiframe alignment to construct mosaic images |
US6393162B1 (en) * | 1998-01-09 | 2002-05-21 | Olympus Optical Co., Ltd. | Image synthesizing apparatus |
US6246413B1 (en) * | 1998-08-17 | 2001-06-12 | Mgi Software Corporation | Method and system for creating panoramas |
US20020135672A1 (en) * | 1999-01-08 | 2002-09-26 | Sezan M. Ibrahim | System for constructing mosaic images |
US6891561B1 (en) * | 1999-03-31 | 2005-05-10 | Vulcan Patents Llc | Providing visual context for a mobile active visual display of a panoramic region |
US20020122113A1 (en) * | 1999-08-09 | 2002-09-05 | Foote Jonathan T. | Method and system for compensating for parallax in multiple camera systems |
US6677981B1 (en) * | 1999-12-31 | 2004-01-13 | Stmicroelectronics, Inc. | Motion play-back of still pictures comprising a panoramic view for simulating perspective |
US20020196340A1 (en) * | 2001-04-24 | 2002-12-26 | Matsushita Electric Industrial Co., Ltd. | Image synthesis display method and apparatus for vehicle camera |
US20050129272A1 (en) * | 2001-11-30 | 2005-06-16 | Frank Rottman | Video monitoring system with object masking |
US7825950B2 (en) * | 2001-11-30 | 2010-11-02 | Robert Bosch Gmbh | Video monitoring system with object masking |
JP2003304444A (en) | 2002-04-08 | 2003-10-24 | Inc Corp | Apparatus and method for compositing image |
US20030234866A1 (en) * | 2002-06-21 | 2003-12-25 | Ross Cutler | System and method for camera color calibration and image stitching |
US7027054B1 (en) * | 2002-08-14 | 2006-04-11 | Avaworks, Incorporated | Do-it-yourself photo realistic talking head creation system and method |
US7227893B1 (en) * | 2002-08-22 | 2007-06-05 | Xlabs Holdings, Llc | Application-specific object-based segmentation and recognition system |
JP2007259477A (en) | 2002-09-30 | 2007-10-04 | Toshiba Corp | Image composition method, apparatus, and program |
US20070104383A1 (en) * | 2002-11-14 | 2007-05-10 | Microsoft Corporation | Stabilization of objects within a video sequence |
US8345963B2 (en) * | 2003-02-26 | 2013-01-01 | Facebook, Inc. | System for image analysis in a network that is structured with multiple layers and differentially weighted neurons |
US20040189691A1 (en) * | 2003-03-28 | 2004-09-30 | Nebojsa Jojic | User interface for adaptive video fast forward |
WO2005024723A1 (en) | 2003-09-08 | 2005-03-17 | Nec Corporation | Image combining system, image combining method, and program |
US20070008499A1 (en) * | 2003-09-08 | 2007-01-11 | Akihiko Iketani | Image combining system, image combining method, and program |
US20050058360A1 (en) * | 2003-09-12 | 2005-03-17 | Thomas Berkey | Imaging system and method for displaying and/or recording undistorted wide-angle image data |
US20050099494A1 (en) * | 2003-11-10 | 2005-05-12 | Yining Deng | Digital camera with panoramic image capture |
WO2006005231A1 (en) * | 2004-07-14 | 2006-01-19 | Kun Ma | A portable panoramic photographic apparatus |
US7602417B2 (en) * | 2005-08-05 | 2009-10-13 | Canon Kabushiki Kaisha | Image processing method, imaging apparatus, and storage medium storing control program of image processing method executable by computer |
US20070182818A1 (en) * | 2005-09-02 | 2007-08-09 | Buehler Christopher J | Object tracking and alerts |
US20070064974A1 (en) * | 2005-09-19 | 2007-03-22 | International Business Machines Corporation | Video data management |
US20070162873A1 (en) * | 2006-01-10 | 2007-07-12 | Nokia Corporation | Apparatus, method and computer program product for generating a thumbnail representation of a video sequence |
US20100272369A1 (en) * | 2006-02-24 | 2010-10-28 | Morpho, Inc. | Image processing apparatus |
US20070206945A1 (en) * | 2006-03-01 | 2007-09-06 | Delorme David M | Method and apparatus for panoramic imaging |
US20080034306A1 (en) | 2006-08-04 | 2008-02-07 | Bas Ording | Motion picture preview icons |
US20080050035A1 (en) * | 2006-08-28 | 2008-02-28 | Shingo Tsurumi | Information Processing Apparatus, Imaging Apparatus, Information Processing System, Device Control Method and Program |
WO2008035146A1 (en) | 2006-09-18 | 2008-03-27 | Sony Ericsson Mobile Communications Ab | Video pattern thumbnails and method |
US20080159708A1 (en) | 2006-12-27 | 2008-07-03 | Kabushiki Kaisha Toshiba | Video Contents Display Apparatus, Video Contents Display Method, and Program Therefor |
JP2008167155A (en) | 2006-12-28 | 2008-07-17 | Sony Corp | Reproducing device and reproduction method, and program |
US20080192125A1 (en) * | 2007-02-14 | 2008-08-14 | Samsung Electronics Co., Ltd. | Panoramic photography method and apparatus |
JP2008245233A (en) | 2007-03-01 | 2008-10-09 | Nippon Hoso Kyokai <Nhk> | Shielding control apparatus, object shielding control system, and shielding control program |
EP1970908A2 (en) | 2007-03-15 | 2008-09-17 | Sony Corporation | Information processing apparatus, imaging apparatus, image display control method and computer program |
US20090208062A1 (en) * | 2008-02-20 | 2009-08-20 | Samsung Electronics Co., Ltd. | Method and a handheld device for capturing motion |
US20090290013A1 (en) * | 2008-05-20 | 2009-11-26 | Sony Corporation | Image pickup apparatus, image pickup method, playback control apparatus, playback control method, and program |
US8345921B1 (en) * | 2009-03-10 | 2013-01-01 | Google Inc. | Object detection with false positive filtering |
US20100302280A1 (en) * | 2009-06-02 | 2010-12-02 | Microsoft Corporation | Rendering aligned perspective images |
US20110057948A1 (en) * | 2009-09-04 | 2011-03-10 | Sony Corporation | Method and apparatus for image alignment |
US20110058749A1 (en) * | 2009-09-04 | 2011-03-10 | Sony Corporation | Method and apparatus for determining the mis-alignment in images |
US20110058014A1 (en) * | 2009-09-10 | 2011-03-10 | Noriyuki Yamashita | Image processing device, image processing method, and program |
US20110149016A1 (en) * | 2009-12-21 | 2011-06-23 | Sony Corporation | Image processing apparatus and method, and program |
US20120133639A1 (en) * | 2010-11-30 | 2012-05-31 | Microsoft Corporation | Strip panorama |
US20130108105A1 (en) * | 2011-10-31 | 2013-05-02 | Electronics And Telecommunications Research Institute | Apparatus and method for masking privacy region based on monitored video image |
US20130106990A1 (en) * | 2011-11-01 | 2013-05-02 | Microsoft Corporation | Planar panorama imagery generation |
Non-Patent Citations (11)
Title |
---|
Communication pursuant to Article 94(3) EPC issued Mar. 23, 2012, in European Patent Application No. 09 171 855.1. |
U.S. Appl. No. 12/447,923, filed Apr. 30, 2009, Tsurumi. |
U.S. Appl. No. 12/524,383, filed Jul. 24, 2009, Wang, et al. |
U.S. Appl. No. 12/524,734, filed Jul. 28, 2009, Tsurumi. |
U.S. Appl. No. 12/524,757, filed Jul. 28, 2009, Tsurumi. |
U.S. Appl. No. 12/524,777, filed Jul. 28, 2009, Fukumoto, et al. |
U.S. Appl. No. 12/524,887, filed Jul. 29, 2009, Tsurumi. |
U.S. Appl. No. 12/526,058, filed Aug. 6, 2009, Tsurumi. |
U.S. Appl. No. 12/558,262, filed Sep. 11, 2009, Onoe, et al. |
U.S. Appl. No. 13/486,209, filed Jun. 1, 2012, Tsurumi. |
Yukinobu Taniguchi, et al., "PanoramaExcerpts: Video Cataloging by Automatic Synthesis and Layout of Panoramic Images", IEICE Transactions on Information and Systems, Information & Systems Society, vol. E83-D. No. 12, XP009093074, Dec. 1, 2000, pp. 2039-2046. |
Cited By (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
USD868093S1 (en) | 2014-04-22 | 2019-11-26 | Google Llc | Display screen with graphical user interface or portion thereof |
USD934281S1 (en) | 2014-04-22 | 2021-10-26 | Google Llc | Display screen with graphical user interface or portion thereof |
USD994696S1 (en) | 2014-04-22 | 2023-08-08 | Google Llc | Display screen with graphical user interface or portion thereof |
USD1006046S1 (en) | 2014-04-22 | 2023-11-28 | Google Llc | Display screen with graphical user interface or portion thereof |
USD933691S1 (en) | 2014-04-22 | 2021-10-19 | Google Llc | Display screen with graphical user interface or portion thereof |
US11860923B2 (en) | 2014-04-22 | 2024-01-02 | Google Llc | Providing a thumbnail image that follows a main image |
USD1008302S1 (en) | 2014-04-22 | 2023-12-19 | Google Llc | Display screen with graphical user interface or portion thereof |
US9972121B2 (en) * | 2014-04-22 | 2018-05-15 | Google Llc | Selecting time-distributed panoramic images for display |
US20150302633A1 (en) * | 2014-04-22 | 2015-10-22 | Google Inc. | Selecting time-distributed panoramic images for display |
USD830407S1 (en) | 2014-04-22 | 2018-10-09 | Google Llc | Display screen with graphical user interface or portion thereof |
USD830399S1 (en) | 2014-04-22 | 2018-10-09 | Google Llc | Display screen with graphical user interface or portion thereof |
USD835147S1 (en) | 2014-04-22 | 2018-12-04 | Google Llc | Display screen with graphical user interface or portion thereof |
USD877765S1 (en) | 2014-04-22 | 2020-03-10 | Google Llc | Display screen with graphical user interface or portion thereof |
US10540804B2 (en) | 2014-04-22 | 2020-01-21 | Google Llc | Selecting time-distributed panoramic images for display |
US11163813B2 (en) | 2014-04-22 | 2021-11-02 | Google Llc | Providing a thumbnail image that follows a main image |
US9934222B2 (en) | 2014-04-22 | 2018-04-03 | Google Llc | Providing a thumbnail image that follows a main image |
USD868092S1 (en) | 2014-04-22 | 2019-11-26 | Google Llc | Display screen with graphical user interface or portion thereof |
US10431258B2 (en) | 2015-10-22 | 2019-10-01 | Gopro, Inc. | Apparatus and methods for embedding metadata into video stream |
US9892760B1 (en) | 2015-10-22 | 2018-02-13 | Gopro, Inc. | Apparatus and methods for embedding metadata into video stream |
US9667859B1 (en) * | 2015-12-28 | 2017-05-30 | Gopro, Inc. | Systems and methods for determining preferences for capture settings of an image capturing device |
US10194073B1 (en) | 2015-12-28 | 2019-01-29 | Gopro, Inc. | Systems and methods for determining preferences for capture settings of an image capturing device |
US10958837B2 (en) | 2015-12-28 | 2021-03-23 | Gopro, Inc. | Systems and methods for determining preferences for capture settings of an image capturing device |
US10469748B2 (en) | 2015-12-28 | 2019-11-05 | Gopro, Inc. | Systems and methods for determining preferences for capture settings of an image capturing device |
US10678844B2 (en) | 2016-01-19 | 2020-06-09 | Gopro, Inc. | Storage of metadata and images |
US9922387B1 (en) | 2016-01-19 | 2018-03-20 | Gopro, Inc. | Storage of metadata and images |
US10469739B2 (en) | 2016-01-22 | 2019-11-05 | Gopro, Inc. | Systems and methods for determining preferences for capture settings of an image capturing device |
US9967457B1 (en) | 2016-01-22 | 2018-05-08 | Gopro, Inc. | Systems and methods for determining preferences for capture settings of an image capturing device |
US11640169B2 (en) | 2016-02-16 | 2023-05-02 | Gopro, Inc. | Systems and methods for determining preferences for control settings of unmanned aerial vehicles |
US10599145B2 (en) | 2016-02-16 | 2020-03-24 | Gopro, Inc. | Systems and methods for determining preferences for flight control settings of an unmanned aerial vehicle |
US12105509B2 (en) | 2016-02-16 | 2024-10-01 | Gopro, Inc. | Systems and methods for determining preferences for flight control settings of an unmanned aerial vehicle |
US9836054B1 (en) | 2016-02-16 | 2017-12-05 | Gopro, Inc. | Systems and methods for determining preferences for flight control settings of an unmanned aerial vehicle |
US9973792B1 (en) | 2016-10-27 | 2018-05-15 | Gopro, Inc. | Systems and methods for presenting visual information during presentation of a video segment |
US10187607B1 (en) | 2017-04-04 | 2019-01-22 | Gopro, Inc. | Systems and methods for using a variable capture frame rate for video capture |
US20210134049A1 (en) * | 2017-08-08 | 2021-05-06 | Sony Corporation | Image processing apparatus and method |
Also Published As
Publication number | Publication date |
---|---|
EP2180698B1 (en) | 2013-02-27 |
EP2180698A1 (en) | 2010-04-28 |
JP4623201B2 (en) | 2011-02-02 |
CN101729784A (en) | 2010-06-09 |
BRPI0904357A2 (en) | 2011-02-01 |
JP2010103878A (en) | 2010-05-06 |
RU2009139617A (en) | 2011-05-10 |
US20100104217A1 (en) | 2010-04-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9106872B2 (en) | Image processing apparatus, image processing method, and program | |
US8368815B2 (en) | Image processing apparatus, image processing method, and program | |
US8488904B2 (en) | Image processing apparatus, image processing method, and program | |
US8963951B2 (en) | Image processing apparatus, moving-image playing apparatus, and processing method and program therefor to allow browsing of a sequence of images | |
US8768097B2 (en) | Image processing apparatus, moving image reproducing apparatus, and processing method and program therefor | |
US8754959B2 (en) | Image processing device, dynamic image reproduction device, and processing method and program in them | |
US8810708B2 (en) | Image processing apparatus, dynamic picture reproduction apparatus, and processing method and program for the same | |
US8350929B2 (en) | Image pickup apparatus, controlling method and program for the same | |
JP4697221B2 (en) | Image processing apparatus, moving image reproduction apparatus, processing method and program therefor | |
KR20130120175A (en) | Apparatus, method and computer readable recording medium for generating a caricature automatically | |
JP2012105205A (en) | Key frame extractor, key frame extraction program, key frame extraction method, imaging apparatus, and server device | |
JP3836814B2 (en) | Image processing method, image processing apparatus, and image processing program | |
CN101617531A (en) | Image processing apparatus, moving image playing device and processing method and program | |
Silva et al. | Fast-Forward Methods for Egocentric Videos: A Review |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SONY CORPORATION,JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TSURUMI, SHINGO;REEL/FRAME:023293/0045 Effective date: 20090911 Owner name: SONY CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TSURUMI, SHINGO;REEL/FRAME:023293/0045 Effective date: 20090911 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |