WO2009081732A1 - 画像処理装置、動画再生装置、これらにおける処理方法およびプログラム - Google Patents
画像処理装置、動画再生装置、これらにおける処理方法およびプログラム Download PDFInfo
- Publication number
- WO2009081732A1 WO2009081732A1 PCT/JP2008/072402 JP2008072402W WO2009081732A1 WO 2009081732 A1 WO2009081732 A1 WO 2009081732A1 JP 2008072402 W JP2008072402 W JP 2008072402W WO 2009081732 A1 WO2009081732 A1 WO 2009081732A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- captured
- captured image
- history
- unit
- Prior art date
Links
- 238000003672 processing method Methods 0.000 title claims description 8
- 239000002131 composite material Substances 0.000 claims abstract description 226
- 238000006243 chemical reaction Methods 0.000 claims abstract description 197
- 238000012937 correction Methods 0.000 claims abstract description 84
- 238000000605 extraction Methods 0.000 claims abstract description 71
- 238000003384 imaging method Methods 0.000 claims abstract description 42
- 239000000203 mixture Substances 0.000 claims abstract description 38
- 239000000284 extract Substances 0.000 claims abstract description 23
- 238000012545 processing Methods 0.000 claims description 138
- 238000000034 method Methods 0.000 claims description 137
- 238000003860 storage Methods 0.000 claims description 84
- 230000008859 change Effects 0.000 claims description 79
- 230000033001 locomotion Effects 0.000 claims description 73
- 238000013519 translation Methods 0.000 claims description 38
- 230000009467 reduction Effects 0.000 claims description 28
- 230000015572 biosynthetic process Effects 0.000 claims description 13
- 230000002194 synthesizing effect Effects 0.000 claims description 13
- 238000003786 synthesis reaction Methods 0.000 claims description 11
- PXFBZOLANLWPMH-UHFFFAOYSA-N 16-Epiaffinine Natural products C1C(C2=CC=CC=C2N2)=C2C(=O)CC2C(=CC)CN(C)C1C2CO PXFBZOLANLWPMH-UHFFFAOYSA-N 0.000 description 313
- 230000009466 transformation Effects 0.000 description 295
- 238000004364 calculation method Methods 0.000 description 107
- 230000003287 optical effect Effects 0.000 description 94
- 239000011159 matrix material Substances 0.000 description 93
- 238000010586 diagram Methods 0.000 description 86
- 230000008569 process Effects 0.000 description 67
- 239000013598 vector Substances 0.000 description 50
- 238000001514 detection method Methods 0.000 description 27
- 230000007704 transition Effects 0.000 description 16
- 238000004091 panning Methods 0.000 description 13
- 238000001914 filtration Methods 0.000 description 8
- 230000000694 effects Effects 0.000 description 6
- 238000012546 transfer Methods 0.000 description 6
- 241000270295 Serpentes Species 0.000 description 4
- 238000009826 distribution Methods 0.000 description 4
- 238000004458 analytical method Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 230000010354 integration Effects 0.000 description 3
- 230000004044 response Effects 0.000 description 3
- 125000002066 L-histidyl group Chemical group [H]N1C([H])=NC(C([H])([H])[C@](C(=O)[*])([H])N([H])[H])=C1[H] 0.000 description 2
- 230000005540 biological transmission Effects 0.000 description 2
- 230000010339 dilation Effects 0.000 description 2
- 230000008707 rearrangement Effects 0.000 description 2
- 230000000717 retained effect Effects 0.000 description 2
- 230000001131 transforming effect Effects 0.000 description 2
- 244000025254 Cannabis sativa Species 0.000 description 1
- 241001465754 Metazoa Species 0.000 description 1
- 241000238370 Sepia Species 0.000 description 1
- 206010047571 Visual impairment Diseases 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 210000001217 buttock Anatomy 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 238000003708 edge detection Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000001172 regenerating effect Effects 0.000 description 1
- 238000011426 transformation method Methods 0.000 description 1
- 238000010626 work up procedure Methods 0.000 description 1
- 230000037303 wrinkles Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/16—Spatio-temporal transformations, e.g. video cubism
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/02—Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
- G11B27/031—Electronic editing of digitised analogue information signals, e.g. audio or video signals
- G11B27/034—Electronic editing of digitised analogue information signals, e.g. audio or video signals on discs
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/19—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
- G11B27/28—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
- G11B27/32—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on separate auxiliary tracks of the same or an auxiliary record carrier
- G11B27/322—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on separate auxiliary tracks of the same or an auxiliary record carrier used signal is digitally coded
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/698—Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/95—Computational photography systems, e.g. light-field imaging systems
- H04N23/951—Computational photography systems, e.g. light-field imaging systems by using two or more images to influence resolution, frame rate or aspect ratio
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/14—Picture signal circuitry for video frequency region
- H04N5/144—Movement detection
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/78—Television signal recording using magnetic recording
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/91—Television signal processing therefor
Definitions
- the present invention relates to an image processing apparatus, and more particularly, to an image processing apparatus capable of reproducing a moving image, a moving image reproducing apparatus, a processing method in these, and a program for causing a computer to execute the method.
- the moving image shot in this way can be reproduced on the display using a moving image reproducing device at home, for example.
- a moving image reproducing device at home For example, when browsing a moving image shot mainly about his / her child, a moving image mainly including his / her child is reproduced.
- the playback time is long and the same target video is being browsed continuously, the viewer's interest in the video being played may decrease as the playback time elapses. Therefore, in order to increase the interest of the viewer, it may be possible to display other images related to the currently displayed image.
- an image display method has been proposed in which a video index (still image) is scroll-displayed along with the moving image along with the moving image (see, for example, Japanese Patent Application Laid-Open No. 11-289517 (FIG. 7)).
- past, present, and future still images based on moving images are displayed as video indexes
- past, present, and future still images can be viewed together with the currently displayed video. Can do. For this reason, for example, when viewing a video taken at a kindergarten event where a child attends a school, even if his child is displayed as the current video, as a still or past still image, There are cases where the scenery of the event related to the current video is displayed. In this case, since the scene of the event can be seen while watching the state of the child, it becomes easy to grasp the state of the event, and the viewer's interest can be increased.
- each image composing the moving image is converted based on the motion information of the imaging device so that the state around the central person can be properly grasped while browsing the central person being photographed. It is conceivable that the converted images are reproduced while being sequentially synthesized.
- the photographer quickly moves the imaging apparatus in the left-right direction, there is a possibility that appropriate motion information cannot be detected.
- the photographer stops the recording operation of the captured image and then resumes the recording operation with a different scene as the subject the appropriate motion information can be detected before and after the scene change. It may not be possible.
- the present invention has an object of easily grasping the contents of a moving image when browsing the moving image taken by the imaging device.
- a first aspect of the present invention is a moving image storage unit that stores a captured moving image captured by an imaging device, and a first imaging that constitutes the captured moving image.
- Conversion information storage means for storing, for each second captured image, conversion information for converting a second captured image positioned after the first captured image on the time axis of the captured moving image on the time axis of the image
- An image holding unit that holds a history image including each captured image located before the second captured image on the time axis, and a feature amount extracting unit that extracts a feature amount related to the second captured image
- Image conversion means for converting the second captured image based on the conversion information
- determination means for determining whether or not the second captured image and the history image need to be combined based on the extracted feature amount; ,Up If the determination means determines that the second captured image and the history image are to be combined, the converted second captured image is combined with the history image held in the image holding means to create a new one.
- the determination unit determines that the second captured image and the history image are not combined, the history image stored in the image holding unit is added to the history image.
- the feature amount related to the second captured image is extracted, the second captured image is converted based on the conversion information, and the necessity of combining the second captured image and the history image is determined based on the feature amount.
- the converted second captured image is combined with the history image to obtain a new history image, and the second captured image and the history image are combined.
- the second captured image is held in the image holding unit without combining the second captured image with the history image.
- the image synthesizing unit holds the history held in the image holding unit when the determining unit determines not to synthesize the second captured image and the history image.
- the image may be erased and the second captured image may be held in the image holding unit as a new history image.
- the history image is erased, and the second captured image is held as a new history image in the image holding unit.
- the image synthesizing unit is arranged to arrange the history image in the image holding unit when the determining unit determines not to synthesize the second captured image and the history image.
- the second captured image may be arranged at a position different from that in the image holding unit and held as a new history image.
- the second captured image is arranged at a position different from the arrangement position of the history image to obtain a new history image. Bring.
- the feature amount extraction unit extracts the feature amount by detecting a constant change in each image constituting the captured moving image, and the determination unit includes the extracted feature.
- the determination unit includes the extracted feature.
- the conversion information includes elements relating to enlargement / reduction, translational movement, and rotation
- the feature amount extraction means is a translation component based on the conversion information related to the second captured image.
- the rotation component and the enlargement / reduction component are extracted, and the determination means compares at least one of the extracted translation component, the rotation component, and the enlargement / reduction component with a predetermined threshold value to compare the synthesized component.
- the necessity may be determined.
- the translation component, the rotation component, and the enlargement / reduction component are extracted based on the conversion information related to the second captured image, and at least one of the extracted translation component, rotation component, and enlargement / reduction component and a predetermined threshold value are extracted. Is brought about to determine the necessity of synthesis.
- the feature amount extraction unit relates to each captured image from a captured image located a predetermined number of times before the second captured image on the time axis to the second captured image. Based on the conversion information, the translation component, the rotation component, and the enlargement / reduction component are extracted, and the average values of the extracted translation component, rotation component, and enlargement / reduction component are respectively calculated. The necessity of the synthesis may be determined by comparing at least one of average values of the translation component, the rotation component, and the enlargement / reduction component with a predetermined threshold value.
- the translation component, the rotation component, and the enlargement / reduction component are calculated based on the conversion information related to each captured image from the captured image located at a predetermined number of times before the second captured image on the time axis to the second captured image. And calculating average values of the extracted translation component, rotation component, and enlargement / reduction component, respectively, and calculating at least one of the calculated average values of the translation component, rotation component, and enlargement / reduction component, and a predetermined threshold value.
- the feature amount extraction unit relates to each captured image from a captured image located a predetermined number of times before the second captured image on the time axis to the second captured image. Based on the conversion information, the translation component, the rotation component, and the enlargement / reduction component are extracted, and the integrated values of the extracted translation component, rotation component, and enlargement / reduction component are respectively calculated. The necessity of the synthesis may be determined by comparing at least one integrated value of the translation component, the rotation component, and the enlargement / reduction component with a predetermined threshold value.
- the translation component, the rotation component, and the enlargement / reduction component are calculated based on the conversion information relating to each captured image from the captured image located before the second captured image to the second captured image on the inter-axis. Extracting, calculating an integrated value of the extracted translation component, rotation component, and enlargement / reduction component, respectively, and calculating at least one of the calculated integration values of the translation component, rotation component, and enlargement / reduction component, and a predetermined threshold value. By comparing the above, it is possible to determine the necessity of synthesis.
- a second aspect of the present invention provides a moving image storage unit that stores a captured moving image captured by the imaging apparatus, and the time axis of the captured moving image based on the first captured image that constitutes the captured moving image.
- Conversion information storage means for storing, for each second captured image, conversion information for converting a second captured image located after one captured image, and more than the second captured image on the time axis.
- Image holding means for holding a history image including each captured image located in front, and synthesis of the second captured image and the history image based on an elapsed time since the history image was held in the image holding means
- Determining means for determining necessity of the image, image converting means for converting the second captured image based on the conversion information, and determining that the second captured image and the history image are synthesized by the determining means.
- the converted second captured image is combined with the history image held in the image holding means and is held in the image holding means as a new history image.
- the second captured image is combined with the history image held in the image holding unit without combining the second captured image with the second captured image.
- An image processing apparatus, an image processing method, and a program for causing a computer to execute the image processing apparatus comprising:
- the converted second captured image is combined with the history image to obtain a new history image, and the second captured image and the history image are combined.
- the second captured image is held in the image holding unit without combining the second captured image with the history image.
- a moving image storage unit for storing a captured moving image captured by the imaging device, and the time axis of the captured moving image based on the first captured image constituting the captured moving image.
- Conversion information storage means for storing, for each second captured image, conversion information for converting a second captured image located after one captured image, and more than the second captured image on the time axis.
- An image holding unit that holds a history image including each captured image located in front, a feature amount extracting unit that extracts a feature amount related to the second captured image, and the conversion information based on the extracted feature amount Conversion information correcting means for correcting, image converting means for converting the second captured image based on the corrected conversion information, and the history image held in the image holding means
- the feature amount related to the second captured image is extracted, the conversion information is corrected based on the extracted feature amount, the second captured image is converted based on the corrected conversion information, and the history image The second captured image after the conversion is combined and the image holding means holds it as a new history image.
- the conversion information correction unit determines whether or not the conversion information needs to be corrected by comparing the extracted feature amount with a predetermined threshold, and corrects the conversion information. When it is determined that the conversion information is necessary, the conversion information is corrected, and when the conversion information is corrected, the image conversion unit converts the second captured image based on the corrected conversion information. If the conversion information is not corrected, the second captured image may be converted based on the conversion information. Thus, by comparing the extracted feature amount with a predetermined threshold value, it is determined whether or not the conversion information needs to be corrected. If it is determined that the conversion information needs to be corrected, the conversion information is corrected and converted. When the information is corrected, the second captured image is converted based on the corrected conversion information, and when the conversion information is not corrected, the second captured image is converted based on the conversion information. Bring about an effect.
- the conversion information correction means may correct the conversion information to predetermined conversion information when it is determined that the conversion information needs to be corrected. Thereby, when it is determined that conversion information needs to be corrected, the conversion information is corrected to predetermined conversion information.
- the conversion information correction unit relates to a captured image located in the forward direction of the second captured image on the time axis when it is determined that the conversion information needs to be corrected. You may make it correct
- a feature amount extracting means for extracting a feature amount relating to a captured image that constitutes a captured moving image captured by an imaging device, and a composite image division based on the extracted feature amount.
- the captured image is combined with the composite image that has already been formed based on the motion information of the imaging device at the time of imaging.
- an image composition means for creating a new composite image different from the composite image already formed based on the motion information when it is determined by the determination means to divide the composite image.
- the feature amount related to the captured image constituting the captured moving image captured by the imaging device is extracted, the necessity of dividing the composite image is determined based on the extracted feature amount, and the composite image is determined not to be divided. If it is determined that the captured image is synthesized with the synthesized image already formed based on the motion information and the synthesized image is divided, the synthesized image already formed based on the motion information Has the effect of creating different new composite images.
- a moving image storage unit for storing a captured moving image captured by the imaging apparatus, and the time axis of the captured moving image with respect to the first captured image constituting the captured moving image.
- Conversion information storage means for storing, for each second captured image, conversion information for converting a second captured image located after one captured image, and more than the second captured image on the time axis.
- An image holding unit that holds a history image including each captured image located in front, a feature amount extracting unit that extracts a feature amount related to the second captured image, and the second captured image based on the conversion information
- An image converting means for converting; a determining means for determining whether or not to synthesize the second captured image and the history image based on the extracted feature quantity; and the second captured image and the above by the determining means.
- Image synthesizing means for holding the two captured images in the image holding means, and display control means for sequentially displaying the new history image held in the image holding means or the second captured image on the display means. And a processing method thereof and a program for causing a computer to execute the method.
- the feature amount related to the second captured image is extracted, the second captured image is converted based on the conversion information, and the necessity of combining the second captured image and the history image is determined based on the feature amount.
- the converted second captured image is combined with the history image to obtain a new history image, and the second captured image and the history image are combined. If it is determined that the second captured image is not combined with the history image, the second captured image is held in the image holding unit without combining the second captured image with the history image, and the new history image held in the image holding unit or the second The captured images are sequentially displayed.
- the present invention when viewing a moving image taken by an imaging device, it is possible to achieve an excellent effect that the content of the moving image can be easily grasped.
- FIG. 1 is a block diagram illustrating a functional configuration example of an image processing apparatus 100 according to an embodiment of the present invention.
- FIG. 2 is a block diagram illustrating a functional configuration example of the camera work detection unit 120 according to the embodiment of the present invention.
- FIG. 3 is a block diagram illustrating a functional configuration example of the feature amount extraction unit 150 according to the embodiment of the present invention.
- FIG. 4 is a diagram schematically showing a case where images corresponding to the respective frames constituting the moving image are arranged in time series.
- FIG. 5 is a diagram schematically showing each file recorded in the moving image storage unit 200 and the metadata storage unit 210 in the embodiment of the present invention.
- FIG. 6 is a diagram illustrating an example of an image corresponding to a frame constituting a moving image.
- FIG. 1 is a block diagram illustrating a functional configuration example of an image processing apparatus 100 according to an embodiment of the present invention.
- FIG. 2 is a block diagram illustrating a functional configuration example of the camera work detection unit 120 according
- FIG. 7 is a diagram showing an image simplified by omitting the background and the like for the image corresponding to the frame constituting the moving image.
- FIG. 8 is a diagram illustrating a case where a rectangle 340 is transformed using predetermined affine transformation parameters.
- FIG. 9 is a flowchart showing a processing procedure of affine transformation parameter detection processing by the image processing apparatus 100 according to the embodiment of the present invention.
- FIG. 10 is a diagram illustrating an example of transition of a moving image captured by a camera.
- FIG. 11 is a diagram illustrating an example of an optical flow detected in each image illustrated in FIG. 10 while an image corresponding to the immediately preceding frame is indicated by a broken line.
- FIG. 12 is a diagram illustrating a display example when a moving image including the images 401 to 403 illustrated in FIG. 10 is reproduced.
- FIG. 13 is a diagram illustrating an example of transition of a moving image captured by a camera.
- FIG. 14 is a diagram showing an example of an optical flow detected in each image shown in FIG. 13 while an image corresponding to the immediately preceding frame is indicated by a broken line.
- FIG. 15 is a diagram illustrating a display example when a moving image including the images 421 to 423 illustrated in FIG. 13 is reproduced.
- FIG. 16 is a diagram illustrating an example of transition of a moving image captured by the camera.
- FIG. 17 is a diagram illustrating an example of an optical flow detected while the image corresponding to the immediately preceding frame is indicated by a broken line in each image illustrated in FIG. 16.
- FIG. 18 is a diagram illustrating a display example when a moving image including the images 441 to 443 illustrated in FIG. 16 is reproduced.
- FIG. 19 is a diagram schematically showing the relationship between each frame of the moving image file stored in the moving image storage unit 200 and the display area in the embodiment of the present invention.
- FIG. 20 is a diagram schematically showing the flow of each frame of the moving image file stored in the moving image storage unit 200 according to the embodiment of the present invention.
- FIG. 21 is a diagram showing a display example (image 480) in the case of reproducing a moving image shot by a camera and an image 484 in a state before the current image 482 in the image 480 is affine transformed.
- FIG. 22 shows the image 485 when the image region surrounded by the frame 483 shown in FIG. 21 is enlarged and the current image after the affine transformation is saved in the display memory 240 in a state where it is saved in the image memory 220.
- FIG. FIG. 23 is a diagram illustrating an example of transition of a moving image shot by a camera.
- FIG. 24 is a diagram illustrating an example of transition of a moving image shot by a camera.
- FIG. 25 is a diagram illustrating an example of transition of a moving image shot by a camera.
- FIG. 26 is a diagram illustrating an example of transition of a moving image shot by a camera.
- FIG. 27 is a diagram schematically illustrating a case where the images 501 to 517 constituting the moving image file 500 are combined.
- FIG. 28 is a diagram schematically showing the relationship between each frame constituting the moving image and the scene change score calculated for these frames.
- FIG. 29 is a diagram schematically illustrating a case where images 551 to 567 constituting the moving image 530 are combined.
- FIG. 30 is a diagram schematically illustrating a case where images 551 to 567 constituting the moving image 530 are combined.
- FIG. 31 is a diagram illustrating an example of a composite image formed by a moving image shot by a camera.
- FIG. 32 is a diagram illustrating an example of a composite image formed by a moving image shot by a camera.
- FIG. 33 is a flowchart showing a processing procedure of moving image reproduction processing by the image processing apparatus 100 according to the embodiment of the present invention.
- FIG. 34 is a flowchart showing the conversion information correction processing procedure in the processing procedure of the moving image reproduction processing by the image processing apparatus 100 according to the embodiment of the present invention.
- FIG. 35 is a flowchart illustrating a composite image division determination processing procedure in a processing procedure of a moving image reproduction process performed by the image processing apparatus 100 according to the embodiment of the present invention.
- FIG. 36 is a diagram showing a configuration example of the multi-core processor 800 in the embodiment of the present invention.
- FIG. 37 is a diagram showing a configuration example of the control processor core 801 in the embodiment of the present invention.
- FIG. 38 is a diagram showing a configuration example of the arithmetic processor core (# 1) 811 in the embodiment of the present invention.
- FIG. 39 is a diagram schematically showing a calculation method of the multi-core processor 800 in the embodiment of the present invention.
- FIG. 40 is a diagram schematically showing the flow of programs and data in the case where operations are performed by the multi-core processor 800 in the embodiment of the present invention.
- FIG. 41 is a diagram schematically illustrating an outline of an arithmetic method in which processing for a plurality of data is performed by each instruction, and an outline of SIMD arithmetic in which processing for a plurality of data is performed by one instruction.
- FIG. 42 is a diagram showing a configuration example of a program executed by the control processor core 801 or the arithmetic processor core (# 1) 811 in the embodiment of the present invention.
- FIG. 43 is a diagram schematically showing a data structure and a processing flow when filtering processing is performed on the image data stored in the main memory 781 using the Sobel filter 830 according to the embodiment of the present invention.
- FIG. 44 is a diagram schematically showing a data flow when SIMD calculation is performed using the Sobel filter 830 on the image data stored in the main memory 781 according to the embodiment of the present invention.
- FIG. 45 schematically shows a vector creation method for creating nine vectors from image data stored in the first buffer 831 when filtering is performed using the Sobel filter 830 according to the embodiment of the present invention.
- FIG. 46 is a diagram schematically showing a vector operation method for performing vector operation using SIMD instructions for the vector data 841 to 849 when performing filtering using the Sobel filter 830 according to the embodiment of the present invention.
- FIG. 47 is a diagram schematically showing the flow of camera work parameter calculation processing in the embodiment of the present invention in time series.
- FIG. 48 is a diagram schematically showing an internal configuration of a Blu-ray player 890 capable of reproducing the Blu-ray disc 880 and each data 881 to 884 recorded on the Blu-ray disc 880 as an example of a recording medium. It is.
- FIG. 1 is a block diagram illustrating a functional configuration example of an image processing apparatus 100 according to an embodiment of the present invention.
- the image processing apparatus 100 includes a moving image input unit 110, a camera work detection unit 120, a recording control unit 130, a file acquisition unit 140, a feature amount extraction unit 150, a conversion information correction unit 160, and a composite image division determination unit. 170, an image conversion unit 180, an image composition unit 190, a moving image storage unit 200, a metadata storage unit 210, an image memory 220, a display area extraction unit 230, a display memory 240, and a display control unit 250. And a display unit 260 and an operation receiving unit 270.
- the image processing apparatus 100 is a personal computer that can extract a feature amount by video analysis from a moving image shot by an imaging device such as a digital video camera and perform various image processing using the extracted feature amount. It can be realized by a computer.
- the moving image input unit 110 is a moving image input unit that inputs a moving image captured by an imaging device such as a digital video camera (hereinafter simply referred to as “camera”).
- the input moving image is input to the camera work detection unit 120. Output.
- the camera work detection unit 120 analyzes the moving image output from the moving image input unit 110 to detect camera movement information (camera work) at the time of shooting, and is calculated based on the camera movement information.
- the affine transformation parameters (camera work parameters) are output to the recording control unit 130. That is, the camera work detection unit 120 extracts a feature point from each image constituting the moving image, extracts an optical flow (motion vector) for the feature point, and analyzes the optical flow for the extracted feature point. A feature point that shows dominant motion is selected, and camera motion is estimated based on the optical flow for the feature point showing dominant motion.
- the dominant movement means a regular movement indicated by a relatively large number of optical flows among the optical flows for a plurality of feature points.
- the recording control unit 130 records the moving image output from the moving image input unit 110 as a moving image file in the moving image storage unit 200, and associates the affine transformation parameters output from the camera work detection unit 120 with the corresponding moving image and frame. This is recorded in the metadata storage unit 210 as a metadata file.
- the moving image storage unit 200 stores the moving image output from the moving image input unit 110 under the control of the recording control unit 130 as a moving image file.
- the moving image storage unit 200 supplies a moving image file to the file acquisition unit 140 in response to a request from the file acquisition unit 140.
- the moving image file stored in the moving image storage unit 200 will be described in detail with reference to FIG.
- the metadata storage unit 210 stores the affine transformation parameters output from the camera work detection unit 120 based on the control of the recording control unit 130 as a metadata file. Further, the metadata storage unit 210 supplies a metadata file to the file acquisition unit 140 in response to a request from the file acquisition unit 140.
- the metadata file stored in the metadata storage unit 210 will be described in detail with reference to FIG.
- the file acquisition unit 140 acquires at least one of the files stored in the moving image storage unit 200 and the metadata storage unit 210 according to the operation input received by the operation reception unit 270, and acquires each file This information is supplied to each part. Specifically, the file acquisition unit 140, when an instruction operation for synthesizing and reproducing a moving image is received by the operation receiving unit 270, the moving image file stored in the instructed moving image storage unit 200 and the moving image file And a metadata file stored in the metadata storage unit 210 in association with the image data, and outputs the moving image of the acquired moving image file and the affine transformation parameters of the metadata file to the feature amount extraction unit 150 and the image conversion unit 180 At the same time, the affine transformation parameters are output to the transformation information correction unit 160.
- the feature amount extraction unit 150 extracts a feature amount based on each image constituting the moving image output from the file acquisition unit 140 and an affine transformation parameter associated with each image.
- the data is output to the conversion information correction unit 160 and the composite image division determination unit 170.
- the feature quantity extracted by the feature quantity extraction unit 150 is the scene change score in each image constituting the moving image and each component of the camera work (X direction and Y direction) extracted from the affine transformation parameters. (Zoom component, translation component, rotation component), an average value of each of these components for a predetermined number of frames, and an integrated value from the first frame to the current frame of each of these components will be described as an example.
- the scene change score is a value calculated based on a color histogram in each image constituting the moving image, and is a value used for detecting a scene change point.
- the scene change point is, for example, a position indicating the boundary point when the shooting scene changes momentarily in the moving image, and an image immediately after the boundary point is set as an image corresponding to the scene change point.
- the feature quantity extraction unit 150 will be described in detail with reference to FIG. Each component of camera work will be described in detail with reference to FIGS. 6 and 7. Further, the scene change point will be described in detail with reference to FIG.
- the conversion information correction unit 160 corrects the affine transformation parameter output from the file acquisition unit 140 based on the feature amount output from the feature amount extraction unit 150, and converts the corrected affine transformation parameter into the image conversion unit. Output to 180. Specifically, the conversion information correction unit 160 compares the feature amount output from the feature amount extraction unit 150 with a threshold value, thereby determining whether or not the affine transformation parameter output from the file acquisition unit 140 needs to be corrected. If it is determined that correction of the affine transformation parameter is necessary, the affine transformation parameter is corrected, and the corrected affine transformation parameter is output to the image conversion unit 180.
- the affine transformation parameter is not corrected and nothing is output to the image conversion unit 180.
- the correction of the affine transformation parameters the correction using the affine transformation parameter matrix as a unit matrix and the affine transformation parameters associated with the frame existing in the front-rear direction of the current frame are used.
- a linear interpolation matrix will be calculated and correction will be described by way of example in which the calculated linear interpolation matrix is replaced with an affine transformation parameter matrix. The correction of the affine transformation parameters will be described in detail with reference to FIG.
- the composite image division determination unit 170 determines whether or not an image corresponding to the current frame is to be divided and reproduced based on the feature amount output from the feature amount extraction unit 150 from the currently formed composite image. When it is determined that the image corresponding to the current frame is divided and reproduced, the fact that the image corresponding to the current frame is divided and reproduced is output to the image conversion unit 180 and the image composition unit 190, and An instruction to delete the camerawork component calculated up to the immediately preceding frame is output to the feature amount extraction unit 150.
- the composite image division determination unit 170 measures the playback time from the first frame to the current frame when the synthesis playback of the moving image is started, and based on this playback time, the image corresponding to the current frame is It is determined whether to divide and reproduce the formed composite image.
- the composite image division determining unit 170 measures the reproduction time from the first frame to the current frame at the time of division, and the reproduction time is Based on this, it is determined whether or not the image corresponding to the current frame is divided and reproduced from the currently formed composite image. Note that image division will be described in detail with reference to FIGS.
- the image conversion unit 180 performs affine transformation for each frame using the affine transformation parameters corresponding to the image of the moving image of the moving image file output from the file acquisition unit 140, and converts the affine transformed image into an image. This is output to the combining unit 190.
- the image conversion unit 180 performs affine transformation for each frame using the corrected affine transformation parameter.
- the composite image division determination unit 170 outputs that the image corresponding to the current frame is divided and reproduced
- the image conversion unit 180 performs affine transformation on the image corresponding to the current frame using a unit matrix. To the image composition unit 190. Note that these image conversions will be described in detail with reference to FIGS.
- the image composition unit 190 synthesizes the image by overwriting the composite image corresponding to each of the previous frames held in the image memory 220 with the image after the affine transformation by the image conversion unit 180, and creates a new composite image.
- a simple composite image is stored in the image memory 220.
- the image composition unit 190 outputs the frames corresponding to the previous frames held in the image memory 220. The corresponding composite image is deleted, and the image corresponding to the current frame is held in the image memory 220.
- the images are synthesized by overwriting the synthesized image held in the image memory 220 with the image after the affine transformation by the image transformation unit 180.
- the composite image division determination unit 170 outputs that the image synthesis unit 190 divides and reproduces the image corresponding to the current frame in response to the operation input received by the operation reception unit 270
- the image corresponding to the current frame is stored in the image memory 220 at a position different from the position where the composite image is retained, with the composite image corresponding to each of the previous frames held in the image memory 220 being left. Can be retained.
- Image composition in the case where the image corresponding to the current frame is divided and reproduced will be described in detail with reference to FIGS.
- the image composition unit 190 based on the position of the current image in the display area output from the display region extraction unit 230, a composite image in which the current image after the affine transformation by the image conversion unit 180 is held in the display memory 240. Synthesize by overwriting to.
- the size of the current image synthesized in the display memory 240 is determined according to the value of the display magnification.
- the composition of the current image in the display memory 240 will be described in detail with reference to FIG.
- the image composition unit 190 compresses the image after the affine transformation by the image conversion unit 180 and overwrites the composite image held in the image memory 220 and overwrites the composite image held in the display memory 240.
- the captured current image is a higher-resolution captured image than the uncompressed image or the compressed history image.
- the history image when the composite image is output can be a compressed image, and the current image can be a non-compressed image or a captured image with a higher resolution than the compressed history image.
- the image memory 220 is a work buffer that holds the synthesized image synthesized by the image synthesizing unit 190, and supplies the held synthesized image to the image synthesizing unit 190 or the display area extracting unit 230. That is, the image memory 220 is an image memory that holds each image constituting a moving image as a history image.
- the display area extraction unit 230 extracts an image existing within the display area, which is a display target area, from the composite image held in the image memory 220, and the extracted image is displayed in the display memory. 240.
- the extraction of an image included in the display area will be described in detail with reference to FIGS. 19 and 20, and the calculation of the position of the current image in the display area will be described with reference to FIG. Will be described in detail.
- the display memory 240 is a display buffer that holds an image extracted from the image memory 220 by the display area extraction unit 230, and the stored image is displayed on the display unit 260.
- the display control unit 250 causes the display unit 260 to sequentially display the composite image held in the display memory 240 for each frame.
- the display unit 260 displays the composite image held in the display memory 240 based on the control of the display control unit 250.
- it can be realized by a display of a personal computer or a television. Note that a composite image display example will be described in detail with reference to FIGS.
- the operation receiving unit 270 includes a keyboard and a mouse (pointing device) including various input keys.
- a mouse pointing device
- the contents of the received operation input are transferred to the file acquisition unit 140, the image composition unit 190, or Are output to the display area extraction unit 230.
- the operation receiving unit 270 and the display unit 260 may be integrally configured as a touch panel.
- FIG. 2 is a block diagram illustrating a functional configuration example of the camera work detection unit 120 according to the embodiment of the present invention.
- the camera work detection unit 120 includes a feature point extraction unit 121, an optical flow calculation unit 122, and a camera work parameter calculation unit 123.
- the feature point extraction unit 121 extracts feature points from an image corresponding to a frame constituting the moving image output from the moving image input unit 110, and outputs the extracted feature points to the optical flow calculation unit 122.
- the feature point extraction unit 121 extracts the feature point from the entire image for the top frame of the frames constituting the video output from the video input unit 110, and for the frames other than the top, A feature point is extracted from a newly photographed area compared with an image corresponding to a frame.
- a point having a strong edge gradient in the vertical direction or the horizontal direction generally referred to as “corner point”.
- corner point can be extracted as the feature point.
- This corner point is a feature point that is strong in optical flow calculation, and can be obtained using edge detection.
- This corner point extraction will be described in detail with reference to FIGS.
- the feature point extraction unit 121 extracts feature points from the entire image for the first frame, and extracts feature points from a newly captured region portion compared to the previous image for frames other than the first frame. Although extraction is performed, feature points may be extracted from the entire image for each frame other than the top in accordance with the processing capability and the like.
- the optical flow calculation unit 122 calculates an optical flow for each feature point output from the feature point extraction unit 121, and outputs the optical flow obtained by the calculation to the camera work parameter calculation unit 123. Specifically, by comparing each image corresponding to two consecutive frames (the current frame and the immediately preceding frame) constituting the moving image output from the moving image input unit 110, the image corresponding to the immediately preceding frame is compared. The optical flow corresponding to each feature point is obtained as the optical flow of the current frame. The optical flow is obtained for each frame constituting the moving image. As a detection method for detecting the optical flow, a detection method such as a gradient method or a block matching method can be used. The calculation of the optical flow will be described in detail with reference to FIG. 6 and FIG.
- the camera work parameter calculation unit 123 performs a camera work parameter calculation process for calculating a camera work parameter by using an optical flow corresponding to each feature point output from the optical flow calculation unit 122.
- the work parameter is output to the recording control unit 130.
- each image constituting the moving image to be reproduced is converted and displayed in accordance with the movement of the camera.
- camera motion is extracted using the optical flow calculated by the optical flow calculation unit 122, and camera work parameters (conversion parameters) are calculated based on the extracted motion.
- affine transformation is used as an image conversion method for converting an image constituting a moving image to be reproduced will be described.
- an affine transformation parameter used as transformation information is not an affine matrix that represents the movement of feature points between successive images, but one of the successive images is used as a reference image. In this case, it is defined as an affine transformation parameter corresponding to an affine matrix indicating where the next image of the reference image moves.
- an affine transformation parameter is used as a camera work parameter
- other image transformation methods such as projective transformation may be used.
- the affine transformation parameters can be obtained by calculation using a three-point vector.
- the projective transformation parameter can be obtained by calculation using a vector of four points.
- the camera work parameter is conversion information for converting another captured image on the basis of at least one captured image constituting the captured moving image, and is described in at least the camera coordinate system. It includes position information and posture information. That is, the camera work parameter includes information on the position and orientation of the camera when the photographer is photographing. Further, based on the affine transformation parameters obtained by the camera work parameter calculation unit 123, it is possible to estimate the movement of the camera by the photographer's operation such as zooming in, zooming out, panning, tilting, rotation, and the like. The calculation of affine transformation parameters will be described in detail with reference to FIGS.
- FIG. 3 is a block diagram illustrating a functional configuration example of the feature amount extraction unit 150 according to the embodiment of the present invention.
- the feature amount extraction unit 150 includes a color histogram extraction unit 151, a continuous frame difference value calculation unit 152, a scene change score calculation unit 153, a camera work component calculation unit 155, a camera work component holding unit 156, and a camera work.
- a component average value calculation unit 157 and a camera work component integrated value calculation unit 158 are provided.
- the color histogram extraction unit 151 extracts, for each frame, a color histogram of the entire screen in this image from the image constituting the moving image output from the file acquisition unit 140. This is output to the value calculation unit 152.
- the difference value calculation unit 152 between consecutive frames calculates the difference value of the color distribution of the color histogram between two consecutive frames for the color histogram for each frame output from the color histogram extraction unit 151.
- the difference value is output to the scene change score calculation unit 153.
- a value obtained by digitizing the similarity calculated using EMD is used as the difference value of the color distribution of the color histogram between two consecutive frames.
- EMD represents the distance between color distributions, and is used for similar image retrieval or the like (see, for example, JP-A-2007-206919).
- the scene change score calculation unit 153 applies a high-pass filter to the difference value of the color distribution of the color histogram output from the continuous inter-frame difference value calculation unit 152, thereby calculating the difference of the offset value of the difference value between frames.
- the value after removal is performed and the value after removal of the DC component is calculated as a scene change score, and the calculated scene change score is output to the conversion information correction unit 160 and the composite image division determination unit 170.
- the difference value between two consecutive frames is calculated using the similarity calculated by EMD, and the scene change score is calculated based on the difference value.
- Another difference value between the frames constituting the moving image on the time axis may be calculated, and the scene change score may be calculated based on the difference value.
- the camerawork component calculation unit 155 calculates each component of camerawork (translational components in the x and y directions, zoom components, and rotation components) from the affine transformation parameters corresponding to the current frame output from the file acquisition unit 140.
- Each component of the calculated camera work is output to the conversion information correction unit 160, the camera work component average value calculation unit 157, and the camera work component integrated value calculation unit 158, and also to the camera work component holding unit 156. Hold.
- the camera work component holding unit 156 holds each component of the camera work up to the current frame calculated by the camera work component calculation unit 155 for each frame, and each component of the held camera work is a camera work component.
- the average value calculation unit 157 and the camera work component integrated value calculation unit 158 are supplied.
- the camera work component holding unit 156 displays each component of the held camera work. to erase.
- the camerawork component average value calculation unit 157 is a predetermined number of frames before the camerawork component corresponding to the current frame output from the camerawork component calculation unit 155 and the previous frame held in the camerawork component holding unit 156.
- the average value of the camera work components is calculated based on each component of the camera work corresponding to each frame up to this frame, and the calculated average value of the camera work components is output to the composite image division determination unit 170. To do.
- the number of frames used when calculating the average value of the camera work components can be set to about 5, for example.
- the camera work component integrated value calculation unit 158 includes each component of the camera work corresponding to the current frame output from the camera work component calculation unit 155, and the frame immediately before the first frame held in the camera work component holding unit 156.
- the integrated value of the camera work component is calculated based on each component of the camera work corresponding to each of the above frames, and the calculated integrated value of the camera work component is output to the composite image division determination unit 170.
- each component of the camera work held in the camera work component holding unit 156 is displayed. Erased.
- the camerawork component integrated value calculation unit 158 corresponds to the current frame output from the camerawork component calculation unit 155 when calculating the integrated value of the camerawork component for the next frame or later after the instruction is given.
- FIG. 4 is a diagram schematically showing a case where images corresponding to each frame constituting the moving image are arranged in time series.
- images 351 to 358 corresponding to the respective frames constituting the moving image 350 photographed by the imaging device are schematically shown. It should be noted that the images 351 to 358 are shown by simplifying the number, subject, and the like for the sake of explanation.
- images 351 to 358 are images included in one moving image 350, but include three scenes having different shooting times and places, so frames in which these scenes are switched. Then, the composition and color of the subject change instantaneously.
- the scene change point is between the two consecutive frames. Determined.
- the determination of the scene change point is performed by the conversion information correction unit 160 and the composite image division determination unit 170.
- the threshold used when determining the scene change point a value common to the conversion information correction unit 160 and the composite image division determination unit 170 may be used, or a different value may be used.
- the threshold value of the composite image division determination unit 170 can be set higher than the threshold value of the conversion information correction unit 160.
- the images 351 to 353 are taken of a person walking in the city, and since the person is moving, the position of the person is slightly different, but the entire image has a similar impression. For this reason, the histogram difference value between two consecutive frames related to the images 351 to 353 does not exceed the threshold value, and it is determined that the two consecutive frames are not scene change points.
- the images 354 to 356 are images of the houses in front of the mountains, and the images are taken by moving the imaging device in the horizontal direction. The overall impression is similar. However, when the image 353 and the image 354 are compared with each other, the entire image is greatly different because it corresponds to a scene switching portion at which the time and place of shooting are different. For this reason, the boundary between the image 353 and the image 354 is determined as a scene change point. Similarly, the boundary between the image 356 and the image 357 is determined to be a scene change point. As described above, when it is determined that the scene change point is set, in the embodiment of the present invention, an image immediately after the boundary point is set as an image corresponding to the scene change point. Images corresponding to the scene change points are, for example, images 354 and 357.
- FIG. 5 is a diagram schematically showing each file recorded in the moving image storage unit 200 and the metadata storage unit 210 in the embodiment of the present invention.
- 5A the moving image files 201 to 204 stored in the moving image storage unit 200 and the metadata files 211 to 213 stored in the metadata storage unit 210 in association with the moving image files 201 to 204 are included.
- a moving image ID that is identification information for identifying each moving image file stored in the moving image storage unit 200 is given to each moving image file. For example, “# 1” is assigned to the moving image file 201, “# 2” is assigned to the moving image file 202, and “#n” is assigned to the moving image file 204.
- FIG. 5B schematically shows a moving image file 201 stored in the moving image storage unit 200 and a metadata file 211 stored in the metadata storage unit 210 in association with the moving image file 201.
- the moving image file 201 is a moving image file composed of n frames, and these n frames are shown as frames “1” 205 to “n” 208.
- a moving image ID 214 In the metadata file 211, a moving image ID 214, a frame number 215, and an affine transformation parameter 216 are stored in association with each other.
- the moving image ID 214 is a moving image ID assigned to the corresponding moving image file. For example, “# 1” assigned to the moving image file 201 is stored.
- the frame number 215 is a serial number of each frame constituting the moving image of the corresponding moving image file. For example, “1” to “n” corresponding to the frames “1” 205 to “n” 208 constituting the moving image of the moving image file 201. Is stored.
- the affine transformation parameter 216 is an affine transformation parameter calculated for each frame of the moving image corresponding to the frame number 215.
- the affine transformation parameter 216 “a1, b1, c1, d1, e1, f1” corresponding to “1” of the frame number 215 is an affine transformation parameter of the unit matrix.
- “am, bm, cm, dm, em, fm” of the affine transformation parameter 216 corresponding to “m (m is an integer of 2 or more)” of the frame number 215 is the frame “m ⁇ ” immediately before the frame “m”. 1 ”is an affine transformation parameter.
- FIGS. 7B and 7C are diagrams showing a simplified image of the image 300 shown in FIG.
- the images 300, 320, and 330 shown in FIGS. 6 and 7 include horse images 301, 321, and 331 on which a person is straddling, and a snake image installed in front of the horse images 301, 321, and 331. 302, 322, and 332 are included. Further, as shown in FIG. 6, there are flags, chairs, etc. in the background of these images, and these flags are fluttering in the wind.
- the image 320 shown in (a) of FIG. 7 is a frame immediately before the frame corresponding to the images 300 and 330 shown in (a) to (c) of FIG. 6 and (b) and (c) of FIG. It is the image which simplified the corresponding image. Also, images 320 and 330 corresponding to two consecutive frames are images showing transitions when the subject in the screen gradually increases. That is, at the time of this photographing, a zoom-in operation that is an operation for gradually increasing the subject in the screen is performed.
- a method of detecting a feature point from an image constituting a moving image and calculating an affine transformation parameter using an optical flow corresponding to the feature point will be described as an example.
- a case where a corner point is used as a feature point will be described.
- FIGS. 7A to 7C a method for calculating affine transformation parameters using an optical flow corresponding to three corner points detected from the images 320 and 330 will be described as an example.
- a corner point 323 near the mouth in the horse image 321, a corner point 324 near the human buttocks in the horse image 321, and a snake image It is assumed that a corner point 325 near the mouth of 322 is detected.
- optical flows 337, 338, and 339 for the corner points 323, 324, and 325 in the image 320 are detected by the gradient method, the block matching method, or the like in the image 330 shown in FIG. 7B. Then, based on the detected optical flows 337, 338, and 339, corner points 333, 334, and 335 corresponding to the corner points 323, 324, and 325 in the image 320 are detected.
- the horse images 321 and 331 and the snake images 322 and 332 included in the images 320 and 330 shown in FIGS. 7A and 7B are installed on the ground. It does not move independently of camera movement. For this reason, the motion of the camera can be accurately estimated based on the optical flow obtained for the corner points detected for the horse images 321 and 331 and the snake images 322 and 332. For example, as shown in FIG. 7C, based on the three optical flows 337 to 339 detected in the image 330, the image 330 is obtained by enlarging the image 320 around the point 336. Can be estimated.
- the camera movement at the time of capturing the image 330 is a zoom-in operation centered on the point 336.
- a corner point is detected for an object that does not move independently of the camera movement, and the camera movement with a certain regularity is accurately determined based on the optical flow obtained for the corner point. Can be detected. Therefore, the affine transformation parameters can be calculated and obtained using the optical flow obtained for these corner points.
- an object that moves independently of the camera movement such as a flag waving in the wind
- the image 300 illustrated in FIG. 6 includes a flag fluttering in the wind.
- the optical flow detected in the image 300 shown in FIG. 6B is indicated by an arrow, and the corner point detected by the optical flow is indicated by a white circle at the tip of the arrow.
- the corner points 303 to 305 are corner points corresponding to the corner points 333 to 335 shown in (b) and (c) of FIG.
- Corner points 306 to 311 are corner points detected for the flag present in the background of the horse image 301. Since these flags flutter in the wind, the movement of the flag due to the influence of the wind is detected as an optical flow. That is, each optical flow corresponding to the corner points 306 to 311 is detected for a flag that moves independently of the movement of the camera. Therefore, when the optical flow corresponding to at least one of the corner points 306 to 311 is included in the three optical flows used when calculating the affine transformation parameters, the accurate camera Unable to detect movement. In this case, accurate affine transformation parameters cannot be calculated.
- the relationship between the optical flow (the optical flows corresponding to the corner points 306 to 311 shown in FIG. 6B) with respect to an object that moves independently of the camera motion and the camera motion.
- an optical flow having a certain regularity an optical flow other than the optical flows corresponding to the corner points 306 to 311 shown in FIG. 6B is detected from the captured image.
- the affine transformation parameter calculation process for calculating the affine transformation parameters based on the three optical flows is performed a plurality of times, a plurality of affine transformation parameters are obtained, and the plurality of affine transformation parameters are calculated.
- An example of selecting an optimum affine transformation parameter from among them will be described. In this example, it is assumed that the size of the moving object included in each image constituting the moving image is relatively small with respect to the area of the image.
- the affine transformation will be briefly described. In two dimensions, when the position of the movement source is (x, y) and the position of the movement destination after the affine transformation is (x ′, y ′), the determinant of the affine transformation is expressed by Equation 1. Can do.
- a to f are affine transformation parameters.
- the affine matrix AM based on this affine transformation parameter can be expressed by the following equation.
- the X direction zoom component XZ, the Y direction zoom component YZ, the X direction translation component XT, the Y direction translation component YT, the X direction rotation component ⁇ x, and the X direction rotation component ⁇ y are respectively expressed by the following equations. Can be obtained.
- each component of the camerawork (X-direction and Y-direction zoom components, translation components, and rotation components) can be obtained from the affine transformation parameters.
- the rotation component will be described with reference to FIG.
- FIG. 8 is a diagram illustrating a case where the rectangle 340 is converted using predetermined affine transformation parameters.
- a predetermined affine transformation parameter is set in an xy coordinate having an apex of a rectangle 340 as an origin and two sides of the rectangle 340 in contact with the apex corresponding to the origin as an x-axis and a y-axis.
- the rectangle after the rectangle 340 is converted is used as a rectangle 341.
- an angle formed by the sides on the x-axis side of the rectangles 340 and 341 is a rotation component ⁇ x
- an angle formed by the sides on the y-axis side of the rectangles 340 and 341 is a rotation component ⁇ y.
- the average value of camerawork components is a value obtained by dividing a value obtained by adding each component of camerawork corresponding to a predetermined number of frames by a predetermined number.
- the predetermined number is 5, and the translational components in the X direction of the camera work components of the frames 1 to 5 to be calculated are XT1 to 5 respectively. In this case, it can be obtained by “(XT1 + XT2 + XT3 + XT4 + XT5) / 5”.
- an average speed obtained by dividing the value obtained by adding each component of camera work by the elapsed time between frames to be calculated may be used. .
- the integrated value of the camera work component is a value obtained by adding each component of the camera work corresponding to each frame from the first frame to the current frame.
- the translational component in the X direction of the camera work component of each of the frames 1 to 5 from the first frame to the current frame is set to XT1 to 5, respectively. In this case, it can be obtained by “XT1 + XT2 + XT3 + XT4 + XT5”.
- three feature points are selected from the feature points where the optical flow has been detected. For example, three corner points are selected at random from the corner points (indicated by white circles) detected in the image 300 shown in FIG.
- projective transformation parameters are used as camera work parameters, four feature points are selected at random.
- the affine transformation parameters are calculated using the three optical flows corresponding to the three selected feature points. For example, an optical flow corresponding to three corner points selected from the corner points (indicated by white circles) in the image 300 shown in FIG. 6B (with arrows connected to the white circles). Is used to calculate the affine transformation parameters.
- This affine transformation parameter can be obtained using Equation 1.
- the score of the affine transformation parameter is calculated based on the obtained affine transformation parameter. Specifically, using the obtained affine transformation parameters, the positions of the movement destinations of all feature points in the image corresponding to the frame immediately before the current frame are obtained. Then, the position of the feature point obtained using the affine transformation parameter is compared with the position of the feature point detected in the current frame, and the difference value between the positions of the two feature points corresponding to each other is determined for each feature point. Is calculated. As the difference value, for example, an absolute distance between the positions of two feature points corresponding to each other is calculated.
- the calculated difference value and a preset threshold value are compared for each feature point, and the number of feature points having the difference value smaller than the threshold value is obtained as a score of the affine transformation parameter.
- three feature points are randomly selected from the feature points in which the optical flow is detected, and a process for calculating the score of the affine transformation parameter based on the optical flow corresponding to these feature points is performed a predetermined number of times.
- a plurality of affine transformation parameter scores are calculated repeatedly. This predetermined number of times may be set as appropriate according to the type of image to be compared, the processing capability of the image processing apparatus 100, or a fixed value. For example, the predetermined number of times can be set to about 20 in consideration of the processing capability of the image processing apparatus 100.
- a case is considered where three corner points including at least one of the corner points 306 to 311 are selected from the corner points detected in the image 300 shown in FIG.
- the three optical flows have a certain regularity. Since an optical flow that is not provided is included, an affine transformation parameter that does not transform an image corresponding to the immediately preceding frame according to a certain rule is obtained. For this reason, the difference value calculated
- the affine transformation parameter having the largest score value is selected as the representative affine transformation parameter from the obtained scores of the plurality of affine transformation parameters.
- an inverse matrix for the selected matrix of representative affine transformation parameters is calculated, and the affine transformation parameters of the inverse matrix are recorded in the metadata storage unit 210 in association with the current frame.
- each image constituting the moving image includes a moving object (animal body) such as a person or a car
- the size of the moving object relative to the area of the image is relatively small.
- the camera motion can be extracted without being affected by the moving object.
- at least translation, rotation, and zoom are used. One conversion can be performed in the form of a rectangular image.
- FIG. 9 is a flowchart showing a processing procedure of affine transformation parameter detection processing by the image processing apparatus 100 according to the embodiment of the present invention.
- a moving image file is input to the moving image input unit 110 (step S900). Subsequently, the moving image file input to the moving image input unit 110 is decoded, and an image of one frame is acquired in time-series order (step S901). Subsequently, it is determined whether or not the acquired one frame is the first frame of the moving image file input to the moving image input unit 110 (step S902). When the acquired one frame is the top frame (step S902), feature points are extracted from the entire image corresponding to the top frame (step S903). For example, as shown in FIG. 6B, a plurality of corner points are extracted from the image. Subsequently, an affine transformation parameter of a unit matrix is selected as an affine transformation parameter (step S904), and the process proceeds to step S914.
- a feature point is extracted from a region newly taken with reference to an image corresponding to the immediately preceding frame (step S905). That is, since the feature points already extracted in the image corresponding to the immediately preceding frame can be obtained by the optical flow corresponding to the feature points, they are not extracted in the image corresponding to the current frame.
- step S906 an optical flow for each feature point extracted from the image corresponding to the immediately preceding frame is calculated. That is, as shown in FIG. 6B, an optical flow for each corner point is calculated.
- variable i is initialized to “1” (step S907).
- M feature points are selected from the feature points from which the optical flow has been detected (step S908). For example, when affine transformation parameters are used as camera work parameters, three feature points are selected at random. When projective transformation parameters are used as camera work parameters, four feature points are selected at random. Subsequently, affine transformation parameters are calculated based on the M optical flows calculated corresponding to the selected M feature points (step S909).
- the score of the affine transformation parameter is calculated based on the affine transformation parameter obtained by calculation (step S910). Specifically, using the affine transformation parameters obtained by calculation, the movement destination positions of all feature points in the image corresponding to the immediately preceding frame are obtained. Then, the position of the feature point obtained using this affine transformation parameter is compared with the position of the feature point in the image corresponding to the current frame obtained when the optical flow is calculated in step S906 to correspond to each other. A difference value between the positions of the two feature points is calculated for each feature point. As the difference value, for example, an absolute distance between two positions corresponding to each other is calculated. Subsequently, the calculated difference value and a preset threshold value are compared for each feature point, and the number of feature points having the difference value smaller than the threshold value is obtained as a score of the affine transformation parameter.
- step S911 “1” is added to the variable i (step S911), and it is determined whether or not the variable i is larger than the constant N (step S912). If the variable i is equal to or less than the constant N (step S912), the process returns to step S908, and the affine transformation parameter score calculation process is repeated (steps S908 to S910).
- 20 can be used as the constant N.
- step S912 when the variable i is larger than the constant N (step S912), the affine transformation parameter having the largest score value is selected as the representative affine transformation parameter among the obtained scores of the affine transformation parameters (step S912). S913). Subsequently, an affine transformation parameter of an inverse matrix for the selected representative affine transformation parameter matrix is recorded in the metadata storage unit 210 in association with the current frame (step S914). If the current frame is the top frame, the affine transformation parameters of the selected unit matrix are recorded in the metadata storage unit 210 in association with the top frame. Subsequently, the image corresponding to the current frame and the feature points in this image are overwritten and saved (step S915).
- step S916 it is determined whether or not the current frame is the last frame of the moving image file input to the moving image input unit 110 (step S916). If the current frame is not the last frame (step S916), the process returns to step S901 to repeat the affine transformation parameter detection process (steps S901 to S915). On the other hand, if the current frame is the last frame (step S916), the affine transformation parameter detection process is terminated.
- the example of detecting the affine transformation parameter based on the optical flow detected in the image constituting the moving image as the detection of the camera work parameter has been described, but a sensor such as an acceleration sensor or a gyro sensor, A zoom button used when performing a zoom operation may be provided on the camera, and the movement amount of the camera at the time of shooting may be detected by the sensor or the zoom button, and the camera work parameter may be obtained based on the movement amount of the camera. .
- the movement amount of the camera detected at the time of shooting can be used when determining whether the camera work parameter obtained by the camera work parameter calculation unit 123 is correct.
- a plurality of camera work parameters are detected by the camera work parameter calculation unit 123, and one camera work parameter is selected from the plurality of camera work parameters based on the movement amount of the camera detected at the time of shooting. You may make it do.
- FIG. 10 is a diagram showing an example of transition of a moving image taken by a camera.
- FIG. 10 is a diagram illustrating images 401 to 403 corresponding to continuous frames included in a moving image when a person 400 is photographed against a mountain background. This example shows a case where the photographer is shooting while moving the direction of the lens of the camera to the right and above. In this case, the person 400 included in the moving image photographed by the camera moves from the right side to the left side and moves downward in the image constituting the moving image.
- FIG. 11 is a diagram showing an example of an optical flow detected in each image shown in FIG. 10 while an image corresponding to the immediately preceding frame is indicated by a broken line.
- An image 401 shown in (a) of FIG. 11 is the same as the image 401 shown in (a) of FIG.
- the solid line portion in the image 402 shown in FIG. 11B is the same as the image 402 shown in FIG. 10B, and the broken line in the image 402 shown in FIG. This part is the same as the solid line part of the image 401 shown in FIG.
- arrows 404 to 406 in the image 402 shown in FIG. 11B indicate an example of the optical flow detected from the image 402.
- FIG. 11C is the same as the image 403 shown in FIG. 10C.
- the broken line portion is the same as the solid line portion of the image 402 shown in FIG.
- arrows 407 to 409 in the image 403 shown in FIG. 11C indicate an example of the optical flow detected from the image 403.
- the person 400 and the background mountain included in the image move in accordance with the movement of the camera. Based on the optical flow detected by this movement, affine transformation parameters can be obtained for each frame.
- FIG. 12 is a diagram showing an image composition example in the case where the moving image including the images 401 to 403 shown in FIG. 10 is synthesized and reproduced.
- the images constituting the two moving images are combined, so that the image displayed on the display unit 260 becomes larger than the normal image as the playback time elapses. For this reason, the first displayed image is displayed with a size relatively smaller than the size of the display area of the display unit 260. Note that the user may specify the size, position, and the like of the first displayed image.
- the image 401 corresponding to the first frame is displayed at first.
- the matrix of affine transformation parameters (3 ⁇ 3 matrix) corresponding to the image 401 is A1
- the value of A1 is obtained, and the position and size of the image 401 of the first frame is used as a reference
- the image 401 is affine transformed with the obtained matrix of A1.
- A is a unit matrix, the position and size of the image 401 are not converted.
- the image 402 corresponding to the next frame is displayed, the image 402 is affine transformed using the affine transformation parameters associated with the frame.
- the matrix of the affine transformation parameters corresponding to the image 402 is A2 and the matrix of the affine transformation parameters corresponding to the image 401 is A1
- the value of A1 ⁇ A2 is obtained, and the image of the first frame
- the image 402 is affine-transformed using the obtained A1 ⁇ A2 matrix with reference to the position and size of 401.
- the image shown in FIG. 12B only the position of the image 402 is converted.
- the image 402 that has been affine transformed with the affine transformation parameters is overwritten so as to overlap the image 401 corresponding to the immediately preceding frame. That is, in the area of the image 401, the area 410 that overlaps the image 402 is overwritten with the image of the image 402.
- the image 401 is synthesized. That is, when the image 402 corresponding to the second frame is displayed, as shown in FIG. 12B, the entire portion of the image 402 and the portion corresponding to the region 411 in the image 401 The combined image is displayed. In addition, an image frame indicating the latest image among the displayed images can be displayed around the image corresponding to the current frame. In FIG. 12B, an image frame is displayed on the image 402. Further, the image conversion unit 180 holds affine transformation parameters obtained by affine transformation of the image 402.
- the image 403 is affine transformed using the affine transformation parameters associated with this frame. That is, the image 403 is affine-transformed using the affine transformation parameters obtained using the affine transformation parameter matrix corresponding to the image 403 and the affine transformation parameter matrix corresponding to the image 402 used for the immediately preceding affine transformation. .
- the matrix of affine transformation parameters corresponding to the image 403 is A3
- the matrix of affine transformation parameters corresponding to the image 402 is A2
- the matrix of affine transformation parameters corresponding to the image 401 is A1
- a value of A1 ⁇ A2 ⁇ A3 is obtained
- the image 403 is affine transformed with the obtained matrix of A1 ⁇ A2 ⁇ A3 based on the position and size of the image 401 of the first frame.
- the image shown in (c) of FIG. 12 only the position of the image 403 is converted. Then, the image 403 that has been affine transformed with the affine transformation parameters is overwritten so as to overlap the composite image of the images 401 and 402 corresponding to the previous frame.
- the composite image of the images 401 and 402 is combined. That is, when the image 403 corresponding to the third frame is displayed, as shown in FIG. 12C, the entire portion of the image 403 and the portion corresponding to the region 411 in the image 401 Then, an image obtained by combining the portion corresponding to the region 412 in the image 402 is displayed.
- the image conversion unit 180 holds affine transformation parameters obtained by affine transformation of the image 403. That is, the affine transformation parameters obtained by multiplying the matrix of affine transformation parameters corresponding to the images 402 and 403 are held in the image transformation unit 180.
- the image corresponding to the current frame is affine transformed, it is obtained using the matrix of affine transformation parameters corresponding to the current frame and the matrix of affine transformation parameters corresponding to each frame up to this point.
- the image corresponding to the current frame is affine transformed by the affine transformation parameters.
- the affine transformation parameters obtained at the time of this affine transformation are held in the image transformation unit 180 and used in the next affine transformation. The same applies to the cases of FIGS. 15 and 18.
- FIG. 13 is a diagram illustrating an example of transition of a moving image shot by a camera.
- FIG. 13 is a diagram illustrating images 421 to 423 corresponding to continuous frames included in a moving image when a person 420 is photographed with a mountain in the background.
- This example shows a case where the photographer is shooting while increasing the magnification of the lens of the camera.
- the person 420 included in the moving image photographed by the camera gradually increases in the images constituting the moving image.
- the camera position may move slightly when the magnification is increased, in this example, description will be made without considering the movement of the camera position.
- FIG. 14 is a diagram showing an example of an optical flow detected in each image shown in FIG. 13 while an image corresponding to the immediately preceding frame is indicated by a broken line.
- An image 421 shown in (a) of FIG. 14 is the same as the image 421 shown in (a) of FIG.
- the solid line portion in the image 422 shown in FIG. 14B is the same as the image 422 shown in FIG. 13B, and the broken line in the image 422 shown in FIG. This part is the same as the solid line part of the image 421 shown in FIG.
- arrows 424 to 426 in the image 422 shown in FIG. 14B indicate an example of the optical flow detected from the image 422.
- FIG. 14C is the same as the image 423 shown in FIG. 13C, and the portion of the image 423 shown in FIG. The broken line portion is the same as the solid line portion of the image 422 shown in FIG.
- arrows 427 to 429 in the image 423 illustrated in FIG. 14C indicate an example of the optical flow detected from the image 423.
- the size of the person 420 and the background mountain included in the image is changed in accordance with the change in magnification. Based on the optical flow detected by this change, affine transformation parameters can be obtained for each frame.
- FIG. 15 is a diagram showing a display example when a moving image including the images 421 to 423 shown in FIG. 13 is reproduced.
- FIG. 15A only the image 421 corresponding to the first frame is displayed at first. Subsequently, when the image 422 corresponding to the next frame is displayed, the image 422 is affine transformed using the affine transformation parameters associated with this frame. In the image shown in FIG. 15B, only the size of the image 422 is converted. Then, the image 422 that has been affine transformed with the affine transformation parameters is overwritten so as to overlap the image 421 corresponding to the immediately preceding frame. That is, in the area of the image 421, the area of the image 422 is overwritten in the area overlapping with the image 422.
- the image 421 overlaps with all the areas of the image 422, all the images of the image 422 are overwritten on the image 421.
- the image of the image 421 is synthesized. That is, when the image 422 corresponding to the second frame is displayed, as shown in FIG. 15B, the entire portion of the image 422 and the portion corresponding to the region 431 in the image 421 The combined image is displayed.
- an image frame indicating the latest image among the displayed images can be displayed around the image corresponding to the current frame. In FIG. 15B, an image frame is displayed on the image 422.
- the image conversion unit 180 holds an affine transformation parameter obtained by affine transformation of the image 422.
- the image 423 corresponding to the next frame is displayed, the image 423 is affine transformed using the affine transformation parameters associated with this frame. That is, the image 423 is affine-transformed with the affine transformation parameters obtained by multiplying the matrix of the affine transformation parameters corresponding to the image 423 and the matrix of the affine transformation parameters corresponding to the image 422 used for the immediately preceding affine transformation.
- the In the image shown in FIG. 15C only the size of the image 423 is converted. Then, the affine transformed image 423 is overwritten so as to overlap the synthesized image of the images 421 and 422 corresponding to the previous frame.
- the image of the image 423 is overwritten in the region of the composite image of the images 421 and 422 that overlaps the image 423.
- the image 423 overlaps with all the areas of the images 421 and 422, all the images of the image 423 are overwritten on the composite image of the images 421 and 422.
- the regions 432 and 433 that do not overlap with the image 423 among the regions of the combined image of the images 421 and 422 the combined image of the images 421 and 422 is combined. That is, when the image 423 corresponding to the third frame is displayed, as shown in FIG.
- the image conversion unit 180 holds an affine transformation parameter obtained by affine transformation of the image 423. That is, the affine transformation parameters obtained using the affine transformation parameters corresponding to the images 422 and 423 are held in the image conversion unit 180.
- FIG. 16 is a diagram illustrating an example of transition of a moving image shot by a camera.
- FIG. 16 is a diagram illustrating images 441 to 443 corresponding to continuous frames included in a moving image when a person 440 is photographed against a mountain background. This example shows a case where the photographer is photographing while rotating the camera around the photographing direction. In this case, the person 440 included in the moving image photographed by the camera rotates in the images constituting the moving image. Note that although the position of the camera may move slightly due to the rotation of the camera, in this example, description will be made without considering the movement of the camera position.
- FIG. 17 is a diagram illustrating an example of an optical flow detected in each image illustrated in FIG. 16 while an image corresponding to the immediately preceding frame is indicated by a broken line.
- An image 441 shown in (a) of FIG. 17 is the same as the image 441 shown in (a) of FIG.
- the solid line portion in the image 442 shown in FIG. 17B is the same as the image 442 shown in FIG. 16B
- the broken line in the image 442 shown in FIG. This part is the same as the solid line part of the image 441 shown in FIG.
- arrows 444 to 446 in the image 442 shown in FIG. 17B indicate an example of the optical flow detected from the image 442.
- 17C is the same as the image 443 shown in FIG. 16C, and the portion of the image 443 shown in FIG. The broken line portion is the same as the solid line portion of the image 442 shown in FIG.
- arrows 447 to 449 in the image 443 illustrated in FIG. 17C indicate an example of the optical flow detected from the image 443.
- the person 440 and the background mountain included in the image rotate in accordance with the rotation of the camera. Based on the optical flow detected by this rotational movement, affine transformation parameters can be obtained for each frame.
- FIG. 18 is a diagram showing a display example when a moving image including the images 441 to 443 shown in FIG. 16 is reproduced.
- the image 441 corresponding to the first frame is displayed at first. Subsequently, when the image 442 corresponding to the next frame is displayed, the image 442 is affine transformed using the affine transformation parameters associated with this frame. In the image shown in FIG. 18B, only the angle of the image 442 is converted. Then, the image 442 that has been affine transformed with the affine transformation parameters is overwritten so as to overlap the image 441 corresponding to the immediately preceding frame. That is, in the region of the image 441, the image of the image 442 is overwritten in the region 450 that overlaps with the image 442.
- the image of the image 441 is synthesized. That is, when the image 442 corresponding to the second frame is displayed, as shown in FIG. 18B, the entire portion of the image 442 and the areas 451 and 452 in the image 441 are corresponded. An image obtained by combining the portions is displayed. In addition, an image frame indicating the latest image among the displayed images can be displayed around the image corresponding to the current frame. In FIG. 18B, an image frame is displayed on the image 442.
- the image conversion unit 180 holds affine transformation parameters obtained by affine transformation of the image 442.
- the image 443 corresponding to the next frame is displayed, the image 443 is affine transformed using the affine transformation parameters associated with this frame. That is, the image 443 is affine-transformed with the affine transformation parameters obtained using the affine transformation parameter matrix corresponding to the image 443 and the affine transformation parameter matrix corresponding to the image 442 used for the immediately preceding affine transformation. . In the image shown in FIG. 18C, only the angle of the image 443 is converted. Then, the affine transformed image 443 is overwritten so as to overlap the synthesized image of the images 441 and 442 corresponding to the previous frame.
- the image of the image 443 is overwritten in areas 453 to 457 that overlap with the image 443 among the areas of the composite image of the images 441 and 442.
- the combined image of the images 441 and 442 is further combined. That is, when the image 443 corresponding to the third frame is displayed, as shown in FIG. 18C, the entire portion of the image 443 and the portion corresponding to the region 459 in the image 441 , An image obtained by combining the portions corresponding to the regions 458 and 461 in the image 442 is displayed.
- the image conversion unit 180 holds affine transformation parameters obtained by affine transformation of the image 443. That is, the affine transformation parameters obtained using the affine transformation parameters corresponding to the images 442 and 443 are held in the image conversion unit 180.
- FIG. 19 is a diagram schematically showing a relationship between each frame of the moving image file stored in the moving image storage unit 200 and the display area in the embodiment of the present invention.
- the operation accepting unit 270, the metadata storage unit 210, and the image memory 220 are illustrated, and illustrations of other configurations are omitted.
- a composite image is created in the image memory 220 using the affine transformation parameters 216 stored in the metadata file 211 for the frames “1” to “3” constituting the moving image file 201 shown in FIG. This will be described as an example.
- FIG. 19 shows an example in which one moving image is held in the image memory 220, but the case where a plurality of moving images are held in the image memory 220 is similarly synthesized.
- FIG. 19A shows a case where frame 1 (205), which is the first frame among the frames constituting the moving image file 201 shown in FIG. 5B, is stored in the image memory 220.
- frame 1 (205) which is the first frame among the frames constituting the moving image file 201 shown in FIG. 5B
- an image 471 corresponding to frame 1 (205) of the moving image file 201 is stored in the image memory 220.
- the position and size at which the image 471 corresponding to the first frame is held in the image memory 220 may be stored in a position designated in advance, and may be designated by the user in the operation reception unit 270. You may make it preserve
- the size of a composite image created by each frame constituting the moving image is calculated, and an image 471 is calculated based on this calculation. You may make it determine the position preserve
- the upper left position of the image 471 arranged on the image memory 220 is the origin, the horizontal direction (horizontal axis) is the x axis, and the vertical direction (vertical axis) is the y axis.
- a display area when an image 471 is arranged on the image memory 220 is a display area 470.
- the display area 470 may be determined according to the value of the display magnification accepted by the operation accepting unit 270, for example, based on the position and size of the composite image created from the moving image.
- the position of the display area 470 with respect to the image 471 can be determined by affine transformation parameters. That is, when the display magnification of “0.5 times” for zooming out the current image is specified, the display area is set using the affine transformation parameters that double the zoom components in the x and y directions. The Even when the display area is translated or rotated with respect to the current image, the position and range of the display area can be determined by using the affine transformation parameters.
- FIG. 19B shows a case where frame 2 (206) of the frames constituting the moving image file 201 shown in FIG. 5B is stored in the image memory 220.
- the image 472 corresponding to the frame 2 (206) using the affine transformation parameters 216 stored in the metadata file 211 in association with the frame numbers 215 “1” and “2”. Are converted and overwritten on the image 471.
- FIG. 19 (c) shows a case where frame 3 of the frames constituting the moving image file 201 shown in FIG. 5 (b) is stored in the image memory 220. Also in this case, as described above, the image 473 corresponding to the frame 3 is converted using the affine transformation parameters 216 stored in the metadata file 211 in association with the frame numbers 215 “1” to “3”. The images 471 and 352 are overwritten and synthesized.
- FIG. 20 is a diagram schematically showing the flow of each frame of the moving image file stored in the moving image storage unit 200 in the embodiment of the present invention.
- the operation receiving unit 270 the moving image storage unit 200
- the metadata storage unit 210 the image memory 220
- the display memory 240 the illustration of the other configurations is omitted.
- FIG. 20 a case where one moving image is displayed on the display unit 260 is shown as an example, but the case where a plurality of moving images are displayed on the display unit 260 is also synthesized.
- FIG. 20 (a) shows a simplified video file 201 and metadata file 211 shown in FIG. 5 (b).
- an image corresponding to the frame i (207) constituting the moving image file 201 is displayed will be described. That is, for the images corresponding to frames 1 to “i ⁇ 1” constituting the moving image file 201, it is assumed that a composite image has been created.
- FIG. 20B schematically shows an image memory 220 that holds a composite image in which images corresponding to the frames constituting the moving image file 201 are combined.
- an image 471 corresponding to frame 1 (661) constituting the moving image file 201 is first held in the image memory 220.
- each image corresponding to the frames 2 to “i ⁇ 1” constituting the moving image file 201 is associated with each of the frames 2 to “i ⁇ 1” and metadata.
- the affine transformation parameters 216 stored in the file 211 are sequentially used for affine transformation, and the affine transformation images are sequentially overwritten and held in the image memory 220.
- the display area extraction unit 230 extracts an image existing in the display area for each frame.
- An image corresponding to frame i (207) constituting the moving image file 201 is displayed in frames 1 to i in a state where a composite image of each image corresponding to frames 1 to “i-1” is held in the image memory 220.
- An affine transformation is performed using the affine transformation parameters 216 stored in the metadata file 211 in association with each other, and the affine transformed current image 474 is overwritten and held in the image memory 220.
- the display area extraction unit 230 extracts an image existing in the display area 470 from the composite image held in the image memory 220, and displays the extracted image as shown in FIG. 20C, for example. Is stored in the memory 240.
- FIG. 20C schematically shows the display memory 240 in which the image extracted by the display area extracting unit 230 is held.
- the current image 475 corresponding to the current frame among the images extracted by the display area extracting unit 230 is not the current image 474 extracted from the image memory 220 by the display area extracting unit 230 but the moving image storage unit 200.
- An image acquired and affine transformed by the image conversion unit 180 is used.
- the storage position of the current image 475 in the display memory 240 can be determined based on the position and size of the current image 474 in the image memory 220 and the position and size of the display area 470 in the image memory 220.
- the matrix of affine transformation parameters stored in the metadata file 211 in association with the frame numbers 215 “1” to “i” is A1,..., Ai, respectively, and the affine transformation parameters for determining the display area 470 are determined.
- a matrix with the image memory 220 as a reference is C, using Inv (C) ⁇ A1 ⁇ ... ⁇ Ai with the position of the image 471 as a reference, The storage location of the current image 475 can be determined.
- the image extracted by the display area extraction unit 230 is held in the display memory 240, and the image extracted by the display area extraction unit 230 is acquired from the moving image storage unit 200.
- the image affine transformed by the image conversion unit 180 is overwritten and held in the display memory 240.
- the image held in the display memory 240 is displayed on the display unit 260.
- a relatively clean current image can be displayed by using an image in a state before being subjected to processing such as reduction after affine transformation and stored in the image memory 220.
- the current image can be displayed in a beautiful state even when it is enlarged by a user operation.
- FIG. 21 is a figure which shows the example of a display in the case of reproducing
- an image 480 is shown during reproduction of a moving image when a parent and child playing in a grassy plaza with a large building are photographed while moving the camera mainly in the left-right direction.
- an image 481 synthesized by images corresponding to the respective claims constituting the moving image is formed in a panoramic shape.
- the image corresponding to the current frame in the image 480 is the current image 482.
- the user can specify a desired display magnification by operating a display magnification specifying key in the operation receiving unit 270.
- a display magnification specifying key for example, as shown in FIG. 21A, when the image 480 is displayed on the display unit 260, when the image area surrounded by the frame 483 is displayed in an enlarged manner, the user uses the operation receiving unit 270.
- the display magnification designation key By operating the display magnification designation key to designate the display magnification and designating the position, the image area surrounded by the frame 483 can be enlarged and displayed.
- FIG. 21 is a diagram showing an image 484 in a state before the current image 482 in the image 480 is subjected to affine transformation.
- FIG. 22A is a diagram showing an image 485 when the image region surrounded by the frame 483 shown in FIG. 21A is enlarged and displayed.
- An image 485 shown in FIG. 22A is an image synthesized in the display memory 240 in a state before the current image after the affine transformation is stored in the image memory 220.
- a relatively fine image in a state before being stored in the image memory 220 is displayed.
- the current image 486 is compared with an area other than this area, the current image 486 that is relatively clearer than the other areas can be seen.
- the 22B is an image stored in the display memory 240 in a state where the current image after affine transformation is stored in the image memory 220. When displayed in this way, an image of the same level as the image of the other region is also displayed in the region of the current image 488. That is, according to the embodiment of the present invention, when an image is combined and displayed, the history image held in the display memory 240 may be compressed, but the current (current) image is an uncompressed image. Alternatively, since an image having a higher resolution than that of the history image can be used, high-quality image composition display can be realized.
- the composite image is displayed only in the display area of the display unit 260 in the area where the image corresponding to the current frame and the previous frame constituting at least one of the moving images is displayed.
- An example of blackening the area is shown.
- a frame is displayed around the image corresponding to the current frame.
- the display example shown below shows a display example from the middle of one moving image being reproduced.
- a composite image is sequentially displayed for each frame, but in the same figure, a display example is shown at a predetermined number of frame intervals, and illustration of the composite image displayed between the frames is omitted. For this reason, the movement of the frame corresponding to the current frame is large in FIG.
- FIG. 23 to FIG. 26 are diagrams showing an example of transition of a moving image taken by a camera.
- the figure shows images 730 to 741 constituting a moving image when a parent and child playing in a grassy square with a large building are photographed while moving the camera mainly in the left-right direction. Note that this moving image is the same as the moving image shown in FIGS. 21 and 22.
- the image corresponding to the current frame is the image 750.
- the current image is indicated by the same reference numeral 750 even if the synthesized images are different.
- an object to be photographed (grass square or the like) included in the photographed image is fixed on the screen, and an image 750 corresponding to the current frame moves on the screen in accordance with the movement of the camera.
- a moving image to be reproduced includes a moving image that has not been edited since it was captured by the camera (an unedited moving image), a moving image that has been captured after being captured by the camera (an edited moving image), and the like.
- These unedited videos include parts that are not in focus on some or all of them, those that include parts where people suddenly crossed the front of the camera during imaging, and cameras that change drastically during imaging. It is conceivable that the change is included in the image, or another scene is imaged while the imaging moving image recording instruction button is pressed.
- the edited moving image for example, one in which different scenes are linked by editing can be considered.
- the affine transformation parameters cannot be detected properly and the affine transformation parameters are erroneously detected.
- the image cannot be appropriately converted.
- the affine transformation parameter is corrected or the synthesized image is divided based on the feature amount extracted from each image constituting the moving image.
- the affine transformation parameter is corrected when the feature amount extracted from each image constituting the moving image satisfies a predetermined condition.
- a predetermined condition relating to the correction of the affine transformation parameter for example, the condition is that the scene change score is equal to or greater than a threshold value, or the case where at least one of each component of the camera work is equal to or greater than the threshold value. Can do. Moreover, you may make it make the conditions satisfy
- an affine transformation parameter is erroneously detected at a position determined to be a scene change point, it is possible to avoid failure as a composite image by performing correction.
- a correction method for correcting the affine transformation parameters a method of correcting by replacing with a unit matrix, or a method of correcting by a matrix (for example, a linear interpolation matrix) that can suppress the camera work speed can be used. Thereby, a high-speed camera work can be suppressed and the browsing property of a moving image can be improved. Further, even when an affine transformation parameter is erroneously detected, it is possible to suppress an influence on the composition of images after the image related to the erroneous detection.
- the affine transformation parameter correction method will be described in detail with reference to FIG.
- a setting for prohibiting correction may be made in accordance with an operation input from the operation reception unit 270.
- the synthesized image when the feature amount extracted from each image constituting the moving image satisfies a predetermined condition, the synthesized image is different from the synthesized image formed by each previous image.
- a discontinuous moving image can be isolate
- the images corresponding to each scene can be dispersed, it is possible to easily search a wide range of moving images over time.
- the predetermined condition relating to the division of the composite image for example, when the scene change score is equal to or greater than a threshold, when at least one of the components of the camera work is equal to or greater than the threshold, each average value of the camera work components When at least one of them is equal to or greater than a threshold value, when at least one of the integrated values of the camera work components is equal to or greater than the threshold value (zoom-in / zoom-out rate threshold value determination, panning or tilting angle threshold value determination, etc.)
- the playback time from the first frame may be a threshold or more. Further, a case where at least two of these are satisfied may be set as a predetermined condition. Note that the composite image dividing method will be described in detail with reference to FIGS.
- the position corresponding to the scene change score can be determined as the scene change point.
- the position determined to be the scene change point may be a cut point (edit point) or a position where the subject of the same scene has changed greatly.
- a camerawork component when using a camerawork component as a condition for dividing a composite image, it is preferable to use a camerawork component for a relatively long time rather than an instantaneous camerawork component. That is, it is preferable to perform threshold determination using an average value of components for the past several frames (a value to which a low-pass filter is applied) rather than a camerawork component related to one frame. By doing so, excessive division can be prevented.
- REC button imaging moving image recording instruction button
- the integrated value of the zoom component among the integrated values of the camera work component is an enlargement of the camera work component corresponding to the current frame with respect to the first frame (in this case, the first frame after division) constituting the moving image.
- the amount of change in the reduction component, and the integrated value of this zoom component can be used for threshold determination. For example, if the zoom-in operation is excessively performed at the time of image capturing by the camera, the current image in the synthesized image being synthesized and reproduced may become too small, which may make it difficult to see the current image. Therefore, when the change amount of the enlargement / reduction component of the camera work component becomes large, it is possible to make a moving image related to the combined reproduction easy to view by setting the combined image as a new combined image.
- the translational component integration value of the integration values of the camerawork component is the translation of the camerawork component corresponding to the current frame with respect to the first frame (in this case, the first frame after division) constituting the moving image.
- the amount of change of the component, and the integrated value of the translation component can be used for threshold determination.
- the integrated value of the translation component can be considered as a panning angle or a tilting angle. For example, when an image is captured by an actual camera, the photographer rotates while taking the photographer holding the camera as the center position, and the camera is directed in one direction while moving horizontally. It is difficult to distinguish from the case where imaging is performed. However, the rotation angle of the camera when the photographer rotates and takes an image is reflected in the size of the composite image.
- the panning angle or tilting angle and the translation component are regarded as the same, and the division process can be performed.
- the panning operation is excessively performed at the time of image pickup by the camera, the combined image being combined and reproduced may become extremely horizontally long.
- the tilting operation is excessively performed at the time of image pickup by the camera, the combined image being combined and reproduced may be extremely long. For this reason, it may be difficult to see an image and a composite image corresponding to the current frame. Therefore, when the amount of change in the translational component of the camera work component becomes large, it is possible to make the moving image related to the combined reproduction easy to see by using a new combined image obtained by dividing the combined image.
- the playback time from the first frame is equal to or greater than the threshold.
- the overlap of each frame increases, so the amount of information related to the moving image May be lacking.
- an image with a relatively new playback time is overwritten on an image with a relatively old playback time.
- the functionality as an index of the composite image is inferior and the viewability is inferior. Therefore, by forcibly dividing the composite image in a certain time (for example, 5 minutes), the functionality as an index of the composite image and the viewability of the composite image can be improved.
- a minimum reproduction time (for example, 5 seconds) may be set. That is, even when the predetermined condition relating to the division of the composite image described above is satisfied, the division processing is not performed within the minimum reproduction time after the division of the composite image.
- the function as an index of the composite image to be displayed can be improved, and the consistency as the video is ensured and the composite image Can be improved.
- these division determinations may be performed independently based on each element, or may be performed only when a plurality of elements satisfy the condition at the same time. Further, for example, when detecting affine transformation parameters for each image constituting a moving image, a detection result indicating that feature points cannot be matched is stored in the metadata storage unit 210 in association with a frame, and this detection is performed. Based on the result, the necessity for correction or the necessity for division may be determined. That is, if the detection result is normal detection although the threshold value is exceeded, it can be determined that correction is not performed.
- FIG. 27 is a diagram schematically illustrating a case where the images 501 to 517 constituting the moving image file 500 are combined.
- FIG. 27A shows a case where the images 501 to 517 are synthesized using the affine transformation parameters associated with the images 501 to 517.
- FIGS. 27B and 27C show the images.
- a case is shown in which images 501 to 517 are synthesized using affine transformation parameters in which some of the affine transformation parameters associated with 501 to 517 are corrected.
- the images 508 to 510 may be converted regardless of the movement of the camera.
- the images 508 to 510 may be converted larger than the conversion amount of other images regardless of the movement of the camera.
- the parts of the images 508 to 510 are greatly different from the part of the actual subject, so that the composite image being reproduced and finally formed. There is a high possibility that the composite image will be difficult to see.
- the composite image is prevented from being significantly different from the actual subject portion, and the composite image being reproduced and the final image are reproduced. It is possible to make it easy to see the synthesized image.
- a linear method based on a correction method for correcting a matrix of affine transformation parameters into a unit matrix and a matrix of affine transformation parameters associated with an image in the front-rear direction A correction method for performing interpolation will be described.
- the images 508 to 510 are converted using the unit matrix, when viewed on the basis of the image 507, the images 508 to 510 are not actually converted and are located at the same position as the immediately preceding image 507. Synthesized. That is, as shown in FIG. 27B, the images 508 to 510 are overwritten and synthesized at the position of the image 507 whose inside is indicated by diagonal lines. Note that the image 511 immediately after the image 510 is subjected to normal conversion using the affine transformation parameters associated with the image 511.
- a front image that is an image positioned before these continuous images and an image that is positioned after these continuous images A side image is extracted.
- the front image and the rear image are images that are adjacent in the front-rear direction of the above-described successive images among images whose feature amounts do not satisfy the correction condition. Then, correction is performed using the average value of the matrix of affine transformation parameters associated with the extracted front image and rear image.
- the images 508 to 510 are converted using the linear interpolation matrix, the images 508 to 510 are converted by the average value of the conversion amounts of the images 507 and 511 located before and after the images 508 to 510, for example, FIG. As shown in (c) of FIG. 5, images 508 to 510 are sequentially synthesized by overwriting an image 507 whose inside is hatched. Subsequently, normal conversion is performed on the image 511 immediately after the image 510 using the affine conversion parameters associated with the image 511.
- the affine transformation parameters associated with the images 508 to 510 are uncertain affine transformation parameters (when the feature amount is equal to or greater than a threshold value).
- the images 507 to 517 are converted in a direction different from the movement of the camera.
- the images 507 to 517 have different directions from the camera movement. Therefore, it is possible to make it easy to see the composite image being reproduced or the composite image created.
- a correction method for correcting an affine transformation parameter a case where the affine transformation matrix is corrected to a unit matrix or a case where a linear interpolation matrix is corrected has been described.
- a matrix of affine transformation parameters associated with images up to a predetermined number before the target image may be sequentially added to calculate an average value of the matrix after the addition, and the average value matrix may be corrected. .
- FIG. 28 is a diagram schematically showing the relationship between each frame constituting the moving image and the scene change score calculated for these frames.
- the graph shown in FIG. 28A is a graph showing the scene change score calculated for each frame constituting the moving picture in time series, the horizontal axis shows the imaging time, and the vertical axis shows the scene change score.
- a dotted line 533 is a dotted line indicating a threshold value for determining whether or not an image corresponding to each frame corresponds to a scene change point in the graph shown in FIG. That is, when the scene change score is equal to or higher than the dotted line 533, an image corresponding to the scene change score is determined as an image corresponding to the scene change point.
- a position indicating a scene change score calculated for the frame 11 (531) among the frames 1 to 20 constituting the moving image 530 is indicated by a point 532.
- the image corresponding to the frame 11 (531) is determined to be an image corresponding to the scene change point.
- the scene change score other than the frame 11 (531) is equal to or less than the dotted line 533, and thus is not determined to be an image corresponding to the scene change point.
- the scene change score is determined by the conversion information correction unit 160 and the composite image division determination unit 170.
- FIG. 29 is a diagram schematically showing a case where the images 551 to 567 composing the moving image 530 shown in FIG. 28B are synthesized.
- FIG. 29A shows a case where images 551 to 567 are synthesized using affine transformation parameters associated with images 551 to 567
- FIGS. 29B and 29C show images. A transition when a part of 551 to 567 is divided and combined is shown.
- images 551 to 567 constituting the moving image 530 shown in FIG. 29 are images corresponding to the frames 1 to 17 shown in FIG. 28B, and an image 561 whose inside is hatched is shown in FIG. The image corresponds to the frame 11 (531) shown in FIG. That is, the image 561 is an image corresponding to a scene change point.
- the scene change score corresponding to the current image is equal to or greater than the threshold value and at least one of the average values of the camera work components is equal to or greater than the threshold value, the scene change score is associated with the current image.
- An example is shown in which the current image is not converted using the affine transformation parameters, and a new composite image is created separately from the current composite image.
- the image 561 is displayed. Is included in the moving image 530, but is likely to be an image captured in a scene different from the images 551 to 560.
- the images 562 to 567 after the image 561 are captured in the same scene as the image 561, there is a high possibility that the images are captured in a scene different from the images 551 to 560.
- FIG. 29 (a) when images 561 to 567 captured in a scene different from images 551 to 560 are overwritten on images 551 to 560, two different scenes are included.
- FIG. 30 is a diagram schematically illustrating a case where the images 551 to 567 composing the moving image 530 illustrated in FIG. 28B are combined.
- the composite image shown in FIG. 30A is the same as the composite image shown in FIG. FIGS. 30B and 30C show transitions in the case where a part of the images 551 to 567 is divided and combined, and the combined image before the division is left. Note that this is the same as the case shown in FIG. 29 except that the composite image before division is left and the current image is arranged at a position different from the position of the composite image before division.
- the arrangement position of the image corresponding to the first frame constituting the moving image is set at the origin on the xy coordinates, and the arrangement position of the image satisfying the division condition is on the xy coordinates. Is set as the arrangement position 580.
- the current image becomes an image 561 that satisfies the division conditions after the composite reproduction up to the image 560 is performed, it is shown in (c) of FIG.
- the image 561 is held at the arrangement position 580 different from the position of the composite image formed by the images 551 to 560.
- an image holding position that satisfies the division condition may be set in advance, and is determined based on the size of the combined image so that the arrangement position is different from the already formed combined image. You may make it do.
- FIG. 30 an example in which the right side of an already formed composite image is set as the holding position of an image that satisfies the division condition has been described, but another direction may be set as the holding position.
- FIG. 31 and FIG. 32 are diagrams illustrating an example of a composite image formed by a moving image photographed by a camera.
- FIG. 31 when a photographer in a building is photographing the outdoors from a building window, the camera is turned leftward by a relatively fast panning operation, and the moving image is taken inside the building.
- Composite images 590, 592, and 593 formed by the above are shown. That is, in the composite image 590, the right image is a relatively old image, and becomes a new image as it proceeds to the left. Further, the portion of the image surrounded by the dotted circle 591 shown in FIG. 31A is a portion in which the orientation of the camera is changed by a relatively fast panning operation.
- the subject before and after the image is synthesized regardless of the imaging space.
- FIG. 31A an outdoor subject that should actually be separated and an indoor subject are combined at a portion surrounded by a dotted circle 591. Therefore, a composite image corresponding to the subject can be created by dividing the composite image described above or correcting the affine transformation parameters.
- the condition for correcting the affine transformation parameter and the condition for dividing the composite image are satisfied in the portion of the image surrounded by the dotted circle 591 shown in FIG.
- a composite image 592 illustrated in FIG. 31B is a composite image corresponding to the right portion of the composite image 590 illustrated in FIG.
- conditions for dividing the composite image in the portion of the image surrounded by the dotted circle 591 shown in FIG. In order to satisfy the above, the composite image 592 shown in FIG. 31B is deleted. Then, an image that satisfies the condition for dividing the composite image is displayed at the arrangement position of the first frame, and the subsequent images are sequentially combined to form a composite image 593.
- FIG. 32 in the same manner as in FIG. 31, in the state where the photographer in the building is photographing the outside from the building window, the direction of the camera is changed to the left side by a relatively fast panning operation to 3 shows synthesized images 590, 594, and 595 formed by moving images. Note that the composite image 590 and the dotted circle 591 shown in FIG. 32A are the same as those shown in FIG.
- a composite image 594 shown in FIG. 32B is a composite image corresponding to the right portion of the composite image 590 shown in FIG.
- the affine transformation parameter correction is performed in the image portion surrounded by the dotted circle 591 shown in FIG.
- the composite image 594 shown in FIG. 32B is deleted. Then, an image that satisfies the condition for dividing the composite image is displayed at the arrangement position of the first frame, and subsequent images are sequentially combined to form a composite image 595.
- the correction for the affine transformation parameter is sequentially performed when the condition for correcting the affine transformation parameter is satisfied.
- the composite image 595 shown in FIG. 32C is different from the composite image 593 shown in FIG. 31C because the affine transformation parameters are sequentially corrected. That is, since the affine transformation parameters are corrected in the left part of the image surrounded by the dotted circle 591 shown in FIG. 32A, the affine transformation parameter is corrected compared to the composite image 593 shown in FIG. The change in the left-right direction can be reduced, and an image portion on which a relatively fast panning operation has been performed can be corrected appropriately.
- the user can view the composite image corresponding to the scene.
- affine transformation parameters that have not been properly detected by the relatively fast panning operation can be corrected to reduce unnecessary changes in the composite image.
- the example in which an already formed composite image is erased and another composite image is played back has been described. However, when another composite image is played back with the already formed composite image remaining. The same applies to.
- FIG. 33 is a flowchart showing a processing procedure of moving image reproduction processing by the image processing apparatus 100 according to the embodiment of the present invention.
- the composite image division determination unit 170 outputs that the image corresponding to the current frame is to be divided and reproduced, the composite image of each image corresponding to the frame before the current frame is displayed. An example of erasing will be described.
- Step S921 a work buffer larger than the size of the image constituting the moving image is secured in the image memory 220.
- the file acquisition unit 140 acquires the moving image file selected by the operation receiving unit 270 from the moving image storage unit 200, and acquires the metadata file associated with these moving image files from the metadata storage unit 210. (Step S922).
- the file acquisition unit 140 decodes the moving image file and acquires a current frame that is one frame constituting the moving image file (step S923). Subsequently, the file acquisition unit 140 acquires an affine transformation parameter corresponding to the acquired current frame from the metadata file (step S924). Here, if the current frame is the first frame, the affine transformation parameters of the unit matrix are acquired.
- the feature amount extraction unit 150 calculates a scene change score for the image corresponding to the current frame (step S925).
- the camera work component calculation unit 155 calculates each component of the camera work from the affine transformation parameters corresponding to the current frame (step S926).
- the camerawork component average value calculation unit 157 applies each component of the camerawork corresponding to the current frame and each frame from the immediately previous frame held in the camerawork component holding unit 156 to a frame a predetermined number of frames ago. Based on the corresponding components of the camera work, an average value of the camera work components is calculated (step S927).
- the camera work component integrated value calculation unit 158 corresponds to each component of the camera work corresponding to the current frame and each frame from the first frame held in the camera work component holding unit 156 to the previous frame. Based on each component of the camera work, an integrated value of the camera work component is calculated (step S928).
- step S940 conversion information correction processing
- step S940 conversion information correction processing
- step S950 composite image division determination process
- the image conversion unit 180 affine-transforms the image corresponding to the current frame using the affine transformation parameters (step S929).
- the composite image division determination unit 170 outputs that the image corresponding to the current frame is to be divided and reproduced, the actual image is not converted because the affine transformation is performed using the unit matrix.
- the conversion information correction unit 160 corrects the affine transformation parameters when the composite image division determination unit 170 does not output that the image corresponding to the current frame is divided and reproduced, the affine transformation after the correction is performed. Affine transformation is performed using the parameters.
- the affine transformation corresponding to the current frame is performed. Affine transformation is performed using the parameters.
- the current frame is the first frame, an affine transformation is performed using a unit matrix, and thus an actual image is not transformed.
- the image composition unit 190 synthesizes an image corresponding to the current frame subjected to the affine transformation by overwriting the composite image of each image corresponding to the frame before the current frame, and corresponds to the current frame.
- the combined image is stored in the image memory 220 (step S930). If the composite image division determination unit 170 outputs that the image corresponding to the current frame is to be divided and reproduced, the composite image of each image corresponding to the frame before the current frame is deleted. Therefore, only the image corresponding to the current frame is stored in the image memory 220.
- the display area extraction unit 230 determines the position and size of the display area using the affine transformation parameters corresponding to the designated display magnification and position (step S931). Subsequently, the display area extraction unit 230 extracts the composite image included in the display area from the image memory 220 (step S932). Subsequently, the display area extraction unit 230 stores the composite image extracted from the image memory 220 in the display memory 240 (step S933).
- the display area extraction unit 230 uses the affine transformation parameter matrix used for transforming the current image and the inverse matrix for the affine transformation parameter matrix used for determining the display area, to display memory 240.
- the position of the current image at is determined (step S934).
- the image composition unit 190 synthesizes the composite image stored in the display memory 240 by overwriting the current image after affine transformation (step S935).
- the composite image stored in the display memory 240 is displayed on the display unit 260 (step S936).
- the composite image division determination unit 170 outputs that the image corresponding to the current frame is to be divided and reproduced, the composite image of each image corresponding to the frame before the current frame is deleted.
- Step S937 it is determined whether or not the current frame is the last frame among the frames constituting the input moving image file. If the current frame is not the last frame (step S937), the process returns to step S923 to repeat the moving image synthesis reproduction process (steps S923 to S936). On the other hand, if the current frame is the last frame (step S937), the secured work buffer is released (step S938), and the moving image playback process is terminated.
- FIG. 34 is a flowchart showing a conversion information correction processing procedure (processing procedure of step S940 shown in FIG. 33) in the processing procedure of the moving image reproduction processing by the image processing apparatus 100 according to the embodiment of the present invention.
- the affine transformation parameter corresponding to the current frame is determined based on whether at least one of the components of the camera work corresponding to the current frame is equal to or greater than the threshold and whether the scene change score is equal to or greater than the threshold.
- An example of correction will be described.
- the conversion information correction unit 160 determines whether or not at least one of the components of the camera work corresponding to the current frame is equal to or greater than a threshold value (step S941). If at least one of the camerawork components corresponding to the current frame is not greater than or equal to the threshold (step S941), the conversion information correction unit 160 determines whether or not the scene change score calculated for the current frame is greater than or equal to the threshold. Judgment is made (step S942). If the scene change score calculated for the current frame is not greater than or equal to the threshold value (step S942), it is not necessary to correct the affine transformation parameters corresponding to the current frame, and the operation of the conversion information correction process is terminated.
- the conversion information correction unit 160 corrects the matrix of affine transformation parameters corresponding to the current frame into a unit matrix (step S943).
- correction is performed using a linear interpolation matrix instead of the unit matrix, if it is determined in step S941 or S942 that the threshold is greater than or equal to the threshold, it corresponds to each frame in the front-rear direction of the current frame in step S943.
- a linear interpolation matrix is calculated using the affine transformation parameters, and correction is performed using this linear interpolation matrix.
- FIG. 35 is a flowchart showing a composite image division determination processing procedure (processing procedure of step S950 shown in FIG. 33) in the processing procedure of moving image reproduction processing by the image processing apparatus 100 according to the embodiment of the present invention.
- at least one of the average values of each component of the camera work corresponding to the current frame is equal to or greater than the threshold value, and whether the scene change score is equal to or greater than the threshold value.
- An example will be described in which it is determined whether or not the image corresponding to the current frame needs to be divided based on whether or not at least one of the integrated values is equal to or greater than the threshold and whether or not the reproduction time is equal to or greater than the threshold.
- the composite image division determination unit 170 determines whether at least one of the average values of the components of the camera work corresponding to the current frame is equal to or greater than a threshold (step S951). If at least one of the average values of the components of the camera work corresponding to the current frame is not equal to or greater than the threshold value (step S951), the process proceeds to step S953. On the other hand, if at least one of the average values of the components of the camera work corresponding to the current frame is equal to or greater than the threshold value (step S951), the composite image division determination unit 170 determines that the scene change score calculated for the current frame is the threshold value. It is determined whether or not the above is true (step S952).
- the composite image division determination unit 170 has at least one of the integrated values of the components of the camera work corresponding to the current frame greater than or equal to the threshold. Whether or not (step S953).
- the composite image division determination unit 170 is included in the current composite image held in the image memory 220. It is determined whether or not the playback time from the first frame of the image to be recorded has exceeded a threshold value (step S954).
- step S954 If the playback time from the first frame of the images included in the current composite image held in the image memory 220 has not exceeded the threshold (step S954), the image corresponding to the current frame is divided. Since it is not necessary, the operation of the composite image division determination process ends.
- step S951 and S952 when at least one of the average values of each component of the camera work corresponding to the current frame is equal to or greater than the threshold and the scene change score calculated for the current frame is equal to or greater than the threshold (steps S951 and S952), When at least one of the integrated values of each component of the corresponding camera work is greater than or equal to the threshold (step S953), or from the first frame of the images included in the current composite image held in the image memory 220 If the reproduction time has exceeded the threshold (step S954), the image composition unit 190 deletes the composite image held in the image memory 220 (step S955). Subsequently, the image conversion unit 180 changes the matrix of affine transformation parameters corresponding to the current frame to a unit matrix (step S956).
- FIG. 36 is a diagram showing a configuration example of the multi-core processor 800 in the embodiment of the present invention.
- the multi-core processor 800 is a processor in which a plurality of different types of processor cores are mounted on one CPU (Central Processing Unit) package. That is, in the multi-core processor 800, in order to maintain the processing performance of each processor core alone and to have a simple configuration, one type of core corresponding to every application (application) and a certain degree of optimization for a predetermined application are used. Two or more types of processor cores with other types of cores are mounted.
- CPU Central Processing Unit
- the multi-core processor 800 includes a control processor core 801, arithmetic processor cores (# 1) 811 to (# 8) 818, and a bus 802, and is connected to the main memory 781.
- the multi-core processor 800 is connected to other devices such as a graphics device 782 and an I / O device 783, for example.
- the multi-core processor 800 for example, “Cell (Cell: Broadband Engine)” which is a microprocessor developed by the applicant of the present application can be employed.
- the control processor core 801 is a control processor core that mainly performs frequent thread switching such as an operating system.
- the control processor core 801 will be described in detail with reference to FIG.
- the arithmetic processor cores (# 1) 811 to (# 8) 818 are simple and small arithmetic processor cores that specialize in multimedia processing.
- the arithmetic processor cores (# 1) 811 to (# 8) 818 will be described in detail with reference to FIG.
- the bus 802 is a high-speed bus called EIB (Element Interconnect Bus), and the control processor core 801 and the arithmetic processor cores (# 1) 811 to (# 8) 818 are connected, and data access by each processor core is performed. This is done via the bus 802.
- EIB Element Interconnect Bus
- the main memory 781 is connected to the bus 802 and stores various programs to be loaded into each processor core and data necessary for processing of each processor core, and also stores data processed by each processor core. is there.
- the graphics device 782 is a graphics device connected to the bus 802, and the I / O device 783 is an external input / output device connected to the bus 802.
- FIG. 37 is a diagram showing a configuration example of the control processor core 801 in the embodiment of the present invention.
- the control processor core 801 includes a control processor unit 803 and a control processor storage system 806.
- the control processor unit 803 is a unit that performs the arithmetic processing of the control processor core 801.
- the control processor unit 803 includes an instruction set based on a microprocessor architecture, and includes an instruction cache 804 and a data cache 805 as primary caches.
- the instruction cache 804 is, for example, a 32 KB instruction cache
- the data cache 805 is, for example, a 32 KB data cache.
- the control processor storage system 806 is a unit that controls data access from the control processor unit 803 to the main memory 781, and is equipped with a 512 KB secondary cache 807 for speeding up memory access from the control processor unit 803. Yes.
- FIG. 38 is a diagram showing a configuration example of the arithmetic processor core (# 1) 811 in the embodiment of the present invention.
- the arithmetic processor core (# 1) 811 includes an arithmetic processor unit 820 and a memory flow controller 822.
- the arithmetic processor cores (# 2) 812 to (# 8) 818 have the same configuration as that of the arithmetic processor core (# 1) 811, and thus the description thereof is omitted here.
- the arithmetic processor unit 820 is a unit that performs the arithmetic processing of the arithmetic processor core (# 1) 811 and includes a unique instruction set different from the control processor unit 803 of the control processor core 801.
- the arithmetic processor unit 820 is equipped with a local store (LS) 821.
- the local store 821 is a dedicated memory of the arithmetic processor unit 820, and is the only memory that can be directly referred to from the arithmetic processor unit 820.
- As the local store 821 for example, a memory having a capacity of 256 Kbytes can be used.
- the memory flow controller 822 is used for the arithmetic processor unit 820 to access a local store on the main memory 781 and other arithmetic processor cores (arithmetic processor cores (# 2) 812 to (# 8) 818). There is a need to.
- the memory flow controller 822 is a unit for exchanging data with the main memory 781 and other arithmetic processor cores, and is a unit called an MFC (Memory Flow Controller).
- MFC Memory Flow Controller
- the arithmetic processor unit 820 requests the memory flow controller 822 for data transfer or the like via an interface called a channel.
- FIG. 39 is a diagram schematically showing a calculation method of the multi-core processor 800 in the embodiment of the present invention.
- the control processor core 801 executes the task 784 using the data 785
- the task 786 necessary for the processing of the task 786 that is a part of the task 784 (a part of the data 785) is used.
- An example in which each arithmetic processor core executes 786 is illustrated.
- the control processor core 801 executes the task 784 using the data 785, the data 787 (part of the data 785) necessary for processing of the task 786 which is a part of the task 784 is obtained. And causes each arithmetic processor core to execute task 786.
- arithmetic processing is performed by each arithmetic processor core for each frame constituting a moving image.
- the calculation processor cores (# 1) 811 to (# 8) 818 can be used in parallel to perform many calculations in a relatively short time.
- SIMD Single Instruction / Multiple Data: single instruction / multiple data
- a relatively large number of instructions can be obtained. Arithmetic processing can be performed.
- the SIMD calculation will be described in detail with reference to FIGS. 43 to 46 and the like.
- FIG. 40 is a diagram schematically showing the flow of programs and data in the case where operations are performed by the multi-core processor 800 in the embodiment of the present invention.
- the arithmetic processor core (# 2) 812 to (# 8) 818 will be described by taking the arithmetic processor core (# 1) 811 of the arithmetic processor cores (# 1) 811 to (# 8) 818 as an example. The same can be done for.
- control processor core 801 sends an instruction to load the arithmetic processor core program 823 stored in the main memory 781 to the local store 821 of the arithmetic processor core (# 1) 811 to the arithmetic processor core (# 1) 811. .
- the arithmetic processor core (# 1) 811 loads the arithmetic processor core program 823 stored in the main memory 781 into the local store 821.
- control processor core 801 instructs the arithmetic processor core (# 1) 811 to execute the arithmetic processor core program 825 stored in the local store 821.
- the arithmetic processor core (# 1) 811 transfers data 824 necessary for execution processing of the arithmetic processor core program 825 stored in the local store 821 from the main memory 781 to the local store 821.
- the arithmetic processor core (# 1) 811 processes the data 826 transferred from the main memory 781 based on the arithmetic processor core program 825 stored in the local store 821, and executes processing according to the conditions.
- the processing result is stored in the local store 821.
- the arithmetic processor core (# 1) 811 transfers the processing result executed based on the arithmetic processor core program 825 stored in the local store 821 from the local store 821 to the main memory 781.
- the arithmetic processor core (# 1) 811 notifies the control processor core 801 of the end of the arithmetic processing.
- SIMD operation is an operation method in which processing for a plurality of data is performed with one instruction.
- FIG. 41 (a) is a diagram schematically showing an outline of an arithmetic method for performing processing on a plurality of data with respective instructions.
- the calculation method shown in (a) of FIG. 41 is a normal calculation method, and is called, for example, scalar calculation.
- the processing result of the data “C1” is obtained by an instruction to add the data “A1” and the data “B1”.
- the instructions for adding the data “A2”, “A3”, “A4” and the data “B2”, “B3”, “B4” in the same row for the other three operations respectively By this instruction, the values of the respective rows are added, and the processing result is obtained as data “C2”, “C3”, “C4”.
- FIG. 41 is a diagram schematically showing an outline of SIMD operation, which is an operation method for performing processing on a plurality of data with one instruction.
- data grouped for SIMD calculation (each data surrounded by dotted lines 827 and 828) may be referred to as vector data.
- SIMD calculation performed using such vector data may be referred to as vector calculation.
- data "A1" and “B1” are added, data “A2” and “B2” are subtracted, data “A3” and “B3” are multiplied, and data "A4" and "B4" Cannot be performed by SIMD calculation. That is, when different processing is performed on each of a plurality of data, processing by SIMD calculation cannot be performed.
- FIG. 42 is a diagram showing a configuration example of a program executed by the control processor core 801 or the arithmetic processor core (# 1) 811 in the embodiment of the present invention. Although only the arithmetic processor core (# 1) 811 is illustrated here, the same processing is performed for the arithmetic processor cores (# 2) 812 to (# 8) 818.
- the control processor core 801 executes decode 852, interlace 853, and resize 854 as decode 851.
- the decode 852 is a process for decoding a moving image file.
- the interlace 853 is a process for removing the interlace for each decoded frame.
- Resizing 854 is a process of reducing each frame from which the interlace is removed.
- control processor core 801 executes instruction transmission 857 and 859 and end notification reception 858 and 860 as the arithmetic processor core management 856.
- the instruction transmissions 857 and 859 are processes for transmitting SIMD calculation execution instructions to the arithmetic processor cores (# 1) 811 to (# 8) 818
- the end notification receptions 858 and 860 are the arithmetic processor cores (# 1) Processing for receiving SIMD calculation end notifications from 811 to (# 8) 818.
- the control processor core 801 executes camera work parameter calculation processing 862 as camera work detection 861.
- the camera work parameter calculation processing 862 is processing for calculating an affine transformation parameter for each frame based on an optical flow calculated by SIMD calculation by the arithmetic processor cores (# 1) 811 to (# 8) 818.
- the arithmetic processor core (# 1) 811 includes, as a feature point extraction process 863, a Sobel filter process 864, a second moment matrix calculation process 865, a separable filter process 866, Harris.
- a corner extraction (Calc Harris) process 867, a dilation process (Dilation) 868, and a rearrangement process (Sort) 869 are executed.
- the Sobel filter process 864 is a process of calculating the value dx in the x direction obtained using the P2 filter (x direction) and the value dy in the y direction obtained using the Y direction filter. The calculation of the value dx in the x direction will be described in detail with reference to FIGS. 43 to 46.
- the second moment matrix calculation process 865 is a process of calculating each value of dx 2 , dy 2 , dx ⁇ dy using dx and dy calculated by the Sobel filter process 864.
- the separable filter process 866 is a process of applying a Gaussian filter (blurring process) to the dx 2 , dy 2 , dx ⁇ dy images calculated by the second moment matrix calculation process 865.
- the Harris corner extraction process 867 is a process of calculating a Harris corner score using each value of dx 2 , dy 2 , dx ⁇ dy subjected to the blurring process by the separable filter process 866.
- the Harris corner score S is calculated by the following equation, for example.
- the expansion process 868 is a process of performing a blurring process on an image composed of the Harris corner score calculated by the Harris corner extraction process 867.
- the rearrangement processing 869 arranges pixels in descending order of the Harris corner score calculated by the Harris corner extraction processing 867, picks up a predetermined number from the higher score, and extracts the picked points as feature points. It is processing.
- the arithmetic processor core (# 1) 811 executes a multi-resolution image creation (Make Pyramid Image) processing 871 and an optical flow calculation (Calc Optical Flow) processing 872 as an optical flow (Optical Flow) processing 870.
- the multi-resolution image creation process 871 is a process for sequentially creating images reduced to a predetermined number of stages from the image size at the time of image capture by the camera, and the created image is called a multi-resolution image.
- the optical flow calculation process 872 calculates the optical flow for the smallest image among the multi-resolution images created by the multi-resolution image creation process 871, and uses this calculation result again for the image of the next higher resolution. This is a process for calculating an optical flow, and this series of processes is repeated until the largest image is reached.
- the feature point extraction process performed by the feature point extraction unit 121 and the optical flow calculation process performed by the optical flow calculation unit 122 shown in FIG. 2 and the like are performed by SIMD calculation using the multi-core processor 800. Processing results can be obtained by parallel processing. Note that the feature point extraction processing and the optical flow calculation processing shown in FIG. 42 and the like are examples, and the multicore processor 800 uses other processing configured by various filter processing, threshold processing, and the like for the images constituting the moving image. SIMD calculation may be performed.
- FIG. 43 shows a filtering process using the Sobel filter 830 for image data (image data corresponding to one frame constituting a moving image captured by the camera) stored in the main memory 781 according to the embodiment of the present invention. It is a figure which shows roughly the data structure in the case of performing, and the flow of a process. Note that the image data stored in the main memory 781 shown in the figure is shown in a simplified manner with 32 horizontal pixels.
- the Sobel filter 830 is a 3 ⁇ 3 edge extraction filter. As shown in the figure, filtering processing using a Sobel filter 830 is performed on the image data stored in the main memory 781, and the result of this filtering processing is output. In this example, an example will be described in which four filter results are obtained at a time using SIMD computation.
- FIG. 44 is a diagram schematically showing a data flow when SIMD computation is performed on the image data stored in the main memory 781 according to the embodiment of the present invention using the Sobel filter 830.
- a predetermined number of lines for example, 3 lines
- DMA Direct Memory Access
- a predetermined number of lines obtained by shifting down each line DMA-transferred to the first buffer 831 by one is DMA-transferred to the second buffer 832.
- DMA Direct Memory Access
- FIG. 45 schematically shows a vector creation method for creating nine vectors from image data stored in the first buffer 831 when filtering is performed using the Sobel filter 830 according to the embodiment of the present invention.
- FIG. 44 after DMA transfer is performed, nine vectors are created from the image data stored in the first buffer 831.
- vector data 841 is created from four data from the left corner in one line of the image data stored in the first buffer 831, and the vector data is obtained by shifting the four data to the right by one.
- the vector data 843 is created from four data obtained by shifting the four data by one to the right.
- the vector data 844 to 849 are generated from the four data in the second and third lines.
- FIG. 46 is a diagram schematically showing a vector operation method for performing vector operation using SIMD instructions for the vector data 841 to 849 when performing filtering using the Sobel filter 830 according to the embodiment of the present invention. . Specifically, SIMD operations are sequentially performed on the vector data 841 to 843 to obtain the vector A. In this SIMD operation, first, a SIMD operation of ““ ⁇ 1 ” ⁇ “ vector data 841 ”” is executed. Subsequently, a SIMD operation ““ 0 ” ⁇ “ vector data 842 ”” is executed, and a SIMD operation ““ 1 ” ⁇ “ vector data 843 ”” is executed.
- vector data 842 can be omitted because the calculation result is determined to be “0”.
- vector data 843 can be omitted because the calculation result is determined to be the same value as “vector data 843”.
- the addition process of the calculation result ““ ⁇ 1 ” ⁇ “ vector data 841 ” and the calculation result ““ 0 ” ⁇ “ vector data 842 ”” is executed by SIMD calculation.
- an addition process between the result of the addition process and the calculation result of ““ 1 ” ⁇ “ vector data 843 ”” is executed by SIMD calculation.
- the calculation of the data structure of “vector data 1” ⁇ “vector data 2” + “vector data 3” can be executed by SIMD calculation.
- the SIMD operation is performed on the vector data 844 to 846 to determine the vector B
- the SIMD operation is performed on the vector data 847 to 849
- the vector C is determined.
- the SIMD operation is performed on the vectors A to C obtained by the SIMD operation, and the vector D is obtained.
- the results corresponding to the number of elements of the vector can be collectively obtained.
- the same processing is repeatedly executed in the image data stored in the first buffer 831 shown in FIG. 44 while shifting the position of the data to be extracted by one to the right. D is calculated sequentially.
- the processing result is DMA transferred to the main memory 781.
- high speed can be realized by performing most of the processing of feature point extraction and optical flow calculation by SIMD calculation.
- FIG. 47 is a diagram schematically showing the flow of camera work parameter calculation processing in the embodiment of the present invention in time series.
- decoding and analysis processing for a moving image can be performed in parallel. For this reason, it is possible to shorten the analysis time of 1 frame which comprises a moving image rather than decoding time.
- t1 indicates the time required for the decoding processing of one frame constituting the moving image by the control processor core 801
- t2 indicates that the arithmetic processor cores (# 1) 811 to (# 8) 818 constitute the moving image.
- the time required for the feature point extraction process for one frame to be performed t3 indicates the time required for the optical flow calculation process for one frame constituting the moving image by the arithmetic processor cores (# 1) 811 to (# 8) 818
- t4 Indicates the time required for the camera processor detection processing of one frame constituting the moving image by the control processor core 801.
- t5 indicates the time required for the camera work detection process for one frame in which the control processor core 801 and the arithmetic processor cores (# 1) 811 to (# 8) 818 constitute a moving image.
- t6 indicates the time required for the process for the control processor core 801 to manage the arithmetic processor cores (# 1) 811 to (# 8) 818.
- t1 can be “25.0 ms”
- t2 can be “7.9 ms”
- t3 can be “6.7 ms”
- t4 can be “1.2 ms”
- t5 can be “15.8 ms”. .
- FIG. 48A is a top view schematically showing a Blu-ray Disc (registered trademark) 880 which is an example of a recording medium
- FIG. 48B is a diagram recorded on the Blu-ray Disc 880. It is a figure which shows typically each data 881 thru
- moving image content 882 that is a moving image captured by a camera or the like
- metadata obtained by analyzing the moving image content 882 for example, (b in FIG.
- FIG. 48 (c) is a diagram schematically showing an internal configuration of a Blu-ray player (Blu-ray Disc Player) 890 capable of playing the Blu-ray Disc 880.
- the Blu-ray player 890 capable of playing a Blu-ray disc has a Java (registered trademark) VM (Java (registered trademark) virtual machine) and a library 893, as well as a CPU 891 and an OS 892, so Java (registered) (Trademark) program. Therefore, by installing the Blu-ray disc 880 in the Blu-ray player 890, the Blu-ray player 890 can load and execute the Java (registered trademark) program 881.
- Java registered trademark
- the Blu-ray player 890 reproduces the moving image content 882, it is possible to reproduce the moving image according to the embodiment of the present invention using the metadata 884. That is, it is possible to realize the moving image reproduction according to the embodiment of the present invention on all Blu-ray players without using dedicated PC software or the like.
- the example of correcting the affine transformation parameters in order to improve the quality of the composite image has been described.
- other information related to the moving image may be corrected to improve the quality of the composite image.
- exposure correction parameters in each frame are acquired as metadata and stored in association with the frames.
- a frame subjected to exposure correction based on the metadata is extracted, and the frame subjected to the exposure correction is subjected to reverse correction by image processing, and before the exposure correction is performed. Correct to the frame.
- the white balance function it is possible to correct a moving image that has been imaged by the auto white balance function. For example, when this moving image is synthesized and reproduced, the white balance is reversely corrected. Thereby, since the white balance of each frame which comprises a moving image differs, it can prevent that the whole synthesized image turns into a blurred image.
- a moving image is an aggregate of a plurality of frames over a predetermined period of time, and normal reproduction is prioritized, and thus the imaging conditions of each frame are often different. For this reason, it is possible to create one high-quality panoramic composite image by reversely correcting these image quality. Note that when the exposure correction is canceled by image processing, for example, there is a possibility that the image may be completely white. Therefore, it is preferable to perform correction according to the content of the moving image.
- each image corresponding to a frame before the currently displayed image is displayed while being combined with the current image.
- the background and the like that were shot in at least a part of the time zone. For this reason, for example, if you want to see the background taken at least in part of the time period again, the background etc. can be displayed at the same time as the currently displayed image without rewinding or searching. Can see.
- the contents of the moving image can be easily grasped. Further, the relative relationship between the moving images can be easily grasped, and the viewer can easily recognize the spatial spread of the moving images.
- an appreciation method for completing a panoramic image while reproducing a plurality of moving images can be provided, so that the viewer can appreciate the moving images with great interest.
- the current image since the images in the state before being stored in the image memory 220 can be sequentially displayed, a relatively beautiful image can be displayed.
- the affine transformation parameters are corrected or the composite image is divided to prevent the video being composited from breaking down. At the same time, the quality of the composite image formed in a panoramic shape can be improved.
- an affine transformation parameter is calculated at the time of reproduction, and the calculated affine transformation parameter is used. May be displayed for playback.
- the affine transformation parameters by SIMD calculation using a multi-core processor, it is possible to calculate one frame of affine transformation parameters within the processing time of one frame decoding. As a result, even when playing back a moving image for which no affine transformation parameter is calculated, it is possible to play the moving image while calculating the affine transformation parameter. it can.
- a plurality of moving image files are stored in the moving image storage unit 200, and an affine transformation parameter corresponding to the moving image is associated with the corresponding moving image and frame as a metadata file as a metadata storage unit.
- the moving image and the affine transformation parameter corresponding to the moving image are associated and recorded as a moving image file in the moving image storage unit, and each piece of information is extracted from the moving image file for use during reproduction. You may do it.
- the synthesized image synthesized by the image synthesizing unit 190 may be recorded on a recording medium or the like and used for other reproduction display.
- a composite image can be stored as a representative image in a metadata file of the metadata storage unit 210, and the composite image can be used for retrieval during reproduction.
- the composite image instead of repeating the composite image creation process for all the frames constituting the moving image file, the composite image for at least a certain number of frames in the frames constituting the moving image file is used.
- the creation process may be repeated to create a composite image, and this composite image may be used as a representative image.
- the composite image creation process is repeated from the first frame constituting the moving image to create the composite image.
- the composite image is directed from the last frame toward the first frame.
- the creation process may be repeated to create a composite image.
- the old image on the time axis is sequentially superimposed on the new image and synthesized.
- the scene immediately before the division may stand out in the vicinity of the division. Therefore, for example, when creating a composite image used as a representative image, it is not necessary to create it with all the frames. You may make it exclude.
- the embodiment of the present invention can be applied to a case where a moving image index image is extracted based on a duplication rate on an imaging space that is a space corresponding to a composite image to be formed. That is, an image after correction or division can be a target for determining the overlap rate.
- the example in which the composite image corresponding to the frame before the current frame is displayed has been described.
- the composite image is sequentially deleted as time passes. Also good. In this case, an effect may be given such that an afterimage is left and erased.
- the image corresponding to the current frame is displayed in color, and the composite image corresponding to the frame before the current frame is provided with an effect of changing from color display to sepia over time. It may be.
- the synthesized image of the scene before the division may be gradually blurred. As a result, it is possible to smoothly switch scenes for the composite image. In this case, the head image of the new scene may be returned to a certain fixed position on the screen.
- the image processing apparatus that displays the image combined by the image combining unit on the display unit has been described as an example.
- the image combined by the image combining unit is displayed in another image display device.
- the embodiment of the present invention can be applied to an image processing apparatus provided with an image output means for outputting image information for display.
- the embodiment of the present invention can be applied to a moving image reproducing device capable of reproducing a moving image, an imaging device such as a digital video camera capable of reproducing a captured moving image, and the like.
- the moving image captured by the camera has been described.
- the edited moving image or the moving image synthesized by the animation when the moving image captured by the camera is edited is also described.
- the embodiment of the present invention can be applied.
- an example in which a part or all of a history image is displayed has been described.
- only a plurality of converted current images may be displayed. That is, only a plurality of current images last stored in the image memory may be sequentially displayed.
- the embodiment of the present invention can also be applied to a case where the size of the moving object is relatively large with respect to the area of the image constituting the moving image.
- the motion of the train is calculated when the above affine transformation parameters are calculated.
- conversion information for converting the images constituting the moving image it is possible to calculate and use motion information regarding the relative motion amount between the camera and the subject at the time of imaging.
- the moving image storage means corresponds to, for example, the moving image storage unit 200.
- the conversion information storage unit corresponds to the metadata storage unit 210, for example.
- An image holding unit corresponds to the image memory 220, for example.
- the feature amount extraction unit corresponds to the feature amount extraction unit 150, for example.
- An image conversion unit corresponds to the image conversion unit 180, for example.
- the determination unit corresponds to the composite image division determination unit 170, for example.
- An image composition unit corresponds to the image composition unit 190, for example.
- moving image storage means corresponds to the moving image storage unit 200, for example.
- the conversion information storage unit corresponds to the metadata storage unit 210, for example.
- An image holding unit corresponds to the image memory 220, for example.
- the determination unit corresponds to the composite image division determination unit 170, for example.
- An image conversion unit corresponds to the image conversion unit 180, for example.
- An image composition unit corresponds to the image composition unit 190, for example.
- moving image storage means corresponds to the moving image storage unit 200, for example.
- the conversion information storage unit corresponds to the metadata storage unit 210, for example.
- An image holding unit corresponds to the image memory 220, for example.
- the feature amount extraction unit corresponds to the feature amount extraction unit 150, for example.
- the conversion information correction unit corresponds to the conversion information correction unit 160, for example.
- An image conversion unit corresponds to the image conversion unit 180, for example.
- An image composition unit corresponds to the image composition unit 190, for example.
- the feature amount extraction means corresponds to, for example, the feature amount extraction unit 150.
- the determination unit corresponds to the composite image division determination unit 170, for example.
- An image composition unit corresponds to the image conversion unit 180 and the image composition unit 190, for example.
- the moving image storage means corresponds to the moving image storage unit 200, for example.
- the conversion information storage unit corresponds to the metadata storage unit 210, for example.
- An image holding unit corresponds to the image memory 220, for example.
- the feature amount extraction unit corresponds to the feature amount extraction unit 150, for example.
- An image conversion unit corresponds to the image conversion unit 180, for example.
- the determination unit corresponds to the composite image division determination unit 170, for example.
- An image composition unit corresponds to the image composition unit 190, for example.
- the display control unit corresponds to the display control unit 250, for example.
- the image conversion procedure corresponds to, for example, step S929.
- the determination procedure corresponds to, for example, step S950.
- the image composition procedure corresponds to, for example, step S930.
- the processing procedure described in the embodiment of the present invention may be regarded as a method having a series of these procedures, and a program for causing a computer to execute these series of procedures or a recording medium storing the program May be taken as
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computing Systems (AREA)
- Image Processing (AREA)
- Studio Devices (AREA)
- Television Signal Processing For Recording (AREA)
- Editing Of Facsimile Originals (AREA)
Abstract
Description
続いて、入力された動画ファイルを構成するフレームの中で、現フレームが最後のフレームであるか否かが判断される(ステップS937)。現フレームが最後のフレームではない場合には(ステップS937)、ステップS923に戻り、動画の合成再生処理を繰り返す(ステップS923乃至S936)。一方、現フレームが最後のフレームである場合には(ステップS937)、確保されているワークバッファを解放して(ステップS938)、動画再生処理を終了する。
膨張処理868は、ハリスコーナー抽出処理867により算出されたハリスコーナーのスコアで構成された画像に対してぼかし処理を行う処理である。
Claims (16)
- 撮像装置により撮像された撮像動画を記憶する動画記憶手段と、
前記撮像動画を構成する第1の撮像画像を基準として前記撮像動画の時間軸において前記第1の撮像画像よりも後に位置する第2の撮像画像を変換するための変換情報を前記第2の撮像画像毎に記憶する変換情報記憶手段と、
前記時間軸において前記第2の撮像画像よりも前に位置する各撮像画像を含む履歴画像を保持する画像保持手段と、
前記第2の撮像画像に関する特徴量を抽出する特徴量抽出手段と、
前記変換情報に基づいて前記第2の撮像画像を変換する画像変換手段と、
前記抽出された特徴量に基づいて前記第2の撮像画像および前記履歴画像の合成の要否を判定する判定手段と、
前記判定手段により前記第2の撮像画像および前記履歴画像を合成すると判定された場合には前記画像保持手段に保持されている前記履歴画像に前記変換された第2の撮像画像を合成して新たな履歴画像として前記画像保持手段に保持させ、前記判定手段により前記第2の撮像画像および前記履歴画像を合成しないと判定された場合には前記画像保持手段に保持されている前記履歴画像に前記第2の撮像画像を合成させずに前記第2の撮像画像を前記画像保持手段に保持させる画像合成手段とを具備することを特徴とする画像処理装置。 - 前記画像合成手段は、前記判定手段により前記第2の撮像画像および前記履歴画像を合成しないと判定された場合には前記画像保持手段に保持されている前記履歴画像を消去して前記第2の撮像画像を新たな履歴画像として前記画像保持手段に保持させることを特徴とする請求項1記載の画像処理装置。
- 前記画像合成手段は、前記判定手段により前記第2の撮像画像および前記履歴画像を合成しないと判定された場合には前記画像保持手段における前記履歴画像の配置位置とは異なる位置に前記第2の撮像画像を配置して新たな履歴画像として前記画像保持手段に保持させることを特徴とする請求項1記載の画像処理装置。
- 前記特徴量抽出手段は、前記撮像動画を構成する各画像における一定の変化を検出することにより前記特徴量を抽出し、
前記判定手段は、前記抽出された特徴量と所定の閾値とを比較することにより前記第2の撮像画像がシーンチェンジ点に該当するか否かを判定して前記第2の撮像画像がシーンチェンジ点に該当すると判定された場合には前記第2の撮像画像および前記履歴画像を合成しないと判定することを特徴とする請求項1記載の画像処理装置。 - 前記変換情報は、拡大縮小と並進移動と回転とに関する要素を含み、
前記特徴量抽出手段は、前記第2の撮像画像に係る前記変換情報に基づいて並進成分と回転成分と拡大縮小成分とを抽出し、
前記判定手段は、前記抽出された前記並進成分と前記回転成分と前記拡大縮小成分との少なくとも1つと所定の閾値とを比較することにより前記合成の要否を判定することを特徴とする請求項1記載の画像処理装置。 - 前記特徴量抽出手段は、前記時間軸において前記第2の撮像画像よりも所定数前に位置する撮像画像から前記第2の撮像画像までの各撮像画像に係る変換情報に基づいて並進成分と回転成分と拡大縮小成分とを抽出して当該抽出された並進成分と回転成分と拡大縮小成分との平均値をそれぞれ算出し、
前記判定手段は、前記算出された前記並進成分と前記回転成分と前記拡大縮小成分との平均値の少なくとも1つと所定の閾値とを比較することにより前記合成の要否を判定することを特徴とする請求項5記載の画像処理装置。 - 前記特徴量抽出手段は、前記時間軸において前記第2の撮像画像よりも所定数前に位置する撮像画像から前記第2の撮像画像までの各撮像画像に係る変換情報に基づいて並進成分と回転成分と拡大縮小成分とを抽出して当該抽出された並進成分と回転成分と拡大縮小成分との積算値をそれぞれ算出し、
前記判定手段は、前記算出された前記並進成分と前記回転成分と前記拡大縮小成分との積算値の少なくとも1つと所定の閾値とを比較することにより前記合成の要否を判定することを特徴とする請求項5記載の画像処理装置。 - 撮像装置により撮像された撮像動画を記憶する動画記憶手段と、
前記撮像動画を構成する第1の撮像画像を基準として前記撮像動画の時間軸において前記第1の撮像画像よりも後に位置する第2の撮像画像を変換するための変換情報を前記第2の撮像画像毎に記憶する変換情報記憶手段と、
前記時間軸において前記第2の撮像画像よりも前に位置する各撮像画像を含む履歴画像を保持する画像保持手段と、
前記履歴画像が前記画像保持手段に保持されてからの経過時間に基づいて前記第2の撮像画像および前記履歴画像の合成の要否を判定する判定手段と、
前記変換情報に基づいて前記第2の撮像画像を変換する画像変換手段と、
前記判定手段により前記第2の撮像画像および前記履歴画像を合成すると判定された場合には前記画像保持手段に保持されている前記履歴画像に前記変換された第2の撮像画像を合成して新たな履歴画像として前記画像保持手段に保持させ、前記判定手段により前記第2の撮像画像および前記履歴画像を合成しないと判定された場合には前記画像保持手段に保持されている前記履歴画像に前記第2の撮像画像を合成させずに前記第2の撮像画像を前記画像保持手段に保持させる画像合成手段とを具備することを特徴とする画像処理装置。 - 撮像装置により撮像された撮像動画を記憶する動画記憶手段と、
前記撮像動画を構成する第1の撮像画像を基準として前記撮像動画の時間軸において前記第1の撮像画像よりも後に位置する第2の撮像画像を変換するための変換情報を前記第2の撮像画像毎に記憶する変換情報記憶手段と、
前記時間軸において前記第2の撮像画像よりも前に位置する各撮像画像を含む履歴画像を保持する画像保持手段と、
前記第2の撮像画像に関する特徴量を抽出する特徴量抽出手段と、
前記抽出された特徴量に基づいて前記変換情報を補正する変換情報補正手段と、
前記補正された前記変換情報に基づいて前記第2の撮像画像を変換する画像変換手段と、
前記画像保持手段に保持されている前記履歴画像に前記変換された第2の撮像画像を合成して新たな履歴画像として前記画像保持手段に保持させる画像合成手段とを具備することを特徴とする画像処理装置。 - 前記変換情報補正手段は、前記抽出された特徴量と所定の閾値とを比較することにより前記変換情報の補正の要否を判定して前記変換情報の補正が必要と判定された場合には前記変換情報を補正し、
前記画像変換手段は、前記変換情報が補正された場合には当該補正された変換情報に基づいて前記第2の撮像画像を変換して前記変換情報が補正されない場合には前記変換情報に基づいて前記第2の撮像画像を変換することを特徴とする請求項9記載の画像処理装置。 - 前記変換情報補正手段は、前記変換情報の補正が必要と判定された場合には前記変換情報を所定の変換情報に補正することを特徴とする請求項10記載の画像処理装置。
- 前記変換情報補正手段は、前記変換情報の補正が必要と判定された場合には前記時間軸において前記第2の撮像画像の前方向に位置する撮像画像に係る変換情報と前記第2の撮像画像の後方向に位置する撮像画像に係る変換情報とに基づいて前記変換情報を補正することを特徴とする請求項10記載の画像処理装置。
- 撮像装置により撮像された撮像動画を構成する撮像画像に関する特徴量を抽出する特徴量抽出手段と、
前記抽出された特徴量に基づいて合成画像の分割の要否を判定する判定手段と、
前記判定手段により合成画像を分割しないと判定された場合には撮像時における前記撮像装置の動き情報に基づいて既に形成されている合成画像に前記撮像画像を合成し、前記判定手段により合成画像を分割すると判定された場合には前記動き情報に基づいて既に形成されている合成画像とは異なる新たな合成画像を作成する画像合成手段とを具備することを特徴とする画像処理装置。 - 撮像装置により撮像された撮像動画を記憶する動画記憶手段と、
前記撮像動画を構成する第1の撮像画像を基準として前記撮像動画の時間軸において前記第1の撮像画像よりも後に位置する第2の撮像画像を変換するための変換情報を前記第2の撮像画像毎に記憶する変換情報記憶手段と、
前記時間軸において前記第2の撮像画像よりも前に位置する各撮像画像を含む履歴画像を保持する画像保持手段と、
前記第2の撮像画像に関する特徴量を抽出する特徴量抽出手段と、
前記変換情報に基づいて前記第2の撮像画像を変換する画像変換手段と、
前記抽出された特徴量に基づいて前記第2の撮像画像および前記履歴画像の合成の要否を判定する判定手段と、
前記判定手段により前記第2の撮像画像および前記履歴画像を合成すると判定された場合には前記画像保持手段に保持されている前記履歴画像に前記変換された第2の撮像画像を合成して新たな履歴画像として前記画像保持手段に保持させ、前記判定手段により前記第2の撮像画像および前記履歴画像を合成しないと判定された場合には前記画像保持手段に保持されている前記履歴画像に前記第2の撮像画像を合成させずに前記第2の撮像画像を前記画像保持手段に保持させる画像合成手段と、
前記画像保持手段に保持された新たな履歴画像または前記第2の撮像画像を表示手段に順次表示させる表示制御手段とを具備することを特徴とする動画再生装置。 - 撮像装置により撮像された撮像動画を記憶する動画記憶手段と、前記撮像動画を構成する第1の撮像画像を基準として前記撮像動画の時間軸において前記第1の撮像画像よりも後に位置する第2の撮像画像を変換するための変換情報を前記第2の撮像画像毎に記憶する変換情報記憶手段と、前記時間軸において前記第2の撮像画像よりも前に位置する各撮像画像を含む履歴画像を保持する画像保持手段と、前記第2の撮像画像に関する特徴量を抽出する特徴量抽出手段とを備える画像処理装置における画像処理方法であって、
前記変換情報に基づいて前記第2の撮像画像を変換する画像変換手順と、
前記抽出された特徴量に基づいて前記第2の撮像画像および前記履歴画像の合成の要否を判定する判定手順と、
前記判定手順で前記第2の撮像画像および前記履歴画像を合成すると判定された場合には前記画像保持手段に保持されている前記履歴画像に前記変換された第2の撮像画像を合成して新たな履歴画像として前記画像保持手段に保持させ、前記判定手順で前記第2の撮像画像および前記履歴画像を合成しないと判定された場合には前記画像保持手段に保持されている前記履歴画像に前記第2の撮像画像を合成させずに前記第2の撮像画像を前記画像保持手段に保持させる画像合成手順とを具備することを特徴とする画像処理方法。 - 撮像装置により撮像された撮像動画を記憶する動画記憶手段と、前記撮像動画を構成する第1の撮像画像を基準として前記撮像動画の時間軸において前記第1の撮像画像よりも後に位置する第2の撮像画像を変換するための変換情報を前記第2の撮像画像毎に記憶する変換情報記憶手段と、前記時間軸において前記第2の撮像画像よりも前に位置する各撮像画像を含む履歴画像を保持する画像保持手段と、前記第2の撮像画像に関する特徴量を抽出する特徴量抽出手段とを備える画像処理装置において、
前記変換情報に基づいて前記第2の撮像画像を変換する画像変換手順と、
前記抽出された特徴量に基づいて前記第2の撮像画像および前記履歴画像の合成の要否を判定する判定手順と、
前記判定手順で前記第2の撮像画像および前記履歴画像を合成すると判定された場合には前記画像保持手段に保持されている前記履歴画像に前記変換された第2の撮像画像を合成して新たな履歴画像として前記画像保持手段に保持させ、前記判定手順で前記第2の撮像画像および前記履歴画像を合成しないと判定された場合には前記画像保持手段に保持されている前記履歴画像に前記第2の撮像画像を合成させずに前記第2の撮像画像を前記画像保持手段に保持させる画像合成手順とをコンピュータに実行させることを特徴とするプログラム。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP08865920A EP2180699A4 (en) | 2007-12-26 | 2008-12-10 | PICTURE PROCESSING APPARATUS, ANIMATION RESTORING UNIT AND PROCESSING METHOD AND PROGRAM FOR THE PROCESSING DEVICE AND THE DEVICE |
US12/524,777 US8515256B2 (en) | 2007-12-26 | 2008-12-10 | Image processing apparatus, moving image reproducing apparatus, and processing method and program therefor |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2007335344A JP4697221B2 (ja) | 2007-12-26 | 2007-12-26 | 画像処理装置、動画再生装置、これらにおける処理方法およびプログラム |
JP2007-335344 | 2007-12-26 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2009081732A1 true WO2009081732A1 (ja) | 2009-07-02 |
Family
ID=40801040
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2008/072402 WO2009081732A1 (ja) | 2007-12-26 | 2008-12-10 | 画像処理装置、動画再生装置、これらにおける処理方法およびプログラム |
Country Status (6)
Country | Link |
---|---|
US (1) | US8515256B2 (ja) |
EP (1) | EP2180699A4 (ja) |
JP (1) | JP4697221B2 (ja) |
KR (1) | KR20100103776A (ja) |
CN (1) | CN101611629A (ja) |
WO (1) | WO2009081732A1 (ja) |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5092722B2 (ja) | 2007-12-07 | 2012-12-05 | ソニー株式会社 | 画像処理装置、画像処理方法およびプログラム |
JP2009151896A (ja) | 2007-12-21 | 2009-07-09 | Sony Corp | 画像処理装置、動画再生装置、これらにおける処理方法およびプログラム |
JP2011139282A (ja) * | 2009-12-28 | 2011-07-14 | Sony Corp | 画像処理装置、撮像装置、画像処理方法およびプログラム |
US8989499B2 (en) * | 2010-10-20 | 2015-03-24 | Comcast Cable Communications, Llc | Detection of transitions between text and non-text frames in a video stream |
CN104661825A (zh) | 2012-06-15 | 2015-05-27 | 海德堡印刷机械股份公司 | 用于将印刷液体间接施加到承印材料上的方法 |
US20190132614A1 (en) * | 2016-04-28 | 2019-05-02 | Sony Corporation | Information processing terminal apparatus and distribution apparatus |
US10607108B2 (en) | 2018-04-30 | 2020-03-31 | International Business Machines Corporation | Techniques for example-based affine registration |
CN109682296A (zh) * | 2018-09-21 | 2019-04-26 | 浙江华诚工程管理有限公司 | 一种建筑工程监理用材料尺寸检测系统 |
Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0962861A (ja) * | 1995-08-21 | 1997-03-07 | Matsushita Electric Ind Co Ltd | パノラマ映像装置 |
JPH11120353A (ja) * | 1997-10-16 | 1999-04-30 | Matsushita Electric Ind Co Ltd | 画像表示装置 |
JPH11134352A (ja) * | 1997-10-30 | 1999-05-21 | Hitachi Ltd | 動画像表示方法 |
JPH11289517A (ja) | 1998-04-02 | 1999-10-19 | Sony Corp | 再生装置及び再生方法 |
JP2002359798A (ja) * | 2001-05-31 | 2002-12-13 | Hitachi Medical Corp | パノラマ画像表示方法、ソフトウエア及び装置 |
JP2004222025A (ja) * | 2003-01-16 | 2004-08-05 | Matsushita Electric Ind Co Ltd | 電子カメラ |
JP2004266670A (ja) * | 2003-03-03 | 2004-09-24 | Sony Corp | 撮像装置及び方法、画像情報提供システム並びにプログラム |
WO2005024723A1 (ja) * | 2003-09-08 | 2005-03-17 | Nec Corporation | 画像合成システム、画像合成方法及びプログラム |
WO2005032125A1 (ja) * | 2003-09-26 | 2005-04-07 | Sharp Kabushiki Kaisha | パノラマ画像作成装置及びパノラマ画像撮影装置 |
JP2005217874A (ja) * | 2004-01-30 | 2005-08-11 | Konica Minolta Photo Imaging Inc | ファイル管理方法 |
JP2007183588A (ja) * | 2006-01-04 | 2007-07-19 | Mitsubishi Electric Research Laboratories Inc | グラフィック画像組をテレビシステム上に提示する方法及びグラフィック画像組を提示するテレビシステム |
JP2007206919A (ja) | 2006-02-01 | 2007-08-16 | Sony Corp | 表示制御装置および方法、プログラム、並びに記録媒体 |
Family Cites Families (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5657402A (en) * | 1991-11-01 | 1997-08-12 | Massachusetts Institute Of Technology | Method of creating a high resolution still image using a plurality of images and apparatus for practice of the method |
JP2989364B2 (ja) * | 1992-03-12 | 1999-12-13 | シャープ株式会社 | 画像処理装置及び画像処理方法 |
US5649032A (en) * | 1994-11-14 | 1997-07-15 | David Sarnoff Research Center, Inc. | System for automatically aligning images to form a mosaic image |
US5706416A (en) * | 1995-11-13 | 1998-01-06 | Massachusetts Institute Of Technology | Method and apparatus for relating and combining multiple images of the same scene or object(s) |
JP2001503895A (ja) * | 1996-11-15 | 2001-03-21 | サーノフ コーポレイション | ビデオ情報を効果的に表示、保存、およびアクセスするための方法および装置 |
US6895126B2 (en) * | 2000-10-06 | 2005-05-17 | Enrico Di Bernardo | System and method for creating, storing, and utilizing composite images of a geographic location |
WO2003084249A1 (en) * | 2002-03-27 | 2003-10-09 | The Trustees Of Columbia University In The City Of New York | Methods for summarizing video through mosaic-based shot and scene clustering |
JP4419499B2 (ja) * | 2003-01-07 | 2010-02-24 | セイコーエプソン株式会社 | 画像生成装置、画像ずれ量検出装置、画像生成方法、画像ずれ量検出方法、画像生成プログラムおよび画像ずれ量検出プログラム |
US7139006B2 (en) * | 2003-02-04 | 2006-11-21 | Mitsubishi Electric Research Laboratories, Inc | System and method for presenting and browsing images serially |
US20070103544A1 (en) * | 2004-08-26 | 2007-05-10 | Naofumi Nakazawa | Panorama image creation device and panorama image imaging device |
JP5260264B2 (ja) | 2008-03-25 | 2013-08-14 | 富士通株式会社 | ビーコン無線通信装置、ビーコン無線通信プログラム、およびビーコン無線通信方法 |
-
2007
- 2007-12-26 JP JP2007335344A patent/JP4697221B2/ja not_active Expired - Fee Related
-
2008
- 2008-12-10 KR KR1020097017680A patent/KR20100103776A/ko not_active Application Discontinuation
- 2008-12-10 CN CNA2008800018845A patent/CN101611629A/zh active Pending
- 2008-12-10 US US12/524,777 patent/US8515256B2/en active Active
- 2008-12-10 EP EP08865920A patent/EP2180699A4/en not_active Ceased
- 2008-12-10 WO PCT/JP2008/072402 patent/WO2009081732A1/ja active Application Filing
Patent Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0962861A (ja) * | 1995-08-21 | 1997-03-07 | Matsushita Electric Ind Co Ltd | パノラマ映像装置 |
JPH11120353A (ja) * | 1997-10-16 | 1999-04-30 | Matsushita Electric Ind Co Ltd | 画像表示装置 |
JPH11134352A (ja) * | 1997-10-30 | 1999-05-21 | Hitachi Ltd | 動画像表示方法 |
JPH11289517A (ja) | 1998-04-02 | 1999-10-19 | Sony Corp | 再生装置及び再生方法 |
JP2002359798A (ja) * | 2001-05-31 | 2002-12-13 | Hitachi Medical Corp | パノラマ画像表示方法、ソフトウエア及び装置 |
JP2004222025A (ja) * | 2003-01-16 | 2004-08-05 | Matsushita Electric Ind Co Ltd | 電子カメラ |
JP2004266670A (ja) * | 2003-03-03 | 2004-09-24 | Sony Corp | 撮像装置及び方法、画像情報提供システム並びにプログラム |
WO2005024723A1 (ja) * | 2003-09-08 | 2005-03-17 | Nec Corporation | 画像合成システム、画像合成方法及びプログラム |
WO2005032125A1 (ja) * | 2003-09-26 | 2005-04-07 | Sharp Kabushiki Kaisha | パノラマ画像作成装置及びパノラマ画像撮影装置 |
JP2005217874A (ja) * | 2004-01-30 | 2005-08-11 | Konica Minolta Photo Imaging Inc | ファイル管理方法 |
JP2007183588A (ja) * | 2006-01-04 | 2007-07-19 | Mitsubishi Electric Research Laboratories Inc | グラフィック画像組をテレビシステム上に提示する方法及びグラフィック画像組を提示するテレビシステム |
JP2007206919A (ja) | 2006-02-01 | 2007-08-16 | Sony Corp | 表示制御装置および方法、プログラム、並びに記録媒体 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2180699A4 |
Also Published As
Publication number | Publication date |
---|---|
CN101611629A (zh) | 2009-12-23 |
JP4697221B2 (ja) | 2011-06-08 |
US8515256B2 (en) | 2013-08-20 |
KR20100103776A (ko) | 2010-09-28 |
EP2180699A4 (en) | 2011-11-30 |
US20100135635A1 (en) | 2010-06-03 |
JP2009159330A (ja) | 2009-07-16 |
EP2180699A1 (en) | 2010-04-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8963951B2 (en) | Image processing apparatus, moving-image playing apparatus, and processing method and program therefor to allow browsing of a sequence of images | |
JP4678404B2 (ja) | 撮像装置、その制御方法およびプログラム | |
JP5223318B2 (ja) | 画像処理装置、画像処理方法およびプログラム | |
WO2009081806A1 (ja) | 画像処理装置、動画再生装置、これらにおける処理方法およびプログラム | |
JP5092722B2 (ja) | 画像処理装置、画像処理方法およびプログラム | |
US8421871B2 (en) | Method and apparatus for image pickup and image processing | |
JP4623199B2 (ja) | 画像処理装置、画像処理方法およびプログラム | |
US8754959B2 (en) | Image processing device, dynamic image reproduction device, and processing method and program in them | |
JP4623201B2 (ja) | 画像処理装置、画像処理方法およびプログラム | |
JP4623200B2 (ja) | 画像処理装置、画像処理方法およびプログラム | |
JP4697221B2 (ja) | 画像処理装置、動画再生装置、これらにおける処理方法およびプログラム | |
JP4692550B2 (ja) | 画像処理装置、その処理方法およびプログラム | |
JP5602818B2 (ja) | カメラ、画像再生装置、画像合成方法、およびプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 200880001884.5 Country of ref document: CN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 12524777 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2008865920 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1020097017680 Country of ref document: KR |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 08865920 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |