WO2021181965A1 - Dispositif de traitement d'image, procédé de traitement d'image, et programme - Google Patents
Dispositif de traitement d'image, procédé de traitement d'image, et programme Download PDFInfo
- Publication number
- WO2021181965A1 WO2021181965A1 PCT/JP2021/004160 JP2021004160W WO2021181965A1 WO 2021181965 A1 WO2021181965 A1 WO 2021181965A1 JP 2021004160 W JP2021004160 W JP 2021004160W WO 2021181965 A1 WO2021181965 A1 WO 2021181965A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- shaking
- image
- area
- processing
- unit
- Prior art date
Links
- 238000012545 processing Methods 0.000 title claims abstract description 305
- 238000003672 processing method Methods 0.000 title claims description 6
- 230000008859 change Effects 0.000 claims abstract description 120
- 238000000034 method Methods 0.000 claims description 131
- 230000008569 process Effects 0.000 claims description 115
- 230000010365 information processing Effects 0.000 claims description 21
- 238000010191 image analysis Methods 0.000 claims description 15
- 230000033001 locomotion Effects 0.000 claims description 13
- 230000000694 effects Effects 0.000 abstract description 35
- 238000012937 correction Methods 0.000 description 71
- 230000006641 stabilisation Effects 0.000 description 43
- 238000011105 stabilization Methods 0.000 description 43
- 238000003384 imaging method Methods 0.000 description 38
- 230000000875 corresponding effect Effects 0.000 description 25
- 230000003287 optical effect Effects 0.000 description 23
- 101001064774 Homo sapiens Peroxidasin-like protein Proteins 0.000 description 21
- 101001038163 Homo sapiens Sperm protamine P1 Proteins 0.000 description 21
- 102100031894 Peroxidasin-like protein Human genes 0.000 description 21
- 238000006243 chemical reaction Methods 0.000 description 20
- 230000006870 function Effects 0.000 description 20
- 230000015654 memory Effects 0.000 description 20
- 238000004891 communication Methods 0.000 description 17
- 101001090148 Homo sapiens Protamine-2 Proteins 0.000 description 16
- 102100034750 Protamine-2 Human genes 0.000 description 16
- 238000004364 calculation method Methods 0.000 description 16
- 238000000605 extraction Methods 0.000 description 16
- 238000005516 engineering process Methods 0.000 description 15
- 230000009467 reduction Effects 0.000 description 15
- 238000007781 pre-processing Methods 0.000 description 10
- 230000001133 acceleration Effects 0.000 description 8
- 230000005484 gravity Effects 0.000 description 8
- 230000007246 mechanism Effects 0.000 description 8
- 101150076088 MTD1 gene Proteins 0.000 description 7
- 230000015572 biosynthetic process Effects 0.000 description 7
- 238000003786 synthesis reaction Methods 0.000 description 7
- 238000003860 storage Methods 0.000 description 6
- 239000013256 coordination polymer Substances 0.000 description 5
- 238000005096 rolling process Methods 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 4
- 238000005070 sampling Methods 0.000 description 4
- 230000000740 bleeding effect Effects 0.000 description 3
- 230000008602 contraction Effects 0.000 description 3
- 238000001514 detection method Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 238000005401 electroluminescence Methods 0.000 description 3
- 238000002955 isolation Methods 0.000 description 3
- 239000000203 mixture Substances 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000002441 reversible effect Effects 0.000 description 3
- 239000004065 semiconductor Substances 0.000 description 3
- 230000001360 synchronised effect Effects 0.000 description 3
- 230000004075 alteration Effects 0.000 description 2
- 238000004590 computer program Methods 0.000 description 2
- 238000005520 cutting process Methods 0.000 description 2
- 230000007423 decrease Effects 0.000 description 2
- 230000008676 import Effects 0.000 description 2
- 108090000237 interleukin-24 Proteins 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 230000014759 maintenance of location Effects 0.000 description 2
- 239000011159 matrix material Substances 0.000 description 2
- 210000001747 pupil Anatomy 0.000 description 2
- 238000011946 reduction process Methods 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 238000000926 separation method Methods 0.000 description 2
- 230000002123 temporal effect Effects 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- 230000009466 transformation Effects 0.000 description 2
- 241000226585 Antennaria plantaginifolia Species 0.000 description 1
- 241000282326 Felis catus Species 0.000 description 1
- 101000911772 Homo sapiens Hsc70-interacting protein Proteins 0.000 description 1
- 101000710013 Homo sapiens Reversion-inducing cysteine-rich protein with Kazal motifs Proteins 0.000 description 1
- 101000661807 Homo sapiens Suppressor of tumorigenicity 14 protein Proteins 0.000 description 1
- 101000585359 Homo sapiens Suppressor of tumorigenicity 20 protein Proteins 0.000 description 1
- 241000283973 Oryctolagus cuniculus Species 0.000 description 1
- 102100029860 Suppressor of tumorigenicity 20 protein Human genes 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 238000012790 confirmation Methods 0.000 description 1
- 230000001276 controlling effect Effects 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 239000004615 ingredient Substances 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000001151 other effect Effects 0.000 description 1
- 238000002360 preparation method Methods 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
- 238000000844 transformation Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
Definitions
- This technology relates to image processing devices, image processing methods, and programs, and particularly to image processing related to image shake.
- Patent Document 1 discloses that vibration isolation processing is performed on moving image data related to captured images, and the influence of vibration isolation processing is removed on the moving image data after the vibration isolation processing.
- adding shaking to an image according to the content of a moving image is a method of producing an effect that expands the image expression.
- image effects such as when there are subjects in the image that are desired to be shaken and subjects that are not desired to be shaken, or when only a part of the image is to be shaken violently. Therefore, the present disclosure proposes a technique capable of generating an image in which shaking is changed according to the image content when adding or removing shaking in a moving image.
- the area setting unit for setting a plurality of areas in the image of the input moving image data and the shaking state appearing in the output moving image data are different for each area set by the area setting unit. It is provided with a shaking change unit that performs a shaking change process so as to be.
- the shaking change process is to change the shaking state by reducing the shaking generated in the moving image or adding the shaking. Then, by dividing one screen of the input video data into a plurality of areas and changing the shaking state for each area, for example, a region without shaking and a region with shaking can be formed in the screen, or the shaking can be formed. It is possible to form a small area and a large shaking area.
- the shaking changing unit adds shaking for one area set by the area setting unit and reduces shaking for the other area set by the area setting unit. It is conceivable to perform change processing. For example, it is possible to obtain a region that does not shake (or has less shake) and a region that shakes (or has a large shake) in the image.
- the shaking changing unit adds shaking for one area set by the area setting unit, and the other area set by the area setting unit is more than the one area. It is conceivable to perform a shaking change process that adds a small amount of shaking. For example, a region with less shaking and a region with large shaking can be obtained in the image.
- the shaking changing unit reduces shaking for one area set by the area setting unit, and is more than the other area set by the area setting unit. It is conceivable to perform a shaking change process that reduces shaking with a large reduction amount. For example, in the image, a region where the shaking is slightly reduced and a region where the shaking is significantly reduced can be obtained.
- the shaking changing unit performs shaking changing processing for adding or reducing shaking to one of one area and the other area set by the area setting unit. ..
- the shaking changing unit performs shaking changing processing for adding or reducing shaking to one of one area and the other area set by the area setting unit. ..
- the image processing device includes a user interface processing unit that detects operation information related to shaking change, and the area setting unit sets a plurality of areas based on the operation information detected by the user interface processing unit. It is possible to set it. For example, the user can specify a boundary or a range for dividing into a plurality of areas in one screen.
- the area setting unit sets a plurality of areas based on the image analysis of the input moving image data. For example, by subject recognition in an image, an area is set based on a pixel range in which a specific subject is captured.
- the shaking changing unit attaches each frame of the input moving image data to the celestial sphere model and rotates each frame using the shaking information corresponding to each frame.
- the shaking information for example, quaternion
- the shaking changing unit changes the shaking amount for each region by moving the coordinate points on the celestial sphere model in each frame. For example, by moving the coordinate points of the image on the celestial sphere model, the enlargement or reduction of the image fluctuates from frame to frame, thereby realizing partial shaking.
- the shaking changing unit moves the coordinate points by setting the coordinate points of the pixels of the boundary region between one region and the other region set by the region setting unit. It is conceivable that it is a process of moving. For example, by expanding and contracting the boundary region of a plurality of regions in the image, the shaking of one region is increased.
- the image processing apparatus sets an area setting process for setting a plurality of areas in the image of the input moving image data, and the shaking state appearing in the output moving image data is an area set in the area setting process.
- a shaking change process that makes the state different for each is performed. This makes it possible to add or remove shake in each area of the image.
- the program according to the present technology is a program that causes an information processing apparatus to execute a process corresponding to such an image processing method. As a result, the image processing of the present disclosure can be executed by various information processing devices.
- hake refers to the interframe shake of the images that make up a moving image. It broadly refers to vibration components (image fluctuations between frames) that occur between frames, such as shaking caused by camera shake in an image captured by a so-called image pickup device, or shaking intentionally added by image processing. It shall be.
- Interframe shake modification refers to changing the state of shaking in an image, such as reducing the shaking occurring in the image or adding shaking to the image.
- This "shake change” shall include the following “interframe shake reduction” and “interframe shake addition”.
- Hande removal refers to eliminating or reducing the shaking that occurs in the image due to camera shake (total removal of shaking) or reduction (partial removal of shaking). That is, it means adjusting so as to reduce the shaking based on the shaking information at the time of imaging.
- image stabilization performed in the image pickup apparatus is to remove the shaking.
- Adding shaking means adding shaking to the image. This includes adding shaking to an image without shaking, and adding shaking to an image with shaking so as to further increase the shaking.
- sway removal including partial sway removal
- sway addition are processes for changing the sway state or obtaining an image with sway as a result, and therefore can be said to be a process for adding a sway effect. ..
- the shaking effect it is assumed that the image is intentionally shaken in order to give power to the moving image scene.
- Hande information during imaging is information related to shaking when an image is taken by an image pickup device, such as motion detection information of the image pickup device, information that can be calculated from the detection information, and posture information indicating the posture of the image pickup device.
- Information such as shift and rotation as the movement of the image pickup device is applicable.
- the quarternion (QD) and the IMU data are given as specific examples of the "shaking information at the time of imaging", but there are other examples such as shift / rotation information, and the present invention is not particularly limited.
- FIG. 1A shows an example of an image source VS and an image processing device (TDx, TDy) that acquires a moving image file MF from the image source VS.
- the moving image file MF includes image data (that is, moving image data) and audio data constituting the moving image. However, there may be an audio file separate from the video file so that it can be synchronized.
- the moving image data also includes a plurality of continuously shot still image data.
- the image processing device TDx is a device that temporarily performs shaking change processing on the moving image data acquired from the image source VS.
- the image processing device TDy is a device that secondarily performs the shaking change processing on the moving image data that has already been subjected to the shaking change processing by another image processing device.
- an image pickup device 1 As the image source VS, an image pickup device 1, a server 4, a recording medium 5, and the like are assumed.
- a mobile terminal 2 such as a smartphone, a personal computer 3 and the like are assumed.
- various devices such as an image editing dedicated device, a cloud server, a television device, and a video recording / playback device are assumed as image processing devices TDx and TDy. These devices can function as any of the image processing devices TDx and TDy.
- the image pickup device 1 as an image source VS is a digital camera or the like capable of performing video imaging, and transfers the video file MF obtained by video imaging to a mobile terminal 2 or a personal computer 3 via wired communication or wireless communication.
- the server 4 may be a local server, a network server, a cloud server, or the like, but refers to a device capable of providing a moving image file MF captured by the image pickup device 1. It is conceivable that the server 4 transfers the moving image file MF to the mobile terminal 2 or the personal computer 3 via some kind of transmission path.
- the recording medium 5 may be a solid-state memory such as a memory card, a disk-shaped recording medium such as an optical disk, or a tape-shaped recording medium such as a magnetic tape, but removable recording in which the moving image file MF captured by the imaging device 1 is recorded. Pointing to the medium. It is conceivable that the moving image file MF read from the recording medium 5 is read by the mobile terminal 2 or the personal computer 3.
- the mobile terminal 2 and the personal computer 3 as the image processing devices TDx and TDy are capable of performing image processing on the moving image file MF acquired from the above image source VS.
- the image processing referred to here includes shaking change processing (shaking addition and shaking removal).
- the shaking change processing is performed, for example, for each frame of moving image data, after the processing of pasting to the celestial sphere model, which will be described later, is performed, and then the rotation is performed using the attitude information corresponding to the frame, or the coordinates of the pixels on the celestial sphere model. It can be done by moving points.
- a certain mobile terminal 2 or personal computer 3 may serve as an image source VS for another mobile terminal 2 or personal computer 3 that functions as an image processing device TDx or TDy.
- FIG. 1B shows an image pickup device 1 and a mobile terminal 2 as one device that can function as both an image source VS and an image processing device TDx.
- a microcomputer or the like inside the image pickup apparatus 1 performs the shaking change processing. That is, the image pickup apparatus 1 can output the image as the result of the image processing in which the shaking is removed or the shaking is added by performing the shaking changing process on the moving image file MF generated by the imaging.
- the mobile terminal 2 since it can be an image source VS by having an imaging function, the image processing result obtained by removing or adding shaking by performing the above-mentioned shaking change processing on the moving image file MF generated by imaging.
- the image can be output as.
- various other devices that can serve as an image source and an image processing device can be considered.
- the image processing device TDx of the embodiment the device functioning as the TDy, and the image source VS are various. And another image processing device TDy will be described as separate devices.
- FIG. 2 shows a state of information transmission in the image processing device TDy of the image source VS and the image processing device TDx.
- the moving image data VD1 and the metadata MTD1 are transmitted from the image source VS to the image processing device TDx via wired communication, wireless communication, or a recording medium.
- the moving image data VD1 and the metadata MTD1 are information transmitted as, for example, a moving image file MF.
- the metadata MTD1 may include a coordinate conversion parameter HP as information on shaking removal at the time of imaging performed, for example, as image stabilization.
- the image processing device TDx can perform various processes by receiving the moving image data VD1, the metadata MTD1, and the coordinate conversion parameter HP. For example, the image processing device TDx can perform the shaking change processing on the moving image data VD1 by using the shaking information at the time of imaging included in the metadata MTD1. Further, for example, the image processing device TDx can cancel the shaking removal applied to the moving image data VD1 at the time of imaging by using the coordinate conversion parameter HP included in the metadata MTD1.
- the image processing device TDx may perform a process of associating the moving image data with the shaking information at the time of imaging and the shaking change information SMI that can specify the processing amount of the shaking change processing.
- Shake change information SMI includes information on shake change for each region, which will be described later. Then, the associated moving image data, the shaking information at the time of imaging, and the shaking change information SMI are collectively or separately transmitted to the image processing device TDy via wired communication, wireless communication, or a recording medium. Can be done.
- the term "associate" means, for example, to make the other information available (linkable) when processing one piece of information (data, commands, programs, etc.).
- the information associated with each other may be collected as one file or the like, or may be individual information.
- the information B associated with the information A may be transmitted on a transmission path different from that of the information A.
- the information B associated with the information A may be recorded on a recording medium (or another recording area of the same recording medium) different from the information A.
- this "association" may be a part of the information, not the entire information.
- an image and information corresponding to the image may be associated with each other in an arbitrary unit such as a plurality of frames, one frame, or a part within the frame.
- FIG. 2 shows moving image data transmitted from the image processing device TDx to the image processing device TDy as moving image data VD2.
- the moving image data VD2 is an image in which the shake removal performed by the image pickup device 1 is canceled, an image in which the shake removal is performed by the image processing device TDx, or a shake change process is performed by the image processing device TDx.
- FIG. 2 shows the metadata MTD2 transmitted from the image processing device TDx to the image processing device TDy.
- the metadata MTD2 may have the same information as the metadata MTD1 or may have some different information.
- the image processing device TDy When the metadata MTD2 includes the shaking information at the time of imaging, the image processing device TDy is in a state where at least the moving image data VD2, the shaking information at the time of imaging included in the metadata MTD2, and the shaking change information SMI are associated with each other. You can get it. It should be noted that a data form in which the shaking change information SMI is also included in the metadata MTD2 can be considered.
- the image pickup apparatus 1 includes, for example, a lens system 11, an image sensor unit 12, a camera signal processing unit 13, a recording control unit 14, a display unit 15, an output unit 16, an operation unit 17, a camera control unit 18, and a memory. It has a unit 19, a driver unit 22, and a sensor unit 23.
- the lens system 11 includes a lens such as a cover lens, a zoom lens, and a focus lens, an aperture mechanism, and the like. Light from the subject (incident light) is guided by the lens system 11 and focused on the image sensor unit 12. Although not shown, the lens system 11 may be provided with an optical image stabilization mechanism that corrects image shake (interframe shake) and blur due to camera shake or the like.
- the image sensor unit 12 includes, for example, an image sensor 12a (imaging element) such as a CMOS (Complementary Metal Oxide Semiconductor) type or a CCD (Charge Coupled Device) type.
- the image sensor unit 12 executes, for example, CDS (Correlated Double Sampling) processing, AGC (Automatic Gain Control) processing, and the like on the electric signal obtained by photoelectric conversion of the light received by the image sensor 12a, and further performs A / D. (Analog / Digital) Perform conversion processing. Then, the image pickup signal as digital data is output to the camera signal processing unit 13 and the camera control unit 18 in the subsequent stage.
- the optical image stabilization mechanism (not shown) is a mechanism that corrects image shake by moving the image sensor 12a side instead of the lens system 11 side, or spatial optical image stabilization using a gimbal. In some cases, it may be a balanced optical image stabilization mechanism, and any method may be used. In the optical image stabilization mechanism, in addition to the interframe shake, the blur in the frame is also corrected as described later.
- the camera signal processing unit 13 is configured as an image processing processor by, for example, a DSP (Digital Signal Processor) or the like.
- the camera signal processing unit 13 performs various signal processing on the digital signal (image image signal) from the image sensor unit 12. For example, as a camera process, the camera signal processing unit 13 performs preprocessing, simultaneous processing, YC generation processing, resolution conversion processing, codec processing, and the like.
- the camera signal processing unit 13 also performs various correction processes. However, it is assumed that the image stabilization may or may not be performed in the image pickup apparatus 1.
- a clamping process for clamping the black level of R, G, B to a predetermined level, a correction process between the color channels of R, G, B, etc. are performed on the captured image signal from the image sensor unit 12. conduct.
- a color separation processing is performed so that the image data for each pixel has all the color components of R, G, and B.
- demosaic processing is performed as color separation processing.
- YC generation process a luminance (Y) signal and a color (C) signal are generated (separated) from the image data of R, G, and B.
- the resolution conversion process the resolution conversion process is executed on the image data subjected to various signal processing.
- FIG. 4 shows an example of various correction processes (internal correction of the image pickup apparatus 1) performed by the camera signal processing unit 13.
- the optical image stabilization performed by the lens system 11 and the correction processing performed by the camera signal processing unit 13 are illustrated by their execution order.
- the in-lens image stabilization by shifting the yaw direction and pitch direction of the lens system 11 and the in-body image stabilization by shifting the yaw direction and pitch direction of the image sensor 12a are performed to perform image stabilization.
- the image of the subject is formed on the image sensor 12a in a state where the influence of the above is physically canceled.
- the in-lens image stabilization and the in-body image stabilization may be only one, or both may be used. When both in-lens image stabilization and in-body image stabilization are used, it is conceivable that the in-body image stabilization does not shift in the yaw direction or pitch direction.
- neither in-lens image stabilization nor in-body image stabilization is adopted, and for image stabilization, only electronic image stabilization or only optical image stabilization may be performed.
- processing from processing F2 to processing F7 is performed by spatial coordinate transformation for each pixel.
- lens distortion correction is performed.
- focal plane distortion correction is performed as one element of electronic image stabilization. It should be noted that this is to correct the distortion when the rolling shutter type reading is performed by, for example, the CMOS type image sensor 12a.
- Roll correction is performed in the process F4. That is, the roll component is corrected as one element of the electronic image stabilization.
- trapezoidal distortion correction is performed for the trapezoidal distortion caused by the electronic image stabilization.
- the keystone distortion caused by electronic image stabilization is perspective distortion caused by cutting out a place away from the center of the image.
- the pitch direction and the yaw direction are shifted and cut out as one element of the electronic image stabilization.
- camera shake correction, lens distortion correction, and trapezoidal distortion correction are performed by the above procedure. It is not essential to carry out all of the processes listed here, and the order of the processes may be changed as appropriate.
- the image data subjected to the above various processing is subjected to, for example, coding processing for recording or communication, and file generation.
- a moving image file MF as an MP4 format used for recording MPEG-4 compliant video / audio is generated. It is also conceivable to generate files in formats such as PEG (Joint Photographic Experts Group), TIFF (Tagged Image File Format), GIF (Graphics Interchange Format), and HEIF (High Efficient Image File) as still image files.
- the camera signal processing unit 13 also generates metadata to be added to the moving image file MF by using the information from the camera control unit 18 and the like.
- the audio processing system is not shown in FIG. 3, it actually has an audio recording system and an audio processing system, and the moving image file MF may include the audio data as well as the moving image data.
- the recording control unit 14 records and reproduces, for example, a recording medium using a non-volatile memory.
- the recording control unit 14 performs a process of recording a moving image file MF such as moving image data or still image data, a thumbnail image, or the like on a recording medium, for example.
- the actual form of the recording control unit 14 can be considered in various ways.
- the recording control unit 14 may be configured as a flash memory built in the image pickup device 1 and a write / read circuit thereof, or a recording medium that can be attached to and detached from the image pickup device 1, such as a memory card (portable flash memory, etc.). ) May be in the form of a card recording / playback unit that performs recording / playback access. Further, it may be realized as an HDD (Hard Disk Drive) or the like as a form built in the image pickup apparatus 1.
- HDD Hard Disk Drive
- the display unit 15 is a display unit that displays various displays to the imager, and is, for example, a display such as a liquid crystal panel (LCD: Liquid Crystal Display) or an organic EL (Electro-Luminescence) display arranged in the housing of the image pickup device 1. It is used as a display panel or view finder depending on the device.
- the display unit 15 causes various displays to be executed on the display screen based on the instruction of the camera control unit 18. For example, the display unit 15 displays a reproduced image of the image data read from the recording medium by the recording control unit 14.
- the display unit 15 is supplied with image data of the captured image whose resolution has been converted by the camera signal processing unit 13 for display, and the display unit 15 is based on the image data of the captured image in response to an instruction from the camera control unit 18. May be displayed. As a result, a so-called through image (subject monitoring image), which is an captured image during composition confirmation, is displayed. Further, the display unit 15 causes various operation menus, icons, messages, etc., that is, display as a GUI (Graphical User Interface) to be executed on the screen based on the instruction of the camera control unit 18.
- GUI Graphic User Interface
- the output unit 16 performs data communication and network communication with an external device by wire or wirelessly.
- the image data for example, a moving image file MF
- the output unit 16 is a network communication unit, it communicates with various networks such as the Internet, a home network, and a LAN (Local Area Network), and transmits and receives various data to and from servers, terminals, and the like on the network. You may do so.
- the operation unit 17 collectively shows input devices for the user to perform various operation inputs. Specifically, the operation unit 17 shows various controls (keys, dials, touch panels, touch pads, etc.) provided in the housing of the image pickup apparatus 1. The operation unit 17 detects the user's operation, and the signal corresponding to the input operation is sent to the camera control unit 18.
- the camera control unit 18 is composed of a microcomputer (arithmetic processing device) provided with a CPU (Central Processing Unit).
- the memory unit 19 stores information and the like used for processing by the camera control unit 18.
- a ROM Read Only Memory
- RAM Random Access Memory
- flash memory and the like are comprehensively shown.
- the memory unit 19 may be a memory area built in the microcomputer chip as the camera control unit 18, or may be configured by a separate memory chip.
- the camera control unit 18 controls the entire image pickup apparatus 1 by executing a program stored in the ROM of the memory unit 19, the flash memory, or the like.
- the camera control unit 18 controls the shutter speed of the image sensor unit 12, gives instructions for various signal processing in the camera signal processing unit 13, captures and records according to the user's operation, and reproduces the recorded moving image file MF and the like.
- the operation of each necessary part is controlled with respect to the operation of the lens system 11 such as zoom, focus, and aperture adjustment in the lens barrel, and the operation of the user interface.
- the RAM in the memory unit 19 is used for temporarily storing data, programs, and the like as a work area for various data processing of the CPU of the camera control unit 18.
- the ROM and flash memory (nonvolatile memory) in the memory unit 19 include an OS (Operating System) for the CPU to control each unit, content files such as a moving image file MF, application programs for various operations, and a firmware. It is used to store clothing and the like.
- the driver unit 22 is provided with, for example, a motor driver for the zoom lens drive motor, a motor driver for the focus lens drive motor, a motor driver for the diaphragm mechanism motor, and the like. These motor drivers apply a drive current to the corresponding driver in response to an instruction from the camera control unit 18, to move the focus lens and the zoom lens, open and close the diaphragm blades of the diaphragm mechanism, and the like.
- the sensor unit 23 comprehensively shows various sensors mounted on the image pickup apparatus.
- an IMU intial measurement unit
- an angular velocity is detected by a three-axis angular velocity (gyro) sensor of pitch-, yaw, and roll
- acceleration is detected by an acceleration sensor. can do.
- the sensor unit 23 may include a sensor capable of detecting camera shake during imaging, and does not need to include both a gyro sensor and an acceleration sensor. Further, the sensor unit 23 may be equipped with a position information sensor, an illuminance sensor, or the like.
- the moving image file MF imaged and generated by the above image pickup device 1 can be transferred to the image processing devices TDx and TDy of the mobile terminal 2 and the like for image processing.
- the mobile terminal 2 and the personal computer 3 serving as the image processing devices TDx and TDy can be realized as, for example, an information processing device having the configuration shown in FIG.
- the server 4 can be realized by the information processing device having the configuration shown in FIG.
- the CPU 71 of the information processing apparatus 70 executes various processes according to a program stored in the ROM 72 or a program loaded from the storage unit 79 into the RAM 73.
- the RAM 73 also appropriately stores data and the like necessary for the CPU 71 to execute various processes.
- the CPU 71, ROM 72, and RAM 73 are connected to each other via a bus 74.
- An input / output interface 75 is also connected to the bus 74.
- An input unit 76 including an operator and an operation device is connected to the input / output interface 75.
- various controls and operation devices such as a keyboard, mouse, keys, dial, touch panel, touch pad, and remote controller are assumed.
- the user's operation is detected by the input unit 76, and the signal corresponding to the input operation is interpreted by the CPU 71.
- a display unit 77 made of an LCD or an organic EL panel and an audio output unit 78 made of a speaker or the like are connected to the input / output interface 75 as one or a separate body.
- the display unit 77 is a display unit that performs various displays, and is composed of, for example, a display device provided in the housing of the information processing device 70, a separate display device connected to the information processing device 70, and the like.
- the display unit 77 executes the display of various images for image processing, moving images to be processed, and the like on the display screen based on the instruction of the CPU 71. Further, the display unit 77 displays various operation menus, icons, messages, etc., that is, as a GUI (Graphical User Interface) based on the instruction of the CPU 71.
- GUI Graphic User Interface
- a storage unit 79 composed of a hard disk, a solid-state memory, or the like, or a communication unit 80 composed of a modem or the like may be connected to the input / output interface 75.
- the communication unit 80 performs communication processing via a transmission line such as the Internet, wire / wireless communication with various devices, bus communication, and the like.
- a drive 82 is also connected to the input / output interface 75, if necessary, and a removable recording medium 81 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory is appropriately mounted.
- the drive 82 can read data files such as a moving image file MF and various computer programs from the removable recording medium 81.
- the read data file is stored in the storage unit 79, and the image and sound included in the data file are output by the display unit 77 and the sound output unit 78. Further, the computer program or the like read from the removable recording medium 81 is installed in the storage unit 79 as needed.
- this information processing device 70 for example, software for image processing as the image processing device of the present disclosure can be installed via network communication by the communication unit 80 or a removable recording medium 81.
- the software may be stored in the ROM 72, the storage unit 79, or the like in advance.
- FIG. 6 shows a function provided as an information processing device 70 that functions as an image processing device TDx. That is, the information processing device 70 (CPU 71) has functions as a shake changing unit 101, a parameter setting unit 102, a user interface processing unit 103, and an area setting unit 104.
- the "user interface” is also referred to as "UI”
- the user interface processing unit 103 is also referred to as "UI processing unit 103" below.
- the shaking changing unit 101 is a function of performing shaking changing processing for changing the shaking state appearing in the output moving image data by using the parameters PRM1 and PRM2 based on the user's input and automatic setting and the adjusted shaking information.
- the shaking changing unit 101 performs processing for removing shaking and adding shaking to obtain an output image as moving image data in which shaking has been removed or added. Further, the shaking changing unit 101 performs shaking changing processing for each area in the image based on the area information iAR.
- the UI processing unit 103 is a function of causing the user to present an operator related to the shaking change and performing a process of acquiring operation information by the operator. For example, the UI processing unit 103 performs a process of displaying an operator, a preview image, information about an image, and the like as a UI image on the display unit 77. Further, the UI processing unit 103 detects the user's operation by the input unit 76. For example, a touch operation on a UI image is detected.
- the parameter setting unit 102 is a function of setting the parameters PRM1 and PRM2 for the shaking change processing based on the operation information acquired by the UI processing unit 103. That is, the user operation content detected by the UI processing unit 103 is converted into the parameters PRM1 and PRM2 for changing the shaking, and the data is supplied to the shaking changing unit 101 to change the shaking of the moving image data according to the user's operation. Allow processing to take place.
- the parameter PRM1 is a parameter for performing sway removal or sway addition, and can be said to be an indicated value of the overall sway removal amount or sway addition amount.
- the parameter PRM2 is an instruction value of the shaking change processing that causes different shaking states for each area, such as the amount of difference in shaking for each of a plurality of areas set by the area setting unit 104 in the image.
- the parameter PRM2 includes the amount of shaking for each region (the difference in the amount of shaking may be used), the period of shaking for each region, the value indicating the direction of shaking for each region, and the like. ..
- the parameter PRM1 used for changing the shaking due to the rotation of the celestial sphere model which will be described later
- the parameter PRM2 used for changing the shaking condition of a part area by enlarging / reducing the image by moving the coordinate points on the celestial sphere are distinguished.
- the use of the parameters PRM1 and PRM2 having such a meaning is only an example.
- the shaking change process may be performed so that the shaking changing unit 101 is in a different shaking state for each region in the image according to the parameter PRM2.
- the parameter setting unit 102 does not necessarily have to set the parameters according to the user operation.
- parameters may be set to add a fixed shaking pattern.
- the parameters PRM1 and PRM2 may be automatically set according to the image content.
- the shaking information may be selected according to the image content, and the parameters PRM1 and PRM2 may be set according to the shaking pattern.
- the user selects a shaking pattern and sets the parameters PRM1 and PRM2 according to the selection.
- the area setting unit 104 performs a process of setting a plurality of areas in the image of the moving image data according to, for example, an operation of specifying the area detected by the UI processing unit 103. Then, the area information iAR indicating the set area is provided to the shaking change unit 101. For example, the area setting unit 104 performs a touch operation such that the user traces an arbitrary range on the preview image displayed on the display unit 77 by the processing of the UI processing unit 103, an operation of designating or drawing some line, and the like. Depending on what is done, it is conceivable to determine the boundary instructed by the user and divide and set the area in the screen. The shaking changing unit 101 grasps the area in the image by the area information iAR, and performs the shaking changing process so that the shaking is different for each area.
- FIG. 7 shows another functional configuration example.
- FIG. 7 shows a configuration example in which the area setting unit 104 sets the area by image analysis instead of user operation.
- the area setting unit 104 analyzes the input moving image data to perform, for example, determination of the subject person, determination of the main subject, determination of the type of the subject such as a moving object / fixed object / background, composition determination, and the like. Set the area division accordingly. Then, the area information iAR indicating the set area area is provided to the shaking change unit 101.
- an example of area setting based on user operation is shown in FIG. 6 and an example of automatic area setting based on image analysis is shown in FIG. 7, but both of them may be used.
- a user performs an operation of designating a subject, it is conceivable to identify the range of the subject or the range of a similar subject by image analysis and generate a boundary of the area.
- a region candidate may be presented to the user based on image analysis, and the user may be able to select the region division pattern to be adopted by an operation.
- the UI processing unit 103 and the parameter setting unit 102 are shown in the configuration examples of FIGS. 6 and 7, these configurations are not essential when changing the shaking for each area.
- the shaking changing unit 101 and the area setting unit 104 By providing at least the shaking changing unit 101 and the area setting unit 104, the shaking change for each area is realized.
- FIG. 8A is an example divided into a region AR1 in which a person riding a bicycle is shown on the lower side of the image and a region AR2 on the upper side considered to be a background image. note that.
- the boundary region is shown as the region AR3, which can be considered to have a slight width at the boundaries of the regions AR1 and AR2.
- the areas AR1 and AR2 are set according to the user performing an operation such as tracing the boundary between the areas AR1 and AR2.
- the area AR1 is automatically set so as to surround the area in which a person is captured, and the other areas are set as the area AR2.
- the area AR1 is an area for expressing power by shaking, and since the area AR2 is a background or the like, it is desired to eliminate or reduce the shaking. In the shaking change process according to the area setting in this way, the shaking is relatively large for the area AR1 and the shaking is reduced for the area AR2.
- Shake change processing such as processing P1 and processing P2 is assumed.
- -Processing P1 Shaking and adding only to the area AR1 (the area AR2 maintains a state without shaking).
- -Processing P2 A large amount of shaking is added to the area AR1, and a small amount of shaking is added to the area AR2.
- shaking changing processing such as the following processing P3, processing P4, processing P5, and processing P6 is assumed.
- -Processing P3 Shake addition to the area AR1 and shaking removal (including partial removal) to the area AR2.
- -Processing P4 A large amount of shaking is added to the area AR1, and a small amount of shaking is added to the area AR2.
- -Processing P5 Shake removal (including partial removal) is performed only in the area AR2 (the area AR1 maintains the shaking state).
- -Processing P6 Partial sway removal for removing a small amount of sway in the area AR1 and sway removal for almost eliminating the sway in the area AR2 are performed.
- the area AR3 which is a boundary area
- FIG. 8B is an example of dividing the area of the course (paved road and dirt) in the image.
- the following is a specific example of the effect, taking as an example the image after the shaking at the time of the original imaging is removed by the shaking removal.
- FIG. 9 schematically shows a processing example of shaking change using the celestial sphere model MT.
- FIG. 9A shows one frame of an image that is shaken due to camera shake or the like at the time of imaging.
- the frame of this two-dimensional image is converted into a three-dimensional image model attached to the virtual celestial sphere model MT as shown in FIG. 9B.
- the corresponding points from two dimensions to three dimensions are obtained from the lens distortion characteristic data.
- this celestial sphere model MT is rotated according to the shaking information at the time of imaging, a three-dimensional image model in which the shaking of FIG. 9C is removed can be obtained.
- the shaking information at the time of imaging is, for example, IMU data described later or a quarternion QD based on the IMU data.
- the specific coordinate range AZ shown by the thick line is cut out from the celestial sphere model MT of FIG. 9C, a two-dimensional image in which shaking and lens distortion are removed can be obtained as shown in FIG. 9D.
- FIGS. 10 and 11 show a state in which a certain frame and another frame are attached to the celestial sphere model MT.
- the radial line is a line corresponding to the longitude in the celestial sphere model MT
- the concentric line is a line corresponding to the latitude.
- the images of FIGS. 10 and 11 are images of different frames that differ in time, and the positions of the images projected onto the celestial sphere model MT as shown in FIGS. 9A to 9B are displaced due to the shaking during imaging. It becomes a thing.
- FIG. 12A shows a pixel range of a certain moving image data and a cutout range CR as an image to be output.
- FIG. 12A shows a state in which there is no enlargement / reduction in the moving image data
- FIG. 12B shows a state in which the shaded area is enlarged
- FIG. 12C shows a state in which the shaded area is reduced.
- the area of the cutting range CR does not change.
- the shaded area can be considered as the area AR3 which is the boundary area in FIG. 8A and the like.
- FIG. 12B by enlarging the boundary region AR3, the coordinates of each pixel in the lower portion (region AR1 in FIG. 8A) move downward.
- FIG. 12C the boundary region AR3 is reduced, so that the coordinates of each pixel in the lower portion (region AR1 in FIG. 8A) move upward.
- FIG. 12B the enlargement processing shown in FIG. 13 is performed in the shaded portion which is the boundary region (AR3).
- FIG. 13A shows a certain four pixels in the boundary region.
- “A”, “b”, “c”, and “d” are pixel values (for example, RGB values).
- FIG. 14A is similar to FIG. 13A.
- the reduction process is performed as shown in FIGS. 14B and 14C.
- FIG. 14B "la”, “lb”, “lc”, and “ld” are the distances from “p” to "a", "b", "c", and “d”.
- the pixel size is "1" and "p”.
- Performing the above-mentioned enlargement / reduction for each frame means that the intersection coordinates of the mesh of the celestial sphere model MT are moved according to the degree of shaking, and the image cut out as the cutout range CR in FIG. 12 is from the boundary area.
- shaking is added to one area. Therefore, for example, with respect to the shaking of the region AR2, the rotation of the celestial sphere model MT removes, partially removes, or adds, and then the movement of the coordinate points due to the enlargement / reduction of the boundary region is added to the pixels of the region AR1.
- AR1 is in a state where a stronger shaking than the region AR2 is added.
- a boundary area in contact with the area AR1 and a boundary area in contact with the area AR2 may be set, and enlargement / reduction may be performed for each frame according to different shaking amounts / cycles / directions.
- the above-mentioned processes P1 to P6 may be realized by combining the sway change due to the rotation of the celestial sphere model MT and the sway change of a specific region due to the coordinate change.
- FIG. 15A shows the data included in the moving image file MF.
- the moving image file MF includes various data as "header”, “sound”, “movie”, and "metadata”.
- “header” information such as a file name and a file size as well as information indicating the presence or absence of metadata are described.
- “Sound” is audio data recorded together with a moving image. For example, 2-channel stereo audio data is stored.
- the “movie” is moving image data, and is composed of image data as each frame (# 1, # 2, # 3, ...) Constituting the moving image.
- the "metadata” additional information associated with each frame (# 1, # 2, # 3, ...) Constituting the moving image is described.
- FIG. 15B An example of the contents of the metadata is shown in FIG. 15B.
- IMU data, coordinate conversion parameter HP, timing information TM, and camera parameter CP are described for one frame. It should be noted that these are a part of the metadata contents, and here, only the information related to the image processing described later is shown.
- the IMU data As the IMU data, a gyro (angular velocity data), an accelerator (acceleration data), and a sampling rate are described.
- the IMU mounted on the image pickup apparatus 1 as the sensor unit 23 outputs angular velocity data and acceleration data at a predetermined sampling rate. Generally, this sampling rate is higher than the frame rate of the captured image, so that many IMU data samples can be obtained in one frame period.
- n samples are associated with each frame, such as gyro sample # 1, gyro sample # 2, ... Gyro sample # n shown in FIG. 15C.
- m samples are associated with each frame, such as accelerator sample # 1, accelerator sample # 2, ... accelerator sample # m.
- n m
- n ⁇ m the metadata is described here as an example associated with each frame, for example, the IMU data may not be completely synchronized with the frame. In such a case, for example, the time information associated with the time information of each frame is provided as the IMU sample timing offset in the timing information TM.
- the coordinate conversion parameter HP is a general term for parameters used for correction accompanied by coordinate conversion of each pixel in the image. It also includes non-linear coordinate transformations such as lens distortion.
- the coordinate conversion parameter HP is a term that can include at least a lens distortion correction parameter, a trapezoidal distortion correction parameter, a focal plane distortion correction parameter, an electronic image stabilization parameter, and an optical image stabilization parameter.
- the lens distortion correction parameter is information for directly or indirectly grasping how the distortion such as barrel aberration and pincushion aberration is corrected and returning the image to the image before the lens distortion correction.
- the metadata regarding the lens distortion correction parameter as one of the metadata will be briefly described.
- FIG. 16A shows the image height Y, the angle ⁇ , the entrance pupil position d1, and the exit pupil position d2 in the schematic diagram of the lens system 11 and the image sensor 12a.
- the lens distortion correction parameter is used in image processing to know the incident angle of each pixel of the image sensor 12a. Therefore, it is sufficient to know the relationship between the image height Y and the angle ⁇ .
- FIG. 16B shows the image 110 before the lens distortion correction and the image 111 after the lens distortion correction.
- the maximum image height H0 is the maximum image height before distortion correction, and is the distance from the center of the optical axis to the farthest point.
- the maximum image height H1 is the maximum image height after distortion correction.
- the metadata required to understand the relationship between the image height Y and the angle ⁇ is the maximum image height H0 before distortion correction and the incident angle data d0, d1, ... d (N-) for each of the N image heights. 1). It is assumed that "N" is about 10 as an example.
- the trapezoidal distortion correction parameter is a correction amount when correcting the trapezoidal distortion caused by shifting the cutout area from the center by electronic image stabilization, and is also a value corresponding to the correction amount of electronic image stabilization.
- the focal plane distortion correction parameter is a value indicating the amount of correction for each line with respect to the focal plane distortion.
- the parameters of lens distortion correction, trapezoidal distortion correction, focal plane distortion correction, and electronic image stabilization are collectively referred to as coordinate conversion parameters, but these correction processes are performed by each of the image sensors 12a of the image sensor unit 12. This is because it is a correction process for an image formed on a pixel and is a parameter of a correction process that involves a coordinate conversion of each pixel.
- Optical image stabilization is also one of the coordinate conversion parameters, but this is because the correction of the fluctuation of the inter-frame component in the optical image stabilization is a process that involves coordinate conversion of each pixel.
- the image data to which the lens distortion correction, the trapezoidal distortion correction, the focal plane distortion correction, the electronic image stabilization, and the optical image stabilization have been performed can be captured before each correction processing, that is, by imaging. It is possible to return to the state when the image sensor 12a of the element unit 12 is imaged.
- the lens distortion correction, trapezoidal distortion correction, and focal plane distortion correction parameters are distortion correction processing for the case where the optical image itself from the subject is an image captured in an optically distorted state, and each of them is an optical distortion. Since it is intended for correction, it is collectively referred to as an optical distortion correction parameter. That is, if the reverse correction is performed using these parameters, the image data to which the lens distortion correction, the trapezoidal distortion correction, and the focal plane distortion correction have been performed can be returned to the state before the optical distortion correction.
- the timing information TM in the metadata includes each information of exposure time (shutter speed), exposure start timing, readout time (curtain speed), number of exposure frames (long exposure information), IMU sample offset, and frame rate. In the image processing of the present embodiment, these are mainly used to associate the line of each frame with the IMU data. However, even if the image sensor 12a is a CCD or a global shutter CMOS, if the exposure center of gravity shifts using an electronic shutter or mechanical shutter, the exposure start timing and curtain speed are also used to match the exposure center of gravity. Correction is possible.
- the camera parameter CP in the metadata As the camera parameter CP in the metadata, the angle of view (focal length), zoom position, and lens distortion information are described.
- FIG. 17 shows the procedures of various processes executed in the information processing device 70 as the image processing device TDx, and shows the relationship of the information used in each process.
- the processes of steps ST13, ST14, ST15, and ST16 enclosed as step ST30 in FIG. 17 are performed.
- the parameter setting process of step ST41 is performed.
- the UI processing in step ST40 is performed.
- the area setting process of step ST42 is performed.
- step ST30 surrounded by a broken line is referred to as "sway change”, and step ST16 is described as "sway change processing”.
- Step ST16 is a celestial sphere model MT for actually changing the state of shaking. It is a process such as movement of coordinate points by rotation or enlargement / reduction of the boundary area, and is a "sway change” in a narrow sense.
- step ST30 is a broadly defined “sway change” including the sway change process of step ST16, the celestial sphere model process as a preparation for the process, and parameter setting.
- the pre-processing is the processing performed when the moving image file MF is imported.
- the term "import" as used herein means that the information processing device 70 targets, for example, a moving image file MF that can be accessed by being imported into a storage unit 79 or the like, and performs image processing by performing preprocessing. It means to develop as possible. For example, it does not mean transferring from the image pickup device 1 to the mobile terminal 2 or the like.
- the CPU 71 imports the moving image file MF designated by the user operation or the like so as to be the image processing target, and also performs processing related to the metadata added to the moving image file MF as preprocessing. For example, a process of extracting and storing metadata corresponding to each frame of a moving image is performed. Specifically, in this preprocessing, metadata extraction (step ST1), all IMU data concatenation (step ST2), metadata retention (step ST3), conversion to quotation (posture information of imaging device 1), and retention. (Step ST4) is performed.
- step ST1 the CPU 71 reads the target moving image file MF and extracts the metadata included in the moving image file MF as described with reference to FIG. Note that part or all of steps ST1, ST2, ST3, and ST4 may be performed on the image source VS side such as the image pickup apparatus 1. In that case, in the pre-processing, the contents after the processing described below are acquired as metadata.
- the CPU 71 performs a concatenation process on the IMU data (angular velocity data (gyro sample) and acceleration data (accelerator sample)) in step ST2.
- This is a process of constructing IMU data corresponding to the entire sequence of moving images by arranging and concatenating all the IMU data associated with all frames in chronological order.
- integration processing is performed on the connected IMU data to calculate a quarternion QD representing the posture of the imaging device 1 at each time point on the sequence of moving images, and this is stored and retained. It is an example that the quarternion QD is calculated. It is also possible to calculate the quarternion QD using only the angular velocity data.
- the CPU 71 performs a process of holding the metadata other than the IMU data, that is, the coordinate conversion parameter HP, the timing information TM, and the camera parameter CP in step ST3. That is, the coordinate conversion parameter HP, the timing information TM, and the camera parameter CP are stored in a state corresponding to each frame.
- the CPU 71 is ready to perform various image processing including the shaking change of the moving image data received as the moving image file MF.
- the routine processing of FIG. 17 shows image processing performed on the moving image data of the moving image file MF that has been preprocessed as described above.
- the CPU 71 takes out one frame of the moving image (step ST11), cancels the internal correction of the image pickup device (step ST12), pastes it on the celestial sphere model (step ST13), synchronizes processing (step ST14), adjusts the shaking information (step ST15), and shakes. Change (step ST16), output area designation (step ST17), plane projection and cutout (step ST18).
- the CPU 71 performs each of the above steps ST11 to ST20 for each frame when reproducing the image of the moving image file MF. Along with these, the CPU 71 performs UI processing (step ST40), parameter setting processing (step ST41), and area setting processing (step ST42) at necessary timings.
- step ST11 the CPU 71 decodes one frame of the moving image (moving image file MF) along the frame number FN. Then, one frame of moving image data PD (#FN) is output. Note that "(#FN)" indicates a frame number and indicates that the information corresponds to that frame. If the moving image is not encoded by compression or the like, the decoding process in step ST11 is not necessary.
- step ST12 the CPU 71 performs a process of canceling the internal correction performed by the image pickup apparatus 1 for the moving image data PD (#FN) of one frame.
- the CPU 71 refers to the coordinate conversion parameter HP (#FN) stored corresponding to the frame number (#FN) at the time of preprocessing, and performs a correction opposite to the correction performed by the image pickup apparatus 1.
- moving image data iPD (#FN) in a state in which lens distortion correction, trapezoidal distortion correction, focal plane distortion correction, electronic camera shake correction, and optical camera shake correction in the image pickup apparatus 1 are canceled is obtained.
- step ST12 it is moving image data in which the shaking removal or the like performed by the image pickup apparatus 1 is canceled and the influence of the shaking such as camera shake at the time of imaging appears as it is.
- the process of canceling the internal correction of the image pickup apparatus as step ST12 may not be performed.
- the process of step ST12 may be skipped and the moving image data PD (#FN) may be output as it is.
- step ST13 the CPU 71 attaches the 1-frame video data iPD (#FN) in a state where various corrections have been canceled to the celestial sphere model MT.
- the camera parameter CP (#FN) stored corresponding to the frame number (#FN), that is, the angle of view, the zoom position, and the lens distortion information are referred to.
- FIG. 18 shows an outline of attachment to the celestial sphere model MT.
- FIG. 18A shows the moving image data iPD.
- the image height h is the distance from the center of the image.
- Each circle in the figure indicates a position where the image heights h are equal.
- the "relationship between the image sensor surface and the incident angle ⁇ " in that frame is calculated, and "data0" at each position on the image sensor surface ... Let's say “dataN-1”. Then, from “data0” ... "dataN-1", it is expressed as a one-dimensional graph of the relationship between the image height h and the incident angle ⁇ as shown in FIG. 18B.
- the incident angle ⁇ is the angle of the light beam (the angle seen from the optical axis).
- This one-dimensional graph is rotated once around the center of the captured image, and the relationship between each pixel and the incident angle is obtained. Accordingly, mapping from the pixel G1 in FIG. 18C to the celestial sphere model MT is performed for each pixel of the moving image data iPD as in the pixel G2 on the celestial sphere coordinates.
- an image (data) of the celestial sphere model MT in which the captured image is attached to the ideal celestial sphere with the lens distortion removed can be obtained.
- this celestial sphere model MT the parameters and distortions peculiar to the image pickup device 1 that originally captured the moving image data iPD are removed, and the range that can be seen by an ideal pinhole camera is pasted on the celestial sphere. Therefore, by rotating the image of the celestial sphere model MT in a predetermined direction in this state, it is possible to realize the overall shaking removal of the image and the shaking change processing as the shaking addition.
- the attitude information (quarterion QD) of the image pickup apparatus 1 is used for the shaking change processing. Therefore, the CPU 71 performs the synchronization process in step ST14.
- a process of identifying and acquiring a quaternion QD (#LN) suitable for each line corresponding to the frame number FN is performed. Note that "(#LN)" indicates a line number in the frame and indicates that the information corresponds to that line.
- the reason why the quarternion QD (#LN) for each line is used is that when the image sensor 12a is a CMOS type and the imaging is performed by the rolling shutter method, the amount of shaking differs for each line.
- a frame-by-frame quarternion QD (#FN) may be used.
- the center of gravity shifts when an electronic shutter (similar to a mechanical shutter) is used, so the center of the exposure period of the frame (shifts according to the shutter speed of the electronic shutter). ) Timing quotation should be used.
- FIG. 19 shows the synchronization signal cV of the image pickup apparatus 1 during the vertical period, the synchronization signal sV of the image sensor 12a generated from the synchronization signal cV, and the sample timing of the IMU data, and also shows the exposure timing range 120.
- the exposure timing range is a parallelogram schematically showing the exposure period of each line of one frame when the exposure time is t4 by the rolling shutter method.
- the temporal offset t0 of the synchronization signal cV and the synchronization signal sV, the IMU sample timing offset t1, the read start timing t2, the read time (shutter speed) t3, and the exposure time t4 are shown.
- the read start timing t2 is the timing after a predetermined time t2of has passed from the synchronization signal sV.
- Each IMU data obtained at each IMU sample timing is associated with a frame.
- the IMU data in the period FH1 is the metadata associated with the current frame indicating the exposure period in a parallelogram
- the IMU data in the period FH1 is the metadata associated with the next frame.
- the IMU data in the period FH1 is the metadata associated with the next frame.
- the association between each frame and the IMU data is released, and the IMU data can be managed in chronological order.
- the IMU data corresponding to the exposure center of gravity (timing of the broken line W) of each line of the current frame is specified. This can be calculated if the temporal relationship between the IMU data and the effective pixel area of the image sensor 12a is known.
- the IMU data corresponding to the exposure center of gravity (timing of the broken line W) of each line is specified by using the information that can be acquired as the timing information TM corresponding to the frame (#FN). That is, it is information on the exposure time, the exposure start timing, the readout time, the number of exposure frames, the IMU sample offset, and the frame rate. Then, the quaternion QD calculated from the IMU data of the exposure center of gravity is specified and used as the quaternion QD (#LN) which is the attitude information for each line.
- This quarternion QD (#LN) is provided for the process of adjusting the shaking information in step ST15.
- the CPU 71 adjusts the quaternion QD according to the input parameter PRM1 for changing the shaking.
- the parameter PRM1 may be a parameter input according to a user operation or a parameter generated by automatic control.
- the user can input to set the parameter PRM1 by the operation method provided in the UI processing of step ST40 so as to add an arbitrary degree of shaking to the image. Further, the CPU 71 can generate the parameter PRM1 by automatic control according to an image analysis, an image type, a user's shaking model selection operation, or the like.
- the user can input an operation to instruct the shaking change. That is, an operation for instructing the shaking as a shaking effect, an operation for instructing the degree of shaking removal, and the like.
- the CPU 71 sets the parameters in step ST41. That is, the shaking change parameter is set according to the user operation, and is used for the shaking information adjustment process in step ST15.
- the user can further perform an operation of dividing the area and an operation input for instructing the amount of shaking to be applied to each area. That is, it is an operation for changing the shaking for each area.
- the parameter PRM2 for changing the shaking for each area is set according to such processing and provided for the shaking changing process of step ST16.
- the CPU 71 sets the area according to the user operation and provides the area information iAR to the shaking change process of step ST16.
- an automatic area setting based on the image analysis result may be performed instead of the area setting according to the user operation. Therefore, in the area setting process, there is a case where the image analysis process is performed on the moving image data iPD, the area is automatically set according to the subject recognition, and the area information iAR is provided for the shaking change process in step ST16.
- the area information iAR may be generated according to both the image analysis and the user input. For example, when a user specifies a certain person on an image, the pixel range of the person is specified by image analysis, and the pixel range is set as one area.
- the CPU 71 adds shaking to the image or increases or decreases the amount of shaking based on the quaker QD which is the shaking information at the time of imaging and the parameter PRM1 set in step ST41. Generate adjusted quaternion eQD for.
- FIG. 20 shows an example in which the adjusted quarternion eQD is generated according to the instruction of the gain for each frequency band by the parameter PRM1.
- the frequency band is a band of fluctuation frequencies.
- the band is divided into three bands: low band, middle band, and high band.
- the number of bands may be 2 or more.
- the low-frequency gain LG, the mid-frequency gain MG, and the high-frequency gain HG are given as the parameters PRM1.
- the adjustment processing system of FIG. 20 includes a low-pass filter 41, a mid-pass filter 42, a high-pass filter 43, gain calculation units 44, 45, 46, and a synthesis unit 47.
- “Quaternion QDs for shaking” are input to this adjustment processing system. This is the conjugate of the quarternion QD as shake information during imaging.
- Each value q for the current frame as the quaternion QDs for shaking and the predetermined frames before and after is input to the low-pass filter 41, and the low-pass component q low is obtained.
- the gain calculation unit 44 gives the low-frequency gain LG to the low-frequency component q low.
- Mean (q, n) in the equation indicates the average value of n before and after q. It goes without saying that this mean (q, n) equation is just an example of a low-pass filter, and other calculation methods may be used. Each equation described below is also an example.
- the value q of the quarternion QDs for shaking is also input to the mid- range passing filter 42, and the mid-range component q mid is obtained.
- q * low is a conjugate of q low.
- x is the quaternion product.
- the gain calculation unit 45 gives the mid-range gain MG to the mid- range component q mid.
- the value q of the quarternion QDs for shaking is input to the high frequency passing filter 43, and the high frequency component q high is obtained.
- q * mid is a conjugate of q mid.
- the gain calculation unit 46 gives a high-frequency gain HG to the high-frequency component q high.
- Such gain calculating section 44, 45 and 46 respectively the low frequency gain LG, midrange gain MG, high frequency gain HG is low frequency component q 'low, midrange component q' given mid, high-frequency components q 'You get high.
- the value q mixed obtained by combining this with the synthesis unit 47 is obtained.
- FIG. 21 shows an example in which the adjusted quarternion eQD is generated according to the instruction of the gain for each direction by the parameter PRM1.
- the direction is the direction of sway, that is, the direction of yaw, pitch, and roll.
- Yaw gain YG, pitch gain PG, and roll gain RG are given as parameters PRM1.
- the adjustment processing system of FIG. 21 includes a yaw component extraction unit 51, a pitch component extraction unit 52, a roll component extraction unit 53, a gain calculation unit 54, 55, 56, and a synthesis unit 57.
- Information on the yaw axis, the pitch axis, and the roll axis is provided to the yaw component extraction unit 51, the pitch component extraction unit 52, and the roll component extraction unit 53, respectively.
- Each value q for the current frame as the quarternion QDs for shaking and the predetermined frames before and after is input to the yaw component extraction unit 51, the pitch component extraction unit 52, and the roll component extraction unit 53, respectively, and the yaw component q yaw and the pitch component q are input.
- Find pitch and roll component q roll In each of these component extraction processes, the input is set to the next “q in ”.
- u is a unit vector representing the direction of axes such as the yaw axis, the pitch axis, and the roll axis.
- the yaw component q yaw , the pitch component q pitch , and the roll component q roll obtained by such component extraction are given the yaw gain YG, the pitch gain PG, and the roll gain RG by the gain calculation units 54, 55, and 56, respectively. ..
- FIG. 22 is an example in which the above frequency bands and directions are combined.
- the adjustment processing system includes a low-pass filter 41, a mid-pass filter 42, a high-pass filter 43, direction-specific processing units 58, 59, 60, a gain calculation unit 44, 45, 46, and a synthesis unit 61.
- low-frequency gain LG, mid-frequency gain MG, high-frequency gain HG, and yaw gain YG, pitch gain PG, and roll gain RG (not shown) are given.
- each value q for the current frame as the quaternion QDs for shaking and the predetermined frames before and after is supplied to the low-pass filter 41, the mid-pass filter 42, and the high-pass filter 43, and the respective bands are supplied. Get the ingredients.
- Each band component is input to the direction-specific processing units 58, 59, 60.
- Each of the direction-specific processing units 58, 59, 60 has a yaw component extraction unit 51, a pitch component extraction unit 52, a roll component extraction unit 53, a gain calculation unit 54, 55, 56, and a synthesis unit 57 in FIG. do.
- the direction-specific processing unit 58 the low-frequency components of the quarternion QDs for shaking are divided into the yaw direction, roll direction, and pitch direction components, and the gain calculation is performed using the yaw gain YG, pitch gain PG, and roll gain RG. After performing the above, synthesize.
- the direction-specific processing unit 59 divides the mid-range components of the quaternion QDs for shaking into the components in the yaw direction, the roll direction, and the pitch direction, performs the same gain calculation, and then synthesizes the components.
- the direction-specific processing unit 60 divides the high-frequency components of the quaternion QDs for shaking into components in the yaw direction, roll direction, and pitch direction, performs gain calculation in the same manner, and then synthesizes the components.
- the gains used in the direction-specific processing units 58, 59, and 60 have different gain values. That is, the direction-specific processing unit 58 uses the low-frequency yaw gain YG, the low-frequency pitch gain PG, and the low-frequency roll gain RG, and the direction-specific processing unit 59 uses the mid-range yaw gain YG and the mid-range.
- the pitch gain PG and the roll gain RG for the mid range are used, and the direction-specific processing unit 60 uses the yaw gain YG for the high range, the pitch gain PG for the high range, and the roll gain RG for the high range. That is, it is conceivable that the direction-specific processing units 58, 59, and 60 use nine gains.
- the outputs of these direction-specific processing units 58, 59, and 60 are supplied to the gain calculation units 44, 45, and 46, respectively, and low-frequency gain LG, mid-frequency gain MG, and high-frequency gain HG are given, respectively. Then, it is synthesized by the synthesis unit 61 and output as the value of the adjusted quarternion eQD.
- processing for each direction is applied for each band component, but the reverse is also possible. That is, after dividing by direction first, processing for each frequency band may be applied for each direction component. In that case, it is conceivable to use nine gains in the processing for each frequency band. For example, in the processing for each frequency band in the yaw direction, the low-frequency gain LG for the yaw direction, the mid-range gain MG for the yaw direction, and the high-frequency gain HG for the yaw direction are used.
- the low-frequency gain LG for the pitch direction In the processing for each frequency band in the pitch direction, the low-frequency gain LG for the pitch direction, the mid-range gain MG for the pitch direction, and the high-frequency gain HG for the pitch direction are used. In the processing for each frequency band in the roll direction, the low-frequency gain LG for the roll direction, the mid-range gain MG for the roll direction, and the high-frequency gain HG for the roll direction are used.
- step ST15 of FIG. 17 for example, the adjusted quarternion eQD is generated by the above processing example. Then, the generated adjusted quarternion eQD is provided for the shaking change processing in step ST16.
- the shaking change processing in step ST16 can be considered to apply, for example, the adjusted quarternion eQD obtained by the processing of FIGS. 20, 21, and 22 to an image to add shaking. If the adjusted quaternion eQD is generated as a value for removing sway or partially removing sway, not for swaying, what is the sway change processing in step ST16? The adjusted quaternion eQD is applied to the image to remove the sway. , Can be considered to be reduced.
- step ST16 the CPU 71 rotates the image of the celestial sphere model MT to which the frame image is pasted in step ST13 by using the quatertern eQD (#LN) after adjusting for each line, thereby rotating the entire image. Try to change the sway.
- the CPU 71 changes the shaking for each area according to the area information iAR and the parameter PRM2. That is, first, the boundary region AR3 is specified from the regions AR1 and AR2 grasped from the region information iAR. Alternatively, the area AR3 is specified from the information of the boundary area AR3 set together with the areas AR1 and AR2 in the area setting process of step ST42 and included in the area information iAR. Further, based on the parameter PRM2, the amount of shaking given to the area AR2 (for example, the difference in the amount of shaking from the area AR1), the period of shaking, the direction of shaking, and the like are grasped.
- the parameter PRM2 the amount of shaking given to the area AR2 (for example, the difference in the amount of shaking from the area AR1), the period of shaking, the direction of shaking, and the like are grasped.
- the celestial sphere model MT of the pixels in the range of the area AR1 for each frame Allow movement of the coordinate points above.
- the sway change from the above-mentioned process P1 to the process P6 is executed by the overall sway change due to the rotation of the celestial sphere model MT and the sway change due to the coordinate point movement.
- the region AR1 is shaken and the region AR2 is not shaken (or slightly shaken).
- processing P1, processing P3, processing P5 Further, by further shaking the region AR1 after adding the shaking by the rotation of the celestial sphere model MT, the region AR1 is shaken greatly and the region AR2 is shaken slightly.
- processing P2, processing P4 Further, by removing the shaking by rotating the celestial sphere model MT and then shaking the areas AR1 and AR2 with different amounts of shaking by moving the coordinate points, the area AR1 can generate a large shaking image and the area AR2 can generate a small shaking image. (For example, processing P2, processing P4)
- step ST18 The image of the celestial sphere model hMT whose shaking has been changed as described above is sent to the processing of step ST18.
- the CPU 71 projects the image of the celestial sphere model hMT whose shaking has been changed onto a plane and cuts it out to obtain an image (output moving image data oPD) whose shaking has been changed.
- the sway change is realized by the rotation of the celestial sphere model MT, and by using the celestial sphere model MT, the trapezoidal shape is not formed no matter where it is cut out, and as a result, the trapezoidal distortion is also eliminated.
- the celestial sphere model MT has no lens distortion because the range that can be seen by an ideal pinhole camera is pasted on the celestial sphere.
- the focal plane distortion correction is also eliminated by rotating the celestial sphere model MT according to the adjusted quarternion eQD (#LN) based on the quarternion QD (#LN) for each line. Furthermore, since the quarternion QD (#LN) corresponds to the exposure center of gravity of each line, the blur is inconspicuous in the image.
- FIG. 23A shows an example of the rectangular coordinate plane 131 to be projected in a plane. Let each coordinate of the image projected on the plane be (x, y).
- the coordinate plane 131 is arranged (normalized) in the three-dimensional space so as to be in contact with the celestial sphere model MT in the center. That is, the center of the coordinate plane 131 is arranged at a position that coincides with the center of the celestial sphere model MT and is in contact with the celestial sphere model MT.
- the coordinates are normalized based on the zoom magnification and the size of the cutout area. For example, when the horizontal coordinates of the coordinate plane 131 are 0 to outh and the vertical coordinates are 0 to outv as shown in FIG. 23A, outh and outv are the image sizes. Then, for example, the coordinates are normalized by the following equation.
- min (A, B) is a function that returns the smaller value of A and B.
- "zoom” is a parameter for controlling enlargement / reduction.
- xnorm, ynorm, and znorm are normalized x, y, and z coordinates. According to each of the above equations (Equation 12), the coordinates of the coordinate plane 131 are normalized to the coordinates on the spherical surface of the hemisphere having a radius of 1.0.
- the coordinate plane 131 is rotated by rotation matrix calculation for the rotation for obtaining the orientation of the cutout region. That is, the following rotation matrix (Equation 13) is used to rotate the pan angle, tilt angle, and roll angle.
- the pan angle is a rotation angle that rotates the coordinates around the z-axis.
- the tilt angle is a rotation angle for rotating the coordinates around the x-axis, and the roll angle is a rotation angle for rotating the coordinates around the y-axis.
- These coordinates are used to calculate the celestial sphere corresponding points in perspective projection.
- the coordinate plane 131 is perspectively projected onto the surface of the celestial sphere (region 132). That is, when a straight line is drawn from the coordinates toward the center of the celestial sphere, the point that intersects the sphere is found.
- Each coordinate is calculated as follows.
- xsph, ysph, and zsph are coordinates obtained by projecting the coordinates on the coordinate plane 131 onto the coordinates on the surface of the celestial sphere model MT. Image data projected in a plane can be obtained in this relationship.
- the cutout area for the image projected on the plane by the above method is set in step ST17 of FIG.
- the cutout area information CRC in the current frame is set based on the tracking process by image analysis (subject recognition) and the cutout area instruction information CRC according to the user operation.
- FIGS. 25A and 25B show the cutout area information CRA set for the image of a certain frame in the state of the frame.
- Such cutout area instruction information CRC is set for each frame.
- the cutout area information CRA also reflects the instruction of the aspect ratio of the image by the user or automatic control.
- the cutout area information CRA is reflected in the process of step ST18. That is, as described above, the region corresponding to the cutout region information CRA is projected on the celestial sphere model MT in a plane, and the output moving image data oPD is obtained.
- the output moving image data oPD obtained in this way is moving image data in which the shaking change processing is performed in step ST16 so that the shaking state is different for each area.
- an image to which shaking is added is displayed as a shaking effect. Therefore, for example, when the user performs an operation for setting the parameters PRM1 and PRM2 and the area, the image is obtained by adding a shaking effect according to the user's intention for each area.
- the image may be an image in which the area is automatically set and the degree of shaking according to the subject is added to each area.
- Such moving image data is displayed or saved as an image with a shaking effect.
- the area setting unit 104 (ST42) for setting a plurality of areas in the image of the input moving image data and the area set by the area setting unit 104 for the shaking state appearing in the output moving image data. It is provided with a shake changing unit 101 (ST30) that performs a shake changing process so as to be in a different state each time. Therefore, for example, it is possible to form a region without shaking and a region with shaking in the screen, or to form a region with small shaking and a region with large shaking.
- the shaking changing unit 101 adds shaking for one area (for example, area AR1) set by the area setting unit 104 (ST42) and reduces shaking for the other area (for example, area AR2).
- An example of performing the shaking change processing is described (for example, processing P3).
- the shaking changing unit 101 adds shaking to one area (for example, area AR1) set by the area setting unit 104 (ST42), and one area for another area (for example, area AR2).
- An example of performing a shaking change process for adding smaller shaking has been described (for example, process P2, process P4).
- process P2, process P4 it is possible to set the magnitude of the shaking according to the subject in the image while increasing the shaking in the entire image, to achieve both visibility and effect, and to realize a new image effect due to the difference in the degree of shaking. ..
- the shaking changing unit 101 reduces the shaking for one area (for example, area AR1) set by the area setting unit 104 (ST42), and one area for another area (for example, area AR2).
- An example of performing a shaking change process for reducing shaking with a larger reduction amount has been described (for example, process P6).
- the shaking changing unit 101 adds or reduces shaking to one of one area (for example, area AR1) and another area (for example, area AR2) set by the area setting unit 104 (ST42).
- An example of performing the shaking change processing has been described (for example, processing P1, processing P5). This also makes it possible to set the magnitude of the shaking in the image according to the subject and the like, to achieve both visibility and a directing effect, and to realize a new image effect due to the difference in the degree of shaking. It is also possible to utilize the original shaking as a shaking effect in the other area.
- the UI processing unit 103 (ST40) for detecting the operation information related to the shaking change is provided, and the area setting unit 104 describes an example of setting a plurality of areas based on the operation information detected by the UI processing unit 103. (See FIG. 6).
- the area setting unit 104 describes an example of setting a plurality of areas based on the operation information detected by the UI processing unit 103. (See FIG. 6).
- a plurality of areas are set according to the user's intention, and an image effect in which the degree of shaking differs in each area is realized. Therefore, it is possible to produce a shaking effect that reflects the user's image editing intention.
- the area setting unit 104 has described an example of setting a plurality of areas based on the image analysis of the input moving image data (see FIG. 7). For example, subject recognition, detection of a specific subject, composition determination, determination of a fixed object such as a background, etc. are performed, and an area that should be shaken and an area that is not shaken are separated, and the area is automatically set. As a result, it is possible to set an appropriate area according to the image content and change the shaking for each area, and the user can easily obtain a shaking effect image in which the shaking is changed for each area.
- the shaking change unit 101 attaches each frame of the input moving image data to be processed to the celestial sphere model MT, and rotates the shaking information (adjusted quaternion eQD) corresponding to each frame to change the shaking.
- the shaking information adjusted quaternion eQD
- an example is given in which the amount of shaking for each region is changed by moving the coordinate points on the celestial sphere model MT in each frame of the moving image data iPD (see FIGS. 9 to 14).
- This makes it possible to partially increase the sway after changing the sway as a whole. Therefore, it is possible to change the shaking state for each region by enlarging / reducing by moving the coordinate points while making use of the shaking change due to the rotation of the celestial sphere model MT.
- the shaking by enlarging and reducing the continuity of the image is not interrupted at the boundary of the region, so that the image quality can be kept good without performing complicated joint processing.
- the movement of the coordinate points is a process of moving the coordinate points of the pixels of the boundary region between one region and the other region set by the region setting unit 104 (FIGS. 9 to 9). 14).
- the position change (that is, shaking) of a part of the area can be realized by a relatively simple process of changing the coordinates of the pixels only in the boundary area, that is, enlarging / reducing the image of the boundary area by moving the coordinate points.
- each shaking can be expressed as a natural image.
- the shaking for each region by rotating the celestial sphere model MT without moving the coordinate points.
- the image of the area AR1 is attached to the first celestial sphere model
- the image of the area AR2 is attached to the second celestial sphere model
- the amount of shaking, the period, and the rotation in the shaking direction indicated by the parameters PRM2 are calculated, respectively. conduct.
- the images projected from each celestial sphere model are combined. If this is performed for each frame, it is possible to obtain output moving image data oPD having a different shaking state for each area.
- the program of the embodiment is a program that causes, for example, a CPU, a DSP, or a device including these to execute the process described with reference to FIG. That is, in the program of the embodiment, the area setting process (ST42) for setting a plurality of areas in the image of the input moving image data and the shaking state appearing in the output moving image data are different for each area set in the area setting process.
- This is a program that causes the information processing apparatus to execute the shaking change processing (ST30) so as to be.
- the above-mentioned image processing device TDx can be realized in a device such as a mobile terminal 2, a personal computer 3, or an image pickup device 1.
- a program that realizes such an image processing device TDx can be recorded in advance in an HDD as a recording medium built in a device such as a computer device, a ROM in a microcomputer having a CPU, or the like.
- a recording medium built in a device such as a computer device, a ROM in a microcomputer having a CPU, or the like.
- flexible discs CD-ROMs (Compact Disc Read Only Memory), MO (Magneto Optical) discs, DVDs (Digital Versatile Discs), Blu-ray discs (Blu-ray Discs (registered trademarks)), magnetic discs, semiconductor memories, It can be temporarily or permanently stored (recorded) on a removable recording medium such as a memory card.
- a removable recording medium can be provided as so-called package software.
- it can also be downloaded from a download site via a network such as a LAN (Local Area Network) or the Internet.
- LAN Local Area Network
- the personal computer or the like can function as the image processing device of the present disclosure. Can be done.
- the present technology can also adopt the following configurations.
- An area setting unit that sets multiple areas in the image of the input video data
- An image processing device including a shaking changing unit that performs shaking changing processing so that the shaking state appearing in the output moving image data is different for each area set by the area setting unit.
- the shaking change part is The image processing apparatus according to (1) above, wherein shaking is added to one area set by the area setting unit, and shaking change processing is performed to reduce the shaking of another area set by the area setting unit.
- the shaking change part is Shake change processing is performed to add shaking to one area set by the area setting unit and to add shaking smaller than the one area to the other areas set by the area setting unit (1) or (2). ).
- the image processing apparatus is
- the shaking change part is Shake change processing is performed to reduce shaking in one area set by the area setting unit and to reduce shaking in another area set by the area setting unit with a reduction amount larger than that in the one area (1).
- the image processing apparatus according to any one of (3).
- the shaking change part is The image processing apparatus according to any one of (1) to (4) above, which performs a shaking change process for adding or reducing shaking to one of one area and the other area set by the area setting unit.
- (6) Equipped with a user interface processing unit that detects operation information related to shaking changes The image processing device according to any one of (1) to (5) above, wherein the area setting unit sets a plurality of areas based on operation information detected by the user interface processing unit.
- the image processing apparatus according to any one of (1) to (6) above, wherein the area setting unit sets a plurality of areas based on image analysis of the input moving image data.
- the shaking changing unit changes the shaking of the entire image by pasting each frame of the input moving image data to the celestial sphere model and rotating each frame using the shaking information corresponding to each frame (1).
- the image processing apparatus according to any one of (7) to (7).
- (9) The image processing apparatus according to (8) above, wherein the shaking changing unit changes the amount of shaking for each region by moving coordinate points on the celestial sphere model in each frame.
- the shaking change part is The image processing apparatus according to (9) above, wherein the movement of the coordinate points is a process of moving the coordinate points of the pixels of the boundary region between one region and the other region set by the region setting unit.
- the image processing device Area setting process to set multiple areas in the image of input video data, An image processing method for performing a shaking change process for making the shaking state appearing in the output moving image data different for each area set in the area setting process.
- Area setting process to set multiple areas in the image of input video data Shake change processing that makes the state of shaking appearing in the output video data different for each area set in the area setting process, and Is a program that causes the information processing device to execute.
- Imaging device 2
- Mobile terminal 3
- Personal computer 4 Server 5
- Recording medium 70
- Information processing device 71
- CPU 101
- Shake change unit 102
- Parameter setting unit 103
- UI processing unit 104
- Area setting unit 104
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Studio Devices (AREA)
- Image Processing (AREA)
Abstract
La présente invention réalise un traitement qui accroît la variété d'effets d'image produits par l'expression d'états de tremblement d'image. Le présent dispositif de traitement d'image est configuré de telle façon que les actions suivantes sont réalisées: traitement de spécification de régions dans lequel une pluralité de régions est spécifiée dans une image qui provient de données d'image animée d'entrée; et traitement de modification de vibrations dans lequel il est fait en sorte que des états de tremblement apparaissant dans des données d'image animée de sortie soient des états différents pour chaque région spécifiée par le traitement de spécification de régions.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2020-039701 | 2020-03-09 | ||
JP2020039701 | 2020-03-09 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021181965A1 true WO2021181965A1 (fr) | 2021-09-16 |
Family
ID=77671404
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2021/004160 WO2021181965A1 (fr) | 2020-03-09 | 2021-02-04 | Dispositif de traitement d'image, procédé de traitement d'image, et programme |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2021181965A1 (fr) |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2011259119A (ja) * | 2010-06-08 | 2011-12-22 | Nippon Signal Co Ltd:The | 画像処理装置 |
JP2012165338A (ja) * | 2011-02-09 | 2012-08-30 | Canon Inc | 画像処理装置、画像処理方法及びプログラム |
JP2016105534A (ja) * | 2014-12-01 | 2016-06-09 | 株式会社ザクティ | 撮像装置、及び撮像装置システム |
WO2018211781A1 (fr) * | 2017-05-18 | 2018-11-22 | ソニー株式会社 | Dispositif de capture d'image et procédé de capture d'image |
JP2018201123A (ja) * | 2017-05-26 | 2018-12-20 | 株式会社リコー | 画像処理装置、画像処理方法、プログラム |
JP2019114143A (ja) * | 2017-12-25 | 2019-07-11 | キヤノン株式会社 | 画像処理装置、画像処理方法、およびプログラム |
-
2021
- 2021-02-04 WO PCT/JP2021/004160 patent/WO2021181965A1/fr active Application Filing
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2011259119A (ja) * | 2010-06-08 | 2011-12-22 | Nippon Signal Co Ltd:The | 画像処理装置 |
JP2012165338A (ja) * | 2011-02-09 | 2012-08-30 | Canon Inc | 画像処理装置、画像処理方法及びプログラム |
JP2016105534A (ja) * | 2014-12-01 | 2016-06-09 | 株式会社ザクティ | 撮像装置、及び撮像装置システム |
WO2018211781A1 (fr) * | 2017-05-18 | 2018-11-22 | ソニー株式会社 | Dispositif de capture d'image et procédé de capture d'image |
JP2018201123A (ja) * | 2017-05-26 | 2018-12-20 | 株式会社リコー | 画像処理装置、画像処理方法、プログラム |
JP2019114143A (ja) * | 2017-12-25 | 2019-07-11 | キヤノン株式会社 | 画像処理装置、画像処理方法、およびプログラム |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7444162B2 (ja) | 画像処理装置、画像処理方法、プログラム | |
US10789676B2 (en) | Image processing device, image processing method, and program | |
CN102739955B (zh) | 能生成动态范围宽的全景图像的数据的摄像装置 | |
JP5428210B2 (ja) | 情報処理装置、撮像システム、録画制御方法及びプログラム | |
JP5133816B2 (ja) | カメラ、画像合成方法、およびプログラム | |
CN108445705B (zh) | 图像处理装置、图像处理方法 | |
JP7405131B2 (ja) | 画像処理装置、画像処理方法、プログラム | |
WO2012046371A1 (fr) | Dispositif d'affichage d'image et procédé d'affichage d'image | |
JP2015156523A (ja) | 画像処理装置、画像処理方法、プログラム | |
JP6304293B2 (ja) | 画像処理装置、画像処理方法及びプログラム | |
JP2012119804A (ja) | 画像記録装置 | |
US8736664B1 (en) | Moving frame display | |
WO2021181965A1 (fr) | Dispositif de traitement d'image, procédé de traitement d'image, et programme | |
WO2021014716A1 (fr) | Dispositif de traitement d'image, procédé de traitement d'image et programme | |
JP2020096349A (ja) | 画像処理装置、撮像装置、動画再生システム、方法およびプログラム | |
WO2021171848A1 (fr) | Dispositif de traitement d'image, procédé de traitement d'image et programme | |
WO2021181966A1 (fr) | Dispositif de traitement d'image, procédé de traitement d'image, et programme | |
JP5602818B2 (ja) | カメラ、画像再生装置、画像合成方法、およびプログラム | |
JP2021002803A (ja) | 画像処理装置、その制御方法、プログラム | |
JP5646033B2 (ja) | 画像表示装置および画像表示方法 | |
JP2010183254A (ja) | 撮像装置および被写体検出プログラム | |
JP2023042257A (ja) | 画像処理装置、画像処理方法およびプログラム | |
JP2014003715A (ja) | 撮像装置、撮像方法、及びプログラム | |
JP2012165247A (ja) | 画像処理装置、撮影装置および画像処理プログラム | |
JP2013078039A (ja) | 三次元画像を取得可能な電子機器、当該電子機器を制御するための方法、および、当該電子機器を制御するためのプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21768481 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21768481 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: JP |