WO2013080245A1 - Video processing device and video processing method - Google Patents

Video processing device and video processing method Download PDF

Info

Publication number
WO2013080245A1
WO2013080245A1 PCT/JP2011/006619 JP2011006619W WO2013080245A1 WO 2013080245 A1 WO2013080245 A1 WO 2013080245A1 JP 2011006619 W JP2011006619 W JP 2011006619W WO 2013080245 A1 WO2013080245 A1 WO 2013080245A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
frame
video signal
output
cutout
Prior art date
Application number
PCT/JP2011/006619
Other languages
French (fr)
Japanese (ja)
Inventor
諭 鈴木
Original Assignee
パナソニック株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by パナソニック株式会社 filed Critical パナソニック株式会社
Priority to PCT/JP2011/006619 priority Critical patent/WO2013080245A1/en
Publication of WO2013080245A1 publication Critical patent/WO2013080245A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/80Camera processing pipelines; Components thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders

Definitions

  • the present invention relates to a video processing apparatus and a video processing method for detecting and removing a background portion from a video acquired from a camera connected to a television or a PC and automatically displaying a subject in an appropriate size at the center of the screen.
  • the video analysis unit that detects the background from the acquired video, the video adjustment unit that adjusts the angle of view, such as cutting out the acquired video from the analysis result, and the camera position and zoom (zoom) are controlled automatically. The angle is adjusted.
  • a video device such as a camera-integrated VCR (Video Cassette Recorder) or a stationary VCR
  • VCR Video Cassette Recorder
  • a stationary VCR by specifying a moving subject within the screen, the subject is always projected at a certain position on the screen, for example, the center position.
  • An image processing apparatus for obtaining a video signal is disclosed (see, for example, Patent Document 1).
  • This conventional image processing device detects a moving vector of a subject based on a CCD (Charge Coupled Device) image signal in order to detect a moving subject, and is set so that the subject is always located at the center of the cutout frame. To do.
  • the conventional image processing apparatus always projects the subject at a fixed position on the screen, for example, the center position.
  • the conventional image processing apparatus detects the motion vector of the subject as the subject changes at times t1, t2, and t3.
  • the conventional image processing apparatus determines a certain area including the subject as a cutout frame from the screen configured by the video signal at time t1 on the screen 1202.
  • the conventional image processing apparatus moves the cutout frame according to the motion vector, and shows that the subject is always displayed at the center position of the screen of the image processing apparatus.
  • the change in the cutout frame may be small due to image processing.
  • the background variation is small.
  • a minute fluctuation of the background occurs in a short time. For a user, it may be accompanied by considerable discomfort when gazing at a screen in which minute background fluctuations are continued.
  • the video processing apparatus of the present invention includes a camera, a video acquisition unit, a video analysis unit, a video adjustment unit, and a video output unit.
  • the camera captures an image including a subject.
  • the video acquisition unit acquires video from the camera and outputs a video signal.
  • the video analysis unit analyzes the video in the time direction with respect to the video signal.
  • the video adjustment unit calculates an enlarged video signal based on the video signal according to the cutout frame calculated by the video analysis unit based on the video signal and the enlargement ratio calculated based on the display screen size signal. Output.
  • the video output unit outputs an output video signal based on the video signal and the enlarged video signal output from the video adjustment unit.
  • the video analysis unit detects, for each successive frame, an area where the difference between successive frames exceeds a predetermined threshold. In addition, the video analysis unit calculates an included cutout frame so as to include all of the plurality of cutout frames in a plurality of predetermined consecutive frames.
  • the video processing apparatus of the present invention captures a video including a subject from a camera, detects a background portion, and substantially fixes the background portion in terms of time.
  • the video processing apparatus can cut out the subject and display it as large as possible on the entire screen. Therefore, even when the user gazes at the screen, discomfort can be less likely to occur.
  • the video processing method of the present invention is a video processing method having the following steps.
  • the video including the subject is shot, the video is acquired, the video is converted into a video signal and output, the video is analyzed in the time direction with respect to the video signal, the cutout frame is calculated based on the video signal, An area where the difference between successive frames exceeds a predetermined threshold is detected for each successive frame as a cutout frame, and the included cutout frame is calculated by including all of the plurality of cutout frames in a plurality of predetermined continuous frames.
  • an enlarged video signal is calculated based on the video signal, and based on the video signal and the enlarged video signal, Output video signal.
  • FIG. 1 is a diagram showing an example of setting cut frames for continuous frames in the embodiment of the present invention.
  • FIG. 2 is a diagram showing an example in which cutout frames of continuous frames are enlarged and displayed as an enlarged display image on the entire screen in the embodiment of the present invention.
  • FIG. 3 is a block diagram showing the configuration of the video processing apparatus according to the embodiment of the present invention.
  • FIG. 4 is a block diagram showing a detailed configuration of the video analysis unit in the embodiment of the present invention.
  • FIG. 5 is a block diagram showing a detailed configuration of the position information determination unit in the present embodiment.
  • FIG. 6 is a diagram for explaining a method of calculating the boundary between the moving pixel region and the still pixel region in the present embodiment.
  • FIG. 1 is a diagram showing an example of setting cut frames for continuous frames in the embodiment of the present invention.
  • FIG. 2 is a diagram showing an example in which cutout frames of continuous frames are enlarged and displayed as an enlarged display image on the entire screen in the embodiment of the present invention
  • FIG. 7 is a diagram showing the relationship between the cutout frame and the included cutout frame in the present embodiment.
  • FIG. 8 is a block diagram showing a detailed configuration of the video adjustment unit in the present embodiment.
  • FIG. 9A shows an example in which an enlarged cutout frame is calculated from an included cutout frame in the present embodiment, and the width of the included cutout frame on the screen in the vertical direction of the screen included in the display screen size signal. It is a figure which shows the case where it is larger than the width
  • FIG. 9B shows an example in which an enlarged cutout frame is calculated from the cutout frame included in the present embodiment, and the horizontal width on the screen of the cutout frame included in the screen includes the display screen size signal.
  • FIG. 10 is a diagram showing a predetermined frame of the enlarged video signal displayed on the screen in the present embodiment.
  • FIG. 11 is a diagram illustrating an example of a display screen during image processing when the video output unit according to the present embodiment does not use an output flag.
  • FIG. 12 is a diagram illustrating an example of a display screen during image processing when the video output unit according to the present embodiment uses an output flag.
  • FIG. 13 is a diagram illustrating an example in which the video output unit according to the present embodiment uses a frame memory to output a stable display screen even during image processing.
  • FIG. 14A is a conceptual diagram for explaining the zoom control parameters output by the video analysis unit in the present embodiment.
  • FIG. 14A is a conceptual diagram for explaining the zoom control parameters output by the video analysis unit in the present embodiment.
  • FIG. 14B is a conceptual diagram for explaining the zoom control parameters output by the video analysis unit in the present embodiment.
  • FIG. 15 is a flowchart showing the procedure of the video processing method in the present embodiment.
  • FIG. 16 is a diagram conceptually showing image processing of a conventional image processing apparatus.
  • FIG. 1 is a diagram showing an example of setting continuous frame cutout frames 1301, 1302, and 1303 in the embodiment of the present invention.
  • a method of setting a moving portion of the subject as cut frames 1301, 1302, and 1303 for each frame can be considered.
  • the clipping frames 1301, 1302, and 1303 including the moving part of the subject can be determined using the image difference between frames. That is, based on the detection result of the image difference between frames, the moving pixel region and the still pixel region of each frame are determined.
  • a rectangular shape including a moving pixel region is determined as the cutout frames 1301, 1302, and 1303.
  • the horizontal axis in FIG. 1 is a frame, and shows only a part of a continuous frame.
  • N and m are arbitrary integers.
  • the positions and sizes of the cutout frames 1301, 1302, and 1303 in the frame N, the frame (N + m), and the frame (N + 2m) are the latest frame preceding each frame.
  • the amount of movement of the subject in each frame is the number of frames of the most recent frame preceding each frame for determining the amount of movement of the subject can be set as appropriate based on the specifications of the system to be applied.
  • FIG. 1 it is assumed that the amount of movement of the subject is relatively small. As shown in FIG. 1, the positions and sizes of the cutting frames 1301, 1302, and 1303 are slightly different.
  • FIG. 2 is a diagram showing an example in which cut-out frames 1301, 1302, and 1303 of continuous frames in the embodiment of the present invention are enlarged by image processing and displayed as an enlarged display image on the entire screen.
  • Screens 1401, 1402, and 1403 correspond to cut frames 1301, 1302, and 1303, respectively.
  • the cutout frames 1301, 1302, and 1303 are enlarged as they are, but the screens 1401, 1402, and 1403 make it easy to determine the subject.
  • the rectangular shape including the moving pixel region is determined as the cutout frames 1301, 1302, and 1303.
  • the rectangular shape including the moving pixel region is not enlarged and displayed as a cutout frame as it is. This is because the background tends to fluctuate between frames as shown in FIG. Details of the method for determining the cutout frames 1301, 1302, and 1303 and the method for determining the enlarged display image will be described below.
  • FIG. 3 is a block diagram showing the configuration of the video processing apparatus 100 according to the embodiment of the present invention.
  • the video processing apparatus 100 includes a camera 101, a video acquisition unit 102, a video analysis unit 103, a video adjustment unit 104, a camera control unit 106, and a video output unit 107.
  • the camera 101 captures a video including the subject and outputs the captured signal to the video acquisition unit 102.
  • the video acquisition unit 102 converts a signal input from the camera 101 into a video signal 301.
  • the video analysis unit 103 analyzes the video in the time direction with respect to the video signal 301. Specifically, the video analysis unit 103 receives the video signal 301 and detects a difference between successive frames of the video signal 301 to separate the background portion from the moving subject. Then, the video analysis unit 103 calculates a cutting frame based on the video signal 301 in accordance with the detected background portion and the moving subject. That is, the video analysis unit 103 detects, for each successive frame, an area where the difference between successive frames exceeds a predetermined threshold as a cutout frame. Then, the video analysis unit 103 calculates the included cutout frame 306 so as to include all of the plurality of cutout frames in a predetermined plurality of frames.
  • the video analysis unit 103 can adjust the zoom of the camera 101 to the wide side (that is, the zoom lens to the wide angle side). Judge as necessary. Here, zoom is performed using a zoom lens, but the image received by the image sensor may be calculated electronically. Then, the video analysis unit 103 outputs the zoom control parameter 307 of the camera 101 to the camera control unit 106. Details of the operation of the video analysis unit 103 will be described later.
  • the camera control unit 106 transmits a zoom control signal to the camera 101.
  • the camera control unit 106 adjusts the zoom value of the camera 101 based on the zoom control parameter 307 output from the video analysis unit 103.
  • the zoom control parameter 307 is used to control the zoom value of the camera 101 when the ratio of the background portion in the video area is small in the video analysis unit 103 and the subject is outside the shooting range. Further, when the moving subject is extremely small with respect to the shooting range and is difficult for the user to visually recognize, the position of the camera 101 may be controlled so that the subject moves to the center of the screen. Then, the zoom may be adjusted to the telephoto side (that is, the zoom lens is set to the telephoto side), and control may be performed so that the subject is displayed larger on the screen. Such control makes it easier for the user to recognize the subject.
  • the video adjustment unit 104 inputs the video signal 301, the included cutout frame 306, and the display screen size signal 105. Then, the video adjustment unit 104 outputs an enlarged video signal 309 based on the video signal 301 according to the included cutout frame 306 and the enlargement ratio calculated based on the display screen size signal 105. That is, the video adjustment unit 104 outputs an enlarged video signal 309 including a subject cut out except for the background portion to the video output unit 107 according to the subject and background portion separated by the video analysis unit 103. Further, the video output unit 107 receives the video signal 301, the enlarged video signal 309, and the output flag 308 output from the video analysis unit 103, and outputs the output video signal 108.
  • the display screen size signal 105 is, for example, a signal input from a control unit (not shown), and the aspect ratio of the screen displayed by the output video signal 108 of the video processing apparatus 100 and the screen display size of the output video signal 108. It contains information indicating.
  • the apparatus on which the output video signal 108 is displayed is an apparatus connected to the video processing apparatus 100 according to the present embodiment by a communication system or the like, and is used by, for example, another user who uses the video conference system. It may be a device that performs. Further, for example, it may be a monitor device directly connected to the video processing device 100 by a video signal connection cable or the like. Furthermore, the present invention can also be applied to a configuration in which the output video signal 108 is displayed on a video processing device in which an image display unit is incorporated.
  • FIG. 4 is a block diagram showing a detailed configuration of the video analysis unit 103 in the embodiment of the present invention.
  • the video analysis unit 103 includes a video storage unit 302, a difference calculation unit 303, a background detection unit 304, a position information determination unit 305, and a cut-out position storage unit 310.
  • the video storage unit 302 and the difference calculation unit 303 each input the video signal 301 output from the video acquisition unit 102.
  • the video storage unit 302 stores one frame of the video signal 301. Then, the video storage unit 302 delays the input video signal 301 by one frame and outputs it to the difference calculation unit 303.
  • the video storage unit 302 is a frame memory, and may be configured by, for example, a semiconductor memory.
  • the difference calculation unit 303 detects a difference between the video signal 301a delayed from the video storage unit 302 by one frame and the original video signal 301 that has not been delayed.
  • the video storage unit 302 outputs a frame (N ⁇ 1) preceding the current frame N.
  • the difference calculation unit 303 detects the difference between the video signals of the frame N and the frame (N ⁇ 1) for each pixel. Then, the difference calculation unit 303 outputs the detected difference result to the background detection unit 304.
  • the background detection unit 304 determines that each pixel is a moving pixel or a still pixel by comparing the difference for each pixel with a predetermined threshold.
  • the predetermined threshold is, for example, about 10 levels.
  • each pixel is expressed by 8 bits (when a video signal is handled with 256 levels of accuracy).
  • the difference for each pixel exceeds a predetermined threshold, the pixel is a moving pixel. That is, the moving pixels in the frame N and the frame (N ⁇ 1) are regarded as moving pixels.
  • the background detection unit 304 divides the moving pixel region and the still pixel region in the current frame N, and outputs them to the position information determination unit 305.
  • FIG. 5 is a block diagram illustrating a detailed configuration of the position information determination unit 305 in the present embodiment.
  • the position information determination unit 305 includes a coordinate calculation unit 351 and a change amount calculation unit 352.
  • the position information determination unit 305 outputs the included cutout frame 306 to the video adjustment unit 104 using the cutout position storage unit 310.
  • FIG. 6 is a diagram for explaining a method of calculating the boundary 632 between the moving pixel region and the stationary pixel region in the present embodiment.
  • the coordinate calculation unit 351 calculates a boundary 632 between the background portion and the subject portion from the moving pixel region and the still pixel region in the current frame N. 6 indicates a boundary 632 between the moving pixel region and the stationary pixel region.
  • a boundary 632 has a rectangular shape indicating the boundary between the background portion and the subject portion, and indicates a cutout frame.
  • the extraction start coordinate 630 and the extraction end coordinate 631 use the upper left end coordinate and the lower right end coordinate of the boundary 632 so that the boundary 632 having a rectangular shape can be specified on the screen.
  • This is an example for reducing the memory amount of the cutout position storage unit 310 as much as possible, and the present embodiment is not limited to using these coordinates.
  • the coordinates of the upper right end and the lower left end of the boundary 632 having a rectangular shape may be used.
  • the number of pixels in the vertical direction and the horizontal direction of the boundary 632 which is a rectangular shape, and the center coordinates may be used. In this way, a cutout frame having a rectangular shape is calculated from the boundary 632 between the background portion and the subject portion.
  • the coordinate calculation unit 351 outputs the cut coordinates 312 including the cut start coordinates 630 and the cut end coordinates 631 to the cut position storage unit 310 and the change amount calculation unit 352.
  • the cut-out position storage unit 310 stores cut-out coordinates 313 for nine frames from the frame (N-1) preceding the current frame N to the frame (N-9).
  • the cutout position storage unit 310 calculates the amount of change of the cutout coordinates 313 for nine frames from the stored frame (N-1) to the frame (N-9).
  • the cut-out position storage unit 310 updates the stored content. That is, the cutout position storage unit 310 stores the cutout coordinates 313 for nine frames from the current frame N to the frame (N-8).
  • the change amount calculation unit 352 applies the cut-out coordinates 312 of each of the current frame N to the frame (N-9) and the latest frame preceding each of the current frame N to the frame (N-9), for example, the current frame N. Compares the cut-out coordinates 313 of the frame (N ⁇ 1) and verifies the movement amount of both. Then, when the amount of movement of both is equal to or less than a predetermined amount of movement threshold, the change amount calculation unit 352 extracts 10 frames of cutout coordinates 312 for 10 frames from the current frame N to a frame (N-9) preceding the current frame N. 313 is output to the coordinate calculation unit 351 as post-determination cut-out coordinates 314.
  • the predetermined movement amount threshold is preferably about 200 pixels, for example.
  • the coordinate calculation unit 351 inputs the post-judgment cut-out coordinates 314, and calculates an included cut-out frame 306 so as to include all of the plurality of cut-out frames in a plurality of predetermined consecutive frames.
  • the predetermined number of consecutive frames is, for example, 10 frames.
  • the coordinate calculation unit 351 inputs post-determination cut-out coordinates 314 for 10 consecutive frames.
  • the coordinate calculation unit 351 includes all of the post-determination cut-out coordinates 314 for 10 frames, and calculates an included cut-out frame 306 that is a wider cut-out frame.
  • the shape of the included cutout frame 306 is a rectangle.
  • the shape of the included cutout frame 306 is simple, the subsequent signal processing can be performed more easily. That is, the shape of the included cutout frame 306 should not be a simple composition of the post-judgment cutout coordinates 314. However, if the signal processing capability of each part of the video processing apparatus 100 in the present embodiment is improved and there is a margin in the signal processing capability thereafter, the shape of the included extraction frame 306 is simply extracted after the determination coordinates. It is good also as composition of 314.
  • the change amount calculation unit 352 includes the cut-out coordinates 312 of each of the current frame N to the frame (N-9) and the latest frame preceding each of the current frame N to the frame (N-9), for example, the current frame N. If the movement amount of both of the frames N exceeds the predetermined movement amount threshold value by comparing the cut coordinates 313 of the frame (N-1), for example, the frame N-1 to the current frame N When the movement amount to exceeds the predetermined movement amount threshold, the cut-out coordinates 312 of the current frame N are excluded from the cut-out coordinates 314 after determination.
  • the change amount calculation unit 352 outputs the cut coordinates 313 of the frame (N ⁇ 1) preceding the current frame N as the cut coordinates 314 after determination instead of the cut coordinates 312 of the current frame N. Further, the change amount calculation unit 352 determines to the coordinate calculation unit 351 the extracted coordinates 312 for nine frames from the frame (N ⁇ 1) preceding the current frame N to the frame (N-9) preceding the current frame N. Output as post-cutout coordinates 314. That is, data is output twice for the cut-out coordinates 313 of the frame (N ⁇ 1).
  • the video analysis unit 103 sets the cut frames of the continuous frames to all of a plurality of predetermined continuous frames. Remove from the cropping frame. Even in this case, the coordinate calculation unit 351 inputs post-determination cut-out coordinates 314 corresponding to 10 frames. Then, the coordinate calculation unit 351 includes all of the post-judgment cutout coordinates 314 corresponding to 10 frames, and calculates an included cutout frame 306 that is a wider cutout frame.
  • FIG. 7 is a diagram showing the relationship between the cut frames 601, 602, and 603 and the included cut frame 306 in the present embodiment.
  • FIG. 7 illustrates the case of three frames as an example for the sake of simplicity.
  • the coordinate calculation unit 351 includes inclusion frames 601, 602, and 603 included when the included extraction frame 306 is calculated. The relationship with the cut out frame 306 is shown. As shown in FIG. 7, the coordinate calculation unit 351 calculates the minimum rectangular shape that includes all the cut frames 601, 602, and 603 as the included cut frame 306.
  • FIG. 8 is a block diagram showing a detailed configuration of the video adjustment unit 104 in the present embodiment.
  • the video adjustment unit 104 includes an enlargement ratio calculation unit 403 and a video enlargement unit 405.
  • the enlargement ratio calculation unit 403 inputs the included cutout frame 306 and the display screen size signal 105, and outputs the enlargement ratio E of the included cutout frame 306 and the enlargement start positions 360a and 360b. Then, the video enlargement unit 405 inputs the enlargement ratio E of the included cutout frame 306, the enlargement start positions 360a and 360b, and the video signal 301, and outputs the post-enlargement video signal 309.
  • the enlargement ratio E is the ratio of the enlarged cutout frame that becomes the display screen size in the post-enlargement video signal 309 and the included cutout frame 306.
  • the enlargement start positions 360a and 360b may be, for example, the upper left end of the enlargement cutout frame.
  • FIG. 9A shows an example in which the enlarged cutout frame 350a is calculated from the cutout frame 306a included in the present embodiment, and the vertical width of the included cutout frame 306a on the screen is included in the display screen size signal. It is a figure which shows the case where it is larger than the width
  • FIG. 9B shows an example in which the enlarged cutout frame 350b is calculated from the cutout frame 306b included in the present embodiment, and the horizontal width on the screen of the cutout frame 306b included is included in the display screen size signal.
  • the coordinates of the upper left end (Xs, Ys) and the coordinates of the lower right end (Xe, Ye) are used to specify the included cutout frames 306a, 306b on the screen.
  • An operation for obtaining each of the enlargement ratio E and the enlargement start positions 360a and 360b in the above two cases will be described.
  • the included cutout frame 306a is expanded in the horizontal or vertical direction of the screen so as to match the aspect ratio of the screen on which the enlarged video signal 309 is displayed.
  • the enlargement ratio E for enlarging the image signal 309 after enlargement so as to match the aspect ratio of the screen on which the image signal 309 is displayed can be expressed as follows.
  • min (A, B) is a function for selecting the smaller numerical value of the argument A and the argument B.
  • E min (Dx ⁇ 1 /
  • E min (Dx ⁇ 1 /
  • the included cutout frame 306a is expanded by twice in the horizontal direction, the enlarged cutout frame 350a can be calculated.
  • the included cutout frame 306a is expanded by an equal width on both sides in the horizontal direction. In this way, the enlargement start position 360a can be obtained.
  • the enlargement ratio E is obtained assuming specific numerical values.
  • E min (1920 ⁇ 1 /
  • the included cutout frame 306a is expanded upward in the vertical direction. This is because, in the present embodiment, the included cutout frames 306a and 306b are along the lower end of the display screen. In this way, the enlargement start position 360b can be obtained.
  • the included cutout frames 306a and 306b do not necessarily have to be along the lower end of the display screen. In that case, the included cutout frame 306b is expanded by an equal width on both sides in the vertical direction.
  • the video enlargement unit 405 can calculate the enlargement cut frames 350a and 350b, that is, the enlargement ratio E and the enlargement start positions 360a and 360b, and output the enlarged image signal 309. That is, the video adjustment unit 104 expands the cutout frame 306 included so as to have an aspect ratio equivalent to the aspect ratio of the display screen, and outputs the expanded cutout video signal 309 to the video output unit 107.
  • FIG. 10 is a diagram in which predetermined frames of the enlarged video signal 309 are displayed on the screens 601a, 602a, and 603a.
  • FIG. 10 shows screens 601a, 602a, and 603a in a frame N, a frame (N + m), and a frame (N + 2m) (N and m are arbitrary integers).
  • a constant region is always displayed as a background even if the frame changes.
  • the user is less likely to experience discomfort even if he / she gazes at the continuous screens 601a, 602a, and 603a.
  • it can be enlarged and displayed almost at the center of the screens 601a, 602a, and 603a for displaying the subject. As a result, the user can accurately capture the subject.
  • the video analysis unit 103 outputs an output flag 308 based on the included enlarged video signal 309 output from the video adjustment unit 104. Specifically, the video analysis unit 103 generates an enlarged video signal from a predetermined continuous frame including the current frame N of the enlarged video signal 309 output from the video adjustment unit 104 and a plurality of frames preceding the current frame N. After calculating 309, the output flag 308 is output. That is, for example, using a predetermined continuous frame, which is 10 frames, the video adjustment unit 104 extends the included cutout frame 306 corresponding to the current frame N and a plurality of frames preceding the current frame N. Then, the video adjusting unit 104 outputs an enlarged video signal 309 based on the video signal 301 in accordance with the plurality of included cutout frames 306 described above.
  • FIG. 11 is a diagram illustrating an example of a display screen during image processing when the video output unit 107 in the present embodiment does not use the output flag 308.
  • the enlarged video signal 309 is also displayed from the video enlargement unit 405 during the period t1 to t2 (during the enlargement process) during signal processing for 10 frames including the current frame N.
  • the video adjustment unit 104 and the video enlargement unit 405 perform the expansion processing of the plurality of included cutout frames 306, the size of the subject changes.
  • the output flag 308 is not used in this way, the enlarged video signal 309 is output while the included extraction frame 306 is being expanded, and thus the user feels uncomfortable when watching such a screen. There is a possibility to give.
  • FIG. 12 is a diagram illustrating an example of a display screen during image processing when the video output unit 107 according to the present embodiment uses the output flag 308.
  • the video enlargement unit 405 outputs the video signal 301.
  • the period t3 expansion process completion
  • the expansion process of the included cut frame 306 corresponding to a plurality of predetermined consecutive frames including the current frame N is completed, the expanded video signal after the expansion process 309 is output.
  • the video output unit 107 calculates the enlarged video signal 309 from a predetermined continuous frame including the current frame N of the video signal 301 and a plurality of frames preceding the current frame N. Thereafter, the output video signal 108 may be switched from the video signal 301 based on the video from the camera 101 to the enlarged video signal 309 based on the output flag 308 output from the video analysis unit 103.
  • the cut frame of the subsequent frame (N ⁇ 1) following the preceding frame N of the continuous frame is changed by the change amount.
  • the video output unit 107 includes, from the video signal 301 based on the video from the camera 101, the included clipping corresponding to a plurality of predetermined consecutive frames that do not include the current frame N and precede the current frame N.
  • the output video signal 108 may be switched to the post-enlargement video signal 309 including the information of the frame 306. In this way, an image in which the size of the subject being enlarged is changed is not displayed.
  • FIG. 13 is a diagram illustrating an example in which the video output unit 107 according to the present embodiment uses the frame memory to output a stable display screen during image processing.
  • the frame memory includes a frame (N ⁇ 1) preceding the current frame N and included clipping frames 306 included in a plurality of predetermined consecutive frames.
  • a corresponding enlarged video signal 309 may be stored.
  • the video enlargement unit 405 corresponds to a plurality of predetermined consecutive frames before the frame (N ⁇ 1) preceding the current frame N.
  • An enlarged video signal 309 based on the information of the included cutout frame 306 is output from the frame memory.
  • the video output unit 107 has a frame memory for one frame, and the frame memory includes information on the included cutout frame 306 corresponding to a plurality of predetermined consecutive frames preceding the current frame N.
  • the post video signal 309 is stored.
  • the video output unit 107 calculates the current frame from the enlarged video signal 309 from the frame memory based on the output flag 308 output from the video analysis unit 103 after the video adjustment unit 104 calculates the enlarged video signal 309.
  • the output video signal 108 may be switched to a post-enlarged video signal 309 including information of the included cutout frame 306 corresponding to a predetermined plurality of consecutive frames including N.
  • the video enlargement unit 405 may output the enlarged video signal 309.
  • the zoom control parameter 307 output from the video analysis unit 103 will be described with reference to FIGS. 14A and 14B.
  • the included cutout frame 306c is indicated by a broken line.
  • upper left end coordinates (Xs, Ys) and lower right end coordinates (Xe, Ye) are shown.
  • the upper left end coordinates (Xs, Ys) and the lower right end coordinates (Xe, Ye) are located at the upper and lower ends of the screen. Therefore, the camera control unit 106 adjusts the zoom value of the camera 101 based on the zoom control parameter 307 output from the video analysis unit 103.
  • the zoom control parameter 307 is when the ratio of the background portion in the video area is small in the video analysis unit 103 and the subject is outside the shooting range. Is larger than a predetermined background threshold value, it is used to control the zoom value of the camera 101.
  • the ratio of the background portion in the video area may be 1 ⁇ 4 of the video area.
  • the video analysis unit 103 calculates the position of the background in the shooting area of the camera 101 based on the calculated cutout frame, and if the range occupied by the background is larger than a predetermined background threshold, the included cutout frame 306c is displayed in the video. Output to the adjustment unit 104. On the other hand, when the range occupied by the background is smaller than the predetermined background threshold, the video analysis unit 103 outputs a control parameter for adjusting the zoom of the camera 101 to the wide side to the camera control unit 106.
  • an included cutout frame 306d is shown.
  • the coordinates (Xe, Ye) of the lower right end of the included cutout frame 306d are located at the right end of the screen.
  • the camera 101 controls the zoom value of the camera 101.
  • a control parameter for adjusting the zoom to the wide side is output to the camera control unit 106.
  • the included cutout frame is displayed small on the screen.
  • the position of the camera 101 may be controlled so that the subject moves to the center of the screen.
  • the zoom may be adjusted to the telephoto side so that the subject is displayed larger on the screen. Such control makes it easier for the user to recognize the subject.
  • FIG. 15 is a flowchart showing the procedure of the video processing method in the present embodiment.
  • the camera 101 captures an image including a subject (step S601).
  • the video acquisition unit 102 acquires the video captured by the camera 101 (step S602), converts the video into the video signal 301, and outputs the video signal 301 to the video analysis unit 103, the video adjustment unit 104, and the video output unit 107 (Ste S603).
  • the video analysis unit 103 analyzes the video in the time direction with respect to the video signal 301 (step S604). Specifically, by inputting the video signal 301 and detecting a difference between frames of the video signal 301, the background portion and the moving subject are separated. That is, the background detection unit 304 determines that each pixel in the current frame N is a moving pixel or a still pixel by comparing the difference for each pixel between successive frames of the video signal 301 with a predetermined threshold.
  • the predetermined threshold is, for example, about 10 levels.
  • each pixel is expressed by 8 bits (when a video signal is handled with 256 levels of accuracy).
  • the coordinate calculation unit 351 calculates the boundary 632 between the background portion and the subject portion from the moving pixel region and the still pixel region in the current frame N.
  • the video analysis unit 103 uses the boundary 632 as a cutting frame.
  • the video analysis unit 103 calculates a cutout frame based on the video signal 301 (step S605).
  • the video analysis unit 103 calculates the cutout frame as a rectangular shape in order to simplify the signal processing.
  • the video analysis unit 103 detects, for each successive frame, an area where the difference between successive frames exceeds a predetermined threshold as a cutout frame (step S606).
  • the video analysis unit 103 calculates a cut frame 306 that is included in a plurality of predetermined frames so as to include all of the plurality of cut frames (step S607).
  • the predetermined number of frames is, for example, 10 frames.
  • the shape of the included cutout frame 306 is a rectangle so that the subsequent signal processing can be performed more easily.
  • the video analysis unit 103 may remove the cut frame from all of the plurality of cut frames in the predetermined continuous frames when the movement amount of the cut frames of the continuous frames exceeds a predetermined movement amount threshold value.
  • the predetermined movement amount threshold is preferably about 200 pixels, for example.
  • the video adjustment unit 104 inputs the video signal 301, the included cutout frame 306, and the display screen size signal 105. Then, the video adjustment unit 104 calculates an enlarged video signal 309 based on the video signal according to the included cutout frame 306 and the enlargement ratio calculated based on the display screen size signal 105 (step S608). ).
  • the display screen size signal 105 is a signal input from a control unit (not shown), for example, and the screen aspect ratio displayed by the output video signal 108 of the video processing apparatus 100 and the screen of the output video signal 108 are displayed. Information indicating the display size is included.
  • the enlargement ratio E is the ratio of the enlarged cutout frame that is the display screen size in the post-enlargement video signal 309 and the included cutout frame 306.
  • the video adjustment unit 104 expands the included cutout frame 306 so as to have an aspect ratio equivalent to the aspect ratio of the display screen, and outputs the expanded video signal 309 to the video output unit 107.
  • the expanded cut-out frame 306b is enlarged so as to be arranged at substantially the center of the screen on which the moving subject is displayed, and the enlarged video signal 309 is calculated.
  • the video adjustment unit 104 outputs the enlarged video signal 309 including the subject cut out from the background part to the video output unit 107 based on the background part separated by the video analysis part 103 (step S609).
  • the video adjustment unit 104 calculates the enlarged video signal 309 from the included cutout frame 306, and the video output unit 107 outputs the enlarged video signal 309 as the output video signal 108.
  • a constant area is always displayed as the background even if the frame changes. As a result, the user is less likely to feel discomfort even if such a screen is watched.
  • the video processing apparatus When the video processing apparatus according to the present invention transmits / receives video between different users, the video transmitted to the other party can be transmitted to the other party without moving the user according to the camera. Can be sent.
  • the background part from the video difference at different times and cutting out the part other than the background part, it is possible to detect it uniformly regardless of the type, color, or number of subjects. This is useful for a video processing apparatus when performing the above.

Abstract

A video processing device is provided with: a video acquisition unit which acquires video from a camera and outputs video signals; a video analysis unit which calculates, from the video signal, an area in which the difference between successive frames exceeds a predetermined threshold value, as a crop region; a video adjustment unit which calculates and outputs zoomed video signals, on the basis of the video signals, in accordance with the crop regions and a zoom percentage; and a video output unit which outputs output video signals on the basis of the video signals and the zoomed video signals. The video analysis unit calculates an inclusive crop region which encompasses all of the plurality of crop regions for a plurality of predetermined successive frames.

Description

映像処理装置および映像処理方法Video processing apparatus and video processing method
 本発明は、テレビやPCに接続したカメラから取得した映像に対して、背景部分を検出・除去し、被写体を画面の中央に適切な大きさに自動で表示させる映像処理装置、映像処理方法に関する。特に、取得映像から背景部分を検出する映像解析部と、解析結果から取得映像を切出すなど画角を調整する映像調整部と、カメラの位置やズーム(Zoom)を制御して、自動で画角調整を行うものである。 The present invention relates to a video processing apparatus and a video processing method for detecting and removing a background portion from a video acquired from a camera connected to a television or a PC and automatically displaying a subject in an appropriate size at the center of the screen. . In particular, the video analysis unit that detects the background from the acquired video, the video adjustment unit that adjusts the angle of view, such as cutting out the acquired video from the analysis result, and the camera position and zoom (zoom) are controlled automatically. The angle is adjusted.
 カメラ一体型VCR(Video Cassette Recorder)、据置き型VCR等の映像機器において、画面内において移動する被写体を特定することによって、被写体が常に画面の一定位置、例えば中央位置に映出されることとなる映像信号を得るための画像処理装置が開示されている(例えば、特許文献1参照)。この従来の画像処理装置は、移動する被写体を検出するために、CCD(Charge Coupled Device)の撮像信号に基づいて被写体の動きベクトルを検出し、被写体が常に切出し枠の中央に位置するように設定する。そして、図16に示すように従来の画像処理装置は、被写体を常に画面の一定位置、例えば中央位置に映出している。 In a video device such as a camera-integrated VCR (Video Cassette Recorder) or a stationary VCR, by specifying a moving subject within the screen, the subject is always projected at a certain position on the screen, for example, the center position. An image processing apparatus for obtaining a video signal is disclosed (see, for example, Patent Document 1). This conventional image processing device detects a moving vector of a subject based on a CCD (Charge Coupled Device) image signal in order to detect a moving subject, and is set so that the subject is always located at the center of the cutout frame. To do. As shown in FIG. 16, the conventional image processing apparatus always projects the subject at a fixed position on the screen, for example, the center position.
 図16に示すように、従来の画像処理装置は、被写体が時刻t1、t2、t3と変化することに伴って、被写体の動きベクトルを検出する。従来の画像処理装置は、画面1202の時刻t1において、映像信号によって構成される画面から被写体を含む一定領域を切出し枠として決定する。図16では、従来の画像処理装置は、切出し枠を動きベクトルに応じて移動させ、被写体が常に画像処理装置の画面の中央位置に映出される様子を示している。 As shown in FIG. 16, the conventional image processing apparatus detects the motion vector of the subject as the subject changes at times t1, t2, and t3. The conventional image processing apparatus determines a certain area including the subject as a cutout frame from the screen configured by the video signal at time t1 on the screen 1202. In FIG. 16, the conventional image processing apparatus moves the cutout frame according to the motion vector, and shows that the subject is always displayed at the center position of the screen of the image processing apparatus.
 ところで、被写体の移動量が比較的小さい場合には、画像処理によって切出し枠の変化が、小さい場合も生じる。この場合、切出し枠の変化が小さいことに伴って、背景の変動も小さいものになる。このような状況では、背景の微小な変動が短時間のうちに生じる。ユーザにとって、このように背景の微小な変動が連続した画面を注視することは、相当な不快感を伴う場合がある。 By the way, when the amount of movement of the subject is relatively small, the change in the cutout frame may be small due to image processing. In this case, as the change in the cutout frame is small, the background variation is small. In such a situation, a minute fluctuation of the background occurs in a short time. For a user, it may be accompanied by considerable discomfort when gazing at a screen in which minute background fluctuations are continued.
特開平06-014323号公報Japanese Patent Laid-Open No. 06-014323
 本発明の映像処理装置は、カメラと映像取得部と映像解析部と映像調整部と映像出力部とを備える。カメラは、被写体を含む映像を撮影する。映像取得部は、カメラから映像を取得し、映像信号を出力する。映像解析部は、映像信号に対して時間方向に映像を解析する。映像調整部は、映像信号に基づいて映像解析部で算出された切出し枠と、表示画面サイズ信号に基づいて算出した拡大率とに応じて、映像信号に基づいて拡大後映像信号を算出して出力する。映像出力部は、映像信号と、映像調整部が出力する拡大後映像信号とに基づいて、出力映像信号を出力する。映像解析部は、連続するフレームの差分が所定の閾値を超える領域を切出し枠として連続するフレーム毎に検出する。また、映像解析部は、所定の連続する複数のフレームにおいて、複数のすべての切出し枠を含むようにして、包含された切出し枠を算出する。 The video processing apparatus of the present invention includes a camera, a video acquisition unit, a video analysis unit, a video adjustment unit, and a video output unit. The camera captures an image including a subject. The video acquisition unit acquires video from the camera and outputs a video signal. The video analysis unit analyzes the video in the time direction with respect to the video signal. The video adjustment unit calculates an enlarged video signal based on the video signal according to the cutout frame calculated by the video analysis unit based on the video signal and the enlargement ratio calculated based on the display screen size signal. Output. The video output unit outputs an output video signal based on the video signal and the enlarged video signal output from the video adjustment unit. The video analysis unit detects, for each successive frame, an area where the difference between successive frames exceeds a predetermined threshold. In addition, the video analysis unit calculates an included cutout frame so as to include all of the plurality of cutout frames in a plurality of predetermined consecutive frames.
 このような構成により、本発明の映像処理装置は、カメラから被写体を含む映像を撮影し、背景部分を検出し、実質的に背景部分を時間的に固定化とする。そして、映像処理装置は、被写体を切り出して、できるだけ画面全体に大きく表示することができる。したがって、ユーザが画面を注視する場合でも、不快感を生じにくくできる。 With such a configuration, the video processing apparatus of the present invention captures a video including a subject from a camera, detects a background portion, and substantially fixes the background portion in terms of time. The video processing apparatus can cut out the subject and display it as large as possible on the entire screen. Therefore, even when the user gazes at the screen, discomfort can be less likely to occur.
 また、本発明の映像処理方法は、以下のステップを有する映像処理方法である。すなわち、被写体を含む映像を撮影し、映像を取得し、映像を映像信号に変換して出力し、映像信号に対して時間方向に映像を解析し、映像信号に基づいて切出し枠を算出し、連続するフレームの差分が所定の閾値を超える領域を切出し枠として連続するフレーム毎に検出し、所定の連続する複数のフレームにおいて、複数のすべての切出し枠を含むようにして、包含された切出し枠を算出し、包含された切出し枠と、表示画面サイズ信号に基づいて算出した拡大率とに応じて、映像信号に基づいて拡大後映像信号を算出し、映像信号と拡大後映像信号とに基づいて、出力映像信号を出力する。 The video processing method of the present invention is a video processing method having the following steps. In other words, the video including the subject is shot, the video is acquired, the video is converted into a video signal and output, the video is analyzed in the time direction with respect to the video signal, the cutout frame is calculated based on the video signal, An area where the difference between successive frames exceeds a predetermined threshold is detected for each successive frame as a cutout frame, and the included cutout frame is calculated by including all of the plurality of cutout frames in a plurality of predetermined continuous frames. In accordance with the included cutout frame and the enlargement ratio calculated based on the display screen size signal, an enlarged video signal is calculated based on the video signal, and based on the video signal and the enlarged video signal, Output video signal.
図1は、本発明の実施の形態における連続するフレームの切出し枠を設定する1例を示す図である。FIG. 1 is a diagram showing an example of setting cut frames for continuous frames in the embodiment of the present invention. 図2は、本発明の実施の形態における連続するフレームの切出し枠を拡大して、拡大表示映像として画面全体に表示する1例を示す図である。FIG. 2 is a diagram showing an example in which cutout frames of continuous frames are enlarged and displayed as an enlarged display image on the entire screen in the embodiment of the present invention. 図3は、本発明の実施の形態における映像処理装置の構成を示したブロック図である。FIG. 3 is a block diagram showing the configuration of the video processing apparatus according to the embodiment of the present invention. 図4は、本発明の実施の形態における映像解析部の詳細な構成を示すブロック図である。FIG. 4 is a block diagram showing a detailed configuration of the video analysis unit in the embodiment of the present invention. 図5は、本実施の形態における位置情報判定部の詳細な構成を示すブロック図である。FIG. 5 is a block diagram showing a detailed configuration of the position information determination unit in the present embodiment. 図6は、本実施の形態における動画素の領域と静止画素の領域との境界を算出する方法を説明するための図である。FIG. 6 is a diagram for explaining a method of calculating the boundary between the moving pixel region and the still pixel region in the present embodiment. 図7は、本実施の形態における切出し枠と包含された切出し枠との関係を示す図である。FIG. 7 is a diagram showing the relationship between the cutout frame and the included cutout frame in the present embodiment. 図8は、本実施の形態における映像調整部の詳細な構成を示すブロック図である。FIG. 8 is a block diagram showing a detailed configuration of the video adjustment unit in the present embodiment. 図9Aは、本実施の形態における包含された切出し枠から拡大切出し枠を算出する1例において、包含された切出し枠の画面上での垂直方向の幅が、表示画面サイズ信号に含まれる画面のアスペクト比により定まる幅よりも大きい場合を示す図である。FIG. 9A shows an example in which an enlarged cutout frame is calculated from an included cutout frame in the present embodiment, and the width of the included cutout frame on the screen in the vertical direction of the screen included in the display screen size signal. It is a figure which shows the case where it is larger than the width | variety decided by an aspect ratio. 図9Bは、本実施の形態における包含された切出し枠から拡大切出し枠を算出する1例において、包含された切出し枠の画面上での水平方向の幅が、表示画面サイズ信号に含まれる画面のアスペクト比により定まる幅よりも大きい場合を示す図である。FIG. 9B shows an example in which an enlarged cutout frame is calculated from the cutout frame included in the present embodiment, and the horizontal width on the screen of the cutout frame included in the screen includes the display screen size signal. It is a figure which shows the case where it is larger than the width | variety decided by an aspect ratio. 図10は、本実施の形態における拡大後映像信号の所定のフレームを画面に表示した図である。FIG. 10 is a diagram showing a predetermined frame of the enlarged video signal displayed on the screen in the present embodiment. 図11は、本実施の形態のおける映像出力部が、出力フラグを用いない場合における、画像処理中での表示画面の例を示す図である。FIG. 11 is a diagram illustrating an example of a display screen during image processing when the video output unit according to the present embodiment does not use an output flag. 図12は、本実施の形態のおける映像出力部が、出力フラグを用いる場合における、画像処理中での表示画面の例を示す図である。FIG. 12 is a diagram illustrating an example of a display screen during image processing when the video output unit according to the present embodiment uses an output flag. 図13は、本実施の形態のおける映像出力部がフレームメモリを用いて、画像処理中にも安定した表示画面を出力する例を示した図である。FIG. 13 is a diagram illustrating an example in which the video output unit according to the present embodiment uses a frame memory to output a stable display screen even during image processing. 図14Aは、本実施の形態のおける映像解析部が出力するズーム制御パラメータについて説明するための概念図である。FIG. 14A is a conceptual diagram for explaining the zoom control parameters output by the video analysis unit in the present embodiment. 図14Bは、本実施の形態のおける映像解析部が出力するズーム制御パラメータについて説明するための概念図である。FIG. 14B is a conceptual diagram for explaining the zoom control parameters output by the video analysis unit in the present embodiment. 図15は、本実施の形態のおける映像処理方法の手順を示すフローチャートである。FIG. 15 is a flowchart showing the procedure of the video processing method in the present embodiment. 図16は、従来の画像処理装置の画像処理を概念的に示した図である。FIG. 16 is a diagram conceptually showing image processing of a conventional image processing apparatus.
 (実施の形態)
 以下、図面を参照しながら、本発明の実施の形態における映像処理装置および映像処理方法について説明する。
(Embodiment)
Hereinafter, a video processing apparatus and a video processing method according to an embodiment of the present invention will be described with reference to the drawings.
 図1は、本発明の実施の形態における連続するフレームの切出し枠1301、1302、1303を設定する1例を示す図である。図1に示すように、カメラなどから取得される映像信号に含まれる被写体が移動している場合、被写体の移動部分を切出し枠1301、1302、1303としてフレームごとに設定する方法が考えられる。この場合、被写体の移動部分を含む切出し枠1301、1302、1303は、フレーム間の画像の差分を用いて決定することができる。すなわち、フレーム間の画像の差分の検出結果に基づいて、それぞれのフレームの動画素の領域と、静止画素の領域を決定する。そして、動画素の領域を含む矩形形状を、切出し枠1301、1302、1303として決定している。 FIG. 1 is a diagram showing an example of setting continuous frame cutout frames 1301, 1302, and 1303 in the embodiment of the present invention. As shown in FIG. 1, when a subject included in a video signal acquired from a camera or the like is moving, a method of setting a moving portion of the subject as cut frames 1301, 1302, and 1303 for each frame can be considered. In this case, the clipping frames 1301, 1302, and 1303 including the moving part of the subject can be determined using the image difference between frames. That is, based on the detection result of the image difference between frames, the moving pixel region and the still pixel region of each frame are determined. A rectangular shape including a moving pixel region is determined as the cutout frames 1301, 1302, and 1303.
 図1の横軸は、フレームであって、連続するフレームの1部のみを示している。ここで、N、mは任意の整数である。図1に示すように、フレームN、フレーム(N+m)、フレーム(N+2m)における切出し枠1301、1302,1303のそれぞれの位置とそれぞれの大きさとは、各フレームに先行する直近のフレームと各フレームとにおける被写体の移動量によって変化する。なお、被写体の移動量を決定するための各フレームに先行する直近のフレームのフレーム数は、適用するシステムの仕様に基づいて適宜設定することができる。 The horizontal axis in FIG. 1 is a frame, and shows only a part of a continuous frame. Here, N and m are arbitrary integers. As shown in FIG. 1, the positions and sizes of the cutout frames 1301, 1302, and 1303 in the frame N, the frame (N + m), and the frame (N + 2m) are the latest frame preceding each frame. And the amount of movement of the subject in each frame. Note that the number of frames of the most recent frame preceding each frame for determining the amount of movement of the subject can be set as appropriate based on the specifications of the system to be applied.
 図1では、被写体の移動量が比較的小さい場合を想定している。図1に示すように、切出し枠1301、1302,1303のそれぞれの位置とそれぞれの大きさとは、すこし異なることとなる。 In FIG. 1, it is assumed that the amount of movement of the subject is relatively small. As shown in FIG. 1, the positions and sizes of the cutting frames 1301, 1302, and 1303 are slightly different.
 図2は、本発明の実施の形態における連続するフレームの切出し枠1301、1302,1303を画像処理によって拡大して、拡大表示映像として画面全体に表示する1例を示す図である。画面1401、1402、1403は、それぞれ切出し枠1301、1302,1303に対応する。図2では、切出し枠1301、1302,1303をそのまま拡大したものであるが、画面1401、1402、1403では被写体を判別しやすくなっている。なお、本実施の形態では、動画素の領域を含む矩形形状を、切出し枠1301、1302、1303として決定している。ただし、動画素の領域を含む矩形形状をそのまま、切出し枠として拡大表示するものではない。そのようにすると、図2に示すようにフレーム間において、背景が変動し易くなるからである。切出し枠1301、1302、1303の決定方法と拡大表示映像の決定方法との詳細について、以下に説明する。 FIG. 2 is a diagram showing an example in which cut- out frames 1301, 1302, and 1303 of continuous frames in the embodiment of the present invention are enlarged by image processing and displayed as an enlarged display image on the entire screen. Screens 1401, 1402, and 1403 correspond to cut frames 1301, 1302, and 1303, respectively. In FIG. 2, the cutout frames 1301, 1302, and 1303 are enlarged as they are, but the screens 1401, 1402, and 1403 make it easy to determine the subject. In this embodiment, the rectangular shape including the moving pixel region is determined as the cutout frames 1301, 1302, and 1303. However, the rectangular shape including the moving pixel region is not enlarged and displayed as a cutout frame as it is. This is because the background tends to fluctuate between frames as shown in FIG. Details of the method for determining the cutout frames 1301, 1302, and 1303 and the method for determining the enlarged display image will be described below.
 図3は、本発明の実施の形態における映像処理装置100の構成を示すブロック図である。本実施の形態における映像処理装置100は、カメラ101と、映像取得部102と、映像解析部103と、映像調整部104と、カメラ制御部106と、映像出力部107とを有する。 FIG. 3 is a block diagram showing the configuration of the video processing apparatus 100 according to the embodiment of the present invention. The video processing apparatus 100 according to the present embodiment includes a camera 101, a video acquisition unit 102, a video analysis unit 103, a video adjustment unit 104, a camera control unit 106, and a video output unit 107.
 カメラ101は、被写体を含む映像を撮影し、映像取得部102へ撮影した信号を出力する。映像取得部102は、カメラ101から入力した信号を映像信号301に変換する。映像解析部103は、映像信号301に対して時間方向に映像を解析する。具体的には、映像解析部103は、映像信号301を入力して、映像信号301の連続するフレーム間の差分を検出することにより、背景部と移動する被写体とを分離する。そして、映像解析部103は、検出された背景部と移動する被写体とに応じて、映像信号301に基づいて切出し枠を算出する。すなわち、映像解析部103は、連続するフレームの差分が所定の閾値を超える領域を切出し枠として、連続するフレーム毎に検出する。そして、映像解析部103は、所定の複数のフレームにおいて、複数のすべての切出し枠を含むようにして、包含された切出し枠306を算出する。 The camera 101 captures a video including the subject and outputs the captured signal to the video acquisition unit 102. The video acquisition unit 102 converts a signal input from the camera 101 into a video signal 301. The video analysis unit 103 analyzes the video in the time direction with respect to the video signal 301. Specifically, the video analysis unit 103 receives the video signal 301 and detects a difference between successive frames of the video signal 301 to separate the background portion from the moving subject. Then, the video analysis unit 103 calculates a cutting frame based on the video signal 301 in accordance with the detected background portion and the moving subject. That is, the video analysis unit 103 detects, for each successive frame, an area where the difference between successive frames exceeds a predetermined threshold as a cutout frame. Then, the video analysis unit 103 calculates the included cutout frame 306 so as to include all of the plurality of cutout frames in a predetermined plurality of frames.
 また、被写体が映像取得部102で取得した映像領域内からはみだしている場合、映像解析部103は、カメラ101のズーム(Zoom)をワイド側(すなわち、ズームレンズを広角側)に調整することが必要と判断する。ここで、ズームはズームレンズを用いるとしたが、撮像素子で受像した映像を電子的に演算して行ってもよい。そして、映像解析部103は、カメラ101のズーム制御パラメータ307を、カメラ制御部106へ出力する。なお、映像解析部103の動作の詳細については、後述する。 Further, when the subject protrudes from the video area acquired by the video acquisition unit 102, the video analysis unit 103 can adjust the zoom of the camera 101 to the wide side (that is, the zoom lens to the wide angle side). Judge as necessary. Here, zoom is performed using a zoom lens, but the image received by the image sensor may be calculated electronically. Then, the video analysis unit 103 outputs the zoom control parameter 307 of the camera 101 to the camera control unit 106. Details of the operation of the video analysis unit 103 will be described later.
 カメラ制御部106は、カメラ101にズームの制御信号を送信する。カメラ制御部106は、映像解析部103から出力されるズーム制御パラメータ307に基づいて、カメラ101のズーム値を調整する。ズーム制御パラメータ307は、映像解析部103で映像領域内における背景部分の占める割合が小さい場合であって、被写体が撮影範囲をはみだしている場合、カメラ101のズーム値を制御するために用いる。また、移動する被写体が撮影範囲に対して、極端に小さく、ユーザが視認しにくい場合、カメラ101の位置を被写体が画面の中央に移動するように制御してもよい。そして、ズームをテレ側(すなわち、ズームレンズを望遠側)に調整し、画面に被写体がより大きく表示されるように制御してもよい。このような制御により、ユーザはより被写体を認識しやすくできる。 The camera control unit 106 transmits a zoom control signal to the camera 101. The camera control unit 106 adjusts the zoom value of the camera 101 based on the zoom control parameter 307 output from the video analysis unit 103. The zoom control parameter 307 is used to control the zoom value of the camera 101 when the ratio of the background portion in the video area is small in the video analysis unit 103 and the subject is outside the shooting range. Further, when the moving subject is extremely small with respect to the shooting range and is difficult for the user to visually recognize, the position of the camera 101 may be controlled so that the subject moves to the center of the screen. Then, the zoom may be adjusted to the telephoto side (that is, the zoom lens is set to the telephoto side), and control may be performed so that the subject is displayed larger on the screen. Such control makes it easier for the user to recognize the subject.
 映像調整部104は、映像信号301と包含された切出し枠306と表示画面サイズ信号105とを入力する。そして、映像調整部104は、包含された切出し枠306と、表示画面サイズ信号105に基づいて算出した拡大率とに応じて、映像信号301に基づいて拡大後映像信号309を出力する。すなわち、映像調整部104は、映像解析部103で分離した被写体と背景部とに応じて、背景部以外を切り出した被写体を含む拡大後映像信号309を、映像出力部107へ出力する。また、映像出力部107は、映像信号301と拡大後映像信号309と映像解析部103から出力される出力フラグ308とを入力して、出力映像信号108を出力する。 The video adjustment unit 104 inputs the video signal 301, the included cutout frame 306, and the display screen size signal 105. Then, the video adjustment unit 104 outputs an enlarged video signal 309 based on the video signal 301 according to the included cutout frame 306 and the enlargement ratio calculated based on the display screen size signal 105. That is, the video adjustment unit 104 outputs an enlarged video signal 309 including a subject cut out except for the background portion to the video output unit 107 according to the subject and background portion separated by the video analysis unit 103. Further, the video output unit 107 receives the video signal 301, the enlarged video signal 309, and the output flag 308 output from the video analysis unit 103, and outputs the output video signal 108.
 表示画面サイズ信号105は、例えば制御部(図示せず)から入力される信号であって、映像処理装置100の出力映像信号108が表示する画面のアスペクト比と出力映像信号108の画面表示サイズとを示す情報を含んでいる。 The display screen size signal 105 is, for example, a signal input from a control unit (not shown), and the aspect ratio of the screen displayed by the output video signal 108 of the video processing apparatus 100 and the screen display size of the output video signal 108. It contains information indicating.
 ここで、制御部は、映像処理装置100の出力映像信号108が表示される装置と通信や情報交換を行い、その装置の画面のアスペクト比と画面表示サイズとを取得しているものと想定している。なお、出力映像信号108が表示される装置は、本実施の形態のおける映像処理装置100に通信システムなどで接続された装置であって、例えばテレビ会議システムを利用している他のユーザが利用する装置であってもよい。また、例えば映像処理装置100に映像信号接続用のケーブルなどで直接接続されたモニター装置であってもよい。さらにまた、画像表示部が組み込まれた映像処理装置に、出力映像信号108を表示する構成であっても、本発明を適用できる。 Here, it is assumed that the control unit communicates and exchanges information with the device on which the output video signal 108 of the video processing device 100 is displayed, and acquires the screen aspect ratio and screen display size of the device. ing. The apparatus on which the output video signal 108 is displayed is an apparatus connected to the video processing apparatus 100 according to the present embodiment by a communication system or the like, and is used by, for example, another user who uses the video conference system. It may be a device that performs. Further, for example, it may be a monitor device directly connected to the video processing device 100 by a video signal connection cable or the like. Furthermore, the present invention can also be applied to a configuration in which the output video signal 108 is displayed on a video processing device in which an image display unit is incorporated.
 次に、映像解析部103の詳細な構成と動作について説明する。図4は、本発明の実施の形態における映像解析部103の詳細な構成を示すブロック図である。映像解析部103は、映像記憶部302と、差分算出部303と、背景検出部304と、位置情報判定部305と、切出し位置記憶部310とを有する。 Next, the detailed configuration and operation of the video analysis unit 103 will be described. FIG. 4 is a block diagram showing a detailed configuration of the video analysis unit 103 in the embodiment of the present invention. The video analysis unit 103 includes a video storage unit 302, a difference calculation unit 303, a background detection unit 304, a position information determination unit 305, and a cut-out position storage unit 310.
 映像記憶部302と差分算出部303とは、映像取得部102から出力される映像信号301を、それぞれ入力する。映像記憶部302は、映像信号301の1フレーム分を記憶する。そして、映像記憶部302は、入力された映像信号301を1フレーム分遅延させて、差分算出部303へ出力する。映像記憶部302は、フレームメモリであって、例えば半導体メモリなどで構成してもよい。 The video storage unit 302 and the difference calculation unit 303 each input the video signal 301 output from the video acquisition unit 102. The video storage unit 302 stores one frame of the video signal 301. Then, the video storage unit 302 delays the input video signal 301 by one frame and outputs it to the difference calculation unit 303. The video storage unit 302 is a frame memory, and may be configured by, for example, a semiconductor memory.
 差分算出部303は、映像記憶部302から出力される1フレーム分遅延された映像信号301aと、遅延されていない元の映像信号301との差分を検出する。ここで、映像信号301の現フレームをフレームN(Nは整数)とすると、映像記憶部302からは、現フレームNに先行するフレーム(N-1)が出力される。差分算出部303は、フレームNとフレーム(N-1)の映像信号の差分を、画素毎に検出する。そして、差分算出部303は、検出した差分結果を背景検出部304へ出力する。 The difference calculation unit 303 detects a difference between the video signal 301a delayed from the video storage unit 302 by one frame and the original video signal 301 that has not been delayed. Here, if the current frame of the video signal 301 is a frame N (N is an integer), the video storage unit 302 outputs a frame (N−1) preceding the current frame N. The difference calculation unit 303 detects the difference between the video signals of the frame N and the frame (N−1) for each pixel. Then, the difference calculation unit 303 outputs the detected difference result to the background detection unit 304.
 背景検出部304は、画素毎の差分を所定の閾値と比較することにより、画素毎に動画素、もしくは静止画素と判定する。所定の閾値は、例えば、10段階程度とする。ここで各画素は、8ビット(256段階の精度で映像信号を扱う場合)で表現されていると想定する。この際、画素毎の差分が所定の閾値を超える場合、その画素は動画素とする。すなわち、フレームNとフレーム(N-1)とで、動きのあった動画素と見做す。また、画素が、所定の閾値以下の差分を有する場合、動きのなかった静止画素と見做す。そして、背景検出部304は、現フレームNにおける動画素の領域と、静止画素の領域とを区分して位置情報判定部305へ出力する。 The background detection unit 304 determines that each pixel is a moving pixel or a still pixel by comparing the difference for each pixel with a predetermined threshold. The predetermined threshold is, for example, about 10 levels. Here, it is assumed that each pixel is expressed by 8 bits (when a video signal is handled with 256 levels of accuracy). At this time, if the difference for each pixel exceeds a predetermined threshold, the pixel is a moving pixel. That is, the moving pixels in the frame N and the frame (N−1) are regarded as moving pixels. Further, when a pixel has a difference equal to or less than a predetermined threshold, it is regarded as a still pixel that has not moved. Then, the background detection unit 304 divides the moving pixel region and the still pixel region in the current frame N, and outputs them to the position information determination unit 305.
 次に、図5を参照しながら、位置情報判定部305の詳細な構成と動作について説明する。図5は、本実施の形態における位置情報判定部305の詳細な構成を示すブロック図である。位置情報判定部305は、図5に示すように、座標算出部351と変化量算出部352とを有する。位置情報判定部305は、切出し位置記憶部310を用いて、包含された切出し枠306を映像調整部104へ出力する。 Next, the detailed configuration and operation of the position information determination unit 305 will be described with reference to FIG. FIG. 5 is a block diagram illustrating a detailed configuration of the position information determination unit 305 in the present embodiment. As illustrated in FIG. 5, the position information determination unit 305 includes a coordinate calculation unit 351 and a change amount calculation unit 352. The position information determination unit 305 outputs the included cutout frame 306 to the video adjustment unit 104 using the cutout position storage unit 310.
 図6は、本実施の形態における動画素の領域と静止画素の領域との境界632を算出する方法を説明するための図である。座標算出部351は、現フレームNにおける動画素の領域と、静止画素の領域とから、背景部分と被写体部分の境界632を算出する。図6に示す破線は、動画素の領域と静止画素の領域との境界632を示している。境界632は、背景部分と被写体部分の境界を示す矩形の形状であって、切出し枠を示している。 FIG. 6 is a diagram for explaining a method of calculating the boundary 632 between the moving pixel region and the stationary pixel region in the present embodiment. The coordinate calculation unit 351 calculates a boundary 632 between the background portion and the subject portion from the moving pixel region and the still pixel region in the current frame N. 6 indicates a boundary 632 between the moving pixel region and the stationary pixel region. A boundary 632 has a rectangular shape indicating the boundary between the background portion and the subject portion, and indicates a cutout frame.
 なお、切出し開始座標630と切出し終了座標631とは、矩形の形状である境界632を画面上で特定できるように境界632の上部左端の座標と下部右端の座標とを用いている。これは、切出し位置記憶部310のメモリ量を極力少なくするための1例であって、本実施の形態は、これらの座標を用いることに限定するものではない。例えば、矩形の形状である境界632の上部右端の座標と下部左端の座標とを用いてもよい。また、矩形の形状である境界632の上下方向および左右方向の画素数と中心座標とを用いてもよい。このようにして、背景部分と被写体部分の境界632から、矩形の形状である切出し枠を算出する。 It should be noted that the extraction start coordinate 630 and the extraction end coordinate 631 use the upper left end coordinate and the lower right end coordinate of the boundary 632 so that the boundary 632 having a rectangular shape can be specified on the screen. This is an example for reducing the memory amount of the cutout position storage unit 310 as much as possible, and the present embodiment is not limited to using these coordinates. For example, the coordinates of the upper right end and the lower left end of the boundary 632 having a rectangular shape may be used. Further, the number of pixels in the vertical direction and the horizontal direction of the boundary 632 which is a rectangular shape, and the center coordinates may be used. In this way, a cutout frame having a rectangular shape is calculated from the boundary 632 between the background portion and the subject portion.
 図5に示すように、座標算出部351は、切出し開始座標630と切出し終了座標631とを含む切出し座標312を、切出し位置記憶部310と変化量算出部352へ出力する。 As shown in FIG. 5, the coordinate calculation unit 351 outputs the cut coordinates 312 including the cut start coordinates 630 and the cut end coordinates 631 to the cut position storage unit 310 and the change amount calculation unit 352.
 切出し位置記憶部310は、現フレームNに先行するフレーム(N-1)からフレーム(N-9)までの9フレーム分の切出し座標313を記憶している。切出し位置記憶部310は、現フレームNの切出し座標312が入力されると、記憶しているフレーム(N-1)からフレーム(N-9)までの9フレーム分の切出し座標313を変化量算出部352へ出力する。そして、切出し位置記憶部310は、記憶している内容を更新する。すなわち、切出し位置記憶部310は、現フレームNからフレーム(N-8)までの9フレーム分の切出し座標313を記憶する。 The cut-out position storage unit 310 stores cut-out coordinates 313 for nine frames from the frame (N-1) preceding the current frame N to the frame (N-9). When the cutout coordinates 312 of the current frame N are input, the cutout position storage unit 310 calculates the amount of change of the cutout coordinates 313 for nine frames from the stored frame (N-1) to the frame (N-9). To the unit 352. Then, the cut-out position storage unit 310 updates the stored content. That is, the cutout position storage unit 310 stores the cutout coordinates 313 for nine frames from the current frame N to the frame (N-8).
 変化量算出部352は、現フレームNからフレーム(N-9)のそれぞれの切出し座標312と現フレームNからフレーム(N-9)のそれぞれに先行する直近のフレーム、例えば現フレームNに対してはフレーム(N-1)、の切出し座標313とを比較して、両者の移動量を検証する。そして、変化量算出部352は、両者の移動量が所定の移動量閾値以下である場合、現フレームNから現フレームNに先行するフレーム(N-9)までの10フレーム分の切出し座標312,313を座標算出部351に判定後切出し座標314として出力する。ここで、本実施の形態における映像処理装置100では、テレビ会議システムを前提とすれば、所定の移動量閾値は、例えば200画素程度が望ましい。 The change amount calculation unit 352 applies the cut-out coordinates 312 of each of the current frame N to the frame (N-9) and the latest frame preceding each of the current frame N to the frame (N-9), for example, the current frame N. Compares the cut-out coordinates 313 of the frame (N−1) and verifies the movement amount of both. Then, when the amount of movement of both is equal to or less than a predetermined amount of movement threshold, the change amount calculation unit 352 extracts 10 frames of cutout coordinates 312 for 10 frames from the current frame N to a frame (N-9) preceding the current frame N. 313 is output to the coordinate calculation unit 351 as post-determination cut-out coordinates 314. Here, in the video processing apparatus 100 according to the present embodiment, if a video conference system is assumed, the predetermined movement amount threshold is preferably about 200 pixels, for example.
 座標算出部351は、判定後切出し座標314を入力して、所定の連続する複数のフレームにおいて、複数のすべての切出し枠を含むようにして、包含された切出し枠306を算出する。ここで、所定の連続する複数のフレームの数は、例えば、10フレームとしている。この場合、座標算出部351は、連続する10フレーム分の判定後切出し座標314を入力する。そして、座標算出部351は、10フレーム分の判定後切出し座標314のすべてを包含し、それらより広い切出し枠である、包含された切出し枠306を算出する。ここで、包含された切出し枠306の形状は、矩形とする。 The coordinate calculation unit 351 inputs the post-judgment cut-out coordinates 314, and calculates an included cut-out frame 306 so as to include all of the plurality of cut-out frames in a plurality of predetermined consecutive frames. Here, the predetermined number of consecutive frames is, for example, 10 frames. In this case, the coordinate calculation unit 351 inputs post-determination cut-out coordinates 314 for 10 consecutive frames. Then, the coordinate calculation unit 351 includes all of the post-determination cut-out coordinates 314 for 10 frames, and calculates an included cut-out frame 306 that is a wider cut-out frame. Here, the shape of the included cutout frame 306 is a rectangle.
 このようにすれば、包含された切出し枠306の形状が単純であるため、以後の信号処理をより簡便に行うことができる。すなわち、包含された切出し枠306の形状を単純な判定後切出し座標314の合成とするべきではない。ただし、本実施の形態における映像処理装置100の各部の信号処理能力が向上し、以降の信号処理能力に余裕が生じれば、上記した包含された切出し枠306の形状を単純な判定後切出し座標314の合成としてもよい。 In this way, since the shape of the included cutout frame 306 is simple, the subsequent signal processing can be performed more easily. That is, the shape of the included cutout frame 306 should not be a simple composition of the post-judgment cutout coordinates 314. However, if the signal processing capability of each part of the video processing apparatus 100 in the present embodiment is improved and there is a margin in the signal processing capability thereafter, the shape of the included extraction frame 306 is simply extracted after the determination coordinates. It is good also as composition of 314.
 一方、変化量算出部352は、現フレームNからフレーム(N-9)のぞれぞれの切出し座標312と現フレームNからフレーム(N-9)のそれぞれに先行する直近のフレーム、例えば現フレームNに対してはフレーム(N-1)、の切出し座標313とを比較して、両者の移動量が所定の移動量閾値を超えている場合、例えばフレーム(N-1)から現フレームNへの移動量が所定の移動量閾値を超えている場合、判定後切出し座標314から、現フレームNの切出し座標312を除く。そして、変化量算出部352は、現フレームNの切出し座標312に変えて、現フレームNに先行するフレーム(N-1)の切出し座標313を判定後切出し座標314として出力する。さらに、変化量算出部352は、現フレームNに先行するフレーム(N-1)から現フレームNに先行するフレーム(N-9)までの9フレーム分の切出し座標312を座標算出部351に判定後切出し座標314として出力する。すなわち、フレーム(N-1)の切出し座標313は、データが2回出力される。 On the other hand, the change amount calculation unit 352 includes the cut-out coordinates 312 of each of the current frame N to the frame (N-9) and the latest frame preceding each of the current frame N to the frame (N-9), for example, the current frame N. If the movement amount of both of the frames N exceeds the predetermined movement amount threshold value by comparing the cut coordinates 313 of the frame (N-1), for example, the frame N-1 to the current frame N When the movement amount to exceeds the predetermined movement amount threshold, the cut-out coordinates 312 of the current frame N are excluded from the cut-out coordinates 314 after determination. Then, the change amount calculation unit 352 outputs the cut coordinates 313 of the frame (N−1) preceding the current frame N as the cut coordinates 314 after determination instead of the cut coordinates 312 of the current frame N. Further, the change amount calculation unit 352 determines to the coordinate calculation unit 351 the extracted coordinates 312 for nine frames from the frame (N−1) preceding the current frame N to the frame (N-9) preceding the current frame N. Output as post-cutout coordinates 314. That is, data is output twice for the cut-out coordinates 313 of the frame (N−1).
 上記したように映像解析部103は、連続するフレームの切出し枠の移動量が所定の移動量閾値よりも超える場合、連続するフレームの切出し枠を、所定の連続する複数のフレームにおける複数のすべての切出し枠から除く。この場合でも、座標算出部351は、10フレーム分に相当する判定後切出し座標314を入力する。そして、座標算出部351は、10フレーム分に相当する判定後切出し座標314のすべてを包含し、それらより広い切出し枠である、包含された切出し枠306を算出する。 As described above, when the movement amount of the cut frames of the continuous frames exceeds the predetermined movement amount threshold, the video analysis unit 103 sets the cut frames of the continuous frames to all of a plurality of predetermined continuous frames. Remove from the cropping frame. Even in this case, the coordinate calculation unit 351 inputs post-determination cut-out coordinates 314 corresponding to 10 frames. Then, the coordinate calculation unit 351 includes all of the post-judgment cutout coordinates 314 corresponding to 10 frames, and calculates an included cutout frame 306 that is a wider cutout frame.
 図7は、本実施の形態における切出し枠601、602、603と包含された切出し枠306との関係を示す図である。図7は、簡便に図示するために、3つのフレームの場合を例にして、座標算出部351が、包含された切出し枠306を算出する際の切出し枠601、602、603と包含された包含された切出し枠306との関係を示している。図7に示すように、座標算出部351は、切出し枠601、602、603をすべて包含する最小の矩形の形状を、包含された切出し枠306として算出している。 FIG. 7 is a diagram showing the relationship between the cut frames 601, 602, and 603 and the included cut frame 306 in the present embodiment. FIG. 7 illustrates the case of three frames as an example for the sake of simplicity. The coordinate calculation unit 351 includes inclusion frames 601, 602, and 603 included when the included extraction frame 306 is calculated. The relationship with the cut out frame 306 is shown. As shown in FIG. 7, the coordinate calculation unit 351 calculates the minimum rectangular shape that includes all the cut frames 601, 602, and 603 as the included cut frame 306.
 次に、映像調整部104の詳細な構成と動作について、図8を参照しながら説明する。図8は、本実施の形態における映像調整部104の詳細な構成を示すブロック図である。映像調整部104は、拡大率算出部403と映像拡大部405とを有する。 Next, the detailed configuration and operation of the video adjustment unit 104 will be described with reference to FIG. FIG. 8 is a block diagram showing a detailed configuration of the video adjustment unit 104 in the present embodiment. The video adjustment unit 104 includes an enlargement ratio calculation unit 403 and a video enlargement unit 405.
 拡大率算出部403は、包含された切出し枠306と表示画面サイズ信号105とを入力して、包含された切出し枠306の拡大率Eと拡大開始位置360a、360bとを出力する。そして、映像拡大部405は、包含された切出し枠306の拡大率Eと拡大開始位置360a、360bと映像信号301とを入力して拡大後映像信号309を出力する。ここで、拡大率Eは、拡大後映像信号309における表示画面サイズとなる拡大切出し枠と包含された切出し枠306との比である。また、拡大開始位置360a、360bは、拡大切出し枠の例えば上部左端としてもよい。 The enlargement ratio calculation unit 403 inputs the included cutout frame 306 and the display screen size signal 105, and outputs the enlargement ratio E of the included cutout frame 306 and the enlargement start positions 360a and 360b. Then, the video enlargement unit 405 inputs the enlargement ratio E of the included cutout frame 306, the enlargement start positions 360a and 360b, and the video signal 301, and outputs the post-enlargement video signal 309. Here, the enlargement ratio E is the ratio of the enlarged cutout frame that becomes the display screen size in the post-enlargement video signal 309 and the included cutout frame 306. Further, the enlargement start positions 360a and 360b may be, for example, the upper left end of the enlargement cutout frame.
 上記した構成によって、拡大後映像信号309を出力する具体的な動作について、図9A、図9Bを用いて説明する。図9Aは、本実施の形態における包含された切出し枠306aから拡大切出し枠350aを算出する1例において、包含された切出し枠306aの画面上での垂直方向の幅が、表示画面サイズ信号に含まれる画面のアスペクト比により定まる幅よりも、大きい場合を示す図である。図9Bは、本実施の形態における包含された切出し枠306bから拡大切出し枠350bを算出する1例において、包含された切出し枠306bの画面上での水平方向の幅が、表示画面サイズ信号に含まれる画面のアスペクト比により定まる幅よりも、大きい場合を示す図である。図9A、図9Bでは、それぞれ包含された切出し枠306a、306bと拡大された映像信号を出力するための拡大切出し枠とを破線で示している。また、図9A、図9Bは、拡大開始位置360a、360bも示している。 Specific operations for outputting the enlarged video signal 309 with the above-described configuration will be described with reference to FIGS. 9A and 9B. FIG. 9A shows an example in which the enlarged cutout frame 350a is calculated from the cutout frame 306a included in the present embodiment, and the vertical width of the included cutout frame 306a on the screen is included in the display screen size signal. It is a figure which shows the case where it is larger than the width | variety determined by the aspect-ratio of the screen to be displayed. FIG. 9B shows an example in which the enlarged cutout frame 350b is calculated from the cutout frame 306b included in the present embodiment, and the horizontal width on the screen of the cutout frame 306b included is included in the display screen size signal. It is a figure which shows the case where it is larger than the width | variety determined by the aspect-ratio of the screen to be displayed. 9A and 9B, the included cutout frames 306a and 306b and the enlarged cutout frame for outputting the enlarged video signal are indicated by broken lines. 9A and 9B also show enlargement start positions 360a and 360b.
 ここで、包含された切出し枠306a、306bを画面上で特定するために、それぞれ上部左端の座標(Xs、Ys)と下部右端の座標(Xe、Ye)とを用いている。上記の2つの場合において、拡大率Eと拡大開始位置360a、360bとのそれぞれを求める動作について説明する。 Here, the coordinates of the upper left end (Xs, Ys) and the coordinates of the lower right end (Xe, Ye) are used to specify the included cutout frames 306a, 306b on the screen. An operation for obtaining each of the enlargement ratio E and the enlargement start positions 360a and 360b in the above two cases will be described.
 具体的には、拡大後映像信号309を表示する画面のアスペクト比に合致するように、包含された切出し枠306aは画面の水平方向あるいは垂直方向に拡張される。ここで、画面の画素数を1920×1080とし、画面の最端部の座標を(Dx,Dy)=(1920、1080)で表す。すると、拡大後映像信号309を表示する画面のアスペクト比に合致するように拡大するための拡大率Eは次のように表すことができる。ここで、min(A、B)は、引数Aと引数Bとの小さいほうの数値を選択する関数とする。 Specifically, the included cutout frame 306a is expanded in the horizontal or vertical direction of the screen so as to match the aspect ratio of the screen on which the enlarged video signal 309 is displayed. Here, the number of pixels on the screen is 1920 × 1080, and the coordinates of the end of the screen are represented by (Dx, Dy) = (1920, 1080). Then, the enlargement ratio E for enlarging the image signal 309 after enlargement so as to match the aspect ratio of the screen on which the image signal 309 is displayed can be expressed as follows. Here, min (A, B) is a function for selecting the smaller numerical value of the argument A and the argument B.
 E=min(Dx×1/|Xe―Xs|、Dy×1/|Ye―Ys|)
 まず、図9Aの場合において、具体的な数値を想定して上式を用いて、拡大率Eを求める。ここで、例えば、拡大切出し枠350aの上部左端の座標(Xs、Ys)=(800、540)、下部右端の座標(Xe、Ye)=(992、1080)とすると、
 E=min(1920×1/|992―800|、1080×1/|1080―540|)
である。したがって、
 E=min(10、2)=2(倍)
と求めることができる。すなわち、包含された切出し枠306aを水平方向に2倍だけ拡張すれば、拡大切出し枠350aが算出できる。この場合、包含された切出し枠306aを、その水平方向の両側に同等の幅だけ拡張する。このようにして拡大開始位置360aを求めることができる。
E = min (Dx × 1 / | Xe−Xs |, Dy × 1 / | Ye−Ys |)
First, in the case of FIG. 9A, the enlargement ratio E is obtained using the above equation assuming specific numerical values. Here, for example, if the upper left end coordinates (Xs, Ys) = (800, 540) and the lower right end coordinates (Xe, Ye) = (992, 1080) of the enlarged cutout frame 350a,
E = min (1920 × 1 / | 992-800 |, 1080 × 1 / | 1080-540 |)
It is. Therefore,
E = min (10, 2) = 2 (times)
It can be asked. That is, if the included cutout frame 306a is expanded by twice in the horizontal direction, the enlarged cutout frame 350a can be calculated. In this case, the included cutout frame 306a is expanded by an equal width on both sides in the horizontal direction. In this way, the enlargement start position 360a can be obtained.
 次に、図9Bの場合において、具体的な数値を想定して拡大率Eを求める。ここで、例えば、拡大切出し枠350bの上部左端の座標(Xs、Ys)=(100、540)、下部右端の座標(Xe、Ye)=(1900、1080)とすると、
 E=min(1920×1/|1920―100|、1080×1/|1080―540|)
である。したがって、
 E=min(1.06、2)=1.06(倍)
と求めることができる。すなわち、包含された切出し枠306bを垂直方向に1.06倍だけ拡張すれば、拡大切出し枠350bが算出できる。この場合、包含された切出し枠306aを垂直方向に上方に拡張する。これは、本実施の形態では、包含された切出し枠306a、306bが表示画面の下端部に沿っているとしているからである。このようにして拡大開始位置360bを求めることができる。
Next, in the case of FIG. 9B, the enlargement ratio E is obtained assuming specific numerical values. Here, for example, if the upper left end coordinates (Xs, Ys) = (100, 540) and the lower right end coordinates (Xe, Ye) = (1900, 1080) of the enlarged cutout frame 350b,
E = min (1920 × 1 / | 1920−100 |, 1080 × 1 / | 1080−540 |)
It is. Therefore,
E = min (1.06, 2) = 1.06 (times)
It can be asked. That is, if the included cutout frame 306b is expanded by 1.06 times in the vertical direction, the enlarged cutout frame 350b can be calculated. In this case, the included cutout frame 306a is expanded upward in the vertical direction. This is because, in the present embodiment, the included cutout frames 306a and 306b are along the lower end of the display screen. In this way, the enlargement start position 360b can be obtained.
 なお、包含された切出し枠306a、306bは、必ずしも表示画面の下端部に沿っている必要はない。その場合では、包含された切出し枠306bを、その垂直方向の両側に同等の幅だけ拡張する。 Note that the included cutout frames 306a and 306b do not necessarily have to be along the lower end of the display screen. In that case, the included cutout frame 306b is expanded by an equal width on both sides in the vertical direction.
 以上のようにして映像拡大部405は、拡大切出し枠350a、350b、すなわち拡大率Eと拡大開始位置360a、360bとを算出し、拡大後映像信号309を出力することができる。すなわち、映像調整部104は、表示画面のアスペクト比と同等のアスペクト比を有するように包含された切出し枠306を拡張し、拡大後映像信号309として映像出力部107に出力する。 As described above, the video enlargement unit 405 can calculate the enlargement cut frames 350a and 350b, that is, the enlargement ratio E and the enlargement start positions 360a and 360b, and output the enlarged image signal 309. That is, the video adjustment unit 104 expands the cutout frame 306 included so as to have an aspect ratio equivalent to the aspect ratio of the display screen, and outputs the expanded cutout video signal 309 to the video output unit 107.
 図10は、拡大後映像信号309の所定のフレームを画面601a、602a、603aに表示した図である。図10では、フレームN、フレーム(N+m)とフレーム(N+2m)とにおける画面601a、602a、603aを示している(N、mは任意の整数)。図10に示すように本実施の形態における映像処理装置100によれば、フレームが変わっても、背景は、一定の領域が常に表示されている。その結果、ユーザにとって、このように連続する画面601a、602a、603aを注視しても不快感は生じにくい。また、被写体を表示する画面601a、602a、603aのほぼ中央に拡大表示できる。その結果、ユーザは被写体を的確に捉えることが可能となる。 FIG. 10 is a diagram in which predetermined frames of the enlarged video signal 309 are displayed on the screens 601a, 602a, and 603a. FIG. 10 shows screens 601a, 602a, and 603a in a frame N, a frame (N + m), and a frame (N + 2m) (N and m are arbitrary integers). As shown in FIG. 10, according to the video processing apparatus 100 in the present embodiment, a constant region is always displayed as a background even if the frame changes. As a result, the user is less likely to experience discomfort even if he / she gazes at the continuous screens 601a, 602a, and 603a. In addition, it can be enlarged and displayed almost at the center of the screens 601a, 602a, and 603a for displaying the subject. As a result, the user can accurately capture the subject.
 次に、図3で示した出力フラグ308について説明する。映像解析部103は、映像調整部104から出力される包含された拡大後映像信号309に基づいて出力フラグ308を出力する。具体的には、映像解析部103は、映像調整部104から出力する拡大後映像信号309の現フレームNと現フレームNに先行する複数のフレームからなる所定の連続するフレームから、拡大後映像信号309を算出した後に、出力フラグ308を出力する。すなわち、例えば10フレームである所定の連続するフレームを用いて、映像調整部104は、現フレームNと現フレームNに先行する複数のフレームに対応する包含された切出し枠306を拡張する。そして、映像調整部104は、上記した複数の包含された切出し枠306に応じて、映像信号301に基づいて拡大後映像信号309を出力する。 Next, the output flag 308 shown in FIG. 3 will be described. The video analysis unit 103 outputs an output flag 308 based on the included enlarged video signal 309 output from the video adjustment unit 104. Specifically, the video analysis unit 103 generates an enlarged video signal from a predetermined continuous frame including the current frame N of the enlarged video signal 309 output from the video adjustment unit 104 and a plurality of frames preceding the current frame N. After calculating 309, the output flag 308 is output. That is, for example, using a predetermined continuous frame, which is 10 frames, the video adjustment unit 104 extends the included cutout frame 306 corresponding to the current frame N and a plurality of frames preceding the current frame N. Then, the video adjusting unit 104 outputs an enlarged video signal 309 based on the video signal 301 in accordance with the plurality of included cutout frames 306 described above.
 このような動作を、図11、図12を用いて説明する。図11は、本実施の形態における映像出力部107が、出力フラグ308を用いない場合における、画像処理中での表示画面の例を示す図である。図11では、現フレームNを含む10フレーム分の信号処理をしている期間t1から期間t2(拡大処理中)も、映像拡大部405から拡大後映像信号309を表示している。この期間は、映像調整部104と映像拡大部405とが、複数の包含された切出し枠306の拡張処理を行っているため、被写体のサイズなどが変化する。その後、期間t3(拡大処理完了)において、その拡張処理が完了すると被写体のサイズなどは確定し、拡大後映像信号309は安定に表示できる。このように出力フラグ308を用いない場合、包含された切出し枠306の拡張処理を行っている間の拡大後映像信号309が出力されるため、ユーザにとって、このような画面を注視すると不快感を与える可能性がある。 Such operation will be described with reference to FIGS. FIG. 11 is a diagram illustrating an example of a display screen during image processing when the video output unit 107 in the present embodiment does not use the output flag 308. In FIG. 11, the enlarged video signal 309 is also displayed from the video enlargement unit 405 during the period t1 to t2 (during the enlargement process) during signal processing for 10 frames including the current frame N. During this period, since the video adjustment unit 104 and the video enlargement unit 405 perform the expansion processing of the plurality of included cutout frames 306, the size of the subject changes. After that, in the period t3 (enlargement process completed), when the extension process is completed, the size of the subject is fixed, and the enlarged video signal 309 can be displayed stably. When the output flag 308 is not used in this way, the enlarged video signal 309 is output while the included extraction frame 306 is being expanded, and thus the user feels uncomfortable when watching such a screen. There is a possibility to give.
 図12は、本実施の形態における映像出力部107が、出力フラグ308を用いる場合における、画像処理中での表示画面の例を示す図である。図12に示すように期間t1から期間t2(拡大処理中)では、映像拡大部405は、映像信号301を出力する。そして、その後、期間t3(拡大処理完了)において、現フレームNを含む所定の連続する複数のフレームに対応する包含された切出し枠306の拡張処理が完了すると、その拡張処理後の拡大後映像信号309を出力する。すなわち、映像出力部107は、映像信号301の現フレームNと現フレームNに先行する複数のフレームからなる、所定の連続するフレームから、拡大後映像信号309を算出する。そして、その後に、映像解析部103から出力される出力フラグ308に基づいて、カメラ101からの映像に基づく映像信号301から、拡大後映像信号309へと出力映像信号108を切替えてもよい。 FIG. 12 is a diagram illustrating an example of a display screen during image processing when the video output unit 107 according to the present embodiment uses the output flag 308. As shown in FIG. 12, during the period t <b> 1 to the period t <b> 2 (during the expansion process), the video enlargement unit 405 outputs the video signal 301. After that, in the period t3 (enlargement process completion), when the expansion process of the included cut frame 306 corresponding to a plurality of predetermined consecutive frames including the current frame N is completed, the expanded video signal after the expansion process 309 is output. That is, the video output unit 107 calculates the enlarged video signal 309 from a predetermined continuous frame including the current frame N of the video signal 301 and a plurality of frames preceding the current frame N. Thereafter, the output video signal 108 may be switched from the video signal 301 based on the video from the camera 101 to the enlarged video signal 309 based on the output flag 308 output from the video analysis unit 103.
 また、現フレームNに連続するフレームの切出し枠の移動量が所定の移動量閾値を超える場合、連続するフレームの先行するフレームNに続く後続のフレーム(N-1)の切出し枠は、変化量算出部352が出力する判定後切出し座標314における所定の連続する複数のフレームの切出し枠から除く。このような場合、映像出力部107は、カメラ101からの映像に基づく映像信号301から、現フレームNを含まない、現フレームNに先行する所定の連続する複数のフレームに対応する包含された切出し枠306の情報を含む拡大後映像信号309へと出力映像信号108を切替えてもよい。このようにすれば、拡大処理中の被写体のサイズなどが変化する映像を表示することはない。 In addition, when the movement amount of the cut frame of the frame subsequent to the current frame N exceeds a predetermined movement amount threshold, the cut frame of the subsequent frame (N−1) following the preceding frame N of the continuous frame is changed by the change amount. Excluded from the cut out frames of a plurality of predetermined consecutive frames in the post-judgment cut out coordinates 314 output by the calculation unit 352. In such a case, the video output unit 107 includes, from the video signal 301 based on the video from the camera 101, the included clipping corresponding to a plurality of predetermined consecutive frames that do not include the current frame N and precede the current frame N. The output video signal 108 may be switched to the post-enlargement video signal 309 including the information of the frame 306. In this way, an image in which the size of the subject being enlarged is changed is not displayed.
 図13は、本実施の形態のおける映像出力部107がフレームメモリを用いて、画像処理中にも安定した表示画面を出力する例を示した図である。映像拡大部405がフレームメモリを有している場合、そのフレームメモリには、現フレームNに先行するフレーム(N-1)から所定の連続する複数のフレームに含まれる包含された切出し枠306に対応する拡大後映像信号309を記憶しておいてもよい。そして、図13に示すように期間t1から期間t2(拡大処理中)では、映像拡大部405は、現フレームNに先行するフレーム(N-1)以前の所定の連続する複数のフレームに対応する包含された切出し枠306の情報に基づいた拡大後映像信号309をフレームメモリから出力する。そして、その後、期間t3(拡大処理完了)において、現フレームを含む所定の連続する複数のフレームに対応する包含された切出し枠306の拡張処理が完了すると、その拡張処理後の拡大後映像信号309を出力する。すなわち、映像出力部107は、1フレーム分のフレームメモリを有し、フレームメモリには、現フレームNに先行する所定の連続する複数のフレームに対応する包含された切出し枠306の情報を含む拡大後映像信号309を格納する。そして、映像出力部107は、映像調整部104が拡大後映像信号309を算出した後に、映像解析部103が出力する出力フラグ308に基づいて、フレームメモリからの拡大後映像信号309から、現フレームNを含む所定の連続する複数のフレームに対応する包含された切出し枠306の情報を含む拡大後映像信号309へと、出力映像信号108を切替えてもよい。 FIG. 13 is a diagram illustrating an example in which the video output unit 107 according to the present embodiment uses the frame memory to output a stable display screen during image processing. When the video enlargement unit 405 has a frame memory, the frame memory includes a frame (N−1) preceding the current frame N and included clipping frames 306 included in a plurality of predetermined consecutive frames. A corresponding enlarged video signal 309 may be stored. As shown in FIG. 13, during the period t1 to the period t2 (during the expansion process), the video enlargement unit 405 corresponds to a plurality of predetermined consecutive frames before the frame (N−1) preceding the current frame N. An enlarged video signal 309 based on the information of the included cutout frame 306 is output from the frame memory. After that, in the period t3 (enlargement process complete), when the expansion process of the included cutout frame 306 corresponding to a plurality of predetermined consecutive frames including the current frame is completed, the expanded video signal 309 after the expansion process is completed. Is output. In other words, the video output unit 107 has a frame memory for one frame, and the frame memory includes information on the included cutout frame 306 corresponding to a plurality of predetermined consecutive frames preceding the current frame N. The post video signal 309 is stored. The video output unit 107 calculates the current frame from the enlarged video signal 309 from the frame memory based on the output flag 308 output from the video analysis unit 103 after the video adjustment unit 104 calculates the enlarged video signal 309. The output video signal 108 may be switched to a post-enlarged video signal 309 including information of the included cutout frame 306 corresponding to a predetermined plurality of consecutive frames including N.
 このようにすれば、映像拡大部405に1フレーム分のフレームメモリが必要となるが、拡大後映像信号309を被写体のサイズなどに関して、ほぼ連続的に安定して表示できる。なお、フレームメモリにデータが蓄積されていない場合、映像拡大部405は、拡大後映像信号309を出力すればよい。 In this way, a frame memory for one frame is required for the video enlargement unit 405, but the enlarged video signal 309 can be displayed almost continuously and stably with respect to the size of the subject. If no data is stored in the frame memory, the video enlargement unit 405 may output the enlarged video signal 309.
 次に、図14A、図14Bを用いて、映像解析部103が出力するズーム制御パラメータ307について説明する。図14Aでは、包含された切出し枠306cを破線で示している。また、包含された切出し枠306cを画面上で特定するために、上部左端の座標(Xs、Ys)と下部右端の座標(Xe、Ye)とを示している。この場合、上部左端の座標(Xs、Ys)と下部右端の座標(Xe、Ye)とは、画面の上端および下端に位置している。したがって、カメラ制御部106は、映像解析部103から出力されるズーム制御パラメータ307に基づいて、カメラ101のズーム値を調整する。具体的には、ズーム制御パラメータ307は、映像解析部103で映像領域内における背景部分の占める割合が小さい場合であって、被写体が撮影範囲をはみだしている場合、さらに言えば、背景の占める範囲が所定の背景閾値より大きい場合、カメラ101のズーム値を制御するために用いる。ここで、所定の背景閾値は、例えば、映像領域内における背景部分の占める割合が映像領域の1/4としてもよい。 Next, the zoom control parameter 307 output from the video analysis unit 103 will be described with reference to FIGS. 14A and 14B. In FIG. 14A, the included cutout frame 306c is indicated by a broken line. In addition, in order to specify the included cutout frame 306c on the screen, upper left end coordinates (Xs, Ys) and lower right end coordinates (Xe, Ye) are shown. In this case, the upper left end coordinates (Xs, Ys) and the lower right end coordinates (Xe, Ye) are located at the upper and lower ends of the screen. Therefore, the camera control unit 106 adjusts the zoom value of the camera 101 based on the zoom control parameter 307 output from the video analysis unit 103. Specifically, the zoom control parameter 307 is when the ratio of the background portion in the video area is small in the video analysis unit 103 and the subject is outside the shooting range. Is larger than a predetermined background threshold value, it is used to control the zoom value of the camera 101. Here, for example, the ratio of the background portion in the video area may be ¼ of the video area.
 すなわち、映像解析部103は、算出した切出し枠に基づいてカメラ101の撮影領域内における背景の位置を算出し、背景の占める範囲が所定の背景閾値より大きい場合、包含された切出し枠306cを映像調整部104に出力する。一方、映像解析部103は、背景の占める範囲が所定の背景閾値より小さい場合、カメラ101のズームをワイド側に調整する制御パラメータをカメラ制御部106に出力する。 That is, the video analysis unit 103 calculates the position of the background in the shooting area of the camera 101 based on the calculated cutout frame, and if the range occupied by the background is larger than a predetermined background threshold, the included cutout frame 306c is displayed in the video. Output to the adjustment unit 104. On the other hand, when the range occupied by the background is smaller than the predetermined background threshold, the video analysis unit 103 outputs a control parameter for adjusting the zoom of the camera 101 to the wide side to the camera control unit 106.
 また、図14Bでは、包含された切出し枠306dを示している。包含された切出し枠306dの下部右端の座標(Xe、Ye)が画面の右端に位置している。このように上部左端の座標(Xs、Ys)、または下部右端の座標(Xe、Ye)が画面の左端または右端に位置している場合にも、カメラ101のズーム値を制御するため、カメラ101のズームをワイド側に調整する制御パラメータをカメラ制御部106に出力する。 In FIG. 14B, an included cutout frame 306d is shown. The coordinates (Xe, Ye) of the lower right end of the included cutout frame 306d are located at the right end of the screen. Thus, even when the coordinates of the upper left end (Xs, Ys) or the coordinates of the lower right end (Xe, Ye) are located at the left end or the right end of the screen, the camera 101 controls the zoom value of the camera 101. A control parameter for adjusting the zoom to the wide side is output to the camera control unit 106.
 なお、被写体が画面に対して小さく表示され、かつ被写体の移動が小さい場合、包含された切出し枠は画面に対して小さく表示される。このような場合、カメラ101の位置を被写体が画面の中央に移動するように制御してもよい。また、ズームをテレ側に調整し、画面に被写体がより大きく表示されるように制御してもよい。このような制御により、ユーザはより被写体を認識しやすくできる。 When the subject is displayed small on the screen and the movement of the subject is small, the included cutout frame is displayed small on the screen. In such a case, the position of the camera 101 may be controlled so that the subject moves to the center of the screen. Alternatively, the zoom may be adjusted to the telephoto side so that the subject is displayed larger on the screen. Such control makes it easier for the user to recognize the subject.
 次に、図15を用いて、本実施の形態のおける映像処理装置100の映像処理方法について説明する。映像処理装置100の構成は、図3に示し、その動作については、既に説明した。図15は、本実施の形態のおける映像処理方法の手順を示すフローチャートである。 Next, a video processing method of the video processing apparatus 100 according to the present embodiment will be described with reference to FIG. The configuration of the video processing apparatus 100 is shown in FIG. 3, and its operation has already been described. FIG. 15 is a flowchart showing the procedure of the video processing method in the present embodiment.
 まず、カメラ101が、被写体を含む映像を撮影する(ステップS601)。映像取得部102は、カメラ101が、撮影した映像を取得し(ステップS602)、映像を映像信号301に変換して、映像解析部103と映像調整部104と映像出力部107とに出力する(ステップS603)。 First, the camera 101 captures an image including a subject (step S601). The video acquisition unit 102 acquires the video captured by the camera 101 (step S602), converts the video into the video signal 301, and outputs the video signal 301 to the video analysis unit 103, the video adjustment unit 104, and the video output unit 107 ( Step S603).
 映像解析部103は、映像信号301に対して時間方向に映像を解析する(ステップS604)。具体的には、映像信号301を入力して、映像信号301のフレーム間の差分を検出することにより、背景部と移動する被写体とを分離する。すなわち、背景検出部304が、映像信号301の連続するフレーム間における画素毎の差分を所定の閾値と比較することにより、現フレームNにおける画素毎に、動画素、もしくは静止画素と判定する。 The video analysis unit 103 analyzes the video in the time direction with respect to the video signal 301 (step S604). Specifically, by inputting the video signal 301 and detecting a difference between frames of the video signal 301, the background portion and the moving subject are separated. That is, the background detection unit 304 determines that each pixel in the current frame N is a moving pixel or a still pixel by comparing the difference for each pixel between successive frames of the video signal 301 with a predetermined threshold.
 ここで、所定の閾値は、例えば、10段階程度とする。ここで各画素は、8ビット(256段階の精度で映像信号を扱う場合)で表現されていると想定する。この際、画素毎の差分が所定の閾値を超える場合、その画素は動画素とする。この判定結果に基づいて、座標算出部351が、現フレームNにおける動画素の領域と、静止画素の領域とから、背景部分と被写体部分の境界632を算出する。映像解析部103は、境界632を切出し枠とする。 Here, the predetermined threshold is, for example, about 10 levels. Here, it is assumed that each pixel is expressed by 8 bits (when a video signal is handled with 256 levels of accuracy). At this time, if the difference for each pixel exceeds a predetermined threshold, the pixel is a moving pixel. Based on the determination result, the coordinate calculation unit 351 calculates the boundary 632 between the background portion and the subject portion from the moving pixel region and the still pixel region in the current frame N. The video analysis unit 103 uses the boundary 632 as a cutting frame.
 すなわち、映像解析部103は、映像信号301に基づいて切出し枠を算出する(ステップS605)。ここで、映像解析部103は、信号処理を簡略化するために、切出し枠を矩形の形状として算出する。そして、映像解析部103は、連続するフレームの差分が所定の閾値を超える領域を切出し枠として連続するフレーム毎に検出する(ステップS606)。そして、映像解析部103は、所定の複数のフレームにおいて、複数のすべての切出し枠を含むようにして包含された切出し枠306を算出する(ステップS607)。 That is, the video analysis unit 103 calculates a cutout frame based on the video signal 301 (step S605). Here, the video analysis unit 103 calculates the cutout frame as a rectangular shape in order to simplify the signal processing. Then, the video analysis unit 103 detects, for each successive frame, an area where the difference between successive frames exceeds a predetermined threshold as a cutout frame (step S606). Then, the video analysis unit 103 calculates a cut frame 306 that is included in a plurality of predetermined frames so as to include all of the plurality of cut frames (step S607).
 ここで、所定の複数のフレームの数は、例えば、10フレームとしている。そして、包含された切出し枠306の形状は、以後の信号処理をより簡便に行うために矩形とする。なお、映像解析部103は、連続するフレームの切出し枠の移動量が所定の移動量閾値よりも超える場合、切出し枠は、所定の連続する複数のフレームにおける複数のすべての切出し枠から除いてもよい。ここで、本実施の形態における映像処理装置100では、テレビ会議システムを前提とすれば、所定の移動量閾値は、例えば200画素程度が望ましい。 Here, the predetermined number of frames is, for example, 10 frames. The shape of the included cutout frame 306 is a rectangle so that the subsequent signal processing can be performed more easily. Note that the video analysis unit 103 may remove the cut frame from all of the plurality of cut frames in the predetermined continuous frames when the movement amount of the cut frames of the continuous frames exceeds a predetermined movement amount threshold value. Good. Here, in the video processing apparatus 100 according to the present embodiment, if a video conference system is assumed, the predetermined movement amount threshold is preferably about 200 pixels, for example.
 映像調整部104は、映像信号301と包含された切出し枠306と表示画面サイズ信号105とを入力する。そして、映像調整部104は、包含された切出し枠306と、表示画面サイズ信号105に基づいて算出した拡大率とに応じて、映像信号に基づいて、拡大後映像信号309を算出する(ステップS608)。ここで、表示画面サイズ信号105は、例えば制御部(図示せず)から入力される信号であって、映像処理装置100の出力映像信号108が表示する画面のアスペクト比と出力映像信号108の画面表示サイズとを示す情報を含んでいる。また、拡大率Eは、拡大後映像信号309における表示画面サイズとなる拡大切出し枠と包含された切出し枠306との比である。 The video adjustment unit 104 inputs the video signal 301, the included cutout frame 306, and the display screen size signal 105. Then, the video adjustment unit 104 calculates an enlarged video signal 309 based on the video signal according to the included cutout frame 306 and the enlargement ratio calculated based on the display screen size signal 105 (step S608). ). Here, the display screen size signal 105 is a signal input from a control unit (not shown), for example, and the screen aspect ratio displayed by the output video signal 108 of the video processing apparatus 100 and the screen of the output video signal 108 are displayed. Information indicating the display size is included. The enlargement ratio E is the ratio of the enlarged cutout frame that is the display screen size in the post-enlargement video signal 309 and the included cutout frame 306.
 そして、映像調整部104は、表示画面のアスペクト比と同等のアスペクト比を有するように包含された切出し枠306を拡張し、拡大後映像信号309として映像出力部107に出力する。この際、包含された切出し枠306bを、移動する被写体を表示する画面のほぼ中央に配置するように拡大して拡大後映像信号309を算出する。 Then, the video adjustment unit 104 expands the included cutout frame 306 so as to have an aspect ratio equivalent to the aspect ratio of the display screen, and outputs the expanded video signal 309 to the video output unit 107. At this time, the expanded cut-out frame 306b is enlarged so as to be arranged at substantially the center of the screen on which the moving subject is displayed, and the enlarged video signal 309 is calculated.
 映像調整部104は、映像解析部103で分離した背景部に基づいて、背景部以外を切り出した被写体を含む拡大後映像信号309を、映像出力部107へ出力する(ステップS609)。 The video adjustment unit 104 outputs the enlarged video signal 309 including the subject cut out from the background part to the video output unit 107 based on the background part separated by the video analysis part 103 (step S609).
 上記したように、映像調整部104が、包含された切出し枠306から、拡大後映像信号309を算出し、映像出力部107が、拡大後映像信号309を出力映像信号108として出力する。その結果、本実施の形態における映像処理方法によれば、フレームが変わっても、背景は、一定の領域が常に表示されている。その結果、ユーザにとって、このような画面を注視しても不快感は生じにくい。 As described above, the video adjustment unit 104 calculates the enlarged video signal 309 from the included cutout frame 306, and the video output unit 107 outputs the enlarged video signal 309 as the output video signal 108. As a result, according to the video processing method of the present embodiment, a constant area is always displayed as the background even if the frame changes. As a result, the user is less likely to feel discomfort even if such a screen is watched.
 本発明による映像処理装置は、異なるユーザ間で映像の送受信をする際、相手側への送信映像について、カメラに合わせてユーザが移動することなく相手に自身の映像を表示画面に対し適切なサイズで送信することが可能である。また異なる時間の映像差分から、背景部分を検出し背景部分以外を切出すことで、被写体の種類、色、数を問わずに一律に検出することが可能であるため、テレビやPCでビデオコミュニケーションを行う際の映像処理装置に有用である。 When the video processing apparatus according to the present invention transmits / receives video between different users, the video transmitted to the other party can be transmitted to the other party without moving the user according to the camera. Can be sent. In addition, by detecting the background part from the video difference at different times and cutting out the part other than the background part, it is possible to detect it uniformly regardless of the type, color, or number of subjects. This is useful for a video processing apparatus when performing the above.
 100  映像処理装置
 101  カメラ
 102  映像取得部
 103  映像解析部
 104  映像調整部
 105  表示画面サイズ信号
 106  カメラ制御部
 107  映像出力部
 108  出力映像信号
 301,301a  映像信号
 302  映像記憶部
 303  差分算出部
 304  背景検出部
 305  位置情報判定部
 306,306a,306b,306c,306d  包含された切出し枠
 307  ズーム制御パラメータ
 308  出力フラグ
 309  拡大後映像信号
 310  切出し位置記憶部
 312,313  切出し座標
 314  判定後切出し座標
 350a,350b  拡大切出し枠
 352  変化量算出部
 360a,360b  拡大開始位置
 405  映像拡大部
 601,602,603,1301,1302,1303  切出し枠
 601a,602a,603a,1401,1402,1403  画面
 632  境界
 E  拡大率
DESCRIPTION OF SYMBOLS 100 Image processing apparatus 101 Camera 102 Image | video acquisition part 103 Image | video analysis part 104 Image | video adjustment part 105 Display screen size signal 106 Camera control part 107 Image | video output part 108 Output image | video signal 301, 301a Image | video signal 302 Image | video storage part 303 Difference calculation part 304 Background Detection unit 305 Position information determination unit 306, 306a, 306b, 306c, 306d Included extraction frame 307 Zoom control parameter 308 Output flag 309 Image signal after enlargement 310 Extraction position storage unit 312, 313 Extraction coordinate 314 Post-determination extraction coordinate 350a, 350b Enlarged cutout frame 352 Change amount calculation unit 360a, 360b Enlargement start position 405 Video enlargement unit 601, 602, 603, 1301, 1302, 1303 Cutout frame 601a, 602a, 03a, 1401,1402,1403 screen 632 boundary E expansion rate

Claims (8)

  1. 被写体を含む映像を撮影するカメラと、
    前記カメラから前記映像を取得し、映像信号を出力する映像取得部と、
    前記映像信号に対して時間方向に映像を解析する映像解析部と、
    前記映像信号に基づいて前記映像解析部で算出された切出し枠と、表示画面サイズ信号に基づいて算出した拡大率とに応じて、前記映像信号に基づいて拡大後映像信号を算出して出力する映像調整部と、
    前記映像信号と、前記映像調整部が出力する前記拡大後映像信号とに基づいて、出力映像信号を出力する映像出力部と、を備え
    前記映像解析部は、
      連続するフレームの差分が所定の閾値を超える領域を前記切出し枠として前記連続するフレーム毎に検出し、
      所定の連続する複数のフレームにおいて、複数のすべての前記切出し枠を含むようにして、包含された切出し枠を算出する
    映像処理装置。
    A camera that shoots images including the subject;
    A video acquisition unit that acquires the video from the camera and outputs a video signal;
    A video analysis unit that analyzes video in a time direction with respect to the video signal;
    Based on the cutout frame calculated by the video analysis unit based on the video signal and the enlargement ratio calculated based on the display screen size signal, an enlarged video signal is calculated based on the video signal and output. A video adjustment unit;
    Based on the video signal and the enlarged video signal output by the video adjustment unit, a video output unit that outputs an output video signal, the video analysis unit,
    An area where the difference between successive frames exceeds a predetermined threshold is detected for each successive frame as the cutout frame,
    A video processing apparatus that calculates an included cutout frame so as to include a plurality of all the cutout frames in a predetermined plurality of consecutive frames.
  2. 前記映像出力部は、
     前記映像調整部が前記拡大後映像信号を算出した後に、前記映像解析部が出力する出力フラグに基づいて、
     前記カメラからの映像に基づく前記映像信号から、前記拡大後映像信号へと前記出力映像信号を切替える
    請求項1記載の映像処理装置。
    The video output unit
    After the video adjustment unit calculates the enlarged video signal, based on the output flag output by the video analysis unit,
    The video processing apparatus according to claim 1, wherein the output video signal is switched from the video signal based on the video from the camera to the enlarged video signal.
  3. 前記映像解析部は、
      前記連続するフレームの前記切出し枠の移動量が所定の移動量閾値よりも超える場合、
        前記切出し枠を、前記所定の連続する複数のフレームにおける前記複数のすべての前記切出し枠から除く
    請求項1に記載の映像処理装置。
    The video analysis unit
    When the movement amount of the cut frame of the continuous frames exceeds a predetermined movement amount threshold,
    The video processing apparatus according to claim 1, wherein the cutout frame is excluded from all of the plurality of cutout frames in the predetermined plurality of consecutive frames.
  4. 映像処理装置は、前記カメラへズームの制御信号を送信するカメラ制御部をさらに備え、
    前記位置情報判定部は、算出した前記切出し枠に基づいて前記カメラの撮影領域内における背景の位置を算出し、
      前記背景の占める範囲が所定の背景閾値より大きい場合、
        前記包含された切出し枠を前記映像調整部に出力し、
      前記背景の占める範囲が前記所定の背景閾値より小さい場合、
        前記カメラの前記ズームをワイド側に調整する制御パラメータを前記カメラ制御部に出力する
    請求項1記載の映像処理装置。
    The video processing apparatus further includes a camera control unit that transmits a zoom control signal to the camera,
    The position information determination unit calculates the position of the background in the shooting area of the camera based on the calculated cutout frame,
    If the range occupied by the background is greater than a predetermined background threshold,
    Outputting the included cutout frame to the video adjustment unit;
    When the range occupied by the background is smaller than the predetermined background threshold,
    The video processing apparatus according to claim 1, wherein a control parameter for adjusting the zoom of the camera to a wide side is output to the camera control unit.
  5. 前記映像出力部は、
     前記映像調整部が前記拡大後映像信号を算出した後に、前記映像解析部が出力する出力フラグに基づいて、
     前記カメラからの前記映像に基づく前記映像信号から、現フレームに先行する所定の連続する複数のフレームに対応する包含された切出し枠の情報を含む前記拡大後映像信号へと、前記出力映像信号を切替える
    請求項1記載の映像処理装置。
    The video output unit
    After the video adjustment unit calculates the enlarged video signal, based on the output flag output by the video analysis unit,
    The output video signal from the video signal based on the video from the camera to the post-enlarged video signal including information on included cutout frames corresponding to a plurality of predetermined consecutive frames preceding the current frame. The video processing apparatus according to claim 1 to be switched.
  6. 前記映像出力部は、
     1フレーム分のフレームメモリを有し、
     前記フレームメモリには、現フレームに先行する所定の連続する複数のフレームに対応する包含された切出し枠の情報を含む前記拡大後映像信号を格納し、
     前記映像調整部が前記拡大後映像信号を算出した後に、前記映像解析部が出力する出力フラグに基づいて、
     前記フレームメモリからの前記拡大後映像信号から、前記現フレームを含む前記所定の連続する前記複数のフレームに対応する前記包含された切出し枠の情報を含む前記拡大後映像信号へと、前記出力映像信号を切替える
    請求項1記載の映像処理装置。
    The video output unit
    It has a frame memory for 1 frame,
    In the frame memory, the enlarged video signal including information of included cutout frames corresponding to a plurality of predetermined consecutive frames preceding the current frame is stored,
    After the video adjustment unit calculates the enlarged video signal, based on the output flag output by the video analysis unit,
    The output video from the enlarged video signal from the frame memory to the enlarged video signal including information of the included cutout frames corresponding to the predetermined consecutive frames including the current frame The video processing apparatus according to claim 1, wherein the signal is switched.
  7. 表示画面サイズ信号は、前記出力映像信号の表示画面のアスペクト比と前記出力映像信号の画面表示サイズとを示す情報を含み、
    前記包含された切出し枠の形状は矩形であり、
    前記映像調整部は、前記表示画面の前記アスペクト比と同等のアスペクト比を有するように前記包含された切出し枠を拡張し、前記拡大後映像信号として前記映像出力部に出力する
    請求項1に記載の映像処理装置。
    The display screen size signal includes information indicating an aspect ratio of the display screen of the output video signal and a screen display size of the output video signal,
    The shape of the included cutting frame is a rectangle,
    The video adjustment unit extends the included clipping frame so as to have an aspect ratio equivalent to the aspect ratio of the display screen, and outputs the expanded cut-out video signal to the video output unit. Video processing equipment.
  8. 被写体を含む映像を撮影し、
    前記映像を取得し、
    前記映像を映像信号に変換して出力し、
    前記映像信号に対して時間方向に前記映像を解析し、
    前記映像信号に基づいて切出し枠を算出し、
    連続するフレームの差分が所定の閾値を超える領域を前記切出し枠として前記連続するフレーム毎に検出し、
    所定の連続する複数のフレームにおいて、複数のすべての前記切出し枠を含むようにして、包含された切出し枠を算出し、
    前記包含された切出し枠と、表示画面サイズ信号に基づいて算出した拡大率とに応じて、前記映像信号に基づいて拡大後映像信号を算出し、
    前記映像信号と前記拡大後映像信号とに基づいて、出力映像信号を出力する
    映像処理方法。
    Take a picture of the subject,
    Get the video,
    Converting the video into a video signal and outputting it,
    Analyzing the video in the time direction with respect to the video signal,
    Calculate a cutting frame based on the video signal,
    An area where the difference between successive frames exceeds a predetermined threshold is detected for each successive frame as the cutout frame,
    Calculating a contained cutout frame so as to include all of the cutout frames in a plurality of predetermined consecutive frames;
    According to the included cutout frame and the enlargement ratio calculated based on the display screen size signal, calculate an enlarged video signal based on the video signal,
    A video processing method for outputting an output video signal based on the video signal and the enlarged video signal.
PCT/JP2011/006619 2011-11-29 2011-11-29 Video processing device and video processing method WO2013080245A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/JP2011/006619 WO2013080245A1 (en) 2011-11-29 2011-11-29 Video processing device and video processing method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2011/006619 WO2013080245A1 (en) 2011-11-29 2011-11-29 Video processing device and video processing method

Publications (1)

Publication Number Publication Date
WO2013080245A1 true WO2013080245A1 (en) 2013-06-06

Family

ID=48534780

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2011/006619 WO2013080245A1 (en) 2011-11-29 2011-11-29 Video processing device and video processing method

Country Status (1)

Country Link
WO (1) WO2013080245A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022065189A1 (en) * 2020-09-23 2022-03-31 株式会社Aiメディカルサービス Examination assistance device, examination assistance method, and examination assistance program

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0765173A (en) * 1993-08-30 1995-03-10 Oki Electric Ind Co Ltd Intruding body recognition method
JP2006279894A (en) * 2005-03-30 2006-10-12 Casio Comput Co Ltd Image processing apparatus, image processing method, and program
JP2007316957A (en) * 2006-05-26 2007-12-06 Fujifilm Corp Image trimming device, image trimming method and program
JP2008141700A (en) * 2006-12-05 2008-06-19 Fujifilm Corp Monitoring system and method, and program

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0765173A (en) * 1993-08-30 1995-03-10 Oki Electric Ind Co Ltd Intruding body recognition method
JP2006279894A (en) * 2005-03-30 2006-10-12 Casio Comput Co Ltd Image processing apparatus, image processing method, and program
JP2007316957A (en) * 2006-05-26 2007-12-06 Fujifilm Corp Image trimming device, image trimming method and program
JP2008141700A (en) * 2006-12-05 2008-06-19 Fujifilm Corp Monitoring system and method, and program

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022065189A1 (en) * 2020-09-23 2022-03-31 株式会社Aiメディカルサービス Examination assistance device, examination assistance method, and examination assistance program

Similar Documents

Publication Publication Date Title
US8040398B2 (en) Imaging apparatus and method for displaying zoom information
US20090268079A1 (en) Image-capturing apparatus and image-capturing method
JP2007189503A (en) Terminal device and program
US9743000B2 (en) Moving image processing apparatus, imaging apparatus, and moving image processing method
US20180035076A1 (en) Video processing apparatus, video processing system, and video processing method
CN113302911A (en) Camera control
JP5599063B2 (en) Display control apparatus, display control method, and program
US20150288949A1 (en) Image generating apparatus, imaging apparatus, and image generating method
JP2005073218A (en) Image processing apparatus
US10911686B2 (en) Zoom control device, zoom control method, and program
JP2006033380A (en) Monitoring system
JP2012173683A (en) Display control device, information display apparatus, and display control method
JP2018061130A (en) Image processing device, image processing method, and program
TW202002606A (en) Image-capturing device and method for operating the same
JP2013179708A (en) Imaging system and imaging method
WO2013080245A1 (en) Video processing device and video processing method
JP2018207415A (en) Imaging apparatus comprising visual line detection function
KR20110090623A (en) Home network terminal being capable of tracking face and method thereof
JP5072103B2 (en) Angle of view control apparatus and angle of view control method
EP2918079A1 (en) Handheld display zoom feature
KR20140106351A (en) Analysis apparatus of smart golf swing
JP6930880B2 (en) Video display device and video production support method
CN115134505A (en) Preview picture generation method and device, electronic equipment and storage medium
KR100917166B1 (en) Leture shooting device and leture shooting method
JP7271749B2 (en) Imaging device with line-of-sight detection function

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11876844

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 11876844

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP