WO2022244314A1 - 情報処理装置、情報処理方法およびプログラム - Google Patents
情報処理装置、情報処理方法およびプログラム Download PDFInfo
- Publication number
- WO2022244314A1 WO2022244314A1 PCT/JP2022/003240 JP2022003240W WO2022244314A1 WO 2022244314 A1 WO2022244314 A1 WO 2022244314A1 JP 2022003240 W JP2022003240 W JP 2022003240W WO 2022244314 A1 WO2022244314 A1 WO 2022244314A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- composition
- subject
- transition mode
- information processing
- video
- Prior art date
Links
- 230000010365 information processing Effects 0.000 title claims abstract description 49
- 238000003672 processing method Methods 0.000 title claims description 14
- 239000000203 mixture Substances 0.000 claims abstract description 378
- 230000007704 transition Effects 0.000 claims abstract description 240
- 238000004364 calculation method Methods 0.000 claims abstract description 39
- 230000008859 change Effects 0.000 claims description 63
- 238000009432 framing Methods 0.000 claims description 38
- 238000010191 image analysis Methods 0.000 claims description 7
- 238000012545 processing Methods 0.000 description 65
- 239000013256 coordination polymer Substances 0.000 description 55
- 238000010586 diagram Methods 0.000 description 39
- 238000000034 method Methods 0.000 description 33
- 238000003384 imaging method Methods 0.000 description 13
- 230000008569 process Effects 0.000 description 12
- 210000000707 wrist Anatomy 0.000 description 12
- 210000003128 head Anatomy 0.000 description 9
- 230000000694 effects Effects 0.000 description 7
- 210000000887 face Anatomy 0.000 description 6
- 230000004048 modification Effects 0.000 description 6
- 238000012986 modification Methods 0.000 description 6
- 230000009467 reduction Effects 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 4
- 238000004891 communication Methods 0.000 description 3
- 230000001815 facial effect Effects 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 210000004247 hand Anatomy 0.000 description 3
- 230000001965 increasing effect Effects 0.000 description 3
- 238000013135 deep learning Methods 0.000 description 2
- 238000006073 displacement reaction Methods 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 210000003127 knee Anatomy 0.000 description 2
- 230000015654 memory Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 101100088247 Picea mariana RPL13A gene Proteins 0.000 description 1
- 101100202463 Schizophyllum commune SC14 gene Proteins 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 230000002708 enhancing effect Effects 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000001151 other effect Effects 0.000 description 1
- 238000004091 panning Methods 0.000 description 1
- 230000000284 resting effect Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000007480 spreading Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/667—Camera operation mode switching, e.g. between still and video, sport and normal or high- and low-resolution modes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/1116—Determining posture transitions
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B15/00—Special procedures for taking photographs; Apparatus therefor
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B17/00—Details of cameras or camera bodies; Accessories therefor
-
- G—PHYSICS
- G03—PHOTOGRAPHY; CINEMATOGRAPHY; ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ELECTROGRAPHY; HOLOGRAPHY
- G03B—APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR
- G03B17/00—Details of cameras or camera bodies; Accessories therefor
- G03B17/56—Accessories
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
- H04N23/611—Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/64—Computer-aided capture of images, e.g. transfer from script file into camera, check of taken image quality, advice or proposal for image composition or decision on when to take image
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/69—Control of means for changing angle of the field of view, e.g. optical zoom objectives or electronic zooming
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/268—Signal distribution or switching
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2503/00—Evaluating a particular growth phase or type of persons or animals
- A61B2503/12—Healthy persons not otherwise provided for, e.g. subjects of a marketing survey
Definitions
- the present invention relates to an information processing device, an information processing method, and a program.
- the conditions such as the direction of movement and orientation are not always the same.
- the subject does not always face right, but turns.
- the timing for changing the orientation also changes from time to time, and there are cases where the orientation changes frequently depending on the content of the lecture and the interaction with the audience.
- a technique for determining the orientation of a subject includes, for example, Patent Document 1.
- Patent Document 1 since framing is not the purpose of Patent Document 1, proper framing as described above cannot be obtained by applying the orientation of the subject obtained here as it is.
- the orientation of the subject is determined using the orientation of the hand, but the orientation in which the hand is extended does not necessarily match the direction in which the subject faces.
- no consideration is given to appropriately capture the continuity and change in the orientation of the subject in the time direction.
- Patent Document 2 Patent Document 3, and Patent Document 4.
- the direction of the line of sight is determined, and the composition of the space in that direction is devised.
- no consideration is given to how changes in the direction of the line of sight are captured in the time direction and applied to framing.
- the present disclosure proposes an information processing device, an information processing method, and a program capable of setting an appropriate composition according to changes in the state of a subject.
- a target composition calculation unit that calculates a target composition based on the state of a subject
- An information processing device having a composition transition determination unit for switching between a smooth composition transition mode in which composition transitions and an instantaneous composition transition mode in which the composition instantaneously transitions toward the target composition is provided. Further, according to the present disclosure, there are provided an information processing method in which the information processing of the information processing device is executed by a computer, and a program for causing the computer to implement the information processing of the information processing device.
- FIG. 5 is a diagram showing an example of posture information acquired by a posture acquisition unit; It is a figure explaining the determination method of the direction of a to-be-photographed object. It is a figure explaining the determination method of the direction of a to-be-photographed object. It is a figure which shows an example of an orientation determination flow.
- FIG. 5 is a diagram showing an example of posture information acquired by a posture acquisition unit; It is a figure explaining the determination method of the direction of a to-be-photographed object. It is a figure explaining the determination method of the direction of a to-be-photographed object. It is a figure which shows an example of an orientation determination flow.
- FIG. 10 is a diagram showing an example in which orientation determination is performed based on the distance between the wrist and the body axis instead of the orientation determination based on the subject area; It is a figure explaining the determination method of the to-be-photographed object's size. It is a figure which shows an example of a size determination flow. It is a figure which shows an example of a motion determination flow. It is a figure explaining the calculation method of target composition.
- FIG. 7 is a diagram illustrating an example of a target composition calculation flow
- FIG. 7 is a diagram illustrating an example of a target composition calculation flow
- FIG. 10 is a diagram showing a smooth transition of composition when starting to walk from a resting state;
- FIG. 10 is a diagram showing an example in which the subject frequently changes direction due to exchange of questions and calls from listeners;
- FIG. 11 is a diagram showing an example of composition transition in an instantaneous composition transition mode; It is a figure which shows the example which performs an instantaneous transition first and then performs a smooth transition. It is a figure which shows an example of the determination flow of a transition mode. It is a figure which shows the imaging system of 2nd Embodiment. It is a figure which shows the imaging system of 3rd Embodiment.
- FIG. 3 is a diagram showing an example of a functional configuration of a video switcher; FIG. FIG.
- FIG. 10 is a diagram showing an example of a selection flow of captured video; It is a figure explaining the outline
- FIG. 21 is a diagram illustrating an overview of framing according to the fifth embodiment;
- FIG. It is a figure which shows an example of the functional structure of a video processing apparatus.
- FIG. 4 is a diagram showing an example of a control flow of a PTZ camera; It is a figure which shows the hardware structural example of an imaging system.
- FIG. 1 is a diagram showing an imaging system CS1 of the first embodiment.
- the photography system CS1 is a system that performs lecture capture (lecture recording).
- the photographing system CS1 automatically tracks the lecturer, who is the subject SU, and automatically records the content of the lecture.
- the imaging system CS1 has a video processing device 10, a camera 20, a display device 30 and a recording device 40.
- the camera 20 can photograph the range in which the subject SU moves around.
- the video processing device 10 performs framing on the captured video CV captured by the camera 20 .
- Framing is a process of cutting out a video area according to the composition from the shot video CV. Since this process is similar to the work of adjusting the angle of view, in the following description, the image area to be clipped will be referred to as angle of view AV (see FIG. 9).
- composition means the composition regarding the position and size of the subject SU captured in the video.
- the screen is vertically and horizontally divided into three, and the subject SU is arranged at the intersection of the division lines DL (see FIG. 15).
- a composition is adopted in which the subject SU is arranged on the left division line DL (a lead room is provided on the right side of the subject SU).
- a composition is adopted in which the subject SU is placed on the right division line DL (a lead room is provided on the left side of the subject SU).
- the size of the subject SU compositions such as long shot, full figure, knee shot, waist shot, bust shot and close-up are known depending on the range of the subject SU to be photographed.
- the video processing device 10 determines a target composition (target composition) based on the state of the subject SU. For example, the video processing device 10 stores a plurality of types of compositions with different positions or sizes of the subject SU as standard compositions. The video processing device 10 selects a specific type of composition according to the state of the subject SU as the target composition.
- the video processing device 10 places the subject SU on the division line DL on the left side of the screen and determines a composition in which the subject SU is framed with a bust shot as the target composition. do.
- the video processing device 10 arranges the subject SU on the division line DL on the right side of the screen and determines a composition in which the subject SU is framed in a full figure as a target composition.
- the state of the subject SU is obtained, for example, by image analysis of the captured video CV of the subject SU.
- the video processing device 10 determines the next composition based on the current composition and the target composition TCP.
- the video processing device 10 generates angle-of-view information indicating the position and size of the angle-of-view AV (image area to be clipped) based on the determined composition.
- the video processing device 10 cuts out a video region based on the angle-of-view information.
- the video processing device 10 outputs the video region cut out from the shot video CV as the framing video OV.
- the display device 30 displays the framing video OV output by the video processing device 10 .
- the recording device 40 records the framing video OV.
- Each device included in the imaging system CS1 may be directly connected via HDMI (registered trademark) (High-Definition Multimedia Interface) or SDI (Serial Digital Interface), or may be connected via a wired or wireless network. may be connected to each other.
- FIG. 2 is a block diagram showing the functional configuration of the video processing device 10. As shown in FIG.
- the video processing device 10 is an information processing device that processes various types of information.
- the video processing device 10 has a posture acquisition unit 11, a subject tracking unit 12, an orientation determination unit 13, a size determination unit 14, a motion determination unit 15, a target composition calculation unit 16, a composition transition determination unit 17, and a video clipping unit 18. .
- Fig. 3 is a flow chart showing the flow of the overall video processing.
- the image processing device 10 After initializing the states of counters required for processing (step SA1), the image processing device 10 acquires the posture of the subject SU and the head area HEA (see FIG. 14) in the posture acquisition unit 11. (Step SA2). Next, the video processing device 10 acquires the motion of the subject SU by the subject tracking section 12 by matching with the previous frame (step SA3). Based on the information acquired in this way, the image processing device 10 determines the orientation of the subject SU, the size of the subject SU, and the movement of the subject SU by the orientation determining section 13, the size determining section 14, and the movement determining section 15. (Steps SA4, SA5, SA6).
- the video processing device 10 calculates the target composition TCP (see FIG. 15) suitable for the input frame in the target composition calculation unit 16 (step SA7). Subsequently, the composition transition determining unit 17 of the video processing apparatus 10 determines the transition from the current composition CP (see FIG. 15) to the target composition TCP, and determines the composition CP for the input frame (step SA8). The video clipping unit 18 clips and outputs the framing video OV based on the determined composition CP (step SA9).
- the video processing device 10 determines whether framing has been performed for all frames (step SA10). If there is a next frame (step SA10: no), the process returns to step SA2 and the above processing is repeated until framing is performed for all frames. If there is no next frame (step SA10: yes), the video processing device 10 terminates the process.
- FIG. 4 is a diagram showing an example of posture information acquired by the posture acquisition unit 11. As shown in FIG.
- the posture acquisition unit 11 acquires the body part BP used by the orientation determination unit 13 and the size determination unit 14 from the captured video CV by image analysis processing.
- the posture acquisition unit 11 extracts posture information of the subject SU by analyzing the image of the body part BP.
- skeleton data SK as shown on the left side of FIG. It is possible to obtain the position of JT.
- the position of the joint JT can be used as part of the posture information.
- the head area HEA can also be used as part of the pose information.
- the position of the joint JT and the head area HEA may be obtained by separate estimation processes, or may be obtained by one estimation process.
- the subject tracking unit 12 detects movement of the subject SU between frames. For this, there is a movement amount calculation process in which an image is directly input, such as an optical flow process that performs matching for each pixel, and a one-time image processing such as a matching process between frames of the body part BP obtained by the posture acquisition unit 11. It is possible to use a movement amount calculation process based on the data obtained by the above. In order to improve robustness, it is also possible to calculate the movement amount using both the image and the matching of the body part BP.
- Orientation determination unit 5 and 6 are diagrams for explaining a method of determining the orientation of the subject SU.
- the orientation determination unit 13 uses the orientation information obtained by the orientation acquisition unit 11 to determine the orientation of the subject SU.
- the determination of the orientation is performed based on, for example, the displacement of the body part BP with respect to the body axis AX of the subject SU.
- the displacement of the position of the nose NS with respect to the body axis AX exceeds the threshold, is determined to be
- it is also possible to use the deviation of other facial parts such as the eyes EY from the body axis AX.
- the orientation determination unit 13 determines the orientation of the subject SU based on the pose of the subject SU. For example, even if the subject SU faces the front, depending on the pose of the subject SU, a well-balanced image may be obtained by shifting the subject SU from the center of the screen.
- FIG. 6 shows an example in which the subject SU spreads his right hand for explanation, but in this example, it is preferable to shift the subject SU to the right side of the screen.
- the orientation determination unit 13 detects the bias in posture based on the pose of the subject SU, and determines the bias in the posture as the orientation of the subject SU.
- the pose-based determination for example, whether or not the distance D1 between the subject SU's right wrist (the fourth joint on the left side in FIG. 4) and the body axis AX exceeds a threshold (left side in FIG. 6), whether or not the subject area SUA Information such as whether or not the distance D2 between the center and the body axis AX exceeds a threshold (right side of FIG. 6) can be used.
- the subject area SUA means an area of the subject SU that includes the main part of the subject SU to be photographed. For example, in a bust shot, a rectangular area including joints of the upper body of the subject SU is the subject area SUA.
- the orientation determination unit 13 determines the orientation using a plurality of calculation means such as body axis AX and facial parts, and body axis AX and pose.
- FIG. 7 is a diagram showing an example of the orientation determination flow.
- the orientation of the subject SU is determined from the relationship between the nose NS, the body axis AX, the shoulder width SW, and the subject area SUA and the body axis AX.
- the orientation determining unit 13 first calculates a body axis AX along the body from the base of the neck (step SB1). Since the posture of the subject SU assumed in lecture capture is a standing posture, the body axis AX is normally a line extending vertically from the base of the neck.
- three types of orientation determination processing are performed. From the left of the figure, orientation determination by face (steps SB2 to SB7), orientation determination by shoulder width SW (steps SB8 to SB11), orientation using subject area SUA. Judgment (steps SB12 to SB17) is made.
- step SB7 it is calculated whether the nose NS is on the left side or the right side of the screen with respect to the body axis AX.
- the nose NS is determined to be on the right side of the body axis AX on the screen. If the distance d is a negative value, the nose NS is determined to be on the left side of the body axis AX.
- d x0-b. If the nose NS is on the right side (x0>b) of the body axis AX, then d>0. If the nose NS is on the left side of the body axis AX (x0 ⁇ b), then d ⁇ 0.
- the orientation determination unit 13 determines whether or not the absolute value of the distance d is greater than a threshold (step SB3). If the absolute value of the distance d is equal to or less than the threshold (step SB3: no), the orientation determination unit 13 determines that the subject SU is facing forward (step SB7). If the absolute value of the distance d is greater than the threshold (step SB3: yes), the orientation determining unit 13 determines whether the direction is rightward or leftward based on the sign of the distance d (steps SB4 to SB6).
- the orientation determination unit 13 calculates the distance between the shoulders as the shoulder width SW (step SB8). The orientation determination unit 13 determines whether or not the absolute value of the shoulder width SW is greater than a threshold (step SB9). If the absolute value of the shoulder width SW is larger than the threshold (step SB9: yes), the orientation determination unit 13 determines that the subject SU is facing forward (step SB10). If the absolute value of the shoulder width SW is equal to or less than the threshold (step SB9: no), the orientation determination unit 13 determines that the subject SU is not facing the front (step SB11).
- the orientation determination unit 13 calculates an area including upper body parts obtained by the posture acquisition unit 11 as the subject area SUA. Orientation determination unit 13 calculates a distance D1 between the center of subject area SUA and body axis AX by a method similar to the method for calculating distance d (step SB12). The orientation determination unit 13 determines whether or not the absolute value of the distance D1 is greater than a threshold (step SB13). If the absolute value of the distance D1 is equal to or less than the threshold (step SB13: no), the orientation determination unit 13 determines that the subject SU is facing forward (step SB17). If the absolute value of the distance D1 is greater than the threshold (step SB13: yes), the direction determination unit 13 determines whether the direction is rightward or leftward based on the sign of the distance D1 (steps SB14 to SB16).
- the orientation determination unit 13 comprehensively determines left/right/front when a plurality (three types) of orientation determination results are obtained (step SB18). Considering a case where the neck position cannot be obtained, the orientation determination unit 13 adopts the orientation when two of the three orientation determination results indicate the same direction, thereby enhancing orientation robustness. If it is determined in step SB11 that the orientation of the subject SU is not the front, the orientation of the subject SU may be right or left.
- FIG. 8 is a diagram showing an example of orientation determination based on the distance D2 between the wrist and the body axis AX instead of the orientation determination based on the subject area SUA.
- Direction determination by the face steps SC2 to SC7
- direction determination by the shoulder width SW steps SC8 to SC11
- the orientation determination unit 13 calculates the distance D2 between the right wrist and the body axis AX by the same method as the method for calculating the distance d (step SC12). The orientation determination unit 13 determines whether or not the absolute value of the distance D2 is greater than a threshold (step SC13). If the absolute value of the distance D2 is equal to or less than the threshold (step SC13: no), the orientation determination unit 13 determines that the subject SU faces the front (step SC17). If the absolute value of the distance D2 is greater than the threshold (step SC13: yes), the direction determining unit 13 determines whether the direction is rightward or leftward based on the sign of the distance D2 (steps SC14 to SC16).
- the orientation determination unit 13 performs the same orientation determination flow for the left wrist as for the right wrist. That is, the orientation determination unit 13 calculates the distance D2 between the left wrist and the body axis AX by the same method as the method for calculating the distance d (step SC18). The orientation determination unit 13 determines whether or not the absolute value of the distance D2 is greater than a threshold (step SC19). When the absolute value of the distance D2 is equal to or less than the threshold (step SC19: no), the orientation determination unit 13 determines that the subject SU faces the front (step SC23). If the absolute value of the distance D2 is greater than the threshold (step SC19: yes), the direction determination unit 13 determines whether the direction is rightward or leftward based on the sign of the distance D2 (steps SC20 to SC22).
- the orientation determination unit 13 uses the four types of orientation determination results to determine the overall orientation (step SC24). At this time, it is desirable to first integrate the determination results using the right wrist and the left wrist, and then integrate the integrated determination results with the other two types of determination results.
- the orientation determination unit 13 determines three types of orientation determination results by combining one type of integrated determination result and the remaining two types of determination results (orientation determination result based on the face and orientation determination result based on the shoulder width SW). to determine the orientation of the subject SU.
- the orientation determination unit 13 outputs the determination result that the orientation of the subject SU is the front as an integrated determination result, and performs left/right/front determination in combination with the other two types of determination results. . If two of the three orientation determination results indicate the same orientation, the orientation determination unit 13 adopts that orientation. This increases orientation robustness.
- the thresholds used in the flows of FIGS. 7 and 8 are set to appropriate values according to the part of the body to be measured.
- the threshold is preferably represented by a ratio based on the size of the specific part of the subject SU. For example, if the threshold is set to 0.2 times the height of the head area HEA, it is possible to cope with changes in the size of a person due to changes in zoom.
- FIG. 9 is a diagram for explaining a method of determining the size of the subject SU.
- the size determination unit 14 determines whether the pose of the subject SU is large or small. On the left side of FIG. 9, the subject SU is shown expressing joy by spreading both hands HA. The right side of FIG. 9 shows the subject SU with a small fist pose. For these subjects SU, the size determination unit 14 extracts, for example, an area surrounding the upper body or the whole body of the subject SU as a subject area SUA, and determines the size of the subject area SUA as the size of the subject SU. The determination result of the size is reflected in the composition CP.
- FIG. 10 is a diagram showing an example of the size determination flow.
- the user often designates an approximate size in advance.
- description will be made on the assumption that the user has designated the extraction of a bust shot.
- the size determination unit 14 uses the position of the body part BP and the head area HEA obtained by the posture acquisition unit 11 to calculate a rectangular area surrounding the upper body of the subject SU as the subject area SUA (step SD1). .
- the size determination unit 14 calculates how much the subject area SUA has increased from the previous frame as a difference from the previous frame (step SD2). As the difference, the difference between the vertical direction and the horizontal direction of the subject SU is obtained, and the size determination unit 14 selects the difference with the larger absolute value, for example.
- the magnitude determination unit 14 determines whether or not the absolute value of the calculated difference is greater than the threshold (step SD3). When the absolute value of the difference is equal to or less than the threshold (step SD3: no), the size determination unit 14 determines that there is no change in size (step SD7). If the absolute value of the difference is greater than the threshold (step SD3: yes), the size determination unit 14 determines whether the size has been reduced or expanded based on the sign of the difference (steps SD4 to SD6). ).
- the size determination unit 14 basically determines the reduction ratio and enlargement ratio of the angle of view AV (image area in the shot image CV to be cut out) so as to include the subject area SUA.
- the reduction ratio and enlargement ratio can also be determined in advance. If the enlargement ratio and the reduction ratio are determined in advance, although it depends on the user's request, the switching of the angle of view AV becomes clear, and the image becomes easy to see.
- the motion determination unit 15 calculates the direction and amount of movement of the subject SU in the current frame with respect to the previous frame from the subject tracking result obtained by the subject tracking unit 12 .
- the motion determination unit 15 determines whether the subject SU has moved based on the calculation result of the movement method and the movement amount.
- FIG. 11 is a diagram showing an example of the motion determination flow.
- the motion determination unit 15 acquires the subject position of the current frame and the subject position of the previous frame based on the tracking result.
- the subject position is obtained, for example, as the coordinates of the subject SU in the camera coordinate system.
- the motion determination unit 15 calculates the amount of movement of the subject SU by subtracting the subject position in the previous frame from the subject position in the current frame (step SE1).
- the amount of movement is calculated as a signed amount of movement in which the right direction is positive. When moving to the right side of the screen, the amount of movement is positive, and when moving to the left side, the amount of movement is negative.
- the motion determination unit 15 determines whether or not the absolute value of the movement amount is greater than the threshold (step SE2). When the absolute value of the movement amount is equal to or less than the threshold (step SE2: no), the motion determination unit 15 determines that the subject SU is not moving (step SE6). When the absolute value of the movement amount is larger than the threshold (step SE2: yes), the movement determination unit 15 determines whether the movement direction is rightward or leftward based on the sign of the movement amount (steps SE3 to SE5).
- FIG. 12 is a diagram illustrating a method of calculating the target composition.
- the state of the subject SU is specified based on the determination results of the orientation determination unit 13, the size determination unit 14, and the motion determination unit 15.
- the state of the subject SU includes, for example, a state related to at least one of the moving direction of the subject SU, the orientation of the subject SU, and the size of the subject SU. For example, in the left example of FIG. 12, a state in which the subject SU faces left and moves leftward is detected.
- the target composition calculation unit 16 calculates the target composition (target composition TCP) of the framing video OV based on the state of the subject SU. For example, the target composition calculation unit 16 determines the size of the target composition TCP (the size of the angle of view AV) based on the determination result of the size. The target composition calculation unit 16 determines the spatial margins on the left and right sides of the subject SU based on the direction and motion determination results. As a result, the target composition TCP including left and right margins is determined.
- the determination result of the movement direction and the determination result of the orientation of the subject SU may differ. Judgment results take precedence. In many cases, the composition CP of the previous frame is continued.
- FIG. 13 and 14 are diagrams showing an example of the target composition calculation flow.
- the target composition calculation unit 16 decrements the counter by a preset set number when the orientation determination unit 13 determines that the direction is leftward, and increases the counter by the same set number when the direction determination unit 13 determines that the direction is rightward.
- the target composition calculation unit 16 does not change the counter.
- the target composition calculation unit 16 determines leftward or rightward based on the sign of the counter.
- the target composition calculation unit 16 When the motion determination unit 15 determines left direction, right direction, or no motion, the target composition calculation unit 16 performs processing to decrease, increase, or keep the motion counter, respectively. When the absolute value of the counter exceeds the threshold, the target composition calculation unit 16 determines that there is movement, and determines leftward or rightward based on the sign of the counter.
- the determination result of direction determination and movement determination is used as an index for creating a margin in the direction of the determination result by shifting the subject SU from the center of the screen.
- the target composition calculation unit 16 determines the shift amount of the subject SU by prioritizing this movement direction.
- the target composition calculation unit 16 gives priority to the direction determination result.
- the target composition calculation unit 16 places the subject SU at the center of the screen when the direction determination determines that the subject is facing forward, and when the direction determination determines that the subject SU faces left or right, the target composition calculation unit 16
- the shift amount of the subject SU is determined so that there is a margin in that direction.
- the target composition calculation unit 16 increases the size counter.
- the target composition calculation unit 16 decrements the size counter when the size determination unit 14 determines that the subject area SUA has been reduced.
- the target composition calculation unit 16 does not change the size of the angle of view AV if the absolute value is less than or equal to the threshold, and if it exceeds the absolute value, based on the sign of the size counter, Determines enlargement and reduction of the angle of view AV.
- the output of the target composition calculation unit 16 is, for example, the size of the angle of view AV and the relative position of the body axis AV starting from the top of the head of the subject SU within the angle of view AV.
- the starting point is obtained as the intersection of the upper side of the head area HEA in FIG. 14 and the body axis AX.
- Composition transition determination unit 15 to 18 are diagrams for explaining the composition CP transition method.
- the current composition CCP, the target composition TCP, and the next composition NCP are shown as image areas (angle of view AV) to be clipped.
- the current composition CCP means the composition CP of the video area cut out as the framing video OV of the current frame.
- the next composition NCP means the composition CP of the video area to be cut out as the framing video OV of the next frame.
- the composition transition determining unit 17 determines the transition mode from the current composition CP to the target composition TCP. The composition transition determining unit 17 determines the next composition NCP based on the determined transition mode.
- the transition modes include a smooth composition transition mode and an instant composition transition mode.
- the smooth composition transition mode is a transition mode in which the composition CP gradually transitions from the current composition CCP to the target composition TCP.
- the instantaneous composition transition mode is a transition mode in which the composition CP instantaneously transitions from the current composition CCP to the target composition TCP.
- the composition CP transitions quickly in a shorter period than in the smooth composition transition mode.
- An instant means, for example, a period within several preset frames. The duration of this transition is, for example, 5 frames or less, preferably 1 frame.
- the composition transition determining unit 17 switches the transition mode to the target composition TCP between the smooth composition transition mode and the instantaneous composition transition mode based on the change in the state of the subject SU.
- the state of change in the state of the subject SU is, for example, the state of at least one of the speed at which the direction of movement of the subject SU is switched, the speed at which the direction of the subject SU is switched, and the change in the size of the subject SU. include.
- a situation in which the transition mode should be switched from the smooth composition transition mode to the instantaneous composition transition mode is, for example, a situation in which the type of the target composition TCP changes while the composition is smoothly transitioning toward the target composition TCP. means.
- the position where the lead room should be provided is from left to right or from right to left. to the left. Therefore, another type of target composition TCP with a different installation direction of the lead room is determined.
- the type of target composition TCP also changes when the pose of the subject SU changes and it becomes appropriate to shoot from a bust shot to a full shot. In such a case, the transition mode is switched from the smooth composition transition mode to the instant composition transition mode.
- the occurrence of a situation in which the transition mode should be switched from the smooth composition transition mode to the instant composition transition mode is defined as a change in the orientation or movement direction of the subject SU within a predetermined reference time that causes a change in composition type as described above. or there is a change in the size of the subject SU by a predetermined reference magnification or more. Note that the reference time and the reference magnification can be appropriately set according to the user.
- FIG. 15 is a diagram showing a smooth transition of the composition CP when walking from a stationary state.
- the target composition TCP is determined based on the rule of thirds.
- the dashed-dotted line in FIG. 15 indicates the position of the division line DL that divides the screen into three.
- the subject SU In the initial stationary state, the subject SU is placed in the center of the angle of view AV.
- the target composition calculation unit 16 calculates the position and size of the angle of view AV (target angle of view) of the target composition TCP so that a margin is formed in the traveling direction.
- the composition transition determining unit 17 calculates the result of interpolation between the angle of view AV of the current composition CCP and the target angle of view as the angle of view AV of the next composition NCP. do.
- the composition transition determination unit 17 generates information about the position and size of the angle of view AV of the next composition NCP as angle of view information, and supplies the information to the video clipping unit 18 .
- the composition transition determination unit 17 gradually brings the position and size of the angle of view AV closer to the target angle of view by obtaining the angle of view AV of the next composition NCP by interpolation processing.
- a smooth transition of the angle of view AV is natural, and is an effective transition method when the movement direction and the orientation of the subject SU do not change frequently.
- FIG. 16 is a diagram showing an example in which the subject SU frequently changes direction due to exchanges of questions and calls by listeners.
- the composition transition determining unit 17 instantaneously transitions the composition CP toward the target composition TCP.
- FIG. 17 is a diagram showing an example of transitioning the composition CP in the instant composition transition mode.
- the composition transition determining unit 17 When a new target composition TCP with a different position and size of the target angle of view is determined while the composition CP is transitioning in the smooth composition transition mode, the composition transition determining unit 17 instantly changes the transition mode. Switch to composition transition mode. For example, immediately after the transition starts in the smooth composition transition mode, or in a situation away from the target composition TCP, the orientation of the subject SU changes and the position and size of the target angle of view change. , the composition transition determining unit 17 instantaneously transitions the position and size of the angle of view AV of the next composition NCP toward the position and size of the target angle of view.
- the composition transition determining unit 17 determines the change rate (enlargement ratio or reduction ratio) of the angle of view AV immediately after the transition mode is switched from the smooth composition transition mode to the instant composition transition mode as the angle of view immediately before switching the transition mode. Differentiate the rate of change of the AV magnitude. This is because by adjusting the size of the angle of view AV as well as the position of the angle of view AV, an instantaneous change in the composition CP is less likely to be recognized as disturbance of the image. For example, if only the position of the angle of view AV is changed without changing the rate of change of the angle of view AV, it may give the impression that the camera 20 has hit something and the image is disturbed. When the size of the angle of view AV is changed, such a sense of incongruity is less likely to occur.
- the composition transition determining unit 17 when the rate of change in the magnitude of the angle of view AV at the instantaneous transition calculated from the size of the target angle of view is different from the rate of change in the magnitude of the angle of view AV immediately before the instantaneous transition, the composition transition determining unit 17 , the size of the target angle of view is set as it is as the size of the angle of view AV of the next composition NCP.
- the composition transition determining unit 17 An angle of view larger than the target angle of view is set as the size of the angle of view AV of the next composition NCP.
- the composition transition determining unit 17 prohibits selection of the instantaneous composition transition mode for a predetermined period after the instantaneous composition transition mode is selected.
- the composition transition determining unit 17 selects a smooth transition instead of an instantaneous transition when a certain period of time has not elapsed since the previous instantaneous transition.
- FIG. 18 is a diagram showing an example in which an instantaneous transition is first performed and then a smooth transition is performed.
- FIG. 19 is a diagram illustrating an example of a transition mode determination flow.
- the composition CP transition is performed using the instantaneous counter and smooth counter cleared in the initialization (step SA1) of FIG.
- the instantaneous counter indicates the number of frames after the transition in the instantaneous composition transition mode.
- the smooth counter indicates the number of frames after transition in the smooth composition transition mode.
- the composition transition determining unit 17 increments the instantaneous counter and the smooth counter by 1 for each frame (steps SG1 to SG2). In the instantaneous composition transition mode, the composition transition determination unit 17 clears only the instantaneous counter (step SG8). In the smooth composition transition mode, the composition transition determining unit 17 clears only the smooth counter (step SG10).
- the composition transition determination unit 17 selects the smooth composition transition mode (step SG9). Even if the composition CP changes due to a change in the orientation of the subject SU, if the smooth counter is equal to or greater than the threshold value or the instantaneous counter is equal to or less than the threshold value, the composition transition determination unit 17 selects the smooth composition transition mode instead of the instantaneous composition transition mode. Select a mode (step SG9).
- the composition transition determining unit 17 selects the instantaneous composition transition mode (step SG7 ).
- the composition transition determining unit 17 determines that the rate of change of the angle of view AV of the next frame calculated from the target angle of view (the rate of change of the angle of view AV at the time of instantaneous transition) is the rate of change of the angle of view AV of the current frame (the rate of change of the angle of view AV immediately before the instantaneous transition). change rate of the angle of view AV), the magnification of the angle of view AV is increased, and the instantaneous composition CP is changed.
- Video clipping section Based on the angle-of-view information acquired from the composition transition determination unit 17, the video clipping unit 18 clips a video region corresponding to the angle of view AV of the next composition NCP from the captured video CV.
- the video clipping unit 18 outputs the clipped video of the video region as a framing video OV.
- the video processing device 10 has a target composition calculation unit 16 and a composition transition determination unit 17 .
- the target composition calculation unit 16 calculates the target composition TCP based on the state of the subject SU.
- the composition transition determining unit 17 switches the transition mode to the target composition TCP between a smooth composition transition mode and an instant composition transition mode based on the change in the state of the subject SU.
- the smooth composition transition mode is a transition mode in which the composition CP gradually transitions toward the target composition TCP.
- the instantaneous composition transition mode is a transition mode in which the composition CP instantaneously transitions toward the target composition TCP.
- the processing of the video processing device 10 is executed by a computer.
- the program of the present embodiment causes a computer to implement the processing of the video processing device 10 .
- the state of the subject SU includes a state related to at least one of the moving direction of the subject SU, the orientation of the subject SU, and the size of the subject SU.
- the state of change in the state of the subject SU includes at least one of the speed at which the moving direction is switched, the speed at which the orientation of the subject SU is switched, and the change in the size of the subject SU.
- the composition transition determining unit 17 determines the rate of change in the magnitude of the angle of view AV immediately after switching the transition mode from the smooth composition transition mode to the instant composition transition mode as the rate of change in the magnitude of the angle of view AV immediately before switching the transition mode. be different from
- the composition transition determining unit 17 switches the transition mode to the instantaneous composition transition mode when a new target composition TCP is determined while the composition CP is transitioning in the smooth composition transition mode.
- the composition transition determining unit 17 prohibits selection of the instantaneous composition transition mode for a predetermined period after the instantaneous composition transition mode is selected.
- the state of the subject SU is obtained by image analysis of the captured video CV of the subject SU.
- the state of the subject SU can be easily obtained without using any sensor other than the camera.
- FIG. 20 is a diagram showing the imaging system CS2 of the second embodiment.
- the difference between this embodiment and the first embodiment is that the state of the subject SU is obtained based on the sensor position information POI of the position sensor SE attached to the subject SU.
- the following description will focus on differences from the first embodiment.
- the sensor position information POI includes information on the position of the body part BP of the subject SU to which the position sensor SE is attached.
- the video processing device 50 acquires the sensor position information POI via the sensor signal receiver RC.
- the video processing device 50 performs calibration with the camera 20 and transforms the position of the body part BP into a position in the camera coordinate system.
- arithmetic processing using the sensor position information POI replaces the processing of the orientation acquisition unit 11 and the subject tracking unit 12 (or both). Therefore, the state of the subject SU can be accurately detected without performing complicated image analysis.
- FIG. 21 is a diagram showing the imaging system CS3 of the third embodiment.
- a plurality of cameras 20 are installed at the shooting site.
- a plurality of cameras 20 photograph the subject SU from different viewpoints.
- the image processing device 60 is installed for each camera 20 .
- the configuration of the video processing device 60 is the same as that of the video processing device 10 of the first embodiment.
- individual cameras 20 are distinguished by numbers attached after the reference numerals.
- Equipment and information corresponding to each camera 20 are given the same numbers as those of the cameras 20 after the reference numerals. This number matches, for example, the video number assigned to each shot video CV.
- the video processing device 60 performs image analysis on the captured video CV to generate angle-of-view information CPI and subject information SUI.
- the angle-of-view information CPI indicates the position and size of the angle-of-view AV (video area cut out from the shot video CV) corresponding to the composition NCP of the next frame.
- the subject information SUI indicates the state of the subject SU (moving direction of the subject SU, direction of the subject SU, size of the subject SU).
- the subject information SUI is generated by the target composition calculator 16 based on the determination results of the orientation determiner 13 , size determiner 14 and motion determiner 15 .
- the imaging system CS3 has a video switcher 70.
- Video processing device 60 outputs angle-of-view information CPI and subject information SUI to video switcher 70 .
- the video switcher 70 switches the captured video CV to be framed based on the subject information SUI acquired from each video processing device 60 .
- the video switcher 70 outputs a framing video OV cut out from the selected captured video CV.
- the framing video OV is cut out from the captured video CV based on the angle of view information CPI.
- the clipping of the framing video OV may be performed by the video processing device 60 or may be performed by the video switcher 70 .
- the video switcher 70 cuts out the framing video OV, the video clipping section 18 of the video processing device 60 is not used.
- FIG. 22 is a diagram showing an example of the functional configuration of the video switcher 70. As shown in FIG. 22
- the video switcher 70 has a selected video determination unit 71 and a video switching unit 72 .
- the selected image determination unit 71 acquires subject information SUI and angle-of-view information CPI from each image processing device 60 .
- the selected image determination unit 71 selects one captured image CV according to the state of the subject SU from the captured images CV of the plurality of subjects SU based on the obtained plurality of pieces of subject information SUI.
- the image switching unit 72 acquires angle of view information CPI corresponding to the selected captured image CV from the selected image determination unit 71 .
- the video switching unit 72 cuts out a video region corresponding to the angle-of-view information CPI from the selected shot video CV.
- the image switching unit 72 outputs the image area cut out from the selected captured image CV as the framing image OV.
- the selected video determination unit 71 selects a shot video CV closer to the front (or a preset direction), for example, using the orientation of the subject SU obtained from the subject information SUI.
- the moving direction of the subject SU obtained from the subject information SUI is used to select the captured image CV of the camera 20 that is close to the approaching direction of the subject SU.
- FIG. 23 is a diagram showing an example of a flow for selecting a shot video CV.
- the user sets the preferred orientation of the subject SU (step SH1). By default, for example, front is set as the preferred orientation.
- the selected video determination unit 71 acquires subject information SUI and angle-of-view information CPI synchronized with the shot video CV from a plurality of video processing devices 60 .
- the selected video determination unit 71 detects the orientation of the subject SU from the subject information SUI for each shot video CV.
- the selected video determination unit 71 obtains the degree of similarity between the orientation of the subject SU and the preferred orientation for each shot video CV (steps SH2 to SH3).
- the orientation similarity can be calculated using, for example, the distance d between the nose NS and the body axis AX. Since the distance d differs depending on whether the face is oriented sideways or obliquely, it is possible to set the orientation more precisely than left, right, or front.
- the selected video determination unit 71 detects the shot video CV that indicates the orientation of the subject SU with the highest degree of similarity.
- the selected video determination unit 71 outputs the video number of the detected captured video CV and the field angle information CPI corresponding to this captured video CV to the video switching unit 72 (step SH4).
- the image switching unit 72 cuts out the captured image CV indicated by the acquired image number at an appropriate angle of view AV based on the angle of view information CPI.
- the image switching unit 72 outputs the clipped image area as the framing image OV.
- the switching of the shot video CV is instantaneous.
- selection of the instantaneous composition transition mode is prohibited for a predetermined period after the instant composition transition mode is selected. This is to reduce discomfort caused by frequent instantaneous transitions of the composition CP.
- the selected video determination unit 71 prohibits selection of another captured video CV for a predetermined period after one captured video CV is selected. For example, the selected video determination unit 71 performs the next switching after several seconds from the previous switching. Such control becomes possible by providing a counter for measuring the switching interval in the flow of FIG.
- the selected video determination unit 71 determines the rate of change in the magnitude of the angle of view AV immediately after another captured video CV is selected as the change in the magnitude of the angle of view AV immediately before the other captured video CV is selected. make the rate different. As a result, an instantaneous change in composition CP is less likely to be recognized as image disturbance.
- the selected video determination unit 71 selects one captured video CV according to the state of the subject SU from a plurality of captured video CVs of the subject SU captured from different viewpoints.
- the image switching unit 72 outputs the image area cut out from the selected captured image CV as the framing image OV.
- the viewpoint (camera 20) is selected based on the state of the subject SU.
- a video with a more appropriate composition CP can be obtained.
- FIG. 26 is a diagram showing an example of the functional configuration of the video processing device 80. As shown in FIG.
- the target composition TCP When calculating the target composition TCP, it may be desired to consider not only the subject SU but also the specific object OB at the same time. For example, in a scene in which a new product is advertised as shown in FIG. 24, it is desirable to determine the composition CP including not only the speaker (subject SU) but also the product (object OB).
- the speaker and the product are the basis for calculating the composition CPB.
- a region containing the speaker and the product is extracted as the subject region, and the center line of the subject region is placed in the center of the screen.
- this composition CPB a well-balanced image can be obtained because the viewer's attention area is arranged in the center of the screen.
- composition CP including not only the main speaker (subject SU) but also the monitor (object OB) showing the teleconference participants.
- a video processing device 80 as shown in FIG. 26, which is obtained by adding an object recognition unit 81 to the video processing device 10 of FIG.
- the object recognition unit 81 detects the subject SU and the object OB that serves as a basis for calculating the target composition TCP from the captured video CV of the subject SU.
- the target composition TCP is calculated based on the object recognition result as well. Therefore, an image with an appropriate composition CP including the target object OB can be obtained.
- FIG. 27 is a diagram for explaining the outline of framing according to the fifth embodiment.
- FIG. 28 is a diagram showing an example of the functional configuration of the video processing device 90. As shown in FIG.
- camera 20 a fixed camera (overhead camera) with a fixed field of view that covers the entire range in which the subject SU moves around is used.
- camera 20 is not limited to a fixed camera.
- the PTZ camera 22 can control the orientation (pan, tilt) and zoom of the lens, it is possible to obtain a higher quality image than a fixed camera.
- FIG. 27 shows the difference in field of view between the fixed camera 21 and the PTZ camera 22 when photographing the subject SU.
- the moving range of the subject SU lecturer
- a range including all of it must be included in the field of view FV1 to output the image of the subject SU during the lecture.
- the PTZ camera 22 can rotate the lens, so it is possible to photograph the subject SU with the narrowed field of view FV2 and the resolution.
- the image A on the right side of FIG. 27 shows the image captured by the fixed camera 21.
- a video area (angle of view AV) near the subject SU is clipped out of a wide range of video, and much of the shooting resolution is not used for output.
- Video B is a video captured by the PTZ camera 22 in consideration of instantaneous transitions. Since the PTZ camera 22 rotates the lens by mechanical control, the composition CP cannot be changed instantaneously. Therefore, taking into account the image range to be framed after the instantaneous transition, the field of view FV2 is photographed with a larger margin than the target composition TCP. If the field of view FV2 is set so as to include the target composition TCP (including the enlargement of the angle of view AV at the time of instantaneous transition), even if the subject SU faces in the opposite direction to the image B, as in the image C, the instantaneous composition CP transition becomes possible.
- the shooting resolution of the PTZ camera 22 is utilized to achieve higher quality framing than the fixed camera 21.
- FIG. 28 shows a video processing device 90 to which a camera control section 91 for controlling the PTZ camera 22 is added.
- the camera control unit 91 controls the field of view (pan, tilt, zoom) of the PTZ camera 22 that captures the subject SU based on the state of the subject SU (moving direction of the subject SU, direction of the subject SU, size of the subject SU). do.
- the camera control unit 91 sets the field of view of the PTZ camera 22 so that there is a spatial margin on the right side of the subject SU.
- the camera control unit 91 sets the field of view of the PTZ camera 22 so that a similar spatial margin is generated on the left side of the subject SU so that the composition CP for leftward can be instantly changed.
- FIG. 29 is a diagram showing an example of the control flow of the PTZ camera 22. As shown in FIG.
- the camera control unit 91 outputs a control signal for the PTZ camera 22 based on the position of the subject SU's body axis AX and the assumed maximum field of view. For example, the camera control unit 91 calculates the field of view when the body axis AX is placed in the left 1/3 of the screen (step SJ1), and calculates the field of view when the body axis AX is placed in the right 1/3 of the screen. is calculated (step SJ2). The camera control unit 91 calculates the margin taking into account the delay in the turning motion of the PTZ camera 22 (step SJ3).
- the camera control unit 91 calculates a field of view that includes the two fields of view calculated in steps SJ1 and SJ2, adds the margins calculated in step SJ3 to this, and calculates the range of the field of view captured by the PTZ camera 22. (step SJ4).
- the camera control unit 91 controls pan, tilt and zoom of the PTZ camera so as to obtain the calculated field of view (step SJ5).
- the camera control unit 91 sets the pan and tilt so that the body axis AX starting from the top of the head is placed in the center of the screen, and sets the zoom so that the size of the field of view obtained is obtained. As a result, an image corresponding to image B in FIG. 27 is acquired.
- the camera control unit 91 appropriately controls the field of view of the PTZ camera 22 . Therefore, a high-resolution video image with an appropriate composition can be obtained.
- the transition mode is switched to the instantaneous composition transition mode when a new target composition TCP is determined while the composition CP is transitioning in the smooth composition transition mode.
- the timing of switching the transition mode to the instantaneous composition transition mode is not limited to this.
- the composition transition determining unit 17 switches the transition mode to the instantaneous composition transition mode when the state of the subject SU greatly changes beyond the permissible standard while the composition CP is transitioning in the smooth composition transition mode. be able to.
- the allowable standard means the standard regarding the amount of change in the state that requires the target composition TCP to be changed. If the state of the subject SU changes significantly beyond the permissible standard, it is determined that the target composition TCP cannot be maintained.
- the composition transition determining unit 17 switches the composition CP in the instant composition transition mode.
- the sense of discomfort caused by the discrepancy between the state of the subject SU and the composition CP is reduced.
- the state of the subject SU changes significantly, it takes time for the change in composition CP to catch up with the state of the subject SU in a smooth transition of the composition CP. If the time period during which the state of the subject SU and the composition CP are different from each other becomes longer, the viewer feels uncomfortable. If the composition CP is instantaneously changed to the target composition TCP when the state of the subject SU changes significantly, such discomfort is less likely to occur.
- the framing method of the fifth embodiment can also be applied to shooting with a hand-held camera.
- the hand-held camera is similar to the PTZ camera 22 in that the field of view can be adjusted by changing the orientation of the lens. Therefore, like the PTZ camera 22, the camera control unit 91 can control the field of view of the camera 20 that captures the subject SU based on the state of the subject SU. For example, the camera control unit 91 automatically controls the zoom of the hand-held camera so that the image can be captured in a field of view that can be reversed in orientation or enlarged or reduced in composition.
- the camera control unit 91 can also notify the user who shoots the subject of recommendation information based on the difference between the field of view of the current frame and the target field of view. For example, the camera control unit 91 presents recommendation information regarding changes in the panning direction, zoom ratio, etc. within the finder in order to bring the current field of view of the hand-held camera closer to the target field of view. Accordingly, it is possible to prompt the user to change the field of view.
- the framing method of the third embodiment can be combined with the control method of the PTZ camera 22 described in the fifth embodiment.
- the next angle of view AV is assumed based on the angle of view AV of the framing video OV currently being output, and the field of view of the PTZ camera 22 not selected is adjusted to the field of view corresponding to the next angle of view AV.
- the camera control unit 91 calculates the field of view of the PTZ camera 22 that captured the non-selected captured video CV based on the angle of view AV of the video area cut out from the selected captured video CV. As a result, it is possible to smoothly deal with changes in the orientation and movement direction of the subject.
- FIG. 30 is a diagram showing a hardware configuration example of the imaging system CS.
- the imaging system CS is implemented by, for example, a computer 1000 configured as shown in FIG.
- the computer 1000 has a CPU 1100 , a RAM 1200 , a ROM (Read Only Memory) 1300 , a HDD (Hard Disk Drive) 1400 , a communication interface 1500 and an input/output interface 1600 .
- Each part of computer 1000 is connected by bus 1050 .
- the CPU 1100 operates based on programs stored in the ROM 1300 or HDD 1400 and controls each section. For example, the CPU 1100 loads programs stored in the ROM 1300 or HDD 1400 into the RAM 1200 and executes processes corresponding to various programs.
- the ROM 1300 stores a boot program such as BIOS (Basic Input Output System) executed by the CPU 1100 when the computer 1000 is started, and programs dependent on the hardware of the computer 1000.
- BIOS Basic Input Output System
- the HDD 1400 is a computer-readable recording medium that non-temporarily records programs executed by the CPU 1100 and data used by such programs.
- HDD 1400 is a recording medium that records an information processing program according to the present disclosure, which is an example of program data 1450 .
- a communication interface 1500 is an interface for connecting the computer 1000 to an external network 1550 (for example, the Internet).
- CPU 1100 receives data from another device via communication interface 1500, and transmits data generated by CPU 1100 to another device.
- the input/output interface 1600 is an interface for connecting the input/output device 1650 and the computer 1000 .
- the CPU 1100 receives data from input devices such as a keyboard and mouse via the input/output interface 1600 .
- the CPU 1100 also transmits data to an output device such as a display, speaker, or printer via the input/output interface 1600 .
- the input/output interface 1600 may function as a media interface for reading a program or the like recorded on a predetermined recording medium.
- Media include, for example, optical recording media such as DVDs (Digital Versatile Discs), magneto-optical recording media such as MOs (Magneto-Optical disks), tape media, magnetic recording media, semiconductor memories, and the like.
- the CPU 1100 of the computer 1000 implements the functions of each unit of the video processing device by executing programs loaded on the RAM 1200 .
- the HDD 1400 stores programs according to the present disclosure and data in the video processing device and the recording device 40 .
- CPU 1100 reads and executes program data 1450 from HDD 1400 , as another example, these programs may be obtained from another device via external network 1550 .
- a target composition calculation unit that calculates a target composition based on the state of the subject;
- a smooth composition transition mode in which the composition gradually transitions toward the target composition and a smooth composition transition mode in which the composition transitions instantaneously toward the target composition are transition modes to the target composition based on the status of the state change.
- composition transition determining unit that switches between an instantaneous composition transition mode;
- Information processing device having (2) the state of the subject includes a state related to at least one of a moving direction of the subject, an orientation of the subject, and a size of the subject;
- the status of the state change includes a status related to at least one of the speed at which the direction of movement is switched, the speed at which the direction is switched, and a change in the size of the subject.
- the information processing apparatus according to (1) above.
- the composition transition determining unit determines a change rate of the angle of view immediately after switching the transition mode from the smooth composition transition mode to the instantaneous composition transition mode as the rate of change of the angle of view immediately before switching the transition mode. to differ from the rate of change, The information processing apparatus according to (1) or (2) above.
- the composition transition determining unit switches the transition mode to the instantaneous composition transition mode when the new target composition is determined while the composition is transitioning in the smooth composition transition mode.
- the information processing apparatus according to any one of (1) to (3) above.
- the composition transition determining unit switches the transition mode to the instantaneous composition transition mode when the state greatly changes beyond an allowable standard while the composition is transitioning in the smooth composition transition mode.
- the information processing apparatus according to any one of (1) to (3) above.
- the composition transition determining unit prohibits selection of the instantaneous composition transition mode for a predetermined period after the instantaneous composition transition mode is selected.
- the information processing apparatus according to any one of (1) to (5) above.
- the state of the subject is obtained by image analysis of the captured video of the subject, The information processing apparatus according to any one of (1) to (6) above.
- the state of the subject is obtained based on sensor position information of a position sensor attached to the subject;
- the information processing apparatus according to any one of (1) to (6) above. (9) a selected image determination unit that selects one captured image according to the state of the subject from a plurality of captured images of the subject that are captured from different viewpoints; a video switching unit that outputs a video region clipped from the selected captured video as a framing video;
- the selected image determination unit prohibits selection of the other captured images for a predetermined period after one of the captured images is selected.
- the selected image determination unit makes the rate of change in the size of the angle of view immediately after the other photographed image is selected different from the rate of change in the size of the angle of view immediately before the other photographed image is selected.
- An object recognition unit that detects an object that serves as a basis for calculating the target composition together with the subject from the captured image of the subject, The information processing apparatus according to any one of (1) to (11) above.
- a camera control unit that controls a field of view of a camera that captures the subject based on the state of the subject, The information processing apparatus according to any one of (1) to (12) above.
- the camera control unit presents recommendation information to the user who shoots the subject based on the difference between the field of view of the current frame and the target field of view.
- the information processing device according to (13) above.
- a camera control unit that calculates the field of view of the camera that captured the unselected captured image based on the angle of view of the image area cut out from the selected captured image, The information processing apparatus according to any one of (9) to (11) above.
- (16) Calculate the target composition based on the subject's condition, A smooth composition transition mode in which the composition gradually transitions toward the target composition and a smooth composition transition mode in which the composition transitions instantaneously toward the target composition are transition modes to the target composition based on the status of the state change.
- switch between instantaneous composition transition mode and A computer-implemented information processing method comprising: (17) Calculate the target composition based on the subject's condition, A smooth composition transition mode in which the composition gradually transitions toward the target composition and a smooth composition transition mode in which the composition transitions instantaneously toward the target composition are transition modes to the target composition based on the status of the state change. switch between instantaneous composition transition mode and A program that makes a computer do something.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- General Physics & Mathematics (AREA)
- Pathology (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Dentistry (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- Medical Informatics (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Surgery (AREA)
- Animal Behavior & Ethology (AREA)
- General Health & Medical Sciences (AREA)
- Public Health (AREA)
- Veterinary Medicine (AREA)
- Physiology (AREA)
- Image Analysis (AREA)
Abstract
Description
[1.第1実施形態]
[1-1.撮影システムの構成]
[1-2.映像処理方法]
[1-2-1.姿勢取得部]
[1-2-2.被写体追尾部]
[1-2-3.向き判定部]
[1-2-4.大きさ判定部]
[1-2-5.動き判定部]
[1-2-6.目標構図算出部]
[1-2-7.構図遷移決定部]
[1-2-8.映像切り出し部]
[1-3.効果]
[2.第2実施形態]
[3.第3実施形態]
[3-1.撮影システムの構成]
[3-2.映像スイッチャの構成]
[3-3.映像処理方法]
[3-4.効果]
[4.第4実施形態]
[5.第5実施形態]
[5-1.映像処理装置の機能構成]
[5-2.映像処理方法]
[5-3.効果]
[6.変形例1]
[7.変形例2]
[8.変形例3]
[9.撮影システムのハードウェア構成例]
[1-1.撮影システムの構成]
以下、図面を参照しながら、撮影システムCSの構成の一例を説明する。図1は、第1実施形態の撮影システムCS1を示す図である。
図2は、映像処理装置10の機能構成を示すブロック図である。
図4は、姿勢取得部11で取得される姿勢情報の一例を示す図である。
被写体追尾部12は、被写体SUのフレーム間での移動を検出する。これには、ピクセル毎のマッチングを行うオプティカルフロー処理のような画像を直接入力とする移動量算出処理や、姿勢取得部11で得た身体部位BPのフレーム間でのマッチング処理といった一度画像を処理して得られたデータによる移動量算出処理を用いることができる。ロバスト性を上げるために画像と身体部位BPのマッチングを両方用いて移動量算出を行うこともできる。
図5および図6は、被写体SUの向きの判定方法を説明する図である。
図9は、被写体SUの大きさの判定方法を説明する図である。
動き判定部15は、被写体追尾部12で得た被写体追尾結果から前フレームに対して現フレームで被写体SUが移動した向きと移動量を算出する。動き判定部15は、移動方およびと移動量の算出結果に基づいて、被写体SUが動いたかどうかを判定する。
図12は、目標構図の算出方法を説明する図である。
図15ないし図18は、構図CPの遷移方法を説明する図である。現在の構図CCP、目標構図TCPおよび次の構図NCPは、切り出しの対象となる映像領域(画角AV)として示されている。現在の構図CCPは、現フレームのフレーミング映像OVとして切り出されている映像領域の構図CPを意味する。次の構図NCPは、次フレームのフレーミング映像OVとして切り出されるべき映像領域の構図CPを意味する。
映像切り出し部18は、構図遷移決定部17から取得した画角情報に基づいて、撮影映像CVから次の構図NCPの画角AVに対応した映像領域を切り出す。映像切り出し部18は、切り出された映像領域の映像をフレーミング映像OVとして出力する。
映像処理装置10は、目標構図算出部16と構図遷移決定部17とを有する。目標構図算出部16は、被写体SUの状態に基づいて目標構図TCPを算出する。構図遷移決定部17は、被写体SUの状態の変化の状況に基づいて、目標構図TCPへの遷移モードをスムーズ構図遷移モードと瞬時構図遷移モードとの間で切り替える。スムーズ構図遷移モードは、目標構図TCPに向けて徐々に構図CPが遷移する遷移モードである。瞬時構図遷移モードは、目標構図TCPに向けて瞬時に構図CPが遷移する遷移モードである。本実施形態の情報処理方法は、映像処理装置10の処理がコンピュータにより実行される。本実施形態のプログラムは、映像処理装置10の処理をコンピュータに実現させる。
図20は、第2実施形態の撮影システムCS2を示す図である。
[3-1.撮影システムの構成]
図21は、第3実施形態の撮影システムCS3を示す図である。
図22は、映像スイッチャ70の機能構成の一例を示す図である。
図23は、撮影映像CVの選択フローの一例を示す図である。
本実施形態では、選択映像判定部71は、異なる視点から撮影された複数の被写体SUの撮影映像CVから、被写体SUの状態に応じた1つの撮影映像CVを選択する。映像切り替え部72は、選択された撮影映像CVから切り出された映像領域をフレーミング映像OVとして出力する。
図24および図25は、第4実施形態のフレーミングの概要を説明する図である。図26は、映像処理装置80の機能構成の一例を示す図である。
[5-1.映像処理装置の機能構成]
図27は、第5実施形態のフレーミングの概要を説明する図である。図28は、映像処理装置90の機能構成の一例を示す図である。
図29は、PTZカメラ22の制御フローの一例を示す図である。
本実施形態では、カメラ制御部91により、PTZカメラ22の視野が適切に制御される。そのため、適切な構図を有する解像度の高い映像が得られる。
第1実施形態では、スムーズ構図遷移モードで構図CPが遷移している最中に新たな目標構図TCPが決定された場合に遷移モードが瞬時構図遷移モードに切り替えられた。しかし、遷移モードを瞬時構図遷移モードに切り替えるタイミングはこれに限定されない。例えば、構図遷移決定部17は、スムーズ構図遷移モードで構図CPが遷移している最中に被写体SUの状態が許容基準を超えて大きく変化した場合には、遷移モードを瞬時構図遷移モードに切り替えることができる。
第5実施形態のフレーミングの手法は、手持ちカメラで撮影を行う場合にも適用できる。手持ちカメラは、レンズの向きによって視野が調整できる点でPTZカメラ22と共通する。そのため、カメラ制御部91は、PTZカメラ22と同様に、被写体SUを撮影するカメラ20の視野を被写体SUの状態に基づいて制御することができる。例えば、カメラ制御部91は、向き反転や構図の拡縮に対応可能な視野で撮影できるように、手持ちカメラのズームの自動制御を行う。
第3実施形態のフレーミングの手法に、第5実施形態で説明したPTZカメラ22の制御手法を組み合わせることができる。この際、現在出力しているフレーミング映像OVの画角AVをもとに次の画角AVを想定し、選択されていないPTZカメラ22の視野を、次の画角AVに対応した視野に合わせこんでおく。例えば、カメラ制御部91は、選択されていない撮影映像CVを撮影したPTZカメラ22の視野を、選択されている撮影映像CVから切り出された映像領域の画角AVに基づいて算出する。これにより、被写体の向きや移動方向が変化したときにスムーズに対応することができる。
図30は、撮影システムCSのハードウェア構成例を示す図である。
なお、本技術は以下のような構成も取ることができる。
(1)
被写体の状態に基づいて目標構図を算出する目標構図算出部と、
前記状態の変化の状況に基づいて、前記目標構図への遷移モードを、前記目標構図に向けて徐々に構図が遷移するスムーズ構図遷移モードと、前記目標構図に向けて瞬時に前記構図が遷移する瞬時構図遷移モードと、の間で切り替える構図遷移決定部と、
を有する情報処理装置。
(2)
前記被写体の状態は、前記被写体の移動方向、前記被写体の向き、および、前記被写体の大きさ、のうちの少なくとも1つに関する状態を含み、
前記状態の変化の状況は、前記移動方向が切り替わる速さ、前記向きが切り替わる速さ、および、前記被写体の大きさの変化、のうちの少なくとも1つに関する状況を含む、
上記(1)に記載の情報処理装置。
(3)
前記構図遷移決定部は、前記遷移モードを前記スムーズ構図遷移モードから前記瞬時構図遷移モードに切り替えた直後の画角の大きさの変化率を、前記遷移モードを切り替える直前の画角の大きさの変化率と異ならせる、
上記(1)または(2)に記載の情報処理装置。
(4)
前記構図遷移決定部は、前記スムーズ構図遷移モードで前記構図が遷移している最中に新たな前記目標構図が決定された場合には、前記遷移モードを前記瞬時構図遷移モードに切り替える、
上記(1)ないし(3)のいずれか1つに記載の情報処理装置。
(5)
前記構図遷移決定部は、前記スムーズ構図遷移モードで前記構図が遷移している最中に前記状態が許容基準を超えて大きく変化した場合には、前記遷移モードを前記瞬時構図遷移モードに切り替える、
上記(1)ないし(3)のいずれか1つに記載の情報処理装置。
(6)
前記構図遷移決定部は、前記瞬時構図遷移モードが選択されてから所定の期間は前記瞬時構図遷移モードの選択を禁止する、
上記(1)ないし(5)のいずれか1つに記載の情報処理装置。
(7)
前記被写体の状態は、前記被写体の撮影映像を画像解析することにより取得される、
上記(1)ないし(6)のいずれか1つに記載の情報処理装置。
(8)
前記被写体の状態は、前記被写体に取り付けられた位置センサのセンサ位置情報に基づいて取得される、
上記(1)ないし(6)のいずれか1つに記載の情報処理装置。
(9)
異なる視点から撮影された複数の前記被写体の撮影映像から、前記被写体の状態に応じた1つの撮影映像を選択する選択映像判定部と、
選択された前記撮影映像から切り出された映像領域をフレーミング映像として出力する映像切り替え部と、
を有する上記(1)ないし(8)のいずれか1つに記載の情報処理装置。
(10)
前記選択映像判定部は、1つの前記撮影映像が選択されてから所定の期間は他の前記撮影映像の選択を禁止する、
上記(9)に記載の情報処理装置。
(11)
前記選択映像判定部は、他の前記撮影映像が選択された直後の画角の大きさの変化率を、他の前記撮影映像が選択される直前の画角の大きさの変化率と異ならせる、
上記(10)に記載の情報処理装置。
(12)
前記被写体の撮影映像から、前記被写体とともに前記目標構図の算出の基礎となる物体を検出する物体認識部を有する、
上記(1)ないし(11)のいずれか1つに記載の情報処理装置。
(13)
前記被写体を撮影するカメラの視野を前記被写体の状態に基づいて制御するカメラ制御部を有する、
上記(1)ないし(12)のいずれか1つに記載の情報処理装置。
(14)
前記カメラ制御部は、現フレームの視野と目標となる視野との差異に基づいて、前記被写体を撮影するユーザにレコメンド情報を提示する、
上記(13)に記載の情報処理装置。
(15)
選択されていない前記撮影映像を撮影したカメラの視野を、選択されている前記撮影映像から切り出された前記映像領域の画角に基づいて算出するカメラ制御部を有する、
上記(9)ないし(11)のいずれか1つに記載の情報処理装置。
(16)
被写体の状態に基づいて目標構図を算出し、
前記状態の変化の状況に基づいて、前記目標構図への遷移モードを、前記目標構図に向けて徐々に構図が遷移するスムーズ構図遷移モードと、前記目標構図に向けて瞬時に前記構図が遷移する瞬時構図遷移モードと、の間で切り替える、
ことを有する、コンピュータにより実行される情報処理方法。
(17)
被写体の状態に基づいて目標構図を算出し、
前記状態の変化の状況に基づいて、前記目標構図への遷移モードを、前記目標構図に向けて徐々に構図が遷移するスムーズ構図遷移モードと、前記目標構図に向けて瞬時に前記構図が遷移する瞬時構図遷移モードと、の間で切り替える、
ことをコンピュータに実現させるプログラム。
20 カメラ
16 目標構図算出部
17 構図遷移決定部
71 選択映像判定部
72 映像切り替え部
81 物体認識部
91 カメラ制御部
1000 コンピュータ
1450 プログラムデータ(プログラム)
AV 画角
CP 構図
CV 撮影映像
OB 物体
OV フレーミング映像
POI センサ位置情報
SE 位置センサ
SU 被写体
TCP 目標構図
Claims (17)
- 被写体の状態に基づいて目標構図を算出する目標構図算出部と、
前記状態の変化の状況に基づいて、前記目標構図への遷移モードを、前記目標構図に向けて徐々に構図が遷移するスムーズ構図遷移モードと、前記目標構図に向けて瞬時に前記構図が遷移する瞬時構図遷移モードと、の間で切り替える構図遷移決定部と、
を有する情報処理装置。 - 前記被写体の状態は、前記被写体の移動方向、前記被写体の向き、および、前記被写体の大きさ、のうちの少なくとも1つに関する状態を含み、
前記状態の変化の状況は、前記移動方向が切り替わる速さ、前記向きが切り替わる速さ、および、前記被写体の大きさの変化、のうちの少なくとも1つに関する状況を含む、
請求項1に記載の情報処理装置。 - 前記構図遷移決定部は、前記遷移モードを前記スムーズ構図遷移モードから前記瞬時構図遷移モードに切り替えた直後の画角の大きさの変化率を、前記遷移モードを切り替える直前の画角の大きさの変化率と異ならせる、
請求項1に記載の情報処理装置。 - 前記構図遷移決定部は、前記スムーズ構図遷移モードで前記構図が遷移している最中に新たな前記目標構図が決定された場合には、前記遷移モードを前記瞬時構図遷移モードに切り替える、
請求項1に記載の情報処理装置。 - 前記構図遷移決定部は、前記スムーズ構図遷移モードで前記構図が遷移している最中に前記状態が許容基準を超えて大きく変化した場合には、前記遷移モードを前記瞬時構図遷移モードに切り替える、
請求項1に記載の情報処理装置。 - 前記構図遷移決定部は、前記瞬時構図遷移モードが選択されてから所定の期間は前記瞬時構図遷移モードの選択を禁止する、
請求項1に記載の情報処理装置。 - 前記被写体の状態は、前記被写体の撮影映像を画像解析することにより取得される、
請求項1に記載の情報処理装置。 - 前記被写体の状態は、前記被写体に取り付けられた位置センサのセンサ位置情報に基づいて取得される、
請求項1に記載の情報処理装置。 - 異なる視点から撮影された複数の前記被写体の撮影映像から、前記被写体の状態に応じた1つの撮影映像を選択する選択映像判定部と、
選択された前記撮影映像から切り出された映像領域をフレーミング映像として出力する映像切り替え部と、
を有する請求項1に記載の情報処理装置。 - 前記選択映像判定部は、1つの前記撮影映像が選択されてから所定の期間は他の前記撮影映像の選択を禁止する、
請求項9に記載の情報処理装置。 - 前記選択映像判定部は、他の前記撮影映像が選択された直後の画角の大きさの変化率を、他の前記撮影映像が選択される直前の画角の大きさの変化率と異ならせる、
請求項10に記載の情報処理装置。 - 前記被写体の撮影映像から、前記被写体とともに前記目標構図の算出の基礎となる物体を検出する物体認識部を有する、
請求項1に記載の情報処理装置。 - 前記被写体を撮影するカメラの視野を前記被写体の状態に基づいて制御するカメラ制御部を有する、
請求項1に記載の情報処理装置。 - 前記カメラ制御部は、現フレームの視野と目標となる視野との差異に基づいて、前記被写体を撮影するユーザにレコメンド情報を提示する、
請求項13に記載の情報処理装置。 - 選択されていない前記撮影映像を撮影したカメラの視野を、選択されている前記撮影映像から切り出された前記映像領域の画角に基づいて算出するカメラ制御部を有する、
請求項9に記載の情報処理装置。 - 被写体の状態に基づいて目標構図を算出し、
前記状態の変化の状況に基づいて、前記目標構図への遷移モードを、前記目標構図に向けて徐々に構図が遷移するスムーズ構図遷移モードと、前記目標構図に向けて瞬時に前記構図が遷移する瞬時構図遷移モードと、の間で切り替える、
ことを有する、コンピュータにより実行される情報処理方法。 - 被写体の状態に基づいて目標構図を算出し、
前記状態の変化の状況に基づいて、前記目標構図への遷移モードを、前記目標構図に向けて徐々に構図が遷移するスムーズ構図遷移モードと、前記目標構図に向けて瞬時に前記構図が遷移する瞬時構図遷移モードと、の間で切り替える、
ことをコンピュータに実現させるプログラム。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US18/557,685 US20240223883A1 (en) | 2021-05-21 | 2022-01-28 | Information processing device, information processing method, and program |
JP2023522211A JPWO2022244314A1 (ja) | 2021-05-21 | 2022-01-28 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2021086376 | 2021-05-21 | ||
JP2021-086376 | 2021-05-21 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022244314A1 true WO2022244314A1 (ja) | 2022-11-24 |
Family
ID=84140455
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2022/003240 WO2022244314A1 (ja) | 2021-05-21 | 2022-01-28 | 情報処理装置、情報処理方法およびプログラム |
Country Status (3)
Country | Link |
---|---|
US (1) | US20240223883A1 (ja) |
JP (1) | JPWO2022244314A1 (ja) |
WO (1) | WO2022244314A1 (ja) |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2007067510A (ja) * | 2005-08-29 | 2007-03-15 | Toshiba Corp | 映像撮影システム |
JP2014007653A (ja) * | 2012-06-26 | 2014-01-16 | Jvc Kenwood Corp | 撮像装置、撮像方法、撮像システム及びプログラム |
-
2022
- 2022-01-28 JP JP2023522211A patent/JPWO2022244314A1/ja active Pending
- 2022-01-28 US US18/557,685 patent/US20240223883A1/en active Pending
- 2022-01-28 WO PCT/JP2022/003240 patent/WO2022244314A1/ja active Application Filing
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2007067510A (ja) * | 2005-08-29 | 2007-03-15 | Toshiba Corp | 映像撮影システム |
JP2014007653A (ja) * | 2012-06-26 | 2014-01-16 | Jvc Kenwood Corp | 撮像装置、撮像方法、撮像システム及びプログラム |
Also Published As
Publication number | Publication date |
---|---|
US20240223883A1 (en) | 2024-07-04 |
JPWO2022244314A1 (ja) | 2022-11-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR100542370B1 (ko) | 보이지 않는 마커를 이용하는 비전기반 증강현실 시스템 | |
CN109960401B (zh) | 一种基于人脸追踪的动向投影方法、装置及其系统 | |
JP4389901B2 (ja) | スポーツ競技におけるカメラ自動制御システム、カメラ自動制御方法、カメラ自動制御装置、およびプログラム | |
JP2007129480A (ja) | 撮像装置 | |
WO2020162193A1 (ja) | 情報処理装置および方法、並びにプログラム | |
CN106873933B (zh) | 显示方法及电子设备 | |
JP6621565B2 (ja) | 表示制御装置、表示制御方法、及びプログラム | |
JP6294054B2 (ja) | 映像表示装置、映像提示方法及びプログラム | |
KR20110132260A (ko) | 모니터 기반 증강현실 시스템 | |
WO2019187732A1 (ja) | 情報処理装置、情報処理方法およびプログラム | |
JP2007251429A (ja) | 動画撮影装置及びズーム調整方法 | |
JP2020042206A (ja) | 画像表示制御装置および画像表示制御用プログラム | |
KR20110103692A (ko) | 사용자의 움직임을 반영한 파노라마 영상 재생 장치 및 파노라마 영상 재생 방법 | |
WO2022244314A1 (ja) | 情報処理装置、情報処理方法およびプログラム | |
WO2021200184A1 (ja) | 情報処理装置、情報処理方法、及び、プログラム | |
KR101810671B1 (ko) | 전방향 영상의 방향 정보를 생성하는 방법 및 이러한 방법을 수행하는 장치 | |
JP6602726B2 (ja) | 仮想環境生成装置、仮想環境生成方法、プログラム | |
US12051148B2 (en) | Image processing apparatus, method for controlling the same, and non-transitory computer-readable storage medium | |
JP2022016929A (ja) | 情報処理装置、情報処理方法、及びプログラム | |
KR20090078085A (ko) | 입체 영상 디스플레이 장치 및 그 입체 영상 디스플레이장치에서의 표시부 위치 조절 방법 | |
KR102178990B1 (ko) | 전방향 영상의 방향 정보를 생성하는 방법 및 이러한 방법을 수행하는 장치 | |
JP2008178089A (ja) | 出力装置、出力方法、及びプログラム | |
WO2020184188A1 (ja) | 画像処理装置、画像処理方法及び画像処理プログラム | |
JP7065708B2 (ja) | 録画再生装置及びプログラム | |
JP6356928B1 (ja) | 画像提供装置、画像提供方法、プログラム、ならびに、非一時的なコンピュータ読取可能な情報記録媒体 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 22804237 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2023522211 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 18557685 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 22804237 Country of ref document: EP Kind code of ref document: A1 |