WO2024084953A1 - Information processing device, information processing method, and program - Google Patents

Information processing device, information processing method, and program Download PDF

Info

Publication number
WO2024084953A1
WO2024084953A1 PCT/JP2023/036016 JP2023036016W WO2024084953A1 WO 2024084953 A1 WO2024084953 A1 WO 2024084953A1 JP 2023036016 W JP2023036016 W JP 2023036016W WO 2024084953 A1 WO2024084953 A1 WO 2024084953A1
Authority
WO
WIPO (PCT)
Prior art keywords
sound
information
sound collection
equipment
orientation
Prior art date
Application number
PCT/JP2023/036016
Other languages
French (fr)
Japanese (ja)
Inventor
隆太郎 渡邉
Original Assignee
ソニーグループ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ソニーグループ株式会社 filed Critical ソニーグループ株式会社
Publication of WO2024084953A1 publication Critical patent/WO2024084953A1/en

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64CAEROPLANES; HELICOPTERS
    • B64C13/00Control systems or transmitting systems for actuating flying-control surfaces, lift-increasing flaps, air brakes, or spoilers
    • B64C13/02Initiating means
    • B64C13/16Initiating means actuated automatically, e.g. responsive to gust detectors
    • B64C13/18Initiating means actuated automatically, e.g. responsive to gust detectors using automatic pilot
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/02Casings; Cabinets ; Supports therefor; Mountings therein
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R3/00Circuits for transducers, loudspeakers or microphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R5/00Stereophonic arrangements
    • H04R5/027Spatial or constructional arrangements of microphones, e.g. in dummy heads

Definitions

  • the present disclosure relates to an information processing device, an information processing method, and a program, and in particular to an information processing device, an information processing method, and a program that enable better audio pickup.
  • directional microphones are sometimes used to capture the voices of actors during filming of movies, dramas, and other video content.
  • the sound staff holding the boom must be careful not to capture themselves, the directional microphone, or the boom in the footage.
  • the positioning of the directional microphone and holding the boom were performed by the sound staff, but in the future, it is expected that this will be replaced by drones, robots, and the like.
  • cited reference 1 discloses an imaging system that determines the positions of multiple moving bodies that capture images of a subject, the moving bodies including a moving body having an imaging unit that captures images of the subject and a moving body having an illumination unit that illuminates the subject with light.
  • cited document 2 discloses an information processing device that acquires status information of an imaging unit possessed by a moving object, calculates the imageable area of the imaging unit based on the status information, and transmits information regarding the imageable area to the moving object.
  • the information processing device of one aspect of the present disclosure includes an information acquisition unit that acquires information including at least information indicating the position or orientation of the sound pickup equipment that picks up the sound of the target sound pickup, and that acquires information necessary to calculate an optimal placement of the sound pickup equipment, a calculation processing unit that calculates, based on the information by the information acquisition unit, an optimal placement of the sound pickup equipment for picking up the sound of the target sound pickup, and an operation control unit that controls the operation of the position or orientation of the sound pickup equipment so as to achieve the placement calculated by the calculation processing unit.
  • An information processing method or program includes at least information indicating the position or orientation of a pickup device that picks up a target sound, and includes acquiring information necessary to calculate an optimal placement of the pickup device, calculating an optimal placement of the pickup device for picking up the target sound based on the information, and controlling the position or orientation of the pickup device to achieve that placement.
  • information is acquired that includes at least information indicating the position or orientation of the pickup equipment that picks up the target sound, and the optimal placement of the pickup equipment is calculated based on that information, and the position or orientation of the pickup equipment is controlled to achieve that placement.
  • FIG. 1 is a diagram showing an example of the configuration of an embodiment of an imaging and sound collecting system to which the present technology is applied; 1 is a block diagram showing an example of the configuration of an information processing device; 11 is a flowchart illustrating a first processing example of an imaging and sound collection process. 13 is a flowchart illustrating a second processing example of the shooting and sound collection processing.
  • FIG. 13 is a diagram illustrating considerations, conditions, and additional input information.
  • FIG. 13 is a diagram illustrating an example in which a noise source is taken into consideration.
  • FIG. 13 is a diagram illustrating an example in which lighting equipment is taken into consideration.
  • FIG. 2 is a diagram illustrating a first example of collecting sounds from a plurality of sound collection targets.
  • FIG. 11 is a diagram illustrating a second example of collecting sounds from a plurality of sound collection targets.
  • FIG. 13 is a diagram illustrating a third example of collecting sounds from a plurality of sound collection targets.
  • FIG. 13 is a diagram illustrating a fourth example of collecting sounds from a plurality of sound collection targets.
  • FIG. 13 is a diagram showing an example of an arrangement screen.
  • FIG. 13 is a diagram showing an example of an arrangement screen.
  • 1A and 1B are diagrams illustrating an example configuration using a microphone cable. 1 is a block diagram showing an example of the configuration of an embodiment of a computer to which the present technology is applied.
  • FIG. 1 is a diagram showing an example of the configuration of an embodiment of an imaging and sound collecting system to which the present technology is applied.
  • the imaging and recording system 11 is configured with a imaging camera 12, a directional microphone 13, a moving device 14, and an information processing device 15.
  • the filming and recording system 11 can be used on location for movies, dramas, etc., and can use the directional microphone 13 to record sounds such as lines spoken by a specific actor who is being filmed by the filming camera 12 as the recording target 21.
  • the filming and recording system 11 can also use the information processing device 15 to control the operation of the mobile device 14 equipped with the directional microphone 13 so that the mobile device 14 is in the optimal position or orientation for recording the sound of the recording target 21.
  • the photographing camera 12 is, for example, photographing equipment that takes photographs with an angle of view that corresponds to the lens used.
  • Directional microphone 13 is, for example, a recording device with directional characteristics that mainly picks up sounds in a narrow range along the sound axis and does not pick up sounds outside that range.
  • the moving device 14 can be, for example, an unmanned aerial vehicle (a so-called drone) that can fly autonomously using multiple propellers, and moves the directional microphone 13 to a desired position or orientation according to the control of the information processing device 15.
  • an unmanned aerial vehicle a so-called drone
  • the moving device 14 can be, for example, an unmanned aerial vehicle (a so-called drone) that can fly autonomously using multiple propellers, and moves the directional microphone 13 to a desired position or orientation according to the control of the information processing device 15.
  • the information processing device 15 calculates the optimal position or orientation of the directional microphone 13 for picking up the sound of the collection target 21 (hereinafter referred to as the optimal arrangement of the directional microphone 13), and controls the operation of the moving device 14 so that the sound of the collection target 21 can be picked up in the optimal arrangement of the directional microphone 13.
  • the information processing device 15 can estimate the reflection range (the range surrounded by the dashed line in FIG. 1) that is the range that is reflected in the image captured by the capturing camera 12, based on the position, orientation, and angle of view of the capturing camera 12.
  • the information processing device 15 can also estimate the recording possible range (the range surrounded by the dashed line in FIG. 1) that is the range in which the direct sound of the sound collection target 21 can be recorded when the sound axis of the directional microphone 13 is directed toward the sound collection target 21, based on the position and orientation of the sound collection target 21.
  • the information processing device 15 calculates the position where the directional microphone 13 can get closest to the sound collection target 21, inside the range where the sound of the sound collection target 21 can be recorded and outside the range of the image capture camera 12, as the optimal position of the directional microphone 13 for picking up the sound of the sound collection target 21. Furthermore, the information processing device 15 calculates the direction in which the sound axis of the directional microphone 13 faces toward the sound collection target 21 at that position as the optimal orientation of the directional microphone 13 for picking up the sound of the sound collection target 21. Then, the information processing device 15 controls the operation of the moving device 14 so that the directional microphone 13 is positioned (position or orientation) optimally.
  • FIG. 2 is a block diagram showing an example configuration of the information processing device 15.
  • the information processing device 15 includes a user input unit 31, an operation control unit 32, an information acquisition unit 33, a positional relationship determination unit 34, a memory unit 35, and a calculation processing unit 36.
  • the user input unit 31 is used by the user of the filming and recording system 11 to input various information to the information processing device 15.
  • the user uses the user input unit 31 to input equipment information related to the equipment used in the filming and recording system 11, such as information indicating the angle of view of the filming camera 12, information indicating the shape and directional characteristics of the directional microphone 13, and information indicating the shape of the mobile device 14.
  • the user input unit 31 supplies this equipment information to the calculation processing unit 36.
  • the user can use the user input unit 31 to specify the directional microphone 13 to be used for collecting sound for each sound collection target 21, and the user input unit 31 supplies information indicating the directional microphone 13 specified for each sound collection target 21 to the calculation processing unit 36. Furthermore, the user can use the user input unit 31 to specify the position or orientation of the directional microphone 13 or adjust the position or orientation before shooting by the shooting camera 12 begins, and the user input unit 31 supplies information indicating the position or orientation of the directional microphone 13 specified or adjusted by the user to the operation control unit 32.
  • the operation control unit 32 controls the operation of the moving device 14 based on information indicating the position or orientation of the directional microphone 13 supplied from the user input unit 31, and moves the directional microphone 13 to a position or orientation according to the information.
  • the operation control unit 32 controls the operation of the moving device 14 based on information indicating the optimal placement of the directional microphone 13 supplied from the calculation processing unit 36, as described below, and moves the directional microphone 13 to a position or orientation according to the information.
  • the information acquisition unit 33 acquires various pieces of information required for the calculation processing unit 36 to calculate the optimal placement of the directional microphone 13.
  • the information acquisition unit 33 acquires information indicating the position and orientation of the imaging camera 12 before imaging by the imaging camera 12 begins (hereinafter referred to as initial camera information), and supplies this to the calculation processing unit 36.
  • the information acquisition unit 33 acquires information indicating the position and orientation of the sound collection target 21 for each sound collection target 21 (hereinafter referred to as initial sound collection target information), and information indicating the position or orientation of the directional microphone 13 (hereinafter referred to as initial microphone information), and supplies this to the positional relationship identification unit 34 and the calculation processing unit 36.
  • the information acquisition unit 33 continues to acquire information indicating the position or orientation of the shooting camera 12, the sound collection target 21, and the directional microphone 13 at all times until the shooting ends.
  • the information acquisition unit 33 then supplies the calculation processing unit 36 with information indicating the position and orientation of the shooting camera 12 at each time point during shooting (hereinafter referred to as camera information), information indicating the position and orientation of the sound collection target 21 at each time point during shooting (hereinafter referred to as sound collection target information), and information indicating the position or orientation of the directional microphone 13 at each time point during shooting (hereinafter referred to as microphone information).
  • the positional relationship determination unit 34 determines the positional relationship between the directional microphone 13 and the sound collection target 21 at the time when shooting by the shooting camera 12 begins, based on the initial microphone information and initial sound collection target information supplied from the information acquisition unit 33. For example, the positional relationship determination unit 34 determines the direction in which the directional microphone 13 is pointed and the distance from the directional microphone 13 to the sound collection target 21 as the positional relationship. Then, the positional relationship determination unit 34 stores the positional relationship information indicating the positional relationship between the directional microphone 13 and the sound collection target 21 in the memory unit 35.
  • the storage unit 35 stores the positional relationship information supplied from the positional relationship identification unit 34.
  • the calculation processing unit 36 is supplied with information indicating the angle of view of the shooting camera 12, information indicating the shape and directional characteristics of the directional microphone 13, and information indicating the shape of the moving device 14 from the user input unit 31. Alternatively, the calculation processing unit 36 may acquire this information by communicating with other equipment. Furthermore, the calculation processing unit 36 is supplied with camera information, sound collection target information, and microphone information from the information acquisition unit 33. Then, based on this information, the calculation processing unit 36 calculates the position or orientation of the directional microphone 13 that is optimal for picking up the sound of the sound collection target 21, and supplies information indicating the optimal placement of the directional microphone 13, which instructs the operation control unit 32 to place the directional microphone 13 in that position or orientation.
  • the calculation processing unit 36 has a recordable range estimation unit 41, a reflection range estimation unit 42, and an optimal placement calculation unit 43.
  • the recording range estimation unit 41 estimates the recording range of the sound of the sound collection target 21 at each point during shooting based on the sound collection target information supplied from the information acquisition unit 33.
  • the reflection range estimation unit 42 estimates the reflection range of the shooting camera 12 at each point during shooting based on information indicating the angle of view of the shooting camera 12 supplied from the user input unit 31 and camera information supplied from the information acquisition unit 33.
  • the optimal placement calculation unit 43 estimates an area that is inside the recordable range estimated by the recordable range estimation unit 41 and outside the reflected range estimated by the reflected range estimation unit 42 as a possible placement area in which the directional microphone 13 can be placed. At this time, the optimal placement calculation unit 43 refers to information indicating the shape of the directional microphone 13 and information indicating the shape of the mobile device 14, and estimates the possible placement area of the directional microphone 13 so that the directional microphone 13 and the mobile device 14 do not enter the angle of view of the shooting camera 12.
  • the optimal placement calculation unit 43 calculates the position that is closest to the sound collection target 21 within the placement area of the directional microphone 13 as the optimal position for picking up the sound of the sound collection target 21. Furthermore, based on the information indicating the directional characteristics of the directional microphone 13 supplied from the user input unit 31 and the sound collection target information supplied from the information acquisition unit 33, the optimal placement calculation unit 43 calculates the direction in which the sound axis of the directional microphone 13 faces the sound collection target 21 at the optimal position for picking up the sound of the sound collection target 21 as the optimal orientation for picking up the sound of the sound collection target 21. In this way, the optimal placement calculation unit 43 can calculate the optimal placement (position or orientation) of the directional microphone 13, and can supply information indicating the optimal placement of the directional microphone 13 to the operation control unit 32.
  • the filming and recording system 11 configured in this manner can record the sound of the recording target 21 with the directional microphone 13 in an optimal position, for example, pointing the sound axis toward the recording target 21 at a position that is inside the recording range and outside the projection range and closest to the recording target 21, and can record the sound of the recording target 21 with the directional microphone 13. This allows the filming and recording system 11 to record the sound of the recording target 21 more effectively.
  • the user can specify the position or orientation of the directional microphone 13 before filming by the filming camera 12 begins, and adjust the position or orientation of the directional microphone 13 as necessary while checking the sound being collected.
  • the user can adjust the sound axis of the directional microphone 13 to point toward the mouth, forehead, chest, etc. of the sound collection target 21, or adjust the position of the directional microphone 13 so that it is at a distance where the sound can be more easily heard.
  • the operation control unit 32 can control the operation of the moving device 14 to read out the positional relationship information stored in the memory unit 35 and reproduce that positional relationship.
  • the positional relationship between the directional microphone 13 and the sound collection target 21 can be reproduced when the same scene is filmed multiple times, including rehearsals, on a filming site for a movie or drama, or when filming is resumed after being interrupted.
  • the information acquisition unit 33 can use the following first to third identification methods to identify the position and orientation of the recording target 21.
  • the position and orientation of the sound collection target 21 can be identified by the sensor.
  • a camera is installed on the mobile device 14 together with the directional microphone 13, the position and orientation of the sound collection target 21 can be identified by analyzing the video captured by the camera.
  • identification of the sound collection target 21 by the sensor and identification of the sound collection target 21 by analyzing the video captured by the camera may be combined.
  • the person at the center of the image captured by the camera 12 can be identified as the sound collection target 21.
  • the sound collection targets 21 For example, when a small number of people (e.g., people who speak at a certain time) are identified as the sound collection targets 21 from among many people, it is necessary to point the directional microphone 13 at the speaker every time they speak. Therefore, by analyzing the image captured by the camera 12, the speaker can be identified as the sound collection target 21.
  • the speaker and their position are identified based on speech picked up by multiple microphones or a microphone array, and the speaker who made the speech can be set as the pickup target 21.
  • the filming and recording system 11 can determine the sound collection position of the directional microphone 13 using information indicating the acoustic radiation characteristics of the sound collection target 21. For example, if the sound collection target 21 has acoustic radiation characteristics, the user can input information regarding the direction from which sound should be collected in advance using the user input unit 31, and the position or orientation of the directional microphone 13 is determined based on that information. In particular, for a sound collection target 21 such as a musical instrument for which there is know-how regarding the sound collection position, it is preferable to store in advance the best information regarding where and from what distance sound should be collected.
  • the orientation of the sound collection target 21 does not need to be taken into consideration when calculating the position or orientation of the directional microphone 13.
  • the filming and sound collection system 11 can automatically perform post-processing on the collected sound by referring to information obtained when the sound of the sound collection target 21 was collected. For example, in the filming and sound collection system 11, if it is desired to remove the propeller noise of the moving device 14 contained in the sound signal in the post-processing of the collected sound, information indicating the rotation speed of the propeller of the moving device 14 is stored. Then, in the post-processing, the frequency of the propeller noise can be identified according to that information, and the propeller noise of the moving device 14 can be removed.
  • the volume and frequency characteristics can be adjusted according to the positional relationship with the sound collection target 21, the directional characteristics of the directional microphone 13, and the radiation characteristics of the sound collection target 21.
  • step S11 the user uses the user input unit 31 to input equipment information including information indicating the angle of view of the imaging camera 12, information indicating the shape and directional characteristics of the directional microphone 13, and information indicating the shape of the mobile device 14.
  • the user input unit 31 supplies the equipment information input by the user to the calculation processing unit 36.
  • step S12 the information acquisition unit 33 acquires initial camera information indicating the position and orientation of the imaging camera 12 before imaging by the imaging camera 12 begins, and supplies this information to the calculation processing unit 36.
  • step S13 the user uses the user input unit 31 to specify the directional microphone 13 to be used to collect sound for each sound collection target 21.
  • the user input unit 31 supplies information indicating the directional microphone 13 specified for each sound collection target 21 to the calculation processing unit 36.
  • steps S14 to S17 pre-imaging processing is performed for each sound collection target 21 for which a directional microphone 13 was specified in step S13.
  • step S14 the information acquisition unit 33 acquires initial sound collection target information indicating the position and orientation of the sound collection target 21 before shooting by the shooting camera 12 begins. Furthermore, for the directional microphone 13 specified in step S13, the information acquisition unit 33 acquires initial microphone information indicating the position or orientation of the directional microphone 13 before shooting by the shooting camera 12 begins.
  • step S15 the user specifies the position or orientation of the directional microphone 13 using the user input unit 31.
  • the user input unit 31 supplies information indicating the position or orientation of the directional microphone 13 specified by the user to the operation control unit 32, and the operation control unit 32 controls the operation of the mobile device 14 based on that information.
  • the directional microphone 13 is positioned so that it is in the position or orientation specified by the user.
  • step S16 the user adjusts the position or orientation of the directional microphone 13 using the user input unit 31 while checking the sound picked up by the directional microphone 13 placed in step S15.
  • the user input unit 31 supplies information indicating the position or orientation of the directional microphone 13 adjusted by the user to the operation control unit 32, and the operation control unit 32 controls the operation of the mobile device 14 based on that information.
  • the directional microphone 13 is placed so that it is in the position or orientation adjusted by the user.
  • the user adjusts the placement of the directional microphone 13 and checks the sound picked up by the directional microphone 13, and the placement of the directional microphone 13 before the shooting camera 12 starts shooting is determined.
  • step S17 the information acquisition unit 33 acquires initial microphone information for the arrangement determined in step S17 and supplies it to the positional relationship determination unit 34. Then, based on the initial microphone information for the arrangement determined in step S17 and the initial sound collection target information supplied in step S14, the positional relationship determination unit 34 determines the positional relationship between the directional microphone 13 and the sound collection target 21 at the time when shooting by the shooting camera 12 begins, and stores the positional relationship information indicating this positional relationship in the storage unit 35.
  • step S18 it is determined whether the positional relationships of all sound collection targets 21 have been stored in the storage unit 35, and if it is determined that the positional relationships of all sound collection targets 21 have not been stored in the storage unit 35, the process returns to step S13. That is, in this case, in step S13, the directional microphone 13 to be used for picking up the sound of the next sound collection target 21 is specified, and the same process is repeated thereafter.
  • step S18 determines whether the positional relationships of all sound collection targets 21 have been stored in the memory unit 35. If it is determined in step S18 that the positional relationships of all sound collection targets 21 have been stored in the memory unit 35, the process proceeds to step S19.
  • step S19 the camera 12 starts capturing images, and the directional microphone 13 starts collecting sound.
  • step S20 the information acquisition unit 33 acquires information indicating the position or orientation of each of the shooting camera 12, the sound collection target 21, and the directional microphone 13, and supplies the camera information, sound collection target information, and microphone information to the calculation processing unit 36.
  • step S21 the calculation processing unit 36 calculates the optimal position or orientation of the directional microphone 13 based on the camera information, sound collection target information, and microphone information supplied from the information acquisition unit 33 in step S20.
  • the calculation processing unit 36 then supplies information indicating the optimal placement of the directional microphone 13 to the operation control unit 32, which instructs the directional microphone 13 to be placed in that position or orientation.
  • step S22 the operation control unit 32 controls the operation of the moving device 14 based on the information indicating the optimal placement of the directional microphone 13 supplied from the calculation processing unit 36 in step S21. As a result, the directional microphone 13 is placed in an optimal position or orientation for picking up the sound of the sound collection target 21.
  • step S23 it is determined whether or not to end image capture by the imaging camera 12. If it is determined that image capture by the imaging camera 12 should not be ended, the process returns to step S20, and the same process is repeated thereafter. On the other hand, if it is determined in step S23 that image capture by the imaging camera 12 should be ended, the process ends.
  • the directional microphone 13 is positioned in the optimal position or orientation for collecting the sound of the sound collection target 21, thereby enabling better sound collection.
  • step S35 the calculation processing unit 36 calculates the optimal position or orientation of the directional microphone 13 based on the initial camera information, the initial sound collection target information, and the initial microphone information, and supplies information indicating that position or orientation to the operation control unit 32.
  • step S35 the operation control unit 32 controls the operation of the moving device 14 based on the information indicating the position or orientation of the directional microphone 13 supplied from the calculation processing unit 36 in step S34. As a result, the directional microphone 13 is positioned so that it is in the optimal position or orientation calculated in step S34.
  • steps S36 to S43 the same processing as in steps S16 to S23 in FIG. 3 is performed.
  • the filming and sound collection system 11 performs filming and sound collection processing, and from the time filming begins, the directional microphone 13 is positioned in the optimal position or orientation for collecting the sound of the sound collection target 21, allowing for better sound collection.
  • the generation position of sounds other than the target 21 and the directional characteristics of the directional microphone 13 are input to the information processing device 15 as additional input information.
  • the position or orientation of the directional microphone 13 When considering lighting equipment, the position or orientation of the directional microphone 13 must be determined so that the directional microphone 13 does not block the lighting light and so that the shadow of the directional microphone 13 is not cast. Therefore, in order to satisfy this condition, in the filming and recording system 11, the position, orientation, and size of the lighting equipment, as well as the directional characteristics and light amount of the lighting light, are input to the information processing device 15 as additional input information.
  • the position, orientation, shape, and size of the reflective object and the reflected object, as well as the reflectance are input to the information processing device 15 as additional input information.
  • the time, latitude, longitude, and weather are input to the information processing device 15 as additional input information.
  • the output of a camera or sensor that observes the surrounding conditions is input to the information processing device 15 as additional input information.
  • Figure 6 is a diagram that explains an example of the optimal position or orientation of the directional microphone 13 when taking into account the noise source.
  • noise from the noise source 22 may be noise from various devices such as air conditioners, refrigerators, ventilation fans, or generators, traffic noise from aircraft or emergency vehicles, construction noise, etc.
  • the optimal placement calculation unit 43 calculates the optimal placement of the directional microphone 13 based on the position of the noise source 22 and the directional characteristics of the directional microphone 13 so as to avoid the sound axis of the directional microphone 13 pointing towards the noise source 22. In other words, the optimal placement calculation unit 43 calculates the optimal placement of the directional microphone 13 at a position that is closest to the sound collection target 21 within the placement area of the directional microphone 13 so that the sound axis of the directional microphone 13 points towards the sound collection target 21 and not towards the noise source 22.
  • FIG. 6A it is preferable to position the directional microphone 13 outside the reflection range so that its sound axis faces the sound collection target 21 and not the noise source 22.
  • FIG. 6B shows an undesirable example in which the directional microphone 13 is located outside the reflection range and its sound axis faces the sound collection target 21, but the sound axis of the directional microphone 13 also faces the noise source 22.
  • the filming and recording system 11 may, for example, analyze the recorded signal and output an alert to notify the user if the signal contains noise other than the voice of the recording target 21.
  • Figure 7 is a diagram illustrating an example of the optimal position or orientation of the directional microphone 13 when taking lighting equipment into consideration.
  • the optimal placement calculation unit 43 calculates the optimal placement of the directional microphone 13 based on the position, orientation, and size of the lighting equipment 16, the directional characteristics of the lighting light, and the amount of light, so as to prevent the directional microphone 13 from blocking the lighting light and to prevent the shadow of the directional microphone 13 from being cast. In other words, the optimal placement calculation unit 43 calculates the optimal placement of the directional microphone 13 at a position that is closest to the sound collection target 21 within the placement area of the directional microphone 13, and so as to prevent the directional microphone 13 from entering the area where the lighting equipment 16 irradiates the lighting light.
  • the directional microphone 13 it is preferable to place the directional microphone 13 in a position outside the reflection range and not in the area irradiated by the illumination light, so that the sound axis of the directional microphone 13 is oriented toward the sound collection target 21.
  • B of FIG. 7 shows an undesirable example in which the directional microphone 13 is placed in a position outside the reflection range but in the area irradiated by the illumination light.
  • Figure 8 shows an example of using one directional microphone 13 in a situation where multiple targets 21 for recording are facing each other and conversing simultaneously.
  • Figure 9 shows an example of using one directional microphone 13 in a situation where multiple recording targets 21 are nearby and taking turns conversing.
  • the sound axis of directional microphone 13 can be directed toward sound collection target 21A at time t1 when sound collection target 21A speaks, and toward sound collection target 21B at time t2 when sound collection target 21B speaks. This allows one directional microphone 13 to pick up the voices of multiple sound collection targets 21 well.
  • Figure 10 shows an example of use of one omnidirectional microphone 17.
  • an omnidirectional microphone 17 when mounted on a mobile device 14, it is preferable to place the omnidirectional microphone 17 at a position equidistant to multiple sound collection targets 21. This allows one omnidirectional microphone 17 to effectively pick up the sounds of multiple sound collection targets 21.
  • FIG. 11 shows an example of using multiple directional microphones 13 for each of multiple sound collection targets 21.
  • two directional microphones 13A and 13B are mounted on a mobile device 14, and the mobile device 14 is positioned halfway between sound collection targets 21A and 21B. It is preferable to orient the sound axis of directional microphone 13A toward sound collection target 21A, and the sound axis of directional microphone 13B toward sound collection target 21B. This allows the multiple directional microphones 13 to effectively pick up the sounds of multiple sound collection targets 21.
  • the filming and sound recording system 11 can visualize the situation at the filming site as if looking down from the sky, and display an arrangement screen 51 showing the arrangement of the filming cameras 12, directional microphones 13, lighting equipment 16, sound recording targets 21, etc. on the display unit of the information processing device 15.
  • the situation at the filming site can be visualized from the side, or in a stereoscopic three-dimensional form, and displayed on the arrangement screen 51.
  • the directional microphone 13 needs to be placed within an area of a recording range that is a range in which the direct sound of the sound collection target 21 can be recorded when the sound axis of the directional microphone 13 is directed toward the sound collection target 21. Also, the directional microphone 13 needs to be placed outside an area of a reflection range that is a range in which the microphone appears in the image captured by the shooting camera 12. Furthermore, the directional microphone 13 needs to be placed so as not to block the illumination light and so as not to cast a shadow.
  • the arrangement screen 51 displays areas that are appropriate for placing the directional microphone 13, areas that are not acceptable because they will be reflected in the image captured by the imaging camera 12, areas that are not acceptable because they will obstruct or cast a shadow on the lighting equipment 16, areas that are not acceptable because they are far from the sound collection target 21, and areas that are not acceptable because reflections from glass or the like will be reflected in the image captured by the imaging camera 12. Furthermore, the arrangement screen 51 displays icons representing the imaging camera 12, the directional microphone 13, the lighting equipment 16, the sound collection target 21, etc. The user can recognize whether the arrangement of the directional microphone 13 is appropriate by referring to the arrangement screen 51.
  • the placement screen 51 shown in FIG. 12A shows an OK example in which the icon of the directional microphone 13 is displayed in an area appropriate for the placement of the directional microphone 13, indicating that the placement of the directional microphone 13 is optimal.
  • the arrangement screen 51 shown in FIG. 12B shows an example of an NG situation in which the icon of a directional microphone 13 is displayed in an area that is NG because it will be reflected in the image captured by the filming camera 12.
  • a message is displayed to inform the user that the directional microphone 13 is NG because it will be reflected in the filming camera 12.
  • the user specifies the position of the directional microphone 13 or mobile device 14 and the position is not appropriate, a message informing the user of the cause can be displayed and an alert can be output.
  • the filming and recording system 11 as in the arrangement screen 51 shown in the upper part of FIG. 13, if there is no suitable area for the arrangement of the directional microphone 13, a message is displayed indicating that it is NG because there is no suitable area for the arrangement of the directional microphone 13. In this case, the filming and recording system 11 can present alternatives indicating which conditions should be relaxed to obtain a suitable area for the arrangement of the directional microphone 13.
  • an icon of the directional microphone 13 is displayed in an area that is appropriate as the layout position of the directional microphone 13 obtained when the position of the lighting equipment 16 is moved.
  • a user interface may be displayed that allows the user to change the position or orientation of those objects on the placement screen 51 and check whether an appropriate area exists for placing the directional microphone 13.
  • the operation control unit 32 controls the moving device 14 to move to a position or orientation that will avoid contact with the other object or person, regardless of whether the device will enter the range of the filming camera 12. In this way, when avoiding contact with other objects or people is given priority, the arrangement screen 51 notifies the user that the arrangement position of the directional microphone 13 is different from that expected.
  • the filming and recording system 11 can be configured to utilize a microphone cable 18, and the directional microphone 13 is connected to a recording system 19 that records sound via the microphone cable 18.
  • the directional microphone 13 can transmit audio signals to the recording system 19 wirelessly, or can transmit audio signals to the recording system 19 wired via the microphone cable 18.
  • the position of the mobile device 14A is determined so that, for example, the distance from the mobile device 14A to the recording system 19 is within a range less than the length of the microphone cable 18.
  • the filming and recording system 11 may also use a moving device 14B for supporting the microphone cable 18 to prevent the microphone cable 18 from appearing in the image captured by the filming camera 12.
  • FIG. 14 illustrates one moving device 14B supporting the microphone cable
  • two or more moving devices 14B can be used to support the microphone cable 18.
  • information indicating the number of multiple moving devices 14B supporting the microphone cable 18 is input to the information acquisition unit 33, and the calculation processing unit 36 controls the operation of the moving device 14A or one or more moving devices 14B so as to prevent the microphone cable 18 from appearing in the image captured by the filming camera 12.
  • the positions of the multiple mobile devices 14B are determined, for example, between the mobile device 14A and the recording system 19 so that the multiple mobile devices 14B support the microphone cable 18 at equal intervals.
  • the positions of the mobile device 14A and the multiple mobile devices 14B are determined so that the distance from the mobile device 14A to the recording system 19 via the multiple mobile devices 14B is within a range less than the length of the microphone cable 18.
  • an omnidirectional microphone 17 when used in the filming and sound collection system 11, there is no need to control the direction. Also, when a spherical array microphone is used in the filming and sound collection system 11, the sound of the sound collection target 21 arriving from any direction can be emphasized in post-processing, so there is no need to control the direction. Also, because the shape and diaphragm position of the microphone differ from microphone to microphone, the filming and sound collection system 11 must take these factors into consideration when calculating the optimal placement of the directional microphone 13.
  • the focal length and sensor size may be used as equipment information related to the filming camera 12.
  • the image itself captured by the filming camera 12 may be analyzed to detect whether the directional microphone 13 is captured.
  • a crane, rover, wire, etc. can be used as a moving object that holds the directional microphone 13.
  • information about these must be input to the information processing device 15 and the optimal placement of the directional microphone 13 must be calculated.
  • the position and orientation of the directional microphone 13 are controlled using information including both the position and orientation of the directional microphone 13.
  • the position of the omnidirectional microphone 17 is controlled using information including at least the position of the omnidirectional microphone 17.
  • FIG. 15 is a block diagram showing an example of the configuration of one embodiment of a computer on which a program that executes the series of processes described above is installed.
  • the program can be pre-recorded on the hard disk 105 or ROM 103 as a recording medium built into the computer.
  • the program can be stored (recorded) on a removable recording medium 111 driven by the drive 109.
  • a removable recording medium 111 can be provided as so-called packaged software.
  • examples of the removable recording medium 111 include a flexible disk, a CD-ROM (Compact Disc Read Only Memory), an MO (Magneto Optical) disk, a DVD (Digital Versatile Disc), a magnetic disk, a semiconductor memory, etc.
  • the program can also be downloaded to the computer via a communication network or broadcasting network and installed in the built-in hard disk 105. That is, the program can be transferred to the computer wirelessly from a download site via an artificial satellite for digital satellite broadcasting, or transferred to the computer via a wired connection via a network such as a LAN (Local Area Network) or the Internet.
  • a network such as a LAN (Local Area Network) or the Internet.
  • the computer has a built-in CPU (Central Processing Unit) 102, to which an input/output interface 110 is connected via a bus 101.
  • CPU Central Processing Unit
  • the CPU 102 executes a program stored in the ROM (Read Only Memory) 103 accordingly.
  • the CPU 102 loads a program stored on the hard disk 105 into the RAM (Random Access Memory) 104 and executes it.
  • the CPU 102 performs processing according to the above-mentioned flowchart, or processing performed by the configuration of the above-mentioned block diagram. Then, the CPU 102 outputs the processing results from the output unit 106 via the input/output interface 110, or transmits them from the communication unit 108, or even records them on the hard disk 105, as necessary.
  • the input unit 107 is composed of a keyboard, mouse, microphone, etc.
  • the output unit 106 is composed of an LCD (Liquid Crystal Display), speaker, etc.
  • the processing performed by a computer according to a program does not necessarily have to be performed in chronological order according to the order described in the flowchart.
  • the processing performed by a computer according to a program also includes processing that is executed in parallel or individually (for example, parallel processing or processing by objects).
  • the program may be processed by one computer (processor), or may be distributed among multiple computers. Furthermore, the program may be transferred to a remote computer for execution.
  • a system refers to a collection of multiple components (devices, modules (parts), etc.), regardless of whether all the components are in the same housing. Therefore, multiple devices housed in separate housings and connected via a network, and a single device in which multiple modules are housed in a single housing, are both systems.
  • the configuration described above as one device (or processing unit) may be divided and configured as multiple devices (or processing units).
  • the configurations described above as multiple devices (or processing units) may be combined and configured as one device (or processing unit).
  • part of the configuration of one device (or processing unit) may be included in the configuration of another device (or other processing unit).
  • this technology can be configured as cloud computing, in which a single function is shared and processed collaboratively by multiple devices via a network.
  • the above-mentioned program can be executed in any device.
  • the device has the necessary functions (functional blocks, etc.) and is capable of obtaining the necessary information.
  • each step described in the above flowchart can be executed by one device, or can be shared and executed by multiple devices.
  • one step includes multiple processes, the multiple processes included in that one step can be executed by one device, or can be shared and executed by multiple devices.
  • multiple processes included in one step can be executed as multiple step processes.
  • processes described as multiple steps can be executed collectively as one step.
  • processing of the steps that describe a program executed by a computer may be executed chronologically in the order described in this specification, or may be executed in parallel, or individually at the required timing, such as when a call is made. In other words, as long as no contradictions arise, the processing of each step may be executed in an order different from the order described above. Furthermore, the processing of the steps that describe this program may be executed in parallel with the processing of other programs, or may be executed in combination with the processing of other programs.
  • the present technology can also be configured as follows. (1) An information acquisition unit that acquires information necessary to calculate an optimal arrangement of the sound collection equipment, the information including at least information indicating a position or orientation of a sound collection device that collects a sound to be collected; A calculation processing unit that calculates an optimal arrangement of the sound pickup equipment for picking up the sound of the sound pickup target based on the information by the information acquisition unit; and an operation control unit that controls the position or orientation of the sound pickup equipment so as to achieve the arrangement calculated by the calculation processing unit.
  • the information acquisition unit acquires information indicating a position and an orientation of an imaging device that captures the sound collection target, and information indicating a position and an orientation of the sound collection target, and supplies the information to the arithmetic processing unit;
  • An input unit for inputting information used by the arithmetic processing unit is further provided, The information processing device described in (2) above, wherein the input unit inputs information indicating the angle of view of the photographing equipment, information indicating the shape and directional characteristics of the sound collection equipment, and information indicating the shape of a moving device that moves the sound collection equipment to the calculation processing unit.
  • the information processing device according to (3) above further comprising: a positional relationship specifying unit that specifies and stores a positional relationship between the sound collection equipment and the sound collection target before the shooting by the shooting equipment is started.
  • An information processing device acquires information including at least information indicating a position or orientation of a sound collection device that collects a sound of a sound collection target, and information necessary for calculating an optimal arrangement of the sound collection device; Calculating an optimal arrangement of the sound pickup equipment for picking up the sound of the sound pickup target based on the information; and controlling the position or orientation of the sound pickup equipment so as to achieve the arrangement.
  • the computer of the information processing device Acquiring information necessary for calculating an optimal arrangement of the sound pickup equipment, the information including at least information indicating a position or orientation of the sound pickup equipment that picks up the sound of the sound pickup target; Calculating an optimal arrangement of the sound pickup equipment for picking up the sound of the sound pickup target based on the information; and controlling the position or orientation of the sound pickup equipment so as to achieve the arrangement.

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Automation & Control Theory (AREA)
  • Aviation & Aerospace Engineering (AREA)
  • Multimedia (AREA)
  • Circuit For Audible Band Transducer (AREA)
  • Studio Devices (AREA)

Abstract

The present disclosure relates to an information processing device, an information processing method, and a program with which improved sound collection can be achieved. An information acquisition unit acquires information necessary for calculating an optimal arrangement of a directional microphone that collects sounds to be collected, the information including at least information indicating the position or orientation of the directional microphone. An arithmetic processing unit calculates, on the basis of the information, the optimal arrangement of the directional microphone that collects sounds to be collected. Further, an operation control unit controls an operation of the position or orientation of a moving device, on which the directional microphone is mounted, so that the directional microphone reaches the arrangement calculated by the arithmetic processing unit. The present technology can be used in, for example, a system for imaging and sound collecting.

Description

情報処理装置および情報処理方法、並びにプログラムInformation processing device, information processing method, and program
 本開示は、情報処理装置および情報処理方法、並びにプログラムに関し、特に、より良好に音声を収音することができるようにした情報処理装置および情報処理方法、並びにプログラムに関する。 The present disclosure relates to an information processing device, an information processing method, and a program, and in particular to an information processing device, an information processing method, and a program that enable better audio pickup.
 従来、映画やドラマなどの映像撮影において、役者の音声を収音するために指向性マイクロフォンが使用されることがある。一般的に、ブームの先端に指向性マイクロフォンを設置して収音するときには、ブームを保持する音声スタッフ(音響スタッフ)は、自身や、指向性マイクロフォン、ブームなどが映り込まないように注意する必要がある。また、従来、指向性マイクロフォンの位置決めや、ブームの保持は、音声スタッフ(音響スタッフ)により行われていたが、将来的には、いわゆるドローンやロボットなどによって代替されることが期待されている。 Traditionally, directional microphones are sometimes used to capture the voices of actors during filming of movies, dramas, and other video content. Generally, when recording sound using a directional microphone set at the end of a boom, the sound staff holding the boom must be careful not to capture themselves, the directional microphone, or the boom in the footage. Also, traditionally, the positioning of the directional microphone and holding the boom were performed by the sound staff, but in the future, it is expected that this will be replaced by drones, robots, and the like.
 例えば、引用文献1には、被写体を撮像する撮像部を有する移動体と、被写体に光を照射する照射部を有する移動体とを含む複数の移動体により被写体を撮像する撮像システムにおいて、それらの移動体の位置を決定する撮像システムが開示されている。 For example, cited reference 1 discloses an imaging system that determines the positions of multiple moving bodies that capture images of a subject, the moving bodies including a moving body having an imaging unit that captures images of the subject and a moving body having an illumination unit that illuminates the subject with light.
 また、引用文献2には、移動物体が有する撮像部の状態情報を取得し、その状態情報に基づいて撮像部の撮影可能領域を算出して、撮影可能領域に関する情報を移動体に送信する情報処理装置が開示されている。 Furthermore, cited document 2 discloses an information processing device that acquires status information of an imaging unit possessed by a moving object, calculates the imageable area of the imaging unit based on the status information, and transmits information regarding the imageable area to the moving object.
国際公開第2019/189370号International Publication No. 2019/189370 国際公開第2021/024627号International Publication No. 2021/024627
 ところで、上述したように、指向性マイクロフォンの位置決めや保持が代替される場合には、良好に音声を収音することができるようにすることが求められている。 As mentioned above, when the positioning or holding of a directional microphone is substituted, it is required to be able to pick up sound well.
 本開示は、このような状況に鑑みてなされたものであり、より良好に音声を収音することができるようにするものである。 This disclosure was made in light of these circumstances, and aims to make it possible to pick up sound more effectively.
 本開示の一側面の情報処理装置は、収音対象の音を収音する収音機材の位置または向きを示す情報を少なくとも含み、最適な前記収音機材の配置を算出するのに必要となる情報を取得する情報取得部と、前記情報取得部により前記情報に基づいて、前記収音対象の音を収音するのに最適な前記収音機材の配置を算出する演算処理部と、前記演算処理部により算出された配置となるように前記収音機材の位置または向きの動作を制御する動作制御部とを備える。 The information processing device of one aspect of the present disclosure includes an information acquisition unit that acquires information including at least information indicating the position or orientation of the sound pickup equipment that picks up the sound of the target sound pickup, and that acquires information necessary to calculate an optimal placement of the sound pickup equipment, a calculation processing unit that calculates, based on the information by the information acquisition unit, an optimal placement of the sound pickup equipment for picking up the sound of the target sound pickup, and an operation control unit that controls the operation of the position or orientation of the sound pickup equipment so as to achieve the placement calculated by the calculation processing unit.
 本開示の一側面の情報処理方法またはプログラムは、収音対象の音を収音する収音機材の位置または向きを示す情報を少なくとも含み、最適な前記収音機材の配置を算出するのに必要となる情報を取得することと、前記情報に基づいて、前記収音対象の音を収音するのに最適な前記収音機材の配置を算出することと、その配置となるように前記収音機材の位置または向きの動作を制御することとを含む。 An information processing method or program according to one aspect of the present disclosure includes at least information indicating the position or orientation of a pickup device that picks up a target sound, and includes acquiring information necessary to calculate an optimal placement of the pickup device, calculating an optimal placement of the pickup device for picking up the target sound based on the information, and controlling the position or orientation of the pickup device to achieve that placement.
 本開示の一側面においては、収音対象の音を収音する収音機材の位置または向きを示す情報を少なくとも含み、最適な収音機材の配置を算出するのに必要となる情報が取得され、その情報に基づいて、収音対象の音を収音するのに最適な収音機材の配置が算出され、その配置となるように収音機材の位置または向きの動作が制御される。 In one aspect of the present disclosure, information is acquired that includes at least information indicating the position or orientation of the pickup equipment that picks up the target sound, and the optimal placement of the pickup equipment is calculated based on that information, and the position or orientation of the pickup equipment is controlled to achieve that placement.
本技術を適用した撮影収音システムの一実施の形態の構成例を示す図である。1 is a diagram showing an example of the configuration of an embodiment of an imaging and sound collecting system to which the present technology is applied; 情報処理装置の構成例を示すブロック図である。1 is a block diagram showing an example of the configuration of an information processing device; 撮影収音処理の第1の処理例を説明するフローチャートである。11 is a flowchart illustrating a first processing example of an imaging and sound collection process. 撮影収音処理の第2の処理例を説明するフローチャートである。13 is a flowchart illustrating a second processing example of the shooting and sound collection processing. 考慮対象、条件、および追加の入力情報について説明する図である。FIG. 13 is a diagram illustrating considerations, conditions, and additional input information. 騒音源を考慮する例について説明する図である。FIG. 13 is a diagram illustrating an example in which a noise source is taken into consideration. 照明機材を考慮する例について説明する図である。FIG. 13 is a diagram illustrating an example in which lighting equipment is taken into consideration. 複数の収音対象の音声を収音する第1の例について説明する図である。FIG. 2 is a diagram illustrating a first example of collecting sounds from a plurality of sound collection targets. 複数の収音対象の音声を収音する第2の例について説明する図である。FIG. 11 is a diagram illustrating a second example of collecting sounds from a plurality of sound collection targets. 複数の収音対象の音声を収音する第3の例について説明する図である。FIG. 13 is a diagram illustrating a third example of collecting sounds from a plurality of sound collection targets. 複数の収音対象の音声を収音する第4の例について説明する図である。FIG. 13 is a diagram illustrating a fourth example of collecting sounds from a plurality of sound collection targets. 配置画面の一例を示す図である。FIG. 13 is a diagram showing an example of an arrangement screen. 配置画面の一例を示す図である。FIG. 13 is a diagram showing an example of an arrangement screen. マイクケーブルを利用した構成例について説明する図である。1A and 1B are diagrams illustrating an example configuration using a microphone cable. 本技術を適用したコンピュータの一実施の形態の構成例を示すブロック図である。1 is a block diagram showing an example of the configuration of an embodiment of a computer to which the present technology is applied.
 以下、本技術を適用した具体的な実施の形態について、図面を参照しながら詳細に説明する。 Below, specific embodiments of the application of this technology will be described in detail with reference to the drawings.
 <撮影収音システムの構成例>
 図1は、本技術を適用した撮影収音システムの一実施の形態の構成例を示す図である。
<Example of a recording and recording system configuration>
FIG. 1 is a diagram showing an example of the configuration of an embodiment of an imaging and sound collecting system to which the present technology is applied.
 図1に示すように、撮影収音システム11は、撮影カメラ12、指向性マイクロフォン13、移動装置14、および情報処理装置15を備えて構成される。 As shown in FIG. 1, the imaging and recording system 11 is configured with a imaging camera 12, a directional microphone 13, a moving device 14, and an information processing device 15.
 例えば、撮影収音システム11は、映画やドラマなどの撮影現場で利用され、撮影カメラ12で撮影される特定の役者を収音対象21として、収音対象21が発話したセリフなどの音声を、指向性マイクロフォン13で収音することができる。そして、撮影収音システム11は、収音対象21の音声を収音するのに最適な位置または向きとなるように、指向性マイクロフォン13が搭載された移動装置14の動作を、情報処理装置15によって制御することができる。 For example, the filming and recording system 11 can be used on location for movies, dramas, etc., and can use the directional microphone 13 to record sounds such as lines spoken by a specific actor who is being filmed by the filming camera 12 as the recording target 21. The filming and recording system 11 can also use the information processing device 15 to control the operation of the mobile device 14 equipped with the directional microphone 13 so that the mobile device 14 is in the optimal position or orientation for recording the sound of the recording target 21.
 撮影カメラ12は、例えば、使用するレンズに応じた画角で撮影を行う撮影機材である。 The photographing camera 12 is, for example, photographing equipment that takes photographs with an angle of view that corresponds to the lens used.
 指向性マイクロフォン13は、例えば、音軸に沿った狭い範囲の音を主に収音し、その範囲以外の音を収音しないような指向性特性を備えた収音機材である。 Directional microphone 13 is, for example, a recording device with directional characteristics that mainly picks up sounds in a narrow range along the sound axis and does not pick up sounds outside that range.
 移動装置14は、例えば、複数のプロペラにより自律して飛行することが可能な無人航空機(いわゆるドローン)を使用することができ、情報処理装置15による制御に従って、所望の位置または向きとなるように指向性マイクロフォン13を移動させる。 The moving device 14 can be, for example, an unmanned aerial vehicle (a so-called drone) that can fly autonomously using multiple propellers, and moves the directional microphone 13 to a desired position or orientation according to the control of the information processing device 15.
 情報処理装置15は、収音対象21の音声を収音するのに最適な指向性マイクロフォン13の位置または向き(以下、最適な指向性マイクロフォン13の配置と称する)を算出し、その最適な指向性マイクロフォン13の配置で収音対象21の音声を収音することができるように移動装置14の動作を制御する。 The information processing device 15 calculates the optimal position or orientation of the directional microphone 13 for picking up the sound of the collection target 21 (hereinafter referred to as the optimal arrangement of the directional microphone 13), and controls the operation of the moving device 14 so that the sound of the collection target 21 can be picked up in the optimal arrangement of the directional microphone 13.
 例えば、情報処理装置15は、撮影カメラ12により撮影される映像に映り込む範囲となる映り込み範囲(図1の一点鎖線で囲まれる範囲)を、撮影カメラ12の位置、向き、および画角に基づいて推定することができる。また、情報処理装置15は、収音対象21に対して指向性マイクロフォン13の音軸を向けた際に収音対象21の直接音を収録することができる範囲となる収録可能範囲(図1の破線で囲まれる範囲)を、収音対象21の位置および向きに基づいて推定することができる。 For example, the information processing device 15 can estimate the reflection range (the range surrounded by the dashed line in FIG. 1) that is the range that is reflected in the image captured by the capturing camera 12, based on the position, orientation, and angle of view of the capturing camera 12. The information processing device 15 can also estimate the recording possible range (the range surrounded by the dashed line in FIG. 1) that is the range in which the direct sound of the sound collection target 21 can be recorded when the sound axis of the directional microphone 13 is directed toward the sound collection target 21, based on the position and orientation of the sound collection target 21.
 従って、情報処理装置15は、収音対象21の音声の収録可能範囲の内側、かつ、指向性マイクロフォン13および移動装置14が撮影カメラ12の映り込み範囲の外側であって、収音対象21に最も近づくことができる位置を、収音対象21の音声を収音するのに最適な指向性マイクロフォン13の位置として算出する。さらに、情報処理装置15は、その位置において収音対象21に対して指向性マイクロフォン13の音軸が向かう方向を、収音対象21の音声を収音するのに最適な指向性マイクロフォン13の向きとして算出する。そして、情報処理装置15は、最適な指向性マイクロフォン13の配置(位置または向き)となるように、移動装置14の動作を制御する。 Therefore, the information processing device 15 calculates the position where the directional microphone 13 can get closest to the sound collection target 21, inside the range where the sound of the sound collection target 21 can be recorded and outside the range of the image capture camera 12, as the optimal position of the directional microphone 13 for picking up the sound of the sound collection target 21. Furthermore, the information processing device 15 calculates the direction in which the sound axis of the directional microphone 13 faces toward the sound collection target 21 at that position as the optimal orientation of the directional microphone 13 for picking up the sound of the sound collection target 21. Then, the information processing device 15 controls the operation of the moving device 14 so that the directional microphone 13 is positioned (position or orientation) optimally.
 図2は、情報処理装置15の構成例を示すブロック図である。 FIG. 2 is a block diagram showing an example configuration of the information processing device 15.
 図2に示すように、情報処理装置15は、ユーザ入力部31、動作制御部32、情報取得部33、位置関係特定部34、記憶部35、および演算処理部36を備えて構成される。 As shown in FIG. 2, the information processing device 15 includes a user input unit 31, an operation control unit 32, an information acquisition unit 33, a positional relationship determination unit 34, a memory unit 35, and a calculation processing unit 36.
 ユーザ入力部31は、撮影収音システム11のユーザが、各種の情報を情報処理装置15に入力するのに用いられる。例えば、ユーザは、ユーザ入力部31を用いて、撮影収音システム11で使用される機材に関する機材情報として、撮影カメラ12の画角を示す情報、指向性マイクロフォン13の形状および指向性特性を示す情報、並びに、移動装置14の形状を示す情報を入力する。ユーザ入力部31は、これらの機材情報を演算処理部36に供給する。 The user input unit 31 is used by the user of the filming and recording system 11 to input various information to the information processing device 15. For example, the user uses the user input unit 31 to input equipment information related to the equipment used in the filming and recording system 11, such as information indicating the angle of view of the filming camera 12, information indicating the shape and directional characteristics of the directional microphone 13, and information indicating the shape of the mobile device 14. The user input unit 31 supplies this equipment information to the calculation processing unit 36.
 さらに、ユーザは、ユーザ入力部31を用いて、収音対象21ごとに収音に使用する指向性マイクロフォン13を指定することができ、ユーザ入力部31は、収音対象21ごと指定された指向性マイクロフォン13を示す情報を演算処理部36に供給する。また、ユーザは、撮影カメラ12による撮影が開始される前に、ユーザ入力部31を用いて、指向性マイクロフォン13の位置または向きを指定すること、その位置または向きを調整することができ、ユーザ入力部31は、ユーザにより指定または調整された指向性マイクロフォン13の位置または向きを示す情報を動作制御部32に供給する。 Furthermore, the user can use the user input unit 31 to specify the directional microphone 13 to be used for collecting sound for each sound collection target 21, and the user input unit 31 supplies information indicating the directional microphone 13 specified for each sound collection target 21 to the calculation processing unit 36. Furthermore, the user can use the user input unit 31 to specify the position or orientation of the directional microphone 13 or adjust the position or orientation before shooting by the shooting camera 12 begins, and the user input unit 31 supplies information indicating the position or orientation of the directional microphone 13 specified or adjusted by the user to the operation control unit 32.
 動作制御部32は、撮影カメラ12による撮影が開始される前に、ユーザ入力部31から供給される指向性マイクロフォン13の位置または向きを示す情報に基づいて移動装置14の動作を制御し、その情報に従った位置または向きとなるように指向性マイクロフォン13を移動させる。また、動作制御部32は、撮影カメラ12による撮影が開始された後には、後述するように演算処理部36から供給される指向性マイクロフォン13の最適配置を示す情報に基づいて移動装置14の動作を制御し、その情報に従った位置または向きとなるように指向性マイクロフォン13を移動させる。 Before the imaging camera 12 starts capturing images, the operation control unit 32 controls the operation of the moving device 14 based on information indicating the position or orientation of the directional microphone 13 supplied from the user input unit 31, and moves the directional microphone 13 to a position or orientation according to the information. After the imaging camera 12 starts capturing images, the operation control unit 32 controls the operation of the moving device 14 based on information indicating the optimal placement of the directional microphone 13 supplied from the calculation processing unit 36, as described below, and moves the directional microphone 13 to a position or orientation according to the information.
 情報取得部33は、最適な指向性マイクロフォン13の配置を演算処理部36が算出するために必要となる各種の情報を取得する。 The information acquisition unit 33 acquires various pieces of information required for the calculation processing unit 36 to calculate the optimal placement of the directional microphone 13.
 例えば、情報取得部33は、撮影カメラ12による撮影が開始される前の撮影カメラ12の位置および向きを示す情報(以下、初期カメラ情報と称する)を取得して、演算処理部36に供給する。また、情報取得部33は、撮影カメラ12による撮影が開始される前に、収音対象21ごとに、収音対象21の位置および向きを示す情報(以下、初期収音対象情報と称する)、および、指向性マイクロフォン13の位置または向きを示す情報(以下、初期マイク情報と称する)を取得して、位置関係特定部34および演算処理部36に供給する。 For example, the information acquisition unit 33 acquires information indicating the position and orientation of the imaging camera 12 before imaging by the imaging camera 12 begins (hereinafter referred to as initial camera information), and supplies this to the calculation processing unit 36. In addition, before imaging by the imaging camera 12 begins, the information acquisition unit 33 acquires information indicating the position and orientation of the sound collection target 21 for each sound collection target 21 (hereinafter referred to as initial sound collection target information), and information indicating the position or orientation of the directional microphone 13 (hereinafter referred to as initial microphone information), and supplies this to the positional relationship identification unit 34 and the calculation processing unit 36.
 また、情報取得部33は、撮影カメラ12による撮影が開始された後、撮影が終了するまで常時、撮影カメラ12、収音対象21、および指向性マイクロフォン13それぞれの位置または向きを示す情報を取得し続ける。そして、情報取得部33は、撮影中の各時点における撮影カメラ12の位置および向きを示す情報(以下、カメラ情報と称する)、撮影中の各時点における収音対象21の位置および向きを示す情報(以下、収音対象情報と称する)、撮影中の各時点における指向性マイクロフォン13の位置または向きを示す情報(以下、マイク情報と称する)を、演算処理部36に供給する。 In addition, after the shooting camera 12 starts shooting, the information acquisition unit 33 continues to acquire information indicating the position or orientation of the shooting camera 12, the sound collection target 21, and the directional microphone 13 at all times until the shooting ends. The information acquisition unit 33 then supplies the calculation processing unit 36 with information indicating the position and orientation of the shooting camera 12 at each time point during shooting (hereinafter referred to as camera information), information indicating the position and orientation of the sound collection target 21 at each time point during shooting (hereinafter referred to as sound collection target information), and information indicating the position or orientation of the directional microphone 13 at each time point during shooting (hereinafter referred to as microphone information).
 位置関係特定部34は、情報取得部33から供給される初期マイク情報と初期収音対象情報とに基づいて、撮影カメラ12による撮影が開始される時点における指向性マイクロフォン13と収音対象21との位置関係を特定する。例えば、位置関係特定部34は、指向性マイクロフォン13を向ける方向と、指向性マイクロフォン13から収音対象21までの距離とを位置関係として特定する。そして、位置関係特定部34は、指向性マイクロフォン13と収音対象21との位置関係を示す位置関係情報を記憶部35に記憶させる。 The positional relationship determination unit 34 determines the positional relationship between the directional microphone 13 and the sound collection target 21 at the time when shooting by the shooting camera 12 begins, based on the initial microphone information and initial sound collection target information supplied from the information acquisition unit 33. For example, the positional relationship determination unit 34 determines the direction in which the directional microphone 13 is pointed and the distance from the directional microphone 13 to the sound collection target 21 as the positional relationship. Then, the positional relationship determination unit 34 stores the positional relationship information indicating the positional relationship between the directional microphone 13 and the sound collection target 21 in the memory unit 35.
 記憶部35は、位置関係特定部34から供給される位置関係情報を記憶する。 The storage unit 35 stores the positional relationship information supplied from the positional relationship identification unit 34.
 演算処理部36には、撮影カメラ12の画角を示す情報、指向性マイクロフォン13の形状および指向性特性を示す情報、移動装置14の形状を示す情報が、ユーザ入力部31から供給される。または、演算処理部36は、これらの情報を他の機材と通信することによって取得してもよい。さらに、演算処理部36には、カメラ情報、収音対象情報、およびマイク情報が、情報取得部33から供給される。そして、演算処理部36は、これらの情報に基づいて、収音対象21の音声を収音するのに最適な指向性マイクロフォン13の位置または向きを算出し、その位置または向きで指向性マイクロフォン13が配置されるように指示する指向性マイクロフォン13の最適配置を示す情報を、動作制御部32に供給する。 The calculation processing unit 36 is supplied with information indicating the angle of view of the shooting camera 12, information indicating the shape and directional characteristics of the directional microphone 13, and information indicating the shape of the moving device 14 from the user input unit 31. Alternatively, the calculation processing unit 36 may acquire this information by communicating with other equipment. Furthermore, the calculation processing unit 36 is supplied with camera information, sound collection target information, and microphone information from the information acquisition unit 33. Then, based on this information, the calculation processing unit 36 calculates the position or orientation of the directional microphone 13 that is optimal for picking up the sound of the sound collection target 21, and supplies information indicating the optimal placement of the directional microphone 13, which instructs the operation control unit 32 to place the directional microphone 13 in that position or orientation.
 例えば、演算処理部36は、収録可能範囲推定部41、映り込み範囲推定部42、および最適配置算出部43を有している。 For example, the calculation processing unit 36 has a recordable range estimation unit 41, a reflection range estimation unit 42, and an optimal placement calculation unit 43.
 収録可能範囲推定部41は、情報取得部33から供給される収音対象情報に基づいて、撮影中の各時点における収音対象21の音声の収録可能範囲を推定する。 The recording range estimation unit 41 estimates the recording range of the sound of the sound collection target 21 at each point during shooting based on the sound collection target information supplied from the information acquisition unit 33.
 映り込み範囲推定部42は、ユーザ入力部31から供給される撮影カメラ12の画角を示す情報、および、情報取得部33から供給されるカメラ情報に基づいて、撮影中の各時点における撮影カメラ12の映り込み範囲を推定する。 The reflection range estimation unit 42 estimates the reflection range of the shooting camera 12 at each point during shooting based on information indicating the angle of view of the shooting camera 12 supplied from the user input unit 31 and camera information supplied from the information acquisition unit 33.
 最適配置算出部43は、収録可能範囲推定部41により推定された収録可能範囲の内側であって、映り込み範囲推定部42により推定された映り込み範囲の外側となる領域を、指向性マイクロフォン13を配置することが可能な配置可能エリアとして推定する。このとき、最適配置算出部43は、指向性マイクロフォン13の形状を示す情報、および、移動装置14の形状を示す情報を参照して、指向性マイクロフォン13および移動装置14が撮影カメラ12の画角内に入り込まないように、指向性マイクロフォン13の配置可能エリアを推定する。 The optimal placement calculation unit 43 estimates an area that is inside the recordable range estimated by the recordable range estimation unit 41 and outside the reflected range estimated by the reflected range estimation unit 42 as a possible placement area in which the directional microphone 13 can be placed. At this time, the optimal placement calculation unit 43 refers to information indicating the shape of the directional microphone 13 and information indicating the shape of the mobile device 14, and estimates the possible placement area of the directional microphone 13 so that the directional microphone 13 and the mobile device 14 do not enter the angle of view of the shooting camera 12.
 そして、最適配置算出部43は、指向性マイクロフォン13の配置可能エリア内であって、収音対象21に最も近づくことができる位置を、収音対象21の音声を収音するのに最適な位置として算出する。また、最適配置算出部43は、ユーザ入力部31から供給される指向性マイクロフォン13の指向性特性を示す情報、および、情報取得部33から供給される収音対象情報に基づいて、収音対象21の音声を収音するのに最適な位置において、指向性マイクロフォン13の音軸が収音対象21に向かう方向を、収音対象21の音声を収音するのに最適な向きとして算出する。このように、最適配置算出部43は、最適な指向性マイクロフォン13の配置(位置または向き)を算出することができ、指向性マイクロフォン13の最適配置を示す情報を動作制御部32に供給することができる。 Then, the optimal placement calculation unit 43 calculates the position that is closest to the sound collection target 21 within the placement area of the directional microphone 13 as the optimal position for picking up the sound of the sound collection target 21. Furthermore, based on the information indicating the directional characteristics of the directional microphone 13 supplied from the user input unit 31 and the sound collection target information supplied from the information acquisition unit 33, the optimal placement calculation unit 43 calculates the direction in which the sound axis of the directional microphone 13 faces the sound collection target 21 at the optimal position for picking up the sound of the sound collection target 21 as the optimal orientation for picking up the sound of the sound collection target 21. In this way, the optimal placement calculation unit 43 can calculate the optimal placement (position or orientation) of the directional microphone 13, and can supply information indicating the optimal placement of the directional microphone 13 to the operation control unit 32.
 このように構成される撮影収音システム11は、最適な配置で指向性マイクロフォン13により収音対象21の音声を収音すること、例えば、収録可能範囲の内側、かつ、映り込み範囲の外側であって、収音対象21に最も近づくことができる位置で、収音対象21に音軸を向けて、指向性マイクロフォン13により収音対象21の音声を収音することができる。これにより、撮影収音システム11は、より良好に収音対象21の音声を収音することが可能となる。 The filming and recording system 11 configured in this manner can record the sound of the recording target 21 with the directional microphone 13 in an optimal position, for example, pointing the sound axis toward the recording target 21 at a position that is inside the recording range and outside the projection range and closest to the recording target 21, and can record the sound of the recording target 21 with the directional microphone 13. This allows the filming and recording system 11 to record the sound of the recording target 21 more effectively.
 また、撮影収音システム11では、ユーザが、撮影カメラ12による撮影が開始される前に指向性マイクロフォン13の位置または向きを指定し、必要に応じて、収音される音を確認しながら指向性マイクロフォン13の位置または向き調整することができる。例えば、ユーザは、収音対象21の口元や、おでこ、胸元などに向けて指向性マイクロフォン13の音軸を調整したり、より音声が聞き取りやすくなる距離となるように指向性マイクロフォン13の位置を調整したりすることができる。 Furthermore, in the filming and sound collection system 11, the user can specify the position or orientation of the directional microphone 13 before filming by the filming camera 12 begins, and adjust the position or orientation of the directional microphone 13 as necessary while checking the sound being collected. For example, the user can adjust the sound axis of the directional microphone 13 to point toward the mouth, forehead, chest, etc. of the sound collection target 21, or adjust the position of the directional microphone 13 so that it is at a distance where the sound can be more easily heard.
 そして、撮影収音システム11では、撮影カメラ12による撮影が開始される時点における指向性マイクロフォン13と収音対象21との位置関係を特定して記憶しておくことで、例えば、収音対象21が移動した際に、撮影カメラ12に映り込まない範囲で、この位置関係を保ったままで収音することが可能となる。つまり、記憶部35に記憶されている位置関係情報を読み出して、その位置関係を再現するように、動作制御部32が、移動装置14の動作を制御することができる。また、撮影収音システム11では、映画やドラマなどの撮影現場でリハーサルを含めて同じシーンの撮影を複数回行う場合や、撮影を中断した後に再開する場合など、指向性マイクロフォン13と収音対象21との位置関係を再現することができる。 Then, in the filming and sound collection system 11, by identifying and storing the positional relationship between the directional microphone 13 and the sound collection target 21 at the time when filming by the filming camera 12 begins, for example, when the sound collection target 21 moves, it is possible to collect sound while maintaining this positional relationship within a range where it is not captured by the filming camera 12. In other words, the operation control unit 32 can control the operation of the moving device 14 to read out the positional relationship information stored in the memory unit 35 and reproduce that positional relationship. In addition, in the filming and sound collection system 11, the positional relationship between the directional microphone 13 and the sound collection target 21 can be reproduced when the same scene is filmed multiple times, including rehearsals, on a filming site for a movie or drama, or when filming is resumed after being interrupted.
 また、撮影収音システム11では、情報取得部33が、収音対象21の位置および向きを特定する特定方法として、次の第1乃至第3の特定方法を用いることができる。 In addition, in the filming and recording system 11, the information acquisition unit 33 can use the following first to third identification methods to identify the position and orientation of the recording target 21.
 第1の特定方法では、事前に指定した収音対象21(役者など)に対してセンサを設置した場合には、そのセンサによって収音対象21の位置および向きを特定することができる。または、指向性マイクロフォン13とともに移動装置14にカメラが設置されている場合には、そのカメラにより撮影された映像を解析することによって、収音対象21の位置および向きを特定することができる。もちろん、センサによる収音対象21の特定と、カメラにより撮影された映像を解析することによる収音対象21の特定とを組み合わせてもよい。 In the first identification method, if a sensor is installed at a pre-specified sound collection target 21 (such as an actor), the position and orientation of the sound collection target 21 can be identified by the sensor. Alternatively, if a camera is installed on the mobile device 14 together with the directional microphone 13, the position and orientation of the sound collection target 21 can be identified by analyzing the video captured by the camera. Of course, identification of the sound collection target 21 by the sensor and identification of the sound collection target 21 by analyzing the video captured by the camera may be combined.
 第2の特定方法では、撮影カメラ12により撮影されている映像の中心にいる人物を、収音対象21として特定することができる。例えば、多数の人物の中から、少数の人物(例えば、ある時刻において発話する人物)を収音対象21とする場合、発話が行われるたびに発話者に指向性マイクロフォン13を向ける必要がある。従って、撮影カメラ12により撮影されている映像を解析して、その発話者を収音対象21として特定することができる。 In the second identification method, the person at the center of the image captured by the camera 12 can be identified as the sound collection target 21. For example, when a small number of people (e.g., people who speak at a certain time) are identified as the sound collection targets 21 from among many people, it is necessary to point the directional microphone 13 at the speaker every time they speak. Therefore, by analyzing the image captured by the camera 12, the speaker can be identified as the sound collection target 21.
 第3の特定方法では、例えば、複数のマイクロフォンやアレイ状のマイクロフォンにより収音された発話に基づいて発話者と、その位置を特定し、その発話を行った発話者を収音対象21とすることができる。 In the third identification method, for example, the speaker and their position are identified based on speech picked up by multiple microphones or a microphone array, and the speaker who made the speech can be set as the pickup target 21.
 また、撮影収音システム11では、収音対象21の音響放射特性を示す情報を用いて、指向性マイクロフォン13の収音位置を決定することができる。例えば、収音対象21に音響放射特性がある場合、その向きから収音されるべきかの情報を、ユーザ入力部31を用いてユーザが事前に入力しておくことができ、その情報に応じて、指向性マイクロフォン13の位置または向きを決定する。特に、楽器などのように収音位置のノウハウがある収音対象21に対して、どの辺りをどの程度の距離から収音するのが良いか最善の情報を予め保持しておくことが好ましい。なお、収音対象21に音響放射特性がない場合、即ち、全方向に音が放射されるとみなせる場合、指向性マイクロフォン13の位置または向きを算出する際に、収音対象21の向きを考慮しなくてもよい。 In addition, the filming and recording system 11 can determine the sound collection position of the directional microphone 13 using information indicating the acoustic radiation characteristics of the sound collection target 21. For example, if the sound collection target 21 has acoustic radiation characteristics, the user can input information regarding the direction from which sound should be collected in advance using the user input unit 31, and the position or orientation of the directional microphone 13 is determined based on that information. In particular, for a sound collection target 21 such as a musical instrument for which there is know-how regarding the sound collection position, it is preferable to store in advance the best information regarding where and from what distance sound should be collected. Note that if the sound collection target 21 does not have acoustic radiation characteristics, i.e., if sound can be considered to be radiated in all directions, the orientation of the sound collection target 21 does not need to be taken into consideration when calculating the position or orientation of the directional microphone 13.
 また、撮影収音システム11では、収音対象21の音声を収音したときの情報を参照し、収音済みの音声に対するポストプロセスを自動で行うことができる。例えば、撮影収音システム11では、収音済みの音声に対するポストプロセスにおいて、音声信号に含まれている移動装置14のプロペラ騒音を除去したい場合、移動装置14のプロペラの回転数を示す情報を保存しておく。そして、ポストプロセスにおいてその情報に従ってプロペラ騒音の周波数を特定し、移動装置14のプロペラ騒音を除去することができる。また、撮影収音システム11では、例えば、収音済みの音声に対するポストプロセスにおいて、収音対象21との位置関係や、指向性マイクロフォン13の指向性特性、収音対象21の放射特性に応じて、音量および周波数の特性の調整を行うことができる。 In addition, the filming and sound collection system 11 can automatically perform post-processing on the collected sound by referring to information obtained when the sound of the sound collection target 21 was collected. For example, in the filming and sound collection system 11, if it is desired to remove the propeller noise of the moving device 14 contained in the sound signal in the post-processing of the collected sound, information indicating the rotation speed of the propeller of the moving device 14 is stored. Then, in the post-processing, the frequency of the propeller noise can be identified according to that information, and the propeller noise of the moving device 14 can be removed. In addition, in the filming and sound collection system 11, for example, in the post-processing of the collected sound, the volume and frequency characteristics can be adjusted according to the positional relationship with the sound collection target 21, the directional characteristics of the directional microphone 13, and the radiation characteristics of the sound collection target 21.
 図3に示すフローチャートを参照して、撮影収音システム11において実行される撮影収音処理の第1の処理例について説明する。 With reference to the flowchart shown in FIG. 3, a first processing example of the imaging and sound recording process executed in the imaging and sound recording system 11 will be described.
 ステップS11において、ユーザは、ユーザ入力部31を用いて、撮影カメラ12の画角を示す情報、指向性マイクロフォン13の形状および指向性特性を示す情報、並びに、移動装置14の形状を示す情報を含む機材情報を入力する。これに応じて、ユーザ入力部31は、ユーザにより入力された機材情報を演算処理部36に供給する。 In step S11, the user uses the user input unit 31 to input equipment information including information indicating the angle of view of the imaging camera 12, information indicating the shape and directional characteristics of the directional microphone 13, and information indicating the shape of the mobile device 14. In response, the user input unit 31 supplies the equipment information input by the user to the calculation processing unit 36.
 ステップS12において、情報取得部33は、撮影カメラ12による撮影が開始される前の撮影カメラ12の位置および向きを示す初期カメラ情報を取得して、演算処理部36に供給する。 In step S12, the information acquisition unit 33 acquires initial camera information indicating the position and orientation of the imaging camera 12 before imaging by the imaging camera 12 begins, and supplies this information to the calculation processing unit 36.
 ステップS13において、ユーザは、ユーザ入力部31を用いて、収音対象21ごとに音声の収音に使用する指向性マイクロフォン13を指定する。これに応じて、ユーザ入力部31は、収音対象21ごと指定された指向性マイクロフォン13を示す情報を演算処理部36に供給する。そして、ステップS14乃至S17では、ステップS13で指向性マイクロフォン13が指定された収音対象21ごとに撮影前の処理が行われる。 In step S13, the user uses the user input unit 31 to specify the directional microphone 13 to be used to collect sound for each sound collection target 21. In response to this, the user input unit 31 supplies information indicating the directional microphone 13 specified for each sound collection target 21 to the calculation processing unit 36. Then, in steps S14 to S17, pre-imaging processing is performed for each sound collection target 21 for which a directional microphone 13 was specified in step S13.
 ステップS14において、情報取得部33は、撮影カメラ12による撮影が開始される前の収音対象21の位置および向きを示す初期収音対象情報を取得する。さらに、情報取得部33は、ステップS13で指定された指向性マイクロフォン13について、撮影カメラ12による撮影が開始される前の指向性マイクロフォン13の位置または向きを示す初期マイク情報を取得する。 In step S14, the information acquisition unit 33 acquires initial sound collection target information indicating the position and orientation of the sound collection target 21 before shooting by the shooting camera 12 begins. Furthermore, for the directional microphone 13 specified in step S13, the information acquisition unit 33 acquires initial microphone information indicating the position or orientation of the directional microphone 13 before shooting by the shooting camera 12 begins.
 ステップS15において、ユーザは、ユーザ入力部31を用いて、指向性マイクロフォン13の位置または向きを指定する。これに応じて、ユーザ入力部31は、ユーザにより指定された指向性マイクロフォン13の位置または向きを示す情報を動作制御部32に供給し、動作制御部32は、その情報に基づいて移動装置14の動作を制御する。これにより、ユーザにより指定された位置または向きとなるように、指向性マイクロフォン13が配置される。 In step S15, the user specifies the position or orientation of the directional microphone 13 using the user input unit 31. In response, the user input unit 31 supplies information indicating the position or orientation of the directional microphone 13 specified by the user to the operation control unit 32, and the operation control unit 32 controls the operation of the mobile device 14 based on that information. As a result, the directional microphone 13 is positioned so that it is in the position or orientation specified by the user.
 ステップS16において、ユーザは、ステップS15で配置された指向性マイクロフォン13により収音された音を確認しながら、ユーザ入力部31を用いて、指向性マイクロフォン13の位置または向きを調整する。これに応じて、ユーザ入力部31は、ユーザにより調整された指向性マイクロフォン13の位置または向きを示す情報を動作制御部32に供給し、動作制御部32は、その情報に基づいて移動装置14の動作を制御する。これにより、ユーザにより調整された位置または向きとなるように、指向性マイクロフォン13が配置される。このように、ユーザが、指向性マイクロフォン13の配置の調整と、指向性マイクロフォン13により収音された音の確認とを行うことで、撮影カメラ12による撮影が開始される前の指向性マイクロフォン13の配置が決定される。 In step S16, the user adjusts the position or orientation of the directional microphone 13 using the user input unit 31 while checking the sound picked up by the directional microphone 13 placed in step S15. In response, the user input unit 31 supplies information indicating the position or orientation of the directional microphone 13 adjusted by the user to the operation control unit 32, and the operation control unit 32 controls the operation of the mobile device 14 based on that information. In this way, the directional microphone 13 is placed so that it is in the position or orientation adjusted by the user. In this way, the user adjusts the placement of the directional microphone 13 and checks the sound picked up by the directional microphone 13, and the placement of the directional microphone 13 before the shooting camera 12 starts shooting is determined.
 ステップS17において、情報取得部33は、ステップS17で決定された配置における初期マイク情報を取得して、位置関係特定部34に供給する。そして、位置関係特定部34は、ステップS17で決定された配置における初期マイク情報とステップS14で供給された初期収音対象情報とに基づいて、撮影カメラ12による撮影が開始される時点における指向性マイクロフォン13と収音対象21との位置関係を特定して、その位置関係を示す位置関係情報を記憶部35に記憶させる。 In step S17, the information acquisition unit 33 acquires initial microphone information for the arrangement determined in step S17 and supplies it to the positional relationship determination unit 34. Then, based on the initial microphone information for the arrangement determined in step S17 and the initial sound collection target information supplied in step S14, the positional relationship determination unit 34 determines the positional relationship between the directional microphone 13 and the sound collection target 21 at the time when shooting by the shooting camera 12 begins, and stores the positional relationship information indicating this positional relationship in the storage unit 35.
 ステップS18において、全ての収音対象21の位置関係が記憶部35に記憶されたか否かが判定され、全ての収音対象21の位置関係が記憶部35に記憶されていないと判定された場合、処理はステップS13に戻る。即ち、この場合、ステップS13において次の収音対象21の音声の収音に使用する指向性マイクロフォン13を指定し、以下、同様の処理が繰り返して行われる。 In step S18, it is determined whether the positional relationships of all sound collection targets 21 have been stored in the storage unit 35, and if it is determined that the positional relationships of all sound collection targets 21 have not been stored in the storage unit 35, the process returns to step S13. That is, in this case, in step S13, the directional microphone 13 to be used for picking up the sound of the next sound collection target 21 is specified, and the same process is repeated thereafter.
 一方、ステップS18において、全ての収音対象21の位置関係が記憶部35に記憶されたと判定された場合、処理はステップS19に進む。 On the other hand, if it is determined in step S18 that the positional relationships of all sound collection targets 21 have been stored in the memory unit 35, the process proceeds to step S19.
 ステップS19において、撮影カメラ12が撮影を開始し、指向性マイクロフォン13が収音を開始する。 In step S19, the camera 12 starts capturing images, and the directional microphone 13 starts collecting sound.
 ステップS20において、情報取得部33は、撮影カメラ12、収音対象21、および指向性マイクロフォン13それぞれの位置または向きを示す情報を取得し、カメラ情報、収音対象情報、およびマイク情報を、演算処理部36に供給する。 In step S20, the information acquisition unit 33 acquires information indicating the position or orientation of each of the shooting camera 12, the sound collection target 21, and the directional microphone 13, and supplies the camera information, sound collection target information, and microphone information to the calculation processing unit 36.
 ステップS21において、演算処理部36は、ステップS20で情報取得部33から供給されたカメラ情報、収音対象情報、およびマイク情報に基づいて、最適な指向性マイクロフォン13の位置または向きを算出する。そして、演算処理部36は、その位置または向きで指向性マイクロフォン13が配置されるように指示する指向性マイクロフォン13の最適配置を示す情報を動作制御部32に供給する。 In step S21, the calculation processing unit 36 calculates the optimal position or orientation of the directional microphone 13 based on the camera information, sound collection target information, and microphone information supplied from the information acquisition unit 33 in step S20. The calculation processing unit 36 then supplies information indicating the optimal placement of the directional microphone 13 to the operation control unit 32, which instructs the directional microphone 13 to be placed in that position or orientation.
 ステップS22において、動作制御部32は、ステップS21で演算処理部36から供給された指向性マイクロフォン13の最適配置を示す情報に基づいて移動装置14の動作を制御する。これにより、収音対象21の音声を収音するのに最適な位置または向きとなるように指向性マイクロフォン13が配置される。 In step S22, the operation control unit 32 controls the operation of the moving device 14 based on the information indicating the optimal placement of the directional microphone 13 supplied from the calculation processing unit 36 in step S21. As a result, the directional microphone 13 is placed in an optimal position or orientation for picking up the sound of the sound collection target 21.
 ステップS23において、撮影カメラ12による撮影を終了するか否かが判定され、撮影カメラ12による撮影を終了しないと判定された場合、処理はステップS20に戻り、以下、同様の処理が繰り返して行われる。一方、ステップS23において、撮影カメラ12による撮影を終了すると判定された場合、処理は終了される。 In step S23, it is determined whether or not to end image capture by the imaging camera 12. If it is determined that image capture by the imaging camera 12 should not be ended, the process returns to step S20, and the same process is repeated thereafter. On the other hand, if it is determined in step S23 that image capture by the imaging camera 12 should be ended, the process ends.
 以上のように、撮影収音システム11では、撮影収音処理を実行することによって、収音対象21の音声を収音するのに最適な位置または向きとなるように指向性マイクロフォン13を配置して、より良好に音声を収音することができる。 As described above, in the filming and sound collection system 11, by performing the filming and sound collection process, the directional microphone 13 is positioned in the optimal position or orientation for collecting the sound of the sound collection target 21, thereby enabling better sound collection.
 図4に示すフローチャートを参照して、撮影収音システム11において実行される撮影収音処理の第2の処理例について説明する。 A second example of the imaging and sound recording process executed in the imaging and sound recording system 11 will be described with reference to the flowchart shown in FIG. 4.
 ステップS31乃至S34において、図3のステップS11乃至S14と同様の処理が行われ、ステップS35において、演算処理部36は、初期カメラ情報、初期収音対象情報、および初期マイク情報に基づいて、最適な指向性マイクロフォン13の位置または向きを算出し、その位置または向きを示す情報を動作制御部32に供給する。 In steps S31 to S34, the same processing as in steps S11 to S14 in FIG. 3 is performed, and in step S35, the calculation processing unit 36 calculates the optimal position or orientation of the directional microphone 13 based on the initial camera information, the initial sound collection target information, and the initial microphone information, and supplies information indicating that position or orientation to the operation control unit 32.
 ステップS35において、動作制御部32は、ステップS34で演算処理部36から供給された指向性マイクロフォン13の位置または向きを示す情報に基づいて、移動装置14の動作を制御する。これにより、ステップS34で算出された最適な位置または向きとなるように、指向性マイクロフォン13が配置される。その後、ステップS36乃至S43において、図3のステップS16乃至S23と同様の処理が行われる。 In step S35, the operation control unit 32 controls the operation of the moving device 14 based on the information indicating the position or orientation of the directional microphone 13 supplied from the calculation processing unit 36 in step S34. As a result, the directional microphone 13 is positioned so that it is in the optimal position or orientation calculated in step S34. After that, in steps S36 to S43, the same processing as in steps S16 to S23 in FIG. 3 is performed.
 以上のように、撮影収音システム11では、撮影収音処理を実行することによって、撮影を開始する時点から、収音対象21の音声を収音するのに最適な位置または向きとなるように指向性マイクロフォン13を配置して、より良好に音声を収音することができる。 As described above, the filming and sound collection system 11 performs filming and sound collection processing, and from the time filming begins, the directional microphone 13 is positioned in the optimal position or orientation for collecting the sound of the sound collection target 21, allowing for better sound collection.
 <考慮対象について>
 図5乃至図7を参照して、撮影カメラ12や、指向性マイクロフォン13、収音対象21などの位置または向きの他に、撮影現場における様々なことを考慮し、指向性マイクロフォン13の最適な位置または向きを決定する方法について説明する。
<About what is to be considered>
5 to 7, a method for determining the optimal position or orientation of the directional microphone 13 will be described, taking into consideration various factors at the shooting site in addition to the positions and orientations of the shooting camera 12, directional microphone 13, sound collection target 21, etc.
 図5に示すように、撮影収音システム11では、騒音源や、照明機材、反射物・映り込む物体、太陽の位置、周囲の状況などを考慮して、それらの影響を排除するように、指向性マイクロフォン13の最適な位置または向きを決定する必要がある。 As shown in Figure 5, in the filming and recording system 11, it is necessary to determine the optimal position or orientation of the directional microphone 13 in consideration of noise sources, lighting equipment, reflective objects/reflected objects, the position of the sun, surrounding conditions, etc., in order to eliminate their effects.
 騒音源を考慮する場合、収音対象21以外の騒音を収音してしまうことを回避するように、指向性マイクロフォン13の位置または向きを決定する必要がある。従って、この条件を満たすように、撮影収音システム11では、収音対象21以外の音の発生位置、および、指向性マイクロフォン13の指向性特性が、追加の入力情報として情報処理装置15に入力される。 When considering the noise source, it is necessary to determine the position or orientation of the directional microphone 13 so as to avoid picking up noise other than the target 21. Therefore, in order to meet this condition, in the filming and recording system 11, the generation position of sounds other than the target 21 and the directional characteristics of the directional microphone 13 are input to the information processing device 15 as additional input information.
 照明機材を考慮する場合、指向性マイクロフォン13が照明光を遮らないように、かつ、指向性マイクロフォン13の影が映り込まないように、指向性マイクロフォン13の位置または向きを決定する必要がある。従って、この条件を満たすように、撮影収音システム11では、照明機材の位置、向き、およびサイズ、並びに、照明光の指向特性および光量が、追加の入力情報として情報処理装置15に入力される。 When considering lighting equipment, the position or orientation of the directional microphone 13 must be determined so that the directional microphone 13 does not block the lighting light and so that the shadow of the directional microphone 13 is not cast. Therefore, in order to satisfy this condition, in the filming and recording system 11, the position, orientation, and size of the lighting equipment, as well as the directional characteristics and light amount of the lighting light, are input to the information processing device 15 as additional input information.
 ガラスや鏡などのような反射物・映り込む物体を考慮する場合、そのような物体に指向性マイクロフォン13が映り込まないように、指向性マイクロフォン13の位置または向きを決定する必要がある。従って、この条件を満たすように、撮影収音システム11では、反射物・映り込む物体の位置、向き、形状、およびサイズ、並びに、反射率が、追加の入力情報として情報処理装置15に入力される。 When considering reflective objects such as glass and mirrors, it is necessary to determine the position or orientation of the directional microphone 13 so that the directional microphone 13 is not reflected in such objects. Therefore, in order to satisfy this condition, in the filming and recording system 11, the position, orientation, shape, and size of the reflective object and the reflected object, as well as the reflectance, are input to the information processing device 15 as additional input information.
 太陽の位置を考慮する場合、指向性マイクロフォン13の影が映り込まないように、指向性マイクロフォン13の位置または向きを決定する必要がある。従って、この条件を満たすように、撮影収音システム11では、時刻、緯度・経度、および天候が、追加の入力情報として情報処理装置15に入力される。 When taking into account the position of the sun, it is necessary to determine the position or orientation of the directional microphone 13 so that the shadow of the directional microphone 13 is not cast. Therefore, in order to satisfy this condition, in the filming and recording system 11, the time, latitude, longitude, and weather are input to the information processing device 15 as additional input information.
 周囲の状況を考慮する場合、指向性マイクロフォン13の周囲にある人物や物体と接触しないように、指向性マイクロフォン13の位置または向きを決定する必要がある。従って、この条件を満たすように、周辺の状況を観察するカメラやセンサなどの出力が、追加の入力情報として情報処理装置15に入力される。 When taking the surrounding conditions into consideration, it is necessary to determine the position or orientation of the directional microphone 13 so as not to come into contact with people or objects around the directional microphone 13. Therefore, to meet this condition, the output of a camera or sensor that observes the surrounding conditions is input to the information processing device 15 as additional input information.
 図6は、騒音源を考慮する場合における指向性マイクロフォン13の最適な位置または向きの一例について説明する図である。 Figure 6 is a diagram that explains an example of the optimal position or orientation of the directional microphone 13 when taking into account the noise source.
 図6に示すように、撮影収音システム11を利用した撮影現場に騒音源22がある場合、指向性マイクロフォン13が騒音源22の騒音を収音してしまうことを回避する必要がある。例えば、騒音源22の騒音としては、空調、冷蔵庫、換気扇、または発電機のような各種の機器の騒音や、航空機または緊急自動車などのような交通騒音、工事騒音などが想定される。 As shown in FIG. 6, if there is a noise source 22 at a shooting location where the shooting and sound collection system 11 is used, it is necessary to prevent the directional microphone 13 from picking up the noise from the noise source 22. For example, noise from the noise source 22 may be noise from various devices such as air conditioners, refrigerators, ventilation fans, or generators, traffic noise from aircraft or emergency vehicles, construction noise, etc.
 そして、最適配置算出部43は、上述したように最適な指向性マイクロフォン13の配置を算出する際に、騒音源22の位置、および、指向性マイクロフォン13の指向性特性に基づいて、指向性マイクロフォン13の音軸が騒音源22を向いてしまうことを回避するように、最適な指向性マイクロフォン13の配置を算出する。つまり、最適配置算出部43は、指向性マイクロフォン13の配置可能エリア内において収音対象21に最も近づくことができる位置で、指向性マイクロフォン13の音軸が騒音源22を向かずに収音対象21に向かうように、最適な指向性マイクロフォン13の配置を算出する。 Then, when calculating the optimal placement of the directional microphone 13 as described above, the optimal placement calculation unit 43 calculates the optimal placement of the directional microphone 13 based on the position of the noise source 22 and the directional characteristics of the directional microphone 13 so as to avoid the sound axis of the directional microphone 13 pointing towards the noise source 22. In other words, the optimal placement calculation unit 43 calculates the optimal placement of the directional microphone 13 at a position that is closest to the sound collection target 21 within the placement area of the directional microphone 13 so that the sound axis of the directional microphone 13 points towards the sound collection target 21 and not towards the noise source 22.
 従って、図6のAに示すように、映り込み範囲の外側の位置であって、指向性マイクロフォン13の音軸が騒音源22を向かずに収音対象21に向かうよう向きとなるように指向性マイクロフォン13を配置することが好適である。なお、図6のBには、映り込み範囲の外側の位置であって、指向性マイクロフォン13の音軸が収音対象21に向かうよう向きとなっているが、指向性マイクロフォン13の音軸が騒音源22にも向いてしまっている好ましくない例が示されている。 Therefore, as shown in FIG. 6A, it is preferable to position the directional microphone 13 outside the reflection range so that its sound axis faces the sound collection target 21 and not the noise source 22. Note that FIG. 6B shows an undesirable example in which the directional microphone 13 is located outside the reflection range and its sound axis faces the sound collection target 21, but the sound axis of the directional microphone 13 also faces the noise source 22.
 なお、撮影収音システム11では、例えば、収音された信号を解析することによって、収音対象21の音声以外の騒音が含まれている場合には、その旨を通知するアラートを出力してもよい。 In addition, the filming and recording system 11 may, for example, analyze the recorded signal and output an alert to notify the user if the signal contains noise other than the voice of the recording target 21.
 図7は、照明機材を考慮する場合における指向性マイクロフォン13の最適な位置または向きの一例について説明する図である。 Figure 7 is a diagram illustrating an example of the optimal position or orientation of the directional microphone 13 when taking lighting equipment into consideration.
 図7に示すように、撮影収音システム11を利用した撮影現場に照明機材16がある場合、指向性マイクロフォン13が照明光を遮ること、および、指向性マイクロフォン13の影が映り込むことを回避する必要がある。 As shown in Figure 7, if lighting equipment 16 is present at a shooting location using the filming and sound recording system 11, it is necessary to prevent the directional microphone 13 from blocking the illumination light and to prevent the shadow of the directional microphone 13 from being cast.
 そして、最適配置算出部43は、上述したように最適な指向性マイクロフォン13の配置を算出する際に、照明機材16の位置、向き、サイズ、照明光の指向特性、および光量に基づいて、指向性マイクロフォン13が照明光を遮ること、および、指向性マイクロフォン13の影が映り込むことを回避するように、最適な指向性マイクロフォン13の配置を算出する。つまり、最適配置算出部43は、指向性マイクロフォン13の配置可能エリア内において収音対象21に最も近づくことができる位置で、指向性マイクロフォン13が照明機材16により照明光が照射される領域に入り込まないように、最適な指向性マイクロフォン13の配置を算出する。 Then, when calculating the optimal placement of the directional microphone 13 as described above, the optimal placement calculation unit 43 calculates the optimal placement of the directional microphone 13 based on the position, orientation, and size of the lighting equipment 16, the directional characteristics of the lighting light, and the amount of light, so as to prevent the directional microphone 13 from blocking the lighting light and to prevent the shadow of the directional microphone 13 from being cast. In other words, the optimal placement calculation unit 43 calculates the optimal placement of the directional microphone 13 at a position that is closest to the sound collection target 21 within the placement area of the directional microphone 13, and so as to prevent the directional microphone 13 from entering the area where the lighting equipment 16 irradiates the lighting light.
 従って、図7のAに示すように、映り込み範囲の外側の位置、かつ、照明光が照射される領域に入り込まない位置であって、指向性マイクロフォン13の音軸が収音対象21に向かうよう向きとなるように指向性マイクロフォン13を配置することが好適である。なお、図7のBには、映り込み範囲の外側の位置ではあるが、照明光が照射される領域に入り込んでしまう位置に指向性マイクロフォン13が配置されてしまっている好ましくない例が示されている。 Therefore, as shown in A of FIG. 7, it is preferable to place the directional microphone 13 in a position outside the reflection range and not in the area irradiated by the illumination light, so that the sound axis of the directional microphone 13 is oriented toward the sound collection target 21. Note that B of FIG. 7 shows an undesirable example in which the directional microphone 13 is placed in a position outside the reflection range but in the area irradiated by the illumination light.
 <複数の収音対象について>
 図8乃至図11を参照して、複数の収音対象21の音声を収音する方法について説明する。
<About multiple recording targets>
A method for collecting sounds from a plurality of sound collection targets 21 will be described with reference to Figs.
 図8には、複数の収音対象21が近くで向かい合って同時に会話をしている状況において、1つの指向性マイクロフォン13を使用する使用例が示されている。 Figure 8 shows an example of using one directional microphone 13 in a situation where multiple targets 21 for recording are facing each other and conversing simultaneously.
 図8に示すように、収音対象21Aおよび収音対象21Bが近くで向かい合って同時に会話をしているとき、収音対象21Aおよび収音対象21Bの中間地点に、指向性マイクロフォン13の音軸を向けることが好ましい。同様に、2人以上の複数の収音対象21が近くに居て、1つの指向性マイクロフォン13を使用する場合には、それらの複数の収音対象21の中間地点に、指向性マイクロフォン13の音軸を向けることが好ましい。これにより、1つの指向性マイクロフォン13で複数の収音対象21の音声を良好に収音することができる。 As shown in FIG. 8, when sound collection targets 21A and 21B are facing each other and conversing at the same time, it is preferable to point the sound axis of the directional microphone 13 at the midpoint between sound collection targets 21A and 21B. Similarly, when two or more sound collection targets 21 are nearby and one directional microphone 13 is used, it is preferable to point the sound axis of the directional microphone 13 at the midpoint between the multiple sound collection targets 21. This allows the single directional microphone 13 to effectively pick up the sounds of the multiple sound collection targets 21.
 図9には、複数の収音対象21が近くで交互に会話をしている状況において、1つの指向性マイクロフォン13を使用する使用例が示されている。 Figure 9 shows an example of using one directional microphone 13 in a situation where multiple recording targets 21 are nearby and taking turns conversing.
 図9に示すように、収音対象21Aおよび収音対象21Bが近くで交互に会話をしているとき、収音対象21Aおよび収音対象21Bそれぞれが発話する時刻に応じて指向性マイクロフォン13の音軸の向きを変更することが好ましい。例えば、事前に用意されているシナリオに従って、収音対象21Aが発話する時刻t1では、収音対象21Aに対して指向性マイクロフォン13の音軸を向け、収音対象21Bが発話する時刻t2では、収音対象21Bに対して指向性マイクロフォン13の音軸を向けることができる。これにより、1つの指向性マイクロフォン13で複数の収音対象21の音声を良好に収音することができる。 As shown in FIG. 9, when sound collection targets 21A and 21B are nearby and alternately conversing, it is preferable to change the direction of the sound axis of directional microphone 13 according to the time when sound collection target 21A and sound collection target 21B each speak. For example, according to a scenario prepared in advance, the sound axis of directional microphone 13 can be directed toward sound collection target 21A at time t1 when sound collection target 21A speaks, and toward sound collection target 21B at time t2 when sound collection target 21B speaks. This allows one directional microphone 13 to pick up the voices of multiple sound collection targets 21 well.
 図10には、1つの無指向性マイクロフォン17を使用する使用例が示されている。 Figure 10 shows an example of use of one omnidirectional microphone 17.
 図10に示すように、無指向性マイクロフォン17を移動装置14に搭載する場合、複数の収音対象21どうしで等距離となる位置に無指向性マイクロフォン17を配置することが好ましい。これにより、1つの無指向性マイクロフォン17で複数の収音対象21の音声を良好に収音することができる。 As shown in FIG. 10, when an omnidirectional microphone 17 is mounted on a mobile device 14, it is preferable to place the omnidirectional microphone 17 at a position equidistant to multiple sound collection targets 21. This allows one omnidirectional microphone 17 to effectively pick up the sounds of multiple sound collection targets 21.
 図11には、複数の収音対象21それぞれに対して複数の指向性マイクロフォン13を使用する使用例が示されている。 FIG. 11 shows an example of using multiple directional microphones 13 for each of multiple sound collection targets 21.
 図11に示すように、2つの指向性マイクロフォン13Aおよび指向性マイクロフォン13Bを移動装置14に搭載し、収音対象21Aと収音対象21Bとの中間となる位置に移動装置14を配置させる。そして、収音対象21Aに対して指向性マイクロフォン13Aの音軸を向け、収音対象21Bに対して指向性マイクロフォン13Bの音軸を向けることが好ましい。これにより、複数の指向性マイクロフォン13で複数の収音対象21の音声を良好に収音することができる。 As shown in FIG. 11, two directional microphones 13A and 13B are mounted on a mobile device 14, and the mobile device 14 is positioned halfway between sound collection targets 21A and 21B. It is preferable to orient the sound axis of directional microphone 13A toward sound collection target 21A, and the sound axis of directional microphone 13B toward sound collection target 21B. This allows the multiple directional microphones 13 to effectively pick up the sounds of multiple sound collection targets 21.
 <配置画面について>
 図12および図13を参照して、情報処理装置15の表示部に表示される配置画面について説明する。
<About the placement screen>
The layout screen displayed on the display unit of the information processing device 15 will be described with reference to FIGS.
 撮影収音システム11では、撮影現場の状況を上空から見下ろしたように視覚化して、撮影カメラ12や、指向性マイクロフォン13、照明機材16、収音対象21などの配置を示す配置画面51を情報処理装置15の表示部に表示することができる。なお、撮影現場の状況を側方から可視化したり、撮影現場の状況を立体的な3次元で可視化したりして配置画面51に表示することができる。 The filming and sound recording system 11 can visualize the situation at the filming site as if looking down from the sky, and display an arrangement screen 51 showing the arrangement of the filming cameras 12, directional microphones 13, lighting equipment 16, sound recording targets 21, etc. on the display unit of the information processing device 15. The situation at the filming site can be visualized from the side, or in a stereoscopic three-dimensional form, and displayed on the arrangement screen 51.
 例えば、指向性マイクロフォン13は、収音対象21に指向性マイクロフォン13の音軸を向けた際に収音対象21の直接音を収録することができる範囲となる収録可能範囲のエリア内に配置する必要がある。また、指向性マイクロフォン13は、撮影カメラ12により撮影される映像に映り込む範囲となる映り込み範囲のエリア外に配置する必要がある。さらに、指向性マイクロフォン13は、照明光を遮らないように、かつ、影が映り込まないように配置する必要がある。 For example, the directional microphone 13 needs to be placed within an area of a recording range that is a range in which the direct sound of the sound collection target 21 can be recorded when the sound axis of the directional microphone 13 is directed toward the sound collection target 21. Also, the directional microphone 13 needs to be placed outside an area of a reflection range that is a range in which the microphone appears in the image captured by the shooting camera 12. Furthermore, the directional microphone 13 needs to be placed so as not to block the illumination light and so as not to cast a shadow.
 そこで、図12に示すように、配置画面51には、指向性マイクロフォン13の配置位置として適切なエリア、撮影カメラ12により撮影される映像に映り込むためNGとなるエリア、照明機材16を妨げる・影をつくるためNGとなるエリア、収音対象21から遠いためNGとなるエリア、ガラスなどに反射したものが撮影カメラ12により撮影される映像に映り込むためNGとなるエリアが表示される。さらに、配置画面51には、撮影カメラ12や、指向性マイクロフォン13、照明機材16、収音対象21などを表すアイコンが表示される。そして、ユーザは、配置画面51を参照することで、指向性マイクロフォン13の配置が適切であるか否かを認識することができる。 As shown in FIG. 12, the arrangement screen 51 displays areas that are appropriate for placing the directional microphone 13, areas that are not acceptable because they will be reflected in the image captured by the imaging camera 12, areas that are not acceptable because they will obstruct or cast a shadow on the lighting equipment 16, areas that are not acceptable because they are far from the sound collection target 21, and areas that are not acceptable because reflections from glass or the like will be reflected in the image captured by the imaging camera 12. Furthermore, the arrangement screen 51 displays icons representing the imaging camera 12, the directional microphone 13, the lighting equipment 16, the sound collection target 21, etc. The user can recognize whether the arrangement of the directional microphone 13 is appropriate by referring to the arrangement screen 51.
 例えば、図12のAに示す配置画面51は、指向性マイクロフォン13の配置位置として適切なエリアに指向性マイクロフォン13のアイコンが表示されているOKの事例が示されており、最適な指向性マイクロフォン13の配置となっていることを表している。 For example, the placement screen 51 shown in FIG. 12A shows an OK example in which the icon of the directional microphone 13 is displayed in an area appropriate for the placement of the directional microphone 13, indicating that the placement of the directional microphone 13 is optimal.
 一方、図12のBに示す配置画面51は、撮影カメラ12により撮影される映像に映り込むためNGとなるエリアに、指向性マイクロフォン13のアイコンが表示されているNGの事例が示されている。そして、撮影カメラ12に指向性マイクロフォン13が映り込むためNGであるというメッセージが表示されている。例えば、撮影収音システム11では、ユーザが指向性マイクロフォン13または移動装置14の位置を指定した場合に、その位置が適切でない場合には、その原因を通知するメッセージを表示するとともにアラートを出力することができる。 On the other hand, the arrangement screen 51 shown in FIG. 12B shows an example of an NG situation in which the icon of a directional microphone 13 is displayed in an area that is NG because it will be reflected in the image captured by the filming camera 12. A message is displayed to inform the user that the directional microphone 13 is NG because it will be reflected in the filming camera 12. For example, in the filming and sound collection system 11, if the user specifies the position of the directional microphone 13 or mobile device 14 and the position is not appropriate, a message informing the user of the cause can be displayed and an alert can be output.
 また、撮影収音システム11では、図13の上側に示す配置画面51のように、指向性マイクロフォン13の配置位置として適切なエリアがない場合には、指向性マイクロフォン13の配置位置として適切なエリアがないためにNGであるというメッセージが表示される。そして、この場合、撮影収音システム11では、どの条件を緩和した場合に、指向性マイクロフォン13の配置位置として適切なエリアが得られるかという代替案を提示することとができる。 In addition, in the filming and recording system 11, as in the arrangement screen 51 shown in the upper part of FIG. 13, if there is no suitable area for the arrangement of the directional microphone 13, a message is displayed indicating that it is NG because there is no suitable area for the arrangement of the directional microphone 13. In this case, the filming and recording system 11 can present alternatives indicating which conditions should be relaxed to obtain a suitable area for the arrangement of the directional microphone 13.
 即ち、図13の下側に示す配置画面51のように、照明機材16の位置を移動した場合に得られる指向性マイクロフォン13の配置位置として適切なエリアに、指向性マイクロフォン13のアイコンを表示する。 In other words, as shown in the layout screen 51 at the bottom of Figure 13, an icon of the directional microphone 13 is displayed in an area that is appropriate as the layout position of the directional microphone 13 obtained when the position of the lighting equipment 16 is moved.
 または、ユーザが、配置画面51上で、それらのオブジェクトの位置または向きを変更して、その際に、指向性マイクロフォン13の配置位置として適切なエリアが存在するかどうかの確認が可能なユーザインタフェースを表示してもよい。 Alternatively, a user interface may be displayed that allows the user to change the position or orientation of those objects on the placement screen 51 and check whether an appropriate area exists for placing the directional microphone 13.
 なお、撮影収音システム11では、移動中の移動装置14が他の物体や人物などと接触しそうになった場合、撮影カメラ12の映り込み範囲に入り込むことになるか否かに関わらず、他の物体や人物などとの接触を回避するような位置または向きに移動装置14を移動させることを優先するように、動作制御部32による制御が行われる。このように、他の物体や人物などとの接触を回避させることを優先した場合には、指向性マイクロフォン13の配置位置が想定とは異なっていることが、配置画面51によりユーザへ通知される。 In addition, in the filming and sound recording system 11, when the moving mobile device 14 is about to come into contact with another object or person, the operation control unit 32 controls the moving device 14 to move to a position or orientation that will avoid contact with the other object or person, regardless of whether the device will enter the range of the filming camera 12. In this way, when avoiding contact with other objects or people is given priority, the arrangement screen 51 notifies the user that the arrangement position of the directional microphone 13 is different from that expected.
 <マイクケーブルを利用した構成例>
 図14を参照して、マイクケーブルを利用した構成例について説明する。
<Example of configuration using a microphone cable>
An example of a configuration using a microphone cable will be described with reference to FIG.
 図14に示すように、撮影収音システム11は、マイクケーブル18を利用した構成とすることができ、音声を録音する録音システム19に、マイクケーブル18を介して指向性マイクロフォン13が接続される。つまり、指向性マイクロフォン13は、ワイヤレスで音声信号を録音システム19に送信することもできるし、マイクケーブル18を介して有線で音声信号を録音システム19に送信することもできる。 As shown in FIG. 14, the filming and recording system 11 can be configured to utilize a microphone cable 18, and the directional microphone 13 is connected to a recording system 19 that records sound via the microphone cable 18. In other words, the directional microphone 13 can transmit audio signals to the recording system 19 wirelessly, or can transmit audio signals to the recording system 19 wired via the microphone cable 18.
 このように構成される撮影収音システム11では、マイクケーブル18の長さを示す情報や、録音システム19の位置を示す情報などが情報取得部33に入力され、演算処理部36は、撮影カメラ12により撮影される映像にマイクケーブル18が映り込むことを回避するように、指向性マイクロフォン13を搭載した移動装置14Aの動作を制御する。このとき、撮影収音システム11では、移動装置14Aの位置は、例えば、移動装置14Aから録音システム19までの距離がマイクケーブル18の長さ以下の範囲内となるように決定される。 In the thus configured filming and recording system 11, information indicating the length of the microphone cable 18, information indicating the position of the recording system 19, and the like are input to the information acquisition unit 33, and the calculation processing unit 36 controls the operation of the mobile device 14A equipped with the directional microphone 13 so as to prevent the microphone cable 18 from appearing in the image captured by the filming camera 12. At this time, in the filming and recording system 11, the position of the mobile device 14A is determined so that, for example, the distance from the mobile device 14A to the recording system 19 is within a range less than the length of the microphone cable 18.
 また、撮影収音システム11は、撮影カメラ12により撮影される映像にマイクケーブル18が映り込むことを回避するために、マイクケーブル18を支持するための移動装置14Bを使用してもよい。 The filming and recording system 11 may also use a moving device 14B for supporting the microphone cable 18 to prevent the microphone cable 18 from appearing in the image captured by the filming camera 12.
 なお、図14では、マイクケーブル18を支持する1つの移動装置14Bが図示されているが、2つ以上の移動装置14Bを使用してマイクケーブル18を支持することができる。そして、撮影収音システム11では、マイクケーブル18を支持する複数の移動装置14Bの個数を示す情報が、情報取得部33に入力され、演算処理部36は、撮影カメラ12により撮影される映像にマイクケーブル18が映り込むことを回避するように、移動装置14A、または、1つ若しくは複数の移動装置14Bの動作を制御する。 Note that while FIG. 14 illustrates one moving device 14B supporting the microphone cable 18, two or more moving devices 14B can be used to support the microphone cable 18. In the filming and sound collection system 11, information indicating the number of multiple moving devices 14B supporting the microphone cable 18 is input to the information acquisition unit 33, and the calculation processing unit 36 controls the operation of the moving device 14A or one or more moving devices 14B so as to prevent the microphone cable 18 from appearing in the image captured by the filming camera 12.
 このとき、撮影収音システム11では、複数の移動装置14Bの位置は、例えば、移動装置14Aから録音システム19までの間で、複数の移動装置14Bが等間隔でマイクケーブル18を支持するように決定される。また、移動装置14Aおよび複数の移動装置14Bの位置は、移動装置14Aから複数の移動装置14Bを介して録音システム19までを結んだ距離が、マイクケーブル18の長さ以下の範囲内となるように決定される。 At this time, in the filming and recording system 11, the positions of the multiple mobile devices 14B are determined, for example, between the mobile device 14A and the recording system 19 so that the multiple mobile devices 14B support the microphone cable 18 at equal intervals. In addition, the positions of the mobile device 14A and the multiple mobile devices 14B are determined so that the distance from the mobile device 14A to the recording system 19 via the multiple mobile devices 14B is within a range less than the length of the microphone cable 18.
 なお、撮影収音システム11では、無指向性マイクロフォン17を使用する場合には、その向きの制御は不要になる。また、撮影収音システム11では、球状アレイマイクを使用する場合には、ポストプロセスで任意方向から到来する収音対象21の音声を強調することができるので、その向き制御は不要になる。また、マイクロフォンの形状やダイアフラム位置などはマイクロフォンごとに異なるため、撮影収音システム11では、それらを考慮して、最適な指向性マイクロフォン13の配置を算出する必要がある。 In addition, when an omnidirectional microphone 17 is used in the filming and sound collection system 11, there is no need to control the direction. Also, when a spherical array microphone is used in the filming and sound collection system 11, the sound of the sound collection target 21 arriving from any direction can be emphasized in post-processing, so there is no need to control the direction. Also, because the shape and diaphragm position of the microphone differ from microphone to microphone, the filming and sound collection system 11 must take these factors into consideration when calculating the optimal placement of the directional microphone 13.
 また、撮影収音システム11では、撮影カメラ12に関する機材情報として、画角を使用する他に、焦点距離およびセンサーサイズを使用してもよい。また、撮影カメラ12により撮影された映像自体を解析することで、指向性マイクロフォン13が映り込んでいるか否かを検出してもよい。 In addition, in the filming and recording system 11, in addition to the angle of view, the focal length and sensor size may be used as equipment information related to the filming camera 12. Furthermore, the image itself captured by the filming camera 12 may be analyzed to detect whether the directional microphone 13 is captured.
 また、撮影収音システム11では、指向性マイクロフォン13を保持する移動体として、移動装置14の他に、クレーンや、ローバー、ワイヤーなどを使用することができる。このとき、クレーンや、ローバー、ワイヤーなどが撮影カメラ12により撮影された映像に映り込むことがないように、これらに関する情報を情報処理装置15に入力して、最適な指向性マイクロフォン13の配置を算出する必要がある。 In addition, in the filming and sound collection system 11, in addition to the moving device 14, a crane, rover, wire, etc. can be used as a moving object that holds the directional microphone 13. In this case, to prevent the crane, rover, wire, etc. from appearing in the image captured by the filming camera 12, information about these must be input to the information processing device 15 and the optimal placement of the directional microphone 13 must be calculated.
 なお、本実施の形態において、指向性マイクロフォン13を使用する場合には、指向性マイクロフォン13の位置と向きとの両方を含む情報を用いて、指向性マイクロフォン13の位置と向きとが制御される。一方、無指向性マイクロフォン17を使用する場合には、無指向性マイクロフォン17の位置を少なくとも含む情報を用いて、無指向性マイクロフォン17の位置が制御される。 In this embodiment, when a directional microphone 13 is used, the position and orientation of the directional microphone 13 are controlled using information including both the position and orientation of the directional microphone 13. On the other hand, when an omnidirectional microphone 17 is used, the position of the omnidirectional microphone 17 is controlled using information including at least the position of the omnidirectional microphone 17.
 <コンピュータの構成例>
 次に、上述した一連の処理(情報処理方法)は、ハードウェアにより行うこともできるし、ソフトウェアにより行うこともできる。一連の処理をソフトウェアによって行う場合には、そのソフトウェアを構成するプログラムが、汎用のコンピュータ等にインストールされる。
<Example of computer configuration>
Next, the above-mentioned series of processes (information processing method) can be performed by hardware or software. When the series of processes is performed by software, a program constituting the software is installed in a general-purpose computer or the like.
 図15は、上述した一連の処理を実行するプログラムがインストールされるコンピュータの一実施の形態の構成例を示すブロック図である。 FIG. 15 is a block diagram showing an example of the configuration of one embodiment of a computer on which a program that executes the series of processes described above is installed.
 プログラムは、コンピュータに内蔵されている記録媒体としてのハードディスク105やROM103に予め記録しておくことができる。 The program can be pre-recorded on the hard disk 105 or ROM 103 as a recording medium built into the computer.
 あるいはまた、プログラムは、ドライブ109によって駆動されるリムーバブル記録媒体111に格納(記録)しておくことができる。このようなリムーバブル記録媒体111は、いわゆるパッケージソフトウェアとして提供することができる。ここで、リムーバブル記録媒体111としては、例えば、フレキシブルディスク、CD-ROM(Compact Disc Read Only Memory),MO(Magneto Optical)ディスク,DVD(Digital Versatile Disc)、磁気ディスク、半導体メモリ等がある。 Alternatively, the program can be stored (recorded) on a removable recording medium 111 driven by the drive 109. Such a removable recording medium 111 can be provided as so-called packaged software. Here, examples of the removable recording medium 111 include a flexible disk, a CD-ROM (Compact Disc Read Only Memory), an MO (Magneto Optical) disk, a DVD (Digital Versatile Disc), a magnetic disk, a semiconductor memory, etc.
 なお、プログラムは、上述したようなリムーバブル記録媒体111からコンピュータにインストールする他、通信網や放送網を介して、コンピュータにダウンロードし、内蔵するハードディスク105にインストールすることができる。すなわち、プログラムは、例えば、ダウンロードサイトから、ディジタル衛星放送用の人工衛星を介して、コンピュータに無線で転送したり、LAN(Local Area Network)、インターネットといったネットワークを介して、コンピュータに有線で転送することができる。 In addition to being installed in a computer from the removable recording medium 111 as described above, the program can also be downloaded to the computer via a communication network or broadcasting network and installed in the built-in hard disk 105. That is, the program can be transferred to the computer wirelessly from a download site via an artificial satellite for digital satellite broadcasting, or transferred to the computer via a wired connection via a network such as a LAN (Local Area Network) or the Internet.
 コンピュータは、CPU(Central Processing Unit)102を内蔵しており、CPU102には、バス101を介して、入出力インタフェース110が接続されている。 The computer has a built-in CPU (Central Processing Unit) 102, to which an input/output interface 110 is connected via a bus 101.
 CPU102は、入出力インタフェース110を介して、ユーザによって、入力部107が操作等されることにより指令が入力されると、それに従って、ROM(Read Only Memory)103に格納されているプログラムを実行する。あるいは、CPU102は、ハードディスク105に格納されたプログラムを、RAM(Random Access Memory)104にロードして実行する。 When a command is input by the user via the input/output interface 110 by operating the input unit 107, the CPU 102 executes a program stored in the ROM (Read Only Memory) 103 accordingly. Alternatively, the CPU 102 loads a program stored on the hard disk 105 into the RAM (Random Access Memory) 104 and executes it.
 これにより、CPU102は、上述したフローチャートにしたがった処理、あるいは上述したブロック図の構成により行われる処理を行う。そして、CPU102は、その処理結果を、必要に応じて、例えば、入出力インタフェース110を介して、出力部106から出力、あるいは、通信部108から送信、さらには、ハードディスク105に記録等させる。 As a result, the CPU 102 performs processing according to the above-mentioned flowchart, or processing performed by the configuration of the above-mentioned block diagram. Then, the CPU 102 outputs the processing results from the output unit 106 via the input/output interface 110, or transmits them from the communication unit 108, or even records them on the hard disk 105, as necessary.
 なお、入力部107は、キーボードや、マウス、マイク等で構成される。また、出力部106は、LCD(Liquid Crystal Display)やスピーカ等で構成される。 The input unit 107 is composed of a keyboard, mouse, microphone, etc. The output unit 106 is composed of an LCD (Liquid Crystal Display), speaker, etc.
 ここで、本明細書において、コンピュータがプログラムに従って行う処理は、必ずしもフローチャートとして記載された順序に沿って時系列に行われる必要はない。すなわち、コンピュータがプログラムに従って行う処理は、並列的あるいは個別に実行される処理(例えば、並列処理あるいはオブジェクトによる処理)も含む。 In this specification, the processing performed by a computer according to a program does not necessarily have to be performed in chronological order according to the order described in the flowchart. In other words, the processing performed by a computer according to a program also includes processing that is executed in parallel or individually (for example, parallel processing or processing by objects).
 また、プログラムは、1のコンピュータ(プロセッサ)により処理されるものであっても良いし、複数のコンピュータによって分散処理されるものであっても良い。さらに、プログラムは、遠方のコンピュータに転送されて実行されるものであっても良い。 The program may be processed by one computer (processor), or may be distributed among multiple computers. Furthermore, the program may be transferred to a remote computer for execution.
 さらに、本明細書において、システムとは、複数の構成要素(装置、モジュール(部品)等)の集合を意味し、すべての構成要素が同一筐体中にあるか否かは問わない。したがって、別個の筐体に収納され、ネットワークを介して接続されている複数の装置、及び、1つの筐体の中に複数のモジュールが収納されている1つの装置は、いずれも、システムである。 Furthermore, in this specification, a system refers to a collection of multiple components (devices, modules (parts), etc.), regardless of whether all the components are in the same housing. Therefore, multiple devices housed in separate housings and connected via a network, and a single device in which multiple modules are housed in a single housing, are both systems.
 また、例えば、1つの装置(または処理部)として説明した構成を分割し、複数の装置(または処理部)として構成するようにしてもよい。逆に、以上において複数の装置(または処理部)として説明した構成をまとめて1つの装置(または処理部)として構成されるようにしてもよい。また、各装置(または各処理部)の構成に上述した以外の構成を付加するようにしてももちろんよい。さらに、システム全体としての構成や動作が実質的に同じであれば、ある装置(または処理部)の構成の一部を他の装置(または他の処理部)の構成に含めるようにしてもよい。 Also, for example, the configuration described above as one device (or processing unit) may be divided and configured as multiple devices (or processing units). Conversely, the configurations described above as multiple devices (or processing units) may be combined and configured as one device (or processing unit). Also, it is of course possible to add configuration other than that described above to the configuration of each device (or each processing unit). Furthermore, as long as the configuration and operation of the system as a whole are substantially the same, part of the configuration of one device (or processing unit) may be included in the configuration of another device (or other processing unit).
 また、例えば、本技術は、1つの機能を、ネットワークを介して複数の装置で分担、共同して処理するクラウドコンピューティングの構成をとることができる。 Also, for example, this technology can be configured as cloud computing, in which a single function is shared and processed collaboratively by multiple devices via a network.
 また、例えば、上述したプログラムは、任意の装置において実行することができる。その場合、その装置が、必要な機能(機能ブロック等)を有し、必要な情報を得ることができるようにすればよい。 Furthermore, for example, the above-mentioned program can be executed in any device. In that case, it is sufficient that the device has the necessary functions (functional blocks, etc.) and is capable of obtaining the necessary information.
 また、例えば、上述のフローチャートで説明した各ステップは、1つの装置で実行する他、複数の装置で分担して実行することができる。さらに、1つのステップに複数の処理が含まれる場合には、その1つのステップに含まれる複数の処理は、1つの装置で実行する他、複数の装置で分担して実行することができる。換言するに、1つのステップに含まれる複数の処理を、複数のステップの処理として実行することもできる。逆に、複数のステップとして説明した処理を1つのステップとしてまとめて実行することもできる。 Furthermore, for example, each step described in the above flowchart can be executed by one device, or can be shared and executed by multiple devices. Furthermore, if one step includes multiple processes, the multiple processes included in that one step can be executed by one device, or can be shared and executed by multiple devices. In other words, multiple processes included in one step can be executed as multiple step processes. Conversely, processes described as multiple steps can be executed collectively as one step.
 なお、コンピュータが実行するプログラムは、プログラムを記述するステップの処理が、本明細書で説明する順序に沿って時系列に実行されるようにしても良いし、並列に、あるいは呼び出しが行われたとき等の必要なタイミングで個別に実行されるようにしても良い。つまり、矛盾が生じない限り、各ステップの処理が上述した順序と異なる順序で実行されるようにしてもよい。さらに、このプログラムを記述するステップの処理が、他のプログラムの処理と並列に実行されるようにしても良いし、他のプログラムの処理と組み合わせて実行されるようにしても良い。 In addition, the processing of the steps that describe a program executed by a computer may be executed chronologically in the order described in this specification, or may be executed in parallel, or individually at the required timing, such as when a call is made. In other words, as long as no contradictions arise, the processing of each step may be executed in an order different from the order described above. Furthermore, the processing of the steps that describe this program may be executed in parallel with the processing of other programs, or may be executed in combination with the processing of other programs.
 なお、本明細書において複数説明した本技術は、矛盾が生じない限り、それぞれ独立に単体で実施することができる。もちろん、任意の複数の本技術を併用して実施することもできる。例えば、いずれかの実施の形態において説明した本技術の一部または全部を、他の実施の形態において説明した本技術の一部または全部と組み合わせて実施することもできる。また、上述した任意の本技術の一部または全部を、上述していない他の技術と併用して実施することもできる。 Note that the multiple present technologies described in this specification can be implemented independently and individually, provided no contradictions arise. Of course, any multiple present technologies can also be implemented in combination. For example, some or all of the present technologies described in any embodiment can be implemented in combination with some or all of the present technologies described in other embodiments. Also, some or all of any of the present technologies described above can be implemented in combination with other technologies not described above.
 <構成の組み合わせ例>
 なお、本技術は以下のような構成も取ることができる。
(1)
 収音対象の音を収音する収音機材の位置または向きを示す情報を少なくとも含み、最適な前記収音機材の配置を算出するのに必要となる情報を取得する情報取得部と、
 前記情報取得部により前記情報に基づいて、前記収音対象の音を収音するのに最適な前記収音機材の配置を算出する演算処理部と、
 前記演算処理部により算出された配置となるように前記収音機材の位置または向きの動作を制御する動作制御部と
 を備える情報処理装置。
(2)
 前記情報取得部は、前記収音対象を撮影する撮影機材の位置および向きを示す情報、および、前記収音対象の位置および向きを示す情報を取得して前記演算処理部に供給し、
 前記演算処理部は、前記収音対象に対して前記収音機材の音軸を向けた際に前記収音対象の直接音を収録することができる範囲となる収録可能範囲の内側、かつ、前記撮影機材により撮影される映像に映り込む範囲となる映り込み範囲の外側であって、前記収音対象に最も近づくことができる位置で前記収音対象に対して向かう方向を、最適な前記収音機材の配置として算出する
 上記(1)に記載の情報処理装置。
(3)
 前記演算処理部が使用する情報を入力する入力部をさらに備え、
 前記入力部は、前記撮影機材の画角を示す情報、前記収音機材の形状および指向性特性を示す情報、並びに、前記収音機材を移動させる移動装置の形状を示す情報を、前記演算処理部に入力する
 上記(2)に記載の情報処理装置。
(4)
 前記撮影機材による撮影が開始される前における前記収音機材と前記収音対象との位置関係を特定して記憶させる位置関係特定部
 をさらに備える上記(3)に記載の情報処理装置。
(5)
 前記動作制御部は、前記撮影機材による撮影が開始される前に、前記入力部に対して入力された前記収音機材の位置または向きの指定または調整に従って、前記収音機材の位置または向きの動作を制御する
 上記(3)に記載の情報処理装置。
(6)
 前記動作制御部は、前記撮影機材による撮影が開始される前に、その撮影前に取得された前記情報に基づいて前記演算処理部により算出された配置となるように、前記収音機材の位置または向きの動作を制御する
 上記(3)に記載の情報処理装置。
(7)
 前記演算処理部は、前記収音対象以外の騒音を出す騒音源の位置を示す情報にも基づいて、前記収音対象の音を収音するのに最適な前記収音機材の配置を算出する
 上記(3)に記載の情報処理装置。
(8)
 前記演算処理部は、前記収音対象に対して照明光を照射する照射機材に関する情報にも基づいて、前記収音対象の音を収音するのに最適な前記収音機材の配置を算出する
 上記(3)に記載の情報処理装置。
(9)
 複数の前記収音対象に対して、指向性を備える1つの前記収音機材を使用する場合、それらの前記収音対象の中間地点に前記収音機材の音軸を向かわせる
 上記(1)から(8)までのいずれかに記載の情報処理装置。
(10)
 複数の前記収音対象に対して、指向性を備える1つの前記収音機材を使用する場合、それぞれの前記収音対象が発話する時刻に従って、それぞれの前記収音対象に前記収音機材の音軸を向かわせる
 上記(1)から(9)までのいずれかに記載の情報処理装置。
(11)
 複数の前記収音対象に対して、無指向性の前記収音機材を使用する
 上記(1)から(10)までのいずれかに記載の情報処理装置。
(12)
 複数の前記収音対象それぞれに対して、指向性を備える複数の前記収音機材を使用する
 上記(1)から(11)までのいずれかに記載の情報処理装置。
(13)
 前記収音機材の配置位置として適切なエリア、および、前記収音機材の配置位置として不適切なエリアを表示するとともに、少なくとも前記収音機材のアイコンを表示する配置画面が表示部に表示される
 上記(1)から(12)までのいずれかに記載の情報処理装置。
(14)
 前記配置画面には、前記収音機材の配置が適切でない場合、その原因を通知するメッセージが表示される
 上記(13)に記載の情報処理装置。
(15)
 前記配置画面には、前記収音機材の配置が適切でない場合、前記収音機材を適切に配置するための代替案が提示される
 上記(13)または(14)に記載の情報処理装置。
(16)
 前記動作制御部は、前記収音機材を移動させる移動装置が他の物体または人物と接触しそうになった場合、前記他の物体または前記人物との接触を回避するように前記移動装置を移動させることを優先する制御を行う
 上記(1)から(15)までのいずれかに記載の情報処理装置。
(17)
 音を録音する録音システムにマイクケーブルを介して前記収音機材が接続されており、 前記演算処理部は、前記撮影機材により撮影される映像に前記マイクケーブルが映り込むことを回避するように、前記収音機材を搭載した移動装置、または、前記マイクケーブルを支持する1つ若しくは複数の移動装置の動作を制御する
 上記(1)から(16)までのいずれかに記載の情報処理装置。
(18)
 指向性を備える前記収音機材を使用する場合、前記収音機材の位置と向きとの両方を含む前記情報が用いられ、
 無指向性の前記収音機材を使用する場合、前記収音機材の位置を少なくとも含む前記情報が用いられる
 上記(1)から(17)までのいずれかに記載の情報処理装置。
(19)
 情報処理装置が
 収音対象の音を収音する収音機材の位置または向きを示す情報を少なくとも含み、最適な前記収音機材の配置を算出するのに必要となる情報を取得することと、
 前記情報に基づいて、前記収音対象の音を収音するのに最適な前記収音機材の配置を算出することと、
 その配置となるように前記収音機材の位置または向きの動作を制御することと
 を含む情報処理方法。
(20)
 情報処理装置のコンピュータに、
 収音対象の音を収音する収音機材の位置または向きを示す情報を少なくとも含み、最適な前記収音機材の配置を算出するのに必要となる情報を取得することと、
 前記情報に基づいて、前記収音対象の音を収音するのに最適な前記収音機材の配置を算出することと、
 その配置となるように前記収音機材の位置または向きの動作を制御することと
 を含む情報処理を実行させるためのプログラム。
<Examples of configuration combinations>
The present technology can also be configured as follows.
(1)
An information acquisition unit that acquires information necessary to calculate an optimal arrangement of the sound collection equipment, the information including at least information indicating a position or orientation of a sound collection device that collects a sound to be collected;
A calculation processing unit that calculates an optimal arrangement of the sound pickup equipment for picking up the sound of the sound pickup target based on the information by the information acquisition unit;
and an operation control unit that controls the position or orientation of the sound pickup equipment so as to achieve the arrangement calculated by the calculation processing unit.
(2)
The information acquisition unit acquires information indicating a position and an orientation of an imaging device that captures the sound collection target, and information indicating a position and an orientation of the sound collection target, and supplies the information to the arithmetic processing unit;
The information processing device described in (1) above, wherein the calculation processing unit calculates the optimal placement of the sound collection equipment as a direction toward the sound collection target at a position that is closest to the sound collection target, inside a recording range in which the direct sound of the sound collection target can be recorded when the sound axis of the sound collection equipment is directed toward the sound collection target, and outside a reflection range in which the sound will be reflected in the image captured by the filming equipment.
(3)
An input unit for inputting information used by the arithmetic processing unit is further provided,
The information processing device described in (2) above, wherein the input unit inputs information indicating the angle of view of the photographing equipment, information indicating the shape and directional characteristics of the sound collection equipment, and information indicating the shape of a moving device that moves the sound collection equipment to the calculation processing unit.
(4)
The information processing device according to (3) above, further comprising: a positional relationship specifying unit that specifies and stores a positional relationship between the sound collection equipment and the sound collection target before the shooting by the shooting equipment is started.
(5)
The information processing device described in (3) above, wherein the operation control unit controls the operation of the position or orientation of the sound collection equipment in accordance with a designation or adjustment of the position or orientation of the sound collection equipment input to the input unit before shooting by the photographing equipment begins.
(6)
The information processing device described in (3) above, wherein the operation control unit controls the operation of the position or orientation of the sound collection equipment so that the position is positioned as calculated by the calculation processing unit based on the information acquired before shooting is started by the photographing equipment.
(7)
The information processing device according to (3) above, wherein the calculation processing unit calculates an optimal arrangement of the sound collection equipment for collecting the sound of the sound collection target based on information indicating the position of a noise source that emits noise other than the sound collection target.
(8)
The information processing device according to (3) above, wherein the calculation processing unit calculates an optimal arrangement of the sound collection equipment for collecting the sound of the sound collection target based on information about an irradiation equipment that irradiates the sound collection target with illumination light.
(9)
The information processing device according to any one of (1) to (8) above, wherein when one directional sound recording device is used for a plurality of sound recording targets, the sound axis of the sound recording device is directed toward a midpoint between the sound recording targets.
(10)
When one directional sound recording device is used for multiple sound recording targets, the sound axis of the sound recording device is directed toward each of the sound recording targets according to the time at which each of the sound recording targets speaks. An information processing device as described in any of (1) to (9) above.
(11)
The information processing device according to any one of (1) to (10) above, wherein the sound collection equipment is omnidirectional for a plurality of the sound collection targets.
(12)
The information processing device according to any one of (1) to (11) above, wherein a plurality of the sound collection devices each having directivity are used for each of the plurality of sound collection targets.
(13)
An information processing device as described in any of (1) to (12) above, wherein a layout screen is displayed on a display unit, which displays areas suitable for placement of the sound recording equipment and areas inappropriate for placement of the sound recording equipment, and which displays at least an icon of the sound recording equipment.
(14)
The information processing device according to (13) above, wherein, if the placement of the sound pickup equipment is not appropriate, a message notifying the user of the cause is displayed on the placement screen.
(15)
The information processing device according to (13) or (14) above, wherein, when the placement of the sound pickup equipment is not appropriate, an alternative for appropriately placing the sound pickup equipment is presented on the placement screen.
(16)
The information processing device described in any of (1) to (15) above, wherein the operation control unit performs control to prioritize moving the moving device so as to avoid contact with another object or person when the moving device moving the sound collection equipment is about to come into contact with the other object or person.
(17)
The information processing device described in any of (1) to (16) above, wherein the sound collection equipment is connected via a microphone cable to a recording system that records sound, and the calculation processing unit controls the operation of a mobile device equipped with the sound collection equipment or one or more mobile devices that support the microphone cable so as to avoid the microphone cable being captured in an image captured by the filming equipment.
(18)
When using the sound pickup equipment having directivity, the information including both the position and the orientation of the sound pickup equipment is used,
The information processing device according to any one of (1) to (17) above, wherein when the sound pickup equipment is non-directional, the information including at least the position of the sound pickup equipment is used.
(19)
An information processing device acquires information including at least information indicating a position or orientation of a sound collection device that collects a sound of a sound collection target, and information necessary for calculating an optimal arrangement of the sound collection device;
Calculating an optimal arrangement of the sound pickup equipment for picking up the sound of the sound pickup target based on the information;
and controlling the position or orientation of the sound pickup equipment so as to achieve the arrangement.
(20)
The computer of the information processing device
Acquiring information necessary for calculating an optimal arrangement of the sound pickup equipment, the information including at least information indicating a position or orientation of the sound pickup equipment that picks up the sound of the sound pickup target;
Calculating an optimal arrangement of the sound pickup equipment for picking up the sound of the sound pickup target based on the information;
and controlling the position or orientation of the sound pickup equipment so as to achieve the arrangement.
 なお、本実施の形態は、上述した実施の形態に限定されるものではなく、本開示の要旨を逸脱しない範囲において種々の変更が可能である。また、本明細書に記載された効果はあくまで例示であって限定されるものではなく、他の効果があってもよい。 Note that this embodiment is not limited to the above-described embodiment, and various modifications are possible without departing from the gist of this disclosure. Furthermore, the effects described in this specification are merely examples and are not limiting, and other effects may also be present.
 11 撮影収音システム, 12 撮影カメラ, 13 指向性マイクロフォンロフォン, 14 移動装置, 15 情報処理装置, 16 照明機材, 17 無指向性マイクロフォン, 18 マイクケーブル, 19 録音システム, 21 収音対象, 22 騒音源, 31 ユーザ入力部, 32 動作制御部, 33 情報取得部, 34 位置関係特定部, 35 記憶部, 36 演算処理部, 41 収録可能範囲推定部, 42 映り込み範囲推定部, 43 最適配置算出部 11. Filming and sound recording system, 12. Filming camera, 13. Directional microphone, 14. Mobile device, 15. Information processing device, 16. Lighting equipment, 17. Omnidirectional microphone, 18. Microphone cable, 19. Recording system, 21. Sound recording target, 22. Noise source, 31. User input unit, 32. Operation control unit, 33. Information acquisition unit, 34. Positional relationship determination unit, 35. Memory unit, 36. Calculation processing unit, 41. Recordable range estimation unit, 42. Reflection range estimation unit, 43. Optimal placement calculation unit

Claims (20)

  1.  収音対象の音を収音する収音機材の位置または向きを示す情報を少なくとも含み、最適な前記収音機材の配置を算出するのに必要となる情報を取得する情報取得部と、
     前記情報取得部により前記情報に基づいて、前記収音対象の音を収音するのに最適な前記収音機材の配置を算出する演算処理部と、
     前記演算処理部により算出された配置となるように前記収音機材の位置または向きの動作を制御する動作制御部と
     を備える情報処理装置。
    An information acquisition unit that acquires information necessary to calculate an optimal arrangement of the sound collection equipment, the information including at least information indicating a position or orientation of a sound collection device that collects a sound to be collected;
    A calculation processing unit that calculates an optimal arrangement of the sound pickup equipment for picking up the sound of the sound pickup target based on the information by the information acquisition unit;
    and an operation control unit that controls the operation of the position or orientation of the sound pickup equipment so as to achieve the arrangement calculated by the calculation processing unit.
  2.  前記情報取得部は、前記収音対象を撮影する撮影機材の位置および向きを示す情報、および、前記収音対象の位置および向きを示す情報を取得して前記演算処理部に供給し、
     前記演算処理部は、前記収音対象に対して前記収音機材の音軸を向けた際に前記収音対象の直接音を収録することができる範囲となる収録可能範囲の内側、かつ、前記撮影機材により撮影される映像に映り込む範囲となる映り込み範囲の外側であって、前記収音対象に最も近づくことができる位置で前記収音対象に対して向かう方向を、最適な前記収音機材の配置として算出する
     請求項1に記載の情報処理装置。
    The information acquisition unit acquires information indicating a position and an orientation of an imaging device that captures the sound collection target, and information indicating a position and an orientation of the sound collection target, and supplies the information to the arithmetic processing unit;
    The information processing device of claim 1, wherein the calculation processing unit calculates the optimal position of the sound collection equipment as a direction toward the sound collection target at a position that is closest to the sound collection target, inside a recording range in which the direct sound of the sound collection target can be recorded when the sound axis of the sound collection equipment is directed toward the sound collection target, and outside a reflection range in which the sound is reflected in the image captured by the photographing equipment.
  3.  前記演算処理部が使用する情報を入力する入力部をさらに備え、
     前記入力部は、前記撮影機材の画角を示す情報、前記収音機材の形状および指向性特性を示す情報、並びに、前記収音機材を移動させる移動装置の形状を示す情報を、前記演算処理部に入力する
     請求項2に記載の情報処理装置。
    An input unit for inputting information used by the arithmetic processing unit is further provided,
    The information processing device according to claim 2 , wherein the input unit inputs information indicating an angle of view of the photographing equipment, information indicating a shape and directional characteristics of the sound collecting equipment, and information indicating a shape of a moving device that moves the sound collecting equipment to the calculation processing unit.
  4.  前記撮影機材による撮影が開始される前における前記収音機材と前記収音対象との位置関係を特定して記憶させる位置関係特定部
     をさらに備える請求項3に記載の情報処理装置。
    The information processing apparatus according to claim 3 , further comprising a positional relationship specification unit that specifies and stores a positional relationship between the sound collection equipment and the sound collection target before shooting is started by the photographing equipment.
  5.  前記動作制御部は、前記撮影機材による撮影が開始される前に、前記入力部に対して入力された前記収音機材の位置または向きの指定または調整に従って、前記収音機材の位置または向きの動作を制御する
     請求項3に記載の情報処理装置。
    The information processing device according to claim 3 , wherein the operation control unit controls the operation of a position or orientation of the sound pickup equipment in accordance with a designation or adjustment of a position or orientation of the sound pickup equipment input to the input unit before shooting by the photographing equipment is started.
  6.  前記動作制御部は、前記撮影機材による撮影が開始される前に、その撮影前に取得された前記情報に基づいて前記演算処理部により算出された配置となるように、前記収音機材の位置または向きの動作を制御する
     請求項3に記載の情報処理装置。
    The information processing device according to claim 3 , wherein the operation control unit controls the operation of the position or orientation of the sound collection equipment so as to achieve a position calculated by the calculation processing unit based on the information acquired before the shooting is started by the photographing equipment.
  7.  前記演算処理部は、前記収音対象以外の騒音を出す騒音源の位置を示す情報にも基づいて、前記収音対象の音を収音するのに最適な前記収音機材の配置を算出する
     請求項3に記載の情報処理装置。
    The information processing device according to claim 3 , wherein the calculation processing unit calculates an optimal arrangement of the sound pickup equipment for picking up the sound of the sound pickup target, based also on information indicating a position of a noise source that emits noise other than the sound pickup target.
  8.  前記演算処理部は、前記収音対象に対して照明光を照射する照射機材に関する情報にも基づいて、前記収音対象の音を収音するのに最適な前記収音機材の配置を算出する
     請求項3に記載の情報処理装置。
    The information processing device according to claim 3 , wherein the calculation processing unit calculates an optimal arrangement of the sound collection equipment for collecting the sound of the sound collection target, based also on information on an irradiation equipment that irradiates the sound collection target with illumination light.
  9.  複数の前記収音対象に対して、指向性を備える1つの前記収音機材を使用する場合、それらの前記収音対象の中間地点に前記収音機材の音軸を向かわせる
     請求項1に記載の情報処理装置。
    The information processing apparatus according to claim 1 , wherein, when one of the sound collection devices having directivity is used for a plurality of the sound collection targets, a sound axis of the sound collection device is directed to a midpoint between the sound collection targets.
  10.  複数の前記収音対象に対して、指向性を備える1つの前記収音機材を使用する場合、それぞれの前記収音対象が発話する時刻に従って、それぞれの前記収音対象に前記収音機材の音軸を向かわせる
     請求項1に記載の情報処理装置。
    The information processing device according to claim 1 , wherein when one sound collection device having directionality is used for a plurality of sound collection targets, the sound axis of the sound collection device is directed toward each of the sound collection targets according to the time at which each of the sound collection targets speaks.
  11.  複数の前記収音対象に対して、無指向性の前記収音機材を使用する
     請求項1に記載の情報処理装置。
    The information processing apparatus according to claim 1 , wherein the sound collection device is non-directional for a plurality of the sound collection targets.
  12.  複数の前記収音対象それぞれに対して、指向性を備える複数の前記収音機材を使用する
     請求項1に記載の情報処理装置。
    The information processing apparatus according to claim 1 , wherein a plurality of the sound collection devices each having directivity are used for a plurality of the sound collection targets.
  13.  前記収音機材の配置位置として適切なエリア、および、前記収音機材の配置位置として不適切なエリアを表示するとともに、少なくとも前記収音機材のアイコンを表示する配置画面が表示部に表示される
     請求項1に記載の情報処理装置。
    The information processing device according to claim 1 , wherein a layout screen is displayed on a display unit, the layout screen displaying areas suitable for placement of the sound pickup equipment and areas unsuitable for placement of the sound pickup equipment, and displaying at least an icon of the sound pickup equipment.
  14.  前記配置画面には、前記収音機材の配置が適切でない場合、その原因を通知するメッセージが表示される
     請求項13に記載の情報処理装置。
    The information processing device according to claim 13 , wherein, when the placement of the sound pickup devices is not appropriate, a message notifying the cause is displayed on the placement screen.
  15.  前記配置画面には、前記収音機材の配置が適切でない場合、前記収音機材を適切に配置するための代替案が提示される
     請求項13に記載の情報処理装置。
    The information processing device according to claim 13 , wherein, when the placement of the sound pickup equipment is not appropriate, an alternative plan for appropriately placing the sound pickup equipment is presented on the placement screen.
  16.  前記動作制御部は、前記収音機材を移動させる移動装置が他の物体または人物と接触しそうになった場合、前記他の物体または前記人物との接触を回避するように前記移動装置を移動させることを優先する制御を行う
     請求項1に記載の情報処理装置。
    The information processing device according to claim 1 , wherein the operation control unit performs control to prioritize moving the moving device so as to avoid contact with another object or person when the moving device moving the sound collection equipment is about to come into contact with the other object or person.
  17.  音を録音する録音システムにマイクケーブルを介して前記収音機材が接続されており、 前記演算処理部は、前記撮影機材により撮影される映像に前記マイクケーブルが映り込むことを回避するように、前記収音機材を搭載した移動装置、または、前記マイクケーブルを支持する1つ若しくは複数の移動装置の動作を制御する
     請求項1に記載の情報処理装置。
    The information processing device according to claim 1, wherein the sound collection equipment is connected to a recording system that records sound via a microphone cable, and the calculation processing unit controls the operation of a mobile device equipped with the sound collection equipment or one or more mobile devices that support the microphone cable so as to prevent the microphone cable from appearing in an image captured by the filming equipment.
  18.  指向性を備える前記収音機材を使用する場合、前記収音機材の位置と向きとの両方を含む前記情報が用いられ、
     無指向性の前記収音機材を使用する場合、前記収音機材の位置を少なくとも含む前記情報が用いられる
     請求項1に記載の情報処理装置。
    When using the sound pickup equipment having directivity, the information including both the position and the orientation of the sound pickup equipment is used,
    The information processing apparatus according to claim 1 , wherein when the sound pickup device is non-directional, the information including at least a position of the sound pickup device is used.
  19.  情報処理装置が
     収音対象の音を収音する収音機材の位置または向きを示す情報を少なくとも含み、最適な前記収音機材の配置を算出するのに必要となる情報を取得することと、
     前記情報に基づいて、前記収音対象の音を収音するのに最適な前記収音機材の配置を算出することと、
     その配置となるように前記収音機材の位置または向きの動作を制御することと
     を含む情報処理方法。
    An information processing device acquires information including at least information indicating a position or orientation of a sound collection device that collects a sound of a sound collection target, and information necessary for calculating an optimal arrangement of the sound collection device;
    Calculating an optimal arrangement of the sound pickup equipment for picking up the sound of the sound pickup target based on the information;
    and controlling the position or orientation of the sound pickup equipment so as to achieve the arrangement.
  20.  情報処理装置のコンピュータに、
     収音対象の音を収音する収音機材の位置または向きを示す情報を少なくとも含み、最適な前記収音機材の配置を算出するのに必要となる情報を取得することと、
     前記情報に基づいて、前記収音対象の音を収音するのに最適な前記収音機材の配置を算出することと、
     その配置となるように前記収音機材の位置または向きの動作を制御することと
     を含む情報処理を実行させるためのプログラム。
    The computer of the information processing device
    Acquiring information necessary for calculating an optimal arrangement of the sound pickup equipment, the information including at least information indicating a position or orientation of the sound pickup equipment that picks up the sound of the sound pickup target;
    Calculating an optimal arrangement of the sound pickup equipment for picking up the sound of the sound pickup target based on the information;
    and controlling the position or orientation of the sound pickup equipment so as to achieve the arrangement.
PCT/JP2023/036016 2022-10-20 2023-10-03 Information processing device, information processing method, and program WO2024084953A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2022168133 2022-10-20
JP2022-168133 2022-10-20

Publications (1)

Publication Number Publication Date
WO2024084953A1 true WO2024084953A1 (en) 2024-04-25

Family

ID=90737305

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2023/036016 WO2024084953A1 (en) 2022-10-20 2023-10-03 Information processing device, information processing method, and program

Country Status (1)

Country Link
WO (1) WO2024084953A1 (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH08508347A (en) * 1993-03-31 1996-09-03 ジャン エリク ホルスコブ ジューレル Method for simultaneous recording of image and sound and camera for realizing this method
JP2020006793A (en) * 2018-07-06 2020-01-16 パナソニックIpマネジメント株式会社 Unmanned air vehicle with sound absorbing function
WO2020202621A1 (en) * 2019-03-29 2020-10-08 パナソニックIpマネジメント株式会社 Unmanned moving body and information processing method
WO2020213198A1 (en) * 2019-04-18 2020-10-22 パナソニックIpマネジメント株式会社 Unmanned flight body, control method, and program
WO2020230377A1 (en) * 2019-05-15 2020-11-19 パナソニックIpマネジメント株式会社 Information processing method, unmanned aerial vehicle, and unmanned aerial vehicle control system

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH08508347A (en) * 1993-03-31 1996-09-03 ジャン エリク ホルスコブ ジューレル Method for simultaneous recording of image and sound and camera for realizing this method
JP2020006793A (en) * 2018-07-06 2020-01-16 パナソニックIpマネジメント株式会社 Unmanned air vehicle with sound absorbing function
WO2020202621A1 (en) * 2019-03-29 2020-10-08 パナソニックIpマネジメント株式会社 Unmanned moving body and information processing method
WO2020213198A1 (en) * 2019-04-18 2020-10-22 パナソニックIpマネジメント株式会社 Unmanned flight body, control method, and program
WO2020230377A1 (en) * 2019-05-15 2020-11-19 パナソニックIpマネジメント株式会社 Information processing method, unmanned aerial vehicle, and unmanned aerial vehicle control system

Similar Documents

Publication Publication Date Title
US10848889B2 (en) Intelligent audio rendering for video recording
US10531220B2 (en) Distributed audio capturing techniques for virtual reality (VR), augmented reality (AR), and mixed reality (MR) systems
US20110193935A1 (en) Controlling a video window position relative to a video camera position
KR20200009005A (en) Imaging device, imaging method
JP2004528766A (en) Method and apparatus for sensing and locating a speaker using sound / image
EP2998935B1 (en) Image processing device, image processing method, and program
US11736802B2 (en) Communication management apparatus, image communication system, communication management method, and recording medium
KR20200009003A (en) Information processing device, information processing method, program
WO2024084953A1 (en) Information processing device, information processing method, and program
CN111325790B (en) Target tracking method, device and system
US10939030B2 (en) Video audio processing system and method of controlling the video audio processing system
WO2020026548A1 (en) Information processing device, information processing method, and acoustic system
CN113676622A (en) Video processing method, image pickup apparatus, video conference system, and storage medium
JP2002101408A (en) Supervisory camera system
JP2020006793A (en) Unmanned air vehicle with sound absorbing function
CN113707165A (en) Audio processing method and device, electronic equipment and storage medium
WO2018173139A1 (en) Imaging/sound acquisition device, sound acquisition control system, method for controlling imaging/sound acquisition device, and method for controlling sound acquisition control system
JP5712599B2 (en) Imaging apparatus and program
US20080170705A1 (en) Recorder that creates stereophonic sound
JP2017175598A (en) Sound collecting device and sound collecting method
KR20170105727A (en) Method for enhancing the security of building by tracing a man who is speaking on the basis of voice analaysis and system using the same
CN114175616A (en) Image processing apparatus, image processing method, and program
CN117098032A (en) Audio processing method, electronic device and computer readable storage medium
JP2005348449A (en) Imaging, displaying, recording, reproducing, transmitting device and recording medium
JP6860178B1 (en) Video processing equipment and video processing method