JP4332365B2 - Metadata display system, video signal recording / reproducing device, imaging device, metadata display method - Google Patents

Metadata display system, video signal recording / reproducing device, imaging device, metadata display method Download PDF

Info

Publication number
JP4332365B2
JP4332365B2 JP2003101837A JP2003101837A JP4332365B2 JP 4332365 B2 JP4332365 B2 JP 4332365B2 JP 2003101837 A JP2003101837 A JP 2003101837A JP 2003101837 A JP2003101837 A JP 2003101837A JP 4332365 B2 JP4332365 B2 JP 4332365B2
Authority
JP
Japan
Prior art keywords
metadata
video signal
device
video
example
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
JP2003101837A
Other languages
Japanese (ja)
Other versions
JP2004312278A (en
Inventor
史夫 中島
弘美 星野
Original Assignee
ソニー株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ソニー株式会社 filed Critical ソニー株式会社
Priority to JP2003101837A priority Critical patent/JP4332365B2/en
Publication of JP2004312278A publication Critical patent/JP2004312278A/en
Application granted granted Critical
Publication of JP4332365B2 publication Critical patent/JP4332365B2/en
Application status is Expired - Fee Related legal-status Critical
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/11Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information not detectable on the record carrier
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
    • G11B27/30Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording
    • G11B27/3027Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording used signal is digitally coded
    • G11B27/3036Time code signal
    • G11B27/3045Time code signal superimposed on the recorded main signal, e.g. burn-in-time code
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
    • G11B27/30Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording
    • G11B27/3027Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording used signal is digitally coded
    • G11B27/3036Time code signal
    • G11B27/3054Vertical Interval Time code [VITC]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment ; Cameras comprising an electronic image sensor, e.g. digital cameras, video cameras, TV cameras, video cameras, camcorders, webcams, camera modules for embedding in other devices, e.g. mobile phones, computers or vehicles
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment ; Cameras comprising an electronic image sensor, e.g. digital cameras, video cameras, TV cameras, video cameras, camcorders, webcams, camera modules for embedding in other devices, e.g. mobile phones, computers or vehicles
    • H04N5/225Television cameras ; Cameras comprising an electronic image sensor, e.g. digital cameras, video cameras, camcorders, webcams, camera modules specially adapted for being embedded in other devices, e.g. mobile phones, computers or vehicles
    • H04N5/232Devices for controlling television cameras, e.g. remote control ; Control of cameras comprising an electronic image sensor
    • H04N5/23293Electronic viewfinders
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/79Processing of colour television signals in connection with recording
    • H04N9/80Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
    • H04N9/82Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
    • H04N9/8205Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B2220/00Record carriers by type
    • G11B2220/90Tape-like record carriers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/765Interface circuits between an apparatus for recording and another apparatus
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/765Interface circuits between an apparatus for recording and another apparatus
    • H04N5/77Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television camera
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/765Interface circuits between an apparatus for recording and another apparatus
    • H04N5/77Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television camera
    • H04N5/772Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television camera the recording apparatus and the television camera being placed in the same enclosure
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/765Interface circuits between an apparatus for recording and another apparatus
    • H04N5/775Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television receiver
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/78Television signal recording using magnetic recording
    • H04N5/781Television signal recording using magnetic recording on disks or drums
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/84Television signal recording using optical recording
    • H04N5/85Television signal recording using optical recording on discs or drums
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/907Television signal recording using static stores, e.g. storage tubes or semiconductor memories
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/79Processing of colour television signals in connection with recording
    • H04N9/7921Processing of colour television signals in connection with recording for more than one processing mode
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/79Processing of colour television signals in connection with recording
    • H04N9/80Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
    • H04N9/804Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
    • H04N9/8042Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components involving data reduction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/79Processing of colour television signals in connection with recording
    • H04N9/80Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
    • H04N9/804Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
    • H04N9/8042Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components involving data reduction
    • H04N9/8047Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components involving data reduction using transform coding

Description

[0001]
BACKGROUND OF THE INVENTION
  The present invention displays metadata related to a captured video signal.METADATA DISPLAY SYSTEM, VIDEO SIGNAL RECORDING / REPRODUCING DEVICE, IMAGING DEVICE, METADATA DISPLAY METHODAbout.
[0002]
[Prior art]
In recent years, in the field of production of video works such as movies and TV programs, effective use of metadata relating to captured video material has been promoted. Examples of the metadata relating to the video material include scene information metadata representing the attributes of the video material such as a scene number, a take number, and a time code of the photographed video material. The scene information metadata is useful information for identifying and managing video material being photographed or photographed by a photographer or director at a photographing site, for example.
[0003]
Conventionally, such scene information metadata has been displayed electronically or handwritten on an electronic clapperboard or the like. The electronic clapperboard displaying the scene information metadata in this way is presented and photographed in front of the imaging device at the start or end of photographing, so that the scene information metadata is directly inserted before and after the video material, It was recorded as data (see, for example, Patent Document 1).
[0004]
However, there are cases where the above-mentioned electronic clapperboard cannot be presented directly in front of the imaging device when shooting with a crane or shooting with a steadicam. In such a case, shooting is performed according to the time code output by the electronic clapperboard and the video material is recorded on a storage medium, while the scene information metadata is input to a terminal such as a personal computer separately from the video material. And recorded with the time code.
[0005]
[Patent Document 1]
JP-A-9-46627
[0006]
[Problems to be solved by the invention]
  However, in the conventional metadata recording method,There is a problem that it is impossible to appropriately extract and display necessary metadata in units of groups according to the purpose of use of metadata.
[0008]
  The present invention has been made in view of the above problems, and the object of the present invention is toAccording to the purpose of using metadata, necessary metadata is extracted and displayed in groups.It is an object of the present invention to provide a new and improved metadata display system, metadata display method, and the like that can be used.
[0009]
[Means for Solving the Problems]
  In order to solve the above problems, according to a first aspect of the present invention, a metadata display system for displaying metadata related to a video signal generated by an imaging apparatus is provided. This metadata display system stores, as metadata, metadata including at least scene information acquired in advance from the metadata input terminal device before the start of photographing, and also photographed from the imaging device, the lens device, and / or the dolly device. A buffer memory that stores metadata that is sent in sequence after the start; for each purpose of use of the metadata, the necessary metadata is extracted from the metadata stored in the buffer memory;2 or moreA metadata packing unit for grouping into metadata groups; and extracting at least a part of the grouped metadata and synthesizing the extracted metadata into a video signal; displayThe metadata group is grouped for the purpose of identifying the video material, the scene information group containing information related to the scene captured by the imaging device, and the grouping for the purpose of grasping the image quality of the video material. A camera setting group including setting information of the imaging device and a lens setting group or setting group including setting information of the lens device or the dolly device included in the imaging device grouped for the purpose of grasping the movement or distance of the subject in the video material. Including setting groupsIt is characterized by that.
  In addition, the metadata packing unit may pack each metadata group by KLV (Key Length Value) encoding.
[0010]
  With this configuration,Depending on the purpose of use of metadata, necessary metadata can be suitably extracted and displayed in groups.Therefore, a photographer, a director, or the like can browse the metadata along with the video during recording or after recording to grasp the contents of shooting, check the video, and the like.
[0011]
Further, the display device may be configured to be provided in the imaging device. With this configuration, the video signal combined with the metadata is provided to the imaging device. For this reason, on the display device such as a viewfinder provided in the imaging device, the metadata corresponding to the video can be displayed simultaneously with the shot video.
[0013]
  In addition, the grouped metadata is added to the video signal in units of frames, and at least a part of the grouped metadata is extracted from the video signal and combined with the video signal. Also good. With this configuration, at least a part can be extracted from the metadata added to the video signal in units of frames, converted into video data, and multiplexed on the video signals in units of frames. The video signal combined with the metadata in this way is provided to the display device. Therefore, the display device can simultaneously display the corresponding metadata in synchronization with the video imaged by the imaging device. Therefore, a photographer, a director, or the like can browse the metadata along with the video during recording or after recording to grasp the contents of shooting, check the video, and the like.
[0014]
The metadata group includes a scene information group including information related to a scene captured by the imaging device, a camera setting group including setting information of the imaging device, a lens setting group including setting information of a lens device included in the imaging device, Or you may comprise so that at least any one of the dolly setting groups containing the setting information of the dolly apparatus with which an imaging device is provided may be included.
[0015]
Further, the metadata group added to the video signal may be configured to be given unique group identification information. With this configuration, it is possible to identify which metadata group is based on the group identification information, so that the extraction process in units of metadata groups can be performed quickly.
[0016]
Further, the metadata group added to the video signal may be configured such that data amount information of the metadata group is given. With this configuration, when executing the metadata extraction process of a certain metadata group, the metadata amount in the metadata group can be grasped in advance based on the data amount information. For this reason, it is possible to quickly perform extraction processing in units of metadata groups.
[0017]
  The above metadata display systemGroupedA video signal recording / reproducing device for recording / reproducing a video signal with metadata added to / from a storage medium is provided. From the video signal reproduced by the video signal recording / reproducing device,GroupedIt may be configured such that at least a part of the metadata is extracted and combined with the reproduced video signal. With this configuration, a video signal generated by the imaging device and having metadata added in frame units can be recorded and recorded on a storage medium by the video signal recording / reproducing device. Furthermore, by reproducing the video signal from the storage medium by the video signal recording / reproducing apparatus, and extracting and synthesizing the metadata from the reproduced video signal, the metadata can be displayed together with the video even after recording. .
[0023]
  In order to solve the above-described problem, according to another aspect of the present invention, at least a scene acquired in advance from the metadata input terminal device as the metadata related to the video signal generated by the imaging device before the start of shooting. A buffer memory for storing metadata including information and storing metadata sequentially transmitted from the imaging device, the lens device and / or the dolly device after the start of photographing; and storing in the buffer memory for each purpose of use of the metadata Extract the necessary metadata from the existing metadata,2 or moreA metadata packing unit for grouping into a metadata group; a recording / reproducing unit for recording / reproducing a video signal to which the grouped metadata is added in units of frames; and a grouped metadata A metadata synthesizer that extracts at least part of the grouped metadata from the video signal added in frame units and synthesizes it into the video signal;The metadata group is grouped for the purpose of identifying the video material, the scene information group containing information related to the scene shot by the imaging device, and the grouping for the purpose of grasping the image quality of the video material. A camera setting group including setting information of the imaging device and a lens setting group or setting group including setting information of the lens device or the dolly device included in the imaging device, which is grouped for the purpose of grasping the movement or distance of the subject in the video material. Including setting groupsA video signal recording / reproducing apparatus is provided. With this configuration, it is possible to provide a video signal recording / reproducing apparatus including the metadata synthesizing apparatus.
  In addition, the metadata packing unit may pack each metadata group by KLV (Key Length Value) encoding.
[0024]
  In addition, the metadata synthesizing device provided in the video signal recording / playback device is configured to use the video signal played back from the storage medium by the recording / playback unit.GroupedIt may be configured to extract at least a part of the metadata and synthesize it with the reproduced video signal. As a result, the video once recorded can be reproduced and the metadata can be synthesized.
[0025]
  In order to solve the above problems, according to another aspect of the present invention, there is provided an imaging apparatus including an imaging unit that captures an image of a subject and generates a video signal, and a display unit that displays the video signal. The The imaging apparatus stores metadata including at least scene information acquired in advance from the metadata input terminal device before starting photographing as metadata related to the video signal generated by the imaging apparatus, A buffer memory that stores metadata that is sequentially transmitted from the device and / or the dolly device after the start of photographing; and extracts necessary metadata from the metadata stored in the buffer memory for each purpose of use of the metadata A metadata packing unit for grouping into one or more metadata groups, and from the video signal to which the grouped metadata is added in units of frames, at least one of the grouped metadata. Connected to a metadata synthesizer that extracts the video and synthesizes it to the video signal. From the device, it receives the video signals metadata is synthesized, displayed on the display unitThe metadata group is grouped for the purpose of identifying the video material, the scene information group containing information related to the scene captured by the imaging device, and the grouping for the purpose of grasping the image quality of the video material. A camera setting group including setting information of the imaging device and a lens setting group or setting group including setting information of the lens device or the dolly device included in the imaging device grouped for the purpose of grasping the movement or distance of the subject in the video material. Including setting groupsIt is characterized by that.
  In addition, the metadata packing unit may pack each metadata group by KLV (Key Length Value) encoding.
[0026]
With this configuration, the imaging apparatus can receive the video signal obtained by synthesizing the metadata by the metadata synthesizing apparatus as a return video of the video signal generated by itself. Thereby, the imaging device can multiplex and display the metadata on the generated video signal.
[0027]
  In order to solve the above problems, according to another aspect of the present invention, a metadata display method for displaying metadata related to a video signal generated by an imaging apparatus is provided. This metadata display method stores, as metadata, metadata including at least scene information acquired in advance from the metadata input terminal device before the start of photographing, and is photographed from the imaging device, the lens device, and / or the dolly device. The metadata that is sent sequentially after the start is stored in the buffer memory, and the necessary metadata is extracted from the metadata stored in the buffer memory for each purpose of use of the metadata.2 or moreGroup into metadata groups, extract at least part of the grouped metadata, combine it with video signal, and display the video signal with the combined metadata on the display deviceThe metadata group is grouped for the purpose of identifying the video material, the scene information group containing information related to the scene captured by the imaging device, and the grouping for the purpose of grasping the image quality of the video material. A camera setting group including setting information of the imaging device and a lens setting group or setting group including setting information of the lens device or the dolly device included in the imaging device grouped for the purpose of grasping the movement or distance of the subject in the video material. Including setting groupsIt is characterized by that.
  Each metadata group may be packed by KLV (Key Length Value) encoding.
[0028]
  The display device may be provided in the imaging device.
  The metadata group includes a scene information group including information related to a scene captured by the imaging device, a camera setting group including setting information of the imaging device, a lens setting group including setting information of a lens device included in the imaging device, Alternatively, at least one of the dolly setting groups including the setting information of the dolly device included in the imaging device may be included.
The grouped metadata may be added to the video signal in units of frames, and at least a part of the grouped metadata may be extracted from the video signal and synthesized with the video signal.
[0029]
  The metadata display method is as follows.GroupedA video signal is reproduced from a storage medium that records the video signal to which metadata is added, and from the reproduced video signal,GroupedAt least a part of the metadata may be extracted and synthesized with the reproduced video signal.
[0030]
DETAILED DESCRIPTION OF THE INVENTION
Hereinafter, preferred embodiments of the present invention will be described in detail with reference to the accompanying drawings. In the present specification and drawings, components having substantially the same functional configuration are denoted by the same reference numerals, and redundant description is omitted.
[0031]
(First embodiment)
The metadata display system, metadata synthesizer, video signal recording / playback device, imaging device, metadata display method, and the like according to the first embodiment of the present invention will be described below.
[0032]
<1. System configuration>
First, an outline of the metadata display system according to the present embodiment will be described. The metadata display system according to the present embodiment is configured as, for example, a video recording system for shooting and recording video material.
[0033]
This video recording system is a system for producing video works such as TV programs, video contents, movies, etc., for example, by a television broadcasting station, a production company of video contents, movies, etc. Video content data of video materials that make up a video work can be taken and recorded. This video content data is content data composed of video data and / or audio data, for example. Of these, the video data is generally, for example, moving image data, but may include still image data such as drawings, photographs or paintings.
[0034]
In addition, this video recording system can generate, for example, various types of metadata related to the shot video material. Furthermore, for example, the video recording system can group such metadata, add it to the video signal constituting the video material for each frame, and record it together with the video signal in a storage medium. Note that this metadata is, for example, high-order data that represents the outline of the video material, attributes, or settings of the imaging device, and functions as index information of the video material, information for specifying shooting conditions, and the like. Will be described later.
[0035]
The metadata display system configured as such a video recording system can extract metadata from the video signal to which the metadata is added and synthesize it as video data with the original video signal. Furthermore, the metadata display system is characterized in that the video signal combined with the metadata can be displayed on the display unit of the imaging apparatus.
[0036]
Next, the overall configuration of the metadata display system according to the present embodiment will be described with reference to FIG. FIG. 1 is a block diagram showing a schematic configuration of the metadata display system 1 according to the present embodiment.
[0037]
As shown in FIG. 1, the metadata display system 1 according to the present embodiment includes, for example, an imaging device 10, a sound collection device 18, a camera control unit (hereinafter referred to as CCU) 20, and metadata input. The terminal device 30, the metadata adding device 40, a video tape recorder (hereinafter referred to as VTR) 50, a metadata composition device 60, and a display device 70 are mainly configured.
[0038]
The imaging device 10 is, for example, a video camera that converts an optical image incident on the lens device 12 into an electrical signal, and can capture and image a subject to generate and output a video signal. The imaging apparatus 10 can photograph each scene (scene) constituting a video work and output the generated video signal to, for example, the CCU 20. This video signal may be generated by, for example, a progressive method or an interlace method.
[0039]
In the present embodiment, the transmission of the video signal from the imaging device 10 to the CCU 20 is performed as an optical signal via, for example, an optical fiber cable. By transmitting a video signal as an optical signal in this way, a long distance transmission (for example, about 1 km) is possible as compared with a case of transmitting in HD SDI (High Definition Serial Digital Interface) format (for example, about 50 m). For this reason, since the imaging device 10 and the CCU 20 and the VTR 50 can be disposed sufficiently apart from each other, the degree of freedom in photographing is increased. However, the present invention is not limited to this example, and the imaging device 10 may transmit a video signal using, for example, an HD SDI cable. In this case, for example, the video signal may be directly transmitted from the imaging device 10 to the metadata adding device 40 or the like without providing the CCU 20.
[0040]
The imaging device 10 collects various setting information (shooting condition information such as shutter speed and gain) in the imaging device 10 at the time of shooting, for example, and sets camera setting metadata based on these setting information. Can be generated. Further, for example, the imaging apparatus 10 can group and pack the camera setting metadata as a camera setting group and add it to each frame of the video signal. Details will be described later.
[0041]
Further, for example, a return video signal (details will be described later) is input to the imaging device 10 via the CCU 20 from a metadata synthesis device described later. For example, the imaging apparatus 10 can display the return video signal on the display unit 108. The display unit 108 is configured as one of display devices that display a video signal in which metadata according to the present embodiment is combined. Details will be described later.
[0042]
In addition, the imaging device 10 includes, for example, a lens device 12 and a dolly device 14.
[0043]
The lens device 12 includes, for example, a plurality of lenses and a driving device that adjusts the distance, aperture, and the like of these lenses, and is suitable for the main body of the imaging device 10 by adjusting zoom, iris, focus, and the like. An optical image can be incident. For example, the lens device 12 can generate various setting information (shooting condition information such as zoom, iris, and focus) in the lens device 12 at the time of shooting for each frame as lens setting metadata.
[0044]
The dolly device 14 is a carriage for placing and moving the main body of the imaging device 10, for example, when shooting with the imaging device 10 approaching or moving away from the subject, or imaging with the moving subject. This is used when the apparatus 10 is moved to take a picture. The dolly device 14 can move at high speed along the subject or the like, for example, by placing a pulley provided in the lower part thereof on a rail. The dolly device 14 can generate, for example, various setting information in the dolly device 14 at the time of shooting (shooting condition information such as the position of the dolly and the camera orientation) as the dolly setting metadata for each frame. . The dolly device 14 does not necessarily have to be provided. For example, when the imaging device 10 is installed on a crane or the like for photographing from above, or when a cameraman takes a picture while holding the imaging device 10, It is unnecessary.
[0045]
The lens setting metadata and the dolly setting metadata generated as described above are output to the metadata adding device 40 via, for example, an RS-232C cable.
[0046]
The sound collector 18 is composed of, for example, a microphone and can generate and output an audio signal. More specifically, the sound collecting device 18 collects sound information such as background sounds and voices of actors at the time of photographing by the imaging device 10 to generate a sound signal. This audio signal is output to the VTR 50, for example. The sound collecting device 18 may be included in the imaging device 10.
[0047]
For example, the CCU 20 can convert a video signal input as an optical signal from the photographing apparatus 10 into an HD SDI signal and output the signal to the metadata adding apparatus 40 via the HD SDI cable. The CCU 20 can also acquire camera setting metadata from the video signal via an optical fiber cable or the like, for example. The CCU 20 can also output the return video signal to the imaging device 10 when a return video signal is input from, for example, a metadata synthesis device 60 described later.
[0048]
Note that the CCU 20 does not necessarily have to be provided as a device configured separately from the imaging device 10, and may be incorporated in the imaging device 10, for example. In particular, for example, when the imaging device 10 is configured to output a video signal in, for example, the HD SDI format, the CCU 20 is not an essential device.
[0049]
The metadata input terminal device 30 includes, for example, an information processing device such as a personal computer and its peripheral devices, and can generate scene information metadata based on user input. This scene information metadata is, for example, metadata relating to a scene photographed by the imaging apparatus 10, and is information (scene number, take number, etc.) described in an electronic clapperboard or the like in conventional photographing. For example, when a scene number or the like of a scene to be photographed is input by a director or the like, the metadata input terminal device 30 generates scene information metadata corresponding to the scene number and the like via an RS-232C cable or the like. To the metadata adding device 40. Note that a photographer, director, or the like may additionally input, for example, a comment at the time of recording a video material (such as a note of a shooting situation) as scene situation metadata by using the metadata input terminal device 30. it can.
[0050]
The metadata adding device 40 can add the metadata in units of frames to the video signal, for example. More specifically, the metadata adding device 40 includes, for example, lens setting metadata, dolly setting metadata, and scene information metadata from the lens device 12, the dolly device 14, and the metadata input terminal device 30, respectively. Etc. are entered. For example, the metadata adding apparatus 40 packs these metadata by grouping them into a plurality of metadata groups such as a lens setting group, a dolly setting group, and a scene information group for each purpose of use. Further, for example, the metadata adding device 40 sequentially adds the metadata of the lens setting group, the dolly setting group, and the scene information group that are grouped in this manner to the blanking area of the video signal input from the CCU 20 for each frame. , Can be inserted and added. In this way, the video signal to which all metadata is added is output to the VTR 50 via, for example, an HD SDI cable.
[0051]
The metadata adding device 40 is supplied with a reference signal (reference synchronization signal) from the reference signal generating device 72 and a time code signal (LTC: linear time code) from the time code signal generating device 74. . Also, such LTC can be output to the VTR 50.
[0052]
The VTR 50 is configured as a video signal recording / reproducing apparatus according to the present embodiment. For example, a video signal input from the metadata adding apparatus 40 or an audio signal input from the sound collecting apparatus 18 is converted into a video tape 52. Or the like. The VTR 50 can also reproduce video signals and the like recorded on the video tape 52. The VTR 50 outputs, for example, the video signal input from the metadata adding device 40 to the metadata combining device 60 as it is, or outputs the video signal reproduced from the video tape 52 to the metadata combining device 60. Can do.
[0053]
In this embodiment, the video tape 52 is used as a storage medium. However, the present invention is not limited to this example. For example, any storage medium such as various magnetic tapes, magnetic disks, optical disks, and memory cards may be used. Good. Further, the video signal recording / reproducing apparatus is not limited to the example of the VTR 50, and can be changed to an apparatus (a disk device, various reader / writers, etc.) corresponding to such various storage media.
[0054]
The metadata synthesizing device 60 is a characteristic device according to the present embodiment. For example, the metadata synthesizing device 60 is a decoder device that extracts and decodes the metadata added to the video signal as described above, and synthesizes it to the video signal. is there. More specifically, the metadata composition device 60 can extract, for example, all or part of the metadata added to the video signal input from the VTR 50 in units of frames. Further, the metadata synthesizing device 60 can synthesize the extracted metadata in units of frames after decoding and rewriting the extracted metadata into video data. This combining means, for example, multiplexing (eg, superimposing) the video signal and the video data of the metadata in units of frames.
[0055]
For example, the metadata synthesizing device 60 transmits the video signal (hereinafter referred to as a metadata synthesized video signal) obtained by synthesizing the metadata to the display device 70 and the CCU 20 via an HD SDI cable or the like. Can be output. Among these, for example, the metadata composite video signal transmitted to the imaging device 10 via the CCU 20 functions as a return video signal, for example. This return video signal is a video signal that is returned from the metadata synthesis device 60 to the imaging device 10 after the video signal is recorded by the VTR 50 or the like.
[0056]
As described above, the metadata synthesis device 60 can distribute the monitoring system and return the metadata synthesis video signal to the imaging device 10 as a return video signal, for example. As a result, the imaging apparatus 10 can display a video on which the metadata is superimposed on the display unit 108 based on the received metadata composite video signal.
[0057]
The display device 70 is a display device such as an LCD (Liquid Crystal Display) or a CRT (Cathode Ray Tube). The display device 70 can display a video in which the metadata is superimposed on the basis of the metadata synthesized video signal input from the metadata synthesizer 60.
[0058]
<2. Content of metadata>
Next, grouped metadata according to the present embodiment will be described in detail. In the present embodiment, for example, various metadata related to the video material as described above are grouped into, for example, four metadata groups according to the purpose of use, and transmitted, recorded, and managed. . Hereinafter, for each of these four metadata groups, details of metadata included in the metadata group will be described in detail.
[0059]
<2.1 Scene information group>
First, scene information metadata included in a scene information group will be described in detail with reference to FIG. FIG. 2 is an explanatory diagram showing a specific example of scene information metadata included in the scene information group according to the present embodiment.
[0060]
As shown in FIG. 2, the scene information metadata included in the scene information group includes, for example, “time code”, “scene number”, “take number”, etc., which are conventionally displayed on an electronic clapperboard (slate) or the like. Various kinds of metadata related to a scene taken by the imaging apparatus 10 including information.
[0061]
“Time code” is time information including hours, minutes, seconds, frame numbers, and the like represented by LTC. Conventionally, this “time code” is recorded in the longitudinal direction of an audio track of the video tape 52, for example. In the present embodiment, this “time code” is generated by the time code signal generation device 74 and is attached to the blanking region of the video signal for each frame by the metadata adding device, for example. The position of the video signal can be specified by this time code. The data amount of this “time code” is, for example, 16 bytes.
“Date” is text information indicating the date on which the image was taken, and the amount of data is, for example, 4 bytes.
“Video work title” is text information representing the title of the video work, and the amount of data is, for example, 30 bytes.
“Photographing team number” is an ID number or the like for identifying the photographing team (crew) in charge of the photographing, and the data amount is 2 bytes, for example.
The “scene number” is a number for specifying a scene in which shooting is performed among a plurality of scenes (Scene: shooting scene) constituting the video work, and the data amount is, for example, 2 bytes. . By referring to the “scene number”, it is possible to identify what scene in the video work corresponds to the captured video material. Note that, for example, a cut number obtained by further subdividing a scene can be added as scene information metadata.
The “take number” is a number for specifying a take that is a continuous video unit from the start of recording once to the end of recording by the imaging apparatus 10, and the data amount is, for example, 2 bytes. is there. By referring to the “take number”, it is possible to identify which take belongs to which scene the recorded video signal corresponds to.
“Roll number” is a number for specifying a roll (Roll), which is a video unit obtained by further subdividing the take, and its data amount is, for example, 2 bytes.
“Cameraman”, “Director”, and “Producer” are text information representing the name of the cameraman, director, and producer who are in charge of shooting, respectively, and the amount of these data is, for example, 16 bytes.
[0062]
In this way, in the scene information group, for example, metadata that can be attribute information and index information of the recorded video is collected. This scene information metadata is useful information for grasping the content of the video material, identifying and managing the video material, for example, in the video recording stage, the post-processing stage, and the editing stage.
[0063]
<2.2 Camera Setting Group>
Next, the camera setting metadata included in the camera setting group will be described in detail with reference to FIG. FIG. 3 is an explanatory diagram showing a specific example of camera setting metadata included in the camera setting group according to the present embodiment.
[0064]
As shown in FIG. 3, the camera setting metadata included in the camera setting group is, for example, metadata representing various shooting conditions mainly including setting information of the imaging device 10 when a video is shot.
[0065]
“Camera ID” is a serial number (device number) for specifying the imaging device 10 that has performed the imaging process, and the data amount is, for example, 4 bytes.
“CHU switch ON / OFF” is bit information indicating whether or not the setting of the imaging apparatus 10 is changed from the standard setting as described below, and the data amount is, for example, 1 byte.
“CCU ID” is a serial number (device number) for specifying the CCU 20 that has performed the imaging process, and its data amount is, for example, 4 bytes.
“Filter setting” is information indicating the filter setting of the imaging apparatus 10 at the time of shooting, and the data amount is, for example, 2 bytes. In the present embodiment, for example, the imaging apparatus 10 is provided with five types of filters in duplicate, and represents which of these two filters are combined and photographed.
“Shutter speed” is information indicating the setting value of the shutter speed of the imaging apparatus 10 at the time of shooting, and the data amount is, for example, 1 byte. In the present embodiment, this “shutter speed” can be set in six stages, for example, between “1/100” and “1/2000” seconds.
“Gain” is information representing the installation value of the gain of the imaging apparatus 10 at the time of shooting, and the amount of data is, for example, 1 byte.
“ECS” is information indicating ON / OFF of the ECS (Extended Clear Scan) function of the imaging apparatus 10 at the time of shooting, and the data amount is, for example, 2 bytes.
“Gamma (master)” is information indicating the setting of the gamma characteristic (gamma curve or the like) of the imaging apparatus 10 at the time of shooting, and the data amount is, for example, 2 bytes.
“Gamma (user setting)” is information indicating the setting of the gamma characteristic when the gamma curve or the like is changed by the user setting, and the data amount is, for example, 1 byte.
“Variable frame rate” is information indicating the frame rate setting value of the video signal shot by the imaging device 10 capable of variable speed shooting, and the data amount is, for example, 1 byte. The imaging apparatus 10 according to the present embodiment can be photographed by changing the frame rate at, for example, 23.98 to 30P, but is not limited to such an example, and is configured to be capable of variable speed photographing at 1 to 60P, for example. May be.
“Video signal white level” is information representing the white level setting value of the video signal by the white balance adjustment processing of the imaging apparatus 10 at the time of shooting, and the data amount is, for example, 6 bytes.
“Video signal black level” is information indicating the set value of the black level of the video signal by the black balance adjustment processing of the imaging apparatus 10 at the time of shooting, and the data amount is, for example, 8 bytes.
“Detail level” is information representing a setting value of the detail level by the detail adjustment processing of the imaging apparatus 10 at the time of shooting, and the data amount is, for example, 2 bytes.
“Knee point” is information indicating the set value of the knee point of the video signal compressed by the knee circuit of the imaging apparatus 10 at the time of shooting, and the amount of data is, for example, 2 bytes.
“Need slope” is information indicating the set value of the knee slope of the video signal compressed by the knee circuit of the imaging device 10 at the time of shooting, and the amount of data is, for example, 2 bytes.
“Recorder status” is information indicating the setting value of the frame rate when the video signal recording / reproducing apparatus such as the VTR 50 records the video signal, and the data amount is, for example, 1 byte. The “recorder status” is determined in correspondence with the “variable frame rate”.
[0066]
In this way, in the camera setting group, for example, metadata regarding shooting conditions such as setting information of the imaging device 10 at the time of shooting is collected. The camera setting metadata is useful information for grasping the image quality (brightness, hue, texture, etc.) of the video material at the post-processing stage of the video material, for example.
[0067]
<2.3 Lens setting group>
Next, based on FIG. 4, the lens setting metadata included in the lens setting group will be described in detail with a specific example. FIG. 4 is an explanatory diagram showing a specific example of lens setting metadata included in the lens setting group according to the present embodiment.
[0068]
As shown in FIG. 4, the lens setting metadata included in the lens setting group is, for example, metadata representing various shooting conditions mainly including setting information of the lens device 12 at the time of video shooting.
[0069]
“Zoom” is information representing a zoom setting value obtained by the photographing magnification adjustment processing of the lens apparatus 12 at the time of photographing, and the data amount is, for example, 2 bytes.
“Focus” is information representing a focus setting value by the focal length adjustment processing of the lens device 12 at the time of photographing, and the data amount is, for example, 2 bytes.
“Iris” is information representing an iris (aperture) setting value by exposure adjustment processing of the lens device 12 at the time of photographing, and the data amount is, for example, 2 bytes.
“Lens ID” is a serial number (device number) for specifying the lens device 12 used for photographing, and the data amount is, for example, 4 bytes.
[0070]
In this way, in the lens setting group, for example, metadata relating to shooting conditions such as setting information of the lens device 12 at the time of shooting is collected. The lens setting metadata is useful information for grasping, for example, the movement of the subject photographed with the video material, the distance from the imaging device 10 and the like in the post-processing stage of the video material.
[0071]
<2.4 Dolly Setting Group>
Next, the dolly setting metadata included in the dolly setting group will be described in detail with reference to FIG. FIG. 5 is an explanatory diagram showing a specific example of the camera setting metadata included in the dolly setting group according to the present embodiment.
[0072]
As shown in FIG. 5, the dolly setting metadata included in the dolly setting group is, for example, metadata representing various shooting conditions and the like mainly including setting information of the dolly device 14 at the time of video shooting.
“GPS” is latitude and longitude information (Global Positioning System information) for specifying the position of the dolly device 14 (that is, the position of the imaging device 10) at the time of shooting, and the data amount is, for example, 12 bytes. .
“Moving direction” is information indicating the moving direction of the dolly device 14 at the time of shooting (that is, the moving direction of the imaging device 10) by an angle, and the data amount is, for example, 4 bytes.
“Movement speed” is information indicating the movement speed of the dolly device 14 at the time of photographing (that is, the movement speed of the imaging device 10), and the data amount is, for example, 4 bytes.
“Camera direction” is information indicating the shooting direction of the imaging apparatus 10 and is expressed by the rotation angle (angle of swinging) of the imaging apparatus 10 with the fixed dolly device 14 as a reference. Specifically, for example, the imaging direction of the imaging apparatus 10 is “pan” (Z-axis direction), “tilt” (Y-axis direction), “roll” (X-axis direction). This is expressed as a rotation angle in three directions. Each of these three data amounts is, for example, 2 bytes.
“Dolly height” is information indicating the height of the dolly device 14, and its data amount is, for example, 2 bytes. With this information, the position of the imaging device 10 in the vertical direction can be specified.
“Dori ID” is a serial number (device number) for specifying the dolly device 14 used for photographing, and its data amount is, for example, 4 bytes.
[0073]
In this way, in the dolly setting group, for example, metadata regarding shooting conditions including setting information such as the position and movement of the dolly device 14 at the time of shooting is collected. This dolly setting metadata is also useful information for grasping the movement, distance, etc. of the subject appearing in the video material in the post-processing stage of the video material, for example, as in the lens setting metadata. .
[0074]
The contents of, for example, four metadata groups according to this embodiment have been described above. By grouping metadata in this way, it is possible to suitably extract only necessary metadata in units of groups according to the purpose of use of metadata, and use, rewrite, or the like.
[0075]
For example, in the video recording stage, the metadata of the scene information group including the scene number and time code is extracted and used for the purpose of identifying and grasping the video that is being recorded or has been recorded. Further, in the post-processing stage of the video material, the metadata of the camera, lens, and dolly setting group is useful when a CG video is synthesized with a live-action video. Specifically, the metadata of the camera setting group is extracted and used for the purpose of grasping the image quality of the video material. Further, the metadata of the lens setting group and the dolly setting group is extracted and used for the purpose of grasping the movement of the subject in the video material.
[0076]
As described above, the purpose of using the metadata of the lens setting group and the dolly setting group is common. For this reason, the lens setting group and the dolly setting group are not configured as separate groups as in the present embodiment. For example, the lens setting metadata and the dolly setting metadata are configured as one lens / dolly setting group. They may be grouped together into one.
[0077]
<3. Metadata format>
Next, a metadata format according to this embodiment will be described with reference to FIG. FIG. 6 is an explanatory diagram for explaining a metadata format according to the present embodiment.
[0078]
As described above, the metadata according to the present embodiment is grouped into, for example, four metadata groups. The metadata grouped in this way is added to the video signal in a predetermined format by the imaging device 10 and the metadata adding device 40, for example.
[0079]
More specifically, as shown in FIG. 6A, the metadata is packaged as ancillary data in, for example, an ancillary data area in the vertical blanking area of the video signal and is frame-by-frame. Inserted. The format of the packaged metadata at the time of transmission, for example, is shown in FIG.
[0080]
As shown in FIG. 6B, the metadata is grouped into four metadata groups, for example, a scene information group, a camera setting group, a lens setting group, and a dolly setting group, and these four metadata groups are continuous. And have a format arranged in series. Each metadata group is KLV (Key Length Value) encoded based on, for example, SMPTE (Society of Motion Picture and Television Engineers) standard (SMPTE291M or the like).
[0081]
“K (Key)” is, for example, a 1-byte key ID (reserved word) assigned to the top of each metadata group. This “K” code is configured as group identification information according to the present embodiment, and functions as a code for identifying each metadata group. For example, in any frame of the video signal, for example, “01” is always given to the scene information group, “02” is always given to the camera setting group, and “02” is always given to the lens setting group. For example, “03” is always given, and “04” is always given to the lens setting group, for example, so that a unique identification code can be unified for each metadata group. In this way, by assigning a “K” code that is unique group identification information to each metadata group, based on the group identification information, only a specific metadata group is selected from a plurality of metadata groups. Each frame can be easily extracted.
[0082]
“L (Length)” is, for example, a 1-byte length code provided after the “K” code. This “L” code is configured as data amount information according to the present embodiment, and functions as a code representing the data amount of the subsequent metadata group. For example, if “L” attached to a scene information group of a certain frame is “124”, this indicates that the data amount of the scene information group in that frame is, for example, 124 bytes. In this way, by adding an “L” code that is data amount information before the contents of each metadata group, the processing efficiency of the metadata extraction or rewriting process is improved. In other words, metadata processing devices such as the metadata adding device 40 and the VTR 50 grasp the data amount of metadata to be extracted or rewritten in advance by referring to the “L” code that is the data amount information. it can. For this reason, the processing efficiency of the extraction or rewriting process is improved.
[0083]
“Element” is, for example, a user data area (Value area) in which metadata of each actual metadata group is stored, and has a variable length.
[0084]
In addition, in front of the metadata group encoded in this way, an “Ancillary Data Flag” ancillary data flag, which is a flag for defining and identifying the type of metadata to be transmitted, and “DID: Data identification” A data ID, “SDID: Secondary Data Identification” secondary data ID, “DC: Data Counter” data counter, and the like are attached. On the other hand, codes such as “CRC: Cyclic Redundancy Check” and “CHECK SUM” for error detection at the time of transmission are attached after the metadata group.
[0085]
By the way, in the SMPTE standard, when packing KLV-encoded metadata into an ancillary data area of a video signal and inserting it, the standard packet size of ancillary data is 255 bytes. . Therefore, in the metadata format according to the present embodiment, the total data amount of the grouped metadata is adjusted to be, for example, 255 bytes or less so as to conform to this standard. Specifically, for example, the metadata amount of the scene information group is 124 bytes or less, the metadata amount of the camera setting group is 40 bytes or less, the metadata amount of the lens setting group is 10 bytes or less, For example, the metadata amount is adjusted to be 52 bytes or less. For this reason, one packet size of the ancillary data is set to be, for example, about 226 bytes or less in the total amount of metadata.
[0086]
As described above, in the metadata format according to the present embodiment, all the metadata is set to fit within one packet size (255 bytes) of the ancillary data. However, the present invention is not limited to this example. For example, a plurality of ancillary data packets may be concatenated, and metadata may be divided and packed into the plurality of packets.
[0087]
As described above, in the metadata format according to the present embodiment, for example, the area allocated to the metadata is divided according to the number of metadata groups, and the metadata of each metadata group is divided into each divided area. It is the structure which inserts. Further, the group identification information and the data amount information are given to the top of each metadata group. With this configuration, it is possible to quickly, easily detect, extract, or rewrite metadata necessary for the purpose of using metadata for each metadata group. For example, by sharing the group identification information as a common identification ID between the recording department and the editing department of the video work, the metadata can be suitably used in the production process of the video work.
[0088]
<4. Configuration of each device>
Next, main devices constituting the metadata display system 1 as described above will be described in detail.
[0089]
<4.1 Imaging device>
First, the imaging apparatus 10 according to the present embodiment will be described in detail based on FIG. FIG. 7 is a block diagram illustrating a configuration of the imaging apparatus 10 according to the present embodiment.
[0090]
As illustrated in FIG. 7, the imaging device 10 includes, for example, a CPU 100, a memory unit 102, an imaging unit 104, a signal processing unit 106, a display unit 108, a camera setting metadata generation unit 110, and metadata addition. Unit 112, transmission / reception unit 114, lens device 12, and dolly device 14.
[0091]
A CPU (Central Processing Unit) 100 functions as an arithmetic processing device and a control device, and can control processing of each unit of the imaging device 10. The memory unit 102 includes, for example, storage devices such as various RAMs, ROMs, flash memories, and hard disks, and has a function of storing or temporarily storing various data related to the processing of the CPU 100, operation programs of the CPU 100, and the like. Have
[0092]
The imaging unit 104 is configured by, for example, OHB (Optical Head Base), and has a function of imaging a subject and generating a video signal. More specifically, the imaging unit 104, for example, splits an optical image incident from the lens device 12 into R, G, and B by a prism (not shown), and uses various filters (not shown). Then, the image is photoelectrically converted at a predetermined shutter speed by an imaging device (not shown) such as a CCD (Charge Coupled Device) to generate a video signal that is an analog electrical signal.
[0093]
The signal processing unit 106 performs gain adjustment (AGC) processing, correlated double sampling processing, A / D conversion processing, error correction processing, white correction on a video signal that is a weak analog electric signal input from the imaging unit 104. Digital video signals can be output by performing balance adjustment processing, dynamic range compression processing, gamma correction processing, shading correction processing, detail adjustment processing, knee processing, and the like. In the present embodiment, for example, an HD (High Definition) digital video signal is generated and output. The signal processing unit 106 can also convert the digital video signal into an analog video signal and output the analog video signal to the display unit 108, for example. In addition, the signal processing unit 106 can change the frame rate of the video signal to be output (for example, 23.98 to 30P) based on, for example, preset conditions or based on a cameraman input operation. .
[0094]
The display unit 108 is, for example, a viewfinder for a cameraman to view a subject, and includes a CRT monitor. The display unit 108 can display and output, for example, an analog video signal input from the signal processing unit 106. The display unit 108 may be composed of various display devices such as an LCD monitor, for example.
[0095]
The camera setting metadata generation unit 110 acquires and manages parameters such as setting information of the imaging unit 104 and signal processing setting information such as gamma, knee, and detail in the signal processing unit 108, for example. Further, the camera setting metadata generation unit 110 generates the camera setting metadata, for example, for each frame of the video signal based on the parameters, and outputs the generated camera setting metadata to the metadata addition unit 112.
[0096]
For example, the metadata adding unit 112 can add camera setting metadata to the video signal frame by frame in accordance with the timing of outputting the video signal to the outside of the imaging device 10. Specifically, the metadata adding unit 112 performs KLV encoding and packing, for example, camera setting metadata input from the camera setting metadata generation unit 110. Further, the metadata adding unit 112 adds the packed camera setting metadata to the area assigned to the camera setting group in the blanking area of the video signal as shown in FIG. Are inserted sequentially.
[0097]
At this time, the metadata adding unit 112 inserts dummy data into areas corresponding to the scene information group, the lens setting group, and the dolly setting group other than the camera setting group, for example, as shown in FIG. Can be kept.
[0098]
The camera setting metadata generation unit 108 and the metadata addition unit 110 as described above may be configured as hardware, for example, or may be configured as software that realizes the above processing functions. May be stored in the memory unit 102 and the CPU 100 may perform actual processing.
[0099]
For example, the transmission / reception unit 114 transmits the video signal to which the camera setting metadata is added as described above to the CCU 20 via the optical fiber cable. Further, a metadata composite video signal is input to the transmission / reception unit 114 as the return video signal from the metadata synthesis device 60. The transmission / reception unit 114 outputs the metadata composite video signal to the signal processing unit 106. Thereby, the signal processing unit 106 converts the metadata composite video signal into, for example, an analog video signal and outputs the analog video signal to the display unit 108. As a result, the display unit 108 can display an image on which the metadata is superimposed.
[0100]
The lens device 12 includes, for example, an optical block 122, a drive system block 124, and a lens setting metadata generation unit 124.
[0101]
The optical system block 122 includes, for example, a plurality of lenses, a diaphragm, and the like, and can make an optical image from a subject incident on the imaging unit 104. The drive system block 124 can adjust zoom, iris, focus, and the like by adjusting the distance between the lenses and the aperture of the optical system block 122, for example.
[0102]
The lens setting metadata generation unit 126 acquires and manages parameters such as lens setting information of the drive system block 124, for example. Further, the lens setting metadata generation unit 126 generates the lens setting metadata, for example, for each frame based on such parameters. The lens setting metadata generated in this way is output to the metadata adding device 40 via, for example, an RS-232C cable.
[0103]
The dolly device 14 includes, for example, a dolly measurement unit 142 and a dolly setting metadata generation unit 144.
[0104]
The dolly measuring unit 142 measures various setting information related to the dolly device 14 such as GPS information, the moving speed and direction of the dolly device 14, and the angle of the imaging device 10, and outputs them to the dolly setting metadata generating unit 144. To do.
[0105]
For example, the dolly setting metadata generation unit 144 generates the dolly setting metadata for each frame, for example, based on the measurement information from the dolly measurement unit 142. The dolly setting metadata generated in this way is output to the metadata adding device 40 via, for example, an RS-232C cable.
[0106]
<4.2 Camera control unit>
Next, the CCU 20 according to the present embodiment will be described in detail based on FIG. FIG. 9 is a block diagram showing the configuration of the CCU 20 according to the present embodiment.
[0107]
As illustrated in FIG. 9, the CCU 20 includes, for example, a CPU 200, a memory unit 202, a transmission / reception unit 204, a signal processing unit 206, a serializer 208, and a deserializer 210.
[0108]
The CPU 200 functions as an arithmetic processing unit and a control unit, and can control processing of each unit of the CCU 20. A reference signal is input to the CPU 200, and video signals can be synchronized with other devices in the metadata display system 1. The memory unit 202 includes, for example, storage devices such as various RAMs, ROMs, flash memories, and hard disks, and has a function of storing or temporarily storing various data related to the processing of the CPU 200, operation programs of the CPU 200, and the like. Have
[0109]
For example, the transmission / reception unit 204 receives a video signal to which camera setting metadata is added from the imaging device 10 and transmits the video signal to the signal processing unit 206.
[0110]
The signal processing unit 206 converts, for example, a video signal input as an optical signal into an HD SDI signal and outputs it to the serializer 208. The signal processing unit 206 can also be configured to have the processing function of the signal processing unit 106 of the imaging apparatus 10.
[0111]
For example, the serializer 208 performs parallel-serial conversion on the video signal received from the signal processing unit 206 and transmits the video signal to the metadata adding device 40 via the HD SDI cable. In the blanking area of the video signal output from the CCU 20, actual metadata is inserted only in the area corresponding to the camera setting group, for example, as shown in FIG. Dummy data is inserted in the area of the metadata group.
[0112]
For example, the deserializer 210 performs serial-parallel conversion on the return video signal transmitted from the metadata synthesis device 60 and outputs the result to the signal processing unit 206.
[0113]
Thereby, the signal processing unit 206 converts, for example, a return video signal input as an HD SDI signal into an optical signal, and outputs the optical signal to the transmission / reception unit 204. The transmission / reception unit 204 transmits the return video signal to the imaging apparatus 10 via, for example, an optical fiber cable.
[0114]
<4.3 Metadata adding device>
Next, the metadata adding device 40 according to the present embodiment will be described in detail with reference to FIG. FIG. 10 is a block diagram showing the configuration of the metadata adding device 40 according to this embodiment.
[0115]
As shown in FIG. 10, the metadata adding device 40 includes, for example, a CPU 400, a memory unit 402, a metadata packing unit 404, a metadata encoder 406, a deserializer 408, a metadata insertion unit 410, and a serializer 412. And.
[0116]
The CPU 400 functions as an arithmetic processing device and a control device, and can control processing of each unit of the metadata adding device 40. A reference signal is input to the CPU 400, and the video signal can be synchronized with other devices in the metadata display system 1. Further, a time code signal (LTC) is input to the CPU 400, and time code information that is one of scene information metadata can be generated based on the LTC and stored in the memory unit 402. . Also, such LTC can be output to the VTR 50.
[0117]
The memory unit 402 includes, for example, storage devices such as various RAMs, ROMs, flash memories, and hard disks, and functions to store or temporarily store various data related to the processing of the CPU 400, operation programs of the CPU 400, and the like. Have The memory unit 402 includes a metadata buffer memory 403 for temporarily storing metadata transmitted from each device, for example.
[0118]
The metadata buffer memory 403 includes, for example, lens setting metadata sequentially transmitted from the lens device 12 after the start of photographing, dolly setting metadata sequentially transmitted from the dolly device 14 after starting photographing, and a metadata input terminal. The scene information metadata previously acquired from the apparatus 30 before the start of photographing, time code information input from the CPU 400, and the like are stored.
[0119]
The metadata packing unit 404 extracts, for example, necessary metadata from various kinds of metadata stored in the metadata buffer memory 403, and sets a lens setting group, a dolly setting group, a scene for each purpose of use. The data is grouped into a plurality of metadata groups such as an information group and repacked into the KLV structure. The metadata packing unit 404 outputs the metadata packed in this way to the metadata encoder 406.
[0120]
The metadata encoder 406 encodes the metadata from the metadata packing unit 404. The metadata input to the metadata adding device 40 as described above is, for example, data in the RS-232C protocol format. Therefore, for example, the metadata encoder 406 converts the format into an ancillary data packet format and encodes the metadata so that it can be inserted into an HDSDI video signal (see FIG. 6). By this encoding, for example, various flags and CRC as described above are added before and after the metadata.
[0121]
The deserializer 408 performs serial-parallel conversion on the video signal input from the CCU 20 and outputs the video signal to the metadata insertion unit 410.
[0122]
The metadata insertion unit 410 sequentially inserts the metadata input from the metadata encoder 406 into the blanking area of the video signal input from the deserializer 408 for each frame.
[0123]
At this time, in the video signal input to the metadata insertion unit 410, for example, an area corresponding to the camera setting group in the blanking area is preliminarily set by the imaging device 10 as shown in FIG. The camera setting metadata of the camera setting group is inserted.
[0124]
On the other hand, dummy data is inserted in areas corresponding to the scene information group, the lens setting group, and the dolly setting group other than the camera setting group. For this reason, as shown in FIG. 8B, the metadata insertion unit 410, for example, rewrites the dummy data into actual scene information metadata, lens setting metadata, dolly setting metadata, and the like. The metadata can be inserted into the video signal. At the time of the metadata rewriting process, the metadata insertion unit 410, for example, based on the group identification information “K” and the data amount information “L” assigned to the corresponding area of each metadata group, Since the detection and rewriting process is performed, the rewriting process can be performed efficiently. Further, when inserting metadata in this way, the metadata insertion unit 410 can also perform, for example, phase alignment of the delay time between the metadata to be inserted and the video signal.
[0125]
The serializer 412 performs parallel-serial conversion on the video signal to which the metadata is added for each frame by the metadata insertion unit 410 as described above, and transmits the video signal to the VTR 50.
[0126]
As described above, the metadata adding apparatus 40 according to the present embodiment further adds scene information metadata, lens setting metadata, and dolly setting metadata to a video signal to which camera setting metadata is added in advance. It can be added in addition.
[0127]
<4.4 Video tape recorder>
Next, the VTR 50 according to the present embodiment will be described in detail based on FIG. FIG. 11 is a block diagram showing the configuration of the VTR 50 according to the present embodiment.
[0128]
As shown in FIG. 11, the VTR 50 includes, for example, a CPU 500, a memory unit 502, a deserializer 504, a signal processing unit 506, a metadata decoder 508, a recording / reproducing block 510, an ECC block 512, and a metadata encoder. 514 and a serializer 516.
[0129]
The CPU 500 functions as an arithmetic processing unit and a control unit, and can control processing of each unit of the VTR 50. A time code signal (LTC) signal is input to the CPU 500. The memory unit 502 includes, for example, storage devices such as various RAMs, ROMs, flash memories, and hard disks, and has a function of storing or temporarily storing various data related to processing of the CPU 500, operation programs of the CPU 500, and the like. Have
[0130]
The deserializer 504 performs serial-parallel conversion on the video signal input from the metadata adding device 40 and outputs the video signal to the signal processing unit 506.
[0131]
For example, the signal processing unit 506 can perform various processes on the video signal in order to suitably record / reproduce the video signal and the like on the video tape 52. For example, the signal processing unit 506 can compress / decompress the video signal based on MPEG (Moving Picture Experts Group phase) 1, MPEG2, MPEG4, or DCT (Discrete Cosine Transform) system, as necessary. In addition, the signal processing unit 506 can adjust the timing of recording / reproduction of each of the above signals, or can separate the video signal and the audio signal and add an ECC (Error Correcting Code). . Further, the signal processing unit 506 can extract, for example, metadata added to the video signal in units of frames, or conversely, can insert decoded metadata in units of frames in the video signals.
[0132]
For example, the signal processing unit 506 can directly output the video signal input from the metadata adding device 40 to the serializer 514 or output the video signal reproduced from the video tape 52 to the serializer 514.
[0133]
The metadata decoder 508 decodes, for example, metadata extracted from the video signal. Specifically, the metadata decoder 508 removes, for example, flags (Flag, DID, SDID, etc.) and CRC attached to the metadata that are unnecessary for recording, and outputs them to the CPU 500. For example, the CPU 500 adds ECC to the metadata in the same manner as the video signal and outputs the metadata to the recording / reproducing block 510.
[0134]
The recording / reproducing block 510 is composed of, for example, a video head and a driving mechanism (both not shown). The recording / reproducing block 510 can actually record / reproduce the video signal to which the metadata is added to the video tape 52. More specifically, the recording / reproducing block 510 can record, for example, a video signal, an audio signal, and metadata in units of one frame and sequentially record them in the recording area of the video tape 52. In addition, the recording / reproducing block 510 can sequentially reproduce, for example, a set of video signals, audio signals, and metadata recorded in the recording area of the video tape 52 in units of one frame.
[0135]
For example, the ECC block 512 performs error detection on a video signal or the like reproduced from the video tape 52 by the recording / reproducing block 510 based on the ECC. After the error detection is completed, the ECC block 512 outputs the reproduced metadata to the metadata encoder 514 via the CPU 500 and the video signal and the audio signal to the signal processing unit 506, for example.
[0136]
The metadata encoder 514 encodes the reproduced metadata in a transmission format (with the above flag, CRC, etc.) and outputs the encoded metadata to the signal processing unit 506. For example, the signal processing unit 506 combines the video signal and audio signal input from the ECC block 512 with the metadata encoded by the metadata encoder 514 and outputs the combined signal to the serializer 516.
[0137]
The serializer 516 performs parallel-serial conversion on the video signal or the like input from the signal processing unit 506 and transmits the converted video signal to the metadata synthesis device 60.
[0138]
As described above, the signal processing unit 506, the metadata decoder 508, the CPU 500, the recording / reproducing block 510, the ECC block 512, the metadata encoder 514, and the like are configured as the recording / reproducing unit according to the present embodiment. The video signal to which the data is added can be recorded / reproduced with respect to the storage medium.
[0139]
<4.5 Metadata synthesizer>
Next, based on FIG. 12, the metadata synthesizing apparatus 60 which is a feature according to the present embodiment will be described in detail. FIG. 12 is a block diagram showing the configuration of the metadata synthesis apparatus 60 according to this embodiment.
[0140]
As shown in FIG. 12, the metadata composition apparatus 60 includes, for example, a CPU 600, a memory unit 602, a deserializer 604, a metadata extraction unit 606, a metadata decoder 608, a metadata video generation unit 610, A data video composition unit 612 and a serializer 614 are provided.
[0141]
The CPU 600 functions as an arithmetic processing device and a control device, and can control processing of each unit of the metadata composition device 60. The memory unit 602 includes, for example, storage devices such as various RAMs, ROMs, flash memories, and hard disks, and stores or temporarily stores various data related to the processing of the CPU 600, operation programs of the CPU 600, and the like. Have
[0142]
The deserializer 604 performs serial-parallel conversion on the video signal input from the VTR 50 and outputs it to the metadata extraction unit 606.
[0143]
For example, the metadata extraction unit 606 extracts the metadata inserted in the blanking area of the video signal for each frame. At this time, for example, the metadata extraction unit 606 does not extract all the metadata inserted in the blanking area, but extracts only the data that can be consumed, for example, a specific metadata group (for example, a scene information group). In addition, only specific metadata (for example, time code, scene number, and take number) in the metadata group may be extracted. In the metadata synthesizing apparatus 60 according to the present embodiment, for example, by setting in advance what kind of metadata is to be extracted, for example, extracting necessary arbitrary metadata from various metadata. Can do.
[0144]
At the time of such metadata extraction processing, the metadata extraction unit 606 performs the extraction of the metadata group to be extracted based on the group identification information “K” and the data amount information “L” given to each metadata group. As a result, the required metadata can be extracted efficiently.
[0145]
For example, the metadata extraction unit 606 outputs the metadata extracted in this way to the metadata decoder 608, while outputting the video signal to the metadata video composition unit 612 as it is.
[0146]
For example, the metadata decoder 608 decodes the metadata input from the metadata extraction unit 606 and outputs the decoded metadata to the metadata video generation unit 610.
[0147]
For example, the metadata video generation unit 610 can rewrite the metadata input from the metadata decoder 608 into video data for superimposition. That is, since the metadata decoded by the metadata decoder 608 is, for example, metadata in a text data format, the metadata video generation unit 610 converts this metadata into a video data format.
[0148]
For example, the metadata video synthesis unit 612 can sequentially synthesize the metadata converted into video data by the metadata video generation unit 610 in units of frames with respect to the video signal input from the metadata extraction unit 606. it can. In other words, for example, the metadata adding unit 612 can superimpose the video signal metadata by multiplexing the video data into a frame unit. As described above, the metadata video synthesis unit 612 synthesizes the metadata extracted from the video signal as a video with the video signal to generate the metadata synthesized video signal.
[0149]
For example, the serializer 612 performs parallel-serial conversion on the metadata composite video signal input from the metadata video composite unit 612. Further, the serializer 612 transmits the converted metadata composite video signal to the display device 70. Further, the serializer 612 transmits the metadata composite video signal as a return video signal to the imaging device 10 via the CCU 20, for example.
[0150]
As described above, the metadata synthesizing device 60 extracts, for example, the metadata inserted in the blanking area from the video signal being shot by the imaging device 10 or the video signal reproduced by the VTR 50, and The video signal can be combined (superimposed) as video data. Thereby, the metadata synthesizing device 60 can generate a metadata synthesized video signal in which the metadata is superimposed. Furthermore, the metadata composition device 60 can transmit the metadata composition video signal to the display device 70, for example.
[0151]
As a result, the display device 70 can acquire such a metadata composite video signal and display a video in which the metadata is superimposed on the subject. As a result, the director or the like can browse, for example, the video being recorded by the imaging device 10 or the video reproduced by the VTR 50 after recording, along with metadata relating to the video.
[0152]
Further, the metadata synthesis device 60 can also transmit the metadata synthesis video signal to the imaging device 10 as a return video signal. As a result, for example, the imaging device 10 can receive the metadata composite video signal as a return video signal corresponding to a video signal that is currently being shot or has been shot in the past. Furthermore, the imaging apparatus 10 can display a video in which the metadata is superimposed on the display unit 108 such as a viewfinder based on the metadata composite video signal. Thereby, for example, a cameraman or the like can browse a video being recorded by the imaging device 10 or a video reproduced by the VTR 50 after the recording, together with metadata about the video.
[0153]
<5. Video recording method>
Next, a video recording method according to the present embodiment using the metadata display system 1 as described above will be described with reference to FIG. FIG. 13 is a timing chart for explaining the video recording method according to the present embodiment.
[0154]
As shown in FIG. 13A, when shooting is started, first, raw images are sequentially incident on the imaging device 10. Then, the imaging apparatus 10 sequentially generates video signals in units of frames of 0 frame, 1 frame, 2 frames,. At this time, the CCD or the like of the imaging device 10 scans the video, for example, by a progressive method, for example. For this reason, the video signal output from the imaging device 10 has a delay of, for example, about one frame with respect to the raw video incident on the imaging device 10. As a result, as shown in FIG. 13B, the output of the CCU 20 is also delayed by about one frame, for example.
[0155]
At substantially the same time as the generation of the video signal, the imaging device 10 generates camera setting metadata for each frame, and, as shown in FIG. 13B, 1 in the blanking area of the video signal of the corresponding frame. Insert sequentially for each frame. Thereby, the imaging device 10 can add the metadata of the camera setting group to the video signal in units of frames while executing the imaging process to generate the video signal.
[0156]
At the same time as the imaging process of the video apparatus 10, the lens apparatus 12 and the dolly apparatus 14 collect setting information at the time of the imaging process, and set the lens setting metadata and the dolly setting metadata to, for example, 1 Each frame is generated and sequentially output to the metadata adding device 40.
[0157]
Further, video signals generated by the imaging device 10 and having camera setting metadata added for each frame are sequentially input to the CCU 20. The CCU 20 sequentially outputs the video signal to the metadata adding device 40 as shown in FIG.
[0158]
Further, as shown in FIG. 13C, the metadata adding device 40, for example, stores scene information metadata, lens setting metadata, and dolly setting metadata in the blanking region of the video signal input from the CCU 20. , One frame at a time. In addition, the metadata adding device 40 adds time code information as one piece of scene information metadata for each frame to the video signal, for example. In this way, the metadata adding device 40 adds the metadata grouped according to the utilization object to the video signal in units of frames in parallel with the photographing process by the imaging device 10. be able to.
[0159]
Further, as shown in FIG. 13 (d), for example, video signals to which metadata is added are sequentially input from the metadata adding device 40 and audio signals are sequentially input from the sound collecting device 18 to the VTR 50. It will be. For example, the audio signal is temporarily stored in the memory unit 502 and recorded in synchronization with the video signal in accordance with the delay of the video signal. The VTR 50 decodes the metadata of the video signal and then records the video signal and the synchronized audio signal in units of frames on the video tape 52 and the like.
[0160]
As described above, in the video recording method according to the present embodiment, for example, various types of metadata are generated and grouped while performing the shooting process by the imaging device 10, and the video signal generated by the shooting process is processed. Thus, the grouped metadata can be added to each frame and recorded on a storage medium.
[0161]
<6. Metadata display method>
Next, a metadata display method according to the present embodiment using the metadata display system 1 as described above will be described with reference to FIGS. 14 and 15 are explanatory diagrams illustrating specific examples of the return video displayed on the display unit 108 of the imaging apparatus 10 according to the present embodiment.
[0162]
<6.1 Metadata display method during shooting / recording>
First, a process of displaying the video signal being recorded on the display unit 108 of the imaging apparatus 10 during the imaging process by the imaging apparatus 10 and the recording process by the VTR 50 will be described.
[0163]
As described above with reference to FIG. 13, during shooting by the imaging device 10, the metadata adding device 40 adds metadata in units of frames to the video signal generated by the imaging device 10. The video signal to which the metadata is added in this way is input to the metadata synthesizing device 60 in real time via the VTR 50, for example. Then, for example, the metadata synthesis device 60 sequentially extracts and decodes predetermined metadata from the video signal in units of frames based on the set extraction condition, and further converts the metadata into video data. The video signal is sequentially synthesized in units of frames to generate and output the metadata synthesized video signal. The metadata composite video signal is input to the imaging device 10 as a return video signal, for example.
[0164]
In this way, when the cameraman is shooting using the imaging device 10, for example, a return video signal is provided from the metadata synthesis device 60 to the imaging device 10. As a result, as shown in FIG. 14, the display unit 108 of the imaging apparatus 10 can display, for example, a video currently being photographed by the cameraman (in the example of FIG. 14, the subject is a lion). Furthermore, together with the video being photographed, for example, the scene information metadata extracted by the metadata synthesizing unit 60 can be displayed in a superimposed manner. Specifically, in the example of FIG. 14, for example, in the upper right corner of the display screen, the time code “12: 34: 10: 14”, the scene number “0017” of the video material being shot, and the video material being shot are displayed. The take number “0005” is superimposed. The time code, scene number, take number, and the like are, for example, information conventionally displayed on an electronic clapperboard or the like, and can be extracted from various metadata by setting the extraction conditions of the metadata synthesizer 60. Metadata.
[0165]
By displaying such a return video, the cameraman can view the scene information data conventionally displayed on the electronic clapperboard or the like together with the video being shot. Therefore, a cameraman or the like can easily identify and confirm what scene, what take, for what time, etc., the video being shot based on the scene information data displayed in a superimposed manner. it can.
[0166]
<6.2 Metadata display method after shooting / recording>
Next, a description will be given of a process in which a video signal recorded on the video tape 52 or the like is played back by the VTR 50 and displayed on the display unit 108 of the imaging apparatus 10 after shooting / recording of a scene is completed.
[0167]
As described above, the video signal generated by the imaging device 10 and added with metadata in units of frames by the metadata synthesizing device 50 or the like is recorded on a storage medium such as the video tape 52 by the VTR 50. When the recorded video material is displayed after such video material is shot and recorded, first, the video signal recorded on the video tape 52 or the like is reproduced by the VTR 50. The video signals reproduced in this way are sequentially input to the metadata composition device 60 in units of frames, for example. Then, the metadata synthesizer 60 generates a metadata synthesized video signal in the same manner as described above, and outputs the metadata synthesized video signal to the imaging device 10 as, for example, a return video signal.
[0168]
In this way, the video signal reproduced by the VTR 50 after shooting and recording of a certain scene is provided as a return video signal from the metadata synthesizing device 60 to the imaging device 10, for example. As a result, as shown in FIG. 15, for example, the metadata extracted by the metadata composition unit 60 can be displayed in a superimposed manner together with the reproduced video, for example. Specifically, in the example of FIG. 15, for example, the same scene information metadata as in the case of FIG. 14 is superimposed on the upper left corner of the display screen. Further, in the upper right corner of the display screen, the shutter speed setting “1/125” of the imaging device 10 at the time of shooting the video material, the zoom setting “10 times” of the lens device 12, and the moving speed “1. Metadata such as “2 m / s” is displayed in a superimposed manner. In this manner, not only the scene information metadata but also the necessary metadata in the camera setting metadata, lens setting metadata, and dolly setting metadata are set by setting the extraction condition of the metadata synthesizing device 60. It can be displayed together with the return video.
[0169]
By displaying such a return video, the photographer or the like can check the quality of the video shot by himself / herself by browsing the video reproduced and displayed on the display unit 108 and referring to the metadata after shooting / recording.
[0170]
At this time, as shown in the example of FIG. 15, the scene information metadata is superimposed on the return video, so that the photographer can play the reproduced video in any scene, at any time, at any time. It is possible to easily identify and confirm whether it is a product. Furthermore, if the desired metadata extracted from the camera setting metadata, scene setting metadata, dolly setting metadata, etc. is displayed in a superimposed manner in the return video, the cameraman can display the image quality of the video and the subject. It is possible to objectively and accurately grasp the movement, shooting conditions, and settings of the imaging device 10 and the like. For this reason, when re-shooting the same scene, the cameraman preferably resets the imaging device 10 or the like with reference to the metadata of the return video, or adjusts the shooting angle or zoom appropriately. it can.
[0171]
(Second Embodiment)
Next, a metadata display system 1 and the like according to the second embodiment of the present invention will be described. Compared with the metadata display system 1 according to the first embodiment, the metadata display system 1 according to the second embodiment does not add the camera setting metadata to the video signal, and the metadata is displayed. The only difference is that the additional device 40 collectively adds all the metadata to the video signal, and the other functional configurations are substantially the same as those in the first embodiment, and the description thereof will be omitted.
[0172]
Below, based on FIG. 16, the whole structure of the metadata display system 1 concerning this embodiment and the characteristic processing function of each apparatus are demonstrated. FIG. 16 is a block diagram showing a schematic configuration of the metadata display system 1 according to the present embodiment.
[0173]
As shown in FIG. 16, in the metadata display system 1 according to the present embodiment, for example, the CCU 20 acquires camera setting metadata acquired from the imaging device 10 separately from the video signal via an RS-232C cable or the like. It is configured to be able to output to the metadata adding device 40 via a route. This configuration is effective when the imaging device 10 is different from the first embodiment in that the metadata cannot be added to the video signal in units of frames.
[0174]
More specifically, the imaging apparatus 10 according to the first embodiment includes the metadata adding unit 112 as illustrated in FIG. 7, but the imaging apparatus 10 according to the present embodiment includes, for example, The metadata adding unit 112 is not provided. For this reason, for example, the imaging apparatus 10 according to the present embodiment transmits the video signal from the transmission / reception unit 114 to the CCU 20 without adding camera setting metadata to the video signal generated by the signal processing unit 106 in units of frames. Output to. However, information such as camera setting metadata of the imaging device 10 is communicated between the imaging device 10 and the CCU 20 via, for example, an optical cable. For this reason, the CCU 20 can acquire the camera setting metadata of the imaging device 10 as necessary. Therefore, the CPU 200 of the CCU 20 can sequentially output the acquired camera setting metadata to the metadata adding device 40 separately from the video via the RC-232C cable, for example. The camera setting metadata output from the CCU 20 in this way is stored in the metadata buffer memory 403 of the metadata adding device 40 together with metadata of other metadata groups.
[0175]
For example, the metadata adding device 40 groups camera setting metadata acquired from the CCU 20 in addition to scene information metadata, lens setting metadata, and dolly setting metadata in the same manner as in the first embodiment. . Further, the metadata adding device 40, for example, packs these four metadata groups and performs KLV encoding, and sequentially inserts them into the blanking area of the video signal input from the CCU 20 in units of one frame. I will do it. In this way, the metadata adding device 40 can simultaneously add all metadata in units of one frame to a video signal, for example. Subsequent processing of the VTR 50 and the like is substantially the same as in the case of the first embodiment, and a description thereof will be omitted.
[0176]
As described above, in the metadata display system according to the second embodiment, for example, scene information metadata, lens setting metadata, dolly setting metadata, and camera setting metadata generated in the system are all stored in the metadata. They can be assembled into the addition device 40 and added together to the video signal. For this reason, for example, even when the imaging apparatus 10 that does not have a function of adding metadata is used, all metadata can be added to the video signal in units of frames.
[0177]
Even when metadata is added to a video signal as described above, the metadata composition device 60 according to the present embodiment is similar to the metadata composition device 60 according to the first embodiment. A data synthesized video signal can be generated and provided to the imaging device 10 as a return video signal.
[0178]
As described above, according to the metadata display system 1 and the video recording method and the metadata display method using the metadata display system 1 according to the first and second embodiments, the video signal generated by the imaging device 10 is generated. On the other hand, metadata related to the video signal can be added in units of frames and recorded in the same storage medium in real time during the photographing process. For this reason, it is not necessary to indirectly link metadata recorded in a terminal device such as a PC and video material recorded in a storage medium with a time code or the like as in the past. The metadata related to the video material can be directly linked and recorded. Therefore, it is convenient because the video material and the metadata can be managed integrally. In addition, since it is not necessary to maintain consistency between video material and metadata when extracting metadata, the necessary metadata can be efficiently extracted and used or rewritten.
[0179]
For example, even if the video material is edited and cut out at the In point and the out point in the editing stage of the video material, the corresponding portion of metadata is also cut out accompanying the video material. For this reason, the metadata can be extracted and displayed continuously in synchronization with the video material even if the consistency between the video material and the metadata is not taken. Further, even when the video material is post-processed thereafter, the metadata can be quickly and easily extracted from the edited video material and used for post-processing. For example, by simultaneously monitoring the video material and the metadata corresponding to the video material, the operator can accurately grasp the image quality, movement, etc. of the video material.
[0180]
Further, even when the frame rate of the recorded video signal is changed by the imaging device 10 performing variable speed shooting processing, metadata is added to the video signal in units of frames. Therefore, there is no deviation between the number of video signal frames per unit time and the number of recorded metadata. Therefore, it is possible to suitably link the video material and the metadata in correspondence with the video material shot at such a variable speed. Therefore, it is possible to suitably extract metadata from a video signal taken at a variable speed and monitor it together with the video.
[0181]
Further, as described above, metadata is grouped according to the purpose of use and added to the video signal, whereby the metadata can be extracted and rewritten in units of groups. For this reason, the amount of metadata to be extracted and rewritten can be reduced, so that the processing efficiency can be improved. In addition, monitoring and replacement of only a specific metadata group can be realized easily and quickly.
[0182]
For example, at the shooting / recording stage, only the scene information metadata is extracted and multiplexed and displayed on the video, so that the cameraman, the director, and the like browse the display unit 107 and the display device 70 of the imaging device 10 and shoot. Index information (time code, scene number, take number, etc.) of the video material being recorded or reproduced after recording can be accurately grasped.
[0183]
Further, when post-processing such as CG composition processing is performed in the editing stage, for example, only the camera setting metadata is extracted and displayed on the display device of the editing device, so that the image quality (brightness, hue, On the other hand, for example, by extracting the lens and dolly setting metadata and displaying them on the display device of the editing device, it is possible to accurately grasp the movement of the imaging device 10 and the subject during photographing. In addition, by adding comments from the photographer or director at the time of recording the video material (such as taking notes of the shooting situation) to the scene information metadata, the willingness regarding the video material between the shooting department and the editing department, etc. Can communicate.
[0184]
In addition, a video signal (metadata combined video signal) obtained by combining various metadata as described above can be returned to the imaging apparatus 10 as a return video. Therefore, it is possible to display an image on which various types of metadata are superimposed on the display unit 108 or the like of the imaging device 10. In other words, the display unit 108 of the imaging apparatus 10 can display the captured video material and metadata related to the video material in synchronization in units of frames.
[0185]
More specifically, scene information metadata (time code, scene number, take number, etc.) previously displayed on an electronic clapperboard or the like is synthesized with the video being shot and recorded, and the imaging device is used as a return video. 10 can be displayed. As a result, the cameraman can browse the scene information metadata and the like together with the video being shot. For this reason, the cameraman can perform shooting suitably while accurately grasping index information of the video being shot.
[0186]
In addition, the video that is played back by the VTR 50 after recording is combined with the necessary metadata among the scene information metadata, camera setting metadata, lens setting metadata, dolly setting metadata, and the like as a return video. Thus, the cameraman can view index information such as the scene number and take number of the recorded video while browsing the video reproduced and displayed on the display unit 108 after shooting and recording. It is possible to check and set objective information of various devices at the time of shooting the video, and the metadata regarding the image quality of the video and the movement of the subject objectively and accurately. For this reason, it is possible to check the quality and correction points of the captured video while referring to the various metadata. For this reason, when re-shooting the same scene, the cameraman preferably resets the imaging device 10 or the like with reference to the metadata of the return video, or adjusts the shooting angle or zoom appropriately. it can.
[0187]
As mentioned above, although preferred embodiment of this invention was described referring an accompanying drawing, this invention is not limited to this example. It will be obvious to those skilled in the art that various changes or modifications can be conceived within the scope of the technical idea described in the claims, and these are naturally within the technical scope of the present invention. It is understood that it belongs.
[0188]
For example, in the metadata display system 1 according to the first and second embodiments, the metadata adding device 40 is configured as separate hardware from the imaging device 10, the CCU 20, the VTR 50, and the like. The present invention is not limited to such an example. For example, the metadata adding device 40 may be changed so as to be integrated with any one or more of the imaging device 10, the CCU 20, the VTR 50, and the like.
[0189]
In the above embodiment, the metadata composition device 60 is configured as hardware separate from the VTR 50 and the like, but the present invention is not limited to such an example. For example, the metadata composition device 60 may be changed so as to be integrated with the VTR 50 or the like.
[0190]
A modification of the metadata adding device 40 and the metadata synthesizing device 60 will be described below with a specific example.
[0191]
First, the configuration of the metadata display system 1 according to the first modification will be described with reference to FIG. FIG. 17 is a block diagram illustrating a schematic configuration of the metadata display system 1 according to the first modification.
[0192]
As shown in FIG. 17, in the metadata display system 1 according to the first modification, the metadata adding device 40 is built in a video signal recording / reproducing device such as a VTR 50. Therefore, scene information metadata, lens setting metadata, dolly setting metadata, and the like are input from the lens device 12, the dolly device 14, the metadata input terminal device 30 and the like to the VTR 50 via an RS-232C cable or the like. ing. The processing function of the metadata adding device 40 built in the VTR 50 is, for example, substantially the same as the processing function of the metadata adding device 40 according to the first embodiment, and various input metadata are grouped. It can be KLV encoded and added to the video signal in frame units.
[0193]
In the metadata display system 1 according to the first modification, the metadata synthesis device 60 is also built in a video signal recording / reproducing device such as the VTR 50. The processing function of the metadata synthesizing device 60 built in the VTR 50 is substantially the same as the processing function of the metadata synthesizing device 60 according to the first embodiment, for example. Can be extracted, decoded, and synthesized into a video signal. As described above, the metadata adding device 40, the metadata synthesizing device 60, and the VTR 50 may be integrated.
[0194]
Next, the configuration of the metadata display system 1 according to the second modification will be described with reference to FIG. FIG. 18 is a block diagram illustrating a schematic configuration of the metadata display system 1 according to the second modification.
[0195]
As shown in FIG. 18, in the metadata display system 1 according to the second modification, the metadata adding device 40 is built in the imaging device 10. Therefore, scene information metadata, lens setting metadata, dolly setting metadata, and the like are transferred from the lens device 12, the dolly device 14, the metadata input terminal device 30 and the like to the imaging device 10 via an RS-232C cable or the like. Have been entered. The processing function of the metadata adding device 40 built in the imaging device 10 includes, for example, the processing function of the metadata adding unit 112 according to the first embodiment (a function of adding camera setting metadata to a video signal), This is a processing function combined with the processing function of the metadata adding device 40 (function of adding scene information setting metadata, lens setting metadata, and lens setting metadata to the video signal). Specifically, the metadata adding device 40 built in the imaging device 10 can group, for example, group the above four metadata and perform KLV encoding, and add them to the video signal in units of frames. In this way, the metadata adding device 40 and the imaging device 10 are integrated, and in the imaging device 10, all of the camera setting metadata, scene information setting metadata, lens setting metadata, and dolly setting metadata are You may make it add to a video signal.
[0196]
Further, in the metadata display system 1 according to the modification example 2, the metadata synthesis apparatus 60 is built in a video signal recording / reproducing apparatus such as the VTR 50 as in the modification example 1. As described above, the metadata synthesis device 60 and the VTR 50 may be integrated.
[0197]
As described above, the number of devices in the system is reduced by incorporating the metadata adding device 40 in the VTR 50 or the like or the imaging device 10 or by incorporating the metadata synthesizing device 50 in the VTR 50 or the like. It is possible to save labor and the like.
[0198]
Further, in the metadata display system 1 according to the modification example 1 and the modification example 2, for example, the CCU 20 is not provided, and a video signal is transmitted from the imaging device 10 to the VTR 50 via an HD SDI cable or the like. Yes. In this way, the imaging device 10 may be configured to have the processing function of the CCU 20 as a matter of course.
[0199]
Further, although not shown, the imaging device 10 may be configured as, for example, an imaging device (camcorder or the like) having a function of recording a video signal in a storage medium. Thereby, the imaging device 10 can also be configured to have all the functions of the CCU 20, the metadata adding device 40, the VTR 50, and the metadata synthesizing device 60, for example.
[0200]
In the above embodiment, the lens setting metadata generated by the lens device 12 is output via an RS-232C cable or the like and added to the video signal by the metadata adding device 40. However, the present invention is such an example. It is not limited. For example, a lens device 12 capable of communicating lens setting information and the like with the imaging device 10 main body is employed, and lens setting metadata and the like are directly input from the lens device 12 to the imaging device 10 main body. It may be configured. Accordingly, the metadata adding unit 112 of the imaging device 10 can be configured to add not only camera setting metadata but also lens setting metadata acquired from the lens device 12 to the video signal, for example.
[0201]
In the above embodiment, RS-232C is used as an interface for communicating various metadata between devices. However, the present invention is not limited to this example. For example, USB (Universal Serial Bus), SCSI (Small) are used. Various interfaces such as Computer System Interface (Serial Computer Interface), serial SCSI, and GP-IB (General Purpose Interface Bus) may be used. In addition, the above devices are not limited to wired communication, and for example, metadata and / or video signals may be transmitted by wireless communication.
[0202]
In the above embodiment, various metadata generated in the metadata display system are grouped into four metadata groups including a scene information group, a camera setting group, a lens setting group, and a dolly setting group. However, the present invention is not limited to such an example. For example, the above four metadata groups may be arbitrarily combined according to the purpose of use, such as combining the lens setting group and the dolly setting group into one lens and dolly setting group. Further, it is not necessary to provide all the upper four metadata groups. For example, one or more metadata groups may be provided.
[0203]
A new metadata group other than the above may be provided. Specifically, for example, a voice information group is provided, and within this voice information group, recording method information (stereo, monaural, surround, etc.), recording content information (microphone 1 records the background sound, microphone 2 records the actor's Voice related information metadata such as voice recording may be grouped.
[0204]
For example, when a plurality of imaging devices 10 are provided in the metadata display system 1, the return video may be provided to the plurality of imaging devices 10. Thereby, the cameraman of each imaging device 10 can start each imaging process with a cue that the return video has been displayed on the display unit 108 of the imaging device 10.
[0205]
【The invention's effect】
  As explained above, according to the present invention,Depending on the purpose of use of metadata, necessary metadata can be suitably extracted and displayed in groups.
[Brief description of the drawings]
FIG. 1 is a block diagram illustrating a schematic configuration of a metadata display system according to a first embodiment;
FIG. 2 is an explanatory diagram illustrating a specific example of scene information metadata included in a scene information group according to the first embodiment;
FIG. 3 is an explanatory diagram illustrating a specific example of camera setting metadata included in a camera setting group according to the first embodiment;
FIG. 4 is an explanatory diagram illustrating a specific example of lens setting metadata included in a lens setting group according to the first embodiment;
FIG. 5 is an explanatory diagram illustrating a specific example of camera setting metadata included in the dolly setting group according to the first embodiment;
FIG. 6 is an explanatory diagram for explaining a metadata format according to the first embodiment;
FIG. 7 is a block diagram illustrating a configuration of the imaging apparatus according to the first embodiment.
FIG. 8 is an explanatory diagram for explaining a mode of adding metadata to a video signal according to the first embodiment;
FIG. 9 is a block diagram illustrating a configuration of a camera control unit according to the first embodiment.
FIG. 10 is a block diagram illustrating a configuration of a metadata adding apparatus according to the first embodiment;
FIG. 11 is a block diagram showing a configuration of a video tape recorder according to the first embodiment.
FIG. 12 is a block diagram illustrating a configuration of a metadata composition device according to the first embodiment;
FIG. 13 is a timing chart for explaining a video recording method according to the first embodiment;
FIG. 14 is an explanatory diagram illustrating a specific example of a return video during shooting and recording displayed on the display unit of the imaging apparatus according to the first embodiment;
FIG. 15 is an explanatory diagram illustrating a specific example of a return video after recording displayed on the display unit of the imaging apparatus according to the first embodiment;
FIG. 16 is a block diagram illustrating a schematic configuration of a metadata display system according to a second embodiment;
FIG. 17 is a block diagram illustrating a schematic configuration of a metadata display system according to a first modification.
FIG. 18 is a block diagram illustrating a schematic configuration of a metadata display system according to a second modification.
[Explanation of symbols]
1: Metadata display system
10: Imaging device
12: Lens device
14: Dolly device
18: Sound collector
20: Camera control unit
30: Terminal device for metadata input
40: Metadata adding device
50: Video tape recorder
52: Video tape
60: Metadata synthesizer
70: Display device
104: Imaging unit
108: Display unit
110: Camera setting metadata generation unit
112: Metadata addition unit
126: Lens setting metadata generation unit
144: Dolly setting metadata generation unit
403: Metadata buffer memory
406: Metadata packing unit
408: Metadata encoder
410: Metadata insertion part
506: Signal processor
508: Metadata decoder
514: Metadata encoder
606: Metadata extraction unit
608: Metadata decoder
610: Metadata video generation unit
612: Metadata video composition unit

Claims (13)

  1. A metadata display system that displays metadata associated with a video signal generated by an imaging device:
    As the metadata, metadata including at least scene information acquired in advance from the metadata input terminal device before the start of photographing is stored, and sequentially transmitted from the imaging device, the lens device, and / or the dolly device after the photographing is started. A buffer memory to store incoming metadata;
    A metadata packing unit that extracts necessary metadata from the metadata stored in the buffer memory for each purpose of use of the metadata, and groups the metadata into two or more metadata groups;
    With
    Extracting at least a portion of the grouped metadata and combining it with the video signal;
    The video signal combined with the metadata is displayed on a display device ,
    The metadata group is
    A scene information group including information related to a scene shot by the imaging device, grouped for the purpose of identifying video material;
    A camera setting group including the setting information of the imaging device grouped for the purpose of grasping the image quality of the video material;
    A lens setting group or a dolly setting group including setting information of a lens device or a dolly device included in the imaging device, grouped for the purpose of grasping a movement or distance of a subject in the video material;
    A metadata display system characterized by including:
  2.   The metadata display system according to claim 1, wherein the metadata packing unit packs the metadata groups by performing KLV (Key Length Value) encoding.
  3.   The metadata display system according to claim 1, wherein the display device is provided in the imaging device.
  4.   The grouped metadata is added to the video signal in units of frames, and at least a part of the grouped metadata is extracted from the video signal and combined with the video signal. The metadata display system according to claim 1.
  5. The metadata display system includes:
    A video signal recording / reproducing device for recording / reproducing the video signal to which the grouped metadata is added to a storage medium;
    5. The method according to claim 4 , wherein at least a part of the grouped metadata is extracted from the video signal reproduced by the video signal recording / reproducing apparatus and synthesized with the reproduced video signal. The described metadata display system.
  6. As metadata related to the video signal generated by the imaging device, metadata including at least scene information acquired in advance from the metadata input terminal device before the start of shooting is stored, and the imaging device, lens device, and / or A buffer memory for storing metadata sequentially transmitted from the dolly device after the start of photographing;
    A metadata packing unit that extracts necessary metadata from the metadata stored in the buffer memory for each purpose of use of the metadata, and groups the metadata into two or more metadata groups;
    A recording / reproducing unit for recording / reproducing the video signal to which the grouped metadata is added in units of frames;
    A metadata synthesizer that extracts at least a part of the grouped metadata from the video signal to which the grouped metadata is added in units of frames and synthesizes the grouped metadata with the video signal;
    Bei to give a,
    The metadata group is
    A scene information group including information related to a scene shot by the imaging device, grouped for the purpose of identifying video material;
    A camera setting group including the setting information of the imaging device grouped for the purpose of grasping the image quality of the video material;
    A lens setting group or a dolly setting group including setting information of a lens device or a dolly device included in the imaging device, grouped for the purpose of grasping a movement or distance of a subject in the video material;
    A video signal recording / reproducing apparatus comprising:
  7. The metadata synthesizer is:
    The at least part of the grouped metadata is extracted from the video signal reproduced from the storage medium by the recording / reproducing unit, and is combined with the reproduced video signal. 6. The video signal recording / reproducing apparatus according to 6.
  8. An imaging apparatus comprising: an imaging unit that images a subject to generate a video signal; and a display unit that displays the video signal;
    As metadata related to the video signal generated by the imaging device, metadata including at least scene information acquired in advance from the metadata input terminal device before the start of shooting is stored, and the imaging device, lens device, and / or Or a buffer memory for storing metadata sequentially transmitted from the dolly device after the start of photographing;
    A metadata packing unit that extracts necessary metadata from the metadata stored in the buffer memory for each purpose of use of the metadata, and groups the metadata into two or more metadata groups;
    With
    Connected to a metadata synthesizer for extracting at least a part of the grouped metadata from the video signal to which the grouped metadata is added in units of frames and combining the extracted metadata with the video signal; And
    From the metadata synthesis device, the video signal synthesized with the metadata is received and displayed on the display unit ,
    The metadata group is
    A scene information group including information related to a scene shot by the imaging device, grouped for the purpose of identifying video material;
    A camera setting group including the setting information of the imaging device grouped for the purpose of grasping the image quality of the video material;
    A lens setting group or a dolly setting group including setting information of a lens device or a dolly device included in the imaging device, grouped for the purpose of grasping a movement or distance of a subject in the video material;
    An imaging device comprising:
  9. A metadata display method for displaying metadata related to a video signal generated by an imaging device, wherein:
    As the metadata, metadata including at least scene information acquired in advance from the metadata input terminal device before the start of photographing is stored, and sequentially transmitted from the imaging device, the lens device, and / or the dolly device after the photographing is started. Store incoming metadata in buffer memory,
    For each purpose of use of the metadata, the necessary metadata is extracted from the metadata stored in the buffer memory and grouped into two or more metadata groups,
    Extracting at least a portion of the grouped metadata and combining it with the video signal;
    The video signal combined with the metadata is displayed on a display device ,
    The metadata group is
    A scene information group including information related to a scene shot by the imaging device, grouped for the purpose of identifying video material;
    A camera setting group including the setting information of the imaging device grouped for the purpose of grasping the image quality of the video material;
    A lens setting group or a dolly setting group including setting information of a lens device or a dolly device included in the imaging device, grouped for the purpose of grasping a movement or distance of a subject in the video material;
    A metadata display method characterized by including :
  10. 10. The metadata display method according to claim 9 , wherein each metadata group is packed by KLV (Key Length Value) encoding.
  11. The metadata display method according to claim 9 , wherein the display device is provided in the imaging device.
  12. The grouped metadata is added to the video signal in units of frames, and at least a part of the grouped metadata is extracted from the video signal and combined with the video signal. The metadata display method according to claim 9 .
  13. The video signal is reproduced from a storage medium on which the video signal to which the grouped metadata is added is recorded, and at least a part of the grouped metadata is extracted from the reproduced video signal. The metadata display method according to claim 12 , wherein the metadata is combined with the reproduced video signal.
JP2003101837A 2003-04-04 2003-04-04 Metadata display system, video signal recording / reproducing device, imaging device, metadata display method Expired - Fee Related JP4332365B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2003101837A JP4332365B2 (en) 2003-04-04 2003-04-04 Metadata display system, video signal recording / reproducing device, imaging device, metadata display method

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2003101837A JP4332365B2 (en) 2003-04-04 2003-04-04 Metadata display system, video signal recording / reproducing device, imaging device, metadata display method
US10/809,476 US7623176B2 (en) 2003-04-04 2004-03-26 Meta-data display system, meta-data synthesis apparatus, video-signal recording/reproduction apparatus, imaging apparatus and meta-data display method

Publications (2)

Publication Number Publication Date
JP2004312278A JP2004312278A (en) 2004-11-04
JP4332365B2 true JP4332365B2 (en) 2009-09-16

Family

ID=33465495

Family Applications (1)

Application Number Title Priority Date Filing Date
JP2003101837A Expired - Fee Related JP4332365B2 (en) 2003-04-04 2003-04-04 Metadata display system, video signal recording / reproducing device, imaging device, metadata display method

Country Status (2)

Country Link
US (1) US7623176B2 (en)
JP (1) JP4332365B2 (en)

Families Citing this family (41)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4550809B2 (en) * 2004-03-29 2010-09-22 パナソニック株式会社 Accumulated display device, interlocking display method and system
EP1799095A2 (en) * 2004-09-30 2007-06-27 Boston Scientific Scimed, Inc. Adapter for use with digital imaging medical device
WO2006101035A1 (en) * 2005-03-18 2006-09-28 Nikon Corporation Digital camera
KR101133125B1 (en) * 2005-06-23 2012-04-06 삼성테크윈 주식회사 A system and method to display filming time
JP4661447B2 (en) * 2005-08-16 2011-03-30 ソニー株式会社 Transmission / reception system and method, transmission device and method, reception device and method, and program
JP4765475B2 (en) * 2005-08-17 2011-09-07 ソニー株式会社 Information signal processing apparatus and processing method
KR100669251B1 (en) * 2005-11-25 2007-01-16 한국전자통신연구원 Apparatus and method for automatically analyzing digital image quality
CA2631803A1 (en) * 2005-12-02 2007-06-07 Thomson Licensing Work flow metadata system and method
US8135684B2 (en) * 2006-04-13 2012-03-13 Eastman Kodak Company Value index from incomplete data
JP4770581B2 (en) 2006-05-17 2011-09-14 ソニー株式会社 Moving image data processing device, stream generation device, imaging device, and moving image data processing method
JP2008011349A (en) * 2006-06-30 2008-01-17 Nikon Corp Camera capable of photographing moving picture
US20080111822A1 (en) * 2006-09-22 2008-05-15 Yahoo, Inc.! Method and system for presenting video
US7912366B2 (en) * 2006-12-11 2011-03-22 Kyocera Corporation System and method for providing photograph location information in a wireless communication device
KR101449430B1 (en) * 2007-08-31 2014-10-14 삼성전자주식회사 Method and apparatus for generating movie-in-short of contents
JP4946789B2 (en) 2007-10-24 2012-06-06 株式会社Jvcケンウッド Digest data creation device and digest data creation method
US8077222B2 (en) * 2007-11-06 2011-12-13 Canon Kabushiki Kaisha Image processing apparatus to transmit moving image data
WO2009114723A1 (en) * 2008-03-12 2009-09-17 Sony Corporation Smart slate
US8819014B2 (en) 2008-08-15 2014-08-26 Apple Inc. Digital slate
JP5159535B2 (en) * 2008-09-19 2013-03-06 株式会社東芝 Subtitle sending apparatus and subtitle editing method
JP2010118739A (en) * 2008-11-11 2010-05-27 Sony Corp Switcher control device, switcher control method and image synthesizing apparatus
KR101545902B1 (en) * 2008-12-12 2015-08-20 삼성전자주식회사 Photographing apparatus recording apparatus reproducing apparatus photographing method recording method reproducing method and storing medium having program to perform the recording method
US8520999B2 (en) * 2008-12-23 2013-08-27 Sony Corporation Camera event logger
JP5287465B2 (en) * 2009-04-21 2013-09-11 ソニー株式会社 Imaging apparatus, shooting setting method and program thereof
US10531062B2 (en) * 2009-10-13 2020-01-07 Vincent Pace Stereographic cinematography metadata recording
US8751693B2 (en) * 2009-12-25 2014-06-10 Samsung Electronics Co., Ltd. Apparatus for and method of processing data
JP5577415B2 (en) * 2010-02-22 2014-08-20 ドルビー ラボラトリーズ ライセンシング コーポレイション Video display with rendering control using metadata embedded in the bitstream
EP2540072B1 (en) 2010-02-22 2014-04-09 Dolby Laboratories Licensing Corporation Video delivery and control by overwriting video data
US20120002061A1 (en) * 2010-07-01 2012-01-05 Gay Michael F Systems and methods to overlay remote and local video feeds
KR20120085474A (en) * 2011-01-24 2012-08-01 삼성전자주식회사 A photographing apparatus, a method for controlling the same, and a computer-readable storage medium
JP2015186235A (en) * 2014-03-26 2015-10-22 ソニー株式会社 Image sensor and electronic apparatus
ES2730404T3 (en) 2014-04-04 2019-11-11 Red Com Llc Camcorder with capture modes
US9959292B2 (en) * 2014-09-30 2018-05-01 Ricoh Company, Ltd. Application program, smart device, information processing apparatus, information processing system, and information processing method
US10062413B2 (en) * 2015-10-16 2018-08-28 Tribune Broadcasting Company, Llc Media-production system with social media content interface feature
US9681111B1 (en) 2015-10-22 2017-06-13 Gopro, Inc. Apparatus and methods for embedding metadata into video stream
US9667859B1 (en) 2015-12-28 2017-05-30 Gopro, Inc. Systems and methods for determining preferences for capture settings of an image capturing device
US9922387B1 (en) 2016-01-19 2018-03-20 Gopro, Inc. Storage of metadata and images
US9967457B1 (en) * 2016-01-22 2018-05-08 Gopro, Inc. Systems and methods for determining preferences for capture settings of an image capturing device
US9665098B1 (en) 2016-02-16 2017-05-30 Gopro, Inc. Systems and methods for determining preferences for flight control settings of an unmanned aerial vehicle
US9973792B1 (en) 2016-10-27 2018-05-15 Gopro, Inc. Systems and methods for presenting visual information during presentation of a video segment
US10187607B1 (en) 2017-04-04 2019-01-22 Gopro, Inc. Systems and methods for using a variable capture frame rate for video capture
JP2019008636A (en) * 2017-06-27 2019-01-17 オリンパス株式会社 Imaging System

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0982937B1 (en) * 1990-10-15 2006-11-29 Canon Kabushiki Kaisha Image pickup apparatus provided with interchangeable lenses
JP3166346B2 (en) 1992-10-08 2001-05-14 ソニー株式会社 Video additional information transmission method
JPH06153028A (en) * 1992-11-12 1994-05-31 Sony Corp Video camera device
JPH0946627A (en) 1995-07-31 1997-02-14 Sony Corp Camcorder
US6532039B2 (en) * 1997-09-17 2003-03-11 Flashpoint Technology, Inc. Method and system for digital image stamping
JP4375684B2 (en) 1997-12-24 2009-12-02 カシオ計算機株式会社 Electronic camera device and imaging information display method
JP3914643B2 (en) 1998-09-10 2007-05-16 日本放送協会 Program information management editing system and hierarchical program information storage management apparatus used therefor
JP4273636B2 (en) * 2000-06-30 2009-06-03 ソニー株式会社 Information recording apparatus and method, and information recording system
US7054508B2 (en) * 2000-08-03 2006-05-30 Canon Kabushiki Kaisha Data editing apparatus and method
US7051048B2 (en) * 2000-09-29 2006-05-23 Canon Kabushiki Kaisha Data management system, data management method, and program
JP4205574B2 (en) 2001-05-31 2009-01-07 キヤノン株式会社 Image processing apparatus and control method thereof
JP3806698B2 (en) * 2003-03-07 2006-08-09 三洋電機株式会社 Electronic camera
US20050104976A1 (en) * 2003-11-17 2005-05-19 Kevin Currans System and method for applying inference information to digital camera metadata to identify digital picture content

Also Published As

Publication number Publication date
US7623176B2 (en) 2009-11-24
US20040249861A1 (en) 2004-12-09
JP2004312278A (en) 2004-11-04

Similar Documents

Publication Publication Date Title
US9451229B2 (en) Video recording and reproducing method, and video reproducing apparatus and method
KR100745304B1 (en) Image receiving apparatus and image reproducing apparatus
JP5293463B2 (en) Image processing apparatus, image processing method, and program
EP2153641B1 (en) Digital cinema camera system for recording, editing and visualizing images
JP4792686B2 (en) Image processing apparatus, image processing method, and recording medium
AU716516B2 (en) Motion picture recording device using digital, computer-readable non-linear media
KR100716142B1 (en) Method for transferring stereoscopic image data
CN103297682B (en) Moving image capture apparatus and the method for using camera system
JP4168837B2 (en) Information generating apparatus, recording apparatus, reproducing apparatus, recording / reproducing system, method thereof, and program
JP2013500544A (en) Improved audio / video method and system
US8208792B2 (en) Content shooting apparatus for generating scene representation metadata
JP4358081B2 (en) Video recording device
US7675555B2 (en) Recording method, recording apparatus, recording medium, reproduction method, reproduction apparatus and image pickup apparatus
JP3525353B2 (en) Digital electronic still camera
EP2619761B1 (en) Enriching digital photographs
CN100334642C (en) Recording method and apparatus, recording medium, reproducing method and apparatus and imaging apparatus
US8508579B2 (en) System and method for generating and reproducing 3D stereoscopic image file including 2D image
JP2007300556A (en) Moving image processor, and moving image processing method
US9661047B2 (en) Method and system for central utilization of remotely generated large media data streams despite network bandwidth limitations
JP4708733B2 (en) Imaging device
US20060050321A1 (en) Record/replay apparatus and method
KR101007508B1 (en) Video signal processor, video signal recorder, video signal reproducer, video signal processor processing method, video signal recorder processing method, video signal reproducer processing method, recording medium
US7319480B2 (en) Method and apparatus for compressing motion image files to provide an improved image navigation display
US8199211B2 (en) Camera direct dailies
EP1420406A2 (en) Content editing assistance system, video processing apparatus, playback apparatus, editing apparatus, computer program, and content processing method

Legal Events

Date Code Title Description
A621 Written request for application examination

Free format text: JAPANESE INTERMEDIATE CODE: A621

Effective date: 20041227

A131 Notification of reasons for refusal

Free format text: JAPANESE INTERMEDIATE CODE: A131

Effective date: 20060404

A521 Written amendment

Free format text: JAPANESE INTERMEDIATE CODE: A523

Effective date: 20060605

A02 Decision of refusal

Free format text: JAPANESE INTERMEDIATE CODE: A02

Effective date: 20060704

A521 Written amendment

Free format text: JAPANESE INTERMEDIATE CODE: A523

Effective date: 20060904

A911 Transfer of reconsideration by examiner before appeal (zenchi)

Free format text: JAPANESE INTERMEDIATE CODE: A911

Effective date: 20060915

A912 Removal of reconsideration by examiner before appeal (zenchi)

Free format text: JAPANESE INTERMEDIATE CODE: A912

Effective date: 20061201

A01 Written decision to grant a patent or to grant a registration (utility model)

Free format text: JAPANESE INTERMEDIATE CODE: A01

A61 First payment of annual fees (during grant procedure)

Free format text: JAPANESE INTERMEDIATE CODE: A61

Effective date: 20090622

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20120626

Year of fee payment: 3

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20130626

Year of fee payment: 4

LAPS Cancellation because of no payment of annual fees