WO2022145215A1 - Dispositif de sélection-synthèse de vidéo, procédé de commande associé, et dispositif de gestion - Google Patents
Dispositif de sélection-synthèse de vidéo, procédé de commande associé, et dispositif de gestion Download PDFInfo
- Publication number
- WO2022145215A1 WO2022145215A1 PCT/JP2021/045930 JP2021045930W WO2022145215A1 WO 2022145215 A1 WO2022145215 A1 WO 2022145215A1 JP 2021045930 W JP2021045930 W JP 2021045930W WO 2022145215 A1 WO2022145215 A1 WO 2022145215A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- video
- input
- format
- selection
- video stream
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 54
- 238000003786 synthesis reaction Methods 0.000 title abstract description 11
- 239000000872 buffer Substances 0.000 claims description 63
- 230000002194 synthesizing effect Effects 0.000 claims description 23
- 238000004519 manufacturing process Methods 0.000 abstract description 40
- 238000005516 engineering process Methods 0.000 abstract description 7
- 238000003860 storage Methods 0.000 description 47
- 230000006870 function Effects 0.000 description 25
- 238000009826 distribution Methods 0.000 description 21
- 238000010586 diagram Methods 0.000 description 20
- 230000005540 biological transmission Effects 0.000 description 19
- 230000015572 biosynthetic process Effects 0.000 description 8
- 238000004891 communication Methods 0.000 description 7
- 230000002093 peripheral effect Effects 0.000 description 6
- 230000003287 optical effect Effects 0.000 description 5
- 230000003139 buffering effect Effects 0.000 description 4
- 238000004590 computer program Methods 0.000 description 3
- 239000000470 constituent Substances 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 238000003491 array Methods 0.000 description 2
- 239000000835 fiber Substances 0.000 description 2
- 238000010295 mobile communication Methods 0.000 description 2
- 230000001902 propagating effect Effects 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 239000007787 solid Substances 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 229910052802 copper Inorganic materials 0.000 description 1
- 239000010949 copper Substances 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 210000003813 thumb Anatomy 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/268—Signal distribution or switching
Definitions
- the present disclosure relates to a video selection synthesizer and its control method, and a management device, and more particularly to a video selection synthesis device and its control method, and a management device capable of appropriately realizing the Production Switcher function.
- Non-Patent Document 4 defines timed metadata metrics for media for transfer in files based on the ISO-based media file format (ISO / IEC 14496-12, see Non-Patent Document 1).
- Non-Patent Document 2 defines a framework for network-based media processing (NBMP).
- ISO / IEC 14496-12 2020
- Information technology-Coding of audio-visual objects-Part 12 ISO base media file format ISO / IEC 23090-8: 2020
- Information technology-Coded representation of immersive media-Part 8 Network based media processing ISO / IEC 23001-10: 2020
- Information technology-MPEG systems technologies-Part 10 Carriage of timed metadata metrics of media in ISO base media file format 3GPP TS 26.238 Technical Specification Group Services and System Aspects; Uplink Streaming (Release 16)
- This disclosure is made in view of such a situation, and makes it possible to appropriately realize the Production Switcher function when each video source uses a plurality of stream formats together.
- the video selection synthesizer of the first aspect of the present technology makes settings based on the format characteristic information of the video stream output from the plurality of video sources, and is selected from the video of the video stream supplied from each of the plurality of video sources. , A selection unit for selecting one or a plurality of images, and a component unit constituting one image screen using the one or a plurality of images selected by the selection unit.
- the video selection / synthesizing device makes settings based on the format characteristic information of the video stream output by the plurality of video sources, and is supplied from each of the plurality of video sources.
- One or a plurality of videos are selected from the videos of the video stream, and one video screen is configured by using the selected video.
- settings are made based on the format characteristic information of the video stream output by the plurality of video sources, and one of the video of the video stream supplied from each of the plurality of video sources is used.
- One or a plurality of images are selected, and one image screen is configured by using the selected one or a plurality of images.
- the management device of the second aspect of the present technology generates format characteristic information of the video stream output by a plurality of video sources and notifies the video selection synthesizer.
- the format characteristic information of the video stream output by a plurality of video sources is generated and notified to the video selection synthesizer.
- each of the video selection and synthesizing device on the first side and the management device on the second side of the present technology can be realized by causing a computer to execute a program.
- the program to be executed by the computer can be provided by transmitting through a transmission medium or by recording on a recording medium.
- Each of the video selection synthesis device and the management device may be an independent device or an internal block constituting one device.
- Non-Patent Documents 1 to 4 For definitions of terms and the like that are not directly defined in the detailed description of the invention of the present specification, the contents described in Non-Patent Documents 1 to 4 above are cited by reference. For example, technical terms such as Parsing, Syntax, Semantics, File Structure described in Non-Patent Document 1, and Interfaces for network media processing standard described in Non-Patent Document 2. The terms used in are used in the same manner as those used in Non-Patent Documents 1 to 4.
- FIG. 1 is a block diagram showing a configuration example of a content production system according to an embodiment of the present disclosure.
- the content production system 1 in FIG. 1 is a system that distributes content such as sports and entertainment live performed at an event venue by utilizing a 5G network.
- the content production system 1 includes a plurality of video sources 21, a CCU (Camera Control Unit) 22 for controlling them, a video selection / synthesis device 23, a transmission system 24, and a management device 25.
- a CCU Computera Control Unit
- the plurality of video sources 21 are arranged at the relay site 11 which is the event venue.
- the video source 21 is, for example, a camera that captures video as content produced by the content production system 1.
- FIG. 1 for the sake of simplicity, three video sources 21-1 to 21-3 are shown, but the number of video sources 21 is arbitrary.
- the CCU 22 corresponding to each of the video sources 21 and the video selection synthesizer 23 are arranged on the cloud of the 5G network 12, which is the network of the 5th generation mobile communication system.
- one or more server devices arranged on the cloud of the 5G network 12 execute the function (processing) as the CCU 22 and the function (processing) as the video selection synthesis device 23. That is, on the cloud of the 5G network 12, the functions of the CCU 22 and the video selection synthesizer 23 are realized as application functions (AF) in the 5G network 12.
- the CCU 22 transmits a control signal for controlling shooting to the corresponding video source 21 via a predetermined base station 26 of the 5G network 12. Further, the CCU 22 acquires a video stream of the video captured by the video source 21 via a predetermined base station 26 and transmits the video stream to the video selection / synthesizing device 23.
- three CCUs 22-1 to 22-3 are arranged corresponding to the three video sources 21-1 to 21-3.
- the video sources 21-1 and 21-2 transmit the video stream obtained by shooting to the corresponding CCU 22-1 and 22-2 via the base station 26-1, and the video source 21-3 shoots.
- the video stream obtained by the above is transmitted to the corresponding CCU 22-3 via the base station 26-2.
- the video selection / synthesizing device 23 selects a desired one or a plurality of videos from the videos taken by each of the plurality of video sources 21 and appropriately lays them out, and further, a computer graphic image (hereinafter, appropriately, as appropriate) on the computer graphic image (hereinafter, appropriately). It is called a CG image), and a process of forming one video screen (video screen for distribution) is executed.
- the function) is the Production Switcher process (function), and hereinafter, it is appropriately referred to as a switcher function.
- the video selection synthesizer 23 has a switcher 31 and a storage 32.
- the switcher 31 appropriately selects images from video sources 21-1 to 21-3, or switches between those images and the images stored in the storage 32 to form one video screen. In addition to such image switching, the switcher 31 also performs CG image superimposition processing.
- the switcher 31 acquires the format characteristic information of the video stream transmitted from each video source 21 from the management device 25, decodes the video stream according to the acquired format characteristic information, and video sources 21-1 to 21. Controls such as switching the video of -3.
- the format characteristic information is information indicating the format and / or characteristics of the video stream, and includes, for example, an encoding method of the video stream, delay time information unique to the encoding method, transmission period information, and the like.
- the video source 21 can dynamically switch between a plurality of formats having different characteristics according to the network conditions, the importance of the video, and the like to generate and output a video stream.
- Video stream formats are used in SMPTE (Society of Motion Picture and Television Engineers) standards such as baseband, JPEG2000 (Joint Photographic Experts Group 2000), LLVC (Low Latency Video Codec), and conventional broadcasting systems.
- SMPTE Society of Motion Picture and Television Engineers
- JPEG2000 Joint Photographic Experts Group 2000
- LLVC Low Latency Video Codec
- the format of the encoding method used is mentioned.
- an encoding format such as HEVC (High Efficiency Vide o Coding) can be mentioned.
- the HEVC (High Efficiency Vide o Coding) encoding method is a mode (hevc-intra) that uses only intra coding (in-frame coding), and inter coding (inter-frame coding) with a short GOP frame length (for example, 3 frames).
- a mode using (encoding) hevc-inter
- an inter coding mode hevc-long
- a long GOP frame length for example, 15 frames
- the video source 21 may transmit only one video stream, but it is possible to simultaneously transmit a plurality of video streams having different formats depending on the purpose and application.
- the video source 21 simultaneously transmits two video streams, a proxy video and a high resolution video, to the video selection synthesizer 23.
- the proxy video is a highly compressed low-resolution video, which is mainly used for cut editing prior to distribution. However, proxy video may be used directly for the production of delivered content for content that requires high immediacy.
- the high-resolution video is a video used for distribution content, and is transmitted to the video selection / synthesizing device 23 after the proxy video. Distribution content is produced by editing the high-resolution video based on the edit point information created based on the proxy video.
- the storage 32 is a storage unit that temporarily stores the video supplied from the video sources 21-1 to 21-3.
- the video stored in the storage 32 is used for, for example, replay.
- the transmission system 24 is a system for distributing (transmitting) video, which is installed in a broadcasting company's station building or the like, and has a master switcher 33 as one of the system configurations.
- the master switcher 33 switches between video and audio (specifically, commercials, extraordinary news, etc.) that are not the main part, and outputs the produced content to the broadcasting network 27 or the network 28.
- the transmission system 24 can simultaneously distribute the produced content by broadcasting on the broadcasting network 27 and by network distribution via a network 28 such as the Internet.
- the management device 25 is a device that manages the transmission of a video stream from the entire content production system 1, for example, each video source 21 to the video selection / synthesis device 23, and the distribution of the production content by the transmission system 24. More specifically, the management device 25 monitors the network status of the 5G network 12 and determines the format characteristics of the video stream generated by each video source 21 according to the network status and the like. Then, the management device 25 notifies (instructs) each video source 21 of the determined format characteristic, and also notifies (transmits) the format characteristic information of the video stream output by each video source 21 to the video selection / synthesizing device 23. Further, the management device 25 controls the master switcher 33 to switch between commercials and extraordinary news.
- the management device 25 may be arranged in the same broadcasting company's station building as the transmission system 24, or may be arranged on the cloud of the 5G network 12.
- the content production system 1 configured as described above distributes content such as sports and entertainment live performed at the event venue by utilizing the 5G network 12.
- the content production system 1 is configured to support various formats of video streams and transmission of a plurality of video streams from one video source 21 so that the switcher function can be appropriately realized.
- the same video source 21 is used.
- the video selection synthesizer 23 needs to treat a plurality of video streams as if they were from the same source (video source 21).
- the format of the video stream input to one input port of the video selection / synthesizing device 23 is switched in the middle depending on the network condition or the like, it is necessary to know the format characteristics of the video stream before and after the switching. be.
- the video selection / synthesizing device 23 switches the video between the video stream input from the predetermined input port and the video stream input from another input port, the format characteristics of the video stream before and after the switching. It is necessary to properly control the switching timing in consideration of the above so as not to impair the video synchronization. Even if the format of the video stream input to one input port switches in the middle, it is necessary to appropriately control the switching timing in consideration of the format characteristics of the video stream before and after switching so as not to impair video synchronization. There is.
- the director or the operator 51 instructs each video source 21 of the format characteristics and controls the operation of the video selection / synthesizing device 23 accordingly. It is also possible to select and combine images taken by a plurality of video sources 21.
- FIG. 4 is a block diagram showing a detailed configuration example of the switcher 31 of the video selection synthesizer 23.
- the switcher 31 includes a plurality of input ports 61, a plurality of decoders 62, a plurality of buffers 63, a selection unit 64, a component unit 65, and an output port 66.
- each of the three video sources 21-1 to 21-3 corresponds to the case where two types of video streams, a proxy video and a high-resolution video, are simultaneously transmitted according to the example of FIG.
- Seven input ports 61A to 61G which are a combination of one input port 61A to 61F and an input port 61G which is an input port for a CG image, are shown. Further, six decoders 62A to 62F and buffers 63A to 63F corresponding to the input ports 61A to 61F into which the video stream is input are shown.
- the number of input ports 61 included in the switcher 31 is not limited to seven, and the input port 61 to be used is appropriately changed according to the number of video sources 21 and the number of input video streams.
- the input ports 61A to 61F each supply the input video stream to the corresponding decoder 62 (62A to 62F).
- the input port 61G supplies the CG image input from another CG image generation device to the component unit 65.
- the decoders 62A to 62F each perform decoding corresponding to the encoding method of the input video stream according to the format characteristic information acquired from the management device 25.
- the decoders 62A to 62F supply the decoded video (video data) to the corresponding buffer 63 (63A to 63F).
- the buffers 63A to 63F each hold the video supplied from the corresponding previous stage decoders 62 (decoders 62A to 62F) for a predetermined time.
- the buffers 63A to 63F appropriately supply the held video to the selection unit 64 in accordance with the request from the selection unit 64.
- the buffer time for each of the buffers 63A to 63F to hold the image supplied from the decoder 62 varies depending on the format characteristic information supplied from the management device 25.
- the selection unit 64 sets each unit in the switcher 31 based on the format characteristic information acquired from the management device 25. Specifically, the selection unit 64 sets the grouping of a plurality of input ports 61 to which the video stream output from the same video source 21 is input. Further, the selection unit 64 sets each decoder 62 with a decoding method corresponding to the encoding method of the input video stream. Further, the selection unit 64 sets the buffer time of each buffer 63. The buffer time of each buffer 63 may be set to correspond to the maximum delay time of the input video stream, or may be set to be smaller than the maximum delay time in order to suppress the buffer amount.
- the selection unit 64 selects one or more of the images stored in the buffers 63A to 63F based on the separately supplied selection layout information (selection + Layout information), and supplies the video to the configuration unit 65.
- the configuration unit 65 appropriately lays out one or more images supplied from the selection unit 64 and a CG image input from another CG image generator via the input port 61G according to the selection layout information. , Configure one video screen.
- the configured video screen (video data) is output to the master switcher 33 of the transmission system 24 via the output port 66.
- the selection layout information supplied to the selection unit 64 and the component unit 65 may be supplied from, for example, the management device 25, or may be supplied from a device other than the management device 25 based on the operation control of the director or the operator 51. May be done.
- each video source 21 also transmits the metadata related to the video stream to be transmitted to the video selection synthesizer 23.
- the metadata may be included in the same file as the video stream and transmitted, or may be transmitted in a file separate from the video stream.
- the switcher 31 has the above configuration.
- the seven input ports 61A to 61F of the video selection synthesizer 23 are of the first input port 61A, the second input port 61B, the third input port 61C, ... It may be described as.
- input ports 61 When it is not necessary to distinguish the input ports 61A to 61F, they are referred to as input ports 61 as described above.
- the six decoders 62A to 62F and the buffers 63A to 63F will be referred to as the first decoder 62A, the second decoder 62B, ..., The first buffer 63A, the second buffer 63B, ... May be done.
- the switcher 31 Based on the format characteristic information acquired from the management device 25, the switcher 31 performs decoding corresponding to the encoding method of the input video stream and buffers according to the delay time of the input video stream. As a result, regardless of whether the video is switched by switching the input port 61 or the format of the video stream input to one input port 61 is switched in the middle, in other words, the video is appropriately used. It is possible to configure and output one video screen without generating a blank.
- each of the video sources 21-1 to 21-3 generates one video stream and transmits it to the video selection / synthesizing device 23.
- the video stream transmitted by the video source 21-1 is input to the first input port 61A
- the video stream transmitted by the video source 21-2 is input to the second input port 61B
- the video stream is input to the third input port 61C.
- a CG image is input to the fourth input port 61D.
- FIG. 5 shows an example of video switching control when the buffer 63 secures a buffer time that matches the maximum delay time of the input video stream.
- the video stream from the video source 21-1 input to the first input port 61A is a first format (format1), for example, a baseband stream, and has the longest delay time. Few.
- the video stream from the video source 21-2 input to the second input port 61B was encoded by a second format (format2) different from the first format, for example HEVC using intra coding. It is a stream, and is input to the video selection / synthesizing device 23 with a delay of 3 frames with respect to the video stream from the video source 21-1.
- the delay time of the video stream input to the second input port 61B is the maximum delay time among the three video streams input to the video selection synthesizer 23.
- the rectangle in which the time code such as “01:23:45:04”, “01:23:45:05”, “01:23:45:06”, etc. is described is one frame.
- the frame in which the same time code is described in the video stream of each input port 61 represents the frame of the same scene.
- the first buffering the video stream input to the first input port 61A When the second buffer 63B buffering the video stream input to the second input port 61B has secured the buffer time of two frames, the first buffering the video stream input to the first input port 61A.
- the buffer 63A secures a buffer time for 5 frames covering 3 frames, which is the maximum delay time.
- the first buffer is used. Five frames whose time codes are from “01:23:45:06" to "01:23:45:10" are buffered in the buffer 63A.
- the switcher 31 bidirectionally (in both directions) between the video of the video stream input to the first input port 61A and the video of the video stream input to the second input port 61B on one of the constituent video screens.
- Input-1 ⁇ Input-2, Input-2 ⁇ Input-1) can be switched without delay.
- the switcher 31 can realize the switcher function without impairing the video synchronization by securing the buffer time according to the maximum delay time of the input video stream based on the format characteristic information.
- FIG. 6 shows an example of video switching control when the buffer 63 secures a buffer time smaller than the maximum delay time of the input video stream.
- the encoding method and delay time of the video stream from are the same as in the example of FIG.
- the buffer time secured by the first buffer 63A is different from the example of FIG. Specifically, the first buffer 63A cannot secure a buffer time sufficient to cover the maximum delay time of 3 frames, and the buffer time is 3 frames.
- the switcher 31 bidirectionally connects the video of the video stream input to the first input port 61A and the video of the video stream input to the second input port 61B on one of the constituent video screens. Synchronous switching is not possible. Specifically, when the video of the video stream input to the first input port 61A is switched to the video of the video stream input to the second input port 61B, the displayed video returns to the past video. It ends up.
- the switcher 31 knows the delay time of each input video stream based on the format characteristic information, for example, the video before switching may be used for a longer time, or a CG image may be inserted. , Effects, etc. can be processed appropriately to form one video screen.
- the buffer time secured by the first buffer 63A can be suppressed, so that the cost can be reduced. Can be done.
- FIG. 7 shows an example of video switching control when the format of the video stream input to the second input port 61B, which is one of the input ports 61, is switched in the middle.
- the video stream from the video source 21-2 input to the second input port 61B (Input-2) is initially the same first format (format 1) video as the video stream input to the first input port 61A. It is a stream and has the same delay time. Then, the video stream from the video source 21-2 is changed to a second format (format 2) different from the first format from a predetermined time after a certain period of time has elapsed.
- the second format is input to the video selection synthesizer 23 with a delay of three frames as compared with the first format.
- the video stream input to the second input port 61B switches from the format with a small delay time to the format with a large delay time.
- the buffer time secured by the first buffer 63A and the second buffer 63B is the time for three frames.
- the switcher 31 switches from the video of the video stream input to the first input port 61A to the video of the video stream input to the second input port 61B on one constituent video screen.
- the format of the video stream input to the second input port 61B has been changed to the second format, when switching from the first input port 61A to the second input port 61B, no video can be obtained. Occurs.
- the switcher 31 knows the delay time of each input video stream based on the format characteristic information, it is possible to take measures for the period during which the video cannot be obtained.
- the selection unit 64 can avoid switching to the video of the second input port 61B during the period when the video cannot be obtained.
- the selection unit 64 may select a video of another video source 21, use the video before switching for a longer period, or insert a CG image during the period when the video cannot be obtained. can. As a result, one video screen can be appropriately configured.
- FIG. 8 shows an example of video switching control when the format of the video stream input to the second input port 61B, which is one of the input ports 61, is switched in the middle.
- the example of FIG. 8 is different from the example of FIG. 7 in that the video stream input to the second input port 61B switches from the format having a large delay time to the format having a small delay time, and is common in other points.
- a video stream of the second format is first input to the second input port 61B (Input-2) from the video source 21-2, and the second input port 61B (Input-2) starts from a predetermined time after a certain period of time.
- a video stream of the first format (format1) different from the format is input.
- the second input port 61B is buffered with video whose time code is up to "01:23:45:07" by the second format
- the input first format video has a time code of "01:23:45:11", and there is a period during which no video can be obtained.
- the switcher 31 knows the period during which the image cannot be obtained based on the format characteristic information, it is possible to take measures for the period during which the image cannot be obtained.
- the selection unit 64 can avoid switching to the video of the second input port 61B during the period when the video cannot be obtained.
- the selection unit 64 may select a video of another video source 21, use the video before switching for a longer period, or insert a CG image during the period when the video cannot be obtained. can. As a result, one video screen can be appropriately configured.
- Parameters for notifying format characteristic information> For notification of format characteristic information from the management device 25 to the video selection synthesizer 23, for example, a general descriptor of ISO / IEC 23090-8 (Non-Patent Document 2) that defines a framework for network-based media processing (NBMP).
- NBMP network-based media processing
- the input-ports object and the configuration descriptor object can be used.
- each input port 61 can be defined.
- the configuration descriptor object it is possible to specify the grouping of each input port 61 defined in the input-ports object, the delay time of the input video stream, and the like.
- each input port 61 Grouping, delay time of input video stream, etc. can be specified.
- FIG. 9 specifies which input port 61 of the video stream is from the same video source 21 when a plurality of video streams are transmitted from one video source 21 and input to the video selection synthesizer 23.
- An example of the parameters to be used is shown.
- the management device 25 specifies the grouping of the input port 61 by transmitting the setting parameter (group parameter) of the input-port-group (input port group) to the video selection synthesizer 23.
- group parameter the setting parameter of the input-port-group (input port group)
- video selection synthesizer 23 the video selection synthesizer 23.
- a plurality of video streams input to a plurality of input ports 61 for which grouping is specified are recognized as being from the same video source.
- input-port-group includes input-group-name, input-ports-by-name, input-type, and maximum-.
- Each parameter of delay (maximum delay) is included.
- the parameter input-group-name defines the name of the group.
- the parameter input-group-name can be set using the camera number or the like.
- the parameter input-ports-by-name defines a list of names of input ports 61 that belong to the group. That is, in the parameter input-ports-by-name, the names of the input ports 61 belonging to the group are arranged by the number of the input ports 61.
- the period during which the video stream is provided at each input port 61 belonging to the group is input to the group. It is defined as many as the number of input ports 61 to which it belongs. For example, “continuous” is specified when the input is always, and “intermittent” is specified when the input is intermittent.
- the maximum delay time (encoding delay time) of the video stream of each input port 61 belonging to the group is defined by the number of input ports 61 belonging to the group.
- the unit of delay time is microsecond.
- FIG. 10 shows an example of a parameter that specifies the format of the video stream of the input port 61 when the format of the video stream input to one input port 61 is switched with time.
- the management device 25 is used in a video stream input from one video source 21 to one input port 61 by transmitting input-formats setting parameters (format parameters) to the video selection synthesizer 23. Gives information about the format to be done.
- Input-formats includes input-port-name, input-format-list, and maximum-delay parameters.
- the parameter input-port-name defines the name of the input port 61.
- the parameter input-format-list defines a list of formats to be used. Since the detailed parameters of the codec are separately defined by InputDescriptor, they are specified as, for example, "intra”, “inter”, and “long-gop”. “Intra” represents, for example, the format of HEVC using intra coding. “Inter” represents a HEVC format using inter-encoding in which a GOP (Group Of Picture) is composed of a short number of frames such as 3 frames of IPB (I picture, P picture, B picture). “Long-gop” represents a HEVC format using inter-encoding in which a GOP is composed of a long number of frames such as 15 frames such as IBBPBB ....
- the parameter maximum-delay defines the maximum delay time (encoding delay time) of the video stream for each format.
- the maximum delay time is defined by the number of formats.
- the unit of delay time is microsecond.
- FIG. 11 shows the relationship between the video streams transmitted by the two video sources 21-1 and 21-2 and the input ports 61 of the switcher 31 to which they are input.
- three types of video streams are transmitted from the video source 21-1 to the switcher 31 of the video selection synthesizer 23.
- the first video stream of the video source 21-1 is intermittently provided in the baseband, and the maximum delay time is 10000 microsecond. This first video stream is input to the first input port 61A of the switcher 31.
- the second video stream of the video source 21-1 is encoded by intra-encoded HEVC (“hevc-intra”) and provided intermittently, and the maximum delay time is 83333 microsecond.
- This second video stream is input to the second input port 61B of the switcher 31.
- the third video stream of the video source 21-1 is always provided by being encoded by inter-encoded HEVC (“hevc-long”) that constitutes GOP with a long number of frames, and the maximum delay time is 500,000 microseconds.
- This third video stream is input to the third input port 61C of the switcher 31.
- the first video stream of the video source 21-2 is intermittently provided by being encoded by inter-encoded HEVC (“hevc-long”) that constitutes GOP with a long number of frames, and the maximum delay time is 83333 microsecond. be.
- This first video stream is input to the fourth input port 61D of the switcher 31.
- the second video stream of the video source 21-2 is always provided, encoded by inter-encoded HEVC (“hevc-long2k”), which reduces the resolution to 2K and configures GOP with a long number of frames, and has a maximum delay time. Is 500000 microseconds.
- This second video stream is input to the fifth input port 61E of the switcher 31.
- the resolution of the above-mentioned "hevc-long” and "hevc-intra” is 4K.
- FIG. 12 shows a description example of the input-ports object and the configuration descriptor object corresponding to FIG.
- the numbers starting from 1 arranged vertically on the left side of FIG. 12 indicate line numbers.
- the "general" on the first line indicates the beginning of the general descriptor object.
- the end of the general descriptor object is the 12th line.
- “Input-ports” on the third line indicates the beginning of the input-ports object.
- the end of the input-ports object is line 10.
- the "port-name" of the 4th input port 61D is "input-4”
- the stream ID of the video stream input to the 4th input port 61D is "454". Indicates that the stream name is “camera-1-hevc-long”.
- the "port-name" of the 5th input port 61E is "input-5", and the stream ID of the video stream input to the 5th input port 61E is "645". Indicates that the stream name is “camera-1-hevc-long2k”.
- “Configuration” on the 14th line indicates the beginning of the configuration object.
- the end of the configuration object is line 22.
- the first input port 61A to the third input port into which the first to third video streams of the video source 21-1 are input using the “input-port-group” of FIG. Grouping of 61C is defined.
- input-type the period during which the video streams of "input-1", “input-2”, and “input-3” are provided is “intermittent”, “intermittent”, and “continuous”, respectively. Is defined as.
- the grouping of 61E is defined.
- the input ports 61 belonging to this group can be "input-4" and “input-5" using the "port-name" of the input-ports object. It is defined.
- the "input-type” defines that the periods during which the "input-4" and "input-5" video streams are provided are “intermittent” and “continuous”, respectively.
- Maximum-delay defines that the maximum delay times for "input-4" and "input-5" are “83333” and "500000”, respectively.
- the point that the video source 21-1 transmits three types of video streams to the first input port 61A to the third input port 61C of the switcher 31 is the first setting of FIG. Since it is the same as the example, the description thereof will be omitted.
- the input to the fourth input port 61D and the fifth input port 61E is different from the first setting example in FIG. Specifically, the video stream transmitted from the video source 21-2 is input to the fourth input port 61D, and the video stream transmitted from the video source 21-3 is input to the fifth input port 61E. Entered.
- the video stream transmitted by the video source 21-2 to the fourth input port 61D is a stream whose format changes with time.
- the video stream transmitted by the video source 21-3 to the fifth input port 61E is also a stream whose format changes with time.
- the video source 21-2 transmits the video stream in the following three formats of the first to third video streams.
- the first video stream transmitted first is a video stream encoded by intra-encoded HEVC (“hevc-intra”), and the maximum delay time is 10000 microsecond.
- the second video stream transmitted second is a video stream encoded by inter-encoded HEVC (“hevc-inter”) with a short number of frames, and the maximum delay time is 83333 microsecond.
- the third video stream transmitted third is a video stream encoded by inter-encoded HEVC (“hevc-long”) with a long number of frames, and the maximum delay time is 500000 microsecond.
- the video source 21-3 transmits the video stream in the following two formats, the first and second video streams.
- the first video stream transmitted first is a video stream encoded by inter-encoded HEVC (“hevc-inter”) with a short number of frames, and the maximum delay time is 83333 microsecond.
- HEVC inter-encoded HEVC
- the second video stream transmitted second is a video stream encoded by inter-encoded HEVC (“hevc-long”) with a long number of frames, and the maximum delay time is 500000 microsecond.
- FIG. 14 shows a description example of the input-ports object and the configuration descriptor object corresponding to FIG.
- the numbers starting from 1 arranged vertically on the left side of FIG. 14 indicate line numbers.
- the "port-name" of the 4th input port 61D is "input-4", and the stream ID of the video stream input to the 4th input port 61D is "454". Indicates that the stream name is "camera-2-multi-form”.
- the "port-name" of the 5th input port 61E is "input-5", and the stream ID of the video stream input to the 5th input port 61E is "645". Indicates that the stream name is “camera-3-multi-form”.
- Lines 18 to 21 define the time-varying formats of the video streams input to the 4th input port 61D and the 5th input port 61E using the "input-formats" of FIG.
- the "input-formats” on the 18th and 19th lines define the formats of the video stream input to the 4th input port 61D.
- “input-port-name” is “input-4” of “port-name” of the input-ports object
- “input-format-list” is “intra”, “inter”, and It is defined to be in the order of "long”, where "maximum-delay” is in the order of "10000", "83333", and "500000”.
- “Intra” of "input-format-list” corresponds to "hevc-intra” mentioned above, "inter” corresponds to "hevc-inter”, and "long” corresponds to "hevc-long”. do.
- the "input-formats” on the 20th and 21st lines define the formats of the video stream input to the 5th input port 61E.
- “input-port-name” is “input-5” of “port-name” of input-ports object
- “input-format-list” is in the order of “inter” and “long”. It is defined that "maximum-delay” is in the order of "83333" and "500000”.
- the time information indicating the timing at which the video stream is input from the video source 21 to the predetermined input port 61 of the switcher 31 of the video selection synthesizer 23 is timed metadata (timed metadata) that stores metadata that changes in time series. ) Can be used to notify the video selection synthesizer 23 from the video source 21.
- ISO / IEC 23001-10 (Non-Patent Document 3) defines media timed metadata metrics for transfer in files based on the ISO-based media file format.
- ISO-based media file format (ISOBMFF)
- data is described by a hierarchical structure of basic data structures called boxes (“boxes”).
- timed metadata stores metadata that has meaning for a specific time in a video stream as a “sample” in one “track”. It defines the method.
- the time information of the video stream is in Movie box-> track box-> media box-> media information box-> sample table box, as in the case of video (video), audio (audio), etc.
- the “coding name” of the sample entry that describes the codec information in the sample description box indicates that the metadata has a specific structure.
- "Movie box-> track box-> media box-> media information box-> sample table box” shows the box hierarchy structure in "Movie box”.
- FIG. 15 shows the syntax of switching information (timed metadata) for a group of input ports 61 that receive a plurality of video streams having different formats from one video source 21.
- StreamSwitchTimingMetaDataSampleEntry can be used by setting the coding name of the sample entry that describes the codec information of the sample description box in the Movie box-> track box-> media box-> media information box-> sample table box to'SSWT'. It shows that the sample structure of the track is metadata with the data structure specified by StreamSwitchingTimingMetadataSample.
- the data structure of StreamSwitchTimingMetaDataSample consists of "ntp_timestamp”, "port_name”, “start_time”, and "end_time” of the number of grouped input ports 61 (number_of_ports).
- Ntp_timestamp represents the time (Wall Clock Time) when the information contained in this sample was generated.
- Port_name represents the name of the input port 61.
- Start_time represents the start time (media time).
- End_time represents the end time (media time).
- FIG. 16 shows the syntax of the format switching information (timed metadata) for the input port 61 in which the format of the video stream is switched.
- FormatSwitchTimingMetaDataSampleEntry is by setting the coding name of the sample entry that describes the codec information of the sample description box in the Movie box-> track box-> media box-> media information box-> sample table box to'FSWT'. It shows that the sample structure of the track is metadata with the data structure specified by FormatSwitchingTimingMetadataSample.
- FormatSwitchTimingMetaDataSample consists of "ntp_timestamp”, “format_switch_time”, and “format_number”.
- Ntp_timestamp represents the time (Wall Clock Time) when the information contained in this sample was generated.
- Form_switch_time represents the time (media time) for switching to the next format indicated by “format_number”.
- Form_number is the number (ordinal number) of the corresponding format in "input-format-list” in the "input formats” parameter in FIG. 10, and represents the number of the format to be switched next.
- the time information of the video stream input from one video source 21 to the plurality of input ports 61 and the time information of the video stream sequentially input while switching in one input port 61 are, for example, a video stream. It is possible to notify sequentially in the same file (MP4 file) using timed metadata. As a result, when the video stream input to each input port 61 is input to the switcher 31 of the video selection synthesizer 23, whether it is constantly input or intermittently input, or , You can know in advance whether it will switch.
- Video screen generation processing for distribution by the content production system > Next, a distribution video screen generation process for generating and outputting a distribution video screen based on format characteristic information will be described with reference to the flowchart of FIG. This process is started, for example, when a command for producing content for distribution is supplied from the management device 25.
- the management device 25 determines the format characteristic information of the video stream generated by each of the plurality of video sources 21 arranged in various places of the relay site 11. For example, in the management device 25, the number of streams to which each video source 21 transmits a video stream according to the priority of the video according to the location of the video source 21 in the event venue, the network status of the 5G network 12, and the like. Determine the encoding method, maximum delay time, etc. Then, the management device 25 transmits the determined format characteristic information to the video selection / synthesis device 23.
- step S2 the selection unit 64 of the switcher 31 of the video selection / synthesis device 23 acquires the format characteristic information transmitted from the management device 25.
- the format characteristic information is transmitted from the management device 25 as a configuration parameter using the input-port-group object shown in FIG. 9 and the like and the input-formats object shown in FIG. 10 and the like.
- the selection unit 64 of the switcher 31 sets each unit in the switcher 31 based on the acquired format characteristic information. Specifically, the selection unit 64 is based on the input-port-group object shown in FIG. 9 and the like, and of the input-group-name, input-ports-by-name, input-type, and maximum-delay. Set each parameter. Further, the selection unit 64 sets each parameter of input-port-name, input-format-list, and maximum-delay based on the input-formats object. As a result, a plurality of input ports 61 to which the video stream output from the same video source 21 is input are grouped.
- each input port 61 The provision period (constant input or intermittent input) of each input port 61 is recognized, and the buffer 63 sets a buffer time corresponding to, for example, the maximum delay time.
- the buffer time of the buffer 63 according to the format of the input video stream and the decoding method of the decoder 62 are set (recognized) in advance. Will be done.
- step S4 the decoder 62 and the buffer 63 corresponding to the input port 61 to which the predetermined video stream is supplied from the video source 21 decode and buffer the input video stream. That is, the decoder 62 decodes by the set decoding method, and the buffer 63 holds the decoded video for the set buffer time. Further, the selection unit 64 supplies the decoded video acquired from the buffer 63 to the storage 32 and stores the video for replay.
- step S4 the selection unit 64 acquires the timed metadata obtained in the same file as the video stream or a different file, switches the intermittent input of the video stream (on or off of the video stream input), and formats the video stream. Recognize the switching of.
- the selection unit 64 changes the decoding method setting of the decoder 62 and the buffer time setting of the buffer 63 at a necessary timing.
- step S5 the selection unit 64 selects the video held in one or more buffers 63 based on the selection layout information separately supplied, and supplies the video to the component unit 65. Further, the selection unit 64 also selects a CG image input from another CG image generation device via an input port 61 different from the video stream, and supplies the CG image to the component unit 65.
- step S6 the component unit 65 appropriately lays out one or more images supplied from the selection unit 64 and the CG image according to the selection layout information, and constitutes one video screen.
- the configuration unit 65 outputs the configured video screen (video data) to the master switcher 33 of the transmission system 24 via the output port 66.
- step S7 the switcher 31 determines whether or not to cancel the production of the content for distribution. For example, the component 65 of the switcher 31 determines that the production of the distribution content is stopped when the command to stop the production of the distribution content is supplied from the management device 25.
- step S7 If it is determined in step S7 that the production of the content for distribution is not stopped yet, the process returns to step S4, and the processes of steps S4 to S7 described above are continuously executed.
- step S7 if it is determined in step S7 that the production of the distribution content is stopped, the distribution content production process of FIG. 17 ends.
- the content production process for distribution by the content production system 1 is executed as described above.
- the switcher function (Production Switcher process) is executed on the server device arranged on the cloud of the 5G network 12 a plurality of video streams from the same video source 21 are identical. It can be properly treated as being from a source (one video source 21). Further, when the format of the video stream input to one input port 61 of the video selection synthesizer 23 is switched in the middle, the format characteristics of the video stream before and after the switching are recognized in advance, and the switching timing is appropriately controlled. can do. This enables appropriate control without impairing the video synchronization of the video screen generated by the switcher function. That is, the Production Switcher function can be appropriately realized.
- the series of processes performed by the video selection and synthesizing device 23 and the management device 25 described above can be executed by hardware or software.
- the programs constituting the software are installed on the computer.
- the computer includes a microcomputer embedded in dedicated hardware and, for example, a general-purpose personal computer capable of executing various functions by installing various programs.
- FIG. 18 is a block diagram showing a configuration example of computer hardware that executes the above-mentioned series of processes programmatically.
- the CPU Central Processing Unit
- ROM Read Only Memory
- RAM Random Access Memory
- An input / output interface 505 is further connected to the bus 504.
- An input unit 506, an output unit 507, a storage unit 508, a communication unit 509, and a drive 510 are connected to the input / output interface 505.
- the input unit 506 includes a keyboard, a mouse, a microphone, a touch panel, an input terminal, and the like.
- the output unit 507 includes a display, a speaker, an output terminal, and the like.
- the storage unit 508 includes a hard disk, a RAM disk, a non-volatile memory, and the like.
- the communication unit 509 includes a network interface and the like.
- the drive 510 drives a removable recording medium 511 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory.
- the CPU 501 loads the program stored in the storage unit 508 into the RAM 503 via the input / output interface 505 and the bus 504 and executes the above-mentioned series. Is processed.
- the RAM 503 also appropriately stores data and the like necessary for the CPU 501 to execute various processes.
- the program executed by the computer can be recorded and provided on a removable recording medium 511 as a package medium or the like, for example. Programs can also be provided via wired or wireless transmission media such as local area networks, the Internet, and digital satellite broadcasts.
- the program can be installed in the storage unit 508 via the input / output interface 505 by mounting the removable recording medium 511 in the drive 510. Further, the program can be received by the communication unit 509 and installed in the storage unit 508 via a wired or wireless transmission medium. In addition, the program can be pre-installed in the ROM 502 or the storage unit 508.
- FIG. 19 shows a block diagram of a computer and cloud computing that can implement the various embodiments described herein.
- the present disclosure can be realized as a system, a method, and / or a computer program.
- the computer program may include a computer-readable storage medium, in which the computer-readable storage medium contains computer-readable program instructions that cause one or more processors to perform aspects of this embodiment. ..
- the computer-readable storage medium can be a tangible device that can store instructions for use by the instruction execution device (processor).
- the computer-readable storage medium can be, for example, an electronic storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of these devices. Not limited. More specific examples of computer-readable storage media include each (and appropriate combination) of the following: flexible disks, hard disks, SSDs (solid state drives), RAMs (random access memory), ROMs (reads): only memory), EPROM (erasable and programmable read only memory) or Flash (flash memory), SRAM (static random access memory), compact disc (CD or CD-ROM), DVD (digitalversatile disc), card type or stick type Memory.
- a computer-readable storage medium as used in the present disclosure is an electromagnetic wave propagating through radio waves or other freely propagating electromagnetic waves, waveguides or other transmission media (eg, optical pulses through fiber optic cables). , Or an electrical signal transmitted over a wire, etc., is not construed as a temporary signal itself.
- Computer-readable program instructions of the present disclosure can be downloaded from a computer-readable storage medium to a suitable computing or processing device, such as a global network such as the Internet, a local area network, a wide area network, and / or. , Can be downloaded to an external computer or external storage device via a wireless network.
- Networks include copper transmission lines, optical communication fibers, wireless transmissions, routers, firewalls, switches, gateway computers, and / or edge servers.
- a network adapter card or network interface within a computing or processing device receives computer-readable program instructions from the network and transfers the computer-readable program instructions within the computing or processing device. It can be stored in a computer-readable storage medium.
- Computer-readable program instructions that perform the processing of the present disclosure include machine language instructions and / or microcodes, which are assembly languages, Basic, Fortran, Java®, Python, R, Compiled or interpreted from source code written in any combination of one or more program languages, including C, C ++, C #, or similar programming languages.
- Computer-readable program instructions can be executed entirely on the user's personal computer, notebook computer, tablet, or smartphone, and can be a remote computer or computer server, or any combination of these computing devices. It can also be fully executed above.
- the remote computer or computer server may be connected to the user's device or device via a computer network such as a local area network, wide area network, or global network (eg, the Internet).
- a computer-readable program instruction in which an electric circuit including a programmable logic circuit, an FPGA (field-programmable gate arrays), and a PLA (programmable logic arrays) constitutes or customizes an electronic circuit.
- computer-readable program instructions can be executed using information from.
- Computer-readable program instructions that can perform the systems and methods described in this disclosure are general purpose computers, dedicated computers, or other programmable device processors (and /) for manufacturing the device. Or used by one or more cores in the processor). Execution of program instructions through the processor of a computer or other programmable device creates a system for realizing the functions described in the flow diagrams and block diagrams of the present disclosure.
- These computer-readable program instructions may also be stored in computer-readable storage media that can instruct computers, programmable devices, and / or other devices to function in a particular way. Accordingly, the computer-readable storage medium in which the instructions are stored is a manufactured article that includes the instructions that implement the functional aspects identified in the flow and block diagrams of the present disclosure.
- Computer-readable program instructions can also be loaded onto a computer, other programmable device, or other device to perform a series of operational steps on the computer, other programmable device, or other device, and the results of the computer's processing. Generate.
- the program instructions are executed on a computer, other programmable device, or other device to realize the functions specified in the flow diagram and block diagram of the present disclosure.
- FIG. 19 is a functional block diagram of a network system 800 in which one or more computers, servers, and the like are connected via a network.
- the hardware and software environment shown in the embodiment of FIG. 19 is shown as an example of providing a platform for realizing the software and / or method according to the present disclosure.
- the network system 800 can include, but is not limited to, a computer 805, a network 810, a remote computer 815, a web server 820, a cloud storage server 825, and a computer server 830. In one embodiment, a plurality of instances of one or more functional blocks shown in FIG. 19 are used.
- FIG. 19 illustrates a more detailed configuration of the computer 805. It should be noted that the functional blocks shown in the computer 805 are shown for establishing exemplary functions, and not all of them are shown. Also, although detailed configurations of the remote computer 815, web server 820, cloud storage server 825, and computer server 830 are not shown, they may include configurations similar to the functional blocks shown for computer 805. can.
- Computer 805 includes personal computers (PCs), desktop computers, laptop computers, tablet computers, netbook computers, personal digital assistants (PDAs), smartphones, or other programs capable of communicating with other devices on the network 810. Any possible electronic device can be used.
- PCs personal computers
- PDAs personal digital assistants
- smartphones or other programs capable of communicating with other devices on the network 810. Any possible electronic device can be used.
- the computer 805 is configured to include a processor 835, a bus 837, a memory 840, a non-volatile storage 845, a network interface 850, a peripheral device interface 855, and a display interface 865.
- a processor 835 a bus 837
- a memory 840 a non-volatile storage 845
- a network interface 850 a peripheral device interface 855
- a display interface 865 a display interface 865.
- Each of these functions is implemented as an individual electronic subsystem (integrated circuit chip or combination of chips and related devices) in one embodiment, and in other embodiments, some functions are combined into a single unit. It may be mounted as a chip (system on chip or SoC (System on Chip)).
- SoC System on Chip
- Processor 835 is one or more single or multi-chip microprocessors designed and / or manufactured by, for example, Intel Corporation, Advanced Micro Devices, Inc. (AMD), Arm Holdings (Arm), Apple Computer, etc. Can be.
- microprocessors are Intel Corporation's Celeron, Pentium®, Core i3, Core i5 and Core i7, AMD's Opteron, Phenom, Athlon, Turion and Ryzen, Arm's Cortex-A, Cortex- Examples include R and Cortex-M.
- Bus 837 can adopt a proprietary or industry standard high-speed parallel or serial peripheral interconnection bus such as ISA, PCI, PCI Express (PCI-e), AGP.
- ISA ISA
- PCI PCI Express
- AGP AGP
- the memory 840 and the non-volatile storage 845 are storage media that can be read by a computer.
- any suitable volatile storage device such as DRAM (Dynamic Random Access Memory) or SRAM (Static RAM) can be adopted.
- the non-volatile storage 845 includes a flexible disk, a hard disk, an SSD (Solid State Drive), a ROM (Read Only Memory), an EPROM (Erasable and Programmable Read Only Memory), a flash memory, a compact disk (CD or CD-ROM), and a DVD (CD or CD-ROM).
- One or more of DigitalVersatileDisc), card type memory, or stick type memory can be adopted.
- program 848 is a set of machine-readable instructions and / or data. This set is stored in non-volatile storage 845 and is used to create, manage, and control the specific software features described in detail in the present disclosure and described in the drawings. In a configuration in which the memory 840 is much faster than the non-volatile storage 845, the program 848 can be transferred from the non-volatile storage 845 to the memory 840 before being executed by the processor 835.
- the computer 805 can communicate and interact with other computers via the network 810 via the network interface 850.
- the network 810 can adopt, for example, a LAN (Local Area Network), a WAN (Wide Area Network) such as the Internet, or a combination of LAN and WAN, including a wired, wireless, or optical fiber connection. ..
- the network 810 consists of any combination of connections and protocols that support communication between two or more computers and related devices.
- the peripheral device interface 855 can input / output data to / from other devices that can be locally connected to the computer 805.
- the peripheral interface 855 provides a connection to an external device 860.
- the external device 860 uses a keyboard, mouse, keypad, touch screen, and / or other suitable input device.
- the external device 860 may also include, for example, a thumb drive, a portable optical or magnetic disk, and a portable computer readable storage medium such as a memory card.
- the software and data that implement the embodiments of the present disclosure, such as program 848, may be stored in such a portable computer readable storage medium.
- the software may be loaded onto the non-volatile storage 845 or instead may be loaded directly onto the memory 840 via the peripheral interface 855.
- Peripheral device interface 855 may use an industry standard such as RS-232 or USB (Universal Serial Bus) for connection with an external device 860.
- the display interface 865 can connect the computer 805 to the display 870, and there is also a form in which the display 870 is used to present a command line or a graphical user interface to the user of the computer 805.
- Display Interface 865 uses one or more of industry standard or dedicated connections such as VGA (Video Graphics Array), DVI (Digital Visual Interface), DisplayPort, HDMI (High-Definition Multimedia Interface) (registered trademark). Can be connected to the display 870.
- the network interface 850 provides communication with another computer, a storage system, or an external device of the computer 805.
- the software programs and data described herein are downloaded from, for example, a remote computer 815, a web server 820, a cloud storage server 825, and a computer server 830 to a non-volatile storage 845 via a network interface 850 and a network 810.
- the systems and methods of the present disclosure can be performed by one or more computers connected to the computer 805 via network interfaces 850 and network 810.
- the systems and methods of the present disclosure are performed by a remote computer 815, a computer server 830, or a combination of interconnected computers on a network 810.
- the data, datasets, and / or databases employed in the embodiments of the systems and methods of the present disclosure are downloaded and stored from a remote computer 815, a web server 820, a cloud storage server 825, and a computer server 830. can do.
- the processes performed by the computer according to the program do not necessarily have to be performed in chronological order in the order described as the flowchart. That is, the processing performed by the computer according to the program includes processing executed in parallel or individually (for example, processing by parallel processing or processing by an object).
- the program may be processed by one computer (processor) or may be distributed processed by a plurality of computers. Further, the program may be transferred to a distant computer and executed.
- the system means a set of a plurality of components (devices, modules (parts), etc.), and it does not matter whether all the components are in the same housing. Therefore, a plurality of devices housed in separate housings and connected via a network, and a device in which a plurality of modules are housed in one housing are both systems. ..
- the configuration described as one device (or processing unit) may be divided and configured as a plurality of devices (or processing units).
- the configurations described above as a plurality of devices (or processing units) may be collectively configured as one device (or processing unit).
- a configuration other than the above may be added to the configuration of each device (or each processing unit).
- a part of the configuration of one device (or processing unit) may be included in the configuration of another device (or other processing unit). ..
- the technique of the present disclosure can have the following configurations.
- a selection unit that makes settings based on the format characteristic information of the video stream output by a plurality of video sources and selects one or a plurality of videos from the videos of the video stream supplied from each of the plurality of video sources.
- An image selection / synthesizing device including a component unit that constitutes one image screen using the one or more images selected by the selection unit.
- the format parameter provides information on a format that switches with time.
- the format characteristic information includes any of the above (1) to (5) including a group parameter for grouping a plurality of video streams input from one video source to a plurality of input ports as being from the same video source.
- the file of the video stream supplied from the video source includes any of the above (1) to (10) including switching information for a group of input ports receiving a plurality of video streams having different formats from one video source.
- the video stream file supplied from the video source includes format switching information for an input port in which the video stream format is switched.
- the format switching information includes a time for switching to the next format.
- a decoder for decoding the video stream and a buffer for holding the decoded video for a predetermined time are further provided for the number of video streams supplied from each of the plurality of video sources.
- the video selection / synthesizing device according to any one of (1) to (15), wherein the selection unit sets the decoding method of the decoder and the buffer time of the buffer based on the format characteristic information of the video stream.
- the selection unit sets the grouping of a plurality of input ports to which a video stream supplied from the same video source is input based on the format characteristic information of the video stream.
- the video selection synthesizer described.
- the video selection synthesizer Make settings based on the format characteristic information of the video stream output by multiple video sources.
- One or more videos are selected from the videos of the video stream supplied from each of the plurality of video sources.
- a control method for a video selection synthesizer that constitutes one video screen using the selected one or more videos.
- a management device that generates format characteristic information for video streams output by multiple video sources and notifies the video selection synthesizer.
- 1 Content production system 11 Relay site, 12 5G network, 21 (21-1 to 21-3) video source, 22 (22-1 to 22-3) CCU, 23 video selection synthesizer, 24 transmission system, 25 management Equipment, 26 (26-1, 26-2) base station, 27 broadcasting network, 28 network, 31 switcher, 32 storage, 33 master switcher, 51 operator, 61 (61A to 61G) input port, 62 (62G to 62F) Decoder, 63 (63A to 63F) buffer, 64 selection unit, 65 configuration unit, 66 output port, 502 ROM, 504 bus, 505 input / output interface, 506 input unit, 507 output unit, 508 storage unit, 509 communication unit
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Studio Circuits (AREA)
Abstract
La présente divulgation concerne un dispositif de sélection-synthèse de vidéo, un procédé de commande associé, et un dispositif de gestion, avec lesquels il est possible de réaliser de manière appropriée une fonction de commutateur de production. Le dispositif de sélection-synthèse de vidéo comprend : une unité de sélection pour effectuer des réglages sur la base des informations de caractéristique de format de flux vidéo délivrés en sortie par une pluralité de sources vidéo et sélectionner une vidéo ou une pluralité de vidéos parmi les vidéos de flux vidéo fournis à partir de chacune de la pluralité de sources vidéo ; et une unité de composition pour composer un écran vidéo au moyen de la ou des vidéos sélectionnées par l'unité de sélection. La technologie de la présente divulgation peut être appliquée, par exemple, à un système de production de contenu, etc., qui délivre en sortie, en tirant parti d'un réseau 5G, un contenu tel que du sport ou un divertissement en direct réalisé au niveau d'un lieu d'événement.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US202063131129P | 2020-12-28 | 2020-12-28 | |
US63/131,129 | 2020-12-28 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022145215A1 true WO2022145215A1 (fr) | 2022-07-07 |
Family
ID=82260461
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2021/045930 WO2022145215A1 (fr) | 2020-12-28 | 2021-12-14 | Dispositif de sélection-synthèse de vidéo, procédé de commande associé, et dispositif de gestion |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2022145215A1 (fr) |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2004091198A1 (fr) * | 2003-04-09 | 2004-10-21 | Matsushita Electric Industrial Co., Ltd. | Procede, dispositif et logiciel de decodage d'image en synthese d'affichage d'ecran, et support d'enregistrement |
JP2012147412A (ja) * | 2010-12-20 | 2012-08-02 | Ricoh Co Ltd | 画像形成装置、画像形成方法、及び集積回路 |
JP2019075724A (ja) * | 2017-10-17 | 2019-05-16 | キヤノン株式会社 | 映像伝送方法、映像伝送システム及び選択装置 |
-
2021
- 2021-12-14 WO PCT/JP2021/045930 patent/WO2022145215A1/fr active Application Filing
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2004091198A1 (fr) * | 2003-04-09 | 2004-10-21 | Matsushita Electric Industrial Co., Ltd. | Procede, dispositif et logiciel de decodage d'image en synthese d'affichage d'ecran, et support d'enregistrement |
JP2012147412A (ja) * | 2010-12-20 | 2012-08-02 | Ricoh Co Ltd | 画像形成装置、画像形成方法、及び集積回路 |
JP2019075724A (ja) * | 2017-10-17 | 2019-05-16 | キヤノン株式会社 | 映像伝送方法、映像伝送システム及び選択装置 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20240283834A1 (en) | Systems and methods for multi-device media broadcasting or recording with low-latency active control | |
CN110402581B (zh) | 控制媒体内容捕获以进行实况视频广播制作系统和方法 | |
CN112752115B (zh) | 直播数据传输方法、装置、设备及介质 | |
US11895352B2 (en) | System and method for operating a transmission network | |
US20160373496A1 (en) | Content supply device, content supply method, program, terminal device, and content supply system | |
US10939150B2 (en) | Content supply device, content supply method, program, terminal device, and content supply system | |
US10264208B2 (en) | Layered display content for wireless display | |
CN111182322B (zh) | 一种导播控制方法、装置、电子设备和存储介质 | |
US20200329387A1 (en) | Miracast Framework Enhancements For Direct Streaming Mode | |
US10469860B1 (en) | Chunk parallel transcoding using overlapping frames | |
US20130250975A1 (en) | Method and device for packetizing a video stream | |
WO2022145215A1 (fr) | Dispositif de sélection-synthèse de vidéo, procédé de commande associé, et dispositif de gestion | |
JP6501127B2 (ja) | 情報処理装置および方法 | |
WO2014196393A1 (fr) | Dispositif de fourniture de contenu, procédé de fourniture de contenu, programme, dispositif terminal, et système de fourniture de contenu | |
US11005908B1 (en) | Supporting high efficiency video coding with HTTP live streaming | |
KR20140050522A (ko) | 멀티비전 가상화 시스템 및 가상화 서비스 제공 방법 | |
US20170374243A1 (en) | Method of reducing latency in a screen mirroring application and a circuit of the same | |
JP2024131376A (ja) | 放送・ネット同時配信システム、マスタ装置及びプログラム | |
JP2016076884A (ja) | マルチメディア同期再生装置及びマルチメディア同期再生方法 | |
CN117278778A (zh) | 一种图像处理方法、装置、拼接控制器及图像处理系统 | |
Cousins | Changing the Game: A Guide to Cost-Efficient Software-Based HEVC Video Processing Deployment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21915076 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21915076 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: JP |