WO2023119488A1 - Video compositing system, video compositing method, and video compositing program - Google Patents

Video compositing system, video compositing method, and video compositing program Download PDF

Info

Publication number
WO2023119488A1
WO2023119488A1 PCT/JP2021/047582 JP2021047582W WO2023119488A1 WO 2023119488 A1 WO2023119488 A1 WO 2023119488A1 JP 2021047582 W JP2021047582 W JP 2021047582W WO 2023119488 A1 WO2023119488 A1 WO 2023119488A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
transmission
unit
information
video
Prior art date
Application number
PCT/JP2021/047582
Other languages
French (fr)
Japanese (ja)
Inventor
広夢 宮下
真二 深津
英一郎 松本
麻衣子 井元
Original Assignee
日本電信電話株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 日本電信電話株式会社 filed Critical 日本電信電話株式会社
Priority to PCT/JP2021/047582 priority Critical patent/WO2023119488A1/en
Publication of WO2023119488A1 publication Critical patent/WO2023119488A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring

Definitions

  • the present invention relates to a video synthesizing system, a video synthesizing method, and a video synthesizing program.
  • a method is known in which a plurality of video signals are input and a desired output image is generated by synthesizing the plurality of video signals.
  • These methods are generally realized by functions such as mixer/keyer provided in devices such as live production switchers.
  • Japanese Patent Laid-Open No. 2004-100000 discloses a multi-point communication conference system in which a bit rate of image data can be freely changed for each client terminal, and means for arbitrarily switching between a synthesized image of a client and a specific personal image is provided, thereby allowing an individual It discloses a technique for viewing images with high definition.
  • the client terminal transmits video to the server device in a scalable bitstream by hierarchical coding, and is returned from the server device. Transmission bandwidth is saved by enabling the client terminal to specify the video.
  • Patent Document 1 assumes control by a scalable bitstream, image quality tends to deteriorate in video with a low bit rate.
  • Patent Document 1 assumes only simultaneous display of a plurality of images with respect to image synthesis by a server device, deterioration in image quality tends to appear on the finally displayed image.
  • image effects such as mixing in which the display area of each image cannot be clearly divided.
  • the object of the present invention is to address the above-mentioned circumstances, and the object of the present invention is to notify in advance of synthesis information regarding video synthesis, and to allow the server on the transmission side to process and compress the input image based on the synthesis information. to provide a technique for realizing high-definition transmission and synthesis of a plurality of images even on a transmission path with a narrow transmission band.
  • a video composition system includes a composition instruction unit that generates composition information indicating what kind of image blocks an output image is composed of, and a composition instruction unit that transmits the composition information.
  • synthesizing information related to video synthesizing is notified in advance, and the server on the transmitting side processes and compresses the input image based on the synthesizing information. It is possible to provide high-definition transmission and composition of video.
  • FIG. 1 is a diagram illustrating a configuration example of a video synthesizing system according to an embodiment.
  • FIG. 2 is a diagram showing a configuration example of the video synthesizing system shown in FIG.
  • FIG. 3 is a schematic diagram illustrating an example of the configuration of a computer functioning as a transmission server according to the embodiment;
  • FIG. 4 is a schematic diagram illustrating an example of the configuration of a computer functioning as a receiving server according to the embodiment;
  • FIG. 5 is a flowchart illustrating an example of processing operations of the transmission server.
  • FIG. 6 is a flow chart showing an example of the processing operation of the receiving server.
  • FIG. 7 is a diagram showing an example of an input image block.
  • FIG. 1 is a diagram illustrating a configuration example of a video synthesizing system according to an embodiment.
  • FIG. 2 is a diagram showing a configuration example of the video synthesizing system shown in FIG.
  • FIG. 3 is a schematic diagram illustrating an example of the configuration of
  • FIG. 8 is a diagram showing an example of a transfer image block in picture-in-picture and a synthesis result.
  • FIG. 9 is a diagram illustrating an example of synthesis information in picture-in-picture.
  • FIG. 10 is a diagram showing an example of synthesis information in alpha blending.
  • FIG. 11 is a diagram showing an example of transferred image blocks and synthesis results in alpha blending.
  • FIG. 12 is a diagram showing an example of composition information when image processing is performed by composition information alpha blending.
  • FIG. 13 is a diagram showing the transmission timing of transmission image blocks and image synthesis in dissolve.
  • FIG. 14 is a diagram illustrating an example of synthesis information in dissolve.
  • FIG. 1 is a diagram illustrating a configuration example of a video synthesizing system according to an embodiment.
  • the video synthesizing system 1 includes a first transmission server 2A located at a first site A, a second transmission server 2B located at a second site B, and a transmission server 2B located at a site C. a third transmission server 2C, a fourth transmission server 2D located at a site D, a reception server 3, and a network 4 for relaying video signals and the like transmitted between them.
  • the first transmission server 2A, the second transmission server 2B, the third transmission server 2C, and the fourth transmission server 2D are simply the transmission server 2 when there is no need to distinguish them. and described.
  • the example shown in FIG. 1 shows a plurality of transmission servers 2 and a single reception server 3, the number of transmission servers 2 may be one, and the number of reception servers 3 may be plural. is of course.
  • the transmission server 2 is located at each base.
  • the transmission server 2 transmits the video shot at each base to the reception server 3 via the network 4 .
  • the receiving server 3 is installed at any location.
  • the reception server 3 can synthesize the video received from the transmission server 2 and output the synthesized image.
  • FIG. 2 is a diagram showing a configuration example of the video synthesizing system 1 shown in FIG. In the example of FIG. 2, only the first transmission server 2A and the second transmission server 2B are shown for simplification of explanation.
  • the first transmission server 2A includes a first imaging section 21A, a deformation section 22, an image transmission section 23, and a synthesis information reception section 24.
  • the second transmission server 2B has the same configuration as the first transmission server 2A, only the second imaging unit 21B is shown in the example of FIG. 2 for simplification.
  • the example of FIG. 2 shows an example in which the first imaging unit 21A and the second imaging unit 21B are arranged outside the first transmission server 2A and the second transmission server 2B.
  • the imaging unit 21A may be arranged inside the first transmission server 2A
  • the second imaging unit 21B may be arranged inside the second transmission server 2B.
  • the first imaging section 21A and the second imaging section 21B are simply referred to as the imaging section 21 when there is no need to distinguish between them.
  • the receiving server 3 includes a first image receiving section 31A, a second image receiving section 31B, an image synthesizing section 32, a synthesizing information transmitting section 33, a display section 34, and a synthesizing instruction section 35.
  • the display unit 34 and the composition instruction unit 35 are arranged outside the receiving server 3, but the display unit 34 and the composition instruction unit 35 may be arranged inside the reception server 3.
  • the first image receiving section 31A and the second image receiving section 31B are simply referred to as the image receiving section 31 when there is no need to distinguish them.
  • the imaging unit 21 shoots images at the base.
  • the imaging unit 21 acquires an input image block from a part of the video frame of the captured video and outputs it to the transformation unit 22 .
  • the transformation unit 22 transforms the input image block received from the imaging unit 21 based on the composition information received from the composition information reception unit 24, which will be described later, to generate the transmission image block.
  • the composition information indicates what kind of image blocks the output image is composed of.
  • the transformation unit 22 then outputs the transmission image block to the image transmission unit 23 . A method for generating transmission image blocks from input image blocks will be described later.
  • the image transmission unit 23 transmits the transmission image block to the first image reception unit 31A of the reception server 3, which will be described later, via the network 4.
  • the combining information receiving unit 24 receives combining information from the receiving server 3 via the network 4 and outputs the received combining information to the transforming unit 22 .
  • the second transmission server 2B is the same as the first transmission server 2A, so the description is omitted here. Note that the image transmission unit 23 of the second transmission server 2B transmits the transmission image block to the second image reception unit 31B of the reception server 3, which will be described later, via the network 4. FIG.
  • the first image reception unit 31A and the second image reception unit 31B output the transmission image blocks received from the first transmission server 2A and the second transmission server 2B to the image composition unit 32, respectively.
  • the image synthesizing unit 32 synthesizes the output image from the transmission image block based on the synthesizing information received from the synthesizing instruction unit 35 to be described later, and outputs the synthesized output image to the display unit 34 .
  • the combining information transmission unit 33 transmits the combining information received from the combining instruction unit 35 (to be described later) to the transmission server 2 .
  • the display unit 34 displays the output image received from the image synthesizing unit 32 .
  • the synthesizing instruction unit 35 generates synthesizing information.
  • the synthesis information may be, for example, manually input or generated by parameters set by the administrator of the receiving server 3, or may be automatically generated by the synthesis instruction unit 35 reading settings stored in the data memory 303 or the like. It may be generated in
  • the network 4 stably secures a band sufficient for transmitting one video stream.
  • the network 4 has a band capable of stably transmitting and receiving 4K UHD (Ultra High Definition) resolution 60 fps video at 12 Gbps.
  • FIG. 3 is a schematic diagram showing an example of the configuration of a computer functioning as the transmission server 2 according to the embodiment.
  • the transmission server 2 as shown in FIG. 3, is composed of a computer device and has a processor 201 such as a CPU.
  • a program memory 202 a data memory 203 , a communication interface 204 and an input/output interface 205 are connected to the processor 201 via a bus 206 .
  • the program memory 202 includes storage media such as EPROM (Erasable Programmable Read Only Memory), HDD (Hard Disk Drive), SSD (Solid State Drive) and other non-volatile memories that can be written and read at any time, and ROM (Read Only Memory) can be used in combination with non-volatile memory.
  • the program memory 202 stores programs necessary for executing various processes. That is, all of the various processes described below can be realized by reading and executing a program stored in the program memory 202 by the processor 201 .
  • the data memory 203 is storage that uses, as a storage medium, a combination of a non-volatile memory that can be written and read at any time, such as an HDD or memory card, and a volatile memory such as a RAM (Random Access Memory).
  • the data memory 203 is used to store data acquired and generated while the processor 201 executes the program and performs various processes.
  • the communication interface 204 includes one or more wired or wireless communication modules.
  • the communication interface 204 includes a communication module for wired or wireless connection with other devices including the receiving server 3 .
  • the communication interface 204 may include a wireless communication module for wirelessly connecting with other devices using short-range wireless technology. That is, the communication interface 204 may be a general communication interface as long as it can communicate with other devices and transmit and receive various information under the control of the processor 201 .
  • An input unit 207 and a display unit 208 are connected to the input/output interface 205 .
  • the input unit 207 includes at least one camera or video camera.
  • the camera or video camera included in the input unit 207 captures a target captured image and outputs the captured image to the input/output interface 205 .
  • the input unit 207 is, for example, an input detection sheet that employs an electrostatic method or a pressure method and is arranged on the display screen of a display device that is the display unit 208.
  • the transmission server 2 administrator's touch position is output to the processor 201 .
  • the display unit 208 is a display device using, for example, liquid crystal, organic EL (Electro Luminescence), etc., and displays images and messages according to signals input from the input/output interface 205 .
  • FIG. 4 is a schematic diagram showing an example of the configuration of a computer functioning as the reception server 3 according to the embodiment.
  • the receiving server 3 as shown in FIG. 4, is composed of a computer device and has a processor 301 such as a CPU.
  • a program memory 302 , a data memory 303 , a communication interface 304 and an input/output interface 305 are connected to the processor 301 via a bus 306 .
  • non-volatile memory such as EPROM, HDD, SSD, etc.
  • non-volatile memory such as ROM
  • the program memory 302 stores programs necessary for executing various processes. That is, all of the various processes described below can be realized by reading and executing a program stored in the program memory 302 by the processor 301 .
  • the data memory 303 is storage that uses, as a storage medium, a combination of a non-volatile memory that can be written and read at any time, such as an HDD or memory card, and a volatile memory such as a RAM.
  • the data memory 303 is used to store data acquired and generated while the processor 301 executes the program and performs various processes.
  • the communication interface 304 includes one or more wired or wireless communication modules.
  • the communication interface 304 includes a communication module that makes wired or wireless connections with other devices, including the transmission server 2 .
  • the communication interface 304 may include a wireless communication module for wirelessly connecting with other devices using short-range wireless technology. That is, the communication interface 304 may be a general communication interface as long as it can communicate with other devices under the control of the processor 301 and transmit and receive various kinds of information.
  • An input unit 307 and a display unit 308 are connected to the input/output interface 305 .
  • the input unit 307 is, for example, an input detection sheet that employs an electrostatic method or a pressure method and is arranged on the display screen of a display device, which is the display unit 308 .
  • the administrator's touch position is output to the processor 301 .
  • the display unit 308 is a display device using, for example, liquid crystal, organic EL (Electro Luminescence), etc., and displays images and messages according to signals input from the input/output interface 305 .
  • FIG. 5 is a flow chart showing an example of the processing operation of the transmission server 2
  • FIG. 6 is a flow chart showing an example of the processing operation of the reception server 3.
  • the processor 201 of the transmission server 2 reads and executes the program stored in the program memory 202
  • the operation of the flowchart shown in FIG. 5 is realized.
  • the processor 301 of the receiving server 3 reads and executes the program stored in the program memory 302, thereby realizing the operation of the flowchart shown in FIG.
  • the operation may be started when it is necessary to create synthesis information, for example, triggered by a video effect (Digital Video Effect, etc.) realized by a live production switcher, etc.
  • the operation may be started at arbitrary timing. For example, it may be initiated when there is a need to create synthetic information.
  • Synthesis information receiving section 24 receives synthesis information (step ST101).
  • the combining information is generated by the combining instruction section 35 of the receiving server 3 .
  • the synthesis information receiving section 24 outputs the received synthesis information to the transformation section 22 . Details of the information included in the synthesis information will be described later.
  • the composite information may be received at any timing before step ST102 is processed. That is, it may be received in advance by the transmission server 2 and stored in the data memory 203 or the like. Then, each unit may be read out from the data memory 203 and used when necessary. Further, step ST101 may of course be processed in parallel with step ST102, which will be described later.
  • the imaging unit 21 acquires an input image block (step ST102).
  • the imaging unit 21 acquires a video frame and acquires a part of the acquired video frame as an input image block.
  • one frame may be buffered to obtain an input image, and then divided into input image blocks.
  • the imaging unit 21 then outputs the input image block to the deformation unit 22 .
  • FIG. 7 is a diagram showing an example of an input image block.
  • FIG. 7 is the input image A acquired by the imaging unit 21 of the first transmission server 2A at the first site A
  • (b) is a diagram showing an example of an input image block.
  • FIG. 7C is an input image B acquired by the imaging unit 21 of the second transmission server 2B at the second site B
  • FIG. 7C is a diagram showing an example of an input image block.
  • (d) as well as (b), it is assumed that the input image block is divided into 16 blocks of 4 vertical ⁇ 4 horizontal, for example, B-01 to B-16.
  • the transformation unit 22 transfers the input image block to the transmission image block (step ST103).
  • the transformation unit 22 transfers the pixels included in the input image block to the transmission image block based on the synthesis information. Further, when it is necessary to transform the input image block according to the content of the synthesis information, the transformation unit 22 may set different pixel coordinates before and after the pixel reference and transfer.
  • the transformation unit 22 compresses the transmission image block (step ST104).
  • a transformation unit 22 compresses the transmission image block based on the synthesis information.
  • the compression method may be a general method, detailed description thereof is omitted here.
  • step ST104 may be skipped if there is no compression flag in the combined information. Then, the transformation unit 22 outputs the compressed or skipped transmission image blocks to the image transmission unit 23 .
  • the image transmission unit 23 transmits the transmission image block to the receiving server 3 (step ST105).
  • the image transmission unit 23 receives the transmission image block from the transformation unit 22 and transmits the received transmission image to the reception server 3 through the network 4 .
  • Synthesis information transmitting section 33 transmits the synthesis information to transmission server 2 (step ST201).
  • the composition instruction unit 35 outputs composition information to the composition information transmission unit 33 and the image composition unit 32 .
  • the combined information transmission unit 33 transmits the combined information to each transmission server 2 through the network 4 .
  • the synthesis information may be transmitted in advance before frame processing is started. In other words, the processing of step ST201 may be performed at any timing as long as the frame is transmitted before or during processing by the transmission server 2 .
  • the image receiving unit 31 receives the transmission image block (step ST202).
  • the first image receiving unit 31A receives the transmission image block generated based on the input image A from the first transmission server 2A at the base A
  • the second image receiving unit 31B receives the transmission image block at the base B.
  • a transmission image block generated based on the input image B is received from the second transmission server 2B.
  • the first image receiving section 31A and the second image receiving section 31B output the received transmission image blocks to the image synthesizing section 32, respectively.
  • the image synthesizing unit 32 determines whether or not the transmission image block is a single block (step ST203). For example, when the transmission image block is received only from the first transmission server 2A, the image synthesizing unit 32 determines that the transmission image block is single. In this case, the process proceeds to step ST204. On the other hand, when transmission image blocks are received from the first transmission server 2A and the second transmission server 2B, it is determined that there are a plurality of transmission image blocks. In this case, the process proceeds to step ST205.
  • the image synthesizing unit 32 inserts the transmission image block into the output image (step ST204).
  • the image synthesizing unit 32 generates an output image by arranging the transmission image blocks in sequence.
  • the image synthesizing unit 32 generates an output image from a plurality of transmission image blocks (step ST205).
  • the image composition unit 32 combines and/or blends the transmission image blocks based on the composition information to generate an output image.
  • the image composition unit 32 outputs the output image to the display unit 34 (step ST206).
  • the image synthesizing section 32 outputs the generated output image to the display section 34 .
  • the display unit 34 displays video frames according to the output image.
  • FIG. 8 is a diagram showing an example of a transfer image block and a synthesis result (output image) in picture-in-picture.
  • (a) of FIG. 8 is the input image A reduced to 1/2, and (b) shows an input image block corresponding to the input image A of (a).
  • (c) is obtained by deleting a portion of the input image B that overlaps with the input image A, and (d) shows an input image block corresponding to the input image B in (c).
  • (e) shows the image of the final synthesis result, that is, the output image, and (f) shows the image block corresponding to (e).
  • R-xx indicates the transmission image block.
  • the first transmission server 2A transmits the transfer image block shown in (b) to the reception server 3, and the second transmission server 2B sends the transfer image block shown in (d) to the reception server 3. Then, the image synthesizing unit 32 of the receiving server 3 combines these transferred image blocks to generate an output image shown in (e).
  • the input image B is assigned to the image block corresponding to the peripheral portion, and the input image A is assigned to the image block corresponding to the central portion of the image.
  • the image synthesizing unit 32 generates an output image by combining the input image A and the input image B based on the synthesizing information.
  • FIG. 9 is a diagram illustrating an example of synthesis information in picture-in-picture.
  • FIG. 9(a) shows how input image A and input image B are transformed to generate transmission image blocks.
  • the first column in (a) indicates the transmission image block
  • the second column indicates which input image block to use for the transmission image block
  • the third column indicates the input image block to use. It shows reference flags that indicate how to transform or not.
  • (b) indicates how to compress the transmitted image block.
  • R-06, R-07, R-10, and R-11 are each assigned an input image block of input image A, and a reference flag for scaling 0.5 is set.
  • Input image blocks of input image B are assigned to R-01 to R-05, R-08 to R-09, and R-12 to R-16, respectively, and reference flags for copying input image B are set. ing. That is, these transmission image blocks are generated by directly transferring the input image B without transforming it.
  • the output image block corresponds to either the transfer image block received from the first transmission server 2A or the second transmission server 2B, so there is no need to perform blending. . Therefore, the image synthesizing unit 32 can generate an output image by arranging and combining the transmission image blocks received from the first transmission server 2A and the second transmission server 2B according to the serial numbers.
  • the first transmission server 2A and the second transmission server 2B transmit the input image A and the input image B without processing, respectively, a transmission band for two streams is required. If the network 4 only has a transmission band for one stream, the network 4 will be congested and the receiving server 3 will either be unable to receive the input image, or will receive a degraded input image. However, in this embodiment, the first transmission server 2A and the second transmission server 2B transmit transmission image blocks processed in advance into image blocks constituting output image blocks. As a result, the total amount of data output from all transmission servers 2 does not exceed the transmission band for one stream. Therefore, since congestion does not occur in the network 4, the reception server 3 can stably receive a plurality of transmission image blocks.
  • each input video is received without any processing, and processing such as blending is executed with all the data in place.
  • processing such as blending
  • video data of n ⁇ 12 Gbps will flow into the video synthesizing server and its preceding transmission line. Therefore, especially when uncompressed video transmission such as ST2110-20 is considered, it is necessary to design a video transmission line on the premise of multi-streaming according to the number of connected devices, which increases the cost.
  • the transmission server 2 is controlled to transmit only the required number of image blocks based on the composition information. Therefore, it is sufficient to secure a transmission band for one stream on all video transmission lines.
  • FIG. 10 is a diagram showing an example of synthesis information in alpha blending.
  • each input image and output image may be, for example, 4K UHD resolution (3840 ⁇ 2160).
  • the x-axis of FIG. 10 indicates pixel position and the y-axis indicates alpha value.
  • the resolution of one transfer block is 960 ⁇ 540.
  • Input image A shown in FIG. 10 has been moved to the left by 960 pixels. In other words, it shows a state of moving from the original x-coordinate 0 to 960, and A-01, A-05, A-09, and A-13 of the transfer blocks are not used for the output image.
  • input image B has been shifted to the right by 960 pixels. In other words, it shows a state of moving from the original x-coordinate of 2880 by 3840, and B-04, B-08, B-12 and B-16 of the transfer blocks are not used for the output image. Therefore, these transfer blocks are controlled so as not to be transmitted from the transmission server 2.
  • FIG. the overlapped portion is synthesized with an alpha value when generating the output image.
  • the pixel value of the input image A at the coordinates (m, n) is A (m, n)
  • the pixel value of the input image B is B (m, n)
  • the alpha value (ranging from 0 to 100%) is
  • ⁇ (m, n) the pixel value R(m, n) of the synthesized image (output image) is calculated as follows.
  • FIG. 11 is a diagram showing an example of transferred image blocks and synthesis results (output images) in alpha blending.
  • FIG. 11(a) is the input image A according to alpha blending
  • (b) is the transferred image block
  • (c) is according to the size ratio according to the numerical values shown at the bottom of (b). Shows a compressed image.
  • (d) is the input image B according to alpha blending
  • (e) is the transferred image block
  • (f) is according to the size ratio according to the numbers shown at the bottom of (e). Shows a compressed image.
  • (h) indicates the image of the final synthesis result, that is, the output image
  • (i) indicates the image block corresponding to (h).
  • transmission image blocks (R-04, R-08, R-12, R-16) corresponding to the right end of the image.
  • pixels are not transferred to the transmission image blocks (R-01, R-05, R-09, R-13) corresponding to the left end of the image.
  • Transmission image blocks (R-02 to R-03, R-06 to R-07, R-10 to R-11, R-14 to R-15), which are overlapping portions, are input image A and input image B are included in the transmission image block.
  • these transmitted image blocks will be compressed based on the synthesis information.
  • FIG. 12 is a diagram showing an example of synthesis information when image processing is performed using alpha blending.
  • (a) of FIG. 12 is combined information referred to by the first transmission server 2A
  • (b) is combined information referred to by the second transmission server 2B.
  • the first column of the synthesis information indicates the transmission image block
  • the second column indicates which input image block to use for the transmission image block
  • the third column indicates how the input image block to be used.
  • a reference flag indicating whether to transform is shown
  • the fourth column shows a compression flag indicating how the transmission image block is compressed (ie, data volume after compression).
  • xx in R-xx-y in the first column indicates the serial number of the image block
  • y indicates the location.
  • the transformation unit 22 of the transmission server 2 reduces the size of the transferred image block based on the value of the compression flag
  • the image synthesis unit 32 of the reception server 3 performs processing to restore the size after receiving the block.
  • the compression flag may be calculated from the alpha value for alpha blending. For example, in the regions corresponding to R-02, R-06, R-10, and R-14, referring to FIG. 11, the average alpha value of input image A is 75%, and the average alpha value of input image B is 25%. , the value corresponding to it is set as the compression flag.
  • the synthesis instruction unit 35 may generate synthesis information so that the compression rate of light portions (portions with low alpha values) is increased and the compression rate of dark portions (portions with high alpha values) is decreased. That is, the compression rate of a transmission image block that is output dark as an output image is lower than the compression rate of a transmission image block that is output light as an output image. For example, pixels output lightly as an output image are considered to be less important than pixels output darkly as an output image. Therefore, even if the image quality is degraded by increasing the compression rate of pixels that are output lightly, there is an effect that the influence of image quality deterioration due to compression and decompression of pixels that are output dark is suppressed.
  • Each transmission server 2 compresses the transmission image block based on the synthesis information and transmits the data to the reception server 3. After receiving the transmission image block, the receiving server 3 decompresses the transmission image block based on the synthesis information, and generates an output image from the transmission image block.
  • the transmission image blocks are compressed by the alpha value, the total amount of data output from all transmission servers 2 does not exceed the transmission band for one 4K UHD stream. Therefore, since congestion does not occur in the network 4, the reception server 3 can stably receive a plurality of transmission image blocks.
  • FIG. 13 is a diagram showing transmission timing of transmission image blocks and image synthesis (output image) in dissolve.
  • the imaging unit 21 at each site outputs a video stream at a frame rate of 60 fps.
  • the synthesizing instruction unit 35 of the receiving server 3 generates synthesizing information so as to control the frame rate transmitted from the transmitting server 2 according to the dissolve transition rate.
  • 13(a) shows a frame transmitted from the first transmission server 2A
  • (b) shows a frame transmitted from the second transmission server 2B
  • (c) is displayed on the display unit 34.
  • An example of an output image is shown.
  • ⁇ in (a) and (b) indicates that the transmission image block is transmitted, and ⁇ indicates that the transmission image block is not transmitted.
  • the transmission timings of the transfer image blocks transmitted by the first transmission server 2A become sparse, and the transmission timings of the transfer image blocks transmitted by the second transmission server 2B become denser. shows an example. By doing so, the video of the output image is switched from the input image A to the input image B gradually.
  • FIG. 14 is a diagram illustrating an example of synthesis information in dissolve.
  • FIG. 14 is a diagram showing an example of synthesis information when a dissolve is instructed from 00:01:00 to 00:07:00, for example.
  • the first column in the table of FIG. 14 indicates the time
  • the second column indicates the frame rate transmitted by the first transmission server 2A
  • the third column indicates the frame rate transmitted by the second transmission server 2B. show.
  • the blending ratio changes as time progresses, so the video frame rate is increased or decreased accordingly.
  • the bandwidth of the transfer image block is saved by controlling the frame rate.
  • the quality of the output image may be degraded.
  • the sending server 2 and the receiving server 3 may be synchronized by peer-to-peer (PTP) or the like.
  • past frames may be used for portions where the frame rate is low.
  • the higher the frame rate the higher the transfer image block update frequency.
  • the receiving server 3 analyzes the video content and generates synthesis information so as to control the compression rate of the image when the motion is large and the frame rate when the image content is simple. You can
  • the transmission capacity of transmission image blocks output from the first transmission server 2A and the second transmission server 2B is proportional to its frame rate.
  • the frame rate By controlling the frame rate as described above, the total amount of data output from all transmission servers 2 does not exceed the transmission band for one stream of 4K UHD. Therefore, since congestion does not occur in the network 4, the reception server 3 can stably receive a plurality of transmission image blocks.
  • the reception server 3 notifies the transmission server 2 in advance of synthesis information, which is information regarding video synthesis.
  • synthesis information which is information regarding video synthesis.
  • this invention is not limited to the said embodiment.
  • the composition instruction unit 35 sets the composition information to relax the compression flag and frame rate restrictions described above, and sets the data volume to a value that increases.
  • the synthesizing instruction unit 35 evaluates the transmission path from the transmitting server 2 to the receiving server 3, and adjusts the compression flag to be included in the synthesizing information and the lower/upper limit value of the frame rate according to the lowest transmission capacity that becomes a bottleneck. You can set the value.
  • the method described in the above embodiment can be executed by a computer (computer) as a program (software means), such as a magnetic disk (floppy (registered trademark) disk, hard disk, etc.), an optical disk (CD-ROM, DVD , MO, etc.), semiconductor memory (ROM, RAM, flash memory, etc.), etc., or can be transmitted and distributed via a communication medium.
  • the programs stored on the medium also include a setting program for configuring software means (including not only execution programs but also tables and data structures) to be executed by the computer.
  • a computer that realizes this apparatus reads a program stored in a storage medium, and in some cases, constructs software means by a setting program, and executes the above-described processes by controlling the operation of the software means.
  • the storage medium referred to in this specification includes storage media such as magnetic disks, semiconductor memories, etc. provided in computers or devices connected via a network, without being limited to those for distribution.
  • the present invention is not limited to the above embodiments, and can be modified in various ways without departing from the gist of the invention at the implementation stage.
  • each embodiment may be implemented in combination as much as possible, and in that case, the combined effect can be obtained.
  • the above-described embodiments include inventions at various stages, and various inventions can be extracted by appropriately combining a plurality of disclosed constituent elements.

Abstract

A video compositing system according to one embodiment comprises: a compositing instruction unit that generates compositing information indicative of the image blocks an output image is to be composed of; a compositing information transmission unit that transmits the compositing information; an image reception unit that receives a plurality of transmission image blocks which have been generated on the basis of the compositing information; an composite image unit that composites the plurality of transmission image blocks into the output image on the basis of the compositing information; and a display unit that displays the output image.

Description

映像合成システム、映像合成方法、及び映像合成プログラムVideo Synthesis System, Video Synthesis Method, and Video Synthesis Program
 この発明は、映像合成システム、映像合成方法、及び映像合成プログラムに関する。 The present invention relates to a video synthesizing system, a video synthesizing method, and a video synthesizing program.
 映像合成技術として、複数の映像信号を入力とし、当該複数の映像信号を合成することで所望の出力画像を生成する手法が知られている。これらの手法は、一般的に、ライブプロダクションスイッチャーなどの機器が備えるミクサー・キーヤ(Mixer/Keyer)といった機能により実現されている。 As a video synthesis technology, a method is known in which a plurality of video signals are input and a desired output image is generated by synthesizing the plurality of video signals. These methods are generally realized by functions such as mixer/keyer provided in devices such as live production switchers.
 特許文献1は、多地点通信会議システムにおいて、クライアント端末毎に画像データのビットレートを自由に変更できるようにし、クライアントの合成画像と特定の個人画像とを任意に切り替える手段を備えることで、個人画像を高精細に視聴できる技術を開示している。 Japanese Patent Laid-Open No. 2004-100000 discloses a multi-point communication conference system in which a bit rate of image data can be freely changed for each client terminal, and means for arbitrarily switching between a synthesized image of a client and a specific personal image is provided, thereby allowing an individual It discloses a technique for viewing images with high definition.
日本国特許第3936707号Japanese Patent No. 3936707
 特許文献1は、ネットワークを介して接続される1台のサーバ装置と複数台のクライアント端末において、クライアント端末は階層符号化によるスケーラブルビットストリームでサーバ装置に映像伝送するとともに、サーバ装置から返却される映像をクライアント端末から指定できるようにすることで、伝送帯域を節約している。 In Patent Document 1, in one server device and a plurality of client terminals connected via a network, the client terminal transmits video to the server device in a scalable bitstream by hierarchical coding, and is returned from the server device. Transmission bandwidth is saved by enabling the client terminal to specify the video.
 しかしながら、特許文献1は、スケーラブルビットストリームによる制御を前提としているため、ビットレートの低い映像において画質劣化しやすい。また、特許文献1は、サーバ装置による画像合成について、複数映像の同時表示しか想定していないため、最終的に表示される映像上で画質劣化が表出しやすい。さらに、各映像の表示領域が明確に区分けできないミックスなどの映像効果には対応できないという問題がある。 However, since Patent Document 1 assumes control by a scalable bitstream, image quality tends to deteriorate in video with a low bit rate. In addition, since Patent Document 1 assumes only simultaneous display of a plurality of images with respect to image synthesis by a server device, deterioration in image quality tends to appear on the finally displayed image. Furthermore, there is a problem that it is not possible to cope with image effects such as mixing in which the display area of each image cannot be clearly divided.
 この発明の課題は、上記事情に着目してなされてもので、その目的とするところは、映像合成に関する合成情報を事前通知し、送信側のサーバが合成情報に基づいて入力画像の加工及び圧縮を実施することで、伝送帯域が狭い伝送路においても複数映像の高精細な送信及び合成を実現する技術を提供することにある。 SUMMARY OF THE INVENTION The object of the present invention is to address the above-mentioned circumstances, and the object of the present invention is to notify in advance of synthesis information regarding video synthesis, and to allow the server on the transmission side to process and compress the input image based on the synthesis information. to provide a technique for realizing high-definition transmission and synthesis of a plurality of images even on a transmission path with a narrow transmission band.
 上記課題を解決するためにこの発明の一態様に係る映像合成システムは、出力画像がどのような画像ブロックで構成されるかを示す合成情報を生成する合成指示部と、前記合成情報を送信する合成情報送信部と、前記合成情報に基づいて生成された複数の伝送画像ブロックを受信する画像受信部と、前記合成情報に基づいて前記複数の伝送画像ブロックから前記出力画像を合成する合成画像部と、前記出力画像を表示する表示部と、を備える映像合成システム。 In order to solve the above problems, a video composition system according to one aspect of the present invention includes a composition instruction unit that generates composition information indicating what kind of image blocks an output image is composed of, and a composition instruction unit that transmits the composition information. A synthesis information transmission unit, an image reception unit that receives a plurality of transmission image blocks generated based on the synthesis information, and a synthesis image unit that synthesizes the output image from the plurality of transmission image blocks based on the synthesis information. and a display section for displaying the output image.
 この発明の一態様によれば、映像合成に関する合成情報を事前通知し、送信側のサーバが合成情報に基づいて入力画像の加工及び圧縮を実施することで、伝送帯域が狭い伝送路においても複数映像の高精細な送信及び合成を提供することが可能になる。 According to one aspect of the present invention, synthesizing information related to video synthesizing is notified in advance, and the server on the transmitting side processes and compresses the input image based on the synthesizing information. It is possible to provide high-definition transmission and composition of video.
図1は、実施形態に係る映像合成システムの構成例を示す図である。FIG. 1 is a diagram illustrating a configuration example of a video synthesizing system according to an embodiment. 図2は、図1で示した映像合成システムの構成例を示す図である。FIG. 2 is a diagram showing a configuration example of the video synthesizing system shown in FIG. 図3は、実施形態に係る送信サーバとして機能するコンピュータの構成の一例を示す模式図である。FIG. 3 is a schematic diagram illustrating an example of the configuration of a computer functioning as a transmission server according to the embodiment; 図4は、実施形態に係る受信サーバとして機能するコンピュータの構成の一例を示す模式図である。FIG. 4 is a schematic diagram illustrating an example of the configuration of a computer functioning as a receiving server according to the embodiment; 図5は、送信サーバの処理動作の一例を示すフローチャートである。FIG. 5 is a flowchart illustrating an example of processing operations of the transmission server. 図6は、受信サーバの処理動作の一例を示すフローチャートである。FIG. 6 is a flow chart showing an example of the processing operation of the receiving server. 図7は、入力画像ブロックの一例を示した図である。FIG. 7 is a diagram showing an example of an input image block. 図8は、ピクチャインピクチャにおける転送画像ブロックと合成結果の例を示した図である。FIG. 8 is a diagram showing an example of a transfer image block in picture-in-picture and a synthesis result. 図9は、ピクチャインピクチャにおける合成情報の一例を示す図である。FIG. 9 is a diagram illustrating an example of synthesis information in picture-in-picture. 図10は、アルファブレンディングにおける合成情報の一例を示す図である。FIG. 10 is a diagram showing an example of synthesis information in alpha blending. 図11は、アルファブレンディングにおける転送画像ブロックと合成結果の例を示した図である。FIG. 11 is a diagram showing an example of transferred image blocks and synthesis results in alpha blending. 図12は、合成情報アルファブレンディングで画像処理する際の合成情報の一例を示す図である。FIG. 12 is a diagram showing an example of composition information when image processing is performed by composition information alpha blending. 図13は、ディゾルブにおける伝送画像ブロックの伝送タイミングと画像合成を示した図である。FIG. 13 is a diagram showing the transmission timing of transmission image blocks and image synthesis in dissolve. 図14は、ディゾルブにおける合成情報の一例を示す図である。FIG. 14 is a diagram illustrating an example of synthesis information in dissolve.
 以下、図面を参照してこの発明に係る実施形態を説明する。なお、以降、説明済みの要素と同一又は類似の要素には同一又は類似の符号を付し、重複する説明については基本的に省略する。 Hereinafter, embodiments according to the present invention will be described with reference to the drawings. Elements that are the same as or similar to elements that have already been explained are denoted by the same or similar reference numerals, and overlapping explanations are basically omitted.
 (構成) 
 図1は、実施形態に係る映像合成システムの構成例を示す図である。 
 本実施形態に係る映像合成システム1は、第1の拠点Aに配置された第1の送信サーバ2Aと、第2の拠点Bに配置された第2の送信サーバ2Bと、拠点Cに配置された第3の送信サーバ2Cと、拠点Dに配置された第4の送信サーバ2Dと、受信サーバ3と、これらの間で伝送される映像信号等を中継するネットワーク4とで構成される。なお、説明の簡単化のため、区別する必要が無い場合、第1の送信サーバ2A、第2の送信サーバ2B、第3の送信サーバ2C、及び第4の送信サーバ2Dは、単に送信サーバ2と記載する。また、図1に示す例では、複数の送信サーバ2及び単一の受信サーバ3を示しているが、送信サーバ2が1つであっても良いし、受信サーバ3が複数あっても良いのは勿論である。
(composition)
FIG. 1 is a diagram illustrating a configuration example of a video synthesizing system according to an embodiment.
The video synthesizing system 1 according to the present embodiment includes a first transmission server 2A located at a first site A, a second transmission server 2B located at a second site B, and a transmission server 2B located at a site C. a third transmission server 2C, a fourth transmission server 2D located at a site D, a reception server 3, and a network 4 for relaying video signals and the like transmitted between them. For simplicity of explanation, the first transmission server 2A, the second transmission server 2B, the third transmission server 2C, and the fourth transmission server 2D are simply the transmission server 2 when there is no need to distinguish them. and described. Also, although the example shown in FIG. 1 shows a plurality of transmission servers 2 and a single reception server 3, the number of transmission servers 2 may be one, and the number of reception servers 3 may be plural. is of course.
 送信サーバ2は、各拠点に配置される。送信サーバ2は、各拠点で撮影された映像を、ネットワーク4を介して受信サーバ3に送信する。 The transmission server 2 is located at each base. The transmission server 2 transmits the video shot at each base to the reception server 3 via the network 4 .
 受信サーバ3は、任意の場所に設置される。受信サーバ3は、送信サーバ2から受信した映像を合成し、合成した画像を出力することが可能である。 The receiving server 3 is installed at any location. The reception server 3 can synthesize the video received from the transmission server 2 and output the synthesized image.
 図2は、図1で示した映像合成システム1の構成例を示す図である。 
 図2の例では、説明の簡単化のため、第1の送信サーバ2A及び第2の送信サーバ2Bのみを示している。
FIG. 2 is a diagram showing a configuration example of the video synthesizing system 1 shown in FIG.
In the example of FIG. 2, only the first transmission server 2A and the second transmission server 2B are shown for simplification of explanation.
 第1の送信サーバ2Aは、第1の撮像部21Aと、変形部22と、画像送信部23と、合成情報受信部24と、を備える。第2の送信サーバ2Bも第1の送信サーバ2Aと同様の構成を備えるが、図2の例では、簡単化のため、第2の撮像部21Bのみを示してある。また、図2の例では、第1の撮像部21A及び第2の撮像部21Bが第1の送信サーバ2A及び第2の送信サーバ2Bの外部に配置される例を示しているが、第1の撮像部21Aが第1の送信サーバ2Aの内部に配置され、第2の撮像部21Bが第2の送信サーバ2Bの内部に配置されても良いのは勿論である。また、説明の簡単化のため、区別する必要が無い場合、第1の撮像部21A及び第2の撮像部21Bは、単に撮像部21と記載する。 The first transmission server 2A includes a first imaging section 21A, a deformation section 22, an image transmission section 23, and a synthesis information reception section 24. Although the second transmission server 2B has the same configuration as the first transmission server 2A, only the second imaging unit 21B is shown in the example of FIG. 2 for simplification. Further, the example of FIG. 2 shows an example in which the first imaging unit 21A and the second imaging unit 21B are arranged outside the first transmission server 2A and the second transmission server 2B. Of course, the imaging unit 21A may be arranged inside the first transmission server 2A, and the second imaging unit 21B may be arranged inside the second transmission server 2B. For simplification of explanation, the first imaging section 21A and the second imaging section 21B are simply referred to as the imaging section 21 when there is no need to distinguish between them.
 受信サーバ3は、第1の画像受信部31Aと、第2の画像受信部31Bと、画像合成部32と、合成情報送信部33と、表示部34と、合成指示部35と、を備える。図2の例では、表示部34及び合成指示部35が受信サーバ3の外部に配置される例を示しているが、表示部34及び合成指示部35が受信サーバ3内に配置されても良いのは勿論である。また、説明の簡単化のため、区別する必要が無い場合、第1の画像受信部31A及び第2の画像受信部31Bは、単に画像受信部31と記載する。 The receiving server 3 includes a first image receiving section 31A, a second image receiving section 31B, an image synthesizing section 32, a synthesizing information transmitting section 33, a display section 34, and a synthesizing instruction section 35. In the example of FIG. 2, the display unit 34 and the composition instruction unit 35 are arranged outside the receiving server 3, but the display unit 34 and the composition instruction unit 35 may be arranged inside the reception server 3. Of course. For simplicity of explanation, the first image receiving section 31A and the second image receiving section 31B are simply referred to as the image receiving section 31 when there is no need to distinguish them.
 撮像部21は、拠点における映像を撮影する。撮像部21は、撮影した映像の映像フレームの一部から入力画像ブロックを取得し変形部22に出力する。 The imaging unit 21 shoots images at the base. The imaging unit 21 acquires an input image block from a part of the video frame of the captured video and outputs it to the transformation unit 22 .
 変形部22は、撮像部21から受信した入力画像ブロックを後述する合成情報受信部24から受信した合成情報に基づいて変形し、伝送画像ブロックを生成する。ここで、合成情報は、出力画像がどのような画像ブロックで構成されるかを示す。そして、変形部22は、伝送画像ブロックを画像送信部23に出力する。なお、入力画像ブロックから伝送画像ブロックを生成する方法は、後述する。 The transformation unit 22 transforms the input image block received from the imaging unit 21 based on the composition information received from the composition information reception unit 24, which will be described later, to generate the transmission image block. Here, the composition information indicates what kind of image blocks the output image is composed of. The transformation unit 22 then outputs the transmission image block to the image transmission unit 23 . A method for generating transmission image blocks from input image blocks will be described later.
 画像送信部23は、ネットワーク4を介して、伝送画像ブロックを後述する受信サーバ3の第1の画像受信部31Aに送信する。 The image transmission unit 23 transmits the transmission image block to the first image reception unit 31A of the reception server 3, which will be described later, via the network 4.
 合成情報受信部24は、ネットワーク4を介して、受信サーバ3から合成情報を受信し、受信した合成情報を変形部22に出力する。 The combining information receiving unit 24 receives combining information from the receiving server 3 via the network 4 and outputs the received combining information to the transforming unit 22 .
 第2の送信サーバ2Bは、第1の送信サーバ2Aと同様のため、ここでの説明を省略する。なお、第2の送信サーバ2Bの画像送信部23は、ネットワーク4を介して、伝送画像ブロックを後述する受信サーバ3の第2の画像受信部31Bに送信する。 The second transmission server 2B is the same as the first transmission server 2A, so the description is omitted here. Note that the image transmission unit 23 of the second transmission server 2B transmits the transmission image block to the second image reception unit 31B of the reception server 3, which will be described later, via the network 4. FIG.
 次に、受信サーバ3について詳細に説明する。 
 第1の画像受信部31A及び第2の画像受信部31Bはそれぞれ、第1の送信サーバ2A及び第2の送信サーバ2Bから受信した伝送画像ブロックを画像合成部32に出力する。
Next, the receiving server 3 will be described in detail.
The first image reception unit 31A and the second image reception unit 31B output the transmission image blocks received from the first transmission server 2A and the second transmission server 2B to the image composition unit 32, respectively.
 画像合成部32は、後述する合成指示部35から受信した合成情報に基づいて伝送画像ブロックから出力画像を合成し、合成した出力画像を表示部34に出力する。 The image synthesizing unit 32 synthesizes the output image from the transmission image block based on the synthesizing information received from the synthesizing instruction unit 35 to be described later, and outputs the synthesized output image to the display unit 34 .
 合成情報送信部33は、後述する合成指示部35から受信した合成情報を送信サーバ2に送信する。 The combining information transmission unit 33 transmits the combining information received from the combining instruction unit 35 (to be described later) to the transmission server 2 .
 表示部34は、画像合成部32にから受信した出力画像を表示する。 The display unit 34 displays the output image received from the image synthesizing unit 32 .
 合成指示部35は、合成情報を生成する。合成情報は、例えば、受信サーバ3の管理者が手動入力又は設定したパラメータにより生成されたものであっても良いし、合成指示部35がデータメモリ303等に記憶された設定を読込んで自動的に生成したものであっても良い。 The synthesizing instruction unit 35 generates synthesizing information. The synthesis information may be, for example, manually input or generated by parameters set by the administrator of the receiving server 3, or may be automatically generated by the synthesis instruction unit 35 reading settings stored in the data memory 303 or the like. It may be generated in
 ネットワーク4は、映像1ストリームを十分に伝送可能な帯域を安定して確保しているとする。例えば、ネットワーク4は、4K UHD(Ultra High Definition)解像度60fpsの映像であれば12Gbpsを安定して送受信可能な帯域を有するものとする。 It is assumed that the network 4 stably secures a band sufficient for transmitting one video stream. For example, it is assumed that the network 4 has a band capable of stably transmitting and receiving 4K UHD (Ultra High Definition) resolution 60 fps video at 12 Gbps.
 図3は、実施形態に係る送信サーバ2として機能するコンピュータの構成の一例を示す模式図である。 
 送信サーバ2は、図3に示すように、コンピュータデバイスにより構成され、CPU等のプロセッサ201を有する。当該プロセッサ201に対し、プログラムメモリ202と、データメモリ203と、通信インタフェース204と、入出力インタフェース205とが、バス206を介して接続される。
FIG. 3 is a schematic diagram showing an example of the configuration of a computer functioning as the transmission server 2 according to the embodiment.
The transmission server 2, as shown in FIG. 3, is composed of a computer device and has a processor 201 such as a CPU. A program memory 202 , a data memory 203 , a communication interface 204 and an input/output interface 205 are connected to the processor 201 via a bus 206 .
 プログラムメモリ202は、記憶媒体として、例えば、EPROM(Erasable Programmable Read Only Memory)、HDD(Hard Disk Drive)、SSD(Solid State Drive)等の随時書込み及び読出しが可能な不揮発性メモリと、ROM(Read Only Memory)等の不揮発性メモリとを組み合わせて使用することができる。プログラムメモリ202は、各種処理を実行するために必要なプログラムを格納している。すなわち、下記で説明する各種処理は、いずれも、プログラムメモリ202に格納されたプログラムを上記プロセッサ201により読み出して実行することにより実現され得る。 The program memory 202 includes storage media such as EPROM (Erasable Programmable Read Only Memory), HDD (Hard Disk Drive), SSD (Solid State Drive) and other non-volatile memories that can be written and read at any time, and ROM (Read Only Memory) can be used in combination with non-volatile memory. The program memory 202 stores programs necessary for executing various processes. That is, all of the various processes described below can be realized by reading and executing a program stored in the program memory 202 by the processor 201 .
 データメモリ203は、記憶媒体として、例えば、HDD、メモリカード等の随時書込み及び読出しが可能な不揮発性メモリと、RAM(Random Access Memory)等の揮発性メモリとを組み合わせて使用したストレージである。データメモリ203は、プロセッサ201がプログラムを実行して各種処理を行う過程で取得及び生成されたデータを記憶するために用いられる。 The data memory 203 is storage that uses, as a storage medium, a combination of a non-volatile memory that can be written and read at any time, such as an HDD or memory card, and a volatile memory such as a RAM (Random Access Memory). The data memory 203 is used to store data acquired and generated while the processor 201 executes the program and performs various processes.
 通信インタフェース204は、1つ以上の有線又は無線の通信モジュールを含む。例えば、通信インタフェース204は、受信サーバ3を含む他の装置と有線又は無線接続する通信モジュールを含む。さらに、通信インタフェース204は、近距離無線技術を利用して他の装置と無線接続するための無線通信モジュールを含んでも良い。すなわち、通信インタフェース204は、プロセッサ201の制御の下、他の装置等との間で通信を行い、各種情報を送受信することができるものであれば一般的な通信インタフェースで良い。 The communication interface 204 includes one or more wired or wireless communication modules. For example, the communication interface 204 includes a communication module for wired or wireless connection with other devices including the receiving server 3 . Additionally, the communication interface 204 may include a wireless communication module for wirelessly connecting with other devices using short-range wireless technology. That is, the communication interface 204 may be a general communication interface as long as it can communicate with other devices and transmit and receive various information under the control of the processor 201 .
 入出力インタフェース205には、入力部207及び表示部208が接続されている。 An input unit 207 and a display unit 208 are connected to the input/output interface 205 .
 入力部207は、少なくとも1台のカメラ又はビデオカメラを含む。例えば、入力部207に含まれるカメラ又はビデオカメラは、対象となる撮影された撮影画像を撮影し、入出力インタフェース205に当該撮影画像を出力する。また、入力部207は、例えば、表示部208である表示デバイスの表示画面上に配置された、静電方式又は圧力方式を採用した入力検知シートであり、入出力インタフェース205を介して、送信サーバ2の管理者のタッチ位置をプロセッサ201に出力する。 The input unit 207 includes at least one camera or video camera. For example, the camera or video camera included in the input unit 207 captures a target captured image and outputs the captured image to the input/output interface 205 . Further, the input unit 207 is, for example, an input detection sheet that employs an electrostatic method or a pressure method and is arranged on the display screen of a display device that is the display unit 208. Through the input/output interface 205, the transmission server 2 administrator's touch position is output to the processor 201 .
 表示部208は、例えば液晶、有機EL(Electro Luminescence)、等を使用した表示デバイスであり、入出力インタフェース205から入力された信号に応じた画像及びメッセージを表示する。 The display unit 208 is a display device using, for example, liquid crystal, organic EL (Electro Luminescence), etc., and displays images and messages according to signals input from the input/output interface 205 .
 図4は、実施形態に係る受信サーバ3として機能するコンピュータの構成の一例を示す模式図である。 
 受信サーバ3は、図4に示すように、コンピュータデバイスにより構成され、CPU等のプロセッサ301を有する。当該プロセッサ301に対し、プログラムメモリ302と、データメモリ303と、通信インタフェース304と、入出力インタフェース305とが、バス306を介して接続される。
FIG. 4 is a schematic diagram showing an example of the configuration of a computer functioning as the reception server 3 according to the embodiment.
The receiving server 3, as shown in FIG. 4, is composed of a computer device and has a processor 301 such as a CPU. A program memory 302 , a data memory 303 , a communication interface 304 and an input/output interface 305 are connected to the processor 301 via a bus 306 .
 プログラムメモリ302は、記憶媒体として、例えば、EPROM、HDD、SSD等の随時書込み及び読出しが可能な不揮発性メモリと、ROM等の不揮発性メモリとを組み合わせて使用することができる。プログラムメモリ302は、各種処理を実行するために必要なプログラムを格納している。すなわち、下記で説明する各種処理は、いずれも、プログラムメモリ302に格納されたプログラムを上記プロセッサ301により読み出して実行することにより実現され得る。 For the program memory 302, as a storage medium, for example, non-volatile memory such as EPROM, HDD, SSD, etc., which can be written and read at any time, and non-volatile memory such as ROM can be used in combination. The program memory 302 stores programs necessary for executing various processes. That is, all of the various processes described below can be realized by reading and executing a program stored in the program memory 302 by the processor 301 .
 データメモリ303は、記憶媒体として、例えば、HDD、メモリカード等の随時書込み及び読出しが可能な不揮発性メモリと、RAM等の揮発性メモリとを組み合わせて使用したストレージである。データメモリ303は、プロセッサ301がプログラムを実行して各種処理を行う過程で取得及び生成されたデータを記憶するために用いられる。 The data memory 303 is storage that uses, as a storage medium, a combination of a non-volatile memory that can be written and read at any time, such as an HDD or memory card, and a volatile memory such as a RAM. The data memory 303 is used to store data acquired and generated while the processor 301 executes the program and performs various processes.
 通信インタフェース304は、1つ以上の有線又は無線の通信モジュールを含む。例えば、通信インタフェース304は、送信サーバ2を含む他の装置と有線又は無線接続する通信モジュールを含む。さらに、通信インタフェース304は、近距離無線技術を利用して他の装置と無線接続するための無線通信モジュールを含んでも良い。すなわち、通信インタフェース304は、プロセッサ301の制御の下、他の装置等との間で通信を行い、各種情報を送受信することができるものであれば一般的な通信インタフェースで良い。 The communication interface 304 includes one or more wired or wireless communication modules. For example, the communication interface 304 includes a communication module that makes wired or wireless connections with other devices, including the transmission server 2 . Additionally, the communication interface 304 may include a wireless communication module for wirelessly connecting with other devices using short-range wireless technology. That is, the communication interface 304 may be a general communication interface as long as it can communicate with other devices under the control of the processor 301 and transmit and receive various kinds of information.
 入出力インタフェース305には、入力部307及び表示部308が接続されている。 An input unit 307 and a display unit 308 are connected to the input/output interface 305 .
 入力部307は、例えば、表示部308である表示デバイスの表示画面上に配置された、静電方式又は圧力方式を採用した入力検知シートであり、入出力インタフェース305を介して、送信サーバ2の管理者のタッチ位置をプロセッサ301に出力する。 The input unit 307 is, for example, an input detection sheet that employs an electrostatic method or a pressure method and is arranged on the display screen of a display device, which is the display unit 308 . The administrator's touch position is output to the processor 301 .
 表示部308は、例えば液晶、有機EL(Electro Luminescence)、等を使用した表示デバイスであり、入出力インタフェース305から入力された信号に応じた画像及びメッセージを表示する。 The display unit 308 is a display device using, for example, liquid crystal, organic EL (Electro Luminescence), etc., and displays images and messages according to signals input from the input/output interface 305 .
 (動作) 
 図5は、送信サーバ2の処理動作の一例を示すフローチャートであり、図6は、受信サーバ3の処理動作の一例を示すフローチャートである。 
 送信サーバ2のプロセッサ201がプログラムメモリ202に記憶されたプログラムを読み出して実行することにより、図5で示すフローチャートの動作が実現される。同様に、受信サーバ3のプロセッサ301がプログラムメモリ302に記憶されたプログラムを読み出して実行することにより、図6で示すフローチャートの動作が実現される。
(motion)
FIG. 5 is a flow chart showing an example of the processing operation of the transmission server 2, and FIG. 6 is a flow chart showing an example of the processing operation of the reception server 3. As shown in FIG.
When the processor 201 of the transmission server 2 reads and executes the program stored in the program memory 202, the operation of the flowchart shown in FIG. 5 is realized. Similarly, the processor 301 of the receiving server 3 reads and executes the program stored in the program memory 302, thereby realizing the operation of the flowchart shown in FIG.
 ここで、図5及び図6に示す処理動作は、1フレーム毎に行われるものとし、フレームが入力される度に連続・並列的に処理を実行するものとする。  Here, it is assumed that the processing operations shown in FIGS. 5 and 6 are performed for each frame, and are performed continuously and in parallel each time a frame is input.
 動作は、例えば、ライブプロダクションスイッチャーなどで実現される映像効果(Digital Video Effect等)をトリガとして、合成情報を作成する必要がある場合に開始して良い。或いは、動作は、任意のタイミングで開始されてよい。例えば、合成情報を作成する必要が生じた際に開始されてよい。 The operation may be started when it is necessary to create synthesis information, for example, triggered by a video effect (Digital Video Effect, etc.) realized by a live production switcher, etc. Alternatively, the operation may be started at arbitrary timing. For example, it may be initiated when there is a need to create synthetic information.
 最初に図6を参照して、送信サーバ2の処理動作について説明する。 
 合成情報受信部24は、合成情報を受信する(ステップST101)。ここで、合成情報は、受信サーバ3の合成指示部35によって生成されたものである。合成情報受信部24は、受信した合成情報を変形部22に出力する。なお、合成情報に含まれる情報の詳細は、後述する。ここで、合成情報は、ステップST102が処理される前の任意のタイミングで受信して良い。すなわち、予め送信サーバ2に受信し、データメモリ203等に記憶しておいてよい。そして、各部が必要な際にデータメモリ203から読み出して使用して良い。また、ステップST101は、後述するステップST102と並列で処理しても良いのは勿論である。
First, the processing operation of the transmission server 2 will be described with reference to FIG.
Synthesis information receiving section 24 receives synthesis information (step ST101). Here, the combining information is generated by the combining instruction section 35 of the receiving server 3 . The synthesis information receiving section 24 outputs the received synthesis information to the transformation section 22 . Details of the information included in the synthesis information will be described later. Here, the composite information may be received at any timing before step ST102 is processed. That is, it may be received in advance by the transmission server 2 and stored in the data memory 203 or the like. Then, each unit may be read out from the data memory 203 and used when necessary. Further, step ST101 may of course be processed in parallel with step ST102, which will be described later.
 撮像部21は、入力画像ブロックを取得する(ステップST102)。撮像部21は、映像フレームを取得し、取得した映像フレームの一部を入力画像ブロックとして取得する。ここで、映像フレームを一部分ずつ取得できない場合、1フレーム分をバッファすることによって入力画像として取得し、これを分割して入力画像ブロックとしても良い。そして、撮像部21は、入力画像ブロックを変形部22に出力する。 The imaging unit 21 acquires an input image block (step ST102). The imaging unit 21 acquires a video frame and acquires a part of the acquired video frame as an input image block. Here, if the video frame cannot be obtained part by part, one frame may be buffered to obtain an input image, and then divided into input image blocks. The imaging unit 21 then outputs the input image block to the deformation unit 22 .
 図7は、入力画像ブロックの一例を示した図である。 FIG. 7 is a diagram showing an example of an input image block.
 図7の(a)が第1の拠点Aで第1の送信サーバ2Aの撮像部21が取得した入力画像Aであり、(b)は、入力画像ブロックの一例を示した図である。(b)に示すように、入力画像ブロックは、例えば、A-01~A-16として縦4×横4の16ブロックに分割されているものとする。また、図7の(c)が第2の拠点Bで第2の送信サーバ2Bの撮像部21が取得した入力画像Bであり、(c)は、入力画像ブロックの一例を示した図である。(d)も(b)と同様に、入力画像ブロックは、例えば、B-01~B-16として縦4×横4の16ブロックに分割されているものとする。 (a) of FIG. 7 is the input image A acquired by the imaging unit 21 of the first transmission server 2A at the first site A, and (b) is a diagram showing an example of an input image block. As shown in (b), it is assumed that the input image block is divided into 16 blocks of 4 vertical×4 horizontal, such as A-01 to A-16. FIG. 7C is an input image B acquired by the imaging unit 21 of the second transmission server 2B at the second site B, and FIG. 7C is a diagram showing an example of an input image block. . In (d) as well as (b), it is assumed that the input image block is divided into 16 blocks of 4 vertical×4 horizontal, for example, B-01 to B-16.
 変形部22は、入力画像ブロックを伝送画像ブロックに転写する(ステップST103)。変形部22は、合成情報に基づいて入力画像ブロックに含まれる画素を伝送画像ブロックに転写する。また、合成情報の内容に従って入力画像ブロックを変形する必要がある場合、変形部22は、画素の参照及び転写の前後で異なる画素座標を設定しても良い。 The transformation unit 22 transfers the input image block to the transmission image block (step ST103). The transformation unit 22 transfers the pixels included in the input image block to the transmission image block based on the synthesis information. Further, when it is necessary to transform the input image block according to the content of the synthesis information, the transformation unit 22 may set different pixel coordinates before and after the pixel reference and transfer.
 変形部22は、伝送画像ブロックを圧縮する(ステップST104)。変形部22は、合成情報に基づいて伝送画像ブロックを圧縮する。ここで圧縮方法は、一般的な方法で良いため、ここでの詳細な説明を省略する。なお、合成情報に圧縮フラグが無い場合、ステップST104は、スキップしても良い。そして、変形部22は、圧縮された又は圧縮をスキップした伝送画像ブロックを画像送信部23に出力する。 The transformation unit 22 compresses the transmission image block (step ST104). A transformation unit 22 compresses the transmission image block based on the synthesis information. Here, since the compression method may be a general method, detailed description thereof is omitted here. Note that step ST104 may be skipped if there is no compression flag in the combined information. Then, the transformation unit 22 outputs the compressed or skipped transmission image blocks to the image transmission unit 23 .
 画像送信部23は、伝送画像ブロックを受信サーバ3に送信する(ステップST105)。画像送信部23は、伝送画像ブロックを変形部22から受信し、受信した伝送画像を、ネットワーク4を通じて、受信サーバ3に送信する。 The image transmission unit 23 transmits the transmission image block to the receiving server 3 (step ST105). The image transmission unit 23 receives the transmission image block from the transformation unit 22 and transmits the received transmission image to the reception server 3 through the network 4 .
 次に、図7を参照して、受信サーバ3の処理動作を説明する。 
 合成情報送信部33は、合成情報を送信サーバ2に送信する(ステップST201)。合成指示部35は、合成情報送信部33及び画像合成部32に合成情報を出力する。合成情報を受信した合成情報送信部33は、ネットワーク4を通じて、各送信サーバ2に合成情報を送信する。なお、合成情報の送信は、フレームの処理が開始される前に予め行っておいても良い。すなわち、ステップST201の処理は、送信サーバ2でフレームが処理される前又は処理中に送信されていれば任意のタイミングで良い。
Next, the processing operation of the receiving server 3 will be described with reference to FIG.
Synthesis information transmitting section 33 transmits the synthesis information to transmission server 2 (step ST201). The composition instruction unit 35 outputs composition information to the composition information transmission unit 33 and the image composition unit 32 . After receiving the combined information, the combined information transmission unit 33 transmits the combined information to each transmission server 2 through the network 4 . Note that the synthesis information may be transmitted in advance before frame processing is started. In other words, the processing of step ST201 may be performed at any timing as long as the frame is transmitted before or during processing by the transmission server 2 .
 画像受信部31は、伝送画像ブロックを受信する(ステップST202)。例えば、第1の画像受信部31Aは、拠点Aの第1の送信サーバ2Aから、入力画像Aに基づいて生成された伝送画像ブロックを受信し、第2の画像受信部31Bは、拠点Bの第2の送信サーバ2Bから、入力画像Bに基づいて生成された伝送画像ブロックを受信する。第1の画像受信部31A及び第2の画像受信部31Bはそれぞれ、受信した伝送画像ブロックを画像合成部32に出力する。 The image receiving unit 31 receives the transmission image block (step ST202). For example, the first image receiving unit 31A receives the transmission image block generated based on the input image A from the first transmission server 2A at the base A, and the second image receiving unit 31B receives the transmission image block at the base B. A transmission image block generated based on the input image B is received from the second transmission server 2B. The first image receiving section 31A and the second image receiving section 31B output the received transmission image blocks to the image synthesizing section 32, respectively.
 画像合成部32は、伝送画像ブロックが単体かどうか判定する(ステップST203)。例えば、伝送画像ブロックが第1の送信サーバ2Aのみから受信する場合、画像合成部32は、伝送画像ブロックが単体であると判定する。この場合、処理は、ステップST204に進む。一方、伝送画像ブロックが第1の送信サーバ2A及び第2の送信サーバ2Bから受信する場合、伝送画像ブロックが複数であると判定する。この場合、処理は、ステップST205に進む。 The image synthesizing unit 32 determines whether or not the transmission image block is a single block (step ST203). For example, when the transmission image block is received only from the first transmission server 2A, the image synthesizing unit 32 determines that the transmission image block is single. In this case, the process proceeds to step ST204. On the other hand, when transmission image blocks are received from the first transmission server 2A and the second transmission server 2B, it is determined that there are a plurality of transmission image blocks. In this case, the process proceeds to step ST205.
 画像合成部32は、伝送画像ブロックを出力画像に挿入する(ステップST204)。画像合成部32は、伝送画像ブロックを通番通りに整列させることによって出力画像を生成する。 The image synthesizing unit 32 inserts the transmission image block into the output image (step ST204). The image synthesizing unit 32 generates an output image by arranging the transmission image blocks in sequence.
 画像合成部32は、複数の伝送画像ブロックから出力画像を生成する(ステップST205)。画像合成部32は、合成情報に基づいて伝送画像ブロックを結合及び/又はブレンドを行って出力画像を生成する。 The image synthesizing unit 32 generates an output image from a plurality of transmission image blocks (step ST205). The image composition unit 32 combines and/or blends the transmission image blocks based on the composition information to generate an output image.
 画像合成部32は、出力画像を表示部34に出力する(ステップST206)。画像合成部32は、生成した出力画像を表示部34に出力する。表示部34は、出力画像に従った映像フレームを表示することになる。 The image composition unit 32 outputs the output image to the display unit 34 (step ST206). The image synthesizing section 32 outputs the generated output image to the display section 34 . The display unit 34 displays video frames according to the output image.
 (ピクチャインピクチャでの合成パターン例) 
 以下では、一般的にピクチャインピクチャと呼ばれる合成パターンに基づいて、合成画像が「入力画像Bの上に1/2に縮小した入力画像Aを重ねる」である場合について説明する。
(Example of composite pattern in picture-in-picture)
In the following, a description will be given of a case in which the synthesized image is "superimposed on the input image B with the input image A reduced to 1/2" based on a synthesis pattern generally called picture-in-picture.
 図8は、ピクチャインピクチャにおける転送画像ブロックと合成結果(出力画像)の例を示した図である。 
 図8の(a)は、入力画像Aを1/2に縮小したものであり、(b)は、(a)の入力画像Aに対応する入力画像ブロックを示す。(c)は、入力画像Bのうち入力画像Aに重なる部分を削除したものであり、(d)は、(c)の入力画像Bに対応する入力画像ブロックを示す。(e)は、最終的な合成結果の画像、すなわち出力画像を示し、(f)は、(e)に対応する画像ブロックを示す。ここで、R-xx(xxは、画像ブロックの通番である)は、伝送画像ブロックを示す。
FIG. 8 is a diagram showing an example of a transfer image block and a synthesis result (output image) in picture-in-picture.
(a) of FIG. 8 is the input image A reduced to 1/2, and (b) shows an input image block corresponding to the input image A of (a). (c) is obtained by deleting a portion of the input image B that overlaps with the input image A, and (d) shows an input image block corresponding to the input image B in (c). (e) shows the image of the final synthesis result, that is, the output image, and (f) shows the image block corresponding to (e). Here, R-xx (xx is the serial number of the image block) indicates the transmission image block.
 すなわち、(b)に示される転送画像ブロックを第1の送信サーバ2Aが受信サーバ3に送信し、(d)に示される転送画像ブロックを第2の送信サーバ2Bが受信サーバ3にする。そして、受信サーバ3の画像合成部32が、これらの転送画像ブロックを結合し、(e)で示す出力画像を生成する。 That is, the first transmission server 2A transmits the transfer image block shown in (b) to the reception server 3, and the second transmission server 2B sends the transfer image block shown in (d) to the reception server 3. Then, the image synthesizing unit 32 of the receiving server 3 combines these transferred image blocks to generate an output image shown in (e).
 (e)に示すように、出力画像のうち、周辺部に対応する画像ブロックは、入力画像Bが、画像の中央部分に対応する画像ブロックは、入力画像Aが割り当てられる。そして、画像合成部32は、合成情報に基づいて入力画像A及び入力画像Bを結合することにより、出力画像を生成している。 As shown in (e), of the output image, the input image B is assigned to the image block corresponding to the peripheral portion, and the input image A is assigned to the image block corresponding to the central portion of the image. Then, the image synthesizing unit 32 generates an output image by combining the input image A and the input image B based on the synthesizing information.
 図9は、ピクチャインピクチャにおける合成情報の一例を示す図である。
 図9の(a)は、入力画像A及び入力画像Bをどのように変形して伝送画像ブロックを生成するかを示す。例えば、(a)の第1列は、伝送画像ブロックを示し、第2列は、伝送画像ブロックに対してどの入力画像ブロックを使用するかを示し、第3列は、使用する入力画像ブロックをどのように変形するかどうかを示す参照フラグを示している。(b)は、伝送画像ブロックをどのように圧縮するかどうかを示す。(a)に示すように、R-06、R-07、R-10、R-11は、それぞれ入力画像Aの入力画像ブロックが割り当てられ、0.5のスケーリングを行う参照フラグが設定されている。また、R-01~R-05、R-08~R-09、R-12~R-16は、それぞれ入力画像Bの入力画像ブロックが割り当てられ、入力画像Bをコピーする参照フラグが設定されている。すなわち、これらの伝送画像ブロックは、入力画像Bを変形せずそのまま転写して生成されることを示している。
FIG. 9 is a diagram illustrating an example of synthesis information in picture-in-picture.
FIG. 9(a) shows how input image A and input image B are transformed to generate transmission image blocks. For example, the first column in (a) indicates the transmission image block, the second column indicates which input image block to use for the transmission image block, and the third column indicates the input image block to use. It shows reference flags that indicate how to transform or not. (b) indicates how to compress the transmitted image block. As shown in (a), R-06, R-07, R-10, and R-11 are each assigned an input image block of input image A, and a reference flag for scaling 0.5 is set. there is Input image blocks of input image B are assigned to R-01 to R-05, R-08 to R-09, and R-12 to R-16, respectively, and reference flags for copying input image B are set. ing. That is, these transmission image blocks are generated by directly transferring the input image B without transforming it.
 なお、図9で説明した合成情報を用いる場合、出力画像ブロックは、第1の送信サーバ2A又は第2の送信サーバ2Bから受信した転送画像ブロックのいずれかに該当するためブレンドを行う必要はない。そのため画像合成部32は、第1の送信サーバ2A及び第2の送信サーバ2Bから受信した伝送画像ブロックを通番通りに整列し、結合することによって出力画像を生成することができる。 It should be noted that when using the composite information described with reference to FIG. 9, the output image block corresponds to either the transfer image block received from the first transmission server 2A or the second transmission server 2B, so there is no need to perform blending. . Therefore, the image synthesizing unit 32 can generate an output image by arranging and combining the transmission image blocks received from the first transmission server 2A and the second transmission server 2B according to the serial numbers.
 ここで、第1の送信サーバ2A及び第2の送信サーバ2Bがそれぞれ入力画像A及び入力画像Bを加工せずに送信した場合、2ストリーム分の伝送帯域が必要となる。ネットワーク4が1ストリーム分の伝送帯域のみを有する場合、ネットワーク4が輻輳し、受信サーバ3が入力画像を受信することができなくなる、又は劣化した入力画像を受信することになる。しかしながら、本実施形態では、第1の送信サーバ2A及び第2の送信サーバ2Bは、予め出力画像ブロックを構成する画像ブロックに加工した伝送画像ブロックを送信する。その結果、すべての送信サーバ2から出力されるデータ総量は、1ストリーム分の伝送帯域を超過しない。したがって、ネットワーク4に輻輳が生じないため、受信サーバ3は、複数の伝送画像ブロックを安定して受信することができる。 Here, if the first transmission server 2A and the second transmission server 2B transmit the input image A and the input image B without processing, respectively, a transmission band for two streams is required. If the network 4 only has a transmission band for one stream, the network 4 will be congested and the receiving server 3 will either be unable to receive the input image, or will receive a degraded input image. However, in this embodiment, the first transmission server 2A and the second transmission server 2B transmit transmission image blocks processed in advance into image blocks constituting output image blocks. As a result, the total amount of data output from all transmission servers 2 does not exceed the transmission band for one stream. Therefore, since congestion does not occur in the network 4, the reception server 3 can stably receive a plurality of transmission image blocks.
 例えば、一般的な映像合成では、各入力映像を加工せずにそれぞれ受信し、すべてのデータが揃っている状態でブレンド等の処理が実行される。しかし4K UHDの撮像部21が複数(n)台あった場合、映像合成サーバやその前段となる伝送路にはn×12Gbpsの映像データが流れ込むことになる。そのため、特にST2110-20などの非圧縮映像伝送を考える場合、接続台数に応じてマルチストリームを前提とした映像伝送路の設計が必要となり、コストが増大してしまう。 For example, in general video synthesis, each input video is received without any processing, and processing such as blending is executed with all the data in place. However, if there are a plurality (n) of 4K UHD imaging units 21, video data of n×12 Gbps will flow into the video synthesizing server and its preceding transmission line. Therefore, especially when uncompressed video transmission such as ST2110-20 is considered, it is necessary to design a video transmission line on the premise of multi-streaming according to the number of connected devices, which increases the cost.
 本実施形態では、送信サーバ2において、合成情報に基づいて必要なだけの画像ブロックしか送信しないように制御する。そのため、すべての映像伝送路で1ストリーム分の伝送帯域を確保すれば良い。 In this embodiment, the transmission server 2 is controlled to transmit only the required number of image blocks based on the composition information. Therefore, it is sufficient to secure a transmission band for one stream on all video transmission lines.
 (アルファブレンディングにおける合成パターン例) 
 以下の例では、一般的にアルファブレンディングと称される合成パターンに基づいて、合成情報が「入力画像Aを左側、入力画像Bを右側に移動し、重複部分が滑らかにつながるようにブレンドする」である場合の例を説明する。
(Example of composite pattern in alpha blending)
In the example below, based on a synthesis pattern commonly referred to as alpha blending, the synthesis information is "moving input image A to the left and input image B to the right, and blending so that the overlapping portions are smoothly connected." An example in the case of is explained.
 図10は、アルファブレンディングにおける合成情報の一例を示す図である。 
 ここで、各入力画像及び出力画像は、例えば、4K UHD解像度(3840×2160)であって良い。図10のx軸は、ピクセル位置を示し、y軸はアルファ値を示す。例えば、図6で示す転送画像ブロックを使用する場合、1つの転送ブロックの解像度は、960×540となる。
FIG. 10 is a diagram showing an example of synthesis information in alpha blending.
Here, each input image and output image may be, for example, 4K UHD resolution (3840×2160). The x-axis of FIG. 10 indicates pixel position and the y-axis indicates alpha value. For example, when using the transfer image blocks shown in FIG. 6, the resolution of one transfer block is 960×540.
 図10に示す入力画像Aは、左側に960ピクセルだけ移動している。つまり、元のx座標0から960だけ移動している状態を示しており、転送ブロックのうちA-01、A-05、A-09、A-13は、出力画像に利用されない。同様に、入力画像Bは、960ピクセルだけ右側に移動している。つまり元のx座標2880から3840だけ移動している状態を示しており、転送ブロックのうちB-04、B-08、B-12、B-16は、出力画像に利用されない。したがって、これらの転送ブロックは、送信サーバ2から送信されないように制御されることになる。また、重複部分は、出力画像を生成する際にアルファ値によって合成される。ここで、座標(m,n)における入力画像Aの画素値をA(m,n)入力画像Bの画素値をB(m,n)、アルファ値(0~100%の範囲を取る)をα(m,n)としとき、合成画像(出力画像)の画素値R(m,n)は、以下のように計算される。 Input image A shown in FIG. 10 has been moved to the left by 960 pixels. In other words, it shows a state of moving from the original x-coordinate 0 to 960, and A-01, A-05, A-09, and A-13 of the transfer blocks are not used for the output image. Similarly, input image B has been shifted to the right by 960 pixels. In other words, it shows a state of moving from the original x-coordinate of 2880 by 3840, and B-04, B-08, B-12 and B-16 of the transfer blocks are not used for the output image. Therefore, these transfer blocks are controlled so as not to be transmitted from the transmission server 2. FIG. Also, the overlapped portion is synthesized with an alpha value when generating the output image. Here, the pixel value of the input image A at the coordinates (m, n) is A (m, n), the pixel value of the input image B is B (m, n) , and the alpha value (ranging from 0 to 100%) is When α (m, n) , the pixel value R(m, n) of the synthesized image (output image) is calculated as follows.
Figure JPOXMLDOC01-appb-M000001
Figure JPOXMLDOC01-appb-M000001
 図11は、アルファブレンディングにおける転送画像ブロックと合成結果(出力画像)の例を示した図である。 
 図11の(a)は、アルファブレンディングに従った入力画像Aであり、(b)は、転送画像ブロックであり、(c)は、(b)の下部に示した数値に従ったサイズ比に従って圧縮した画像を示す。同様に、(d)は、アルファブレンディングに従った入力画像Bであり、(e)は、転送画像ブロックであり、(f)は、(e)の下部に示した数値に従ったサイズ比に従って圧縮した画像を示す。また、(h)は、最終的な合成結果の画像、すなわち出力画像を示し、(i)は、(h)に対応する画像ブロックを示す。
FIG. 11 is a diagram showing an example of transferred image blocks and synthesis results (output images) in alpha blending.
FIG. 11(a) is the input image A according to alpha blending, (b) is the transferred image block, and (c) is according to the size ratio according to the numerical values shown at the bottom of (b). Shows a compressed image. Similarly, (d) is the input image B according to alpha blending, (e) is the transferred image block, and (f) is according to the size ratio according to the numbers shown at the bottom of (e). Shows a compressed image. Also, (h) indicates the image of the final synthesis result, that is, the output image, and (i) indicates the image block corresponding to (h).
 入力画像Aは、上述したように、左側に移動しているため、画像右端に該当する伝送画像ブロック(R-04、R-08、R-12、R-16)に画素が転写されない。同様に入力画像Bは、画像左端に該当する伝送画像ブロック(R-01、R-05、R-09、R-13)に画素が転写されない。また重複部分である伝送画像ブロック(R-02~R-03、R-06~R-07、R-10~R-11、R-14~R-15)は、入力画像Aと入力画像Bの両方を参照する必要があるため、どちらも伝送画像ブロックに含まれる。ただし、これらの伝送画像ブロックは、合成情報に基づいて圧縮されることになる。 Since the input image A has moved to the left as described above, no pixels are transferred to the transmission image blocks (R-04, R-08, R-12, R-16) corresponding to the right end of the image. Similarly, in the input image B, pixels are not transferred to the transmission image blocks (R-01, R-05, R-09, R-13) corresponding to the left end of the image. Transmission image blocks (R-02 to R-03, R-06 to R-07, R-10 to R-11, R-14 to R-15), which are overlapping portions, are input image A and input image B are included in the transmission image block. However, these transmitted image blocks will be compressed based on the synthesis information.
 図12は、アルファブレンディングで画像処理する際の合成情報の一例を示す図である。 
 図12の(a)は、第1の送信サーバ2Aが参照する合成情報であり、(b)は、第2の送信サーバ2Bが参照する合成情報である。合成情報の第1列は、伝送画像ブロックを示し、第2列は、伝送画像ブロックに対してどの入力画像ブロックを使用するかを示し、第3列は、使用する入力画像ブロックをどのように変形するかどうかを示す参照フラグを示し、第4列は、伝送画像ブロックをどのように圧縮するか(すなわち圧縮後のデータ容量)を示す圧縮フラグを示している。ここで、第1の列のR-xx-yにおけるxxは、画像ブロックの通番、yは、拠点を示す。また、圧縮フラグの値に基づいて送信サーバ2の変形部22が転送画像ブロックのサイズを縮小し、受信サーバ3の画像合成部32が受信した後にサイズを戻す処理を実施する。例えば、圧縮フラグは、アルファブレンディングのアルファ値から算出して良い。例えば、R-02、R-06、R-10、R-14に該当する領域において、図11を参照すると、入力画像Aのアルファ値は平均75%、入力画像Bのアルファ値は平均25%を示しているため、それに相当する値を圧縮フラグとして設定する。
FIG. 12 is a diagram showing an example of synthesis information when image processing is performed using alpha blending.
(a) of FIG. 12 is combined information referred to by the first transmission server 2A, and (b) is combined information referred to by the second transmission server 2B. The first column of the synthesis information indicates the transmission image block, the second column indicates which input image block to use for the transmission image block, and the third column indicates how the input image block to be used. A reference flag indicating whether to transform is shown, and the fourth column shows a compression flag indicating how the transmission image block is compressed (ie, data volume after compression). Here, xx in R-xx-y in the first column indicates the serial number of the image block, and y indicates the location. Also, the transformation unit 22 of the transmission server 2 reduces the size of the transferred image block based on the value of the compression flag, and the image synthesis unit 32 of the reception server 3 performs processing to restore the size after receiving the block. For example, the compression flag may be calculated from the alpha value for alpha blending. For example, in the regions corresponding to R-02, R-06, R-10, and R-14, referring to FIG. 11, the average alpha value of input image A is 75%, and the average alpha value of input image B is 25%. , the value corresponding to it is set as the compression flag.
 本実施例では、合成指示部35は、薄い部分(アルファ値が低い部分)の圧縮率を上げ、濃い部分(アルファ値が高い部分)の圧縮率を下げるように合成情報を生成して良い。すなわち、出力画像として濃く出力される伝送画像ブロックの圧縮率が出力画像として薄く出力される伝送画像ブロックの圧縮率よりも小さくなる。例えば、出力画像として薄く出力される画素が出力画像として濃く出力される画素よりも重要度が低いと考えられる。そのため、薄く出力される画素の圧縮率を上げて画質が劣化したとしても、濃く出力される画素の圧縮及び伸長による画質劣化の影響が抑えられるという効果がある。 In this embodiment, the synthesis instruction unit 35 may generate synthesis information so that the compression rate of light portions (portions with low alpha values) is increased and the compression rate of dark portions (portions with high alpha values) is decreased. That is, the compression rate of a transmission image block that is output dark as an output image is lower than the compression rate of a transmission image block that is output light as an output image. For example, pixels output lightly as an output image are considered to be less important than pixels output darkly as an output image. Therefore, even if the image quality is degraded by increasing the compression rate of pixels that are output lightly, there is an effect that the influence of image quality deterioration due to compression and decompression of pixels that are output dark is suppressed.
 各送信サーバ2は、伝送画像ブロックを合成情報に基づいて圧縮し、そのデータを受信サーバ3に対して送信する。また、受信サーバ3は、伝送画像ブロックを受信した後、合成情報に基づいて伝送画像ブロックを伸長し、伝送画像ブロックから出力画像を生成する。 Each transmission server 2 compresses the transmission image block based on the synthesis information and transmits the data to the reception server 3. After receiving the transmission image block, the receiving server 3 decompresses the transmission image block based on the synthesis information, and generates an output image from the transmission image block.
 伝送画像ブロックは、アルファ値によって圧縮されているため、すべての送信サーバ2から出力されるデータ総量は、4K UHDの1ストリーム分の伝送帯域を超過しない。したがって、ネットワーク4に輻輳が生じないため、受信サーバ3は、複数の伝送画像ブロックを安定して受信することができる。 Since the transmission image blocks are compressed by the alpha value, the total amount of data output from all transmission servers 2 does not exceed the transmission band for one 4K UHD stream. Therefore, since congestion does not occur in the network 4, the reception server 3 can stably receive a plurality of transmission image blocks.
 (ディゾルブにおける合成パターン例) 
 以下では、一般的にディゾルブと呼ばれる合成パターンに基づいて、合成情報が「フレームが遷移するにつれ、入力画像Aから入力画像Bに滑らかに遷移する」である場合の例を説明する。
(Example of composite pattern in dissolve)
Below, based on a synthesis pattern generally called a dissolve, an example in which the synthesis information is "smooth transition from input image A to input image B as frames transition" will be described.
 図13は、ディゾルブにおける伝送画像ブロックの伝送タイミングと画像合成(出力画像)を示した図である。 
 図13で示す例では、各拠点の撮像部21は、60fpsのフレームレートで映像ストリームを出力しているものとする。受信サーバ3の合成指示部35は、ディゾルブの遷移率に従って送信サーバ2から送信するフレームレートを制御するよう、合成情報を生成する。図13(a)は、第1の送信サーバ2Aから送信するフレームを示し、(b)は、第2の送信サーバ2Bから送信するフレームを示し、(c)は、表示部34に表示される出力画像の一例を示している。ここで、(a)及び(b)の■は、伝送画像ブロックを送信することを示し、□は、伝送画像ブロックを送信しないことを示す。図13の例では、遷移率が増える毎に第1の送信サーバ2Aが送信する転送画像ブロックの送信タイミングを疎にし、第2の送信サーバ2Bが送信する転送画像ブロックの伝送タイミングを密にする例を示している。この様にすることで、入力画像Aから段々と入力画像Bへと出力画像の映像が切り替わっていくことになる。
FIG. 13 is a diagram showing transmission timing of transmission image blocks and image synthesis (output image) in dissolve.
In the example shown in FIG. 13, it is assumed that the imaging unit 21 at each site outputs a video stream at a frame rate of 60 fps. The synthesizing instruction unit 35 of the receiving server 3 generates synthesizing information so as to control the frame rate transmitted from the transmitting server 2 according to the dissolve transition rate. 13(a) shows a frame transmitted from the first transmission server 2A, (b) shows a frame transmitted from the second transmission server 2B, and (c) is displayed on the display unit 34. An example of an output image is shown. Here, ▪ in (a) and (b) indicates that the transmission image block is transmitted, and □ indicates that the transmission image block is not transmitted. In the example of FIG. 13, as the transition rate increases, the transmission timings of the transfer image blocks transmitted by the first transmission server 2A become sparse, and the transmission timings of the transfer image blocks transmitted by the second transmission server 2B become denser. shows an example. By doing so, the video of the output image is switched from the input image A to the input image B gradually.
 図14は、ディゾルブにおける合成情報の一例を示す図である。 
 図14は、例えば、時刻00:01:00から00:07:00にかけてディゾルブが指示された場合の合成情報の一例を示した図である。図14の表の第1列は、時刻を示し、第2列は、第1の送信サーバ2Aが送信するフレームレートを示し、第3列は、第2の送信サーバ2Bが送信するフレームレートを示す。図13に示すように、時刻が進むにしたがってブレンドの比率が変化するため、それに従って映像のフレームレートを増減させる。ここで、元のフレームレートを超過しないように設定されるものとする。
FIG. 14 is a diagram illustrating an example of synthesis information in dissolve.
FIG. 14 is a diagram showing an example of synthesis information when a dissolve is instructed from 00:01:00 to 00:07:00, for example. The first column in the table of FIG. 14 indicates the time, the second column indicates the frame rate transmitted by the first transmission server 2A, and the third column indicates the frame rate transmitted by the second transmission server 2B. show. As shown in FIG. 13, the blending ratio changes as time progresses, so the video frame rate is increased or decreased accordingly. Here, it shall be set so as not to exceed the original frame rate.
 本実施形態では、例えば、遷移率50%の時に入力画像を1/2に圧縮するのではなく、フレームレートを制御することで転送画像ブロックの帯域を節約する。例えば、入力画像を圧縮する場合、出力画像の品質が劣化する恐れがある。しかしながら、フレームレートを変更して転送画像ブロックの帯域を節約する場合、出力画像の品質の劣化がない。ここで、送信サーバ2及び受信サーバ3はそれぞれ、ピアツーピア(PTP)などにより同期していて良い。また、フレームレートが低い部分は、過去のフレームを流用して良い。さらに、フレームレートが高くなるほど転送画像ブロックの更新頻度を高くしても良い。 In this embodiment, for example, instead of compressing the input image to 1/2 when the transition rate is 50%, the bandwidth of the transfer image block is saved by controlling the frame rate. For example, when compressing an input image, the quality of the output image may be degraded. However, when changing the frame rate to save the bandwidth of the transferred image blocks, there is no degradation in the quality of the output image. Here, the sending server 2 and the receiving server 3 may be synchronized by peer-to-peer (PTP) or the like. Also, past frames may be used for portions where the frame rate is low. Furthermore, the higher the frame rate, the higher the transfer image block update frequency.
 ディゾルブを用いて画像を遷移させる際、受信サーバ3は、映像内容を解析し、動きが大きいときは画像の圧縮率を、画像内容が単純なときはフレームレートを制御するように合成情報を生成しても良い。 When transitioning images using dissolve, the receiving server 3 analyzes the video content and generates synthesis information so as to control the compression rate of the image when the motion is large and the frame rate when the image content is simple. You can
 このように第1の送信サーバ2Aと第2の送信サーバ2Bから出力される伝送画像ブロックの伝送容量は、そのフレームレートに比例する。上述したようなフレームレートの制御によって、すべての送信サーバ2から出力されるデータ総量は、4K UHDの1ストリーム分の伝送帯域を超過しない。したがって、ネットワーク4に輻輳が生じないため、受信サーバ3は、複数の伝送画像ブロックを安定して受信することができる。 Thus, the transmission capacity of transmission image blocks output from the first transmission server 2A and the second transmission server 2B is proportional to its frame rate. By controlling the frame rate as described above, the total amount of data output from all transmission servers 2 does not exceed the transmission band for one stream of 4K UHD. Therefore, since congestion does not occur in the network 4, the reception server 3 can stably receive a plurality of transmission image blocks.
 (作用効果) 
 実施形態によれば、複数の送信サーバ2と受信サーバ3から構成される通信システム構成において、受信サーバ3は、映像合成に関する情報である合成情報を送信サーバ2に事前に通知する。そして、送信サーバ2が当該合成情報に基づいて入力画像を加工及び圧縮することで、伝送帯域が狭い伝送路においても複数映像の高精細な送信及び合成を実現する技術を提供することができる。
(Effect)
According to the embodiment, in a communication system configuration including a plurality of transmission servers 2 and reception servers 3, the reception server 3 notifies the transmission server 2 in advance of synthesis information, which is information regarding video synthesis. By having the transmission server 2 process and compress the input image based on the synthesis information, it is possible to provide a technique for realizing high-definition transmission and synthesis of multiple images even on a transmission path with a narrow transmission band.
 [他の実施形態]
 なお、この発明は上記実施形態に限定されるものではない。例えば、本実施形態で示した例では、元の映像の1ストリーム分の伝送帯域を超過しないように制御する例を説明した。しかしながら、伝送帯域が十分に確保されている場合、合成指示部35は、合成情報を、上述の圧縮フラグやフレームレートの制約を緩和するように設定し、よりデータ容量が大きくなる値に設定しても良い。すなわち、合成指示部35は、送信サーバ2から受信サーバ3までの伝送路を評価し、ボトルネックとなる最低の伝送容量にあわせて、合成情報に含ませる圧縮フラグやフレームレートの下限値/上限値を設定して良い。
[Other embodiments]
In addition, this invention is not limited to the said embodiment. For example, in the example shown in this embodiment, an example has been described in which control is performed so as not to exceed the transmission band for one stream of the original video. However, when the transmission band is sufficiently secured, the composition instruction unit 35 sets the composition information to relax the compression flag and frame rate restrictions described above, and sets the data volume to a value that increases. can be That is, the synthesizing instruction unit 35 evaluates the transmission path from the transmitting server 2 to the receiving server 3, and adjusts the compression flag to be included in the synthesizing information and the lower/upper limit value of the frame rate according to the lowest transmission capacity that becomes a bottleneck. You can set the value.
 また、前記実施形態に記載した手法は、計算機(コンピュータ)に実行させることができるプログラム(ソフトウェア手段)として、例えば磁気ディスク(フロッピー(登録商標)ディスク、ハードディスク等)、光ディスク(CD-ROM、DVD、MO等)、半導体メモリ(ROM、RAM、フラッシュメモリ等)等の記憶媒体に格納し、また通信媒体により伝送して頒布することもできる。なお、媒体側に格納されるプログラムには、計算機に実行させるソフトウェア手段(実行プログラムのみならずテーブル、データ構造も含む)を計算機内に構成させる設定プログラムをも含む。本装置を実現する計算機は、記憶媒体に記憶されたプログラムを読み込み、また場合により設定プログラムによりソフトウェア手段を構築し、このソフトウェア手段によって動作が制御されることにより上述した処理を実行する。なお、本明細書で言う記憶媒体は、頒布用に限らず、計算機内部或いはネットワークを介して接続される機器に設けられた磁気ディスク、半導体メモリ等の記憶媒体を含むものである。 In addition, the method described in the above embodiment can be executed by a computer (computer) as a program (software means), such as a magnetic disk (floppy (registered trademark) disk, hard disk, etc.), an optical disk (CD-ROM, DVD , MO, etc.), semiconductor memory (ROM, RAM, flash memory, etc.), etc., or can be transmitted and distributed via a communication medium. The programs stored on the medium also include a setting program for configuring software means (including not only execution programs but also tables and data structures) to be executed by the computer. A computer that realizes this apparatus reads a program stored in a storage medium, and in some cases, constructs software means by a setting program, and executes the above-described processes by controlling the operation of the software means. Note that the storage medium referred to in this specification includes storage media such as magnetic disks, semiconductor memories, etc. provided in computers or devices connected via a network, without being limited to those for distribution.
 要するに、この発明は上記実施形態に限定されるものではなく、実施段階ではその要旨を逸脱しない範囲で種々に変形することが可能である。また、各実施形態は可能な限り適宜組み合わせて実施してもよく、その場合組み合わせた効果が得られる。さらに、上記実施形態には種々の段階の発明が含まれており、開示される複数の構成要件における適当な組み合わせにより種々の発明が抽出され得る。 In short, the present invention is not limited to the above embodiments, and can be modified in various ways without departing from the gist of the invention at the implementation stage. Moreover, each embodiment may be implemented in combination as much as possible, and in that case, the combined effect can be obtained. Furthermore, the above-described embodiments include inventions at various stages, and various inventions can be extracted by appropriately combining a plurality of disclosed constituent elements.
 1…映像合成システム
 2…送信サーバ
 21…撮像部
 22…変形部
 23…画像送信部
 24…合成情報受信部
 201…プロセッサ
 202…プログラムメモリ
 203…データメモリ
 204…通信インタフェース
 205…入出力インタフェース
 206…バス
 207…入力部
 208…表示部
 3…受信サーバ
 31…画像受信部
 32…画像合成部
 33…合成情報送信部
 34…表示部
 35…合成指示部
 301…プロセッサ
 302…プログラムメモリ
 303…データメモリ
 304…通信インタフェース
 305…入出力インタフェース
 306…バス
 307…入力部
 308…表示部
 4…ネットワーク
REFERENCE SIGNS LIST 1 video synthesis system 2 transmission server 21 imaging unit 22 transformation unit 23 image transmission unit 24 synthesis information reception unit 201 processor 202 program memory 203 data memory 204 communication interface 205 input/output interface 206 Bus 207 Input unit 208 Display unit 3 Receiving server 31 Image receiving unit 32 Image synthesis unit 33 Synthesis information transmission unit 34 Display unit 35 Synthesis instruction unit 301 Processor 302 Program memory 303 Data memory 304 Communication interface 305 Input/output interface 306 Bus 307 Input unit 308 Display unit 4 Network

Claims (8)

  1.  出力画像がどのような画像ブロックで構成されるかを示す合成情報を生成する合成指示部と、
     前記合成情報を送信する合成情報送信部と、
     前記合成情報に基づいて生成された複数の伝送画像ブロックを受信する画像受信部と、
     前記合成情報に基づいて前記複数の伝送画像ブロックから前記出力画像を合成する合成画像部と、
     前記出力画像を表示する表示部と、
     を備える映像合成システム。
    a synthesizing instruction unit that generates synthesizing information indicating what kind of image blocks an output image is composed of;
    a combination information transmission unit that transmits the combination information;
    an image receiving unit that receives a plurality of transmission image blocks generated based on the synthesis information;
    a composite image unit that composites the output image from the plurality of transmission image blocks based on the composite information;
    a display unit that displays the output image;
    A video compositing system with
  2.  複数の情報処理装置を備え、
     前記複数の情報処理装置はそれぞれ、
     入力画像を取得する撮像部と、
     前記合成情報を受信する合成情報受信部と、
     前記合成情報に基づいて入力画像を変形し、前記出力画像を構成する画像ブロックを備える前記伝送画像ブロックを生成する変形部と、
     前記伝送画像ブロックを送信する画像送信部と、
     を備える、請求項1に記載の映像合成システム。
    Equipped with multiple information processing devices,
    Each of the plurality of information processing devices
    an imaging unit that acquires an input image;
    a combination information receiving unit that receives the combination information;
    a transformation unit that transforms an input image based on the synthesis information to generate the transmission image block that includes image blocks that form the output image;
    an image transmission unit that transmits the transmission image block;
    2. The video composition system of claim 1, comprising:
  3.  前記変形部は、前記合成情報に基づいて、前記入力画像から出力画像に必要な画像ブロックを取得し、前記画像ブロックを圧縮して前記伝送画像ブロックを生成する、請求項2に記載の映像合成システム。 3. The video composition according to claim 2, wherein said transformation unit acquires image blocks necessary for an output image from said input image based on said composition information, and compresses said image blocks to generate said transmission image blocks. system.
  4.  前記複数の伝送画像ブロックのデータ総量は、1ストリーム分の伝送帯域で伝送可能なデータ量である、請求項1乃至3のいずれか1項に記載の映像合成システム。 The video synthesizing system according to any one of claims 1 to 3, wherein the total amount of data of the plurality of transmission image blocks is the amount of data that can be transmitted in a transmission band for one stream.
  5.  前記出力画像は、前記複数の伝送画像ブロックがブレンドされて生成され、前記ブレンドされる伝送画像ブロックのうち、前記出力画像として濃く出力される伝送画像ブロックの圧縮率が前記出力画像として薄く出力される伝送画像ブロックの圧縮率よりも小さい、請求項1乃至4のいずれか1項に記載の映像合成システム。 The output image is generated by blending the plurality of transmission image blocks, and among the transmission image blocks to be blended, a compression ratio of a transmission image block output dark as the output image is output light as the output image. 5. The video synthesizing system according to claim 1, wherein the compression rate of the transmitted image block is smaller than the compression rate of the transmitted image block.
  6.  前記出力画像は、前記複数の伝送画像ブロック全体をブレンドして生成され、前記複数の伝送画像はそれぞれ、フレームレートが変化する、請求項1乃至5のいずれか1項に記載の映像合成システム。 The video synthesizing system according to any one of claims 1 to 5, wherein the output image is generated by blending all of the plurality of transmission image blocks, and each of the plurality of transmission images has a varying frame rate.
  7.  プロセッサを備える映像後上氏システムが実行する映像合成方法であって、
     前記プロセッサが、出力画像がどのような画像ブロックで構成されるかを示す合成情報を生成することと、
     前記プロセッサが、前記合成情報を送信することと、
     前記プロセッサが、前記合成情報に基づいて生成された複数の伝送画像ブロックを受信することと、
     前記プロセッサが、前記合成情報に基づいて前記複数の伝送画像ブロックから前記出力画像を合成することと、
     前記プロセッサが、前記出力画像を表示することと、
     を備える、映像合成方法。
    A video synthesizing method executed by a video post-upper system comprising a processor, comprising:
    the processor generating synthesis information indicating what image blocks an output image is composed of;
    the processor transmitting the combined information;
    the processor receiving a plurality of transmission image blocks generated based on the synthesis information;
    the processor synthesizing the output image from the plurality of transmission image blocks based on the synthesizing information;
    the processor displaying the output image;
    A video synthesis method comprising:
  8.  請求項1乃至6のいずれか1項に記載の映像合成システムが備える各部としてコンピュータを機能させる映像合成プログラム。 A video synthesizing program that causes a computer to function as each unit included in the video synthesizing system according to any one of claims 1 to 6.
PCT/JP2021/047582 2021-12-22 2021-12-22 Video compositing system, video compositing method, and video compositing program WO2023119488A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/JP2021/047582 WO2023119488A1 (en) 2021-12-22 2021-12-22 Video compositing system, video compositing method, and video compositing program

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2021/047582 WO2023119488A1 (en) 2021-12-22 2021-12-22 Video compositing system, video compositing method, and video compositing program

Publications (1)

Publication Number Publication Date
WO2023119488A1 true WO2023119488A1 (en) 2023-06-29

Family

ID=86901605

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2021/047582 WO2023119488A1 (en) 2021-12-22 2021-12-22 Video compositing system, video compositing method, and video compositing program

Country Status (1)

Country Link
WO (1) WO2023119488A1 (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2018530210A (en) * 2015-08-20 2018-10-11 コニンクリーケ・ケイピーエヌ・ナムローゼ・フェンノートシャップ Forming tiled video based on media streams
JP2020025280A (en) * 2014-01-15 2020-02-13 アビジロン コーポレイション Streaming multiple encoded objects encoded using different encoding parameters

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2020025280A (en) * 2014-01-15 2020-02-13 アビジロン コーポレイション Streaming multiple encoded objects encoded using different encoding parameters
JP2018530210A (en) * 2015-08-20 2018-10-11 コニンクリーケ・ケイピーエヌ・ナムローゼ・フェンノートシャップ Forming tiled video based on media streams

Similar Documents

Publication Publication Date Title
US11120677B2 (en) Transcoding mixing and distribution system and method for a video security system
JP5830496B2 (en) Display controller and screen transfer device
CN112204993B (en) Adaptive panoramic video streaming using overlapping partitioned segments
EP3804349B1 (en) Adaptive panoramic video streaming using composite pictures
JP5326234B2 (en) Image transmitting apparatus, image transmitting method, and image transmitting system
US11792463B2 (en) Method of video transmission and display
US10389908B2 (en) Image processing device, image processing method, and program with reduction and enlargement scaling of image data
CA2737842A1 (en) Remote user interface
JP2010130544A (en) Image processing device, receiving device and display device
US10963987B2 (en) Method and apparatus for decoding projection based frame with 360-degree content represented by triangular projection faces packed in triangle-based projection layout
WO2023119488A1 (en) Video compositing system, video compositing method, and video compositing program
JP2006279893A (en) Image processing apparatus and image processing method
US20160142633A1 (en) Capture apparatuses of video images
JP2007057586A (en) Image processing apparatus method therefor, and television receiver
JP2012222399A (en) Video playback system, video playback method, and video playback device and program
JP3828010B2 (en) Image receiving system
KR20200129593A (en) Apparatus and method for distributing of ultra high definition videos using scalers
JP2007201816A (en) Video image display system and video image receiver
US20150091928A1 (en) Image processing device and method thereof
CN113094008B (en) Data display control method and device and computer readable storage medium
JP6067085B2 (en) Screen transfer device
JP2015023466A (en) Distribution system, distribution method and program
JP2021192470A (en) Content distribution system and content distribution method, and program
GB2559396A (en) Method of video transmission and display
JPH10271501A (en) Image processor

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21968925

Country of ref document: EP

Kind code of ref document: A1