WO2011116611A1 - 用于电视会议的视频播放方法 - Google Patents

用于电视会议的视频播放方法 Download PDF

Info

Publication number
WO2011116611A1
WO2011116611A1 PCT/CN2010/079190 CN2010079190W WO2011116611A1 WO 2011116611 A1 WO2011116611 A1 WO 2011116611A1 CN 2010079190 W CN2010079190 W CN 2010079190W WO 2011116611 A1 WO2011116611 A1 WO 2011116611A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
video stream
conference
stream
data
Prior art date
Application number
PCT/CN2010/079190
Other languages
English (en)
French (fr)
Inventor
郑双明
杨伯辉
Original Assignee
中兴通讯股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中兴通讯股份有限公司 filed Critical 中兴通讯股份有限公司
Publication of WO2011116611A1 publication Critical patent/WO2011116611A1/zh

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/141Systems for two-way working between two video terminals, e.g. videophone
    • H04N7/147Communication arrangements, e.g. identifying the communication as a video-communication, intermediate storage of the signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems

Definitions

  • BACKGROUND OF THE INVENTION The present invention relates to the field of video conference communication, and more particularly to a video playback method for a video conference.
  • BACKGROUND At present, a conference television system realizes that a plurality of conference sites are displayed on a display device of each conference site in the form of a multi-picture-like video wall. With the development of conference television technology, the conference television can handle multiple conference venues.
  • the second video (secondary video) related to the current conference sent by a conference terminal may be processed, and the processed main video display data and the auxiliary video display data are respectively sent to each conference terminal for display.
  • the following two methods are usually used: 1.
  • a set of display devices is separately configured to display the auxiliary video.
  • the parties involved in the conference need to additionally configure the display device, which not only increases the cost of the video conference, but also is not convenient for the conference participants to watch.
  • the main video and the auxiliary video are superimposed in different windows.
  • a superimposed display mode is as follows: Main video (first video) As the bottom window, each conference live video is displayed in the form of a video wall, and the secondary video (second video) is superimposed on the upper small window. Displayed above the video. With this display mode, a sub-picture of the main video is necessarily occluded. As shown in FIG.
  • another superimposed display mode is as follows: the bottom window is a secondary video, and the upper small window is a main video. In this display mode, part of the secondary video is blocked by the main video, and the multiple sub-pictures of the main video are displayed smaller, which affects the participants of the conference, and can only be solved by enlarging the screen of the display device. It can be seen from the above analysis that no matter which of the above-mentioned prior art technologies are used to display two channels of video, there is a problem that the device cost is high and the user experience is poor.
  • a main object of the present invention is to provide a video playing method for a video conference, so as to at least solve the problem that the existing video conferencing system displays two channels of video and has high equipment cost and poor user experience.
  • a video playing method for a video conference including: a video conferencing system receives a first video stream and a second video stream; parses the first video stream to obtain first video data, and parses the second The video stream obtains the second video data; the first video data and the second video data are combined and encoded to obtain a third video stream; and the third video stream is sent to the conference terminal.
  • the bandwidth of the first video stream is an arbitrary value smaller than the total bandwidth; the second video stream The bandwidth is any value less than the total bandwidth.
  • combining the first video data with the second video data to obtain the third video stream includes: respectively determining a display area of the first video data and a display area of the second video data; by displaying the display area of the first video data and The display area of the second video data is merged into one display area, and the first video data and the second video data are combined and encoded to obtain a third video stream.
  • the method before receiving the first video stream and the second video stream, the method includes: the video conferencing system allocates the first video stream storage space and the second video stream storage space. Further, before parsing the first video stream to obtain the first video data, and parsing the second video stream to obtain the second video data, the method includes: acquiring the first video stream from the first video stream storage space, and storing the space from the second video stream. Get the second video stream. Further, the first video stream includes a video stream of each conference site. Further, the first video stream storage space includes: a plurality of sub-storage spaces, where the video streams of the respective conference sites are separately stored.
  • parsing the first video stream to obtain the first video data comprises: parsing the video stream of the conference site to obtain the first video data; determining the display area of the first video data comprises: determining a display area of the conference site. Further, parsing the video stream of the conference site to obtain the first video data comprises: parsing the video stream of one or more conference sites to obtain the first video data.
  • the second video stream includes other video streams sent by the conference terminal. Further, the second video stream storage space includes: a plurality of sub-storage spaces, respectively storing other video streams from the respective conference sites.
  • parsing the second video stream to obtain the second video data includes: parsing the conference terminal to send The other video streams obtain the second video data; determining the display area of the second video data includes: determining a display area of the other video streams. Further, parsing the other video streams sent by the conference terminal to obtain the second video data includes: parsing other video streams sent by one or more conference terminals to obtain second video data.
  • the video conference system separately parses and merges the two video streams, and outputs one video stream to the conference terminal, which solves the problem that the existing video conference system outputs two video streams to the conference terminal, resulting in high conference terminal cost.
  • FIG. 1 A problem of poor user experience, thereby effectively utilizing the remaining resources of the video conferencing system, without replacing or adding the display device of the conference terminal, and without changing the display mode and layout of the original conference terminal, on the same layer and multiple screens
  • the window displays two high-definition video images, which satisfies the user's demand for simultaneous display of HD main video and HD auxiliary video.
  • FIG. 1 is a two-way video display effect diagram of a video conference terminal according to the prior art
  • FIG. 2 is a two-way video display effect diagram of another video conference terminal according to the prior art
  • FIG. 4 is a flowchart of a video playing method of a video conference according to an embodiment of the present invention
  • FIG. 5 is a flowchart of a video conference scheduling and WEB processing according to the first embodiment of the present invention
  • 6 is a flow chart of allocating resources according to the multi-point control unit according to the first embodiment of the present invention
  • FIG. 7 is a flowchart of image processing according to the first embodiment of the present invention
  • FIG. 8 is a second embodiment of the present invention.
  • FIG. 9 is a two-way video display effect diagram of a video conference terminal according to the third embodiment of the present invention.
  • FIG. 10 is a two-way manner of a conference-terminal according to Embodiment 4 of the present invention;
  • FIG. 11 is a two-way video display effect diagram of a conference site two terminal according to Embodiment 4 of the present invention;
  • FIG. 12 is a two-way video display effect diagram of a conference site three terminal according to Embodiment 4 of the present invention;
  • FIG. 13 is a two-channel video display effect diagram of a conference site four terminal according to Embodiment 4 of the present invention.
  • FIG. 3 is a schematic diagram of a video conference networking for applying a video playing method for a video conference according to an embodiment of the present invention.
  • the video conference system is connected to multiple conference terminals, one of which serves as the conference initiator, and the conference system holds a video conference. After the conference is successfully held, any conference terminal can provide the first conference to the video conference system. Road video (main video), any conference terminal can also provide a second video (secondary video) to the video conferencing system.
  • the video conferencing system After processing the main video and the auxiliary video, the video conferencing system sends the re-encoded new video stream (including the main video picture and the auxiliary video picture) to each conference terminal for display.
  • a video playing method for a video conference is provided. As shown in FIG. 4, the method includes: Step 4: S402, the video conferencing system receives the first video stream and the second video stream; Step 1 Step S404, parsing the first video stream to obtain the first video data, and parsing the second video stream to obtain the second video data; Step S406, combining the first video data and the second video data to obtain the third video stream; Step S408, sending The third video stream is sent to the conference terminal.
  • the conference terminal is divided into two layers of windows to display the main video and the auxiliary video respectively, because the video conferencing system sends two video streams to the conference terminal, and the conference terminal does not have The function of combining and encoding two video streams, therefore, can only be displayed hierarchically.
  • two video streams (a first video stream and a second video stream) are decoded in a video conference system, and two video pictures are merged into one display picture as a sub-picture.
  • the encoding operation thereby forming a unique video stream (third video stream), is sent to the conference terminal. Therefore, the conference terminal can receive only one video stream and can be in a layer of window.
  • the multipoint control unit of the video conferencing system controls the access unit to receive the first video stream and the second video stream.
  • the first video stream can be considered a primary video stream and the second frequency stream can be considered a library i video stream.
  • the main video stream and the video stream are distinguished by the video ten-based negotiation standard, for example, the main video stream is encoded by the H323 standard protocol, and the auxiliary video stream is encoded by the H239 standard protocol.
  • the access unit performs unified unpacking processing on the first video stream and the second video stream, and removes IP, UDP, and the like headers of the first video stream and the second video stream to obtain a standard video stream.
  • the multipoint control unit of the video conference system controls the image processing unit to decode the first video stream and the second video stream respectively, since the first video stream and the second video stream are encoded by different protocol standards, When decoding, the first video stream and the second video stream are separately decoded according to different protocol standards, and after decoding, two types of basic display data are formed, including the first video data and the second video data.
  • an ES stream is sent to a chip of the image processing unit through the McBsp serial port of the image processing unit, and the chip assumes the decoding task of the ES stream.
  • the H264 header information of each chip is decompressed to obtain a frame stream, and then sent to the decoder to parse out the YUV data.
  • the decoder chip divides it into small units of 16*16 pixels. The area is then scaled to the original size of 1/N according to the number N of the final displayed pictures. For example, if the final composition is four pictures, then each sub-picture size is 1/4 of the whole, and the reduced size is YUV data is transmitted to the encoder chip via the SRIO line for encoding.
  • the foregoing decoding method is a preferred decoding method.
  • the ES stream is not limited to the above method, and other existing video stream parsing methods may be used.
  • the image processing unit under the control of the multipoint control unit of the video conferencing system, the image processing unit combines and decodes the first video data and the second video data to form a third video stream displayed in the same hierarchical window.
  • the merged code is basically the reverse process of the decoding process. After the encoder chip collects all the scaled YUV data, it combines it into a complete YUV data, and then encodes it into a frame of data by the encoder.
  • the encoding chip puts the data on the H264 header to compress it. After that, it is sent to the specified storage space through the McBsp serial port.
  • the foregoing merge coding method is a preferred coding method.
  • the video data combination coding is not limited to the foregoing coding method, and other existing video stream coding methods may also be used.
  • the videoconferencing system transmits the third video stream to each conference terminal through its access unit. After receiving the third video stream, the conference terminal directly plays the third video stream on the display device, and the display form is a video wall form, that is, only one layer of the window, the window screen is divided into multiple sub-pictures, and one sub-picture plays the auxiliary video image.
  • the remaining sub-pictures play the main video image.
  • two video streams encoded by different protocols are parsed and re-encoded to form a new video stream, so that the conference terminal displays the two video screens in the same level window in the form of a video wall.
  • the above method effectively utilizes the multi-point control unit, the image processing unit and the access unit in the existing video conferencing system to combine the two channels of video to form one channel video output.
  • the conference terminal does not need to increase or enlarge the display device.
  • the method displays two high-definition video images in the same layer window on the premise of the display mode and layout of the video wall of the original conference terminal, which satisfies the user's demand for simultaneous display of the HD main video and the HD auxiliary video.
  • the bandwidth can be arbitrarily set according to the user's meeting subscription requirements. ratio. For example, if the user has high requirements on the quality of the main video display, the bandwidth of the primary video stream is increased, and the bandwidth of the secondary video stream is decreased.
  • combining the first video data with the second video data to obtain the third video stream comprises: respectively determining a display area of the first video data and a display area of the second video data; by displaying the display area of the first video data and The display area of the second video data is merged into one display area, and the first video data and the second video data are combined and encoded to obtain a third video stream.
  • the conference organizer can flexibly select the display mode of the two channels of video, for example, playing the main video or the auxiliary video in any sub-picture of the display window, and the size of the main video or the # video play screen can be adjusted.
  • the method may further include: the video conferencing system allocates the first video stream storage space and the second video stream storage space.
  • the video conferencing system allocates the first video stream storage space and the second video stream storage space.
  • the multi-point control unit of the video conference system pre-allocates a corresponding storage space for the main video and the secondary video, so that the received sub-genus
  • the primary video stream and the secondary video stream of different protocols are stored in respective storage spaces, so that the image processing module does not need to distinguish whether the image data in the storage area is a primary video stream or a secondary video stream.
  • the foregoing method can also be used in the video playing method of the video conference provided by the embodiment of the present invention as a supplementary means.
  • the multi-point control unit of the video conference system increases the display mode judgment function, and before starting the video play function, first determines whether the user selects to display the auxiliary video in a multi-screen video wall, and if so, image processing
  • the unit obtains the main video stream and the auxiliary video stream for subsequent processing under the control of the multi-point control unit; if not, the image processing unit only acquires the main video stream for subsequent processing, and the auxiliary video stream does not need to be processed by the image processing unit, directly
  • the second video stream is sent to each conference terminal via the access unit.
  • the image screen displayed by each conference terminal is divided into two layers, one layer being a main video image and the other layer being a secondary video image.
  • the method includes: acquiring the first video stream from the first video stream storage space, and storing the space from the second video stream. Get the second video stream.
  • the multipoint control unit of the video conferencing system controls the image processing unit to acquire a corresponding ES code stream from different storage spaces, and prepares the image processing unit for ES stream analysis.
  • the multi-point control unit determines that the user instructs to display the secondary video in the multi-layer video wall, after the conference reservation passes and is successfully held, the multi-point control unit notifies the image processing unit of the main view.
  • the storage location of the frequency stream and the storage location of the secondary video stream acquires the primary video stream from the storage location of the primary video stream for analysis, and obtains the secondary video stream from the storage location of the secondary video stream for analysis.
  • the main video stream and the auxiliary video stream are obtained by the above preferred method.
  • the video conferencing system can determine which video stream to parse according to the playback mode selected by the user, and does not need to parse all the video streams every time, thereby effectively saving.
  • the first video stream may comprise a video stream of each conference site.
  • the main video may include video streams of video images of the conference sites, and the video streams are encoded by the same video protocol, for example, using the H323 protocol, so that the video of each conference site can be combined. Displayed in a multi-screen video wall in the same layer of windows.
  • the video stream of each conference site is unified as the processing method of the main video stream, so that the main video image is not limited to one video picture, as long as multiple video streams are encoded by the same video protocol, multiple video streams can be regarded as a whole.
  • the main video stream makes the display mode of the conference terminal more diverse and more friendly.
  • the first video stream storage space may include: a plurality of sub-storage spaces, respectively storing video streams of respective conference sites.
  • the storage space of the primary video stream ie, A video stream storage space can also be subdivided into multiple sub-storage spaces to store video streams of each conference site.
  • the storage space of the main video stream is divided into multiple sub-storage spaces, which can provide a more flexible display manner for the user.
  • the image processing unit does not need to The video stream is obtained in the sub-storage space of the video stream corresponding to the conference site, and only the sub-storage space of the video stream of the conference site to be displayed is obtained, and the corresponding video stream is obtained for parsing, thereby further saving system resources.
  • the parsing the first video stream to obtain the first video data may include: parsing the video stream of the conference site to obtain the first video data; determining the display area of the first video data comprises: determining the display area i of the conference site.
  • the first video 3 ⁇ 4 ⁇ 4 includes three conference video 3 ⁇ 4 ⁇ and a second video stream.
  • the access unit of the video conference system performs IP packet and UDP packet processing, a total of four channels are generated.
  • ES stream solved by four chips of the image processing unit of the video conferencing system code.
  • the H264 header information of each chip is decompressed to obtain a frame stream, and then sent to the decoder to parse out the YUV data.
  • the decoder chip divides it into small units of 16*16 pixels.
  • the area is then scaled to 1/4 of the original size (since the final picture is four pictures, each sub-picture size is 1/4 of the whole), and is encoded by the SRIO line to the encoding chip.
  • the processing of the four decoding chips is the same. .
  • the encoder chip is assembled to all the scaled YUV data, the display area of the YUV data of the different conference live video and the display area of the YUV data of the auxiliary video are determined, and then the YUV of each conference live video is determined according to the determined display area. It is synthesized into a complete YUV data, which is then encoded into a frame of data by the encoder.
  • the encoder chip compresses the data into the H264 header and then sends it to the specified storage space through the McBsp serial port.
  • the above decoding and decoding method is a preferred method. In a specific implementation process, the above decoding and encoding methods are not limited, and other existing video stream decoding and encoding methods may be used.
  • the image processing unit acquires the video streams of all the conference sites for parsing, and the formed first video data includes all the conferences.
  • the live video shows the data.
  • the image processing unit only acquires the video of the conference site selected by the user.
  • the stream is parsed, and the formed first video data includes video display data of the conference site selected by the user.
  • the process of combining and encoding the auxiliary video display data it is necessary to determine the display area of the selected conference live video image, thereby satisfying the user's subscription to the specific display layout.
  • the video conferencing system can flexibly process the video stream of each conference site in the main video stream, thereby satisfying the diversity of the user customized display layout.
  • the second video stream includes other video streams sent by one or more conference terminals.
  • the secondary video may include other video transmitted by one or more conference terminals, such as one or more video materials associated with the conference subject.
  • the video stream is encoded in the same video protocol, for example, all encoded with the H239 protocol, so that one or more other video images can be merged into the same layer of multi-view video wall display.
  • the layer window is displayed.
  • a plurality of other video streams are encoded by the same video protocol, a plurality of other video streams can be regarded as a secondary video stream as a whole, so that the display mode of the conference terminal is more diverse and more friendly.
  • the second video stream storage space may also include: multiple sub-storage spaces, and other video streams from multiple conference terminals respectively.
  • the secondary video stream includes multiple video streams encoded by the same protocol, for example, a video stream sent by multiple conference terminals encoded by the same protocol
  • the storage space of the video stream ie, the second video stream storage space
  • the storage space of the secondary video stream is divided into multiple sub-storage spaces, which can provide a more flexible display manner to the user.
  • the user can select to display all other videos, and the image processing unit is from the auxiliary video storage space. All other videos are obtained in the corresponding sub-storage space for parsing.
  • parsing the second video stream to obtain the second video data comprises: parsing the other video streams to obtain the second video data; determining the display area of the second video data comprises: determining a display area of the other video streams.
  • the image processing unit acquires the video streams of all other videos for parsing, and the formed second video data includes all other videos. Display Data.
  • the image processing unit only acquires the video stream of other videos selected by the user for parsing, and forms the second video.
  • the data includes display data for other videos selected by the user.
  • the video conferencing system can flexibly process a plurality of other video streams in the i video stream, thereby satisfying the diversity of the user customized display layout.
  • the following describes a specific implementation manner of displaying two channels of video in a multi-screen window in a conference television system according to a specific embodiment.
  • Embodiment 1 the processing flow of the video conference system is described in detail as follows: Step S I.
  • the WEB page determines whether it is necessary to add a secondary video option in the multi-screen content selection according to the conference capability;
  • the description shows that if the user configures the primary video and the secondary video in the conference template in the HD 720P format, the secondary video is the H239 standard dual-stream protocol, and the secondary video rate is smaller than the primary video rate (that is, the secondary video bandwidth is smaller than the conference). Total bandwidth), then when the conference is held, the multi-point control unit sends a multi-screen content message to the WEB, and WEB adds "secondary video" to the sub-screen content option of the multi-screen, so that if a conference site sends a secondary video, The user can select "secondary video" to watch. Step S2.
  • the multipoint control unit allocates corresponding resources for the dual stream multi-screen conference, and performs resource scheduling in the conference. As shown in Fig. 6, when the conference conforms to the dual stream multi-screen When the appointment is made and after the conference is successfully held, the multi-point control unit allocates resources. In the case of a conference reservation, as with other dual-stream conferences, the multipoint control unit pre-allocates the corresponding storage resources for the main video and the secondary video respectively; after the conference reservation passes and is successfully held, the multi-point processing layer sends the image data source message to the image processing.
  • the unit (using the HDPU board here) informs the storage location where the main video is located, and performs the decoding process of the main video; if the user selects the video to join the multi-picture, the multi-point processing unit sends a message to the image board to inform the auxiliary
  • the storage location of the video is used to perform the decoding process of the secondary video.
  • Step S3 The processing flow of each board when the content of the multi-screen includes the auxiliary video. As shown in FIG. 7, the auxiliary video is the same as the ordinary multi-screen conference.
  • the multi-point processing unit reports the storage resource bus range of the image data to be decoded of the image board (including the storage location where the main video and the auxiliary video are located), and the image processing unit does not have to distinguish whether the image data is
  • the main video or the auxiliary video only needs to take image data from the corresponding position for decoding, and then the image processing unit performs internal processing to encode the main video and the auxiliary video together into a multi-picture image, and puts the image output into the multi-point processing unit.
  • the video stream is transmitted by the access unit to each site terminal, and the actual effect of the dual stream multi-screen is displayed on the display devices of each site.
  • the images of the multi-screen in the conference display the images of the secondary video, and the multiple images can be displayed as the secondary video of the same content, or can be displayed as the secondary video of different content.
  • the number of pictures supported by the two-way video multi-view conference can reach 16, and also supports matching at different rates, mixed lines of various dedicated lines and IP access, satisfying different video bandwidths, different access methods, and different number of pictures. The requirements are handled in a similar way to the above.
  • Embodiment 2 a four-screen video conference is taken as an example, and four flute two-way video (main video and i-video) are combined and processed by the video conference system to display a display effect on the conference terminal.
  • the conference initiator initiates a four-party conference and subscribes to the display screen.
  • the specific display manner is to display the live video (main video) of the conference site 1, the conference site 2, and the conference site 3 in a video wall window, and
  • the video material (auxiliary video) related to the conference theme sent by the venue is displayed.
  • the conference initiator can update the display content of the four screens in the specific situation of the conference, for example, after the video is played, the live video of the conference site 4 is redisplayed.
  • the multi-point control unit of the video conference system determines that the user-customized play mode is a sub-picture display in which the auxiliary video is added to the same-view window of the main video, and the multi-point control unit controls the image processing unit from the main video.
  • the video stream of the site video selected by the user (the plurality of site video streams is used as the main video stream) is obtained in the corresponding sub-storage space of the storage space, and the secondary video stream is obtained from the corresponding sub-storage space of the video storage space, and the main video stream is
  • the secondary video streams are separately decoded and then combined, and a video stream of a layer of window displayed as a multi-picture is generated and output to each conference terminal.
  • Embodiment 3 In this embodiment, a four-screen video conference is taken as an example, and four flute two-way video (main video and i-video) are combined and processed by the video conference system to display a display effect on the conference terminal.
  • the conference initiator initiates a four-party conference, and only three parties participate in the conference, and the display screen is customized.
  • the specific display manner is to display the live video of the conference site 1, the conference site 2, and the conference site 3 in a video wall window. (main video), and display video material (auxiliary video) related to the conference theme sent by the venue.
  • the conference initiator can update the display content of the four screens according to the specific situation of the conference, for example, switching the sub-screen display position of the main video or video or the display size of a certain sub-screen.
  • the specific display effect is shown in Figure 9.
  • the multi-point control unit of the video conference system determines that the user-customized play mode is a sub-picture display in which the auxiliary video is added to the same-view window of the main video, and the multi-point control unit controls the image processing unit from the main video.
  • the video stream of all the site videos is obtained in the corresponding sub-storage space of the storage space (all the site video streams are used as the main video stream), and the secondary video stream is obtained from the corresponding sub-storage space of the secondary video storage space, and the primary video stream and the secondary video stream are obtained.
  • a video stream of a layer of window displayed as a multi-picture is generated and output to each conference terminal.
  • Embodiment 4 In this embodiment, a four-screen video conference is taken as an example, and four flute two-way video (main video and i-video) are combined and processed by the video conference system to display a display effect on the conference terminal.
  • the conference initiator initiates a four-party conference, and provides a different display screen for each conference site.
  • the specific display manner is that the conference screen is provided to the conference site, and the live video of the conference site 2, the conference site 3, and the conference site 4 is displayed.
  • the display screen provided to the conference site 2 is the live video (main video) of the conference site 1, the conference site 3, and the conference site 4, and the display is performed by
  • the video material (auxiliary video) related to the theme of the conference is sent by the venue;
  • the display screen provided to the conference site 3 is the live video (main video) of the conference site 1, the conference site 2 and the conference site 4, and the conference theme transmitted by the conference site 1 Related video material (auxiliary video).
  • the conference initiator can update the display content of each display screen according to the specific situation of the conference, for example, switching the sub-screen display position or a sub-screen of the main video or the auxiliary video of one or all display screens.
  • the multi-point control unit of the video conference system determines that the user-customized play mode is a sub-screen display in which the secondary video is added to the same-view window of the main video.
  • the multi-point control unit controls the image processing unit to obtain the video stream of the remaining site video except the conference site from the corresponding sub-storage space of the main video storage space (the remaining conference video streams are collectively referred to as the main video stream), Obtain a secondary video stream from the corresponding sub-storage space of the secondary video storage space, for the main video
  • the stream and the secondary video stream are respectively decoded and then combined, and a video stream for each video terminal is generated, and the separately generated video stream is output to the corresponding conference terminal.
  • the flow to the conference terminal shows that the resulting conference terminal has high cost and poor user experience, thereby effectively utilizing the remaining resources of the video conference system, without replacing or adding the display device of the conference terminal, and without changing the display of the original conference terminal.
  • two high-definition video images are displayed in the same layer of window, which satisfies the user's demand for simultaneous display of high-definition main video and high-definition auxiliary video.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Telephonic Communication Services (AREA)

Abstract

本发明公开了一种用于电视会议的视频播放方法,包括:电视会议系统接收第一视频流和第二视频流;解析第一视频流得到第一视频数据,解析第二视频流得到第二视频数据;将第一视频数据与第二视频数据合并编码得到第三视频流;发送第三视频流给会议终端。采用本发明提供的技术方案,可以不改变原有会议终端显示方式和布局,实现在同一层多画面视窗显示两路高清视频图像,满足了用户对高清主视频和高清辅视频同时显示的需求。

Description

用于电见会汉的视频番放方法 技术领域 本发明属于电视会议通讯领域, 具体而言, 涉及用于电视会议的视频播 放方法。 背景技术 目前, 会议电视系统实现了多个会场的画面以一种多画面类似电视墙的 形式在各个会场的显示设备上显示, 随着会议电视技术的发展, 会议电视除 了可以处理包括多个会场画面的主视频外, 还可以处理某个会议终端发送的 与当前会议相关的第二路视频(辅视频), 并将处理后的主视频显示数据和辅 视频显示数据分别发送给各个会议终端显示。 各个会议终端在显示两路视频 时, 通常釆用以下两种方法: 一、 单独再配置一套显示设备进行辅视频的显示。 釆用这种方法, 需要 参加会议的各方额外配置显示设备, 不仅增加了电视会议的成本, 也不方便 会议参加者观看。 二、 主视频和辅视频以不同的视窗叠加显示。 如图 1所示, 一种叠加显 示方式为: 主视频 (第一路视频) 作为底层视窗以电视墙的形式显示各个会 议现场视频, 辅视频(第二路视频)作为上层小视窗叠加在主视频之上显示。 釆用这种显示方式, 主视频的一个子画面必然被遮挡。 如图 2所示, 另一种 叠加显示方式为: 底层视窗为辅视频, 上层小视窗为主视频。 釆用这种显示 方式, 辅视频的部分画面被主视频遮挡, 而且, 主视频的多个子画面显示的 更小, 影响会议参加者观看, 只能通过扩大显示设备的屏幕来解决。 由上述分析可知, 无论釆取上述哪种现有技术显示两路视频, 都存在设 备成本高、 用户体验度差的问题。 发明内容 本发明的主要目的在于提供一种用于电视会议的视频播放方法, 以至少 解决现有的电视会议系统显示两路视频存在设备成本高、 用户体验度差的问 题。 根据本发明的一个方面, 提供了一种用于电视会议的视频播放方法, 包 括: 电视会议系统接收第一视频流和第二视频流; 解析第一视频流得到第一 视频数据, 解析第二视频流得到第二视频数据; 将第一视频数据与第二视频 数据合并编码得到第三视频流; 发送第三视频流给会议终端。 进一步地, 在第一视频流的带宽与第二视频流的带宽的总和不超过电视 会议系统支持的总带宽的前提下,第一视频流的带宽为小于总带宽的任意值; 第二视频流的带宽为小于总带宽的任意值。 进一步地, 将第一视频数据与第二视频数据合并编码得到第三视频流包 括: 分别确定第一视频数据的显示区域以及第二视频数据的显示区域; 通过 将第一视频数据的显示区域和第二视频数据的显示区域合并到一个显示区 域, 对第一视频数据和第二视频数据合并编码得到第三视频流。 进一步地, 在接收第一视频流和第二视频流之前, 方法包括: 电视会议 系统分配第一视频流存储空间和第二视频流存储空间。 进一步地, 在解析第一视频流得到第一视频数据, 解析第二视频流得到 第二视频数据之前, 方法包括: 从第一视频流存储空间获取第一视频流, 从 第二视频流存储空间获取第二视频流。 进一步地, 第一视频流包括各个会议现场的视频流。 进一步地, 第一视频流存储空间包括: 多个子存储空间, 分别存储各个 会议现场的视频流。 进一步地, 解析第一视频流得到第一视频数据包括: 解析会议现场的视 频流得到第一视频数据; 确定第一视频数据的显示区域包括: 确定会议现场 的显示区域。 进一步地, 解析会议现场的视频流得到第一视频数据包括: 解析一个或 多个会议现场的视频流得到第一视频数据。 进一步地, 第二视频流包括会议终端发送的其他视频流。 进一步地, 第 二视频流存储空间包括: 多个子存储空间, 分别存储来自各个会议现场的其 他视频流。 进一步地, 解析第二视频流得到第二视频数据包括: 解析会议终端发送 的其他视频流得到第二视频数据; 确定第二视频数据的显示区域包括: 确定 其他视频流的显示区域。 进一步地, 解析会议终端发送的其他视频流得到第二视频数据包括: 解 析一个或多个会议终端发送的其他视频流得到第二视频数据。 通过本发明, 电视会议系统分别对两路视频流解析与合并编码, 输出一 路视频流给会议终端, 解决了现有的电视会议系统输出两路视频流给会议终 端显示, 导致的会议终端成本高, 用户体验度差的问题, 从而有效利用了电 视会议系统的剩余资源, 无需更换或增加会议终端的显示设备, 而且在不改 变原有会议终端显示方式和布局的前提下, 在同一层多画面视窗显示两路高 清视频图像, 满足了用户对高清主视频和高清辅视频同时显示的需求。 附图说明 此处所说明的附图用来提供对本发明的进一步理解, 构成本申请的一部 分, 本发明的示意性实施例及其说明用于解释本发明, 并不构成对本发明的 不当限定。 在附图中: 图 1是根据现有技术的电视会议终端的两路视频显示效果图; 图 2是才艮据现有技术的另一种电视会议终端的两路视频显示效果图; 图 3是根据本发明实施例的电视会议组网示意图; 图 4是根据本发明实施例的电视会议的视频播放方法的流程图; 图 5是根据本发明实施例一的电视会议调度和 WEB处理流程图; 图 6是才艮据本发明实施例一的多点控制单元分配资源的流程图; 图 7是根据本发明实施例一的图像处理流程图; 图 8是才艮据本发明实施例二的电视会议终端的两路视频显示效果图; 图 9是才艮据本发明实施例三的电视会议终端的两路视频显示效果图; 图 10是根据本发明实施例四的会场一终端的两路视频显示效果图; 图 11是根据本发明实施例四的会场二终端的两路视频显示效果图; 图 12是根据本发明实施例四的会场三终端的两路视频显示效果图; 图 13是根据本发明实施例四的会场四终端的两路视频显示效果图。 具体实施方式 下文中将参考附图并结合实施例来详细说明本发明。 需要说明的是, 在 不冲突的情况下, 本申请中的实施例及实施例中的特征可以相互组合。 图 3是适用本发明实施例提供的电视会议的视频播放方法的电视会议组 网示意图。 如图 3所示, 电视会议系统分别连接多个会议终端, 其中一个会 议终端作为会议发起者, 通过会议系统召开电视会议, 在会议召开成功后, 任何一个会议终端可以向电视会议系统提供第一路视频(主视频),任何一个 会议终端也可以向电视会议系统提供第二路视频(辅视频)。 电视会议系统经 过对主视频和辅视频的处理后, 将再编码后形成的新视频流 (包括主视频画 面和辅视频画面) 发送给各会议终端显示。 根据本发明实施例, 提供了一种用于电视会议的视频播放方法, 如图 4 所示, 该方法包括: 步 4聚 S402、 电视会议系统接收第一视频流和第二视频流; 步一骤 S404、 解析第一视频流得到第一视频数据, 解析第二视频流得到第 二视频数据; 步骤 S406、 将第一视频数据与第二视频数据合并编码得到第三视频流; 步骤 S408、 发送第三视频流给会议终端。 现有电视会议技术, 针对两路视频图像, 会议终端之所以会分为两层视 窗分别显示主视频和辅视频, 是由于电视会议系统向会议终端发送了两路视 频流, 而会议终端不具备对两路视频流合并编码的功能, 因此, 只能分层显 示。 本发明实施例提供的上述方法, 在电视会议系统中对两路视频流 (第一 视频流和第二视频流) 进行解码, 并通过将两个视频画面作为子画面合并入 一个显示画面进行再编码操作, 从而形成唯一的视频流 (第三视频流) 发送 给会议终端。 因此, 会议终端只接收了一路视频流就可以在一层视窗中以子 画面的形式同时显示两路视频。 在上述步 4聚 S402 中, 电视会议系统的多点控制单元控制接入单元接收 第一视频流和第二视频流。 可以将第一视频流视为主视频流, 将第二频流视 为库 i视频流。 其中, 主视频流与 视频流以所基于的视频十办议标准区分, 例 如主视频流釆用 H323标准协议编码, 辅视频流釆用 H239标准协议编码。 在具体的实施过程中, 接入单元对第一视频流和第二视频流进行统一拆 包处理, 将第一视频流和第二视频流的 IP、 UDP等包头去除, 得到标准的视 频码流 (例如 H264 视频码流;)。 这些视频码流统称基本码流 (Elementary Stream, 简称 ES ), 例如是经过 H264技术压缩过的一帧一帧的码流。 在上述步骤 S404 中, 电视会议系统的多点控制单元控制图像处理单元 分别对第一视频流和第二视频流解码, 由于第一视频流和第二视频流釆用不 同的协议标准编码, 因此, 在解码时, 要才艮据不同的协议标准对第一视频流 和第二视频流分别解码, 解码后形成两类基本的显示数据, 包括第一视频数 据及第二视频数据。 在具体的实施过程中, 某一 ES流通过图像处理单元的 McBsp串口发送 到图像处理单元的某一芯片, 该芯片就承担了该 ES流的解码任务。 每个芯片拆除 ES流的 H264包头信息解压获得一帧码流,然后送给解码 器解析出 YUV数据, 对于解出的 YUV数据, 解码芯片将其划分成以 16* 16 像素大小为单位的小区域, 然后根据最终显示的画面个数 N, 将解出的 YUV 数据缩放成原大小的 1/N, 例如最后合成的是四画面, 则每个子画面大小是 整体的 1/4, 缩小后的 YUV数据通过 SRIO线路传送给编码芯片进行编码。 需要说明的是, 上述解码方法为一种优选的解码方法,在具体的实施过程中, 不限于釆用上述方法对 ES 流进行解析, 也可以釆用其他现有的视频流解析 方法。 在上述步骤 S406 中, 在电视会议系统的多点控制单元的控制下, 图像 处理单元将解码后形成第一视频数据及第二视频数据合并编码, 形成显示于 在同一层次视窗的第三视频流。 在具体的实施过程中, 合并编码基本是解码过程的反过程, 编码芯片在 收齐到所有缩放后的 YUV数据后, 将其合成为一路完整的 YUV数据, 然后 由编码器编码成一帧数据, 编码芯片将该数据打上 H264 包头进行压缩, 然 后通过 McBsp串口发送到指定的存储空间中去。 需要说明的是, 上述合并编 码方法为一种优选的编码方法, 在具体的实施过程中, 不限于釆用上述编码 方法进行视频数据合并编码, 也可以釆用其他现有的视频流编码方法。 在上述步 4聚 S408 中, 电视会议系统通过其接入单元将第三视频流发送 给各个会议终端。 会议终端接收到第三视频流后直接在显示设备上播放第三 视频流, 其显示形式为电视墙形式, 即只有一层视窗, 视窗画面分为多个子 画面, 其中一个子画面播放辅视频图像, 其余子画面播放主视频图像。 通过上述电视会议的视频播放方法, 将两路不同协议编码的视频流进行 解析和再编码, 形成新的视频流, 使得会议终端将两路视频画面显示以电视 墙的形式显示在同一层次视窗中, 与现有的电视会议的视频播放方式相比, 上述方法有效利用了现有电视会议系统中的多点控制单元、 图像处理单元以 及接入单元对两路视频进行合并处理, 形成一路视频输出至会议终端, 使得 会议终端的无需增加或扩大显示设备。 而且该方法在继 了原有会议终端的 电视墙显示方式和布局的前提下, 在同一层视窗显示两路高清视频图像, 满 足了用户对高清主视频和高清辅视频同时显示的需求。 优选地, 在具体实施过程中, 在保证第一视频流与第二视频流的带宽之 和不超过电视会议系统能够支持的总带宽的前提下, 可以根据用户的会议订 制要求, 任意设置带宽比。 例如, 如果用户对主视频显示质量的要求高, 则 增加主视频流的带宽, 相应减小辅视频流的带宽, 如果用户要求提高辅视频 的显示质量, 则可以为辅视频流增加带宽, 相应减少主视频流的带宽。 通过 对两路视频的带宽设置进行灵活处理, 满足了用户对于某一路视频显示质量 的要求。 优选地,将第一视频数据与第二视频数据合并编码得到第三视频流包括: 分别确定第一视频数据的显示区域以及第二视频数据的显示区域; 通过将第 一视频数据的显示区域和第二视频数据的显示区域合并到一个显示区域, 对 第一视频数据和第二视频数据合并编码得到第三视频流。 由于对主视频数据和辅视频数据进行合并编码的目的是合并编码形成的 数据流在播放时, 可以在同一层视窗的不同显示区域内分别显示两路视频, 因此,在编码前要分别确定主视频数据的显示区域和辅视频数据的显示区域, 并分别将主视频数据和辅视频数据编入相应的显示区域中。 釆用上述的优选合并编码方法, 会议组织者可以灵活的选择两路视频的 显示方式, 例如显示视窗的任意子画面中播放主视频或辅视频, 可以调整主 视频或 #视频播放画面的大小。 优选地, 在接收第一视频流和第二视频流之前, 上述方法还可以包括: 电视会议系统分配第一视频流存储空间和第二视频流存储空间。 在具体的实施过程中,经过电视会议系统的接入单元进行去 IP包和 UDP 包处理之后, 分别得到第一视频流的 ES流和第二视频流的 ES流, 电视会议 系统的多点控制单元控制这些 ES流放置在相应的存储空间。 在具体的实施过程中, 当有用户作为会议发起者进行两路视频电视会议 预约时, 电视会议系统的多点控制单元为主视频和辅视频预先分配相应的存 储空间, 这样接收到的分属不同协议的主视频流和辅视频流就存储在各自的 存储空间, 使得图像处理模块无需区分存储区中的图像数据是主视频流还是 辅视频流。 优选地, 上述方法还可以将现有的两路视频播放方式作为补充手段, 集 合入本发明实施例提供的电视会议的视频播放方法中。在具体的实施过程中, 电视会议系统的多点控制单元增加显示方式判断功能, 在启动视频播放功能 之前, 首先判断用户是否选择在一层多画面电视墙中显示辅视频, 如果是, 图像处理单元在多点控制单元的控制下, 分别获取主视频流和辅视频流进行 后续处理; 如果否, 则图像处理单元只获取主视频流进行后续处理, 辅视频 流无需图像处理单元处理, 直接作为第二路视频流经接入单元发送至各个会 议终端。 在这种处理方式下, 各个会议终端显示的图像画面分为两层, 一层 为主视频图像, 另一层为辅视频图像。 优选地, 在解析第一视频流得到第一视频数据, 解析第二视频流得到第 二视频数据之前, 方法包括: 从第一视频流存储空间获取第一视频流, 从第 二视频流存储空间获取第二视频流。 在具体的实施过程中, 电视会议系统的多点控制单元控制图像处理单元 从不同的存储空间获取相应的 ES码流,为图像处理单元进行 ES流解析做好 准备。 在多点控制单元判断出用户指示在一层多画面电视墙中显示辅视频的情 况下, 当会议预约通过并召开成功后, 多点控制单元通知图像处理单元主视 频流的存储位置以及辅视频流的存储位置, 图像处理单元从主视频流的存储 位置获取主视频流进行解析 ,从辅视频流的存储位置获取辅视频流进行解析。 通过上述优选方法获取主视频流和辅视频流, 电视会议系统可以根据用户选 择的播放方式决定具体对哪一路视频流进行解析工作, 无需每次对所有的视 频流进行解析处理, 有效地节省了系统的资源。 优选地, 第一视频流可以包括各个会议现场的视频流。 结合图 3所示的 组网图, 主视频可以包括各个会议现场视频图像的视频流, 这些视频流以同 一视频协议编码, 例如都釆用 H323协议编码, 这样就可以将各个会议现场 的视频合并入同一层视窗的多画面电视墙中显示。 将各个会议现场的视频流统一作为主视频流的处理方法, 使得主视频图 像不局限于一个视频画面, 只要多个视频流釆用同一种视频协议编码, 即可 将多个视频流整体视为主视频流, 使得会议终端的显示方式更加多样化, 更 具友好性。 优选地, 第一视频流存储空间可以包括: 多个子存储空间, 分别存储各 个会议现场的视频流。 在具体的实施过程中, 如果主视频流包括了多个釆用相同协议编码的视 频流, 例如, 釆用相同协议编码的多个会议现场的视频流则上述主视频流的 存储空间 (即第一视频流存储空间) 还可以细分为多个子存储空间, 分别存 储各个会议现场的视频流。 釆用该优选的方式将主视频流的存储空间划分为 多个子存储空间, 可以向用户提供更加灵活的显示方式, 例如, 用户可选择 某个会议现场的视频画面无需显示, 则图像处理单元无需到与该会议现场对 应的视频流的子存储空间中获取视频流, 只需到需要显示的会议现场的视频 流的子存储空间获取相应的视频流进行解析, 从而进一步地节省了系统的资 源。 优选地, 解析第一视频流得到第一视频数据可以包括: 解析会议现场的 视频流得到第一视频数据; 确定第一视频数据的显示区域包括: 确定会议现 场的显示区 i或。 以三会场四画面的会议为例, 第一视频¾¾包括三个会场的视频 ¾ϊ和一个 第二视频流, 经过电视会议系统的接入单元进行去 IP包和 UDP包处理后, 共生成四路 ES 流, 分别由电视会议系统的图像处理单元的四个芯片进行解 码。 每个芯片拆除 ES流的 H264包头信息解压获得一帧码流, 然后送给解码 器解析出 YUV数据, 对于解出的 YUV数据, 解码芯片将其划分成以 16* 16 像素大小为单位的小区域, 然后缩放成原大小的 1/4 (因最后合成的是四画 面, 每个子画面大小是整体的 1/4 ), 通过 SRIO线路传送给编码芯片进行编 码, 四个解码芯片的处理过程相同。 编码芯片在收齐到所有缩放后的 YUV数据后, 确定不同会议现场视频 的 YUV数据的显示区域以及辅视频的 YUV数据的显示区域,之后根据确定 好的显示区域将其各个会议现场视频的 YUV合成为一路完整的 YUV数据, 然后由编码器编码成一帧数据, 编码芯片将该数据打上 H264包头进行压缩, 然后通过 McBsp串口发送到指定的存储空间中去。 需要说明的是, 上述解码和解码的方法为优选方法, 在具体的实施过程 中, 不限于釆用上述解码及编码方法, 也可以釆用其他现有的视频流解码和 编码方法。 优选地, 在具体实施过程中, 如果用户选择在多画面电视墙中显示所有 会议现场的视频画面,则图像处理单元获取所有会议现场的视频流进行解析, 形成的第一视频数据包括了所有会议现场的视频显示数据。 并且在与辅视频 显示数据合并编码的过程中, 需要确定每个会议现场视频图像的显示区域, 从而满足用户对具体显示布局的订制。 优选地, 在具体实施过程中, 如果用户选择在多画面电视墙中只显示某 一个会议现场的视频画面或某几个特定的会议现场, 则图像处理单元只获取 用户选定的会议现场的视频流进行解析, 形成的第一视频数据包括了用户选 定的会议现场的视频显示数据。并且在与辅视频显示数据合并编码的过程中, 需要确定选定的会议现场视频图像的显示区域, 从而满足用户对具体显示布 局的订制。 通过上述优选的对主视频流的解析方法和优选的合并编码方法, 可以使 电视会议系统对主视频流中的各个会议现场的视频流灵活处理, 从而满足了 用户订制显示布局的多样性。 优选地, 第二视频流包括一个或多个会议终端发送的其他视频流。 结合 图 3所示的组网图, 辅视频 (第二视频流) 可以包括一个或多个会议终端发 送的其他视频, 例如与会议主体相关的一个或多个影像资料。 这些其他视频 的视频流以同一视频协议编码, 例如都釆用 H239协议编码, 这样就可以将 一个或多个其他视频图像合并入同一层多画面电视墙显示。 将一个或多个釆用同一视频协议编码的其他视频流统一作为辅视频流的 处理方法, 使得辅视频图像不局限于一个视频画面, 并且可以经过电视会议 系统处理, 与主视频图像合并入一层视窗显示。 只要多个其他视频流釆用同 一种视频协议编码, 即可将多个其他视频流整体视为辅视频流, 使得会议终 端的显示方式更加多样化, 更具友好性。 优选地, 第二视频流存储空间也可以包括: 多个子存储空间, 分别来自 多个会议终端的其他视频流。 在具体的实施过程中, 如果辅视频流 (第二视频流) 包括了多个釆用相 同协议编码的视频流, 例如, 釆用相同协议编码的多个会议终端发送的视频 流, 则上述辅视频流的存储空间 (即第二视频流存储空间) 还可以细分为多 个子存储空间, 分别存储来至多个会议终端的其他视频流。 釆用该优选的方 式将辅视频流的存储空间划分为多个子存储空间, 可以向用户提供更加灵活 的显示方式, 例如, 用户可选择显示所有其他视频, 则图像处理单元从辅视 频存储空间的相应子存储空间中获取所有其他视频进行解析处理, 如果用户 选择只显示其中的某些其他视频, 则图像处理单元只需到需要显示的其他视 频的子存储空间获取相应的视频流进行解析, 从而进一步地节省了系统的资 源。 优选地, 解析第二视频流得到第二视频数据包括: 解析其他视频流得到 第二视频数据; 确定第二视频数据的显示区域包括: 确定其他视频流的显示 区域。 优选地, 在具体实施过程中, 如果用户选择在多画面电视墙中显示所有 其他视频画面, 则图像处理单元获取所有其他视频的视频流进行解析, 形成 的第二视频数据包括了所有会其他视频显示数据。 并且在与主视频显示数据 合并编码的过程中, 需要确定每个其他频图像的显示区域, 从而满足用户对 具体显示布局的订制。 优选地, 在具体实施过程中, 如果用户选择在多画面电视墙中只显示某 些特定的其他视频, 则图像处理单元只获取用户选定的其他视频的视频流进 行解析, 形成的第二视频数据包括了用户选定的其他视频的显示数据。 并且 在与主视频显示数据合并编码的过程中, 需要确定选定的其他视频的显示区 域, 从而满足用户对具体显示布局的订制。 通过上述优选的对辅视频的解析方法和优选的合并编码方法, 可以使电 视会议系统对 i视频流中多个其他视频流灵活处理, 从而满足了用户订制显 示布局的多样性。 以下结合具体实施例介绍在会议电视系统中实现一层多画面视窗中显示 两路视频的具体实施方式。 实施例一 在本实施例中, 详细介绍了电视会议系统处理流程, 具体如下: 步骤 S I. WEB 页面根据会议能力判断是否需要在多画面内容选择中加 入辅视频的选项; 如图 5所示, 图中所描述的是如果用户在会议模板中配置主视频和辅视 频的格式均为高清 720P格式, 辅视频为 H239标准双流协议, 且辅视频速率 小于主视频速率 (即辅视频带宽小于会议总带宽), 那么在召开会议的时候, 多点控制单元发送多画面内容消息给 WEB , WEB在多画面的子画面内容选 项中加入 "辅视频,,, 这样与会某一会场若发送辅视频, 用户就可以选择 "辅 视频" 进行观看。 步骤 S2. 多点控制单元为双流多画面的会议分配相应的资源, 并且在会 议中进行资源调度。 如图 6所示, 当符合双流多画面的会议预约时以及会议召开成功后, 多 点控制单元分配资源的情况。 当有用户进行双流多画面会议预约时, 与其他 双流会议一样,多点控制单元分别为主视频和辅视频预先分配相应存储资源; 当会议预约通过并召开成功后, 多点处理层下发图像数据来源消息给图像处 理单元 (这里使用 HDPU板), 告知其主视频所在的存储位置, 由其进行主 视频的解码处理; 若用户选择 视频加入多画面中, 多点处理单元则下发消 息给图像板, 告知其辅视频的存储位置, 由其进行辅视频的解码处理。 步骤 S3、 在选择多画面的内容包括辅视频时, 各个单板的处理流程。 如图 7所示, 与普通多画面会议一样, 辅视频占用一定的系统资源, 当 用户选择辅视频加入多画面时, 多点处理单元告之图像板待解码图像数据的 存储资源总线范围(包括主视频和辅视频的所在的存储位置), 图像处理单元 则不必区分图像数据是否为主视频或辅视频, 只须从相应的位置取图像数据 进行解码即可, 然后图像处理单元经过内部处理, 将主视频和辅视频一起编 码合成多画面图像, 放入多点处理单元的图像输出总线上, 由接入单元传送 视频码流给各会场终端, 在各会场的显示设备上显示出双流多画面的实际效 果。(若用户将所有多画面的内容都选择为辅视频, 则会议中多画面的图像均 显示辅视频的图像, 多个画面可以显示为同一内容的辅视频, 也可以显示为 不同内容的辅视频)。 此夕卜, 双路视频多画面会议支持的画面数可达到 16, 同时也支持不同速 率的匹配, 各种专线和 IP接入的混网, 满足不同视频带宽、 不同接入方式、 不同画面数的需求, 处理方式与上述类似。 实施例二 在本实施例中, 以四画面电视会议为例, 4笛述两路视频 (主视频和 i视 频) 经过电视会议系统合并处理后, 在会议终端上的一种显示效果。 在本实施例中, 会议发起者发起四方会议, 并订制显示画面, 具体的显 示方式为, 在一层电视墙视窗中显示会场一、 会场二以及会场三的现场视频 (主视频), 以及显示由会场一发送的与会议主题相关的视频资料(辅视频)。 在会议进行的过程中, 会议发起者可以 居会议进行的具体情况, 更新四画 面的显示内容, 例如在 视频播放结束后, 重新显示会场四的现场视频。 具 体显示效果如图 8所示。 为达到上述显示效果, 电视会议系统的多点控制单元判断用户订制的播 放方式是将辅视频加入到主视频的同层视窗中的子画面显示, 多点控制单元 控制图像处理单元从主视频存储空间的相应子存储空间中获取用户选择显示 的会场视频的视频流 (多个会场视频流作为主视频流 ) ,从 视频存储空间的 相应子存储空间中获取辅视频流, 对主视频流和辅视频流分别解码后再合并 编码, 生成显示为多画面的一层视窗的视频流输出给各个会议终端。 实施例三 在本实施例中, 以四画面电视会议为例, 4笛述两路视频 (主视频和 i视 频) 经过电视会议系统合并处理后, 在会议终端上的一种显示效果。 在本实施例中, 会议发起者发起四方会议, 参加会议的只有三方, 则订 制显示画面, 具体的显示方式为, 在一层电视墙视窗中显示会场一、 会场二 以及会场三的现场视频(主视频), 以及显示由会场一发送的与会议主题相关 的视频资料(辅视频)。 在会议进行的过程中, 会议发起者可以才艮据会议进行 的具体情况, 更新四画面的显示内容, 例如切换主视频或 视频的子画面显 示位置或某个子画面的显示大小。 具体显示效果如图 9所示。 为达到上述显示效果, 电视会议系统的多点控制单元判断用户订制的播 放方式是将辅视频加入到主视频的同层视窗中的子画面显示, 多点控制单元 控制图像处理单元从主视频存储空间的相应子存储空间中获取所有会场视频 的视频流 (所有会场视频流作为主视频流 ),从辅视频存储空间的相应子存储 空间中获取辅视频流, 对主视频流和辅视频流分别解码后再合并编码, 生成 显示为多画面的一层视窗的视频流输出给各个会议终端。 实施例四 在本实施例中, 以四画面电视会议为例, 4笛述两路视频 (主视频和 i视 频) 经过电视会议系统合并处理后, 在会议终端上的一种显示效果。 在本实施例中, 会议发起者发起四方会议, 针对每个会场提供不同的显 示画面, 具体的显示方式为, 提供给会场一的显示画面, 显示会场二、 会场 三以及会场四的现场视频(主视频), 以及显示由会场一发送的与会议主题相 关的视频资料(辅视频); 提供给会场二的显示画面为会场一、 会场三以及会 场四的现场视频(主视频), 以及显示由会场一发送的与会议主题相关的视频 资料(辅视频); 提供给会场三的显示画面为会场一、 会场二以及会场四的现 场视频(主视频), 以及显示由会场一发送的与会议主题相关的视频资料(辅 视频)。 在会议进行的过程中, 会议发起者可以根据会议进行的具体情况, 分 别更新各个显示画面的显示内容, 例如切换某个或所有显示画面的主视频或 辅视频的子画面显示位置或某个子画面的显示大小。 四个会议现场的具体显 示效果如图 10 -图 13所示。 为达到上述显示效果, 电视会议系统的多点控制单元判断用户订制的播 放方式是将辅视频加入到主视频的同层视窗中的子画面显示。 针对每一个会 议终端, 多点控制单元控制图像处理单元从主视频存储空间的相应子存储空 间中获取除该会议现场之外的其余会场视频的视频流 (其余会场视频流统称 主视频流),从辅视频存储空间的相应子存储空间中获取辅视频流,对主视频 流和辅视频流分别解码后再合并编码,分别生成针对各个视频终端的视频流, 并将分别生成的视频流输出给相应的会议终端。 从以上的描述中, 可以看出, 本发明实施例提供的电视会议系统分别对 两路视频流解析与合并编码, 输出一路视频流给会议终端, 解决了现有的电 视会议系统输出两路视频流给会议终端显示, 导致的会议终端成本高, 用户 体验度差的问题, 从而有效利用了电视会议系统的剩余资源, 无需更换或增 加会议终端的显示设备, 而且在不改变原有会议终端显示方式和布局的前提 下, 在同一层视窗显示两路高清视频图像, 满足了用户对高清主视频和高清 辅视频同时显示的需求。 以上所述仅为本发明的优选实施例而已, 并不用于限制本发明, 对于本 领域的技术人员来说, 本发明可以有各种更改和变化。 凡在本发明的 ^"神和 原则之内, 所作的任何修改、 等同替换、 改进等, 均应包含在本发明的保护 范围之内。

Claims

权 利 要 求 书 一种用于电视会议的视频播放方法, 其特征在于, 包括:
电视会议系统接收第一视频流和第二视频流;
解析所述第一视频流得到第一视频数据, 解析所述第二视频流得到 第二视频数据;
将所述第一视频数据与所述第二视频数据合并编码得到第三视频 流;
发送所述第三视频流给会议终端。 根据权利要求 1所述的方法, 其特征在于, 在所述第一视频流的带宽与 所述第二视频流的带宽的总和不超过所述电视会议系统支持的总带宽的 前提下,
所述第一视频流的带宽为小于所述总带宽的任意值;
所述第二视频流的带宽为小于所述总带宽的任意值。 才艮据权利要求 1所述的方法, 其特征在于, 将所述第一视频数据与所述 第二视频数据合并编码得到第三视频流的步骤包括:
分别确定所述第一视频数据的显示区域以及第二视频数据的显示区 域;
通过将所述第一视频数据的显示区域和第二视频数据的显示区域合 并到一个显示区域, 对所述第一视频数据和第二视频数据合并编码得到 所述第三视频流。 才艮据权利要求 3所述的方法, 其特征在于, 在接收所述第一视频流和所 述第二视频流之前, 所述方法还包括: 所述电视会议系统分配第一视频 流存储空间和第二视频流存储空间。 才艮据权利要求 4所述的方法, 其特征在于, 在解析所述第一视频流得到 第一视频数据, 解析所述第二视频流得到第二视频数据之前, 所述方法 还包括: 从所述第一视频流存储空间获取所述第一视频流, 从所述第二 视频流存储空间获取所述第二视频流。
6. 根据权利要求 1至 5中任一项所述的方法, 其特征在于, 所述第一视频 流包括各个会议现场的视频流。
7. 根据权利要求 6所述的方法, 其特征在于, 所述第一视频流存储空间包 括: 多个子存储空间, 分别存储各个会议现场的视频流。
8. 根据权利要求 6所述的方法, 其特征在于,
解析所述第一视频流得到第一视频数据包括: 解析会议现场的视频 流得到所述第一视频数据;
确定所述第一视频数据的显示区域包括: 确定所述会议现场的显示 区域。
9. 根据权利要求 8所述的方法, 其特征在于, 解析会议现场的视频流得到 所述第一视频数据包括: 解析一个或多个会议现场的视频流得到所述第 一视频数据。
10. 根据权利要求 1至 5中任一项所述的方法, 其特征在于, 所述第二视频 流包括一个或多个会议终端发送的其他视频流。
11. 根据权利要求 10所述的方法, 其特征在于, 所述第二视频流存储空间包 括: 多个子存储空间, 分别存储来自各个会议现场的所述其他视频流。
12. 根据权利要求 10所述的方法, 其特征在于,
解析所述第二视频流得到第二视频数据包括: 解析会议终端发送的 其他视频流得到所述第二视频数据;
确定所述第二视频数据的显示区域包括: 确定所述其他视频流的显 示区域。
13. 居权利要求 12所述的方法, 其特征在于, 解析会议终端发送的其他视 频流得到所述第二视频数据包括: 解析一个或多个会议终端发送的其他 视频流得到所述第二视频数据。
PCT/CN2010/079190 2010-03-22 2010-11-26 用于电视会议的视频播放方法 WO2011116611A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201010149661.2 2010-03-22
CN201010149661A CN101820524A (zh) 2010-03-22 2010-03-22 用于电视会议的视频播放方法

Publications (1)

Publication Number Publication Date
WO2011116611A1 true WO2011116611A1 (zh) 2011-09-29

Family

ID=42655437

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2010/079190 WO2011116611A1 (zh) 2010-03-22 2010-11-26 用于电视会议的视频播放方法

Country Status (2)

Country Link
CN (1) CN101820524A (zh)
WO (1) WO2011116611A1 (zh)

Families Citing this family (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101820524A (zh) * 2010-03-22 2010-09-01 中兴通讯股份有限公司 用于电视会议的视频播放方法
CN102457701B (zh) * 2010-11-02 2014-03-12 华为终端有限公司 图像显示处理方法及装置
CN102118602B (zh) * 2011-03-15 2013-08-21 深圳市捷视飞通科技有限公司 一种在多画面中显示辅流视频的方法及系统
CN103686219B (zh) * 2012-09-24 2017-09-29 华为技术有限公司 一种视频会议录播的方法、设备及系统
CN103345373A (zh) * 2013-06-21 2013-10-09 北京小米科技有限责任公司 内容展示方法、内容提供方法、装置和设备
CN104539872A (zh) * 2014-12-03 2015-04-22 宁波Gqy视讯股份有限公司 会议终端
CN106341622B (zh) * 2015-07-06 2020-01-24 阿里巴巴集团控股有限公司 一种多路视频流的编码方法及装置
CN106357415B (zh) * 2015-07-16 2019-08-23 华为技术有限公司 抗丢包处理方法和装置
CN106817615B (zh) * 2015-12-02 2019-10-01 南宁富桂精密工业有限公司 多路信号处理方法及应用该方法的流媒体装置
CN106131564A (zh) * 2016-07-28 2016-11-16 广东西奥物联网科技股份有限公司 一种对摄像机视频流进行处理的方法及装置
CN106935180A (zh) * 2017-05-17 2017-07-07 南京巨鲨显示科技有限公司 一种控制矩阵从而控制显示器快捷智能开窗的实现方法
CN109240677B (zh) * 2017-07-07 2020-10-16 阿里巴巴集团控股有限公司 一种图层处理方法及装置
CN110324563A (zh) * 2018-03-29 2019-10-11 北京视联动力国际信息技术有限公司 一种视频会议的主辅源切换方法和系统
CN110536077B (zh) * 2018-05-25 2020-12-25 杭州海康威视系统技术有限公司 一种视频合成和播放方法、装置及设备
CN110740286A (zh) * 2018-07-18 2020-01-31 中兴通讯股份有限公司 一种视频会议控制方法、多点控制单元及视频会议终端
CN109756695A (zh) * 2018-12-25 2019-05-14 视联动力信息技术股份有限公司 一种会议视频流显示的方法和装置
CN110401820A (zh) * 2019-08-15 2019-11-01 北京迈格威科技有限公司 多路视频处理方法、装置、介质及电子设备
CN111787267A (zh) * 2020-07-01 2020-10-16 广州科天视畅信息科技有限公司 会议视频字幕合成系统和方法
CN112153412B (zh) * 2020-08-20 2022-10-21 深圳市捷视飞通科技股份有限公司 视频图像切换的控制方法、装置、计算机设备和存储介质
TWI772933B (zh) * 2020-10-23 2022-08-01 宏正自動科技股份有限公司 影像處理裝置及多畫面顯示的影像處理方法
CN112383738B (zh) * 2020-11-11 2023-03-03 浙江讯盟科技有限公司 一种省流量及资源消耗低的多人音视频会议方法和系统
US11582422B2 (en) * 2021-02-24 2023-02-14 Gn Audio A/S Conference device with multi-videostream capability
CN115988258B (zh) * 2023-03-17 2023-06-23 广州佰锐网络科技有限公司 一种基于IoT设备的视频通信方法、存储介质及系统

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN2509797Y (zh) * 2001-09-29 2002-09-04 张健 具有画面分割功能的图像编码器
CN1422078A (zh) * 2001-11-22 2003-06-04 丽台科技股份有限公司 视频会议系统
CN1571508A (zh) * 2003-07-19 2005-01-26 华为技术有限公司 一种实现多画面的方法
CN1889686A (zh) * 2006-07-14 2007-01-03 北京时越网络技术有限公司 一种同时显示多路视频信息的方法
CN101820524A (zh) * 2010-03-22 2010-09-01 中兴通讯股份有限公司 用于电视会议的视频播放方法

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1484918A (zh) * 2000-12-01 2004-03-24 ������·�ɷ����޹�˾ 视频终端,视频终端通信系统,以及电视会议系统
CN101370114B (zh) * 2008-09-28 2011-02-02 华为终端有限公司 视频及音频处理方法、多点控制单元和视频会议系统

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN2509797Y (zh) * 2001-09-29 2002-09-04 张健 具有画面分割功能的图像编码器
CN1422078A (zh) * 2001-11-22 2003-06-04 丽台科技股份有限公司 视频会议系统
CN1571508A (zh) * 2003-07-19 2005-01-26 华为技术有限公司 一种实现多画面的方法
CN1889686A (zh) * 2006-07-14 2007-01-03 北京时越网络技术有限公司 一种同时显示多路视频信息的方法
CN101820524A (zh) * 2010-03-22 2010-09-01 中兴通讯股份有限公司 用于电视会议的视频播放方法

Also Published As

Publication number Publication date
CN101820524A (zh) 2010-09-01

Similar Documents

Publication Publication Date Title
WO2011116611A1 (zh) 用于电视会议的视频播放方法
AU2011258272B2 (en) Systems and methods for scalable video communication using multiple cameras and multiple monitors
JP5852157B2 (ja) スケーラブルなビデオ通信システムおよび音声通信システムにおけるビューレイアウト管理の改良
KR101224097B1 (ko) 다자간 회의의 제어 방법 및 장치
US8233026B2 (en) Scalable video encoding in a multi-view camera system
EP2214410B1 (en) Method and system for conducting continuous presence conferences
US7561179B2 (en) Distributed real-time media composer
JP4921488B2 (ja) スケーラブルなビデオ符号化を用いて、またスケーラブルなテレビ会議サーバを複合してテレビ会議を行うためのシステムおよび方法
US8237765B2 (en) Video conferencing device which performs multi-way conferencing
US20180343419A1 (en) Method and system for conducting video conferences of diverse participating devices
WO2008131644A1 (fr) Procédé, dispositif et système pour réaliser une commutation d'images dans le service vidéo
US20070200923A1 (en) System and method for videoconferencing using scalable video coding and compositing scalable video conferencing servers
US9497390B2 (en) Video processing method, apparatus, and system
US8836753B2 (en) Method, apparatus, and system for processing cascade conference sites in cascade conference
WO2012079424A1 (zh) 分布式视频的处理方法、系统及多点控制单元
JP2013042492A (ja) 常駐表示式ビデオ会議においてビデオストリームを切替える方法およびシステム
CN115734028A (zh) 一种基于级联编码的媒体流推送方法及系统
Iglesias Gracia Development of an integrated interface between SAGE and Ultragrid

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10848273

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 10848273

Country of ref document: EP

Kind code of ref document: A1