CN112969039A - Video fusion method, device and equipment and readable storage medium - Google Patents

Video fusion method, device and equipment and readable storage medium Download PDF

Info

Publication number
CN112969039A
CN112969039A CN202110536986.4A CN202110536986A CN112969039A CN 112969039 A CN112969039 A CN 112969039A CN 202110536986 A CN202110536986 A CN 202110536986A CN 112969039 A CN112969039 A CN 112969039A
Authority
CN
China
Prior art keywords
videos
terminal
video
attributes
fused
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110536986.4A
Other languages
Chinese (zh)
Other versions
CN112969039B (en
Inventor
罗开清
许磊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang Huachuang Video Signal Technology Co Ltd
Original Assignee
Zhejiang Huachuang Video Signal Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang Huachuang Video Signal Technology Co Ltd filed Critical Zhejiang Huachuang Video Signal Technology Co Ltd
Priority to CN202110536986.4A priority Critical patent/CN112969039B/en
Publication of CN112969039A publication Critical patent/CN112969039A/en
Application granted granted Critical
Publication of CN112969039B publication Critical patent/CN112969039B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/265Mixing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The invention discloses a video fusion method, a device, equipment and a readable storage medium, wherein the method comprises the following steps: acquiring videos of all terminals, wherein the videos comprise terminal identifications; searching a terminal attribute corresponding to the identifier of the terminal from the terminal attribute table; classifying videos input by each terminal according to single attributes and/or attribute combinations in the terminal attributes; determining videos corresponding to a main picture when various videos are fused concurrently; and for each type of video, selecting one or more paths of videos to be fused with the video corresponding to the main picture to obtain a corresponding path of first fused video. By implementing the method and the device, the videos can be fused concurrently according to the single attribute and/or the combination of the attributes in the terminal attributes to obtain various types of fused videos, so that a terminal user can select the fused videos as required, the terminal videos participating in the fusion can be dynamically adjusted during video fusion, and the user experience is improved.

Description

Video fusion method, device and equipment and readable storage medium
Technical Field
The present application relates to the field of video conference monitoring technologies, and in particular, to a video fusion method, apparatus, device, and readable storage medium.
Background
With the popularization of the mobile internet, the demand of the video conference is more and more increased. In a large video conference, video conference terminals acquire videos and encode the videos and then send the videos to a video conference server, and the video conference server fuses or un-fuses the videos sent by a plurality of video conference terminals according to conference requirements and then sends the videos to all the video conference terminals, so that video calls of multi-party participants are realized. In the scheme of video fusion, a video conference server side fuses multiple paths of videos sent by a plurality of video conference terminals into a fused video according to fixed configuration parameters, and then sends the fused video to each video conference terminal, so that the fused videos received by each video conference terminal are the same, and the video conference terminals cannot select the fused video according to requirements.
Disclosure of Invention
In view of this, embodiments of the present invention provide a video fusion method, apparatus, device and readable storage medium to solve the problem that the existing video fusion scheme cannot meet the requirement of selecting a fusion video as required by a video conference terminal.
In order to solve the above problem, in a first aspect, an embodiment of the present invention provides a video fusion method, including: acquiring videos of all terminals, wherein the videos comprise terminal identifications; searching a terminal attribute corresponding to the identifier of the terminal from the terminal attribute table; classifying videos input by each terminal according to single attributes and/or attribute combinations in the terminal attributes; determining videos corresponding to a main picture when various videos are fused concurrently; and for each type of video, selecting one or more paths of videos from the videos to be fused with the video corresponding to the main picture to obtain a corresponding path of first fused video.
Optionally, the terminal attribute includes a terminal participation level, a terminal located area, and a service type of a participant corresponding to the terminal.
Optionally, for each type of video, determining the arrangement form of the sub-pictures and the number of the sub-pictures with various terminal attributes during video fusion; randomly selecting videos with various terminal attributes corresponding to the number of the sub-pictures with various terminal attributes from the videos, and fusing the randomly selected videos and the videos corresponding to the main pictures according to the arrangement form of the sub-pictures to obtain a corresponding path of first fused video; or, acquiring a video of a specific terminal from the video, determining the difference value between the number of the sub-pictures of the terminal attribute and the number of the specific terminals of the terminal attribute for the sub-pictures of each terminal attribute, and randomly selecting the video of the terminal attribute corresponding to the difference value from the video; fusing the randomly selected video, the video of the specific terminal and the video corresponding to the main picture according to the arrangement form of the sub-pictures to obtain a corresponding path of first fused video; or, acquiring videos of a specific terminal with the same number of sub-pictures with various terminal attributes from the videos, and fusing the videos of the specific terminal and the videos corresponding to the main pictures according to the arrangement form of the sub-pictures to obtain a corresponding path of first fused video.
Optionally, randomly selecting videos of various terminal attributes corresponding to the number of sub-pictures of various terminal attributes from the videos of the category and randomly selecting videos of the terminal attributes corresponding to the difference value from the videos of the category, including: and randomly selecting videos with various terminal attributes corresponding to the number of the sub-pictures with various terminal attributes from the videos and randomly selecting videos with the terminal attributes corresponding to the difference values from the videos according to a preset period.
Optionally, after selecting one or more paths of videos to be fused with the video corresponding to the main picture, the video fusion method further includes: performing quality detection on one or more paths of videos and videos corresponding to the main picture; and if one or more paths of videos and/or videos with abnormal quality exist in the videos corresponding to the main picture, randomly selecting the videos corresponding to the number of the videos with abnormal quality, and replacing the videos with abnormal quality with the reselected videos.
Optionally, the video fusion method further includes: and selecting one or more paths of first fusion videos to be fused with the video corresponding to the main picture to obtain a corresponding path of second fusion video.
In a second aspect, an embodiment of the present invention provides a video fusion apparatus, including: the device comprises an acquisition unit, a processing unit and a display unit, wherein the acquisition unit is used for acquiring videos of all terminals, and the videos comprise terminal identifications; the searching unit is used for searching the terminal attribute corresponding to the identifier of the terminal from the terminal attribute table; the classification unit is used for classifying the video input by each terminal according to a single attribute and/or a combination of attributes in the second attributes of the terminals; the determining unit is used for determining videos corresponding to the main picture when various videos are fused concurrently; and the fusion unit is used for selecting one or more paths of videos from each type of videos to fuse with the videos corresponding to the main picture to obtain a corresponding path of first fusion video.
Optionally, the terminal attribute includes a terminal participation level, a terminal located area, and a service type of a participant corresponding to the terminal.
In a third aspect, an embodiment of the present invention provides an electronic device, including: at least one processor; and a memory communicatively coupled to the at least one processor; wherein the memory stores instructions executable by the at least one processor, the instructions being executable by the at least one processor to cause the at least one processor to perform the video fusion method as in the first aspect or any of the embodiments of the first aspect.
In a fourth aspect, an embodiment of the present invention provides a computer-readable storage medium, where computer instructions are stored, and the computer instructions are configured to cause a computer to execute a video fusion method as in the first aspect or any implementation manner of the first aspect.
According to the video fusion method, the video fusion device, the video fusion equipment and the readable storage medium, the video of each terminal is obtained, and the video comprises the identifier of the terminal; searching a terminal attribute corresponding to the identifier of the terminal from the terminal attribute table; classifying videos input by each terminal according to single attributes and/or attribute combinations in the terminal attributes; determining videos corresponding to a main picture when various videos are fused concurrently; for each type of video, selecting one or more paths of videos to be fused with the video corresponding to the main picture to obtain a corresponding path of first fused video; therefore, concurrent fusion of videos can be carried out according to single attributes and/or attribute combinations in the terminal attributes, multiple types of fused videos can be obtained, so that a terminal user can select the fused videos as required, the terminal videos participating in fusion can be dynamically adjusted during video fusion, and user experience is improved.
The foregoing description is only an overview of the technical solutions of the present application, and the present application can be implemented according to the content of the description in order to make the technical means of the present application more clearly understood, and the following detailed description of the present application is given in order to make the above and other objects, features, and advantages of the present application more clearly understandable.
Drawings
Fig. 1 is a schematic flow chart of a video fusion method according to an embodiment of the present invention;
fig. 2 is a schematic diagram of interaction between a cloud processing center and a terminal in an embodiment of the present invention;
FIG. 3 is a diagram of a frame layout structure during video fusion according to an embodiment of the present invention;
FIG. 4 is a schematic flow chart of secondary video fusion according to an embodiment of the present invention;
fig. 5 is a schematic structural diagram of a video fusion apparatus according to an embodiment of the present invention;
fig. 6 is a schematic diagram of a hardware structure of an electronic device according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, but not all, embodiments of the present invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The embodiment of the invention provides a video fusion method which is particularly suitable for a video conference system with a large number of terminals participating. The method can be applied to a cloud processing center, a high-performance server or a cluster formed by the servers, wherein the cluster formed by the cloud processing center, the high-performance server and the servers has the capacity of accessing and processing a large number of videos. As shown in fig. 1, the video fusion method includes:
s101, obtaining videos of all terminals, wherein the videos comprise terminal identifications; specifically, the video is a path of code stream with sound and pictures, and the video includes a unique identifier of the terminal. As shown in fig. 2, the cloud processing center is connected to each terminal, and after the video is collected by each terminal, the video is uploaded to the cloud processing center, and the cloud processing center can receive the video of each terminal.
S102, finding out a terminal attribute corresponding to the identifier of the terminal from the terminal attribute table; specifically, a plurality of attributes of the terminals participating in the video conference may be set in the cloud processing center, so as to obtain a terminal attribute table. The terminal attribute table comprises the terminal identification and the terminal attribute, and the corresponding terminal attribute can be found through the terminal identification. The terminal attributes may include, but are not limited to, the participation level of the terminal, the area where the terminal is located, and the service type of the participant corresponding to the terminal. For example: the terminal 1, the participant level is "level 1", the area is "headquarters", the business type of the participant is "market reward portion"; the terminal 2, the participant level is "level 1", the area is "provincial subarea", and the business type of the participant is "general manager office"; the terminal 3, the participant level is "level 2", the area is "provincial district", the business type of the participant is "market reward portion"; and the terminal 4 has a participant grade of 3, the area of the participant is a city zone, and the service type of the participant is market reward part. And setting the attributes of the terminal, wherein the attributes can be used for classifying the video of the terminal.
S103, classifying the videos input by each terminal according to a single attribute and/or a combination of attributes in the terminal attributes; specifically, if the video corresponds to the terminal and the terminal corresponds to the terminal attribute, the video corresponds to the terminal attribute, and then the video input by each terminal can be classified according to a single attribute and/or a combination of attributes in the terminal attribute, for example, i. the video input by each terminal is classified according to the terminal participation level, the area where the terminal is located, and the service type where the participant corresponding to the terminal is located; and ii, classifying videos input by each terminal according to the participation grade of the terminal and the service type of the participant corresponding to the terminal. Videos input by each terminal can be classified according to a single attribute and/or a combination of attributes in the terminal attributes, and then videos of multiple categories can be obtained.
S104, determining videos corresponding to the main picture when various videos are fused concurrently; specifically, a determination policy of a video corresponding to a main picture during video fusion can be preset, so that after the videos of each terminal are received, the video corresponding to the main picture during video fusion can be determined from the videos of each terminal according to the preset determination policy. The preset determination strategy can be a video corresponding to a main picture when determining that the video of the terminal of the current speaker is various videos and is fused or a video corresponding to a main picture when determining that the video of a specific terminal is various videos and is fused.
And S105, for each type of video, selecting one or more paths of videos from the videos and fusing the videos corresponding to the main picture to obtain a corresponding path of first fused video. Specifically, for each type of video, the number of video paths for video fusion may be set, for example, 9 video paths, 16 video paths, 24 video paths, and the like are configured. When the videos are fused, one or more paths of videos can be selected from each type of videos according to the set number of video paths during video fusion, and the videos corresponding to the main picture are fused to obtain a path of corresponding first fusion video. After each type of video is fused in the cloud processing center, all first fused videos participated by the terminal can be seen by each participating terminal, and a terminal user can select one of the first fused videos by himself to play on a large screen. The large-screen playing process of each terminal is as follows: 1) and acquiring all fused video lists participated by the current terminal, wherein each path of fused video has corresponding fused arrangement information. 2) And the terminal automatically controls whether the fused video is played. 3) And after the terminal selects the fused video, the fused video is played by itself. 4) Furthermore, the terminal sets alternative fusion videos, and when a certain fusion video fails, the alternative fusion video is automatically switched to play.
According to the video fusion method provided by the embodiment of the invention, videos of all terminals are obtained, and the videos comprise the identification of the terminals; searching a terminal attribute corresponding to the identifier of the terminal from the terminal attribute table; classifying videos input by each terminal according to single attributes and/or attribute combinations in the terminal attributes; determining videos corresponding to a main picture when various videos are fused concurrently; for each type of video, selecting one or more paths of videos to be fused with the video corresponding to the main picture to obtain a corresponding path of first fused video; therefore, concurrent fusion of videos can be carried out according to single attributes and/or attribute combinations in the terminal attributes, multiple types of fused videos can be obtained, so that a terminal user can select the fused videos as required, the terminal videos participating in fusion can be dynamically adjusted during video fusion, and user experience is improved.
In an optional embodiment, for each type of video, selecting one or more paths of videos from the videos and fusing the selected videos with videos corresponding to the main picture to obtain a corresponding path of first fused video, including: for each type of video, determining the arrangement form of the sub-pictures and the number of the sub-pictures with various terminal attributes during video fusion; randomly selecting videos with various terminal attributes corresponding to the number of the sub-pictures with various terminal attributes from the videos, and fusing the randomly selected videos and the videos corresponding to the main pictures according to the arrangement form of the sub-pictures to obtain a corresponding path of first fused video; or, acquiring videos of specific terminals from the videos, determining the difference value between the number of the sub-pictures of the terminal attributes and the number of the specific terminals of the terminal attributes for the sub-pictures of each terminal attribute, and randomly selecting the videos of the terminal attributes corresponding to the difference value from the videos; fusing the randomly selected video, the video of the specific terminal and the video corresponding to the main picture according to the arrangement form of the sub-pictures to obtain a corresponding path of first fused video; or, acquiring videos of a specific terminal with the same number of sub-pictures with various terminal attributes from the videos, and fusing the videos of the specific terminal and the videos corresponding to the main pictures according to the arrangement form of the sub-pictures to obtain a corresponding path of first fused video.
Specifically, for each type of video, the number of video paths and the arrangement form of the sub-pictures used in video fusion can be set. The arrangement form of the sub-pictures comprises the terminal attributes corresponding to the sub-pictures and the arrangement positions of the sub-pictures with various terminal attributes. When each type of video fusion is performed, the video paths and the arrangement forms of the sub-pictures during the video fusion can be the same or different. And subtracting the main picture path number (one path) from the video path number during video fusion to obtain the number of the sub-pictures during video fusion. After the number of the sub-pictures during video fusion is determined, the video corresponding to the sub-pictures can be selected by the following three methods, including:
1. randomly selecting videos corresponding to the number of the sub-pictures from the videos; when the video corresponding to the number of the sub-pictures is randomly selected, the number of the sub-pictures of each level in the terminal participating level, the number of the sub-pictures of each service type in the service type of the participant corresponding to the terminal or the number of the sub-pictures in each area in the area of the terminal can be determined according to the preset ratio of each level in the terminal participating level, the ratio of each service type in the service type of the participant corresponding to the terminal or the ratio of each area in the area of the terminal and the number of the sub-pictures, and then the video of the corresponding terminal attribute and the corresponding number is randomly selected. And, a third level selection policy may also be set. For example, i, after videos of each terminal are classified according to the area where the terminal is located, when the videos are merged, a sub-picture selection strategy is a terminal participation level and a proportion of the participation level, and a third-level selection strategy is a service type (which may not be selected); after classifying the videos of each terminal according to the service type of the participant corresponding to the terminal, when the videos are merged, the sub-picture selection strategy is the participation level of the terminal and the proportion of the participation level, and the third-level selection strategy is the area where the terminal is located (which can be not selected); and iii, after the videos of all the terminals are classified according to the participation levels of the terminals, when the videos are merged, the sub-picture selection strategy is the service type of the participant corresponding to the terminal and the proportion of each service type, and the third-level selection strategy is the area where the terminal is located (which can be not selected).
2. Acquiring a video of a specific terminal from the video, wherein the video serves as a video corresponding to a part of sub-pictures, determining the difference value between the number of the sub-pictures and the number of the specific terminals, and randomly selecting a video corresponding to the difference value from the video; the video corresponding to part of the sub-pictures can be set as the video of the specific terminal, that is, part of the video of the specific terminal can be fixed, and the rest part of the video is randomly selected from the videos. Similarly, when randomly selecting, determining the number of the sub-pictures of each grade in the participation grade of the terminal, the number of the sub-pictures of each service type in the service type of the participant corresponding to the terminal or the number of the sub-pictures in each area of the terminal according to the preset ratio of each grade in the participation grade of the terminal, the ratio of each service type in the service type of the participant corresponding to the terminal or the ratio of each area of the terminal and the number of the sub-pictures, and then according to the terminal attribute of the video of the specific terminal and the number of the videos of the specific terminal, calculating the residual number of the sub-pictures of each grade in the participation grade of the terminal, the residual number of the sub-pictures of each service type in the service type of the participant corresponding to the terminal or the residual number of the sub-pictures of each area in the area where the terminal is located, and randomly selecting the videos of the corresponding terminal attribute and the corresponding number. And, a third level selection policy may also be set.
For example, when the service type of the participant corresponding to the terminal classifies the video of each terminal, the fusion video corresponding screen is as shown in fig. 3, and the following description is given (the same reason is given by classifying other attributes):
1) main picture: from the terminal where the current speaker/speaker is located or the terminal where the conference organizer is located; 2) configuring the video selection proportion of the terminals of each grade when participating in video fusion, wherein the proportion of each grade can be 0-100%, and the sum of the proportions is 100%; 3) video sources corresponding to each level of sprite: a portion of the fixed terminal's video may be selected and the remaining video randomly selected and periodically updated within the same class of video. 4) And intelligently detecting the abnormality aiming at the video corresponding to any picture, and automatically removing and replacing the abnormal video by the cloud processing center.
3. Acquiring videos of a specific terminal with the same number as that of the sub-pictures from the videos; the video corresponding to all the sub-pictures can be set as the video of the specific terminal.
After the video corresponding to the sub-picture is selected, the video corresponding to the sub-picture and the video corresponding to the main picture can be fused to obtain a path of fused video.
In the embodiment of the invention, when the videos are fused, the arrangement form of the sub-pictures and the source and the position of the fused videos can be arranged, so that the user experience can be improved.
In an alternative embodiment, the randomly selecting videos of various terminal attributes corresponding to the number of sub-pictures of various terminal attributes from the videos of the category and randomly selecting videos of the terminal attributes corresponding to the difference value from the videos of the category includes: and randomly selecting videos with various terminal attributes corresponding to the number of the sub-pictures with various terminal attributes from the videos and randomly selecting videos with the terminal attributes corresponding to the difference values from the videos according to a preset period.
In the embodiment of the invention, videos with various terminal attributes corresponding to the number of the sub-pictures with various terminal attributes are randomly selected from the videos and videos with the terminal attributes corresponding to the difference value are randomly selected from the videos, so that the videos corresponding to the sub-pictures in the fusion video can be dynamically adjusted periodically, and the user experience is improved.
In an optional embodiment, after selecting one or more videos to be fused with a video corresponding to the main picture, the video fusion method further includes: performing quality detection on one or more paths of videos and videos corresponding to the main picture; and if one or more paths of videos and/or videos with abnormal quality exist in the videos corresponding to the main picture, randomly selecting the videos corresponding to the number of the videos with abnormal quality, and replacing the videos with abnormal quality with the reselected videos.
Specifically, in the video fusion process, the cloud processing center can automatically detect the quality of videos participating in video fusion, can timely identify abnormal videos, randomly select videos corresponding to the number of videos with abnormal quality, and replace the videos with abnormal quality. During quality detection, the anomaly duration can be configured, for example, if the code stream quality is found to be poor continuously for 3 seconds, the code stream is considered to have the quality anomaly, and then the code stream is automatically replaced. The quality of the video participating in the video fusion is detected and the video with the quality problem is replaced, so that the fused video can be smoothly played.
In an optional embodiment, the video fusion method further includes: and selecting one or more paths of first fusion videos to be fused with the video corresponding to the main picture to obtain a corresponding path of second fusion video.
Specifically, for the multi-channel video fusion mechanism, secondary fusion may be performed again based on one or more channels of the fused video, as shown in fig. 4, on the premise that the main picture is not changed during the selected video fusion, the selection of each sub-picture may be derived from different fused videos and videos of the terminal. The second fused video can be arranged on the picture by adopting the same mechanism as the first fused video, namely, the number of the fused videos is selected, the sub-pictures can be randomly selected, selected according to attributes, selected according to proportions, selected fixedly and the like. And, the second fused video can be fused again, and the process can be circulated.
In the embodiment of the invention, one or more paths of first fusion videos are selected to be fused with the video corresponding to the main picture to obtain one path of corresponding second fusion video, so that for a conference in which a large number of terminals participate, one second fusion video shows most or all terminal participation conditions.
An embodiment of the present invention further provides a video fusion apparatus, as shown in fig. 5, including: an obtaining unit 21, configured to obtain a video of each terminal, where the video includes an identifier of the terminal; the detailed description of the specific implementation manner is given in step S101 of the above method embodiment, and is not repeated herein. A searching unit 22, configured to search a terminal attribute corresponding to the identifier of the terminal from the terminal attribute table; the detailed description of the specific implementation manner is given in step S102 of the above method embodiment, and is not repeated herein. A classifying unit 23, configured to classify the video input by each terminal according to a single attribute and/or a combination of attributes in the second attributes of the terminals; the detailed description of the specific implementation manner is given in step S103 of the above method embodiment, and is not repeated herein. The determining unit 24 is configured to determine a video corresponding to a main picture when various videos are fused concurrently; the detailed description of the specific implementation manner is given in step S104 of the above method embodiment, and is not repeated herein. And the fusion unit 25 is configured to select one or more channels of videos from each type of video to fuse with the video corresponding to the main picture, so as to obtain a corresponding channel of first fusion video. The detailed description of the specific implementation manner is given in step S105 of the above method embodiment, and is not repeated herein.
According to the video fusion device provided by the embodiment of the invention, videos of all terminals are obtained, and the videos comprise the identification of the terminals; searching a terminal attribute corresponding to the identifier of the terminal from the terminal attribute table; classifying videos input by each terminal according to single attributes and/or attribute combinations in the terminal attributes; determining videos corresponding to a main picture when various videos are fused concurrently; for each type of video, selecting one or more paths of videos to be fused with the video corresponding to the main picture to obtain a corresponding path of first fused video; therefore, concurrent fusion of videos can be carried out according to single attributes and/or attribute combinations in the terminal attributes, multiple types of fused videos can be obtained, so that a terminal user can select the fused videos as required, the terminal videos participating in fusion can be dynamically adjusted during video fusion, and user experience is improved.
In an optional embodiment, the terminal attribute includes a terminal participation level, a terminal located area, and a service type of a participant corresponding to the terminal.
Specifically, the terminal attribute may include, but is not limited to, a terminal participation level, an area where the terminal is located, and a service type where a participant corresponding to the terminal is located. For example: the terminal 1, the participant level is "level 1", the area is "headquarters", the business type of the participant is "market reward portion"; the terminal 2, the participant level is "level 1", the area is "provincial subarea", and the business type of the participant is "general manager office"; the terminal 3, the participant level is "level 2", the area is "provincial district", the business type of the participant is "market reward portion"; and the terminal 4 has a participant grade of 3, the area of the participant is a city zone, and the service type of the participant is market reward part. And setting the attributes of the terminal, wherein the attributes can be used for classifying the video of the terminal.
Based on the same inventive concept as the video fusion method in the foregoing embodiment, an embodiment of the present invention further provides an electronic device, as shown in fig. 6, including: a processor 31 and a memory 32, wherein the processor 31 and the memory 32 may be connected by a bus or other means, and the connection by the bus is illustrated in fig. 6 as an example.
The processor 31 may be a Central Processing Unit (CPU). The Processor 31 may also be other general purpose processors, Digital Signal Processors (DSPs), Application Specific Integrated Circuits (ASICs), Field Programmable Gate Arrays (FPGAs) or other Programmable logic devices, discrete Gate or transistor logic devices, discrete hardware components, or combinations thereof.
The memory 32, which is a non-transitory computer readable storage medium, may be used for storing non-transitory software programs, non-transitory computer executable programs, and modules, such as program instructions/modules corresponding to the video fusion method in the embodiment of the present invention. The processor 31 executes various functional applications and data processing of the processor by running non-transitory software programs, instructions and modules stored in the memory 32, that is, the video fusion method in the above-described method embodiment is implemented.
The memory 32 may include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application program required for at least one function; the storage data area may store data created by the processor 31, and the like. Further, the memory 32 may include high speed random access memory, and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid state storage device. In some embodiments, the memory 32 may optionally include memory located remotely from the processor 31, and these remote memories may be connected to the processor 31 via a network. Examples of such networks include, but are not limited to, the internet, intranets, local area networks, mobile communication networks, and combinations thereof.
One or more of the modules described above are stored in the memory 32 and, when executed by the processor 31, perform the video fusion method in the embodiment shown in fig. 1.
The details of the electronic device may be understood with reference to the corresponding related description and effects in the embodiment shown in fig. 1, and are not described herein again.
It will be understood by those skilled in the art that all or part of the processes of the methods of the embodiments described above can be implemented by a computer program, which can be stored in a computer-readable storage medium, and when executed, can include the processes of the embodiments of the methods described above. The storage medium may be a magnetic Disk, an optical Disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a Flash Memory (Flash Memory), a Hard Disk (Hard Disk Drive, abbreviated as HDD), a Solid State Drive (SSD), or the like; the storage medium may also comprise a combination of memories of the kind described above.
As will be appreciated by one skilled in the art, embodiments of the present invention may be provided as a method, system, or computer program product. Accordingly, the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present invention may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present invention is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable information processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable information processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable information processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable information processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
It will be apparent to those skilled in the art that various changes and modifications may be made in the present invention without departing from the spirit and scope of the invention. Thus, if such modifications and variations of the present invention fall within the scope of the claims of the present invention and their equivalents, the present invention is also intended to include such modifications and variations.

Claims (10)

1. A method of video fusion, comprising:
acquiring videos of all terminals, wherein the videos comprise terminal identifications;
searching a terminal attribute corresponding to the identifier of the terminal from the terminal attribute table;
classifying videos input by each terminal according to single attributes and/or attribute combinations in the terminal attributes;
determining videos corresponding to a main picture when various videos are fused concurrently;
and for each type of video, selecting one or more paths of videos from the videos and fusing the videos corresponding to the main picture to obtain a corresponding path of first fused video.
2. The video fusion method according to claim 1, wherein the terminal attributes include a terminal participation level, a terminal located area, and a service type of a participant corresponding to the terminal.
3. The video fusion method according to claim 2, wherein for each type of video, selecting one or more videos from the videos and fusing the selected videos with the video corresponding to the main picture to obtain a corresponding one of the first fused videos, comprises:
for each type of video, determining the arrangement form of the sub-pictures and the number of the sub-pictures with various terminal attributes during video fusion;
randomly selecting videos with various terminal attributes corresponding to the number of sub-pictures with various terminal attributes from the videos, and fusing the randomly selected videos and the videos corresponding to the main picture according to the arrangement form of the sub-pictures to obtain a corresponding path of first fused video; alternatively, the first and second electrodes may be,
acquiring videos of specific terminals from the videos, determining the difference value between the number of the sub-pictures of the terminal attributes and the number of the specific terminals of the terminal attributes for the sub-pictures of each terminal attribute, and randomly selecting the videos of the terminal attributes corresponding to the difference value from the videos; fusing the randomly selected video, the video of the specific terminal and the video corresponding to the main picture according to the arrangement form of the sub-pictures to obtain a corresponding path of first fused video; alternatively, the first and second electrodes may be,
and acquiring videos of a specific terminal with the same number of sub-pictures with various terminal attributes from the videos, and fusing the videos of the specific terminal and the videos corresponding to the main pictures according to the arrangement form of the sub-pictures to obtain a corresponding path of first fused video.
4. The video fusion method according to claim 3, wherein randomly selecting videos of various terminal attributes corresponding to the number of sub-pictures of various terminal attributes from the videos of the category and randomly selecting videos of the terminal attributes corresponding to the difference from the videos of the category comprises:
and randomly selecting videos with various terminal attributes corresponding to the number of the sub-pictures with various terminal attributes from the videos and randomly selecting the videos with the terminal attributes corresponding to the difference value from the videos according to a preset period.
5. The video fusion method according to claim 1, further comprising, after selecting one or more videos to fuse with the video corresponding to the main picture:
performing quality detection on the one or more paths of videos and the video corresponding to the main picture;
and if the one or more paths of videos and/or videos corresponding to the main picture have abnormal quality, randomly selecting the videos corresponding to the quantity of the videos with abnormal quality, and replacing the videos with abnormal quality with the reselected videos.
6. The video fusion method of claim 1, further comprising:
and selecting one or more paths of first fusion videos to be fused with the video corresponding to the main picture to obtain a corresponding path of second fusion video.
7. A video fusion apparatus, comprising:
the device comprises an acquisition unit, a processing unit and a display unit, wherein the acquisition unit is used for acquiring videos of all terminals, and the videos comprise terminal identifications;
the searching unit is used for searching the terminal attribute corresponding to the identifier of the terminal from the terminal attribute table;
the classification unit is used for classifying the video input by each terminal according to a single attribute and/or a combination of attributes in the second attributes of the terminals;
the determining unit is used for determining videos corresponding to the main picture when various videos are fused concurrently;
and the fusion unit is used for selecting one or more paths of videos from each type of videos to fuse with the video corresponding to the main picture to obtain a corresponding path of first fusion video.
8. The video fusion device of claim 7, wherein the terminal attributes include a terminal participation level, a terminal location area, and a service type of a participant corresponding to the terminal.
9. An electronic device, comprising:
at least one processor; and a memory communicatively coupled to the at least one processor; wherein the memory stores instructions executable by the at least one processor to cause the at least one processor to perform the video fusion method of any one of claims 1-6.
10. A computer-readable storage medium storing computer instructions for causing a computer to perform the video fusion method according to any one of claims 1 to 6.
CN202110536986.4A 2021-05-18 2021-05-18 Video fusion method, device and equipment and readable storage medium Active CN112969039B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110536986.4A CN112969039B (en) 2021-05-18 2021-05-18 Video fusion method, device and equipment and readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110536986.4A CN112969039B (en) 2021-05-18 2021-05-18 Video fusion method, device and equipment and readable storage medium

Publications (2)

Publication Number Publication Date
CN112969039A true CN112969039A (en) 2021-06-15
CN112969039B CN112969039B (en) 2021-08-03

Family

ID=76279600

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110536986.4A Active CN112969039B (en) 2021-05-18 2021-05-18 Video fusion method, device and equipment and readable storage medium

Country Status (1)

Country Link
CN (1) CN112969039B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107945592A (en) * 2017-11-03 2018-04-20 杨斌 A kind of synchronization mutual assistance Teaching System
CN109040619A (en) * 2018-08-24 2018-12-18 合肥景彰科技有限公司 A kind of video fusion method and apparatus
CN109688365A (en) * 2018-12-27 2019-04-26 北京真视通科技股份有限公司 The processing method and computer readable storage medium of video conference
US20200304547A1 (en) * 2019-03-20 2020-09-24 Zoom Video Communications, Inc. Method and System for Elevating A Phone Call into A Video Conferencing Session
EP3772851A1 (en) * 2019-08-09 2021-02-10 Mitel Networks (International) Limited Method and system for adapted modality conferencing

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107945592A (en) * 2017-11-03 2018-04-20 杨斌 A kind of synchronization mutual assistance Teaching System
CN109040619A (en) * 2018-08-24 2018-12-18 合肥景彰科技有限公司 A kind of video fusion method and apparatus
CN109688365A (en) * 2018-12-27 2019-04-26 北京真视通科技股份有限公司 The processing method and computer readable storage medium of video conference
US20200304547A1 (en) * 2019-03-20 2020-09-24 Zoom Video Communications, Inc. Method and System for Elevating A Phone Call into A Video Conferencing Session
EP3772851A1 (en) * 2019-08-09 2021-02-10 Mitel Networks (International) Limited Method and system for adapted modality conferencing

Also Published As

Publication number Publication date
CN112969039B (en) 2021-08-03

Similar Documents

Publication Publication Date Title
US8285810B2 (en) Aggregating media collections between participants of a sharing network utilizing bridging
US9396196B2 (en) Method and system for aggregating media collections between participants of a sharing network
US8285811B2 (en) Aggregating media collections to provide a primary list and sorted sub-lists
US8484311B2 (en) Pruning an aggregate media collection
CN103718166B (en) Messaging device, information processing method
US9686329B2 (en) Method and apparatus for displaying webcast rooms
US20140047049A1 (en) Methods and systems for linking and prioritizing chat messages
CN105933375B (en) Method and device for monitoring microphone connection session and server
CN106506391B (en) Network traffic distribution method and router
CN107105309A (en) Live dispatching method and device
CN102215375B (en) The system of selection of the video source of the sprite of more pictures and device in multimedia conferencing
CN111432246B (en) Method, device and storage medium for pushing video data
CN109819285A (en) A kind of live broadcasting method, device, electronic equipment and storage medium
RU2609072C2 (en) Recommender system and control method therefor
US11089072B2 (en) Real time optimized content delivery framework
CN107205132A (en) The seamless transitions of video session between network topology and centralized bridge topology
CN112752037A (en) Processing method and device for mixed audio and video stream
CA2951525A1 (en) Communication apparatus, communication system, communication management system, and communication control method
CN112969039B (en) Video fusion method, device and equipment and readable storage medium
CN111131757A (en) Video conference display method, device and storage medium
CN108322787A (en) Video stream distributing method, device and electronic equipment
CN105979225A (en) Monitoring method and device for group video room
CN111405308B (en) Method and device for sending live audio and video data
CN114827675A (en) Video data processing method and device for application program
CN114051146A (en) Method and system for preventing repeated plug flow

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant