CN108965783B - Video data processing method and video network recording and playing terminal - Google Patents

Video data processing method and video network recording and playing terminal Download PDF

Info

Publication number
CN108965783B
CN108965783B CN201711447037.9A CN201711447037A CN108965783B CN 108965783 B CN108965783 B CN 108965783B CN 201711447037 A CN201711447037 A CN 201711447037A CN 108965783 B CN108965783 B CN 108965783B
Authority
CN
China
Prior art keywords
video
subtitle
target
image
conference
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201711447037.9A
Other languages
Chinese (zh)
Other versions
CN108965783A (en
Inventor
何晓辉
朱道彦
韩杰
王金龙
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hainan Shilian Communication Technology Co.,Ltd.
Original Assignee
Visionvera Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Visionvera Information Technology Co Ltd filed Critical Visionvera Information Technology Co Ltd
Priority to CN201711447037.9A priority Critical patent/CN108965783B/en
Publication of CN108965783A publication Critical patent/CN108965783A/en
Application granted granted Critical
Publication of CN108965783B publication Critical patent/CN108965783B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems
    • H04N7/155Conference systems involving storage of or access to video conference sessions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/40Support for services or applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • H04L65/75Media network packet handling
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/278Subtitling

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Television Signal Processing For Recording (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The embodiment of the invention discloses a video data processing method and a video network recording and broadcasting terminal, wherein the method comprises the following steps: receiving video of the video networking conference and subtitle texts to be converted; the video networking conference video comprises a plurality of frames of video images; extracting a target subtitle image corresponding to the subtitle text to be converted; and adding the target subtitle image to at least one frame of video image in the video networking conference video to obtain the target video networking conference video. According to the embodiment of the invention, the problem that the user cannot acquire the conference related information from the recorded video networking conference video is solved.

Description

Video data processing method and video network recording and playing terminal
Technical Field
The embodiment of the invention relates to the field of communication, in particular to a video data processing method and a video network recording and broadcasting terminal.
Background
At present, with the rapid development of network technologies, bidirectional communications such as video conferences and video teaching are widely popularized in the aspects of life, work, learning and the like of users, and video networking technologies are increasingly applied to various technical fields.
When video networking technology is used for video networking conferences, videos of the video networking conferences are often required to be recorded and played for storage. When recording and broadcasting video networking conference, the video of the video networking conference recording the audio and video of the conference is obtained by storing the video of the whole conference process.
However, when a video networking conference is performed, participants often publish text information in the conference, and the stored video data of the video networking conference does not contain the text information, so that a user cannot completely watch the recorded video networking conference, and the user cannot comprehensively know the situation of the conference due to the fact that the text information published during the conference cannot be seen, so that the recorded video networking conference video cannot effectively transmit the related information of the conference to the user.
Therefore, the video data processing method in the prior art has the problem that a user cannot acquire the conference related information from the recorded video networking conference video.
Disclosure of Invention
The invention provides a video data processing method and a video network recording and playing terminal, and aims to solve the problem that a user cannot acquire conference related information from recorded and played video network conference videos in the video data processing method in the prior art.
In order to solve the above technical problem, an embodiment of the present invention provides a video data processing method, which is applied to a video recording and playing terminal, where the video recording and playing terminal prestores a plurality of subtitle texts and a plurality of subtitle images respectively corresponding to the subtitle texts, and the method includes:
receiving video of the video networking conference and subtitle texts to be converted; the video networking conference video comprises a plurality of frames of video images;
extracting a target subtitle image corresponding to the subtitle text to be converted;
and adding the target subtitle image to at least one frame of video image in the video networking conference video to obtain the target video networking conference video.
Optionally, the method further comprises:
sending the target video networking conference video to a video networking recording and playing server; and the video network recording and broadcasting server is used for storing the target video network conference video and sending the target video network conference video to a video network terminal.
Optionally, the video networking conference video has a characteristic video format, and before the step of adding the target subtitle image to at least one frame of video image in the video networking conference video to obtain the target video networking conference video, the method further includes:
and converting the target subtitle image into the target subtitle image conforming to the characteristic video format.
Optionally, the subtitle text has a plurality of corresponding candidate subtitle images, the candidate subtitle images have a first subtitle font type, and before the step of extracting the target subtitle image mapped by the subtitle text to be converted, the method further includes:
receiving a caption control command of a conference control terminal; the subtitle control command carries a second subtitle font type;
the step of extracting the target subtitle image corresponding to the subtitle text to be converted comprises the following steps:
determining a plurality of candidate subtitle images mapped by the subtitle text to be converted;
matching the second subtitle font type with a first subtitle font type of the candidate subtitle images;
and extracting candidate subtitle images matched with the first subtitle font type and the second subtitle font type to serve as the target subtitle images.
Optionally, the subtitle control command further carries subtitle appearance characteristic information, and the method further includes:
and adjusting the appearance characteristics of the target subtitle image according to the subtitle appearance characteristic information.
In order to solve the above technical problem, an embodiment of the present invention further provides a video network recording and playing terminal, where the video network recording and playing terminal prestores a plurality of subtitle texts and a plurality of subtitle images respectively corresponding to the subtitle texts, and the terminal includes:
the video and text receiving module is used for receiving video of the video networking conference and the subtitle text to be converted; the video networking conference video comprises a plurality of frames of video images;
the subtitle image extraction module is used for extracting a target subtitle image corresponding to the subtitle text to be converted;
and the subtitle image adding module is used for adding the target subtitle image into at least one frame of video image in the video networking conference video to obtain the target video networking conference video.
Optionally, the terminal further includes:
the conference video sending module is used for sending the target video networking conference video to a video networking recording and playing server; and the video network recording and broadcasting server is used for storing the target video network conference video and sending the target video network conference video to a video network terminal.
Optionally, the video conference over internet has a feature video format, and the terminal further includes:
and the image format conversion module is used for converting the target subtitle image into the target subtitle image conforming to the characteristic video format.
Optionally, the subtitle text has a plurality of corresponding candidate subtitle images, and the candidate subtitle images have a first subtitle font type, and the terminal further includes:
the caption control command receiving module is used for receiving a caption control command of the conference control terminal; the subtitle control command carries a second subtitle font type;
the subtitle image extraction module comprises:
the candidate subtitle image determining submodule is used for determining a plurality of candidate subtitle images mapped by the subtitle text to be converted;
a font type matching sub-module, configured to match the second subtitle font type with the first subtitle font type of the multiple candidate subtitle images;
and the candidate subtitle image extraction sub-module is used for extracting the candidate subtitle image of which the first subtitle font type is matched with the second subtitle font type to serve as the target subtitle image.
Optionally, the subtitle control command further carries subtitle appearance characteristic information, and the terminal further includes:
and the appearance characteristic adjusting module is used for adjusting the appearance characteristic of the target subtitle image according to the subtitle appearance characteristic information.
According to the embodiment of the invention, the caption text is converted into the target caption image and is embedded into the video of the video networking conference, so that the obtained video of the video networking conference contains the caption information, the related information of the conference can be transmitted to a user through the recorded video of the video networking conference, and the problem that the user cannot acquire the related information of the conference from the recorded video of the video networking conference is solved.
Drawings
Fig. 1 is a flowchart of a video data processing method according to an embodiment of the present invention;
fig. 2 is a flowchart of a video data processing method according to a second embodiment of the present invention;
fig. 3 is a block diagram of a video network recording and playing terminal according to a third embodiment of the present invention;
fig. 4 is a block diagram of a video network recording and broadcasting terminal according to a fourth embodiment of the present invention;
fig. 5 is a schematic data flow diagram of a recorded and played conference video according to an embodiment of the present invention;
FIG. 6 is a networking schematic of a video network of the present invention;
FIG. 7 is a diagram of a hardware architecture of a node server according to the present invention;
fig. 8 is a schematic diagram of a hardware architecture of an access switch of the present invention;
fig. 9 is a schematic diagram of a hardware structure of an ethernet protocol conversion gateway according to the present invention.
Detailed Description
Exemplary embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While exemplary embodiments of the present disclosure are shown in the drawings, it should be understood that the present disclosure may be embodied in various forms and should not be limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the disclosure to those skilled in the art.
Example one
Fig. 1 is a flowchart of a video data processing method according to an embodiment of the present invention, and is applied to a video recording and playing terminal, where the video recording and playing terminal prestores a plurality of subtitle texts and a plurality of subtitle images respectively corresponding to the subtitle texts, and the method may specifically include:
step 110, receiving video of a video network conference and a subtitle text to be converted; the video networking conference video comprises a plurality of frames of video images.
It should be noted that the video data processing method provided by the embodiment of the present invention may be applied to a video recording and playing terminal. In the application scene of carrying out the video networking meeting, the user can shoot the video through video networking participant terminal, as the video networking meeting video, and video networking participant terminal can be through the video networking, with video networking meeting video transmission to video networking recorded broadcast terminal. The user can also input the subtitle text through the video network participant terminal, and the video network participant terminal can send the subtitle text to the video network recording and broadcasting terminal through the video network.
Therefore, the video network recording and broadcasting terminal can receive video network conference videos and subtitle texts sent by the multiple video network participant terminals through the video network.
In order to distinguish the description, the caption text received by the video recording and broadcasting terminal is named as the caption text to be converted.
And step 120, extracting a target subtitle image corresponding to the subtitle text to be converted.
In the above video recording and playing terminal, a text image mapping relationship database may be preset, and the database stores a mapping relationship between a subtitle text and a subtitle image.
In specific implementation, after receiving the subtitle text to be converted, the subtitle image mapped by the subtitle text to be converted can be searched in the text image mapping relation database to serve as a target subtitle image, and the target subtitle image is extracted so as to be embedded into the video of the video network conference subsequently.
In practical application, the user can also set the color, type and transparency of the subtitles through the video network participant terminal. For example, a user may submit a subtitle control command to the video recording and playing terminal, and the video recording and playing terminal may adjust the subtitle color, the text type, the transparency of the subtitle in the video, and the like of the target subtitle image according to the subtitle control command.
And step 130, adding the target subtitle image to at least one frame of video image in the video networking conference video to obtain a target video networking conference video.
Video conferencing may consist of a succession of multiple frames of video images. For one or more frames of video images, a target subtitle image may be added thereto. In practical applications, the adding position of the target subtitle image in the video image can be specified.
The video networking conference video added with the target subtitle image is used as a target video networking conference video and can be stored, the obtained target video networking conference video contains the text information published by the participant user, and even after the conference, the user can comprehensively know the situation of the conference through the recorded and broadcasted target video networking conference video.
According to the embodiment of the invention, the caption text is converted into the target caption image and is embedded into the video of the video networking conference, so that the obtained video of the video networking conference contains the caption information, the related information of the conference can be transmitted to a user through the recorded video of the video networking conference, and the problem that the user cannot acquire the related information of the conference from the recorded video of the video networking conference is solved.
Example two
Fig. 2 is a flowchart of a video data processing method according to a second embodiment of the present invention, which is applied to a video recording and playing terminal, where the video recording and playing terminal prestores a plurality of subtitle texts and a plurality of subtitle images respectively corresponding to the subtitle texts, and the method may specifically include:
step 210, receiving video of a video network conference and a subtitle text to be converted; the video networking conference video comprises a plurality of frames of video images.
The video data processing method provided by the embodiment of the invention can be applied to video network recording and broadcasting terminals. In specific implementation, the video networking recording and playing terminal can receive video networking conference videos and subtitle texts sent by the multiple video networking participant terminals through the video networking. The video conference may include a plurality of frames of video images.
Step 220, extracting a target subtitle image corresponding to the subtitle text to be converted.
Optionally, the subtitle text has a corresponding plurality of candidate subtitle images having a first subtitle font type, before the step 220, the method further includes:
receiving a caption control command of a conference control terminal; the subtitle control command carries a second subtitle font type;
the step 220 may specifically include:
determining a plurality of candidate subtitle images mapped by the subtitle text to be converted;
matching the second subtitle font type with a first subtitle font type of the candidate subtitle images;
and extracting candidate subtitle images matched with the first subtitle font type and the second subtitle font type to serve as the target subtitle images.
It should be noted that a text image mapping relationship database may be preset in the video recording and playing terminal, and the database stores the mapping relationship between the subtitle text and the subtitle image. In addition, the subtitle text may have a corresponding plurality of candidate subtitle images having a subtitle font type, for example, a font type such as a song style, a regular style, and the like. For the purpose of distinguishing the description, the preset subtitle font type is named as a first subtitle font type.
In a specific implementation, a user can submit a subtitle control command through a terminal of a participant in the video network, and the subtitle control command can carry a subtitle font type. For the sake of distinguishing the explanation, the subtitle font type carried in the subtitle control command is named as a second subtitle font type.
After receiving the subtitle control command, the video network recording and broadcasting terminal can extract a second subtitle font type from the subtitle control command, match the second subtitle font type with a first subtitle font type of a plurality of candidate subtitle images mapped by the subtitle text to be converted, and extract the candidate subtitle image with the first subtitle font type matched with the second subtitle font type as a target subtitle image. So that the target subtitle image having the font type designated by the user is embedded into the conference video.
Therefore, the user can adjust the font type of the subtitles in the video of the video network conference through the subtitle control command.
Optionally, the subtitle control command further carries subtitle appearance characteristic information, and the method further includes:
and adjusting the appearance characteristics of the target subtitle image according to the subtitle appearance characteristic information.
In specific implementation, a user can also submit a subtitle control command carrying subtitle appearance characteristic information through a video network participant terminal. The subtitle appearance feature information may be used to adjust appearance features such as color and transparency of the subtitle.
After receiving the subtitle control command, the video network recording and playing terminal can adjust the appearance characteristics of the target subtitle image according to the subtitle appearance characteristic information.
Optionally, the video conference via video network has a characteristic video format, and the method may further include:
and converting the target subtitle image into the target subtitle image conforming to the characteristic video format.
In a specific implementation, video conferencing over video typically has a particular characteristic video format, such as the YUV (european color coding method) video format. Therefore, after the target subtitle image is acquired, the target subtitle image can be converted into a characteristic video format, and the target subtitle image can be embedded into the conference video conveniently.
Step 230, adding the target subtitle image to at least one frame of video image in the video networking conference video to obtain a target video networking conference video.
In practical application, a user can also control whether to embed the subtitle image into the video of the video networking conference or not through the video networking participant terminal, so that the flexibility of subtitle processing is improved.
Step 240, sending the target video networking conference video to a video networking recording and playing server; and the video network recording and broadcasting server is used for storing the target video network conference video and sending the target video network conference video to a video network terminal.
In specific implementation, after the target video networking conference video is acquired, the video networking recording and playing terminal can send the target video networking conference video to the video networking recording and playing server, and the video networking recording and playing server stores the target video networking conference video. The video network recording and playing server can send the target video network conference video to a plurality of video network participant terminals for users to browse.
According to the embodiment of the invention, the font type and the appearance characteristic of the caption in the conference video are determined according to the caption control command, so that a user can adjust the caption in the recorded and played video network conference video according to the conference requirement, and the flexibility of caption processing is improved.
EXAMPLE III
Fig. 3 is a block diagram of a video network recording and broadcasting terminal according to a third embodiment of the present invention, where the video network recording and broadcasting terminal prestores a plurality of subtitle texts and a plurality of subtitle images respectively corresponding to the subtitle texts, and the video network recording and broadcasting terminal 300 may specifically include the following modules:
a video and text receiving module 310, configured to receive a video of a video conference and a subtitle text to be converted; the video networking conference video comprises a plurality of frames of video images;
the subtitle image extracting module 320 is configured to extract a target subtitle image corresponding to the subtitle text to be converted;
and the subtitle image adding module 330 is configured to add the target subtitle image to at least one frame of video image in the video networking conference video to obtain a target video networking conference video.
According to the embodiment of the invention, the caption text is converted into the target caption image and is embedded into the video of the video networking conference, so that the obtained video of the video networking conference contains the caption information, the related information of the conference can be transmitted to a user through the recorded video of the video networking conference, and the problem that the user cannot acquire the related information of the conference from the recorded video of the video networking conference is solved.
Example four
Fig. 4 is a block diagram of a video network recording and broadcasting terminal according to a fourth embodiment of the present invention, where the video network recording and broadcasting terminal prestores a plurality of subtitle texts and a plurality of subtitle images respectively corresponding to the subtitle texts, and the video network recording and broadcasting terminal 400 may specifically include the following modules:
a video and text receiving module 410, configured to receive a video of a video conference and a subtitle text to be converted; the video networking conference video comprises a plurality of frames of video images;
the subtitle image extracting module 420 is configured to extract a target subtitle image corresponding to the subtitle text to be converted;
and a subtitle image adding module 430, configured to add the target subtitle image to at least one frame of video image in the video networking conference video to obtain a target video networking conference video.
A conference video sending module 440, configured to send the target video networking conference video to a video networking recording and playing server; and the video network recording and broadcasting server is used for storing the target video network conference video and sending the target video network conference video to a video network terminal.
Optionally, the video conference video has a characteristic video format, and the video recording and playing terminal may further include:
and the image format conversion module is used for converting the target subtitle image into the target subtitle image conforming to the characteristic video format.
Optionally, the subtitle text has a plurality of corresponding candidate subtitle images, the candidate subtitle images have a first subtitle font type, and the video network recording and playing terminal may further include:
the caption control command receiving module is used for receiving a caption control command of the conference control terminal; the subtitle control command carries a second subtitle font type;
the subtitle image extracting module 420 may specifically include:
the candidate subtitle image determining submodule is used for determining a plurality of candidate subtitle images mapped by the subtitle text to be converted;
a font type matching sub-module, configured to match the second subtitle font type with the first subtitle font type of the multiple candidate subtitle images;
and the candidate subtitle image extraction sub-module is used for extracting the candidate subtitle image of which the first subtitle font type is matched with the second subtitle font type to serve as the target subtitle image.
Optionally, the subtitle control command further carries subtitle appearance characteristic information, and the video network recording and playing terminal may further include:
and the appearance characteristic adjusting module is used for adjusting the appearance characteristic of the target subtitle image according to the subtitle appearance characteristic information.
According to the embodiment of the invention, the font type and the appearance characteristic of the caption in the conference video are determined according to the caption control command, so that a user can adjust the caption in the recorded and played video network conference video according to the conference requirement, and the flexibility of caption processing is improved.
Since the processing procedure described in the device embodiment has been described in detail in the method embodiment, it is not described herein again.
In order to facilitate those skilled in the art to understand the video data processing method according to the embodiment of the present invention, the following description will be made with reference to the specific example of fig. 5. Fig. 5 is a schematic data flow diagram of a recorded broadcast conference video according to an embodiment of the present invention, and it can be seen from the diagram that, in the process of performing a video networking conference, a plurality of video networking participant terminals can send a conference video, a subtitle text and a subtitle control command to a video networking recorded broadcast terminal, the video networking recorded broadcast terminal can convert the subtitle text into a BMP bitmap in a manner of retrieving a database, then convert the BMP bitmap into YUV data, then transmit the YUV data to a DSP (Digital Signal Processing) through a shared memory, and embed the YUV data into a video image of the conference video by the DSP to obtain a target conference video, and send the target conference video to a video encoder. In addition, the video network recording and playing terminal can adjust the font type, color, transparency and the like of the subtitles in the conference video according to the subtitle control command. And finally, sending the conference video to a video network recording and broadcasting server for long-term storage.
It should be added that the above embodiments of the present invention can be applied to a communication network of a video network. The video networking is an important milestone for network development, is a real-time network, can realize high-definition video real-time transmission, and pushes a plurality of internet applications to high-definition video, and high-definition faces each other.
The video networking adopts a real-time high-definition video exchange technology, can integrate required services such as dozens of services of video, voice, pictures, characters, communication, data and the like on a system platform on a network platform, such as high-definition video conference, video monitoring, intelligent monitoring analysis, emergency command, digital broadcast television, delayed television, network teaching, live broadcast, VOD on demand, television mail, Personal Video Recorder (PVR), intranet (self-office) channels, intelligent video broadcast control, information distribution and the like, and realizes high-definition quality video broadcast through a television or a computer.
To better understand the embodiments of the present invention, the following description refers to the internet of view:
some of the technologies applied in the video networking are as follows:
network Technology (Network Technology)
Network technology innovation in video networking has improved over traditional Ethernet (Ethernet) to face the potentially enormous video traffic on the network. Unlike pure network Packet Switching (Packet Switching) or network circuit Switching (circuit Switching), the Packet Switching is adopted by the technology of the video networking to meet the Streaming requirement. The video networking technology has the advantages of flexibility, simplicity and low price of packet switching, and simultaneously has the quality and safety guarantee of circuit switching, thereby realizing the seamless connection of the whole network switching type virtual circuit and the data format.
Switching Technology (Switching Technology)
The video network adopts two advantages of asynchronism and packet switching of the Ethernet, eliminates the defects of the Ethernet on the premise of full compatibility, has end-to-end seamless connection of the whole network, is directly communicated with a user terminal, and directly bears an IP data packet. The user data does not require any format conversion across the entire network. The video networking is a higher-level form of the Ethernet, is a real-time exchange platform, can realize the real-time transmission of the whole-network large-scale high-definition video which cannot be realized by the existing Internet, and pushes a plurality of network video applications to high-definition and unification.
Server Technology (Server Technology)
The server technology on the video networking and unified video platform is different from the traditional server, the streaming media transmission of the video networking and unified video platform is established on the basis of connection orientation, the data processing capacity of the video networking and unified video platform is independent of flow and communication time, and a single network layer can contain signaling and data transmission. For voice and video services, the complexity of video networking and unified video platform streaming media processing is much simpler than that of data processing, and the efficiency is greatly improved by more than one hundred times compared with that of a traditional server.
Storage Technology (Storage Technology)
The super-high speed storage technology of the unified video platform adopts the most advanced real-time operating system in order to adapt to the media content with super-large capacity and super-large flow, the program information in the server instruction is mapped to the specific hard disk space, the media content is not passed through the server any more, and is directly sent to the user terminal instantly, and the general waiting time of the user is less than 0.2 second. The optimized sector distribution greatly reduces the mechanical motion of the magnetic head track seeking of the hard disk, the resource consumption only accounts for 20% of that of the IP internet of the same grade, but concurrent flow which is 3 times larger than that of the traditional hard disk array is generated, and the comprehensive efficiency is improved by more than 10 times.
Network Security Technology (Network Security Technology)
The structural design of the video network completely eliminates the network security problem troubling the internet structurally by the modes of independent service permission control each time, complete isolation of equipment and user data and the like, generally does not need antivirus programs and firewalls, avoids the attack of hackers and viruses, and provides a structural carefree security network for users.
Service Innovation Technology (Service Innovation Technology)
The unified video platform integrates services and transmission, and is not only automatically connected once whether a single user, a private network user or a network aggregate. The user terminal, the set-top box or the PC are directly connected to the unified video platform to obtain various multimedia video services in various forms. The unified video platform adopts a menu type configuration table mode to replace the traditional complex application programming, can realize complex application by using very few codes, and realizes infinite new service innovation.
Networking of the video network is as follows:
the video network is a centralized control network structure, and the network can be a tree network, a star network, a ring network and the like, but on the basis of the centralized control node, the whole network is controlled by the centralized control node in the network.
As shown in fig. 6, the video network is divided into an access network and a metropolitan network.
The devices of the access network part can be mainly classified into 3 types: node server, access switch, terminal (including various set-top boxes, coding boards, memories, etc.). The node server is connected to an access switch, which may be connected to a plurality of terminals and may be connected to an ethernet network.
The node server is a node which plays a centralized control function in the access network and can control the access switch and the terminal. The node server can be directly connected with the access switch or directly connected with the terminal.
Similarly, devices of the metropolitan network portion may also be classified into 3 types: a metropolitan area server, a node switch and a node server. The metro server is connected to a node switch, which may be connected to a plurality of node servers.
The node server is a node server of the access network part, namely the node server belongs to both the access network part and the metropolitan area network part.
The metropolitan area server is a node which plays a centralized control function in the metropolitan area network and can control a node switch and a node server. The metropolitan area server can be directly connected with the node switch or directly connected with the node server.
Therefore, the whole video network is a network structure with layered centralized control, and the network controlled by the node server and the metropolitan area server can be in various structures such as tree, star and ring.
The access network part can form a unified video platform (the part in the dotted circle), and a plurality of unified video platforms can form a video network; each unified video platform may be interconnected via metropolitan area and wide area video networking.
Video networking device classification
1.1 devices in the video network of the embodiment of the present invention can be mainly classified into 3 types: servers, switches (including ethernet gateways), terminals (including various set-top boxes, code boards, memories, etc.). The video network as a whole can be divided into a metropolitan area network (or national network, global network, etc.) and an access network.
1.2 wherein the devices of the access network part can be mainly classified into 3 types: node servers, access switches (including ethernet gateways), terminals (including various set-top boxes, code boards, memories, etc.).
The specific hardware structure of each access network device is as follows:
a node server:
as shown in fig. 7, the system mainly includes a network interface module 701, a switching engine module 702, a CPU module 703, and a disk array module 704;
the network interface module 701, the CPU module 703 and the disk array module 704 enter the switching engine module 702; the switching engine module 702 performs an operation of looking up the address table 705 on the incoming packet, thereby obtaining the direction information of the packet; and stores the packet in a corresponding queue of the packet buffer 706 based on the packet's steering information; if the queue of the packet buffer 706 is nearly full, discard; the switching engine module 702 polls all packet buffer queues for forwarding if the following conditions are met: 1) the port send buffer is not full; 2) the queue packet counter is greater than zero. The disk array module 704 mainly implements control over the hard disk, including initialization, read-write, and other operations; the CPU module 703 is mainly responsible for protocol processing with an access switch and a terminal (not shown in the figure), configuring an address table 705 (including a downlink protocol packet address table, an uplink protocol packet address table, and a data packet address table), and configuring the disk array module 704.
The access switch:
as shown in fig. 8, the network interface module mainly includes a network interface module (a downlink network interface module 801, an uplink network interface module 802), a switching engine module 803, and a CPU module 804;
wherein, the packet (uplink data) coming from the downlink network interface module 801 enters the packet detection module 805; the packet detection module 805 detects whether the Destination Address (DA), the Source Address (SA), the packet type, and the packet length of the packet meet the requirements, and if so, allocates a corresponding stream identifier (stream-id) and enters the switching engine module 803, otherwise, discards the stream identifier; the packet (downstream data) coming from the upstream network interface module 802 enters the switching engine module 803; the incoming data packet from the CPU module 804 enters the switching engine module 803; the switching engine module 803 performs an operation of looking up the address table 806 on the incoming packet, thereby obtaining the direction information of the packet; if the packet entering the switching engine module 803 is from the downstream network interface to the upstream network interface, the packet is stored in a queue of the corresponding packet buffer 807 in association with a stream-id; if the queue of the packet buffer 807 is nearly full, it is discarded; if the packet entering the switching engine module 803 is not from the downlink network interface to the uplink network interface, the data packet is stored in the queue of the corresponding packet buffer 807 according to the packet guiding information; if the queue of the packet buffer 807 is nearly full, it is discarded.
The switching engine module 803 polls all packet buffer queues, which in this embodiment of the invention is divided into two cases:
if the queue is from the downlink network interface to the uplink network interface, the following conditions are met for forwarding: 1) the port send buffer is not full; 2) the queued packet counter is greater than zero; 3) obtaining a token generated by a code rate control module;
if the queue is not from the downlink network interface to the uplink network interface, the following conditions are met for forwarding: 1) the port send buffer is not full; 2) the queue packet counter is greater than zero.
The rate control module 808 is configured by the CPU module 804, and generates tokens for packet buffer queues from all downlink network interfaces to uplink network interfaces at programmable intervals to control the rate of uplink forwarding.
The CPU module 804 is mainly responsible for protocol processing with the node server, configuration of the address table 806, and configuration of the code rate control module 808.
Ethernet protocol conversion gateway
As shown in fig. 9, the system mainly includes a network interface module (a downlink network interface module 901 and an uplink network interface module 902), a switching engine module 903, a CPU module 904, a packet detection module 905, a rate control module 908, an address table 906, a packet buffer 907, a MAC adding module 909, and a MAC deleting module 910.
Wherein, the data packet coming from the downlink network interface module 901 enters the packet detection module 905; the packet detection module 905 detects whether the ethernet MAC DA, the ethernet MAC SA, the ethernet length or frame type, the video network destination address DA, the video network source address SA, the video network packet type, and the packet length of the packet meet the requirements, and if so, allocates a corresponding stream identifier (stream-id); then, the MAC deleting module 910 subtracts MAC DA, MAC SA, length or frame type (2byte), and enters the corresponding receiving buffer, otherwise, discards it;
the downlink network interface module 901 detects the sending buffer of the port, and if there is a packet, obtains the ethernet MAC DA of the corresponding terminal according to the destination address DA of the packet, adds the ethernet MAC DA of the terminal, the MACSA of the ethernet coordination gateway, and the ethernet length or frame type, and sends the packet.
The other modules in the ethernet protocol gateway function similarly to the access switch.
A terminal:
the system mainly comprises a network interface module, a service processing module and a CPU module; for example, the set-top box mainly comprises a network interface module, a video and audio coding and decoding engine module and a CPU module; the coding board mainly comprises a network interface module, a video and audio coding engine module and a CPU module; the memory mainly comprises a network interface module, a CPU module and a disk array module.
1.3 devices of the metropolitan area network part can be mainly classified into 2 types: node server, node exchanger, metropolitan area server. The node switch mainly comprises a network interface module, a switching engine module and a CPU module; the metropolitan area server mainly comprises a network interface module, a switching engine module and a CPU module.
2. Video networking packet definition
2.1 Access network packet definition
The data packet of the access network mainly comprises the following parts: destination Address (DA), Source Address (SA), reserved bytes, payload (pdu), CRC.
As shown in the following table, the data packet of the access network mainly includes the following parts:
Figure BDA0001527767140000161
wherein:
the Destination Address (DA) is composed of 8 bytes (byte), the first byte represents the type of the data packet (such as various protocol packets, multicast data packets, unicast data packets, etc.), there are 256 possibilities at most, the second byte to the sixth byte are metropolitan area network addresses, and the seventh byte and the eighth byte are access network addresses;
the Source Address (SA) is also composed of 8 bytes (byte), defined as the same as the Destination Address (DA);
the reserved byte consists of 2 bytes;
the payload part has different lengths according to different types of datagrams, and is 64 bytes if the datagram is various types of protocol packets, and is 32+1024 or 1056 bytes if the datagram is a unicast packet, of course, the length is not limited to the above 2 types;
the CRC consists of 4 bytes and is calculated in accordance with the standard ethernet CRC algorithm.
2.2 metropolitan area network packet definition
The topology of a metropolitan area network is a graph and there may be 2, or even more than 2, connections between two devices, i.e., there may be more than 2 connections between a node switch and a node server, a node switch and a node switch, and a node switch and a node server. However, the metro network address of the metro network device is unique, and in order to accurately describe the connection relationship between the metro network devices, parameters are introduced in the embodiment of the present invention: a label to uniquely describe a metropolitan area network device.
In this specification, the definition of the Label is similar to that of the Label of MPLS (Multi-Protocol Label Switch), and assuming that there are two connections between the device a and the device B, there are 2 labels for the packet from the device a to the device B, and 2 labels for the packet from the device B to the device a. The label is classified into an incoming label and an outgoing label, and assuming that the label (incoming label) of the packet entering the device a is 0x0000, the label (outgoing label) of the packet leaving the device a may become 0x 0001. The network access process of the metro network is a network access process under centralized control, that is, address allocation and label allocation of the metro network are both dominated by the metro server, and the node switch and the node server are both passively executed, which is different from label allocation of MPLS, and label allocation of MPLS is a result of mutual negotiation between the switch and the server.
As shown in the following table, the data packet of the metro network mainly includes the following parts:
Figure BDA0001527767140000171
namely Destination Address (DA), Source Address (SA), Reserved byte (Reserved), tag, payload (pdu), CRC. The format of the tag may be defined by reference to the following: the tag is 32 bits with the upper 16 bits reserved and only the lower 16 bits used, and its position is between the reserved bytes and payload of the packet.
The algorithms and displays presented herein are not inherently related to any particular computer, virtual machine, or other apparatus. Various general purpose systems may also be used with the teachings herein. The required structure for constructing such a system will be apparent from the description above. Moreover, the present invention is not directed to any particular programming language. It is appreciated that a variety of programming languages may be used to implement the teachings of the present invention as described herein, and any descriptions of specific languages are provided above to disclose the best mode of the invention.
In the description provided herein, numerous specific details are set forth. It is understood, however, that embodiments of the invention may be practiced without these specific details. In some instances, well-known methods, structures and techniques have not been shown in detail in order not to obscure an understanding of this description.
Similarly, it should be appreciated that in the foregoing description of exemplary embodiments of the invention, various features of the invention are sometimes grouped together in a single embodiment, figure, or description thereof for the purpose of streamlining the disclosure and aiding in the understanding of one or more of the various inventive aspects. However, the disclosed method should not be interpreted as reflecting an intention that: that the invention as claimed requires more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive aspects lie in less than all features of a single foregoing disclosed embodiment. Thus, the claims following the detailed description are hereby expressly incorporated into this detailed description, with each claim standing on its own as a separate embodiment of this invention.
Those skilled in the art will appreciate that the modules in the device in an embodiment may be adaptively changed and disposed in one or more devices different from the embodiment. The modules or units or components of the embodiments may be combined into one module or unit or component, and furthermore they may be divided into a plurality of sub-modules or sub-units or sub-components. All of the features disclosed in this specification (including any accompanying claims, abstract and drawings), and all of the processes or elements of any method or apparatus so disclosed, may be combined in any combination, except combinations where at least some of such features and/or processes or elements are mutually exclusive. Each feature disclosed in this specification (including any accompanying claims, abstract and drawings) may be replaced by alternative features serving the same, equivalent or similar purpose, unless expressly stated otherwise.
Furthermore, those skilled in the art will appreciate that while some embodiments described herein include some features included in other embodiments, rather than other features, combinations of features of different embodiments are meant to be within the scope of the invention and form different embodiments. For example, in the following claims, any of the claimed embodiments may be used in any combination.
The various component embodiments of the invention may be implemented in hardware, or in software modules running on one or more processors, or in a combination thereof. Those skilled in the art will appreciate that a microprocessor or Digital Signal Processor (DSP) may be used in practice to implement some or all of the functions of some or all of the components in a server, terminal, or the like, according to embodiments of the present invention. The present invention may also be embodied as apparatus or device programs (e.g., computer programs and computer program products) for performing a portion or all of the methods described herein. Such programs implementing the present invention may be stored on computer-readable media or may be in the form of one or more signals. Such a signal may be downloaded from an internet website or provided on a carrier signal or in any other form.
It should be noted that the above-mentioned embodiments illustrate rather than limit the invention, and that those skilled in the art will be able to design alternative embodiments without departing from the scope of the appended claims. In the claims, any reference signs placed between parentheses shall not be construed as limiting the claim. The word "comprising" does not exclude the presence of elements or steps not listed in a claim. The word "a" or "an" preceding an element does not exclude the presence of a plurality of such elements. The invention may be implemented by means of hardware comprising several distinct elements, and by means of a suitably programmed computer. In the unit claims enumerating several means, several of these means may be embodied by one and the same item of hardware. The usage of the words first, second and third, etcetera do not indicate any ordering. These words may be interpreted as names.

Claims (6)

1. A video data processing method is applied to a video network recording and broadcasting terminal, and is characterized in that a plurality of subtitle texts and a plurality of subtitle images respectively corresponding to the subtitle texts are prestored in the video network recording and broadcasting terminal, and the method comprises the following steps:
receiving video of the video networking conference and subtitle texts to be converted; the video networking conference video comprises a plurality of frames of video images;
extracting a target subtitle image corresponding to the subtitle text to be converted;
adding the target subtitle image to at least one frame of video image in the video networking conference video to obtain a target video networking conference video;
wherein the video networking conference video has a characteristic video format, and before the step of adding the target subtitle image to at least one frame of video image in the video networking conference video to obtain the target video networking conference video, the method further comprises:
converting the target subtitle image into a target subtitle image conforming to the characteristic video format;
wherein the subtitle text has a plurality of corresponding candidate subtitle images, the candidate subtitle images have a first subtitle font type, and before the step of extracting the target subtitle image mapped by the subtitle text to be converted, the method further includes:
receiving a caption control command of a conference control terminal; the subtitle control command carries a second subtitle font type;
the step of extracting the target subtitle image corresponding to the subtitle text to be converted comprises the following steps:
determining a plurality of candidate subtitle images mapped by the subtitle text to be converted;
matching the second subtitle font type with a first subtitle font type of the candidate subtitle images;
and extracting candidate subtitle images matched with the first subtitle font type and the second subtitle font type to serve as the target subtitle images.
2. The method of claim 1, further comprising:
sending the target video networking conference video to a video networking recording and playing server; and the video network recording and broadcasting server is used for storing the target video network conference video and sending the target video network conference video to a video network terminal.
3. The method of claim 1, wherein the subtitle control command further carries subtitle appearance characteristic information, the method further comprising:
and adjusting the appearance characteristics of the target subtitle image according to the subtitle appearance characteristic information.
4. The utility model provides a video networking recorded broadcast terminal which characterized in that, a plurality of subtitle texts and a plurality of subtitle images that correspond respectively rather than prestored in video networking recorded broadcast terminal, the terminal includes:
the video and text receiving module is used for receiving video of the video networking conference and the subtitle text to be converted; the video networking conference video comprises a plurality of frames of video images;
the subtitle image extraction module is used for extracting a target subtitle image corresponding to the subtitle text to be converted;
the subtitle image adding module is used for adding the target subtitle image into at least one frame of video image in the video networking conference video to obtain a target video networking conference video;
wherein, the video networking conference video has a characteristic video format, and the terminal further comprises:
the image format conversion module is used for converting the target subtitle image into a target subtitle image conforming to the characteristic video format;
wherein the subtitle text has a plurality of corresponding candidate subtitle images having a first subtitle font type, the terminal further comprising:
the caption control command receiving module is used for receiving a caption control command of the conference control terminal; the subtitle control command carries a second subtitle font type;
the subtitle image extraction module comprises:
the candidate subtitle image determining submodule is used for determining a plurality of candidate subtitle images mapped by the subtitle text to be converted;
a font type matching sub-module, configured to match the second subtitle font type with the first subtitle font type of the multiple candidate subtitle images;
and the candidate subtitle image extraction sub-module is used for extracting the candidate subtitle image of which the first subtitle font type is matched with the second subtitle font type to serve as the target subtitle image.
5. The terminal of claim 4, further comprising:
the conference video sending module is used for sending the target video networking conference video to a video networking recording and playing server; and the video network recording and broadcasting server is used for storing the target video network conference video and sending the target video network conference video to a video network terminal.
6. The terminal according to claim 4, wherein the caption control command further carries caption appearance characteristic information, the terminal further comprising:
and the appearance characteristic adjusting module is used for adjusting the appearance characteristic of the target subtitle image according to the subtitle appearance characteristic information.
CN201711447037.9A 2017-12-27 2017-12-27 Video data processing method and video network recording and playing terminal Active CN108965783B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711447037.9A CN108965783B (en) 2017-12-27 2017-12-27 Video data processing method and video network recording and playing terminal

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711447037.9A CN108965783B (en) 2017-12-27 2017-12-27 Video data processing method and video network recording and playing terminal

Publications (2)

Publication Number Publication Date
CN108965783A CN108965783A (en) 2018-12-07
CN108965783B true CN108965783B (en) 2020-05-26

Family

ID=64495671

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711447037.9A Active CN108965783B (en) 2017-12-27 2017-12-27 Video data processing method and video network recording and playing terminal

Country Status (1)

Country Link
CN (1) CN108965783B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109982027A (en) * 2019-02-26 2019-07-05 视联动力信息技术股份有限公司 A kind of method of adjustment, the device and system of Subtitle Demonstration parameter
CN111327961A (en) * 2020-03-30 2020-06-23 上海句石智能科技有限公司 Video subtitle switching method and system
CN113518248A (en) * 2020-04-09 2021-10-19 西安诺瓦星云科技股份有限公司 Program publishing method, device and system

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1218574C (en) * 2001-10-15 2005-09-07 华为技术有限公司 Interactive video equipment and its caption superposition method
CN101068345B (en) * 2007-05-24 2010-06-23 杭州华三通信技术有限公司 Video frequency monitoring method and system and network transmission equipment
JP5201050B2 (en) * 2009-03-27 2013-06-05 ブラザー工業株式会社 Conference support device, conference support method, conference system, conference support program
CN101867733A (en) * 2009-04-14 2010-10-20 联发科技(新加坡)私人有限公司 Processing method of subtitle data stream of video programme and video displaying system
CN101710961A (en) * 2009-12-09 2010-05-19 中兴通讯股份有限公司 Control method and device for generating title in video conference
CN102006453B (en) * 2010-11-30 2013-08-07 华为终端有限公司 Superposition method and device for auxiliary information of video signals
CN103686219B (en) * 2012-09-24 2017-09-29 华为技术有限公司 A kind of method, equipment and the system of video conference recorded broadcast
CN102892032B (en) * 2012-11-02 2015-06-03 湖南正海智慧网真设备有限公司 Real-time interactive high definition network video communication system

Also Published As

Publication number Publication date
CN108965783A (en) 2018-12-07

Similar Documents

Publication Publication Date Title
CN108737768B (en) Monitoring method and monitoring device based on monitoring system
CN110166728B (en) Video networking conference opening method and device
CN108965224B (en) Video-on-demand method and device
CN108881815B (en) Video data transmission method and device
CN109547163B (en) Method and device for controlling data transmission rate
CN110166433B (en) Method and system for acquiring video data
CN108965227B (en) Data processing method and video networking conference server
CN109788235B (en) Video networking-based conference recording information processing method and system
CN108965930B (en) Video data processing method and device
CN110769310A (en) Video processing method and device based on video network
CN108965783B (en) Video data processing method and video network recording and playing terminal
CN108810457B (en) Method and system for controlling video network monitoring camera
CN110113564B (en) Data acquisition method and video networking system
CN110149305B (en) Video network-based multi-party audio and video playing method and transfer server
CN109743284B (en) Video processing method and system based on video network
CN109905616B (en) Method and device for switching video pictures
CN110769179B (en) Audio and video data stream processing method and system
CN108874844B (en) Form data processing method and video network server
CN110769297A (en) Audio and video data processing method and system
CN110493149B (en) Message processing method and device
CN110022500B (en) Packet loss processing method and device
CN110086773B (en) Audio and video data processing method and system
CN109889516B (en) Method and device for establishing session channel
CN108574655B (en) Conference monitoring and broadcasting method and device
CN110661749A (en) Video signal processing method and video networking terminal

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information

Address after: 100000 Dongcheng District, Beijing, Qinglong Hutong 1, 1103 house of Ge Hua building.

Applicant after: Video Link Power Information Technology Co., Ltd.

Address before: 100000 Beijing Dongcheng District Qinglong Hutong 1 Song Hua Building A1103-1113

Applicant before: BEIJING VISIONVERA INTERNATIONAL INFORMATION TECHNOLOGY CO., LTD.

CB02 Change of applicant information
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right

Effective date of registration: 20201230

Address after: 571924 building C07, Zone C, Hainan Ecological Software Park, hi tech Industrial Demonstration Zone, old town, Haikou City, Hainan Province

Patentee after: Hainan Shilian Communication Technology Co.,Ltd.

Address before: 100000 Dongcheng District, Beijing, Qinglong Hutong 1, 1103 house of Ge Hua building.

Patentee before: VISIONVERA INFORMATION TECHNOLOGY Co.,Ltd.

TR01 Transfer of patent right